├── .clang_format.hook ├── .gitignore ├── .pre-commit-config.yaml ├── LICENSE ├── README.md ├── einops ├── __init__.py ├── _backends.py ├── _torch_specific.py ├── einops.py ├── layers │ ├── __init__.py │ ├── _einmix.py │ ├── chainer.py │ ├── gluon.py │ ├── keras.py │ ├── paddle.py │ ├── tensorflow.py │ └── torch.py └── parsing.py ├── images ├── all-attention.png ├── attention-on-attention.png ├── collaborative-attention.png ├── cosine-sim-attention.png ├── enhanced-recurrence.png ├── ffglu.png ├── gate_values.png ├── gating.png ├── macaron-1.png ├── macaron-2.png ├── memory-transformer.png ├── normformer.png ├── pia.png ├── residual_attn.png ├── rezero.png ├── rotary.png ├── sandwich-2.png ├── sandwich.png ├── sandwich_norm.png ├── scalenorm.png ├── talking-heads.png └── topk-attention.png └── pd_x_transformers ├── __init__.py ├── autoregressive_wrapper.py ├── entmax15.py ├── entmax_bisect.py └── x_transformers.py /.clang_format.hook: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/.clang_format.hook -------------------------------------------------------------------------------- /.gitignore: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/.gitignore -------------------------------------------------------------------------------- /.pre-commit-config.yaml: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/.pre-commit-config.yaml -------------------------------------------------------------------------------- /LICENSE: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/LICENSE -------------------------------------------------------------------------------- /README.md: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/README.md -------------------------------------------------------------------------------- /einops/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/einops/__init__.py -------------------------------------------------------------------------------- /einops/_backends.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/einops/_backends.py -------------------------------------------------------------------------------- /einops/_torch_specific.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/einops/_torch_specific.py -------------------------------------------------------------------------------- /einops/einops.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/einops/einops.py -------------------------------------------------------------------------------- /einops/layers/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/einops/layers/__init__.py -------------------------------------------------------------------------------- /einops/layers/_einmix.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/einops/layers/_einmix.py -------------------------------------------------------------------------------- /einops/layers/chainer.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/einops/layers/chainer.py -------------------------------------------------------------------------------- /einops/layers/gluon.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/einops/layers/gluon.py -------------------------------------------------------------------------------- /einops/layers/keras.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/einops/layers/keras.py -------------------------------------------------------------------------------- /einops/layers/paddle.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/einops/layers/paddle.py -------------------------------------------------------------------------------- /einops/layers/tensorflow.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/einops/layers/tensorflow.py -------------------------------------------------------------------------------- /einops/layers/torch.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/einops/layers/torch.py -------------------------------------------------------------------------------- /einops/parsing.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/einops/parsing.py -------------------------------------------------------------------------------- /images/all-attention.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/images/all-attention.png -------------------------------------------------------------------------------- /images/attention-on-attention.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/images/attention-on-attention.png -------------------------------------------------------------------------------- /images/collaborative-attention.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/images/collaborative-attention.png -------------------------------------------------------------------------------- /images/cosine-sim-attention.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/images/cosine-sim-attention.png -------------------------------------------------------------------------------- /images/enhanced-recurrence.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/images/enhanced-recurrence.png -------------------------------------------------------------------------------- /images/ffglu.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/images/ffglu.png -------------------------------------------------------------------------------- /images/gate_values.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/images/gate_values.png -------------------------------------------------------------------------------- /images/gating.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/images/gating.png -------------------------------------------------------------------------------- /images/macaron-1.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/images/macaron-1.png -------------------------------------------------------------------------------- /images/macaron-2.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/images/macaron-2.png -------------------------------------------------------------------------------- /images/memory-transformer.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/images/memory-transformer.png -------------------------------------------------------------------------------- /images/normformer.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/images/normformer.png -------------------------------------------------------------------------------- /images/pia.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/images/pia.png -------------------------------------------------------------------------------- /images/residual_attn.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/images/residual_attn.png -------------------------------------------------------------------------------- /images/rezero.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/images/rezero.png -------------------------------------------------------------------------------- /images/rotary.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/images/rotary.png -------------------------------------------------------------------------------- /images/sandwich-2.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/images/sandwich-2.png -------------------------------------------------------------------------------- /images/sandwich.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/images/sandwich.png -------------------------------------------------------------------------------- /images/sandwich_norm.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/images/sandwich_norm.png -------------------------------------------------------------------------------- /images/scalenorm.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/images/scalenorm.png -------------------------------------------------------------------------------- /images/talking-heads.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/images/talking-heads.png -------------------------------------------------------------------------------- /images/topk-attention.png: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/images/topk-attention.png -------------------------------------------------------------------------------- /pd_x_transformers/__init__.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/pd_x_transformers/__init__.py -------------------------------------------------------------------------------- /pd_x_transformers/autoregressive_wrapper.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/pd_x_transformers/autoregressive_wrapper.py -------------------------------------------------------------------------------- /pd_x_transformers/entmax15.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/pd_x_transformers/entmax15.py -------------------------------------------------------------------------------- /pd_x_transformers/entmax_bisect.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/pd_x_transformers/entmax_bisect.py -------------------------------------------------------------------------------- /pd_x_transformers/x_transformers.py: -------------------------------------------------------------------------------- https://raw.githubusercontent.com/JunnYu/x-transformers-paddle/HEAD/pd_x_transformers/x_transformers.py --------------------------------------------------------------------------------