GitHub - facebookresearch/xformers: Hackable and optimized Transformers building blocks, supporting a composable construction. (original) (raw)

Open in Colab
CircleCI Codecov black
PRs welcome


xFormers - Toolbox to Accelerate Research on Transformers

xFormers is:

Installing xFormers

[linux & win] cuda 12.6 version

pip3 install -U xformers --index-url https://download.pytorch.org/whl/cu126

[linux & win] cuda 12.8 version

pip3 install -U xformers --index-url https://download.pytorch.org/whl/cu128

[linux & win] cuda 13.0 version

pip3 install -U xformers --index-url https://download.pytorch.org/whl/cu130

[linux only] (EXPERIMENTAL) rocm 7.1 version

pip3 install -U xformers --index-url https://download.pytorch.org/whl/rocm7.1

Same requirements as for the stable version above

pip install --pre -U xformers

(Optional) Makes the build much faster

pip install ninja

Set TORCH_CUDA_ARCH_LIST if running and building on different GPU types

NOTE: pytorch must already be installed!

pip install -v --no-build-isolation -U git+https://github.com/facebookresearch/xformers.git@main#egg=xformers

(this can take dozens of minutes)

Benchmarks

Memory-efficient MHA Benchmarks for ViTS Setup: A100 on f16, measured total time for a forward+backward pass

Note that this is exact attention, not an approximation, just by calling xformers.ops.memory_efficient_attention

More benchmarks

xFormers provides many components, and more benchmarks are available in BENCHMARKS.md.

(Optional) Testing the installation

This command will provide information on an xFormers installation, and what kernels are built/available:

Using xFormers

Key Features

  1. Optimized building blocks, beyond PyTorch primitives
    1. Memory-efficient exact attention - up to 10x faster
    2. sparse attention
    3. block-sparse attention
    4. fused softmax
    5. fused linear layer
    6. fused layer norm
    7. fused dropout(activation(x+bias))
    8. fused SwiGLU

Install troubleshooting

License

xFormers has a BSD-style license, as found in the LICENSE file. It includes code from the triton-lang/kernels repo.

Citing xFormers

If you use xFormers in your publication, please cite it by using the following BibTeX entry.

@Misc{xFormers2022, author = {Benjamin Lefaudeux and Francisco Massa and Diana Liskovich and Wenhan Xiong and Vittorio Caggiano and Sean Naren and Min Xu and Jieru Hu and Marta Tintore and Susan Zhang and Patrick Labatut and Daniel Haziza and Luca Wehrstedt and Jeremy Reizenstein and Grigory Sizov}, title = {xFormers: A modular and hackable Transformer modelling library}, howpublished = {\url{https://github.com/facebookresearch/xformers}}, year = {2022} }

Credits

The following repositories are used in xFormers, either in close to original form or as an inspiration: