facebookresearch/xformers
Hackable and optimized Transformers building blocks, supporting a composable construction.
PythonNOASSERTION
Pinned issues
Issues
- 0
Can't install xformers
#1148 opened by GUEST-1001 - 6
Are there any version for Torch 2.5 including dev? like 0.0.29.dev921?
#1121 opened by FurkanGozukara - 0
Customization of BlockDiagonalMask or LowerTriangularMask
#1147 opened by kc334 - 1
Can I use xformers with torch-2.4.1?
#1146 opened by LukeLIN-web - 1
scaled_dot_product_attention output is different from memory_efficient_attention
#1114 opened by aenoca - 1
[Bug] Unexpected behavior of `memory_efficient_attention` with `BlockDiagonalMask`
#1122 opened by xiangxu-google - 1
incorrect causal mask in global attention
#1135 opened by davidqqq - 1
Is there an efficient way to use memory_efficient_attention with a causal mask that has a small rectangle of zeros?
#1131 opened by arilato - 1
RuntimeError: CUDA error: CUBLAS_STATUS_NOT_INITIALIZED when calling `cublasCreate(handle)`
#1144 opened by antony-frolov - 1
build without flash attn?
#1145 opened by sipie800 - 5
Cant use powershell command to build for Nightly PyTorch on Win - filename too long
#1132 opened by Mescalamba - 0
Generalization of dual_gemm_silu_identity_mul to use custom activation function, not only SiLU
#1140 opened by warpuv - 1
# ❓ Questions and Help NotImplementedError: No operator found for `memory_efficient_attention_forward` with inputs:
#1143 opened by Was-11-wendy - 1
How to install to aarch64
#1142 opened by gongchangsui - 11
Hey, you! The people that made this thing. You know that Torch v 2.4.1 is out, right??
#1096 opened by Xephier102 - 4
torch2.5.1 support
#1137 opened by luohao123 - 1
Cann't install xformers==0.0.28.post2
#1139 opened by Wiselnn570 - 4
vllm 0.6.3 createLLM error TypeError: autotune() got an unexpected keyword argument 'use_cuda_graph' on windows
#1138 opened by xiezhipeng-git - 0
Activation checkpointing is not working on SwiGLU
#1126 opened by warpuv - 1
timm.models.layers is deprecated/resume_download is deprecated
#1134 opened by zd391 - 2
AMD64 version of Xformers
#1130 opened by pandayummy - 1
AttributeError: module 'xformers.ops' has no attribute 'AttentionOpDispatch'
#1129 opened by LarsDoorenbos - 5
I tried upgrade visual studio,cuda 12.4 and 12.6,but it still can't build the wheel,and return this
#1117 opened by neutronslime - 3
`ScaledDotProduct` with attention mask returns different result as standard attention
#1089 opened by amyxlu - 0
- 0
- 1
Incorrect attention output with SparseCS mask
#1124 opened by francois-rozet - 0
- 2
CUTLASS Fused multi head attention
#1112 opened by yoon5862 - 14
- 7
Why xformers 0.0.28.post1 doesn't have pre-compiled wheel for Windows?
#1118 opened by FurkanGozukara - 0
Does memory efficient attention cutlass kernel support various seq len inputs for q/k/v + tensor bias?
#1120 opened by ShijunK - 2
- 2
no module 'xformers'. Processing without...
#1102 opened by ZeroCool22 - 1
Enabling softcap option
#1113 opened by SpyrosMouselinos - 1
what version about torch in FLUX bf16?
#1111 opened by Zhuangvictor0 - 4
compile for rocm w/ gfx1032 card
#1110 opened by brcisna - 5
unable to pip install 0.0.28.post1 on windows
#1104 opened by ajkessel - 1
- 2
`memory_efficient_attention` is slower than `scaled_dot_product_attention` of PyTorch?
#1107 opened by QinlongHuang - 3
Prebuilt wheel for Windows
#1098 opened by KohakuBlueleaf - 1
xformers Installation failed
#1095 opened by wzgrx - 1
why conflicting release notes?
#1097 opened by BBC-Esq - 2
BlockDiagonalGappyKeysMask backward support
#1093 opened by cumulo-autumn - 0
flash attn bug
#1090 opened by zhw-zhang - 2
Sparse attention will not reduce peak memory usage
#1086 opened by ThisisBillhe - 6
- 14
- 0
How to get Q @ K^T similarity?
#1080 opened by volcverse - 1
Local attention mask size mismatch
#1076 opened by samuelwheeler