xformers icon indicating copy to clipboard operation
xformers copied to clipboard

Hackable and optimized Transformers building blocks, supporting a composable construction.

Results 158 xformers issues
Sort by recently updated
recently updated
newest added

在运行CUDA时报告: WARNING[XFORMERS]: xFormers can't load C++/CUDA extensions. xFormers was built for: PyTorch 2.1.0+cu118 with CUDA 1106 (you have 2.0.1+cu118) Python 3.9.16 (you have 3.10.12) Please reinstall xformers (see https://github.com/facebookresearch/xformers#installing-xformers) Memory-efficient...

# ❓ Questions and Help I've been trying to install xFormers in a auxilliary Kaggle notebook (in the /kaggle/working folder) which will be used as a utility script in another...

# 🐛 Bug RuntimeError: unsupported output type: int, from operator: xformers::efficient_attention_forward_cutlass ## Command import torch import xformers.ops as xops import torch.nn as nn class Attention(nn.Module): def forward(self, q,k,v): x =...

# ❓ Questions and Help memory_efficient_attention fw produce inconsistent results not sure what was going on? incorrect built? some specific versions combinations? for some combinations: xformers torch CUDA GPU CUDA...

## What does this PR do? Fixes a performance regress for FA decode ## Before submitting - [x] Did you have fun? - [x] Did you read the [contributor guideline](https://github.com/facebookresearch/xformers/blob/master/CONTRIBUTING.md)?...

CLA Signed

A small PR. This PR addresses the issue of disabling fused swiglu in swiglu_op.py. The original code directly disabled the support for fused swiglu, with the comment "Let's disable autocast...

CLA Signed

# 🐛 Bug ## Command ## To Reproduce Steps to reproduce the behavior: 1. 2. 3. ## Expected behavior ## Environment Please copy and paste the output from the environment...

# 🚀 Feature I think that it would be very helpful to allow sparse tensors to be used as masks (`attn_bias`) in the `memory_efficient_attention` function. ## Motivation As the documentation...