This CUDA extension implements fused dropout + residual + LayerNorm from the flash-attention repo.

Downloads last month
89
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support
danieldk/wrong-sha1 · Hugging Face

This CUDA extension implements fused dropout + residual + LayerNorm from the flash-attention repo.

Downloads last month
89
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support