Skip to content
View efsotr's full-sized avatar

Block or report efsotr

Report abuse

Contact GitHub support about this user’s behavior. Learn more about reporting abuse.

Report abuse

Popular repositories Loading

  1. flash-attention-w-tree-attn flash-attention-w-tree-attn Public

    Forked from Dao-AILab/flash-attention

    Fast and memory-efficient exact attention

    Python 4

  2. nano-patch-sequence-pack nano-patch-sequence-pack Public

    Just a few lines to combine 🤗 Transformers, Flash Attention 2, and torch.compile — simple, clean, fast ⚡

    Python 2

  3. nano-dpo nano-dpo Public

    A minimal implementation of Direct Preference Optimization (DPO) in Chinese

    Jupyter Notebook 1

  4. fast-hadamard-transform fast-hadamard-transform Public

    Forked from Dao-AILab/fast-hadamard-transform

    Fast Hadamard transform in CUDA, with a PyTorch interface

    C 1

  5. Megatron-LM-NEO Megatron-LM-NEO Public

    Python

  6. transformers transformers Public

    Forked from huggingface/transformers

    🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.

    Python