Skip to content
@Dao-AILab

Dao AI Lab

We are an AI research group led by Prof. Tri Dao

Popular repositories Loading

  1. flash-attention flash-attention Public

    Fast and memory-efficient exact attention

    Python 17.1k 1.6k

  2. causal-conv1d causal-conv1d Public

    Causal depthwise conv1d in CUDA, with a PyTorch interface

    Cuda 440 88

  3. fast-hadamard-transform fast-hadamard-transform Public

    Fast Hadamard transform in CUDA, with a PyTorch interface

    C 181 23

  4. gemm-cublas gemm-cublas Public

    Python 15

  5. cutlass cutlass Public

    Forked from NVIDIA/cutlass

    CUDA Templates for Linear Algebra Subroutines

    C++ 1

Repositories

Showing 5 of 5 repositories
  • gemm-cublas Public
    Dao-AILab/gemm-cublas’s past year of commit activity
    Python 15 Apache-2.0 0 0 0 Updated Apr 26, 2025
  • flash-attention Public

    Fast and memory-efficient exact attention

    Dao-AILab/flash-attention’s past year of commit activity
    Python 17,137 BSD-3-Clause 1,647 701 58 Updated Apr 25, 2025
  • cutlass Public Forked from NVIDIA/cutlass

    CUDA Templates for Linear Algebra Subroutines

    Dao-AILab/cutlass’s past year of commit activity
    C++ 1 1,215 0 0 Updated Apr 4, 2025
  • causal-conv1d Public

    Causal depthwise conv1d in CUDA, with a PyTorch interface

    Dao-AILab/causal-conv1d’s past year of commit activity
    Cuda 440 BSD-3-Clause 88 22 6 Updated Dec 6, 2024
  • fast-hadamard-transform Public

    Fast Hadamard transform in CUDA, with a PyTorch interface

    Dao-AILab/fast-hadamard-transform’s past year of commit activity
    C 181 BSD-3-Clause 23 6 2 Updated May 24, 2024

Top languages

Loading…

Most used topics

Loading…