Highlights
- Pro
Stars
A bibliography and survey of the papers surrounding o1
[NeurIPS 2024] Official Repository of The Mamba in the Llama: Distilling and Accelerating Hybrid Models
Official PyTorch Implementation of "SiT: Exploring Flow and Diffusion-based Generative Models with Scalable Interpolant Transformers"
Efficient implementations of state-of-the-art linear attention models in Pytorch and Triton
Autoregressive Model Beats Diffusion: 🦙 Llama for Scalable Image Generation
Following master Karpathy with GPT-2 implementation and training, writing lots of comments cause I have memory of a goldfish
Open-Sora: Democratizing Efficient Video Production for All
Cambrian-1 is a family of multimodal LLMs with a vision-centric design.
Official implementation of "Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling"
Reading list for research topics in state-space models
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Latte: Latent Diffusion Transformer for Video Generation.
Simple, minimal implementation of the Mamba SSM in one file of PyTorch.
utilities for decoding deep representations (like sentence embeddings) back to text
FlashFFTConv: Efficient Convolutions for Long Sequences with Tensor Cores
Official inference library for Mistral models
Code for Fast Training of Diffusion Models with Masked Transformers
Chapyter: ChatGPT Code Interpreter in Jupyter Notebooks
Generative Models by Stability AI
An open-source visual programming environment for battle-testing prompts to LLMs.
Implementation of MEGABYTE, Predicting Million-byte Sequences with Multiscale Transformers, in Pytorch