Stars
Code for the paper: "Fine-Tuning Discrete Diffusion Models with Policy Gradient Methods"
A library that scrapes Linkedin for user data
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention & RFT)
S-LoRA: Serving Thousands of Concurrent LoRA Adapters
Fast and memory-efficient exact attention
PyTorch implementation of FractalGen https://arxiv.org/abs/2502.17437
🔥 A minimal training framework for scaling FLA models
🚀 Efficient implementations of state-of-the-art linear attention models in Torch and Triton
The official implementation of Self-Play Preference Optimization (SPPO)
A PyTorch library for all things Reinforcement Learning (RL) for Combinatorial Optimization (CO)
Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"
[RSS 2024] 3D Diffusion Policy: Generalizable Visuomotor Policy Learning via Simple 3D Representations
Build and share delightful machine learning apps, all in Python. 🌟 Star to support our work!
Autoregressive Model Beats Diffusion: 🦙 Llama for Scalable Image Generation
Official Code for Paper "Think While You Generate: Discrete Diffusion with Planned Denoising" [ICLR 2025]
Google Research
Elucidating the Design Space of Diffusion-Based Generative Models (EDM)
A Flexible Framework for Experiencing Cutting-edge LLM Inference Optimizations
RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose". So it's combining the best of RN…
[NeurIPS 2024 Best Paper][GPT beats diffusion🔥] [scaling laws in visual generation📈] Official impl. of "Visual Autoregressive Modeling: Scalable Image Generation via Next-Scale Prediction". An *ult…
A high-throughput and memory-efficient inference and serving engine for LLMs
A brief and partial summary of RLHF algorithms.
[ICLR 2025] Rectified Diffusion: Straightness Is Not Your Need
Official Pytorch repository for Extreme Compression of Large Language Models via Additive Quantization https://arxiv.org/pdf/2401.06118.pdf and PV-Tuning: Beyond Straight-Through Estimation for Ext…
[ICML 2024] CLLMs: Consistency Large Language Models
WikiChat is an improved RAG. It stops the hallucination of large language models by retrieving data from a corpus.
verl: Volcano Engine Reinforcement Learning for LLMs
[ICML 2024 Best Paper] Discrete Diffusion Modeling by Estimating the Ratios of the Data Distribution (https://arxiv.org/abs/2310.16834)