-
Anyscale
- San Francisco
-
07:50
(UTC -07:00) - https://www.linkedin.com/in/yunxuan-xiao/
Highlights
- Pro
Stars
Efficient Triton Kernels for LLM Training
Large Language Model (LLM) Systems Paper List
FlashInfer: Kernel Library for LLM Serving
PyTorch native quantization and sparsity for training and inference
LLM training code for Databricks foundation models
VILA is a family of state-of-the-art vision language models (VLMs) for diverse multimodal AI tasks across the edge, data center, and cloud.
Code release for "Learning Video Representations from Large Language Models"
CoreNet: A library for training deep neural networks
Letta (formerly MemGPT) is the stateful agents framework with memory, reasoning, and context management.
A general and accurate MACs / FLOPs profiler for PyTorch models
LAVIS - A One-stop Library for Language-Vision Intelligence
PyTorch - FID calculation with proper image resizing and quantization steps [CVPR 2022]
Minimalistic large language model 3D-parallelism training
Dify is an open-source LLM app development platform. Dify's intuitive interface combines AI workflow, RAG pipeline, agent capabilities, model management, observability features and more, letting yo…
Welcome to the Llama Cookbook! This is your go to guide for Building with Llama: Getting started with Inference, Fine-Tuning, RAG. We also show you how to solve end to end problems using Llama mode…
Latency and Memory Analysis of Transformer Models for Training and Inference
DLRover: An Automatic Distributed Deep Learning System
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
It counts how many times your GitHub profile has been viewed. Free cloud micro-service.
A high-throughput and memory-efficient inference and serving engine for LLMs
Tune efficiently any LLM model from HuggingFace using distributed training (multiple GPU) and DeepSpeed. Uses Ray AIR to orchestrate the training on multiple AWS GPU instances
🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (including fp8), and easy-to-configure FSDP and DeepSpeed support
Flexible components pairing 🤗 Transformers with ⚡ Pytorch Lightning