Stars
AutoGPT is the vision of accessible AI for everyone, to use and to build on. Our mission is to provide the tools, so that you can focus on what matters.
Stable Diffusion web UI
🧑🏫 60+ Implementations/tutorials of deep learning papers with side-by-side notes 📝; including transformers (original, xl, switch, feedback, vit, ...), optimizers (adam, adabelief, sophia, ...), ga…
CLI platform to experiment with codegen. Precursor to: https://lovable.dev
🌟 The Multi-Agent Framework: First AI Software Company, Towards Natural Language Programming
Unified Efficient Fine-Tuning of 100+ LLMs & VLMs (ACL 2024)
ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型
Making large AI models cheaper, faster and more accessible
High-Resolution Image Synthesis with Latent Diffusion Models
TensorFlow code and pre-trained models for BERT
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Code and documentation to train Stanford's Alpaca models, and generate the data.
🤗 Diffusers: State-of-the-art diffusion models for image, video, and audio generation in PyTorch and FLAX.
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
中文LLaMA&Alpaca大语言模型+本地CPU/GPU训练部署 (Chinese LLaMA & Alpaca LLMs)
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
DALL·E Mini - Generate images from a text prompt
Implementation of DALL-E 2, OpenAI's updated text-to-image synthesis neural network, in Pytorch
LLMs build upon Evol Insturct: WizardLM, WizardCoder, WizardMath
Implementation of Imagen, Google's Text-to-Image Neural Network, in Pytorch
Code for the paper "Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer"
Firefly: 大模型训练工具,支持训练Qwen2.5、Qwen2、Yi1.5、Phi-3、Llama3、Gemma、MiniCPM、Yi、Deepseek、Orion、Xverse、Mixtral-8x7B、Zephyr、Mistral、Baichuan2、Llma2、Llama、Qwen、Baichuan、ChatGLM2、InternLM、Ziya2、Vicuna、Bloom等大模型
OpenMMLab Pose Estimation Toolbox and Benchmark.
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
A large-scale 7B pretraining language model developed by BaiChuan-Inc.
An Open-source Framework for Data-centric, Self-evolving Autonomous Language Agents