You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Implements pre-training, supervised fine-tuning (SFT), and reinforcement learning from human feedback (RLHF), to train and fine-tune the LLaMA2 model to follow human instructions, similar to InstructGPT or ChatGPT, but on a much smaller scale.
is a comprehensive, educational project dedicated to building a Large Language Model (LLM) from the ground up. It serves as the official code repository for the book Build a Large Language Model (From Scratch), guiding developers step-by-step through the process of developing, pretraining, finetuning, and aligning a GPT-like LLM using PyTorch.