Achieve the llama3 inference step-by-step, grasp the core concepts, master the process derivation, implement the code.
-
Updated
Feb 24, 2025 - Jupyter Notebook
Achieve the llama3 inference step-by-step, grasp the core concepts, master the process derivation, implement the code.
This repository contains an implementation of the LLaMA 2 (Large Language Model Meta AI) model, a Generative Pretrained Transformer (GPT) variant. The implementation focuses on the model architecture and the inference process. The code is restructured and heavily commented to facilitate easy understanding of the key parts of the architecture.
Image Captioning With MobileNet-LLaMA 3
Criando um modelo Transformer do zero com variações como Multi-Head Attention e Grouped Query Attention em livros de Machado de Assis.
Criando um modelo Transformer do zero com Positional Encoding / Posições treináveis, MultiHead Attention, KV Cache e Grouped Attention com alguns livros brasileiros.
Add a description, image, and links to the rms-norm topic page so that developers can more easily learn about it.
To associate your repository with the rms-norm topic, visit your repo's landing page and select "manage topics."