Achieve the llama3 inference step-by-step, grasp the core concepts, master the process derivation, implement the code.
-
Updated
Feb 24, 2025 - Jupyter Notebook
Achieve the llama3 inference step-by-step, grasp the core concepts, master the process derivation, implement the code.
Notes about LLaMA 2 model
One Diffusion model implementation base on LibTorch
PyTorch implementation of Rotary Spatial Embeddings
A production-grade LLM architecture built from scratch in PyTorch. Features Multi-Head Latent Attention (MLA), Mixture of Experts (MoE), GRPO alignment, and a complete 31-part educational course.
Add a description, image, and links to the rotary-position-encoding topic page so that developers can more easily learn about it.
To associate your repository with the rotary-position-encoding topic, visit your repo's landing page and select "manage topics."