What's new
Warez.Ge

This is a sample guest message. Register a free account today to become a member! Once signed in, you'll be able to participate on this site by adding your own topics and posts, as well as connect with other members through your own private inbox!

Code with the Author of Build an LLM (From Scratch) by Sebastian Raschka

voska89

Moderator
Staff member
Top Poster Of Month
e71e7d7c5e97fca2e2a56d455c69e1c0.webp

Free Download Code with the Author of Build an LLM (From Scratch) by Sebastian Raschka
Released 5/2025
By Sebastian Raschka
MP4 | Video: h264, 1280x720 | Audio: AAC, 44.1 KHz, 2 Ch
Genre: eLearning | Language: English | Duration: 13h 35m | Size: 2.72 GB​

Master the inner workings of how large language models like GPT really work with hands-on coding sessions led by bestselling author Sebastian Raschka. These companion videos to Build a Large Language Model from Scratch walk you through real-world implementation, with each session ending in a "test yourself" challenge to solidify your skills and deepen your understanding.
Table of contents
Chapter 1. Python Environment Setup
Chapter 2. Tokenizing text
Chapter 2. Converting tokens into token IDs
Chapter 2. Adding special context tokens
Chapter 2. Byte pair encoding
Chapter 2. Data sampling with a sliding window
Chapter 2. Creating token embeddings
Chapter 2. Encoding word positions
Chapter 3. A simple self-attention mechanism without trainable weights | Part 1
Chapter 3. A simple self-attention mechanism without trainable weights | Part 2
Chapter 3. Computing the attention weights step by step
Chapter 3. Implementing a compact self-attention Python class
Chapter 3. Applying a causal attention mask
Chapter 3. Masking additional attention weights with dropout
Chapter 3. Implementing a compact causal self-attention class
Chapter 3. Stacking multiple single-head attention layers
Chapter 3. Implementing multi-head attention with weight splits
Chapter 4. Coding an LLM architecture
Chapter 4. Normalizing activations with layer normalization
Chapter 4. Implementing a feed forward network with GELU activations
Chapter 4. Adding shortcut connections
Chapter 4. Connecting attention and linear layers in a transformer block
Chapter 4. Coding the GPT model
Chapter 4. Generating text
Chapter 5. Using GPT to generate text
Chapter 5. Calculating the text generation loss: cross entropy and perplexity
Chapter 5. Calculating the training and validation set losses
Chapter 5. Training an LLM
Chapter 5. Decoding strategies to control randomness
Chapter 5. Temperature scaling
Chapter 5. Top-k sampling
Chapter 5. Modifying the text generation function
Chapter 5. Loading and saving model weights in PyTorch
Chapter 5. Loading pretrained weights from OpenAI
Chapter 6. Preparing the dataset
Chapter 6. Creating data loaders
Chapter 6. Initializing a model with pretrained weights
Chapter 6. Adding a classification head
Chapter 6. Calculating the classification loss and accuracy
Chapter 6. Fine-tuning the model on supervised data
Chapter 6. Using the LLM as a spam classifier
Chapter 7. Preparing a dataset for supervised instruction fine-tuning
Chapter 7. Organizing data into training batches
Chapter 7. Creating data loaders for an instruction dataset
Chapter 7. Loading a pretrained LLM
Chapter 7. Fine-tuning the LLM on instruction data
Chapter 7. Extracting and saving responses
Chapter 7. Evaluating the fine-tuned LLM


Recommend Download Link Hight Speed | Please Say Thanks Keep Topic Live
No Password - Links are Interchangeable
 

Users who are viewing this thread

Back
Top