A short project to implement a small GPT like model from scratch. Most of it is inspired by minBPE and nanoGPT.
- Create and activate a virtual environment (optional)
python -m venv project_venv
source project_venv/bin/activate- Setup the project and download the requirements
pip install -e .
pip install -r requirements.txt- Run the code
python main.py --training_iterations=5000 --text=shakespeare --train_model=False --task=generation-
In coming:
- Fine-tuning (LoRA + RLHF)
- (TODO) Provide good training parameters fo translation
-
Done:
- Tokenizer (byte and character level)
- Full Transformer architecture (Encoder + Decoder)
- Training and inference pipeline for generation (Lorem Ipsum and Shakespeare)
- Training and inference pipeline for translation (en → fr)