Welcome to my Neural Networks Training Repo! This repository captures everything I learn as I dive into the Zero to Hero course, building neural networks, language models, and even GPT from scratch.
- Build a micro neural network framework: micrograd.
- Learn backpropagation step-by-step.
- 📝 Lecture 1 Notebook | 🔗 Micrograd Repo
- Create a character-level bigram language model.
- Explore model training, sampling, and loss evaluation.
- 📝 Lecture 2 Notebook | 🔗 Makemore Repo
- Implement and train an MLP language model.
- Learn hyperparameter tuning, evaluation, and avoiding overfitting.
- 📝 Lecture 3 Notebook
- Dive into activations, gradients, and BatchNorm.
- Visualize and debug your deep networks.
- 📝 Lecture 4 Notebook
- Manually implement backpropagation for an MLP.
- Gain an intuitive understanding of gradient flow.
- 📝 Lecture 5 Notebook | 🔗 Google Colab Exercise
- Create a hierarchical CNN inspired by WaveNet.
- Explore
torch.nn
and efficient model development. - 📝 Lecture 6 Notebook
- Implement a GPT following the "Attention is All You Need" paper.
- Learn the building blocks of transformers.
- 📝 Lecture 7 Notebook
- Build and understand the GPT Tokenizer.
- Explore how tokenization impacts LLMs.
- 📝 Lecture 8 Notebook | 🔗 minBPE Code
- 📖 Master neural networks, language models, and GPT.
- 💻 Gain hands-on experience with PyTorch and deep learning tools.
- 🧠 Build intuition and confidence in machine learning.
-
Clone the Repository:
git clone https://github.com/Cluab/neural-networks-training.git cd neural-networks-training
-
Set Up the Environment:
python -m venv venv source venv/Scripts/activate # Ensure the virtual environment is active before proceeding pip install -r requirements.txt
-
Explore the Content:
- Navigate to the
lectures
folder to access notebooks and code.
- Navigate to the
-
Follow the Course:
- Watch the YouTube Playlist for guided learning.
-
Engage with Exercises:
- Attempt the exercises independently before reviewing the provided solutions.
- Languages: Python, PyTorch
- Guides:
This repository follows the MIT License. See the LICENSE file for details.