Complete Reference Materials β Videos, Repos, Courses, PapersΒΆ
Curated resources aligned to each learning phase. Free resources are marked with (FREE).
How to Use This FileΒΆ
Donβt try to consume all of these β use them as a menu
For each phase, pick 1-2 video resources and 1-2 repos to study
βMust-watchβ items are marked with **
Video Courses & YouTube ChannelsΒΆ
Mathematics & FoundationsΒΆ
Resource |
Description |
Link |
|---|---|---|
3Blue1Brown: Essence of Linear Algebra |
Best visual explanation of linear algebra |
https://www.youtube.com/playlist?list=PLZHQObOWTQDPD3MizzM2xVFitgF8hE_ab |
3Blue1Brown: Essence of Calculus |
Intuitive calculus from scratch |
https://www.youtube.com/playlist?list=PLZHQObOWTQDMsr9K-rj53DwVRMYO3t5Yr |
3Blue1Brown: Neural Networks |
4-part series, best visual intro to neural networks |
https://www.youtube.com/playlist?list=PLZHQObOWTQDNU6R1_67000Dx_ZCJB-3pi |
StatQuest with Josh Starmer |
Every ML concept explained simply with cartoons |
|
MIT 18.065 (Gilbert Strang) |
Full MIT linear algebra course (more rigorous) |
https://www.youtube.com/playlist?list=PLUl4u3cNGP63oMNUHXqIUcrkS2PivhN3k |
Stanford CS229 (Andrew Ng) |
Classic ML course, full lectures |
https://www.youtube.com/playlist?list=PLoROMvodv4rMiGQp3WXShtMGgzqpfVfbU |
Deep Learning & Neural NetworksΒΆ
Resource |
Description |
Link |
|---|---|---|
Andrej Karpathy: Neural Networks Zero to Hero |
Build GPT from scratch, best hands-on series |
https://www.youtube.com/playlist?list=PLAqhIrjkxbuWI23v9cThsA9GvCAUhRvKZ |
Andrej Karpathy: makemore series |
Build character-level LM from scratch |
|
MIT 6.S191: Intro to Deep Learning |
Annual MIT deep learning course |
|
fast.ai Practical Deep Learning |
Top-down approach, very practical |
https://www.youtube.com/playlist?list=PLfYUBJiXbdtSvpQjSnJJ_PmDQB_VyT5iU |
Stanford CS231N: CNNs for Visual Recognition |
Best course on computer vision |
https://www.youtube.com/playlist?list=PL3FW7Lu3i5JvHM8ljYj-zLfQRF3EO8sYv |
Stanford CS224N: NLP with Deep Learning |
Best course on NLP/transformers |
https://www.youtube.com/playlist?list=PLoROMvodv4rOSH4v6133s9LFPRHjEmbmJ |
LLMs, Transformers, and Generative AIΒΆ
Resource |
Description |
Link |
|---|---|---|
Andrej Karpathy: Letβs build GPT |
Build GPT-2 from scratch in 2 hours |
|
Andrej Karpathy: Letβs build the GPT Tokenizer |
Hands-on tokenization |
|
Yannic Kilcher: Paper Explanations |
Deep research paper walkthroughs |
|
The AI Epiphany |
Clear explanations of recent AI papers |
|
Umar Jamil: Transformers from scratch |
Code implementations of major transformer papers |
|
Two Minute Papers |
3-5 min summaries of top AI research |
RAG, Agents, and Applied LLMsΒΆ
Resource |
Description |
Link |
|---|---|---|
Sam Witteveen (Red Dragon AI) |
RAG, LangChain, agents hands-on |
|
James Briggs |
LangChain, RAG, Pinecone tutorials |
|
Data Independent |
LangChain + OpenAI practical tutorials |
|
LangChain Official |
Official LangChain tutorials and walkthroughs |
|
Matt Williams (Ollama) |
Local LLMs with Ollama |
Fine-tuning and MLOpsΒΆ
Resource |
Description |
Link |
|---|---|---|
Trelis Research |
LLM fine-tuning, quantization, deployment |
|
Weights & Biases (W&B) |
MLOps, experiment tracking tutorials |
|
Made With ML |
Full MLOps course (free) |
|
Patrick Loeber |
PyTorch tutorials from scratch |
|
Sentdex |
Python, ML, reinforcement learning |
Free Online CoursesΒΆ
Must-Complete (Curated for Job Readiness)ΒΆ
Course |
Platform |
Duration |
Cost |
|---|---|---|---|
Hugging Face NLP Course |
HuggingFace |
4-6 weeks |
FREE |
fast.ai Practical Deep Learning |
8 weeks |
FREE |
|
DeepLearning.AI: ChatGPT Prompt Engineering |
Coursera |
1 week |
FREE audit |
Made With ML |
6-8 weeks |
FREE |
|
Stanford CS229 Machine Learning |
Stanford |
12 weeks |
FREE |
Stanford CS224N NLP with Deep Learning |
Stanford |
10 weeks |
FREE |
MIT OpenCourseWare 6.036 |
MIT |
12 weeks |
FREE |
Google Machine Learning Crash Course |
3-4 weeks |
FREE |
|
Microsoft AI for Beginners |
GitHub |
24 lessons |
FREE |
Microsoft ML for Beginners |
GitHub |
26 lessons |
FREE |
Microsoft Generative AI for Beginners |
GitHub |
18 lessons |
FREE |
Microsoft AI Agents for Beginners |
GitHub |
Self-paced |
FREE |
Paid (Worth the Investment)ΒΆ
Course |
Platform |
Duration |
Cost |
|---|---|---|---|
Deep Learning Specialization (Andrew Ng) |
Coursera |
3-4 months |
$49/mo |
LLM Fine-tuning with Hugging Face |
Coursera/DL.AI |
2 weeks |
$49/mo |
LangChain for LLM Applications |
Coursera/DL.AI |
1 week |
$49/mo |
Building AI Agents (DeepLearning.AI) |
Coursera |
1-2 weeks |
$49/mo |
MLOps Specialization |
Coursera |
4 months |
$49/mo |
Full Stack LLM Bootcamp |
FSDL |
Self-paced |
$200 |
GitHub Repositories to StudyΒΆ
Foundational MLΒΆ
Repo |
Stars |
Why Study It |
|---|---|---|
60k+ |
Industry-standard ML library; study examples folder |
|
28k+ |
Hands-On ML book notebooks (excellent quality) |
|
5k+ |
ML with scikit-learn, Keras, TensorFlow book code |
|
70k+ |
26 lessons, curriculum-style |
Deep Learning & PyTorchΒΆ
Repo |
Stars |
Why Study It |
|---|---|---|
38k+ |
Minimal, readable GPT-2 implementation β must read |
|
10k+ |
Minimal BPE tokenizer implementation |
|
8k+ |
Official PyTorch tutorials |
|
26k+ |
high-level DL library, great for learning patterns |
|
4k+ |
Clean transformer implementations |
LLMs and NLPΒΆ
Repo |
Stars |
Why Study It |
|---|---|---|
140k+ |
Industry standard; study examples and model cards |
|
17k+ |
LoRA/QLoRA fine-tuning library |
|
12k+ |
SFT, DPO, RLHF training |
|
62k+ |
Official OpenAI examples and best practices |
|
8k+ |
Official Claude/Anthropic examples |
|
42k+ |
Excellent LLM engineering course with notebooks |
RAG and Vector SearchΒΆ
Repo |
Stars |
Why Study It |
|---|---|---|
97k+ |
Most popular LLM orchestration framework |
|
38k+ |
Alternative to LangChain, strong for RAG |
|
16k+ |
Most popular local vector database |
|
21k+ |
High-performance vector DB for production |
|
13k+ |
Vector search in PostgreSQL |
|
16k+ |
Unified interface to 100+ LLM APIs |
Agents and Tool UseΒΆ
Repo |
Stars |
Why Study It |
|---|---|---|
10k+ |
Graph-based stateful agent framework |
|
18k+ |
OpenAIβs lightweight multi-agent framework |
|
37k+ |
Multi-agent conversation framework |
|
26k+ |
Role-based multi-agent framework |
|
Growing |
MCP standard for tool integration |
Fine-tuningΒΆ
Repo |
Stars |
Why Study It |
|---|---|---|
22k+ |
2-4x faster fine-tuning with less memory |
|
9k+ |
Powerful fine-tuning config framework |
|
5k+ |
Official HF recipes for SFT and DPO |
|
40k+ |
Easy fine-tuning for many models |
|
2k+ |
Curated LLM fine-tuning datasets |
MLOps and DeploymentΒΆ
Repo |
Stars |
Why Study It |
|---|---|---|
19k+ |
Experiment tracking and model registry |
|
9k+ |
Weights & Biases experiment tracking |
|
44k+ |
Fast LLM inference and serving |
|
80k+ |
Fast API framework for ML serving |
|
34k+ |
Build ML demos in minutes |
|
36k+ |
Data apps with pure Python |
EvaluationΒΆ
Repo |
Stars |
Why Study It |
|---|---|---|
7k+ |
RAG evaluation framework |
|
14k+ |
OpenAIβs evaluation framework |
|
7k+ |
Industry standard LLM benchmarking |
|
2k+ |
Metrics for ML models |
Blogs and ReadingΒΆ
Must-Read BlogsΒΆ
Blog |
Why Read It |
Link |
|---|---|---|
Lilian Weng (OpenAI) |
Deeply researched posts on transformer internals, RL, agents |
|
Sebastian Ruder |
NLP research trends and summaries |
|
Jay Alammar |
Visual explanations of transformer and BERT |
|
Andrej Karpathy |
Occasional deep-dive blog posts |
|
The Gradient |
Long-form ML research coverage |
|
Import AI (Jack Clark) |
Weekly AI news for practitioners |
|
Simon Willison |
Practical LLM and tool use |
|
Eugene Yan |
Applied ML in production |
Research Papers (Required Reading for Senior Roles)ΒΆ
Start with the abstracts and conclusions, read fully for the ones most relevant to your role.
Foundation papers:
Attention Is All You Need (2017) β the Transformer
BERT (2018) β bidirectional pre-training
Language Models are Few-Shot Learners / GPT-3 (2020) β scaling + prompting
An Image is Worth 16x16 Words / ViT (2020) β Vision Transformers
RAG and retrieval:
Retrieval-Augmented Generation (2020) β original RAG paper
Precise Zero-Shot Dense Retrieval / HyDE (2022) β HyDE technique
Fine-tuning and alignment:
LoRA (2021) β low-rank adaptation
QLoRA (2023) β quantized LoRA on consumer hardware
InstructGPT / RLHF (2022) β training with human feedback
Direct Preference Optimization / DPO (2023) β alignment without RL
Agents and tools:
ReAct: Synergizing Reasoning and Acting (2022) β ReAct agents
Toolformer (2023) β teaching LLMs to use tools
Prompting:
Chain-of-Thought Prompting (2022) β step-by-step reasoning
Large Language Models are Zero-Shot Reasoners (2022) β βLetβs think step by stepβ
Datasets for Practice ProjectsΒΆ
Dataset |
Use Case |
Where to Find |
|---|---|---|
HuggingFace Datasets Hub |
Everything β 100k+ datasets |
|
Kaggle Datasets |
Competition-grade structured data |
|
Common Crawl |
Web text for LLM training |
|
The Pile |
Diverse text dataset |
|
OpenAssistant OASST2 |
Conversation data for fine-tuning |
|
Alpaca dataset |
Instruction-following data |
|
SQUAD v2 |
Question answering |
|
MS MARCO |
Information retrieval |
|
BEIR benchmark |
RAG/retrieval evaluation |
Developer Tools to KnowΒΆ
APIs and PlatformsΒΆ
Tool |
What It Does |
Why Learn It |
|---|---|---|
OpenAI API |
GPT-4, embeddings, DALL-E |
Most common LLM API in production |
Anthropic API |
Claude models |
Strong at reasoning, long context |
Hugging Face Hub |
Model hosting and sharing |
Industry standard for open-source models |
Ollama |
Local LLM serving |
Dev/testing without cloud costs |
Replicate |
Run models via API |
Easy access to open-source models |
Modal |
Serverless GPU computing |
Run GPU workloads without managing servers |
Together AI |
Fast inference for open models |
Good OpenAI-compatible API alternative |
Development ToolsΒΆ
Tool |
Use |
Link |
|---|---|---|
LangSmith |
LangChain observability and tracing |
|
Weights & Biases |
Experiment tracking, visualizations |
|
MLflow |
Open-source experiment tracking |
|
LabelStudio |
Data labeling for fine-tuning |
|
Chainlit |
Build chat UIs for LLM apps |
|
Gradio |
Quick ML demos |
Learning CommunitiesΒΆ
Community |
Platform |
Value |
|---|---|---|
Hugging Face Forums |
HuggingFace |
Best for transformers/fine-tuning questions |
r/MachineLearning |
Research and industry news |
|
r/LocalLLaMA |
Local LLM community, very active |
|
ML Twitter/X |
Follow researchers and engineers for news |
|
Discord: Eleuther AI |
Discord |
Open-source LLM research community |
Discord: fast.ai |
Discord |
Practical DL community |
Discord: LangChain |
Discord |
LangChain help and announcements |
Kaggle forums |
Kaggle |
Competition strategies and feedback |
Phase-to-Resource MappingΒΆ
Quick reference β what to use for each phase:
Phase |
Video |
GitHub Repo |
Blog/Other |
|---|---|---|---|
0: Foundations |
StatQuest channel |
microsoft/ML-For-Beginners |
|
1: Python/Data Science |
StatQuest |
scikit-learn/scikit-learn examples |
ageron/handson-ml3 |
2: Mathematics |
3Blue1Brown Linear Algebra + Calculus |
karpathy/nanoGPT (for context) |
Jay Alammar blog |
3: Tokenization |
Karpathy: GPT Tokenizer |
karpathy/minbpe |
HF tokenizers docs |
4: Embeddings |
Sam Witteveen |
sentence-transformers repo |
Jay Alammar: Illustrated Word2Vec |
5: Neural Networks |
3Blue1Brown NN series + Karpathy Zero to Hero |
karpathy/nanoGPT |
Lilian Weng blog |
6: Vector Databases |
James Briggs |
chroma-core/chroma |
Pinecone docs |
7: RAG |
Sam Witteveen, James Briggs |
openai/openai-cookbook |
Lilian Weng RAG post |
8: MLOps |
Made With ML, W&B |
mlflow/mlflow |
Eugene Yan blog |
9: Specializations |
Fast.ai (for CV), Stanford CS224N (for NLP) |
langchain-ai/langgraph |
10-specializations/README.md |
10: Prompt Engineering |
DeepLearning.AI course |
openai/openai-cookbook |
Lilian Weng prompting |
11: Fine-tuning |
Trelis Research |
unslothai/unsloth |
LoRA + QLoRA papers |
12: Multimodal |
Two Minute Papers |
huggingface/transformers |
ViT paper, CLIP paper |
13: Local LLMs |
Matt Williams (Ollama) |
ollama/ollama |
r/LocalLLaMA |
14: AI Agents |
LangChain official |
langchain-ai/langgraph |
ReAct + Toolformer papers |
15: Streaming |
FastAPI docs |
tiangolo/fastapi |
SSE spec, WebSocket MDN |
16: Model Evaluation |
StatQuest: ROC/AUC |
explodinggradients/ragas |
Hugging Face evaluate docs |
17: Debugging |
Made With ML |
whylogs, evidently |
Eugene Yan debugging post |
18: Low-Code Tools |
Gradio official channel |
gradio-app/gradio |
Hugging Face Spaces docs |
19: AI Safety |
Anthropic research blog |
microsoft/promptbench |
Lilian Weng adversarial post |
24: Advanced DL |
Yannic Kilcher |
lucidrains/x-transformers |
Lilian Weng GAN/VAE posts |
25: Reinforcement Learning |
DeepMind YouTube |
openai/gym (now Gymnasium) |
Lilian Weng RL post |
26: Time Series |
StatQuest time series |
facebook/prophet |
Rob Hyndman blog |
27: Causal Inference |
Brady Neal causal course |
microsoft/EconML |
Judea Pearl writings |