Skip to content

SouravAggarwal/LLM-Learning

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

4 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

🧠 LLM-Learning

A Comprehensive Journey into Large Language Models, Generative AI & Autonomous Agents

GitHub Python Jupyter LLM License

πŸš€ 939+ Notebooks Β· πŸ“¦ 428 Python Scripts Β· πŸ“š 5 Learning Tracks Β· πŸ€– 8+ AI Frameworks


From LLM fundamentals to production-grade autonomous agents β€” everything in one repository.


πŸ“‘ Table of Contents


🌟 Overview

This repository is a structured, hands-on learning hub covering the full spectrum of modern AI engineering β€” from the theory behind Large Language Models to building production-grade RAG systems and autonomous multi-agent workflows.

πŸ“š Learning Progression
━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━
  LLM Basics ──▢ Prompt Engineering ──▢ Open-Source Models
       β”‚                                       β”‚
       β–Ό                                       β–Ό
  RAG Systems ──▢ Fine-Tuning (Closed + Open) ──▢ Agentic AI
       β”‚                                            β”‚
       β–Ό                                            β–Ό
  Vector DBs ──▢ Evaluation & Benchmarks ──▢ Multi-Agent Systems

πŸ“‚ Repository Structure

LLM-Learning/
β”‚
β”œβ”€β”€ πŸ““ 1_MyNotebooks/           # Personal consolidated notes & projects
β”‚   β”œβ”€β”€ 1_LLM/                  # Core LLM concepts (13 notebooks)
β”‚   β”œβ”€β”€ 2_RAG/                  # RAG fundamentals (3 notebooks)
β”‚   β”œβ”€β”€ 3_AgentAI/              # Agent architecture (1 notebook)
β”‚   └── 4_Projects/             # Practical projects (1 notebook)
β”‚
β”œβ”€β”€ πŸŽ“ 3_Generative_AI/         # GenAI, ChatGPT & Copilot course (19 modules)
β”‚   β”œβ”€β”€ Module 0 – Module 18    # Data wrangling, EDA, visualization
β”‚   └── Course Outline.xlsx     # Full course syllabus
β”‚
β”œβ”€β”€ βš™οΈ 4_LLM_Engineering/       # 8-week LLM Engineering bootcamp (900+ notebooks)
β”‚   β”œβ”€β”€ week1/                  # Web scraping & summarization
β”‚   β”œβ”€β”€ week2_Gradio_And_ChatBot/
β”‚   β”œβ”€β”€ week3_HuggingFace_OpenSource_Solutions/
β”‚   β”œβ”€β”€ week4_Model_Evaluation/
β”‚   β”œβ”€β”€ week5_RAG/
β”‚   β”œβ”€β”€ week6_FineTuneing_ClosedSourceLLM/
β”‚   β”œβ”€β”€ week7_FineTuning_OpenSource/
β”‚   β”œβ”€β”€ week8_AgentAI/
β”‚   β”œβ”€β”€ guides/                 # 14 foundational guides
β”‚   β”œβ”€β”€ community-contributions/
β”‚   └── extras/
β”‚
β”œβ”€β”€ πŸ” 5_LLM_RAG/               # RAG deep-dive (in progress)
β”‚
└── πŸ€– 6_AI_Agents/             # Agent frameworks & courses
    β”œβ”€β”€ Deeplearning.ai_AutoGen/
    β”œβ”€β”€ Deeplearning.ai_Crew.ai/
    β”œβ”€β”€ Google_Kaggle/
    └── HuggingFace/

πŸ““ 1 β€” My Notebooks (Personal Notes)

Consolidated, self-authored notes covering LLM fundamentals, RAG, Agents, and hands-on projects.

πŸ”· LLM Fundamentals (1_MyNotebooks/1_LLM/)

# Notebook Topic
0 πŸ“ Questions Key questions & interview prep
1 πŸ“˜ Basics LLM fundamentals, models, tokenization
2 πŸ”§ Setup LLM Environment setup & configuration
3.1 ⚑ LLM Engineering Prompt engineering & API usage
3.2 🎯 Choosing a Model Model selection & comparison
4.1 πŸ¦™ Model Internals β€” LLAMA Deep dive into LLAMA architecture
5.1 πŸ€— HuggingFace Library HuggingFace Transformers integration
6.1 πŸ‹οΈ Training Fine-tuning & training concepts
9.1 πŸ’Ό Business Implementation Real-world enterprise use cases
9.2 πŸ’° Commercial (Theory) Commercial LLM landscape
10.1 ☁️ Inference on Cloud Cloud deployment strategies
10.2 πŸ›’ Amazon Price Agent Price estimation agent project

πŸ”· RAG β€” Retrieval-Augmented Generation (1_MyNotebooks/2_RAG/)

# Notebook Topic
1 πŸ“˜ Basics RAG fundamentals, vector stores, embeddings
2 πŸ”„ RAG Workflow End-to-end RAG pipeline
3 🦜 LangChain LangChain library deep-dive

πŸ”· Agent AI (1_MyNotebooks/3_AgentAI/)

# Notebook Topic
1 πŸ€– Basics Agent architecture & frameworks

πŸ”· Projects (1_MyNotebooks/4_Projects/)

# Notebook Topic
1 πŸ’¬ Self-Twin ChatBot Personal AI chatbot project

πŸŽ“ 3 β€” Generative AI Course

GenAI, ChatGPT, CoPilot & Agents for Business β€” A comprehensive 19-module course covering data science foundations and generative AI applications.

πŸ“Š Notebooks

Module Notebook Topic
3 πŸ“Š Data Wrangling (Stemplicity) Data wrangling & analysis with Pandas
3 πŸ” EDA with Pandas Exploratory data analysis project
4 πŸ“ˆ Data Visualization Data visualization with Seaborn

πŸ“‹ Full course outline: GenAI, ChatGPT, CoPilot & Agents for Business - Course Outline.xlsx


βš™οΈ 4 β€” LLM Engineering (8-Week Bootcamp)

900+ notebooks across an intensive 8-week program progressing from web scraping to autonomous agents. Includes community contributions from 40+ learners.

πŸ“… Week 1 β€” Web Scraping & Summarization

Notebook Topic
πŸ“„ Day 1 Web scraping fundamentals
πŸ“„ Day 2 Text summarization with LLMs
πŸ“„ Day 4 Content generation techniques
πŸ“„ Day 5 Brochure generation project
πŸ“ Week 1 Exercise Weekly challenge

πŸ“… Week 2 β€” Gradio & ChatBot Development

Notebook Topic
πŸ’¬ Day 1 Gradio UI basics
πŸ’¬ Day 2 Building chatbots
πŸ’¬ Day 3 Multi-model chat interfaces
πŸ’¬ Day 4 Advanced Gradio features
πŸ“ Week 2 Exercise Weekly challenge

πŸ“… Week 3 β€” HuggingFace & Open-Source Solutions

Notebook Topic
πŸ€— Day 1 HuggingFace ecosystem introduction
πŸ€— Day 2 Pipelines API
πŸ€— Day 2 β€” Pipelines HuggingFace Pipelines deep-dive
πŸ€— Day 3 Tokenizers
πŸ€— Day 3 β€” Tokenizers Tokenizer internals
πŸ€— Day 4 Models & inference
πŸ€— Day 4 β€” Models Model loading & configuration
πŸ€— Day 5 Meeting minutes product
πŸ€— Day 5 β€” Meeting Minutes Meeting minutes generator

πŸ“… Week 4 β€” Model Evaluation & Code Generation

Notebook Topic
πŸ“Š Day 3 Code language conversion
πŸ“Š Day 4 Open-source code conversion
πŸ“Š Day 5 Evaluation & benchmarking

πŸ“… Week 5 β€” RAG (Retrieval-Augmented Generation)

Notebook Topic
πŸ” Day 1 RAG core concepts
πŸ” Day 2 Document chunking strategies
πŸ” Day 3 Basic RAG implementation
πŸ” Day 4 RAG testing & quality
πŸ” Day 5 RAG without LangChain

πŸ“… Week 6 β€” Fine-Tuning Closed-Source LLMs

Notebook Topic
πŸ”₯ Day 1 Fine-tuning fundamentals
πŸ”₯ Day 2 Data preprocessing
πŸ”₯ Day 3 Machine learning foundations
πŸ”₯ Day 4 Neural network & LLM training
πŸ”₯ Day 5 LLM fine-tuning with OpenAI
πŸ“ˆ Results Training results analysis
πŸ” Redemption Train Improved training run
πŸ” Redemption Run Improved inference run

πŸ“… Week 7 β€” Fine-Tuning Open-Source Models

Notebook Topic
πŸ¦™ Day 1 Open-source fine-tuning intro
πŸ¦™ Day 1 β€” QLoRA Memory QLoRA & memory optimization
πŸ¦™ Day 2 Base model evaluation
πŸ¦™ Day 2 β€” OS Evaluation Open-source model benchmarking
πŸ¦™ Day 3 & 4 Training with PEFT/LoRA
πŸ¦™ Day 3 β€” Training Full training pipeline
πŸ¦™ Day 5 Testing fine-tuned models
πŸ¦™ Day 5 β€” Testing Fine-tuned model evaluation
πŸ“ˆ Results Training results

πŸ“… Week 8 β€” Autonomous Agent AI

Notebook Topic
πŸ€– Day 1 Agent fundamentals
πŸ€– Day 2 Tool usage & function calling
πŸ€– Day 3 Planning & reasoning
πŸ€– Day 4 Multi-agent orchestration
πŸ€– Day 5 Production deployment
πŸ“ˆ Results Agent performance analysis

πŸ“– Foundational Guides

# Guide Topic
01 πŸš€ Intro Course introduction
02 πŸ’» Command Line Terminal & CLI basics
03 πŸ”€ Git & GitHub Version control essentials
04 πŸ—οΈ Technical Foundations CS fundamentals for AI
05 πŸ““ Notebooks Jupyter notebook mastery
06 🐍 Python Foundations Python for AI/ML
07 ✨ Vibe Coding & Debugging AI-assisted development
08 πŸ› Debugging Debugging strategies
09 πŸ”Œ AI APIs & Ollama API setup & local models
10 🐍 Intermediate Python Advanced Python patterns
11 ⚑ Async Python Asynchronous programming
12 🎯 Starting Your Project Project scaffolding
13 🌐 Frontend Frontend for AI apps
14 🐳 Docker & Terraform Containerization & IaC

🌍 Community Contributions & Extras

Resource Description
πŸ“‚ Community Contributions 87+ notebooks from 40+ contributors
πŸ“‚ Extras β€” Signal Prototype Signal processing prototype
πŸ“‚ Extras β€” Trading Prototype Trading system prototype

πŸ” 5 β€” LLM RAG (Work in Progress)

🚧 This section is under active development β€” Deep-dive into advanced RAG patterns and techniques.


πŸ€– 6 β€” AI Agents

Hands-on courses from DeepLearning.AI, Google/Kaggle, and HuggingFace covering leading agent frameworks.

πŸ”΅ AutoGen β€” Multi-Agent Conversations (Deeplearning.ai_AutoGen/)

Lesson Notebook Topic
L1 🎭 Multi-Agent Conversation Multi-agent conversations & stand-up comedy
L2 πŸ”— Sequential Chats Sequential chats & customer onboarding
L3 πŸͺž Reflection Reflection & blog post writing
L4 β™ŸοΈ Tool Use Tool use & conversational chess
L5 πŸ’Ή Coding & Finance Coding & financial analysis
L6 πŸ“‹ Planning Planning & stock report generation

🟣 Crew.ai β€” Multi-Agent Orchestration (Deeplearning.ai_Crew.ai/)

Lesson Notebook Topic
L0 πŸ“˜ Basics Crew.ai fundamentals
L2 πŸ“° Research & Articles Research & article writing agents
L3 🎧 Customer Support Customer support automation
L4 πŸ“ž Customer Outreach Tools & customer outreach
L4 πŸ“Š Sales Pipeline Sales pipeline management
L5 πŸŽ‰ Event Planning Task-based event planning
L5 πŸ“‹ Support Reports Executive support reports
L6 πŸ’° Financial Analysis Collaborative financial analysis
L6 πŸ“ Content at Scale Content creation at scale
L7 πŸ“„ Job Application Job application crew
L11 πŸ—‚οΈ Project Planning Project planning agents

🟒 Google Kaggle β€” Agent Fundamentals (Google_Kaggle/)

Notebook Topic
πŸ€– Basic Agents Introduction to AI agents

🟑 HuggingFace (HuggingFace/)

Resource Description
πŸ† Certificate Course completion certificate

πŸ› οΈ Technologies & Frameworks

Category Technologies
🧠 LLM APIs OpenAI (GPT), Anthropic (Claude), Google (Gemini)
πŸ¦™ Open-Source Models LLaMA, Mistral, DeepSeek, Qwen
πŸ€— ML Platform HuggingFace Transformers, Datasets, PEFT
🦜 Orchestration LangChain, LiteLLM
πŸ€– Agent Frameworks AutoGen, Crew.ai, Custom agents
πŸ’Ύ Vector Databases ChromaDB, Pinecone, Qdrant, FAISS
πŸ–₯️ UI Frameworks Gradio, Streamlit
πŸ‹οΈ Fine-Tuning QLoRA, PEFT, LoRA, OpenAI Fine-tuning API
☁️ Cloud & Deploy Google Colab, Modal, Docker, Terraform
🐍 Languages Python, C/C++, JavaScript
πŸ”§ Dev Tools Jupyter, Ollama, uv, Cursor

πŸš€ Getting Started

Prerequisites

  • Python 3.10+
  • Jupyter Notebook / JupyterLab
  • API keys for OpenAI / Anthropic / Google (optional β€” free alternatives available)

Quick Start

# Clone the repository
git clone https://github.com/SouravAggarwal/LLM-Learning.git
cd LLM-Learning

# Start with personal notes
jupyter notebook 1_MyNotebooks/1_LLM/1_Basics.ipynb

# Or jump into the LLM Engineering bootcamp
cd 4_LLM_Engineering
pip install -r requirements.txt
jupyter notebook week1/day1.ipynb

🏠 Run Local Models with Ollama

# Install Ollama from https://ollama.com
ollama run llama3.2         # Standard (3B parameters)
ollama run llama3.2:1b      # Lightweight for smaller machines

πŸ’‘ Tip: See Guide 09 β€” AI APIs & Ollama for detailed setup with Ollama, Gemini, OpenRouter, and more free alternatives.


πŸ“Š Repository Stats

Metric Count
πŸ““ Jupyter Notebooks 939+
🐍 Python Scripts 428
πŸ“– Markdown Docs 208
πŸ“ Total Files 1,600+
πŸ“… Weekly Modules 8 weeks
πŸ—οΈ Projects 100+ community projects
πŸ‘₯ Contributors 40+
πŸ€– AI Frameworks 8+

⭐ Found this useful? Give it a star!

Built with ❀️ for the AI/ML community

GitHub Stars

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors