đ Workshop Schedule (10 AM - 2 PM)
10:00-11:00: Memory Foundation
11:00-12:00: Manual RAG Demo
12:00-1:00: Why Vector Hype?
1:00-2:00: Vector Search in Action
Hour 1: Memory Foundation (10:00-11:00)
The Conversation Analogy (30 minutes)
Core Teaching Example
âRight now, our conversation is going in my short-term memoryâ
âI remember you asked about Day 2 workshop planning earlierâ
âBut when you ask about vector databases, I search my knowledge base like searching a bookâ
âThat's long-term memory through retrievalâ
Interactive Demo
We'll demonstrate memory limitations in real-time:
- Start fresh ChatGPT conversation
- Ask it to remember 5 customer names throughout chat
- Keep chatting until it forgets first names (context limit)
- Show contrast with RAG system that never forgets
Bridge to AI Systems
- Short-term = conversation context (limited, expensive)
- Long-term = knowledge retrieval (unlimited, searchable)
Hour 2: Manual RAG Demonstration (11:00-12:00)
Breaking Down R-A-G Live (45 minutes)
âI'll be the AI agent - let's build RAG step by stepâ
Step 1: The Query
Student asks: âWhat's the latest AI safety regulations?â
Step 2: Retrieval (R) - Live Google Search
- âI need external knowledge. Let me search...â
- Live demo: Google search âAI safety regulations 2024â
- Show results on screen
- âI'm the Retrieval componentâ
Step 3: Augmentation (A) - Building Context
- Show original prompt vs augmented prompt with search results
- âSee how I augmented my knowledge with fresh information?â
Step 4: Generation (G) - Create Response
- Put augmented prompt into ChatGPT
- Show the enhanced response
- âAI now gives current, accurate informationâ
Student Exercise (15 minutes)
Students become âhuman RAG systemsâ:
- Each gets different query (iPhone 16, mortgage rates, etc.)
- They manually perform R-A-G process
Hour 3: Why All the Vector Hype? (12:00-1:00)
âSo why is everyone obsessed with vector search?â (10 minutes)
Set the stage:
- âWe just saw 6 different retrieval methodsâ
- âBut 90% of RAG tutorials only show vectors. Why?â
- âLet's understand what makes vectors specialâ
The Magic of Understanding Meaning (15 minutes)
Real examples to show the power:
SQL search: ârefund policyâ won't find âreturn procedureâ
Keyword search: âcheapâ won't find âaffordableâ or âbudget-friendlyâ
Vector search: understands these are all related concepts
Key insight: âVectors capture meaning, not just wordsâ
How Sentences Become Vectors (20 minutes)
The Transformation Journey:
Step 1: Sentence â Tokens
âThe refund policy is generousâ
Becomes: [âTheâ, ârefundâ, âpolicyâ, âisâ, âgenerousâ]
Show tokenization in action
Step 2: Tokens â Numbers (Embeddings)
Each token gets converted to a vector (list of numbers)
ârefundâ might become [0.2, -0.1, 0.8, 0.3, ...]
Analogy: âLike GPS coordinates, but for wordsâ
Step 3: Understanding Embeddings as Locations
- âEvery word has a location in high-dimensional spaceâ
- Similar words live near each other
- Visual: Show 2D representation of word clusters
The Famous King-Queen Example (15 minutes)
Classic demonstration:
- âkingâ has coordinates in vector space
- âmanâ has coordinates in vector space
- âwomanâ has coordinates in vector space
- âqueenâ has coordinates in vector space
The math magic:
king - man + woman â queen
âRemove the âmalenessâ from king, add âfemalenessâ, get queenâ
Mind-blowing moment: âThe AI discovered gender relationships just from text!â
More examples:
Paris - France + Italy â Rome
walking - walk + ran â running
- âThe AI learns relationships between conceptsâ
Hour 4: Vector Search in Action (1:00-2:00)
How LLMs Work: The Simple Truth (20 minutes)
âLet me show you how every LLM actually works:â
The Universal LLM Process:
Sentence â Tokens â Embeddings â BLACK BOX â Embeddings â Tokens â Sentence
Breaking it down:
- Input: âWhat's your refund policy?â
- Tokenize: [âWhatâ, â'sâ, âyourâ, ârefundâ, âpolicyâ, â?â]
- Convert to embeddings: [numbers, numbers, numbers...]
- BLACK BOX: Complex neural network processing
- Output embeddings: [different numbers...]
- Convert back: [âOurâ, ârefundâ, âpolicyâ, âallowsâ, â30â, âdays...â]
What's in the Black Box?
- Pattern Recognition: âI've seen similar questions beforeâ
- Attention Mechanisms: âWhat words are important here?â
- Context Understanding: âBased on what we've discussed...â
- Knowledge Retrieval: âI remember from training that...â
- Reasoning Circuits: Logical connections between concepts
- Language Generation: âHere's how to say this clearlyâ
Key Message: âYou don't need to understand the black box to use it - like driving a car!â
The Vector Search Connection (20 minutes)
âHere's the beautiful part - we can use these same embeddings for search!â
The Realization:
- LLMs convert text to embeddings anyway
- Those embeddings capture meaning
- We can use them for similarity search
- âIt's the same numbers, just used differently!â
Vector Search Process:
- Query: ârefund policyâ â embeddings
- Documents: All converted to embeddings beforehand
- Search: Find closest embeddings (distance calculation)
- Same embeddings, different purpose!
Why This is Powerful:
- No need for separate search technology
- Same understanding of meaning as the LLM
- Perfect compatibility between search and generation
Embedding Search = Distance Calculation (20 minutes)
Simple explanation:
- âEmbedding search is just measuring distances between pointsâ
- Query vector: âWhat's your return policy?â
- Document vectors: All company documents
- Find documents with shortest distance to query
- Visual: Show dots on a map, finding closest ones
đŻ Interactive Vector Search Animation
Let's see exactly how this works! Follow along as we transform âThe quick brown foxâ into vectors and search for âa fast dark-colored foxâ. Use the controls below to step through the process.
Vector Search Animation
Understanding how semantic search works under the hood
Step 1: Sentence to Tokens
A sentence is broken down into individual words or 'tokens'
Why distance = similarity:
- Close in space = similar in meaning
- Far apart = different concepts
- Demo: Show how ârefundâ and âreturnâ are close, but ârefundâ and âpizzaâ are far
Hands-on Activities
Hands-on Vector Exploration (20 minutes)
- Use embedding playground/tool to show word relationships
- Students input their own examples
- See how different words cluster together
- âAha momentâ: Watch them discover semantic relationships
Vector Database Implementation
Why Start with Vectors? (20 minutes)
- Most flexible for natural language
- Handles ambiguity well
- Good starting point for document-heavy use cases
- But remember - just one tool in retrieval toolkit
Hands-on Vector Setup (40 minutes)
- Use pre-built functions approach
- Students create knowledge base with their documents
- Test different search queries
- Compare results with manual Google search approach
Key Takeaways
đŻ What You'll Walk Away With
- Memory Systems Understanding: Clear distinction between short-term conversation context and long-term knowledge retrieval
- RAG Demystified: Hands-on experience building RAG systems manually before automation
- Vector Intuition: Deep understanding of why vectors capture meaning and enable semantic search
- Practical Skills: Ability to implement vector-based retrieval systems using real tools
- Mathematical Insights: Understanding how embeddings work without getting lost in complexity
đ Ready for Implementation?
By the end of this workshop, you'll have a solid foundation in memory systems and vector search. You'll understand not just how to use these technologies, but why they work and when to apply them.
The combination of practical demonstrations, interactive exercises, and hands-on coding gives you the confidence to build your own RAG systems from scratch.