Milestone 3: RAG Pipeline#
1. Deliverable Description#
The goal of this milestone is to design, implement, and present your Retrieval-Augmented Generation (RAG) pipeline. Your team should demonstrate how you integrated retrieval and generation components, how your system handles user queries, and how you evaluated and refined your pipeline.
There are three deliverables:
Group Presentation – Showcase your RAG pipeline (due Oct 30)
Technical Report – Around 500–800 words (2–3 pages) (due Nov 1)
Peer Reviews – Review other groups’ RAG presentations or reports (due Nov 1)
2. Presentation Guidelines#
Time: Maximum 8 minutes + 2 minutes for Q&A
Slides: Maximum 6 slides
Participation: All team members must contribute
Content:
Present your RAG pipeline diagram (data flow, retriever, generator)
Explain the query transformation techniques you applied
Show examples of prompt templates used in your pipeline
Demonstrate your retrieval and generation process using LangSmith traces or logs
Reflect on challenges and design decisions
3. Suggested 6-Slide Template#
Slide 1: Title & Team#
Slide 2: RAG Pipeline Overview#
High-level workflow diagram:
Document ingestion → Embedding model → Vector store → Retriever → LLM prompt → Response
Indicate tools used (e.g., LangChain, MongoDB Atlas Vector Search, Gemini/OpenAI API)
Highlight any unique design choices
Slide 3: Query Transformation Techniques#
Describe techniques used to improve retrieval (e.g., MultiQuery, RAG-Fusion, Query Decomposition)
Show examples of transformed queries
Explain why your team chose these techniques
Slide 4: Prompt Engineering#
Show your prompt templates and variable placeholders
Explain how you format retrieved context into prompts
Discuss prompt refinement and lessons learned
Slide 5: Example Query & “Under the Hood” View#
Example user query (e.g., “What policies did TRU release on AI ethics in 2024?”)
Show how it is transformed → retrieved → generated
Screenshot from LangSmith trace showing:
Steps of retrieval
Prompt construction
Model response
Interpret what happened and how your RAG pipeline performed
Slide 6: Reflection & Next Steps#
Key insights and takeaways
Challenges (e.g., noisy retrievals, hallucinations, latency)
Planned improvements (e.g., rerankers, prompt optimization, hybrid retrieval)
4. Technical Report Components#
4.1. Introduction#
Briefly describe your project’s goal and how RAG fits into it
Define the purpose of your pipeline (e.g., Q&A, summarization, recommendation)
4.2. Pipeline Architecture#
Include a system diagram and explain each component:
Data ingestion
Embedding generation
Vector storage (MongoDB Atlas)
Retriever configuration
LLM generation
Mention models and libraries used
4.3. Query Transformation#
Detail the transformation techniques you implemented
Include sample code or screenshots showing before–after queries
Justify your design decisions
4.4. Prompt Engineering#
Provide your base prompt(s) and explain how they evolved
Discuss context window management (e.g., truncation, top-k docs)
4.5. Example Walkthrough (with LangSmith)#
Choose one representative query
Show the trace or log of its processing steps:
User query
Transformed query
Retrieved documents (with metadata)
Prompt sent to LLM
Model output
Comment on performance, accuracy, and relevance
4.6. Reflection#
What worked well in your pipeline?
What limitations did you observe?
How could it be improved (e.g., reranker, metadata filtering, evaluation metrics)?
4.7. References#
Cite any frameworks, tutorials, or models used
4.8. Format#
Around 500–800 words Write in Markdown or Jupyter Notebook format (.md or .ipynb) Include diagrams, screenshots, and traces where appropriate
5. Peer Review Activity#
Each student will review 2 RAG presentations or reports. Provide half a page of feedback for each using GitHub Issues. Peer review accounts for 5% of your grade.
Submission Steps:#
Visit the group’s GitHub repository.
Create a GitHub Issue using the peer review template.
Add your comments and submit.