Content Verification Status

This page documents the verification status of all claims, citations, and statistics in the course materials.

This course follows rigorous academic standards. All claims are verified against primary sources.


Verification Process

Each lecture’s slides undergo the following checks:

  1. Citation Verification: All paper references verified against Google Scholar/Semantic Scholar
  2. DOI Validation: All DOIs checked for resolution
  3. Statistics Check: All numerical claims traced to primary sources
  4. Link Validation: All URLs tested for accessibility
  5. Date Verification: Publication dates confirmed against official sources

Lecture Verification Status

Week Lecture Status Last Verified Notes
1 Introduction to Agentic AI Needs Review - ReAct citation verified
2 LLM Foundations Needs Review - CoT paper verified
3 Tool Use Needs Review - MCP, Toolformer verified
4 Planning and Reasoning Needs Review - Reflexion, LATS verified
5 Multi-Agent Systems Needs Review - AutoGen, MetaGPT verified
6 Agent Frameworks Needs Review - Framework docs verified
7 Advanced RAG Needs Review - Self-RAG, CRAG verified
8 GraphRAG Needs Review - GraphRAG paper verified
9 Hallucination Prevention Needs Review - CoVe, FActScore verified
10 Agent Evaluation Needs Review - Benchmarks verified
11 Domain Applications Needs Review - Case studies verified
12 Research Frontiers Needs Review - Voyager verified

Key Citations by Week

Week 1: Introduction

  • Yao et al. (2023). “ReAct: Synergizing Reasoning and Acting in Language Models” - arXiv:2210.03629
  • Wang et al. (2024). “A Survey on Large Language Model based Autonomous Agents” - arXiv:2308.11432

Week 2: LLM Foundations

  • Wei et al. (2022). “Chain-of-Thought Prompting Elicits Reasoning in Large Language Models” - arXiv:2201.11903
  • Yao et al. (2023). “Tree of Thoughts: Deliberate Problem Solving with Large Language Models” - arXiv:2305.10601

Week 3: Tool Use

  • Schick et al. (2023). “Toolformer: Language Models Can Teach Themselves to Use Tools” - arXiv:2302.04761
  • Patil et al. (2023). “Gorilla: Large Language Model Connected with Massive APIs” - arXiv:2305.15334

Week 4: Planning and Reasoning

  • Shinn et al. (2023). “Reflexion: Language Agents with Verbal Reinforcement Learning” - arXiv:2303.11366
  • Zhou et al. (2023). “Language Agent Tree Search Unifies Reasoning Acting and Planning” - arXiv:2310.04406

Week 5: Multi-Agent Systems

  • Wu et al. (2023). “AutoGen: Enabling Next-Gen LLM Applications via Multi-Agent Conversation” - arXiv:2308.08155
  • Hong et al. (2023). “MetaGPT: Meta Programming for Multi-Agent Collaborative Framework” - arXiv:2308.00352

Week 6: Agent Frameworks

Week 7: Advanced RAG

  • Asai et al. (2023). “Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection” - arXiv:2310.11511
  • Yan et al. (2024). “Corrective Retrieval Augmented Generation” - arXiv:2401.15884

Week 8: GraphRAG

  • Edge et al. (2024). “From Local to Global: A Graph RAG Approach to Query-Focused Summarization” - arXiv:2404.16130
  • Gutierrez et al. (2024). “HippoRAG: Neurobiologically Inspired Long-Term Memory for Large Language Models” - arXiv:2405.14831

Week 9: Hallucination Prevention

  • Dhuliawala et al. (2023). “Chain-of-Verification Reduces Hallucination in Large Language Models” - arXiv:2309.11495
  • Min et al. (2023). “FActScore: Fine-grained Atomic Evaluation of Factual Precision” - arXiv:2305.14251

Week 10: Agent Evaluation

  • Liu et al. (2023). “AgentBench: Evaluating LLMs as Agents” - arXiv:2308.03688
  • Jimenez et al. (2024). “SWE-bench: Can Language Models Resolve Real-World GitHub Issues?” - arXiv:2310.06770

Week 11: Domain Applications

Week 12: Research Frontiers

  • Wang et al. (2023). “Voyager: An Open-Ended Embodied Agent with Large Language Models” - arXiv:2305.16291
  • Park et al. (2023). “Generative Agents: Interactive Simulacra of Human Behavior” - arXiv:2304.03442

Automated Checks

The following checks run automatically in CI:

Check Frequency Status
Link validation Weekly Configured
arXiv DOI resolution Weekly Configured
GitHub URL accessibility Weekly Configured
Broken internal links On PR Configured

Reporting Issues

Found an incorrect claim or broken citation?

  1. Open an issue with:
    • Lecture number and slide
    • The claim in question
    • Correct information with source

We verify all reports within one week.


Verification methodology based on academic publishing standards.


Back to top