Capstone Project for Google Agents Intensive
Multi-Agent Autonomous AI System for Literature Review & Research
Agents for Good Category
Academic research and professional knowledge work rely heavily on comprehensive literature reviews.
However, the traditional manual process presents significant challenges:
A thorough literature review typically requires 40–80 hours of researcher time, involving:
- Searching across multiple databases (Google Scholar, PubMed, IEEE Xplore, arXiv)
- Reading and summarizing 20–50+ papers
- Identifying thematic patterns manually
- Cross-referencing methodologies and findings
- Synthesizing insights across diverse sources
- Formatting citations and references
Manual reviews are susceptible to:
- Confirmation bias (favoring papers supporting existing hypotheses)
- Coverage gaps (limited by search skills and database access)
- Incomplete synthesis (difficult to spot subtle cross-paper patterns)
- Citation errors & formatting inconsistencies
Not everyone has equal access to:
- Expensive academic database subscriptions
- Time to conduct thorough reviews
- Training in systematic review methodologies
- Tools for managing large volumes of research
This project addresses the Agents for Good challenge by creating an AI-driven research assistant that:
Allows students, independent researchers, and professionals in developing regions to conduct high-quality literature reviews without costly subscriptions or specialized expertise.
Reduces hours of mechanical effort (searching, summarizing, formatting) so researchers can focus on innovation, hypothesis-building, and analysis.
Ensures consistent, unbiased, and comprehensive analysis—eliminating human cognitive limitations.
Gives smaller institutions the same analytical capabilities enjoyed by well-funded universities.
A system is required that:
- Automates the mechanical aspects of literature review
- Maintains academic rigor
- Reduces time and effort
- Makes quality research accessible to all
Mukti Scholar delivers this capability at scale.
This capstone project presents an autonomous multi-agent AI system that transforms a research topic into a publication-ready literature review in under 2 minutes.
Searches multiple academic databases in parallel, identifying relevant papers across:
- Google Scholar
- Semantic Scholar
- arXiv
Uses advanced NLP to:
- Summarize each paper’s methodology, findings, and contributions
- Cluster papers into thematic groups via embeddings
- Identify methodological patterns and contradictions
- Detect research gaps using systematic cross-paper analysis
Generates:
- A coherent literature review with academic structure
- Critical evaluation of evidence across papers
- Trend & pattern identification
- Fully formatted citations (APA, Harvard, IEEE)
Produces:
- A comprehensive PDF report with executive summary
- Quality metrics: coverage, coherence, completeness
- Exportable structured data for further research
Below is the overall architecture.
This notebook implements a comprehensive multi-agent system for automated literature review generation, demonstrating all ADK concepts from the 5‑day course.
- 10 Specialized Agents + 1 Orchestrator
- Multi-agent patterns: Sequential, Parallel, Loop
- Tools: OpenAPI, MCP, Custom Tools, Built‑in Tools
- Sessions & Memory: State management + long-term Memory Bank
- Observability: Logging, Tracing, Metrics
- Deployment-ready: Vertex AI Agent Engine compatible
- LLM Agents: Topic Understanding, Gap Analysis, Review Writer
- Parallel Agents: Multi-source paper search, parallel summarization
- Sequential Agents: Comparative analysis workflow
- Loop Agents: PDF retrieval with retry logic
- Orchestrator Agent: Coordinates all 10 specialist agents
-
Custom Tools:
search_google_scholarcluster_embeddingsformat_citation
-
Built-in Tools:
- Google Search
- Code Execution (for clustering and evaluations)
-
MCP Tools:
- PDF parsing & file operations (production-ready simulation)
-
OpenAPI Tools:
- Scholar API
- arXiv API
- Semantic Scholar API
-
Agent-as-a-Tool:
- Sub-agents callable by orchestrator
-
Long‑Running Operations:
- Resume/pause via ResumabilityConfig
- InMemorySessionService for per-run state
- Memory Bank for long-term preferences, topics, history
- Vector Store for embeddings + semantic search
- RAG Pattern for literature review writing
- Context Compaction via layered summarization
- Embedding-based thematic clustering
- Structured Logging
- Tracing with IDs
- Metrics:
- Papers found
- Processing duration
- Cluster coherence
- Summary quality
- Coverage metrics
- Cluster coherence scoring
- Writing quality analysis
- Gap analysis quality
- End-to-end grading (A–F scale)
- Functions exposed as A2A agent
- Remote agent consumption examples
- Automatically generated Agent Card
- Vertex AI Agent Engine configuration
- Scaling & resource limits
- Environment configuration
- Deployment commands ready
- Complete 10-Agent Pipeline with clear roles
- Production-quality code with type hints & robust error handling
- Extensive comments & explanations
- End-to-end demo with sample outputs
- Test suite for automated validations
- State export/import for reproducibility
Role: Master coordinator managing the entire workflow
Type: LLM Agent with sub-agents as tools
Model: Gemini 2.5 Flash Lite
- Execute agents in sequence (1 → 2 → … → 10)
- Track workflow state
- Handle errors (retry/skip/abort logic)
- Maintain session state
- Report progress to the user
- All 10 specialized agents wrapped as AgentTool
- Access to session + memory services
- Determine if each stage succeeded
- Adjust parameters (e.g., reduce paper count on timeout)
- Decide when to continue vs. wait for human input
Type: LLM Agent
Model: Gemini 2.5 Flash Lite
- Analyze user topic
- Extract 10–15 core keywords
- Identify 3–5 subdomains
- Generate 15–20 optimized search queries
Tools: Built-in LLM
Output Example:
{
"expanded_topic": "Detailed expanded topic",
"keywords": ["machine learning", "risk modeling"],
"subdomains": ["credit scoring", "deep learning"],
"search_queries": ["ML credit risk modeling", "AI banking prediction"]
}Type: Parallel Agent (3 sub-agents)
Sub-Agents:
- Google Scholar Agent
- arXiv Agent
- Semantic Scholar Agent
- Execute search queries across all sources
- Gather metadata
- Deduplicate by DOI/title
- Rank and select top 10–20 papers
- Custom OpenAPI tools:
search_google_scholar()search_arxiv()search_semantic_scholar()
Reduces search time from 45s → 15s.
Type: Loop Agent
Model: Gemini 2.5 Flash Lite
- Download PDFs
- Extract text + detect sections
- Parse tables & figure metadata
- Normalize text
- Retry failed downloads (up to 3 times)
- MCP:
download_and_extract_pdf(url) - Production tools: PyPDF2 / GROBID
Type: Parallel Agent
Model: Gemini 2.5 Flash Lite
- 20-word micro-summary
- 150-word full summary
- Extract methodology & findings
- Identify contributions & limitations
- Assess relevance
{
"micro_summary": "...",
"long_summary": "...",
"methodology": "...",
"findings": "...",
"contributions": "...",
"limitations": "...",
"relevance_notes": "..."
}Type: LLM Agent + Code Execution
Uses k-means to cluster paper embeddings.
- Generate embeddings
- Perform clustering
- Create theme labels
- Write theme descriptions
cluster_embeddings()- Built-in code execution
Type: Sequential Agent
Sub-Agents: Theme Analyzer + Cross-Theme Comparator
- Compare within-theme methodologies
- Identify contradictions
- Highlight patterns
- Produce comparison matrices
Type: LLM Agent
- Detect methodological, empirical, theoretical, geographic gaps
- Provide evidence for each gap
- Generate research questions
Type: LLM Agent with RAG
- Write structured lit review: intro, themes, comparisons, gaps, conclusion
- Retrieve relevant summaries from vector DB
- Cite papers using internal IDs
Type: Deterministic Agent
- Replace citation markers
- Generate
.bibfile - Format APA/Harvard/IEEE references
Type: Assembly Agent
- Generate PDF review
- Create visuals (theme clusters, comparison tables)
- Export reference lists
The Mukti Scholar Agent delivers a level of speed, scale, consistency, and accessibility that manual literature reviews cannot match.
Manual reviews take 40–70+ hours.
The automated system completes the same workflow in 1–2 minutes, delivering ~99% time savings and enabling rapid iteration.
Manual searches are limited by skills, database access, and time—typically covering 20–30 papers.
The automated system scans 50+ papers across multiple sources, generates optimized queries, and uncovers cross-domain insights using semantic similarity.
Human reviews vary with fatigue and bias.
The agent applies uniform, unbiased analysis, ensures standardized summarization, and produces error-free citations every time.
Manual reviews often miss contradictory or adjacent-field evidence.
The agent uses relevance scoring and clustering to surface both supporting and opposing findings, ensuring balanced synthesis.
Human reviews differ across researchers.
The system produces consistent, audit-ready outputs, easily re-run when new papers appear.
Manual reviews demand subscriptions, training, and weeks of effort.
The agent works with open sources, requires no specialized training, and completes the task in minutes, not weeks.
- Rapid background reviews for grant proposals
- High-quality gap identification
- Quick literature updates before submission
- Strong thesis-ready literature reviews
- Faster ramp-up in new research domains
- Access to professional-grade analytical tools
- Democratized access to deep literature analysis
- Standardized quality across teams
- Reduced barriers for under-resourced researchers
- Faster knowledge synthesis
- Better identification of unmet research needs
- Accelerated scientific progress
The agent doesn’t replace human judgment—it amplifies it. Researchers remain the interpreters and innovators, but they operate from a foundation of comprehensive, unbiased, and instantly generated insights, transforming research productivity.
The Mukti Scholar Agent provides significant time and productivity gains, enabling faster, more efficient academic work.
- Students reduce a ~17-week manual literature review process to ~3 days, accelerating thesis progress by approx. 4 months and enabling exploration of more research directions within the same academic cycle.
- Researchers cut a ~50-hour background review for proposals to ~4 hours, saving approx. 46 hours per proposal — equal to approx. 3.5–4.5 workweeks per year for those submitting multiple proposals.
- Eliminates mechanical tasks (searching, screening, summarizing)
- Allows researchers to focus on analysis, insight generation, and expert judgment
- Supports rapid evaluation of multiple literature landscapes
- Levels academic capability for institutions with limited resources
- 16+ weeks saved per student literature review
- Over 90% reduction in research preparation time
- Massively increases throughput for both students and researchers
