|
1 | | -##CONTEXT |
| 1 | +###Agent Laboratory Research Track & Real-Time Visualization |
2 | 2 |
|
3 | | -I have noticed that there has been quite some difficulty in developing this part of the application. This is not acceptable because the LMM researcher is the most important component. It is a critical success factor. I see on the readme file on GitHub that you can choose between the fast track, a quick AI-generated proposal that takes 30 seconds, which is great, and the research track, which involves deep autonomous research with real-time team visualisations. However, it states that the research track currently uses a mock implementation and not real AI researchers, as those require additional Python setup. This issue needs to be addressed for the entire application to function properly. To assist with this, I will provide the website, the academic paper, an example, and the code base in the links below Your task is to study all of these j |
4 | 3 |
|
5 | | -## RESOURCES |
| 4 | +###Context |
6 | 5 |
|
7 | | -# Website: https://agentlaboratory.github.io/ |
| 6 | +The current application distinguishes between two paths: |
8 | 7 |
|
9 | | -# Code: https://github.com/SamuelSchmidgall/AgentLaboratory.git |
| 8 | +1.Fast track – quick AI-generated proposal in ≈30 seconds. |
| 9 | +2.Research track – deep autonomous research with real-time multi-agent visualisations. |
10 | 10 |
|
11 | | -Use the installation guide to ensure everything is developed correctly. I will also leave the installation guide as its own document in the Resource agent research folder. |
| 11 | +At the moment, the research track relies on a mock implementation rather than real LLM researchers, because the full Agent Laboratory Python pipeline has not been wired in. This is unacceptable given that the LLM researcher is a critical success factor for the entire system. This repository aims to replace the mock path with a faithful integration of Agent Laboratory: Using LLM Agents as Research Assistants.-https://agentlaboratory.github.io/ |
12 | 12 |
|
13 | | -# Installation Guide: https://gitdocs1.s3.amazonaws.com/digests/samuelschmidgall-agentlaboratory.git/4bc5a4a6-93e9-447c-ae02-1b72fd2a7da1.txt |
| 13 | +###Upstream Specification & Resources |
14 | 14 |
|
15 | | -# Citations: |
| 15 | +Agent Laboratory is an autonomous LLM-based framework that takes a human research idea and produces a research report plus a code repository by progressing through three main phases: (1) Literature Review, (2) Experimentation, and (3) Report Writing - https://github.com/SamuelSchmidgall/AgentLaboratory |
16 | 16 |
|
17 | | -@misc{schmidgall2025agentlaboratoryusingllm, |
18 | | - title={Agent Laboratory: Using LLM Agents as Research Assistants}, |
19 | | - author={Samuel Schmidgall and Yusheng Su and Ze Wang and Ximeng Sun and Jialian Wu and Xiaodong Yu and Jiang Liu and Michael Moor and Zicheng Liu and Emad Barsoum}, |
20 | | - year={2025}, |
21 | | - eprint={2501.04227}, |
22 | | - archivePrefix={arXiv}, |
23 | | - primaryClass={cs.HC}, |
24 | | - url={https://arxiv.org/abs/2501.04227}, |
| 17 | +###Authoritative resources: |
| 18 | + |
| 19 | +##Website: https://agentlaboratory.github.io/ – high-level overview of workflow, phases, and agent roles. |
| 20 | + |
| 21 | +##Code (reference implementation): https://github.com/SamuelSchmidgall/AgentLaboratory.git – Python framework and CLI for running the full pipeline. - https://github.com/SamuelSchmidgall/AgentLaboratory |
| 22 | + |
| 23 | +##Installation guide: (mirrored in resources/agent-research/installation-guide.txt) – step-by-step environment setup, dependencies, and commands. |
| 24 | + |
| 25 | +##Paper: Schmidgall et al., Agent Laboratory: Using LLM Agents as Research Assistants, arXiv:2501.04227. - https://arxiv.org/pdf/2501.04227 |
| 26 | + |
| 27 | +This project treats the above implementation and paper as the source of truth for the research workflow and agent behavior. |
| 28 | + |
| 29 | +##Goal for This Repository |
| 30 | +Implement the full research track by correctly installing and invoking the upstream Agent Laboratory Python workflow (three phases, mle-solver, paper-solver) from this application. |
| 31 | + |
| 32 | +Provide a real-time, low-fidelity visualization of the agent team and research stages, so humans can see and grant lab workers in the process as it runs. |
| 33 | + |
| 34 | +The LLM researcher path must be fully functional (no mocks) for the overall system to be considered production-ready. |
| 35 | + |
| 36 | +### Agent Laboratory Real-Time Visualization |
| 37 | + |
| 38 | +Vision |
| 39 | +Low-fidelity pixel-art game view showing the multi-agent research team working in real time, with a breadcrumb trail that demonstrates progress through the 7 research stages. The breadcrumb mockup is stored in the breadcrumb/ folder. The visual goal is to resemble a real academic lab monitoring screen. |
| 40 | + |
| 41 | +Technical Stack |
| 42 | +Engine: Litecanvas (~4KB HTML5 canvas engine) – minimal game-dev complexity. - https://github.com/litecanvas/game-engine.git |
| 43 | + |
| 44 | +Game logic: Driven by an SSE stream using the standard EventSource API; each character reflects backend state so users can see how data flows through the system. |
| 45 | + |
| 46 | +Alternative: Plain Canvas loop (~100 lines of code) if Litecanvas is not desired. |
| 47 | + |
| 48 | +Data source: SSE stream from `/api/research/:id/stream` (emits Agent Laboratory stage and agent updates). |
| 49 | + |
| 50 | +Visual Elements |
| 51 | +Base tiles and props from the Land of Pixels laboratory tileset (included in this repo), plus custom character sprites. |
| 52 | + |
| 53 | +1. Horizontal breadcrumb timeline (7 stages across the top) styled to match the tileset. |
| 54 | + |
| 55 | +2. Agent sprites below the timeline (4–5 researcher characters representing key Agent Laboratory roles). |
| 56 | + |
| 57 | +3. Real-time state updates driven by SSE events. |
| 58 | + |
| 59 | +4. Clicking a stage opens a panel with logs and artefacts for that stage (e.g., retrieved papers, experiment runs, report drafts). |
| 60 | + |
| 61 | +5. There is a section on the side that brings the text following to the visual UI/UX so the user can follow what is happening not only visually. |
| 62 | + |
| 63 | +SSE Event Schema |
| 64 | +json |
| 65 | + |
| 66 | +{ |
| 67 | + "type": "stage_started", // stage_started | stage_completed | agent_active | agent_idle |
| 68 | + "stage": "Literature Review", // one of the 7 research stages |
| 69 | + "agent": "lit_reviewer", // agent identifier |
| 70 | + "message": "Found 12 relevant papers..." |
25 | 71 | } |
26 | 72 |
|
27 | | -Overview. Agent Laboratory begins with the independent collection and analysis of relevant research papers, progresses through collaborative planning and data preparation, and results in automated experimentation and comprehensive report generation. As shown in Figure 2, the overall workflow consists of three primary phases: (1) Literature Review, (2) Experimentation, and (3) Report Writing. In this section, we will introduce these phases in detail along with the corresponding involved agents. Furthermore, in Section 4, we will conduct qualitative and quantitative analyses to demonstrate the strengths of Agent Laboratory and its ability to generate research. |
| 73 | +All visualisation logic consumes events in this format, independent of the upstream Python implementation. |
| 74 | + |
| 75 | +## Implementation Priority |
| 76 | +Phase 3 (stretch goal) within the broader roadmap. |
| 77 | + |
| 78 | +Non-blocking for Phase 1 (Discovery) and Phase 2 (Research), but required for a complete research-track experience. During discovery and guidelines, documents, topics, and guidelines must be passed to the agent lab team. |
| 79 | + |
| 80 | +Target outcome: approximately 5× faster innovation of new business models and significantly better and more original answers to the questions the research grants are trying to solve. |
| 81 | + |
| 82 | +## Why This Approach Works |
| 83 | +Plug-and-play: Litecanvas is a tiny (~4KB gzipped) drop-in HTML5 canvas engine, so no full game engine is required. |
| 84 | + |
| 85 | +Real-time by design: SSE is a natural fit for streaming agent events into the viewer without polling, keeping the implementation simple and efficient. |
| 86 | + |
| 87 | +Debuggable: Agent visualisation and trace tools (such as AgentPrism) show that timeline-style views of agent behaviour can cut debugging time dramatically, motivating this design choice. |
| 88 | + |
| 89 | +Reusable: The viewer and SSE schema are generic; other Agent Lab–style projects can reuse this component with minimal changes. |
| 90 | + |
| 91 | +## Overview |
| 92 | + |
| 93 | +The top-down game view gives you a live dashboard showing which agents are active, what stage they're in, and lets users drill into artefacts at each stage - turning the "black box" AI research into a transparent, engaging process that builds trust and makes debugging 80% faster. To facilitate this you will create |
| 94 | + |
0 commit comments