A full-stack application that provides batch inference capabilities for math problem solving.
Most of the frontend is developed using Cursor
- Batch Inference
- Upload CSV files with math problems
- Configurable parameters (model name, system prompt, max sequences, max length)
- Progress tracking and results visualization
- Export results to CSV
- Python 3.8+
- Node.js 14+
- CUDA-capable GPU
- SGLang library
- Your model
- Clone the repository:
git clone https://github.com/bogoconic1/aimo-inference-service.git
cd aimo-inference-service
- Initial Setup:
bash setup.sh
- Start the backend server:
cd backend
python main.py
- Start the frontend development server:
cd frontend
npm start
The application will be available at:
- Frontend: http://localhost:3000
- Backend API: http://localhost:8000
- Upload a CSV file with 'id', 'problem', and 'answer' columns
- Configure max_num_seqs (default: 3) and max_length (default: 1000) parameters
- Click Process Batch
- View results in the table
- Download results as CSV
The input CSV file should have the following columns:
- id: A unique identifier for each problem
- problem: The math problem text
- answer: The correct answer
Example:
id,problem,answer
1,"What is 2 + 2?",4
2,"Solve for x: 3x + 5 = 20",5
Click "Submit" and leave it all to the server after that! Intermediate progress will be shown as the evaluation goes!
Apache 2.0