- Python 3.10+
- Ollama running locally
- Pull a model:
ollama pull llama3.1:8b
python -m venv .venv
. .venv/Scripts/activate # Windows PowerShell
pip install -r requirements.txt
copy .env.example .env # adjust if needed
python app.py
Service listens on http://localhost:8080.
- add relevant apis for metadata
- improve metadata framing for llm model
- add proper error handling
- improve metadata services architecture design
- return complex types (images / list of travel suggestions)