This project demonstrates:
- Native C++ Performance - Zero Python overhead in runtime
- OpenVINO Integration - Full utilization of Intel's inference framework
- Multimodal Capabilities - Seamless text and image understanding
- Production-Ready Features - Caching, error handling, clean architecture
# Build from source
mkdir build && cd build
cmake .. && ninja
# Run the application
./ovasearch
# Test queries
"What's in the dog image?"
"Summarize my documents"
"reload" # Show cache performance- 384-dimensional embeddings with cosine similarity search
- Persistent cache system reducing startup time to milliseconds
- Sliding window chunking with configurable overlap
- Automatic document extraction from PDF/DOCX/PPTX
- Clean C++17 code with RAII patterns
- Document processing: 50-100 docs/second
- Query latency: <100ms search + 1-3s generation
- Memory usage: ~500MB base
- Cache loading: <10ms for thousands of embeddings
main.cpp- Core implementationCMakeLists.txt- Build configurationprepare_documents.py- Document preprocessingpull_model.py- Model download utility
Created for GSoC 2026 - OpenVINO Organization GitHub: @lagmator22