Skip to content

ctournas/multimodal_rag

Repository files navigation

📊 Multimodal Data Analysis for Starbucks Finance

🚀 Project Overview

This project explores multimodal AI by integrating audio transcription, NLP, and finance-related data processing. Using Whisper for audio transcription, Sentence Transformers for NLP, and LangChain/OpenAI models for analysis, the project aims to extract insights from multiple data types—text, audio, and PDFs.

🛠️ Technologies Used

  • OpenAI Whisper – Speech-to-text transcription
  • LangChain & OpenAI APIs – NLP and LLM-based analysis
  • Sentence Transformers – Semantic similarity search
  • PDF Processing – Extracting text and images from financial documents
  • Torch & Sklearn – Machine learning utilities

📌 Features

  • 🔊 Audio Transcription: Convert spoken content into text
  • 📄 PDF Parsing: Extract structured data from documents
  • 🔍 Semantic Search: Identify key insights using embeddings
  • 📈 Finance-Oriented Analysis: Apply ML/NLP techniques to financial data

📂 Dataset

The project utilizes Starbucks financial data and multimodal content, including audio, text, and documents.

🔧 Installation & Setup

pip install openai langchain langchain-openai langchain-community openai-whisper sentence-transformers pdf2image
apt-get install poppler-utils
pip install --upgrade Pillow

🚀 Usage

  1. Clone the repository:
    git clone https://github.com/ctournas/multimodal-starbucks-finance.git
    cd multimodal-starbucks-finance
  2. Install dependencies (see installation section above)
  3. Run the Jupyter notebook

📜 License

This project is licensed under the MIT License. Feel free to modify and use it for your own projects!

🤝 Contributing

Pull requests are welcome! For major changes, please open an issue first to discuss what you would like to change.

📬 Contact

For questions or collaborations, feel free to reach out via GitHub or email at [email protected].

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published