Skip to content

TT-Studio : An all-in-one platform to deploy and manage AI models optimized for Tenstorrent hardware with dedicated front-end demo applications.

License

Notifications You must be signed in to change notification settings

tenstorrent/tt-studio

TT-Studio

To use TT-Studio's deployment features, you need access to a Tenstorrent AI accelerator.
Alternatively, you can connect to remote endpoints running models on Tenstorrent cards without local hardware.

TL;DR: TT-Studio is an easy-to-use web interface for running AI models on Tenstorrent hardware. It handles all the technical setup automatically and gives you a simple GUI to deploy models, chat with models, and more.


TT-Studio combines TT Inference Server's core packaging setup, containerization, and deployment automation with TT-Metal's model execution framework specifically optimized for Tenstorrent hardware and provides an intuitive GUI for model management and interaction.

Prerequisites

Before you start, make sure you have:

⚠️ IMPORTANT: Complete the base Tenstorrent software installation first:

Follow the Tenstorrent Getting Started Guide

This guide covers hardware setup, driver installation, and system configuration. You must complete this before using TT-Studio.

Also ensure you have:

⚠️ IMPORTANT - Docker Group Setup:

After installing Docker, you must add your user to the docker group to run Docker commands without sudo:

sudo usermod -aG docker $USER

Then log out and log back in (or restart your system) for the group change to take effect.

To verify your user is in the docker group:

groups | grep docker

If you see "docker" in the output, you're ready to go! If not, make sure you've logged out and back in after running the usermod command.

πŸ“š Choose Your Path

πŸ‘€ I'm a Normal User

Want to start using AI models right away on your Tenstorrent hardware? This is for you!

Quick Setup:

git clone https://github.com/tenstorrent/tt-studio.git && cd tt-studio && python3 run.py --easy

What happens step by step:

  1. Downloads TT-Studio - Gets the code from GitHub
  2. Enters the directory - Changes to the tt-studio folder
  3. Runs the setup script - Automatically configures everything (easy mode: only prompts for your Hugging Face token; uses defaults for the rest)
  4. Initializes submodules - Downloads TT Inference Server and dependencies
  5. Builds containers - Sets up Docker environments for frontend and backend
  6. Starts all services - Launches the web interface and backend server

⚠️ Security Note: Easy mode uses default values that are NOT secure for production. Use this mode only for development, testing, and quick evaluation. For production deployments, use the standard setup (python3 run.py) or development mode (python3 run.py --dev).

πŸ“– More Details: See the Complete run.py Guide for a full comparison of setup modes.

After Setup:

To Stop TT-Studio:

python3 run.py --cleanup

Note: This command will stop and remove all running Docker containers, including any currently deployed models. It cleans up containers and networks but preserves your data and configuration files.

🎯 What Can You Do Next?

Once TT-Studio is running:

  1. Deploy a Model - Go to the Model Deployment page and deploy a model to start using AI features
  2. Use AI Features:
    • πŸ’¬ Chat with AI models - Upload documents and ask questions
    • πŸ–ΌοΈ Generate images - Create art with Stable Diffusion
    • 🎀 Process speech - Convert speech to text with Whisper
    • πŸ‘οΈ Analyze images - Detect objects with YOLO models
    • πŸ“š RAG (Retrieval-Augmented Generation) - Query your documents with AI-powered search
    • πŸ€– AI Agent - Autonomous AI assistant for complex tasks

πŸ“– Learn More: Check out our Model Interface Guide for detailed tutorials.

πŸ†˜ Need Help?

πŸ› οΈ I'm a Developer

Want to contribute to TT-Studio or modify it?

Development Mode Setup:

git clone https://github.com/tenstorrent/tt-studio.git
cd tt-studio
python3 run.py --dev

Development Features:

  • Hot Reload: Code changes automatically trigger rebuilds
  • Container Mounting: Local files mounted for real-time development
  • Automatic Setup: All submodules and dependencies handled automatically

Get Started:

Resources:


Remote Access

Running TT-Studio on a remote server? Use SSH port forwarding to access it from your local browser:

ssh -L 3000:localhost:3000 -L 8001:localhost:8001 -L 7000-7010:localhost:7000-7010 username@your-server

Note: Port range 7000-7010 forwards the model inference ports where individual AI models run.

Then open http://localhost:3000 in your local browser.


About TT-Studio

Hardware Requirements: Tenstorrent AI accelerator hardware is automatically detected when available. You can also connect to remote endpoints if you don't have direct hardware access.

TT-Studio combines TT Inference Server and TT-Metal to provide:

  • Modern Web Interface: React-based UI for easy model interaction
  • Django Backend: Robust backend service for model management and deployment
  • Vector Database: ChromaDB for document storage and semantic search
  • Multiple AI Models: Chat, vision, speech, and image generation
  • Model Isolation: Each AI model runs on separate ports (7000+) for better resource management
  • Hardware Optimization: Specifically optimized for Tenstorrent devices
  • Docker Containers: Isolated environments for frontend, backend, and inference services

Supported AI Models

  • Language Models (LLMs): Chat, Q&A, text completion
  • Computer Vision: Object detection with YOLO
  • Speech Processing: Speech-to-text with Whisper
  • Image Generation: Create images with Stable Diffusion

πŸ› οΈ For Developers

Want to contribute or customize TT-Studio?

Get Started:

Development Mode:

python3 run.py --dev  # Enables hot reload for development

Development Features:

  • Hot Reload: Code changes automatically trigger rebuilds
  • Container Mounting: Local files mounted for real-time development
  • Automatic Setup: All submodules and dependencies handled automatically

Resources:


πŸ“‹ Additional Resources

Documentation

Community & Support

⚠️ Note: The startup.sh script is deprecated. Always use python3 run.py for setup and management.

About

TT-Studio : An all-in-one platform to deploy and manage AI models optimized for Tenstorrent hardware with dedicated front-end demo applications.

Resources

License

Code of conduct

Contributing

Stars

Watchers

Forks

Packages

No packages published

Contributors 11