Skip to content

TT-Studio : An all-in-one platform to deploy and manage AI models optimized for Tenstorrent hardware with dedicated front-end demo applications.

License

Notifications You must be signed in to change notification settings

NeerajPise/tt-studio

 
 

Repository files navigation

TT-Studio

To use TT-Studio's deployment features, you need access to a Tenstorrent AI accelerator.
Alternatively, you can connect to remote endpoints running models on Tenstorrent cards without local hardware.

TL;DR: TT-Studio is an easy-to-use web interface for running AI models on Tenstorrent hardware. It handles all the technical setup automatically and gives you a simple GUI to deploy models, chat with models, and more.


TT-Studio combines TT Inference Server's core packaging setup, containerization, and deployment automation with TT-Metal's model execution framework specifically optimized for Tenstorrent hardware and provides an intuitive GUI for model management and interaction.

Prerequisites

Before you start, make sure you have:

⚠️ IMPORTANT: Complete the base Tenstorrent software installation first:

Follow the Tenstorrent Getting Started Guide

This guide covers hardware setup, driver installation, and system configuration. You must complete this before using TT-Studio.

Also ensure you have:

📚 Choose Your Path

👤 I'm a Normal User

Want to start using AI models right away on your Tenstorrent hardware? This is for you!

Quick Setup:

git clone https://github.com/tenstorrent/tt-studio.git && cd tt-studio && python3 run.py

What happens step by step:

  1. Downloads TT-Studio - Gets the code from GitHub
  2. Enters the directory - Changes to the tt-studio folder
  3. Runs the setup script - Automatically configures everything
  4. Initializes submodules - Downloads TT Inference Server and dependencies
  5. Prompts for configuration - Asks for your Hugging Face token and generates security keys
  6. Builds containers - Sets up Docker environments for frontend and backend
  7. Starts all services - Launches the web interface and backend server

After Setup:

To Stop TT-Studio:

python3 run.py --cleanup

Note: This command will stop and remove all running Docker containers, including any currently deployed models. It cleans up containers and networks but preserves your data and configuration files.

🎯 What Can You Do Next?

Once TT-Studio is running:

  1. Deploy a Model - Go to the Model Deployment page and deploy a model to start using AI features
  2. Use AI Features:
    • 💬 Chat with AI models - Upload documents and ask questions
    • 🖼️ Generate images - Create art with Stable Diffusion
    • 🎤 Process speech - Convert speech to text with Whisper
    • 👁️ Analyze images - Detect objects with YOLO models
    • 📚 RAG (Retrieval-Augmented Generation) - Query your documents with AI-powered search
    • 🤖 AI Agent - Autonomous AI assistant for complex tasks

📖 Learn More: Check out our Model Interface Guide for detailed tutorials.

🆘 Need Help?

🛠️ I'm a Developer

Want to contribute to TT-Studio or modify it?

Development Mode Setup:

git clone https://github.com/tenstorrent/tt-studio.git
cd tt-studio
python3 run.py --dev

Development Features:

  • Hot Reload: Code changes automatically trigger rebuilds
  • Container Mounting: Local files mounted for real-time development
  • Automatic Setup: All submodules and dependencies handled automatically

Get Started:

Resources:


Remote Access

Running TT-Studio on a remote server? Use SSH port forwarding to access it from your local browser:

ssh -L 3000:localhost:3000 -L 8001:localhost:8001 -L 7000-7010:localhost:7000-7010 username@your-server

Note: Port range 7000-7010 forwards the model inference ports where individual AI models run.

Then open http://localhost:3000 in your local browser.


About TT-Studio

Hardware Requirements: Tenstorrent AI accelerator hardware is automatically detected when available. You can also connect to remote endpoints if you don't have direct hardware access.

TT-Studio combines TT Inference Server and TT-Metal to provide:

  • Modern Web Interface: React-based UI for easy model interaction
  • Django Backend: Robust backend service for model management and deployment
  • Vector Database: ChromaDB for document storage and semantic search
  • Multiple AI Models: Chat, vision, speech, and image generation
  • Model Isolation: Each AI model runs on separate ports (7000+) for better resource management
  • Hardware Optimization: Specifically optimized for Tenstorrent devices
  • Docker Containers: Isolated environments for frontend, backend, and inference services

Supported AI Models

  • Language Models (LLMs): Chat, Q&A, text completion
  • Computer Vision: Object detection with YOLO
  • Speech Processing: Speech-to-text with Whisper
  • Image Generation: Create images with Stable Diffusion

🛠️ For Developers

Want to contribute or customize TT-Studio?

Get Started:

Development Mode:

python3 run.py --dev  # Enables hot reload for development

Development Features:

  • Hot Reload: Code changes automatically trigger rebuilds
  • Container Mounting: Local files mounted for real-time development
  • Automatic Setup: All submodules and dependencies handled automatically

Resources:


📋 Additional Resources

Documentation

Community & Support

⚠️ Note: The startup.sh script is deprecated. Always use python3 run.py for setup and management.

About

TT-Studio : An all-in-one platform to deploy and manage AI models optimized for Tenstorrent hardware with dedicated front-end demo applications.

Resources

License

Code of conduct

Contributing

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • TypeScript 65.7%
  • Python 31.3%
  • Shell 1.6%
  • JavaScript 0.6%
  • CSS 0.5%
  • Dockerfile 0.2%
  • HTML 0.1%