Private AI First · Optional Global Scale

Your Private AI in 30 Seconds.
Then Scale to a Global Supercomputer.

Install once. Run fully offline. No accounts. No lock-in.
When you want more power, TuTu can tap distributed compute — on your terms.

Terminal
# Install TuTu (macOS)
$ curl -fsSL https://tutuengine.tech/install.sh | sh
# Run your first AI model
$ tutu run llama3.2
pulling manifest... done
pulling 8934d96d3f08... 100% ████████████ 750 MB
success
>>> Hello!
Hello! How can I help you today?
# Install TuTu (Linux)
$ curl -fsSL https://tutuengine.tech/install.sh | sh
# Run your first AI model
$ tutu run llama3.2
pulling manifest... done
pulling 8934d96d3f08... 100% ████████████ 750 MB
success
>>> Hello!
Hello! How can I help you today?
# Install TuTu (Windows PowerShell)
PS> irm tutuengine.tech/install.ps1 | iex
# Run your first AI model
PS> tutu run llama3.2
pulling manifest... done
pulling 8934d96d3f08... 100% ████████████ 750 MB
success
>>> Hello!
Hello! How can I help you today?
30s
From Install to First Reply
<50MB
Idle Runtime Footprint
0
Accounts or API Keys Needed
100%
You Control Data + Compute

Built to Make You Faster, Smarter, and Independent.

TuTu removes friction at every step: instant local AI, zero cloud lock-in, and optional distributed scale when you need it.

Instant Local AI

Go from install to answers in under a minute. TuTu auto-downloads and optimizes top models for your hardware.

🔒

100% Offline by Default

No accounts, no cloud calls, no hidden telemetry. Sensitive prompts and data stay on your machine.

🌐

OpenAI-Compatible API

Drop-in replacement for OpenAI's API on localhost:11434. Works with LangChain, LlamaIndex, every SDK.

📦

TuTufile Packaging

Create custom models with a TuTufile — set system prompts, parameters, and adapters. Like Dockerfile for AI.

🚀

Distributed Supercomputer

Need more throughput? Opt in and unlock shared compute. Contribute idle GPU cycles and earn credits.

🛡

Enterprise-Grade Security

Ed25519 identity, gVisor sandbox, supply chain verification, Byzantine fault tolerance. Built for trust.

🤖

MCP Gateway

Enterprise AI tool access via Model Context Protocol. SLA tiers, usage metering, rate limiting built-in.

🏆

Engagement Engine

XP levels, achievements, weekly quests, passive earning. Max 1 notification/day. Respect-first design.

🌍

Planet-Scale Architecture

Built for 10M+ nodes across 7 continents. Hierarchical gossip, regional sharding, ML-driven scheduling.

From Zero to AI in 30 Seconds

No configuration. No accounts. No PhD required.

1

Install TuTu

One command on macOS, Linux, or Windows. Single binary, zero dependencies.

2

Run a Model

tutu run llama3.2 — auto-downloads and starts chatting instantly.

3

Build with the API

Use the OpenAI-compatible API at localhost:11434 with any SDK.

4

Join the Network

Your idle GPU earns credits. Every node makes AI faster and cheaper for everyone.

Drop-in Compatible

TuTu speaks OpenAI's protocol. Every tool that works with OpenAI works with TuTu — just change the URL.

🤖 OpenAI SDK
🔗 LangChain
📸 LlamaIndex
💻 Continue.dev
Cursor
🔧 AutoGen
🛠 CrewAI
🎓 Hugging Face
📈 Grafana
📊 Prometheus
🚀 Docker
Kubernetes

Install in One Command

Choose your platform. Be running AI models in under a minute.

🍎 macOS

curl -fsSL https://tutuengine.tech/install.sh | sh

Supports Intel & Apple Silicon.

🐧 Linux

curl -fsSL https://tutuengine.tech/install.sh | sh

Supports x86_64 and ARM64. APT/RPM repos coming soon.

🌏 Windows

irm tutuengine.tech/install.ps1 | iex
winget install tutu-network.tutu

PowerShell or WinGet. Requires Windows 10+.

🛠 Build from Source

git clone https://github.com/NikeGunn/tutu.git
cd tutu && go build -o tutu ./cmd/tutu

Requires Go 1.24+. Single binary, no CGO required.

Building the World's AI Supercomputer

AI should be a utility, not a tax. TuTu gives developers private local intelligence first, then optional distributed scale.

Our Mission

TuTu Network is building a distributed AI supercomputer powered by real-world idle devices. We make advanced AI accessible without forcing users into expensive cloud contracts.

Start local in one command, keep full privacy, and stay productive with an OpenAI-compatible API that drops into your existing stack.

When extra capacity matters, opt into the network to access shared compute and contribute your idle GPU for credits. You keep control over when, how, and why your machine participates.

Why Teams Switch to TuTu

Ship faster: no account setup, no API billing setup, no integration rewrite.
Reduce risk: offline-first runtime keeps sensitive prompts and customer data local.
Escape lock-in: OpenAI-compatible interfaces let you keep your current tools.
Scale on demand: add distributed compute only when you need more throughput.
🔒

Privacy First

Zero telemetry. Your data never leaves your machine. Period.

🌐

Open Source

MIT licensed. Fully transparent. Community-driven development.

Performance

Written in Go. Single binary. Sub-50MB idle footprint.

🤝

Community

Every contributor matters. Building together for everyone.

Releases & Changelog

Track our progress, download latest versions, and see what's new in every release.

📦

Follow Our Releases on GitHub

Every release includes pre-built binaries for macOS, Linux, and Windows, detailed changelogs, and upgrade instructions.

1

Visit Releases Page

Go to our GitHub releases page to see all versions, release notes, and download assets.

2

Watch the Repository

Click "Watch" → "Custom" → "Releases" on GitHub to get notified of every new version.

3

Upgrade Anytime

Run the install command again or download the latest binary. TuTu handles migrations automatically.

OpenAI-Compatible. Drop-In Ready.

Replace your OpenAI calls by pointing at localhost. Works with every SDK and framework.

🐍 Python

from openai import OpenAI client = OpenAI( base_url="http://localhost:11434/v1", api_key="tutu" # any string works ) resp = client.chat.completions.create( model="llama3.2", messages=[{ "role": "user", "content": "Hello!" }] ) print(resp.choices[0].message.content)

🔌 curl

curl \ http://localhost:11434/v1/chat/completions \ -H "Content-Type: application/json" \ -d '{ "model": "llama3.2", "messages": [{ "role": "user", "content": "Hello!" }] }'

Model Context Protocol — USB-C for AI

TuTu implements the MCP 2025-03-26 specification. Connect any AI client to any tool through one universal protocol.

🔌

Universal AI Connector

MCP is an open standard that lets AI models interact with databases, APIs, file systems, and any external service — through one consistent interface.

🏗

Enterprise Use Cases

AI coding assistants connected to your CI/CD. Support bots with CRM access. Data pipelines with SQL tools. DevOps agents managing Kubernetes — all via MCP.

📈

Metered & SLA-Backed

4 SLA tiers (Free to Enterprise). Rate limiting, usage metering, and latency targets built-in. Scale from prototype to production without changing code.

Streamable HTTP Transport

JSON-RPC 2.0 over Streamable HTTP. Real-time tool execution, resource access, and prompt serving with session management and progress notifications.

How Companies Use TuTu's MCP Gateway

1

Start TuTu Engine

Run tutu serve — the MCP endpoint is live at /mcp automatically.

2

Connect Your AI Client

Point Claude, ChatGPT, or any MCP-compatible client to http://localhost:11434/mcp.

3

AI Uses Your Tools

The AI can now run models (tutu_run), list models (tutu_list), pull new models (tutu_pull), and check status (tutu_status).

curl -X POST http://localhost:11434/mcp -H "Content-Type: application/json" -d '{"jsonrpc":"2.0","id":1,"method":"initialize","params":{"protocolVersion":"2025-03-26","clientInfo":{"name":"my-app","version":"1.0"}}}'
SLA Tier Rate Limit Latency Price
Free 10 req/min Best effort $0
Pro 100 req/min < 500ms Credits
Business 1,000 req/min < 200ms Credits
Enterprise 10,000 req/min < 100ms Credits

Earn Credits. Use AI. No Blockchain.

Contribute GPU time, earn credits. Spend credits on network AI, fine-tuning, and priority access. Simple, fair, transparent.

💰

Earn by Contributing

Your idle GPU earns credits automatically. High-end GPUs earn 2.5× more. 99%+ uptime? 1.3× reliability bonus. Early adopters get 1.5× for the first 30 days.

💸

Spend on AI

Use credits for network inference (0.001/token), fine-tuning jobs (10/hour), MCP Pro tier (50/month), and priority queue access (5/request).

🔒

Anti-Fraud Protection

Double-entry bookkeeping, velocity checks, Benford's Law analysis, and full audit trails. Every transaction is balanced and verifiable.

💲

Buy Credits (Optional)

Don't want to contribute GPU? Buy credits starting at $9.99/mo for 5,000 credits. Enterprise packages available for organizations.

Credit Pricing

Package Credits Price Best For
Starter 500 Free Everyone. Included automatically.
Developer 5,000 $9.99/mo Individual developers
Team 25,000 $39.99/mo Small teams
Enterprise 100,000 $149.99/mo Organizations
Custom Unlimited Contact us Large-scale deployments

Local AI is always 100% free. Credits are only for distributed network features. 500 free credits included for every user. No surprise charges.

Fine-Tune AI Models — Locally or Distributed

Customize any model for your use case. Use your own hardware or leverage the distributed network's GPU power with credits.

🎯

Local Fine-Tuning

Fine-tune models on your own hardware using TuTufile. Define adapters, system prompts, and training parameters — all in one declarative file. Zero cost.

🌐

Distributed Fine-Tuning

Submit LoRA/QLoRA jobs to the network. Tasks are distributed across capable peers. Pay with credits. Get results faster than training alone.

📊

LoRA & QLoRA Support

Full fine-tune (48GB+), LoRA (8GB+), or QLoRA (4GB+). Choose the method that fits your hardware and budget. Adapter merging lets you combine results.

💰

Earn While Training

Contribute GPU time for other users' fine-tuning jobs and earn credits. Your hardware works for you even when you're not using it.

Fine-Tuning & Credits: How It Works

1

Create a TuTufile

Define your base model, training data, system prompt, and adapter parameters in a simple Tutufile.

2

Train Locally or Submit to Network

Run tutu create my-model -f Tutufile locally, or tutu agent finetune --budget 100 to distribute across the network using credits.

3

Run Your Custom Model

Your fine-tuned model is ready. Run it with tutu run my-model. Share it on the marketplace if you want.

# Local fine-tuning (free)
cat > Tutufile <<EOF FROM llama3 SYSTEM "You are an expert customer support agent." ADAPTER ./my-lora-weights PARAMETER temperature 0.7 EOF
tutu create support-bot -f Tutufile
# Distributed fine-tuning (with credits)
tutu agent finetune --base-model llama3 --dataset ./data.jsonl --method lora --budget 100

Loved by Developers

From solo hackers to enterprise teams, TuTu makes AI accessible to everyone.

★★★★★

"Finally, an AI runtime that respects my privacy. Installed in 10 seconds, running Llama 3.2 offline. No accounts, no telemetry. This is how software should be."

SK
Sarah K.
ML Engineer
★★★★★

"The OpenAI-compatible API is seamless. Switched my entire LangChain pipeline from GPT-4 to local Mistral by just changing the URL. Zero code changes."

MR
Marcus R.
Full-Stack Developer
★★★★★

"The distributed compute vision is incredible. My gaming rig earns credits overnight while I sleep. It's like mining but actually useful — powering AI inference."

JL
Jason L.
Open Source Contributor

Free. Forever. For Local AI.

Local AI is 100% Free

Run any model locally without limits. No accounts, no subscriptions, no token meters. Open source forever.

Free & Open Source
Coming Soon: The distributed network uses a credit system for compute sharing. Credits are earned by contributing idle GPU time. You'll get 500 free credits + 30 days before any prompt to contribute. No surprise charges. No paywalls for local features. Ever.

Join the Movement

Be part of building the world's largest distributed AI supercomputer.