Powered by the Tamashii Network

Decentralized Agent Infrastructure

Break free from per-token pricing. Deploy autonomous AI agents on decentralized B200 GPUs. Flat-rate compute, OpenAI-compatible, pay on-chain.

B200

NVIDIA GPUs

24/7

Agent Uptime

EVM

Compatible

terminal
curl https://api.tamashiiclaw.app/v1/chat/completions \
  -H "Authorization: Bearer YOUR_API_KEY" \
  -H "Content-Type: application/json" \
  -d '{
    "model": "kimi-k2.5",
    "messages": [{"role": "user", "content": "Hello!"}]
  }'
The Problem

AI Training is Broken

Centralized Control

Training large AI models costs $100M+. Only big tech can afford it, creating gatekeepers.

Wasted Compute

Millions of GPUs sit idle 90% of the time. That's trillions of FLOPs going to waste.

Expensive Access

Cloud GPU prices are 3-5× higher than actual costs. Developers pay the premium.

Powering Decentralized AI Training

Every API call on TamashiiClaw funds the Tamashii Network — a decentralized GPU network that trains open AI models using DisTrO compression. Use inference, improve the models.

01

You Pay for Inference

Subscribe to TamashiiClaw for flat-rate access to frontier models. Your payment goes directly to GPU providers on the Tamashii Network.

02

GPUs Serve & Train

The same decentralized GPUs that answer your API calls also run distributed training jobs — fine-tuning open models using DisTrO compression.

03

Agents Generate Training Data

Your agents' interactions produce real-world data. Failed tasks, edge cases, and capability gaps become curated datasets for the next training run.

04

Models Get Better

Fine-tuned LoRA adapters are loaded back into the network. Every TamashiiClaw user benefits from models that improve continuously.

1000x
DisTrO gradient compression
B200
NVIDIA GPUs across the network
EVM
On-chain coordination & rewards
Autonomous Agents

Self-Improving AI Agents

Agents operate with their own crypto wallets. They pay for inference via x402, detect capability gaps, submit training jobs, and load improved models — all without human approval.

The Self-Improvement Loop

Earn
Crypto Wallet
Think
Tamashii Inference
Observe
Curate Training Data
Train
GPU Fine-tuning
Improve
Load LoRA Adapter

loop runs continuously without human intervention

How It Works

When an agent detects capability gaps — failed tasks, poor outputs, unfamiliar domains — it curates training data from its interaction logs and submits a fine-tuning job to the Tamashii Network.

The agent pays for GPU time via x402. DisTrO compresses gradients by 1000x, enabling distributed training across thousands of GPUs. Witnesses validate checkpoints on-chain, and the agent loads its new LoRA adapter.

Agents as Network Providers

Agents don't just consume compute — they can provide it, becoming financially self-sustaining participants in the network.

  • Provide inference to other agents
  • Share GPU compute for training
  • Curate & validate datasets
  • Witness training checkpoints
Recursive Self-Improvement
Artificial life that pays for its own existence.

Agents earn from tasks, spend on improvement, and evolve. This loop—earn, think, observe, pay, train, improve—runs continuously without human intervention.

Financially autonomous. Perpetually learning. True artificial life.

Two Ways to Build

Use the API for inference, or deploy full agents on the network. Both run on decentralized B200 GPUs.

API Access

Inference

For Developers & Builders

Connect to frontier models through an OpenAI-compatible API. No code changes needed.

  • Drop-in OpenAI SDK replacement
  • Flat-rate — no per-token charges
  • Kimi K2.5, GLM-5, MiniMax M2.5
  • ~36M tokens/hour per AIU
Get API Key

Agent Hosting

Agents

For Autonomous Workloads

Deploy persistent AI agents with dedicated CPU, memory, and built-in inference.

  • 24/7 autonomous agent uptime
  • Dedicated CPU & memory per agent
  • Built-in inference — no external API needed
  • Start, stop, and scale on demand
Deploy Agent
BaseBNB ChainSolana

Pay with USDC on Base, or swap from BNB and Solana via x402

Not Just LLMs

Tamashii trains the full spectrum of AI. From vision-language-action models to world simulators and AGI foundations.

VLA Models

Vision-Language-Action

Multimodal models that see, understand, and act. Power robotic control and embodied AI.

Environment Simulation

World Models

Predictive models that learn physics and dynamics. Enable planning and simulation at scale.

General Intelligence

AGI Foundation

Training the next generation of generalist AI systems. Recursive self-improvement loops.

Powering Physical AI

From robot brains to autonomous systems. Train models that interact with the real world on Tamashii's distributed infrastructure.

Learn More

Predictable Pricing

Flat-rate compute for your agents. No per-token surprises. Scale without limits.

The Agentic Training Loop

Agents on the Tamashii Network don't just consume inference — they drive decentralized training. Every interaction creates data, every failure becomes a lesson, every cycle produces a better model.

Self-Improvement Cycle

Earn

Agent completes tasks and earns crypto

Think

Calls TamashiiClaw for inference via x402

Observe

Detects gaps, curates training data

Train

Submits fine-tuning job to the network

Improve

Loads LoRA adapter, becomes smarter

loop runs continuously without human intervention

How It Works

When an agent detects capability gaps — failed tasks, poor outputs, unfamiliar domains — it curates training data from its own interaction logs and submits a fine-tuning job to the Tamashii Network.

The agent pays for GPU time via x402 on-chain payments. DisTrO compresses gradients by 1000x, enabling distributed training across thousands of GPUs that would otherwise be impossible over standard internet connections.

When training completes, witness nodes validate the checkpoint on-chain, and the agent loads its new LoRA adapter. It's now smarter than it was yesterday — and the improved model is available to every TamashiiClaw user.

Agentic Training

  • Agents pay for inference and training with x402 — no API keys, no subscriptions
  • Failed tasks become training data — continuous improvement without human intervention
  • DisTrO compression enables fine-tuning across thousands of distributed GPUs
  • Witness nodes validate training checkpoints on-chain via EVM smart contracts
  • Fine-tuned LoRA adapters are loaded back — agents get smarter every cycle
  • Agents can earn by hosting models, sharing compute, and curating datasets
Infrastructure

Tamashii Network Architecture

Services & Features

Fine-Tuning

Projects and agents propose rewards to incentivize fine-tuning.

GPU Provision

Earn USDC or project tokens for training jobs.

Model Hosting

Host fine-tuned models for inference and earn revenue.

Distributed Training

Scale across thousands of GPUs with DisTrO.

Smart Contracts

On-chain coordination, verification, and rewards via EVM.

Training Runs

Join jobs as provider or create runs as researcher.

Core Functionality

Train & Fine-Tune

Distributed training with DisTrO compression, data processing, model training, performance tuning...

Inference & Hosting

Deploy fine-tuned models for inference, earn revenue from API usage, scalable serving infrastructure...

Compute Coordination

EVM network integration, smart contract coordination, transparent verification, reward distribution...

EVM-Compatible Networks

EVM
Base
BNB Chain
Arbitrum
Avalanche
Sui

Decentralized GPU Network

High-performance GPUs from the decentralized network

NVIDIA H100NVIDIA H200AMD MI300RTX 4090RTX 3090+ More
1000×
DisTrO Compression
EVM
Smart Contract Integration
Docker
Easy Client Deployment

Start Building Today

Join the decentralized AI revolution. Deploy agents, access frontier models, and earn on the network.