Radium Tycho 1.0

Ultra-efficient inference for high-throughput applications,
early development, and lightweight agents.

Built for workloads where every token counts. Clarke delivers the lowest per-request cost in the Radium lineup without sacrificing the speed production interfaces demand.

Our Technology

High-efficiency Scale

Radium's fastest and most cost-efficient model for high-volume production workloads.

Best for

Classification and 
routing

Ticket triage, intent detection, document labeling, workflow decisioning at scale

Extraction and summarization

Structured fields from emails, PDFs, contracts, call transcripts, and support logs

Support automation

Draft responses, escalate edge cases, reduce repetitive queue volume

Fast, efficient model for high-volume classification, extraction, and automation.

Radium's most cost-efficient model for production workloads where speed, throughput, and token economics matter more than reasoning depth.

Tycho 1.0

High-efficiency Scale

Ultra-efficient inference for high-throughput applications, early development, and lightweight agents.

Pricing
input Tokens
$0.50 / MTok
output Tokens
$2.25 / MTok
Cost optimization

Significant savings with prompt caching (up to 90%) on repeated prompts.

When to use Hal

Tycho 1.0 is best suited for:

  • Classification, extraction, and summarization
  • Routing and intent detection
  • Support automation and ticket triage
  • Lightweight agent tasks
  • High-volume, cost-sensitive workloads
Equivilancy

Comparable to Opus 4.7
and GPT-5.5

Where delivery
efficiency becomes
pricing power.

Switch in one line of code and run your existing AI workloads more efficiently without rebuilding your application.

Create an account

Improve unit economics
at production scale

Case Studies

Radium is used by teams shipping
AI into real-world systems

Square’s R&D team used Radium to prototype early (pre-Gen AI) text-to-video and text-to-speech applications.

EQTY Lab used Radium to train a state-of-the-art climate model that was presented at COP28, the United Nations climate change conference.

Realbotix uses Radium to power low-latency, real-time AI interactions on its humanoid robotics platform. Radium enables responsive inference at the speed required for live human–AI interaction.

A leader in generative AI for law, Alexi used Radium to train domain-specific retrieval models. Alexi’s advanced AI platform generates legal memos, arguments, and answers to general litigation queries.

Get Started

One line of code to switch.
A different class of performance.

Swap OpenAI for Radium in your API call. That's it.