Skip to main content

Documentation Index

Fetch the complete documentation index at: https://docs.tera.gw/llms.txt

Use this file to discover all available pages before exploring further.

What is Tera?

Tera is a hosted inference provider for open-weight language models — Llama, Qwen, DeepSeek, Mistral, and more. The API is OpenAI-compatible, so existing OpenAI SDKs and tools work without modification — point them at https://api.tera.gw with your Tera API key. We’re a pure execution layer for the open ecosystem. US team, US-owned GPUs, US data centers. No proprietary model to upsell you to — we focus on per-token economics on the models developers actually ship. Prompts and completions are processed and discarded: zero retention, zero training, zero human review.

What’s available today

  • Chat completionsPOST /v1/chat/completions
  • Text completionsPOST /v1/completions
  • Model catalogGET /v1/models
  • Text-to-speechPOST /v1/audio/speech (Kokoro)
  • Streaming via Server-Sent Events
  • Reasoning models — Qwen3.5 surfaces chain-of-thought traces as a separate reasoning_content field
  • Tool calling — OpenAI-compatible tools and tool_choice

Where to start

Quickstart

Make your first request in 60 seconds with curl, Python, or Node.

Authentication

How API keys work and how to request one.

Models

What’s deployed, context lengths, pricing.

OpenAI compatibility

What carries over from OpenAI clients, what’s different.