Documentation Index
Fetch the complete documentation index at: https://docs.tera.gw/llms.txt
Use this file to discover all available pages before exploring further.
What is Tera?
Tera is a hosted inference provider for open-weight language models — Llama, Qwen, DeepSeek, Mistral, and more. The API is OpenAI-compatible, so existing OpenAI SDKs and tools work without modification — point them athttps://api.tera.gw with your Tera API key.
We’re a pure execution layer for the open ecosystem. US team, US-owned GPUs, US data centers. No proprietary model to upsell you to — we focus on per-token economics on the models developers actually ship. Prompts and completions are processed and discarded: zero retention, zero training, zero human review.
What’s available today
- Chat completions —
POST /v1/chat/completions - Text completions —
POST /v1/completions - Model catalog —
GET /v1/models - Text-to-speech —
POST /v1/audio/speech(Kokoro) - Streaming via Server-Sent Events
- Reasoning models — Qwen3.5 surfaces chain-of-thought traces as a separate
reasoning_contentfield - Tool calling — OpenAI-compatible
toolsandtool_choice
Where to start
Quickstart
Make your first request in 60 seconds with curl, Python, or Node.
Authentication
How API keys work and how to request one.
Models
What’s deployed, context lengths, pricing.
OpenAI compatibility
What carries over from OpenAI clients, what’s different.