Skip to main content
LLM billing follows the vendor’s input and output pricing model, then applies the TTAPI channel discount where applicable. Supported models change over time, so this page focuses on the main platforms and pricing rules rather than every single version.

Pricing rules

PlatformCommon modelsBilling rule
OpenAIgpt-4o
gpt-4o-mini
gpt-4.1
gpt-4.1-mini
o1
o3-mini
gpt-5
gpt-5-mini
gpt-5-nano
gpt-5-chat-latest
gpt-5.1
gpt-5.2
gpt-5-nano-2025-08-07
gpt-5-pro
gpt-5.2-pro
gpt-5.4-mini
gpt-5.4-nano
Official list price
Claudeclaude-sonnet-4-20250514
claude-opus-4-20250514
claude-opus-4-1-20250805
claude-haiku-4-5-20251001
claude-sonnet-4-5-20250929
claude-opus-4-5-20251101
claude-opus-4-5
claude-sonnet-4-5
claude-haiku-4-5
claude-opus-4-6
claude-sonnet-4-6
10% off official pricing
Geminigemini-2.5-flash
gemini-2.5-pro
gemini-3-pro
gemini-3-pro-preview
gemini-3-flash-preview
gemini-3.1-pro-preview
40% off official pricing
X.AIgrok-3
grok-3-fast
grok-3-mini
grok-3-mini-fast
grok-4
grok-4-fast
grok-4-1-fast-reasoning
grok-4-1-fast-non-reasoning
20% off official pricing

Notes

If a platform offers multiple versions of the same model family, the same discount rule usually applies across that platform unless TTAPI announces a special promotion or a separate pricing exception. Final cost per request still depends on actual input tokens, output tokens, context size, and model version.
Last modified on March 26, 2026