Together AI Provider

Together AI is a platform for running large language models in the cloud with fast inference.

Available Models

DeepSeek V4 Pro

deepseek
deepseek-v4-pro
Streaming
Tools
Reasoning
JSON Output
JSON Schema
Together AI
Context: 163.8k
Input
$2.1
/M tokens
Cached
$0.2
/M tokens
Output
$4.4
/M tokens

Kimi K2.6

moonshot
kimi-k2.6
Streaming
Vision
Reasoning
Together AI
Context: 262.1k
Input
$1.2
/M tokens
Cached
$0.2
/M tokens
Output
$4.5
/M tokens

GLM-5.1

glm
glm-5.1
Streaming
Tools
Reasoning
JSON Output
JSON Schema
Together AI
Context: 202.8k
Input
$1.4
/M tokens
Cached
/M tokens
Output
$4.4
/M tokens

MiniMax M2.5

minimaxModel Deactivated
minimax-m2.5
Streaming
Reasoning
JSON Output
JSON Schema
Together AI
Context: 228.7k
Deactivated since Apr 27, 2026
Input
$0.3
/M tokens
Cached
/M tokens
Output
$1.2
/M tokens

GLM-5

glmModel Deactivated
glm-5
Streaming
Tools
Reasoning
JSON Output
JSON Schema
Together AI
Context: 202.8k
Deactivated since Apr 22, 2026
Input
$1
/M tokens
Cached
/M tokens
Output
$3.2
/M tokens

Kimi K2.5

moonshot
kimi-k2.5
Streaming
Vision
Reasoning
Together AI
Context: 262.1k
Input
$0.5
/M tokens
Cached
/M tokens
Output
$2.8
/M tokens

GLM-4.7

glm
glm-4.7
Streaming
Reasoning
Together AI
Context: 202.8k
Input
$0.45
/M tokens
Cached
/M tokens
Output
$2
/M tokens

GPT OSS 120B

openai
gpt-oss-120b
Streaming
Tools
Reasoning
Together AI
Context: 131.1k
Input
$0.15
/M tokens
Cached
/M tokens
Output
$0.6
/M tokens

GPT OSS 20B

openai
gpt-oss-20b
Streaming
Tools
Reasoning
Together AI
Context: 131.1k
Input
$0.05
/M tokens
Cached
/M tokens
Output
$0.2
/M tokens

Llama 4 Scout

metaModel Deactivated
llama-4-scout
Streaming
Tools
Together AI
Context: 32.8k
Deactivated since Apr 25, 2026
Input
$0.18
/M tokens
Cached
/M tokens
Output
$0.59
/M tokens

Llama 3.1 8B Instruct

metaModel Deactivated
llama-3.1-8b-instruct
Streaming
Tools
Together AI
Context: 128k
Deactivated since Mar 27, 2026
Input
$0.06
/M tokens
Cached
/M tokens
Output
$0.06
/M tokens

MiniMax M2.7

minimax
minimax-m2.7
Streaming
Reasoning
JSON Output
JSON Schema
Together AI
Context: 196.6k
Input
$0.3
/M tokens
Cached
$0.06
/M tokens
Output
$1.2
/M tokens