Models / Qwen
Chat
Reasoning

Qwen3 235B A22B Instruct 2507 FP8

235B MoE model with 22B activation featuring enhanced instruction following, reasoning, and 262K context for cost-efficient high-throughput inference.

Performance benchmarks

Model

AIME 2025

GPQA Diamond

HLE

LiveCodeBench

MATH500

SWE-bench verified

65.9%

Related open-source models

Competitor closed-source models

Claude Opus 4.6

90.5%

34.2%

78.7%

OpenAI o3

83.3%

24.9%

99.2%

62.3%

OpenAI o1

76.8%

96.4%

48.9%

GPT-4o

49.2%

2.7%

32.3%

89.3%

31.0%

This model is not available on Together’s Serverless API.

Deploy this model on an on-demand Dedicated Endpoint or pick a supported alternative from the Model Library.

Related models
  • Model provider
    Qwen
  • Type
    Chat
    Reasoning
  • Main use cases
    Chat
    Small & Fast
    Medium General Purpose
    Function Calling
  • Features
    Function Calling
    JSON Mode
  • Parameters
    235B
  • Context length
    262K