Models / Chat / Qwen QwQ 32B Preview
Qwen QwQ 32B Preview
LLM
Experimental research model by Alibaba's Qwen team focused on enhancing AI reasoning capabilities.
Try our Qwen QwQ API

API Usage
Endpoint
Qwen/QwQ-32B-Preview
RUN INFERENCE
curl -X POST "https://api.together.xyz/v1/chat/completions" \
-H "Authorization: Bearer $TOGETHER_API_KEY" \
-H "Content-Type: application/json" \
-d '{
"model": "Qwen/QwQ-32B-Preview",
"messages": [{"role": "user", "content": "What are some fun things to do in New York?"}]
}'
JSON RESPONSE
RUN INFERENCE
from together import Together
client = Together()
response = client.chat.completions.create(
model="Qwen/QwQ-32B-Preview",
messages=[{"role": "user", "content": "What are some fun things to do in New York?"}],
)
print(response.choices[0].message.content)
JSON RESPONSE
RUN INFERENCE
import Together from "together-ai";
const together = new Together();
const response = await together.chat.completions.create({
messages: [{"role": "user", "content": "What are some fun things to do in New York?"}],
model: "Qwen/QwQ-32B-Preview",
});
console.log(response.choices[0].message.content)
JSON RESPONSE
Model Provider:
Qwen
Type:
Chat
Variant:
Preview
Parameters:
32B
Deployment:
✔ Serverless
Quantization
FP16
Context length:
32768
Pricing:
$1.20
Run in playground
Deploy model
Quickstart docs
Looking for production scale? Deploy on a dedicated endpoint
Deploy Qwen QwQ 32B Preview on a dedicated endpoint with custom hardware configuration, as many instances as you need, and auto-scaling.
