Models / Chat / Llama 3.3 70B Instruct Turbo Free
Llama 3.3 70B Instruct Turbo Free
Free
LLM
Free endpoint to try this 70B multilingual LLM optimized for dialogue, excelling in benchmarks and surpassing many chat models.
Try our Llama 3.3 Free API

Free
API Usage
How to use Llama 3.3 70B Instruct Turbo FreeModel CardPrompting Llama 3.3 70B Instruct Turbo FreeApplications & Use CasesAPI Usage
Endpoint
meta-llama/Llama-3.3-70B-Instruct-Turbo-Free
RUN INFERENCE
curl -X POST "https://api.together.xyz/v1/chat/completions" \
-H "Authorization: Bearer $TOGETHER_API_KEY" \
-H "Content-Type: application/json" \
-d '{
"model": "meta-llama/Llama-3.3-70B-Instruct-Turbo-Free",
"messages": [{"role": "user", "content": "What are some fun things to do in New York?"}]
}'
JSON RESPONSE
RUN INFERENCE
from together import Together
client = Together()
response = client.chat.completions.create(
model="meta-llama/Llama-3.3-70B-Instruct-Turbo-Free",
messages=[{"role": "user", "content": "What are some fun things to do in New York?"}],
)
print(response.choices[0].message.content)
JSON RESPONSE
RUN INFERENCE
import Together from "together-ai";
const together = new Together();
const response = await together.chat.completions.create({
messages: [{"role": "user", "content": "What are some fun things to do in New York?"}],
model: "meta-llama/Llama-3.3-70B-Instruct-Turbo-Free",
});
console.log(response.choices[0].message.content)
JSON RESPONSE
Model Provider:
Meta
Type:
Chat
Variant:
Instruct
Parameters:
70B
Deployment:
✔ Serverless
Quantization
FP8
Context length:
128K
Pricing:
Free
Run in playground
Deploy model
Quickstart docs
How to use Llama 3.3 70B Instruct Turbo Free
Model details
Prompting Llama 3.3 70B Instruct Turbo Free
Applications & Use Cases
Looking for production scale? Deploy on a dedicated endpoint
Deploy Llama 3.3 70B Instruct Turbo Free on a dedicated endpoint with custom hardware configuration, as many instances as you need, and auto-scaling.
