Models / Vision / Llama 3.2 11B Free
Llama 3.2 11B Free
Free endpoint to test this auto-regressive language model that uses an optimized transformer architecture.
![](https://cdn.prod.website-files.com/650c3b59079d92475f37b68f/6798c7d256b428d5c7991fef_66f41918314a4184b51788ed_meta-logo.png)
Lorem ipsum dolor sit amet, consectetur adipiscing elit, sed do eiusmod tempor incididunt ut labore et dolore magna aliqua. Ut enim ad minim veniam, quis nostrud exercitation eiusmod tempor incididunt ut labore sed do eiusmod tempor incididunt.
API Usage
Endpoint
RUN INFERENCE
curl -X POST "https://api.together.xyz/v1/chat/completions" \
-H "Authorization: Bearer $TOGETHER_API_KEY" \
-H "Content-Type: application/json" \
-d '{
"model": "meta-llama/Llama-Vision-Free",
"messages": [{"role": "user", "content": "What are some fun things to do in New York?"}]
}'
JSON RESPONSE
RUN INFERENCE
from together import Together
client = Together()
response = client.chat.completions.create(
model="meta-llama/Llama-Vision-Free",
messages=[{"role": "user", "content": "What are some fun things to do in New York?"}],
)
print(response.choices[0].message.content)
JSON RESPONSE
RUN INFERENCE
import Together from "together-ai";
const together = new Together();
const response = await together.chat.completions.create({
messages: [{"role": "user", "content": "What are some fun things to do in New York?"}],
model: "meta-llama/Llama-Vision-Free",
});
console.log(response.choices[0].message.content)
JSON RESPONSE
Model Provider:
Meta
Type:
Vision
Variant:
Vision
Parameters:
11B
Deployment:
✔ Serverless
Quantization
Context length:
128K
Pricing:
Free
Run in playground
Deploy model
Quickstart docs
Looking for production scale? Deploy on a dedicated endpoint
Deploy Llama 3.2 11B Free on a dedicated endpoint with custom hardware configuration, as many instances as you need, and auto-scaling.
![](https://cdn.prod.website-files.com/64f6f2c0e3f4c5a91c1e823a/6751dcee391c160684337c7e_models_visual.jpg)