Models / OpenAI Open Weight
OpenAI
OpenAI Open Models. Deploy instantly.
Immediately start using gpt-oss-120B on Together AI's production-ready infrastructure with proven performance, high reliability and comprehensive developer tooling.

Deploy OpenAI's Open Models in Minutes
No infrastructure setup. No vendor lock-in. Complete freedom to fine-tune for specialized domains and build commercial applications with full model ownership.
Why OpenAI's Open-Weight Models Matter
Frontier AI with complete model ownership and deployment flexibility
These aren't just scaled-up versions of existing models—they're purpose-built reasoning models released under Apache 2.0 licensing while maintaining the safety standards that enterprise deployment requires.
OPEN-WEIGHT MODEL RELEASE
- OpenAI's gpt-oss-120B and gpt-oss-20B under Apache 2.0 license
- Complete freedom to fine-tune for specialized domains, deploy anywhere, and build commercial applications
- Apache 2.0 license allows unlimited modification and deployment globally
ENABLING GLOBAL DEPLOYMENT
- Enables deployment in emerging markets, resource-constrained sectors, and smaller organizations
- Enables government digital sovereignty with air-gapped deployments
- Provides frontier reasoning capabilities with complete deployment control
Safety-first Approach
- Comprehensive safety training and evaluation for each model
- Rigorously tested maliciously fine-tuned versions under OpenAI's Preparedness Framework
- Verified safety even after weight release to open community
OpenAI's Open Reasoning Models
Purpose-built reasoning models released under Apache 2.0 license for unlimited deployment flexibility.
Built for Democratic AI Infrastructure
OpenAI's open models represent breakthrough approaches to making frontier AI truly accessible and customizable.
Harmony Response Format
Models trained on OpenAI's harmony response format for consistent, structured outputs. Enables full chain-of-thought access and improved reasoning transparency.
Complete reasoning process visibility
Optimized Cloud Inference
Native MXFP4 quantization and efficient architecture designed for high-throughput cloud deployment on Together AI's infrastructure.
Superior cloud economics and performance
Apache 2.0 Licensing
Permissive licensing allows unlimited modification, commercial deployment, and redistribution. No vendor dependencies or usage restrictions.
Complete intellectual freedom
Enterprise-Grade Safety
Comprehensive safety training and malicious fine-tuning testing under OpenAI's Preparedness Framework ensures production-ready deployment.
Enterprise-grade safety assurance
Native Agentic Capabilities
Built-in function calling, web browsing, Python code execution, and Structured Outputs. No additional training required for tool use.
Ready-to-use agent functionality
Frontier Performance, Open Access
See how OpenAI's open models deliver competitive performance while maintaining complete deployment freedom.
OpenAI Open Models vs Competition
Benchmark results demonstrate competitive frontier performance with complete deployment freedom
Deploy on Together's Proven Infrastructure
Access OpenAI's open models through Together's optimized inference platform with enterprise-grade security and performance guarantees.
Serverless Endpoints
Pay-per-token pricing with automatic scaling. No infrastructure setup, no throttling.
Best for:
Getting started
Variable workloads
Prototyping
Cost optimization
gpt-oss-120B:
$0.15 input/1M tokens
$0.60 output/1M tokens
gpt-oss-20B:
$0.05 input/1M tokens
$0.20 output/1M tokensDedicated Endpoints
Guaranteed performance with custom fine-tuning capabilities. Complete control over model deployment.
Best for:
Production applications
Custom fine-tuning
Predictable latency
Enterprise SLA
Starting at:
$0.056 per minuteBatch API
Cost-effective processing for large workloads, distillation, and synthetic data generation.
Best for:
Large-scale processing
Model distillation
Data generation
Research workloads
50% discount on standard pricing
Enterprise-Grade Security
Your data and models remain fully under your control with industry-leading security standards.
99.9% Uptime SLA
Multi-region deployment with proven reliability at scale
SOC 2 Type II Compliant
Enterprise security standards with comprehensive auditing
Complete Model Ownership
Deploy anywhere with full rights to model weights and fine-tuned versions
North American Infrastructure
Data sovereignty with US-based secure server deployment
"We've been consistently impressed by Together's ability to quickly provide us with access to the latest open source models - often on day 1 of release - all while delivering exceptional performance and cost-efficiency."
— Nick Huber, AI Ecosystem Lead, Poe

"Together AI's dedicated infrastructure delivers the enterprise reliability that The Washington Post requires for deploying AI applications into production. When breakthrough models like OpenAI's open source models become available, Together AI's platform ensures we can deploy them on day 1 with the performance standards our organization needs."
— Sam Han, Chief AI Officer, The Washington Post
Frequently Asked Questions
How do these models differ from OpenAI's API models?
These are OpenAI's open-weight models, released under Apache 2.0 license. Unlike API models, you can download, modify, fine-tune, and deploy them anywhere without restrictions. You own the model weights completely.
What does Apache 2.0 licensing mean for my business?
Apache 2.0 is a permissive license allowing unlimited commercial use, modification, and redistribution. You can fine-tune for specialized domains, deploy on your infrastructure, and build commercial applications without licensing fees or usage restrictions.
Can I fine-tune these models for my specific use case?
Yes! Complete freedom to fine-tune for specialized domains. Use Together AI's Fine-Tuning API or download the weights to customize on your own infrastructure. All fine-tuned versions remain fully owned by you.
How do I migrate from OpenAI's API to these open models?
Migration is seamless with OpenAI-compatible API endpoints on Together AI. Simply change the model name in your existing code. You get the same reasoning capabilities with deployment freedom and predictable costs
What safety measures are included?
Each model underwent comprehensive safety training and evaluation. OpenAI rigorously tested maliciously fine-tuned versions under their Preparedness Framework to verify safety even after weight release to the community.
What are the current pricing rates?
gpt-oss-120B: $0.16 input / $0.60 output per million tokens and gpt-oss-20B: $0.05 input / $0.20 output per million tokens. Use our Batch API for 50% discounts on large workloads. No licensing fees ever.
Deploy OpenAI's Open-Weight Models
Access frontier reasoning models with complete deployment freedom. Join our technical webinar to understand these breakthrough models.