Together AI Announces $305M Series B to Scale AI Acceleration Cloud for Open Source and Enterprise AI

Today marks an exciting milestone in Together AI's journey. We're thrilled to announce our $305 million Series B funding round, led by General Catalyst and co-led by Prosperity7.
The round saw participation from a distinguished group of global institutional and strategic investors including Salesforce Ventures, DAMAC Capital, NVIDIA, Kleiner Perkins, March Capital, Emergence Capital, Lux Capital, SE Ventures, Greycroft, Coatue, Definition, Cadenza Ventures, Long Journey Ventures, Brave Capital, Scott Banister, SK Telecom, and technology pioneer John Chambers.
This investment will accelerate our leadership as the preferred AI Cloud for building modern AI applications with open source models, and for training custom models with our upcoming large-scale deployment of NVIDIA Blackwell GPUs.
Our AI Acceleration Cloud has already transformed how over 450,000 AI developers, AI-native companies, and global enterprises like Salesforce, Zoom, SK Telecom, Hedra, Cognition, Zomato, Krea, Cartesia, and The Washington Post build modern AI applications.

Making Open Source AI Accessible to All
AI is transforming every industry, creating unprecedented efficiencies and enabling entirely new classes of products. At Together AI, we believe the future of AI is open source, and we have built a cloud company for this AI-first world by combining state-of-the-art open source models and high-performance infrastructure with frontier research in AI efficiency and scalability.
Open source models like DeepSeek-R1 and Meta's Llama have emerged as formidable alternatives to proprietary solutions, marking a decisive shift in the AI landscape. Together AI has established itself as the definitive platform powering this transformation, delivering the fastest DeepSeek-R1 and Llama inference for NVIDIA GPUs at production scale through our secure, highly optimized infrastructure and research innovations.
Our AI Acceleration Cloud uniquely spans the entire AI lifecycle, delivering enterprise-grade inference solutions, training and fine-tuning for frontier foundational models, agentic workflows with built-in code interpretation, and synthetic data generation. It enables organizations to build complete AI applications with the performance, security, accuracy, and model ownership that enterprises demand.
Supporting over 200 open source models across all modalities — chat, image, audio, vision, code, and embeddings — the platform is powered by Together AI's proprietary Inference engine and built on research innovations including FlashAttention-3 kernels and advanced quantization techniques. It delivers 2-3x faster inference than today's hyperscaler solutions.
Expanding Our Infrastructure
To support our rapidly growing ecosystem, we're dramatically expanding our infrastructure. We've secured 200 MW of power capacity and are deploying optimized clusters of NVIDIA Blackwell GPUs across multiple North American data centers. Our recent partnership with Hypertec to co-build a cluster of 36,000 NVIDIA GB200 NVL72 GPUs further strengthens our position as the leading AI Cloud provider. We also announced immediate access to Together GPU Clusters accelerated by NVIDIA HGX B200 GPUs and the Together Kernel Collection, delivering 90% faster training performance than previous generation infrastructure.
Innovation and Research at Our Core
Research drives everything we do at Together AI. Our research lab continues to pioneer breakthrough methods at the intersection of AI and systems optimization, with innovations like our Mixture of Agents, Medusa, Sequoia, Hyena, and Mamba that optimize AI accuracy, performance, and efficiencies. Together Kernel Collection, developed under the leadership of our Chief Scientist Tri Dao, creator of FlashAttention has enabled 24% faster training operations while significantly reducing costs for our customers.
Recent Milestones and Future Vision
In 2024, we've achieved significant milestones that demonstrate our momentum. We deployed DeepSeek models in North American data centers with full opt-out privacy controls, launched the Together Enterprise Platform, and announced AWS Marketplace availability. Our partnership with Cartesia has enabled ultra-low latency voice AI through Sonic model integration, while our acquisition of CodeSandbox brings built-in code interpretation capabilities to our platform. We've also strengthened our leadership team with the addition of go-to-market veteran Kai Mak as CRO, and research pioneer James Zou.
This investment will accelerate our mission to make open source AI accessible to developers and enterprises globally. We're committed to advancing the frontier of AI through open collaboration, innovation, and transparency, while ensuring powerful AI systems remain accessible and cost-effective.
To learn more about opportunities at Together AI, visit our careers page. For media inquiries, please reach out to press@together.ai
- Lower
Cost20% - faster
training4x - network
compression117x
Q: Should I use the RedPajama-V2 Dataset out of the box?
RedPajama-V2 is conceptualized as a pool of data that serves as a foundation for creating high quality datasets. The dataset is thus not intended to be used out of the box and, depending on the application, data should be filtered out using the quality signals that accompany the data. With this dataset, we take the view that the optimal filtering of data is dependent on the intended use. Our goal is to provide all the signals and tooling that enables this.
article