Together AI and Vultr Partner to Scale Generative AI at the Edge

We recently announced that we are embarking on an exciting new partnership with Together AI, a platform as a service (PaaS) provider that unlocks the power of open source large language models (LLMS) to enable generative AI at scale. Together AI will leverage our global array of cloud GPUs as the compute backbone to enable worldwide inference at the edge.

Together AI had a significant announcement of its own last week – they closed on a $102.5 million Series A financing round, led by Kleiner Perkins and with major contributors from NVIDIA and Emergence Capital. We congratulate our colleagues at Together AI for the strong endorsement this investment provides.

Our budding partnership is an incredible validation of Vultr’s unique position within the cloud infrastructure marketplace. As the only global, independent alternative to the hyperscalers and as an ardent supporter of open architecture and composability, Vultr is the natural choice for any enterprise looking to combine the power of generative AI and the flexibility and control of open-source LLMs. Together AI has recognized the incredible potential of partnering with Vultr, so we’re off to the races, enabling inference everywhere.

Together AI will tap into Vultr’s broad array of state-of-the-art NVIDIA GPUs, including our vast inventory of the brand new NVIDIA GH200 superchips, available at all 32 of our cloud data center locations across six continents.

Why This Matters Now

The momentum that companies like Together AI and Vultr are experiencing reflects the sonic boom unleashed one year ago when OpenAI introduced the world to ChatGPT. The events of the past 12 months, including the debut of several competing generative AI applications and the rapid innovation in GPU technology by NVIDIA and others, have demonstrated to the world the art of the possible.

Now, enterprises are rushing en masse to find ways to accelerate AI transformation – all intending to deliver better outcomes for their customers, employees, and shareholders. But without insistence on openness, observability, governance, and reproducibility of inferences and the models that generate them, too many companies will experience missteps when they try to go beyond proof of concept to inference at scale.

Together AI provides the platform for building formidable generative AI applications on top of open source LLMs. Vultr delivers the composable GPU infrastructure that delivers the compute power. In the coming months and years, the partnership will open doors for innovators worldwide to leverage generative AI in ways that will fundamentally reinvent businesses individually and collectively.

Moving Forward Into the Future

As we look ahead to 2024 and beyond, inference at the edge will emerge as the predominant paradigm for content delivery. Smaller, specialized LLMs will be put into service to address end users' specific needs in different locations. Together AI has the platform for enabling AI engineers and developers to train, fine-tune, and run large scale generative AI applications. Vultr has the cloud GPU infrastructure and the global deployment of this infrastructure to power services like Together AI, while meeting the highest demands for performance and reliability.

It’s an exciting time, and Vultr is exceptionally proud to stand shoulder-to-shoulder with Together AI to extend the power and the promise of generative AI to all corners of the world.