Customer stories

We’re creating a platform for progressive AI companies to build their products in the fastest, most performant infrastructure available.

Get in touch
Logo

Descript provides an AI-powered editing platform that enables users to easily edit videos, podcasts, and screen recordings. Their tools transcribe and allow manipulation of audio by editing the corresponding text.

descript.com
Logo

Gamma enables swift generation of AI-driven documents, presentations, and webpages, offering one-click templates and no-code editing to transform text into engaging content, enhancing clarity and audience engagement.

gamma.app
Logo

PicnicHealth develops software to simplify medical records access and enhance research. It offers a comprehensive health timeline service, integrating AI and machine learning for seamless health data management.

picnichealth.com
Logo

Cartesia's mission is to build the next generation of AI: ubiquitous, interactive intelligence that runs wherever you are. They're pioneering the model architectures that will make it possible. To date, Cartesia has built the theory behind SSMs and scaled it up to achieve state-of-the-art results in modalities as diverse as text, audio, video, images, and time-series data.

cartesia.ai
Logo

Rime develops a text-to-speech platform, focusing on natural-sounding voice generation with features like demographic tailoring and fast responses. It specializes in AI and Big Data to enhance conversational speech technologies.

rime.ai
Logo

Robust Intelligence offers a platform to eliminate AI failures, ensuring the robustness and reliability of AI systems. It focuses on detecting vulnerabilities and preventing adverse outcomes throughout the AI lifecycle.

robustintelligence.com
Logo

Wispr Flow is voice dictation built for professionals. Speak naturally, and Flow writes what you would’ve written. It’s like having a superpower that lets you write at the speed of thought. And it works everywhere you do: Slack, Email, Docs, Messages, ChatGPT, and more.

wispr.ai
Logo

Baseten gets the stuff we don't want to do out of the way. Now, our small, scrappy team can punch above our weight. It's everything from model serving, to auto-scaling, to iterating on products around those models, so we can deliver value to our customers and not worry about ML infrastructure.

Read Patreon case study

Baseten enabled us to achieve something remarkable—delivering real-time AI phone calls with sub-400 millisecond response times. That level of speed set us apart from every competitor.

Isaiah Granet, CEO and Co-Founder of Bland AI
Logo

A week ago we reached out with a hefty goal and within days your team helped us get set up and stable for a launch. It went smoothly, entirely thanks to you guys. 100% couldn’t have gone live without the software and hardware support you guys worked through the weekend to get us. The custom optimized Whisper on Baseten’s autoscaling L4 GPUs saved us.

Vincent Wilmet, Co-founder and CTO @ toby
Logo

Inference for custom-built LLMs could be a major headache. Thanks to Baseten, we’re getting cost-effective high-performance model serving without any extra burden on our internal engineering teams. Instead, we get to focus our expertise on creating the best possible domain-specific LLMs for our customers.

Waseem Alshikh, CTO and Co-Founder of Writer
Logo

Baseten provides an easy way for us to host our models, iterate on them, and experiment without worrying about any of the DevOps involved.

Faaez Ul Haq, Head of Data Science at Pipe
Logo

Baseten has allowed us to efficiently build an entirely new machine learning platform in just 4 months. By not needing to worry about managing our model infrastructure, Laurel has been able to drastically reduce our time to develop new predictive features and maintain more than double the number of models from our old platform.

Andrew Ward, VP of Machine Learning at Laurel
Logo

Explore Baseten today

We love partnering with companies developing innovative AI products by providing the most customizable model deployment with the lowest latency.