Meet the performance-obsessed teams shaping the future
Baseten is the infrastructure choice for teams shipping high-stakes, high-performance AI products.
How World Labs is building large world models, pushing the boundaries of 3D
How Gamma makes building presentations criminally fun
How OpenEvidence trains accurate, domain-specific models with Baseten Training
How Writer helps businesses transform with AI
How Rime.ai achieved state-of-the-art p99 latencies on Baseten
Superhuman achieves 80% faster embedding model inference with Baseten
How Sully.ai returned 30M+ clinical minutes to healthcare using open-source models.
How Sully.ai addressed their latency, cost and quality challenges by transitioning its Inference Stack to open-source models running on Baseten.
90%
Inference cost savings
65%
Lower median latency
OpenEvidence delivers instant, accurate medical information with the Baseten Inference Stack
Read moreWispr Flow creates effortless voice dictation with Llama on Baseten
Read moreLatent delivers pharmaceutical search with 99.999% uptime on Baseten
Read moreBuilding AI Agents, Open Code, and Open Source Coding with Dax Raad
Watch nowPraktika delivers ultra-low-latency transcription for global language education with Baseten
Read moreFrom datasets to deployed models: How Oxen helps companies train faster
Read moreScaled Cognition offers ultra-fast AI agents you can trust
Read moreZed Industries serves 2x faster code completions with the Baseten Inference Stack
Read morePatreon saves nearly $600k/year in ML resources with Baseten
Read moreChosen by the world's most ambitious builders
Case study
Case study
Case study
Case study
Case study
Case study
Case study
Case study
Case study
Case study
Case study
Case study

































