Customer stories
We’re creating a platform for progressive AI companies to build their products in the fastest, most performant infrastructure available.
Get in touchDescript provides an AI-powered editing platform that enables users to easily edit videos, podcasts, and screen recordings. Their tools transcribe and allow manipulation of audio by editing the corresponding text.
Gamma enables swift generation of AI-driven documents, presentations, and webpages, offering one-click templates and no-code editing to transform text into engaging content, enhancing clarity and audience engagement.
PicnicHealth develops software to simplify medical records access and enhance research. It offers a comprehensive health timeline service, integrating AI and machine learning for seamless health data management.
Cartesia's mission is to build the next generation of AI: ubiquitous, interactive intelligence that runs wherever you are. They're pioneering the model architectures that will make it possible. To date, Cartesia has built the theory behind SSMs and scaled it up to achieve state-of-the-art results in modalities as diverse as text, audio, video, images, and time-series data.
Robust Intelligence offers a platform to eliminate AI failures, ensuring the robustness and reliability of AI systems. It focuses on detecting vulnerabilities and preventing adverse outcomes throughout the AI lifecycle.
Wispr Flow is voice dictation built for professionals. Speak naturally, and Flow writes what you would’ve written. It’s like having a superpower that lets you write at the speed of thought. And it works everywhere you do: Slack, Email, Docs, Messages, ChatGPT, and more.
Read Patreon case studyBaseten gets the stuff we don't want to do out of the way. Now, our small, scrappy team can punch above our weight. It's everything from model serving, to auto-scaling, to iterating on products around those models, so we can deliver value to our customers and not worry about ML infrastructure.
Rime’s state-of-the-art p99 latency and 100% uptime over 2024 is driven by our shared laser focus on fundamentals, and we’re excited to push the frontier even further with Baseten.
Baseten enabled us to achieve something remarkable—delivering real-time AI phone calls with sub-400 millisecond response times. That level of speed set us apart from every competitor.
A week ago we reached out with a hefty goal and within days your team helped us get set up and stable for a launch. It went smoothly, entirely thanks to you guys. 100% couldn’t have gone live without the software and hardware support you guys worked through the weekend to get us. The custom optimized Whisper on Baseten’s autoscaling L4 GPUs saved us.
Inference for custom-built LLMs could be a major headache. Thanks to Baseten, we’re getting cost-effective high-performance model serving without any extra burden on our internal engineering teams. Instead, we get to focus our expertise on creating the best possible domain-specific LLMs for our customers.
Baseten provides an easy way for us to host our models, iterate on them, and experiment without worrying about any of the DevOps involved.
Baseten has allowed us to efficiently build an entirely new machine learning platform in just 4 months. By not needing to worry about managing our model infrastructure, Laurel has been able to drastically reduce our time to develop new predictive features and maintain more than double the number of models from our old platform.
Explore Baseten today
We love partnering with companies developing innovative AI products by providing the most customizable model deployment with the lowest latency.