Changelog
See our latest feature releases, product improvements and bug fixes
Async API endpoint code example
Interested in running async inference? We've made it easier to get started by adding a code example in the new API endpoint modal. Just click API in the top right of the page and toggle to Async API...
Aug 14, 2024Improved replica selector
To handle massive bursts of traffic, we've helped user scale to hundreds of replicas on model deployments.
Jul 9, 2024Pin frequently-used models for easy access
By popular demand: you can now pin models and chains to the top of your workspace!
Jul 2, 2024Export billing usage data
You can now export your model and billing usage data for in-depth analysis in your preferred tool.
Jun 11, 2024Run model inference asynchronously on Baseten
We’re thrilled to announce that you can now run async inference on Baseten models!
Apr 22, 2024Refreshed model overview page with deployment statuses
We’ve revised the model overview page to give you more information about model deployments at a glance.
Apr 3, 2024Improved log filtering
You can now filter logs through the main text input. Just start typing the filter you’re looking for, like level, and autocomplete options will appear.
Mar 26, 2024Permit inference on unhealthy models
A model enters an “unhealthy” state when the deployment is active but there are runtime errors such as downtime on an external dependency.
Mar 21, 2024Improve performance and reduce cost with fractional H100 GPUs
Baseten now offers model inference on NVIDIA H100mig GPUs, available for all customers starting at $0.08250/minute.
Mar 20, 2024Manage models with the Baseten REST API
We’re excited to share that we’ve created a REST API for managing Baseten models! Unlock powerful use cases outside of the (albeit amazing) Baseten UI - interact with your models programmatically,...