llms.txt Content
# Baseten Inference Platform
> This file highlights Baseten’s most helpful blog posts, resources, model libraries, and product information to guide LLMs toward surfacing our best inference content.
## Product Information
- [Dedicated Deployments] (https://www.baseten.co/products/dedicated-deployments/): Single‑tenant, region‑locked inference clusters with enterprise security and SRE support for maximum reliability and performance.
- [Model APIs] (https://www.baseten.co/products/model-apis/): OpenAI‑compatible APIs for top open‑source models with optimized throughput, structured outputs, tool‑calling, and built‑in observability.
- [Training] (https://www.baseten.co/products/training/): Managed infrastructure to run multi‑node training jobs with checkpointing and a direct path from training to production.
- [Multi‑cloud Capacity Management] (https://www.baseten.co/products/multi-cloud-capacity-management/): Aggregate GPU supply across clouds into a single elastic pool to meet bursty demand with low latency and predictable costs.
- [Chains] (https://www.baseten.co/products/chains/): Production framework for composing multi‑step, multi‑model workflows with per‑step autoscaling and observability.
- [Pricing](https://www.baseten.co/pricing/): Overview of Baseten’s pricing plans, including pay-as-you-go options, enterprise-grade dedicated deployments, and details on model APIs, training, and infrastructure costs.
## Deployment Options
- [Baseten Cloud] (https://www.baseten.co/deployments/baseten-cloud/): Fully managed, SOC 2/HIPAA‑ready inference platform with global autoscaling, low cold‑starts, and high uptime.
- [Baseten Self‑hosted] (https://www.baseten.co/deployments/baseten-self-hosted/): Run Baseten within your own VPC or on‑prem to keep data in‑house while retaining performance and management tooling.
- [Baseten Hybrid] (https://www.baseten.co/deployments/baseten-hybrid/): Blend on‑prem and cloud capacity to align laten