Out-of-the-box infrastructure for fine-tuning and inferencing open-source LLMs, enabling rapid scaling and deployment with per-token pricing.