changelog / post

Training is now GA!

Go back

Since launching the beta of Baseten Training in May, we’ve introduced a ton of improvements, including;

  • A more robust ML Cookbook, great starting points for:

    • Training a coding model with GRPO

    • Long-context training using multi-node with Qwen3 30B A3B

    • A variety of examples with Qwen3, gpt-oss, Gemma3, and Llama

  • Resume from checkpoint: Launch jobs that pick up right where you left off

  • A ton of other improvements, including:

    • Broader checkpoint recognition across FSDP, VeRL, and Megatron checkpointing formats

    • More availability for InfiniBand-backed multi-node training runs

    • Improved management and handling of the training cache

    • Per-GPU metric visibility and improved logs

    • Quality of life improvements around Training Cache

    • And much more!

After months of positive feedback from early users and thousands of training runs completed, Baseten Training is now immediately available for anyone on Baseten. Get started here.