Papers
arxiv:2605.13779

MinT: Managed Infrastructure for Training and Serving Millions of LLMs

Published on May 13
· Submitted by
Andrew Chen
on May 14
#1 Paper of the day
Authors:
,
,
,
,
,
,
,
,
,
,
,
,
,
,
,
,
,
,

Abstract

MinT is a managed infrastructure system that enables efficient low-rank adaptation training and serving by keeping base models resident and moving lightweight adapter revisions, scaling across multiple dimensions including large model architectures, reduced storage requirements, and distributed policy management.

AI-generated summary

We present MindLab Toolkit (MinT), a managed infrastructure system for Low-Rank Adaptation (LoRA) post-training and online serving. MinT targets a setting where many trained policies are produced over a small number of expensive base-model deployments. Instead of materializing each policy as a merged full checkpoint, MinT keeps the base model resident and moves exported LoRA adapter revisions through rollout, update, export, evaluation, serving, and rollback, hiding distributed training, serving, scheduling, and data movement behind a service interface. MinT scales this path along three axes. Scale Up extends LoRA RL to frontier-scale dense and MoE architectures, including MLA and DSA attention paths, with training and serving validated beyond 1T total parameters. Scale Down moves only the exported LoRA adapter, which can be under 1% of base-model size in rank-1 settings; adapter-only handoff reduces the measured step by 18.3x on a 4B dense model and 2.85x on a 30B MoE, while concurrent multi-policy GRPO shortens wall time by 1.77x and 1.45x without raising peak memory. Scale Out separates durable policy addressability from CPU/GPU working sets: a tensor-parallel deployment supports 10^6-scale addressable catalogs (measured single-engine sweeps through 100K) and thousand-adapter active waves at cluster scale, with cold loading treated as scheduled service work and packed MoE LoRA tensors improving live engine loading by 8.5-8.7x. MinT thus manages million-scale LoRA policy catalogs while training and serving selected adapter revisions over shared 1T-class base models.

Community

Paper author Paper submitter

MinT: Managed Infrastructure for Training and Serving Millions of LLMs

Thank you for your patience, support, and testing during MinT’s internal testing period.

Since the start of internal testing, we have been continuously improving MinT’s stability, reliability, and overall user experience. We are happy to share that MinT has now entered a long-term stable phase and is ready for broader external use.

To help everyone get started, we have allocated 5M tokens of usage to each registered user. You can currently use the full Qwen3 model family on MinT. For more details about supported models, please visit:

https://mint-doc.macaron.im/en/community/get-started/supported-models

In addition, MinT also supports model families such as GLM-5.1, MiniMax, and OpenPI. If you need access to these models or have specific model requirements, please contact us at: sales@mindlab.ltd

You can learn more about MinT here:

https://macaron.im/mindlab/mint

To register and start using MinT, please visit:

https://mint-console.macaron.xin/login

You can also explore our doc and training examples here:

https://mint-doc.macaron.im

We hope MinT can help accelerate your model training workflows and make post-training easier, more stable, and more accessible.

Thank you again for being part of the MinT testing journey.

Community version @ Github➡️https://github.com/verl-project/verl-mint/

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2605.13779 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2605.13779 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2605.13779 in a Space README.md to link it from this page.

Collections including this paper 2