MinT: Managed Infrastructure for Training and Serving Millions of LLMs
Abstract
MinT is a managed infrastructure system that enables efficient low-rank adaptation training and serving by keeping base models resident and moving lightweight adapter revisions, scaling across multiple dimensions including large model architectures, reduced storage requirements, and distributed policy management.
We present MindLab Toolkit (MinT), a managed infrastructure system for Low-Rank Adaptation (LoRA) post-training and online serving. MinT targets a setting where many trained policies are produced over a small number of expensive base-model deployments. Instead of materializing each policy as a merged full checkpoint, MinT keeps the base model resident and moves exported LoRA adapter revisions through rollout, update, export, evaluation, serving, and rollback, hiding distributed training, serving, scheduling, and data movement behind a service interface. MinT scales this path along three axes. Scale Up extends LoRA RL to frontier-scale dense and MoE architectures, including MLA and DSA attention paths, with training and serving validated beyond 1T total parameters. Scale Down moves only the exported LoRA adapter, which can be under 1% of base-model size in rank-1 settings; adapter-only handoff reduces the measured step by 18.3x on a 4B dense model and 2.85x on a 30B MoE, while concurrent multi-policy GRPO shortens wall time by 1.77x and 1.45x without raising peak memory. Scale Out separates durable policy addressability from CPU/GPU working sets: a tensor-parallel deployment supports 10^6-scale addressable catalogs (measured single-engine sweeps through 100K) and thousand-adapter active waves at cluster scale, with cold loading treated as scheduled service work and packed MoE LoRA tensors improving live engine loading by 8.5-8.7x. MinT thus manages million-scale LoRA policy catalogs while training and serving selected adapter revisions over shared 1T-class base models.
Community
MinT: Managed Infrastructure for Training and Serving Millions of LLMs
Thank you for your patience, support, and testing during MinT’s internal testing period.
Since the start of internal testing, we have been continuously improving MinT’s stability, reliability, and overall user experience. We are happy to share that MinT has now entered a long-term stable phase and is ready for broader external use.
To help everyone get started, we have allocated 5M tokens of usage to each registered user. You can currently use the full Qwen3 model family on MinT. For more details about supported models, please visit:
https://mint-doc.macaron.im/en/community/get-started/supported-models
In addition, MinT also supports model families such as GLM-5.1, MiniMax, and OpenPI. If you need access to these models or have specific model requirements, please contact us at: sales@mindlab.ltd
You can learn more about MinT here:
https://macaron.im/mindlab/mint
To register and start using MinT, please visit:
https://mint-console.macaron.xin/login
You can also explore our doc and training examples here:
We hope MinT can help accelerate your model training workflows and make post-training easier, more stable, and more accessible.
Thank you again for being part of the MinT testing journey.
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper