Introduction
MindSpore is a high-performance AI framework optimized for Ascend NPUs. This doc guides users to run MindSpore models in SGLang.Requirements
MindSpore currently only supports Ascend NPU devices. Users need to first install Ascend CANN software packages. The CANN software packages can be downloaded from the Ascend Official Website. The recommended version is 8.3.RC2.Supported Models
Currently, the following models are supported:- Qwen3: Dense and MoE models
- DeepSeek V3/R1
- More models coming soon…
Installation
Note: Currently, MindSpore models are provided by an independent packagesgl-mindspore. Support for MindSpore is built upon current SGLang support for Ascend NPU platform. Please first install SGLang for Ascend NPU and then installsgl-mindspore:
Run Model
Current SGLang-MindSpore supports Qwen3 and DeepSeek V3/R1 models. This doc uses Qwen3-8B as an example.Offline infer
Use the following script for offline infer:Start server
Launch a server with MindSpore backend:Troubleshooting
Debug Mode
Enable sglang debug logging by log-level argument.Explicitly select devices
Use the following environment variable to explicitly select the devices to use.Some communication environment issues
In case of some environment with special communication environment, users need set some environment variables.Some dependencies of protobuf
In case of some environment with special protobuf version, users need set some environment variables to avoid binary version mismatch.Support
For MindSpore-specific issues:- Refer to the MindSpore documentation
