Prerequisites
Before starting, ensure the following:- NVIDIA Jetson AGX Orin Devkit is set up with JetPack 6.1 or later.
- CUDA Toolkit and cuDNN are installed.
- Verify that the Jetson AGX Orin is in high-performance mode:
Installing and running SGLang with Jetson Containers
Clone the jetson-containers github repository:Running Inference
Launch the server:--dtype half --context-length 8192) are due to the limited computational resources in Nvidia jetson kit. A detailed explanation can be found in Server Arguments.
After launching the engine, refer to Chat completions to test the usability.
Running quantization with TorchAO
TorchAO is suggested to NVIDIA Jetson Orin.--torchao-config int4wo-128 is also for memory efficiency.
Structured output with XGrammar
Please refer to SGLang doc structured output.Thanks to the support from Nurgaliyev Shakhizat, Dustin Franklin and Johnny Núñez Cano.
