Dynamo supports disaggregated serving of gpt-oss-120b with TensorRT-LLM. This guide demonstrates how to deploy gpt-oss-120b using disaggregated prefill/decode serving on a single B200 node with 8 GPUs, running 1 prefill worker on 4 GPUs and 1 decode worker on 4 GPUs.
This deployment uses disaggregated serving in TensorRT-LLM where:
- Prefill Worker: Processes input prompts efficiently using 4 GPUs with tensor parallelism
- Decode Worker: Generates output tokens using 4 GPUs, optimized for token generation throughput
- Frontend: Provides OpenAI-compatible API endpoint with round-robin routing
The disaggregated approach optimizes for both low-latency (maximizing tokens per second per user) and high-throughput (maximizing total tokens per GPU per second) use cases by separating the compute-intensive prefill phase from the memory-bound decode phase.
- 1x NVIDIA B200 node with 8 GPUs (this guide focuses on single-node B200 deployment)
- CUDA Toolkit 12.8 or later
- Docker with NVIDIA Container Toolkit installed
- Fast SSD storage for model weights (~240GB required)
- HuggingFace account and access token
- HuggingFace CLI
Ensure that the etcd and nats services are running with the following command:
docker compose -f deploy/docker-compose.yml upexport MODEL_PATH=<LOCAL_MODEL_DIRECTORY>
export HF_TOKEN=<INSERT_TOKEN_HERE>
pip install -U "huggingface_hub[cli]"
huggingface-cli download openai/gpt-oss-120b --exclude "original/*" --exclude "metal/*" --local-dir $MODEL_PATHSet the container image:
export DYNAMO_CONTAINER_IMAGE=nvcr.io/nvidia/ai-dynamo/tensorrtllm-runtime:my-tagLaunch the Dynamo TensorRT-LLM container with the necessary configurations:
docker run \
--gpus all \
-it \
--rm \
--network host \
--volume $MODEL_PATH:/model \
--volume $PWD:/workspace \
--shm-size=10G \
--ulimit memlock=-1 \
--ulimit stack=67108864 \
--ulimit nofile=65536:65536 \
--cap-add CAP_SYS_PTRACE \
--ipc host \
-e HF_TOKEN=$HF_TOKEN \
-e TRTLLM_ENABLE_PDL=1 \
-e TRT_LLM_DISABLE_LOAD_WEIGHTS_IN_PARALLEL=True \
$DYNAMO_CONTAINER_IMAGEThis command:
- Automatically removes the container when stopped (
--rm) - Allows container to interact with host's IPC resources for optimal performance (
--ipc=host) - Runs the container in interactive mode (
-it) - Sets up shared memory and stack limits for optimal performance
- Mounts your model directory into the container at
/model - Mounts the current Dynamo workspace into the container at
/workspace/dynamo - Enables PDL and disables parallel weight loading
- Sets HuggingFace token as environment variable in the container
The deployment uses configuration files and command-line arguments to control behavior:
Prefill Configuration (examples/backends/trtllm/engine_configs/gpt-oss-120b/prefill.yaml):
enable_attention_dp: false- Attention data parallelism disabled for prefillenable_chunked_prefill: true- Enables efficient chunked prefill processingmoe_config.backend: CUTLASS- Uses optimized CUTLASS kernels for MoE layerscache_transceiver_config.backend: ucx- Uses UCX for efficient KV cache transfercuda_graph_config.max_batch_size: 32- Maximum batch size for CUDA graphs
Decode Configuration (examples/backends/trtllm/engine_configs/gpt-oss-120b/decode.yaml):
enable_attention_dp: true- Attention data parallelism enabled for decodedisable_overlap_scheduler: false- Enables overlapping for decode efficiencymoe_config.backend: CUTLASS- Uses optimized CUTLASS kernels for MoE layerscache_transceiver_config.backend: ucx- Uses UCX for efficient KV cache transfercuda_graph_config.max_batch_size: 128- Maximum batch size for CUDA graphs
Both workers receive these key arguments:
--tensor-parallel-size 4- Uses 4 GPUs for tensor parallelism--expert-parallel-size 4- Expert parallelism across 4 GPUs--free-gpu-memory-fraction 0.9- Allocates 90% of GPU memory
Prefill-specific arguments:
--max-num-tokens 20000- Maximum tokens for prefill processing--max-batch-size 32- Maximum batch size for prefill
Decode-specific arguments:
--max-num-tokens 16384- Maximum tokens for decode processing--max-batch-size 128- Maximum batch size for decode
Note that GPT-OSS is a reasoning model with tool calling support. To ensure the response is being processed correctly, the worker should be launched with proper --dyn-reasoning-parser and --dyn-tool-call-parser.
You can use the provided launch script or run the components manually:
cd /workspace/examples/backends/trtllm
./launch/gpt_oss_disagg.sh- Start frontend:
# Start frontend with round-robin routing
python3 -m dynamo.frontend --router-mode round-robin --http-port 8000 &- Launch prefill worker:
CUDA_VISIBLE_DEVICES=0,1,2,3 python3 -m dynamo.trtllm \
--model-path /model \
--served-model-name openai/gpt-oss-120b \
--extra-engine-args examples/backends/trtllm/engine_configs/gpt-oss-120b/prefill.yaml \
--dyn-reasoning-parser gpt_oss \
--dyn-tool-call-parser harmony \
--disaggregation-mode prefill \
--max-num-tokens 20000 \
--max-batch-size 32 \
--free-gpu-memory-fraction 0.9 \
--tensor-parallel-size 4 \
--expert-parallel-size 4 &- Launch decode worker:
CUDA_VISIBLE_DEVICES=4,5,6,7 python3 -m dynamo.trtllm \
--model-path /model \
--served-model-name openai/gpt-oss-120b \
--extra-engine-args examples/backends/trtllm/engine_configs/gpt-oss-120b/decode.yaml \
--dyn-reasoning-parser gpt_oss \
--dyn-tool-call-parser harmony \
--disaggregation-mode decode \
--max-num-tokens 16384 \
--free-gpu-memory-fraction 0.9 \
--tensor-parallel-size 4 \
--expert-parallel-size 4Poll the /health endpoint to verify that both the prefill and decode worker endpoints have started:
curl http://localhost:8000/health
Make sure that both of the endpoints are available before sending an inference request:
{
"endpoints": [
"dyn://dynamo.tensorrt_llm.generate",
"dyn://dynamo.prefill.generate"
],
"status": "healthy"
}
If only one worker endpoint is listed, the other may still be starting up. Monitor the worker logs to track startup progress.
Send a test request to verify the deployment:
curl -X POST http://localhost:8000/v1/responses \
-H "Content-Type: application/json" \
-d '{
"model": "openai/gpt-oss-120b",
"input": "Explain the concept of disaggregated serving in LLM inference in 3 sentences.",
"max_output_tokens": 200,
"stream": false
}'The server exposes a standard OpenAI-compatible API endpoint that accepts JSON requests. You can adjust parameters like max_tokens, temperature, and others according to your needs.
Dynamo has supported reasoning and tool calling in OpenAI Chat Completion endpoint. A typical workflow for application built on top of Dynamo is that the application has a set of tools to aid the assistant provide accurate answer, and it is ususally multi-turn as it involves tool selection and generation based on the tool result.
In addition, the reasoning effort can be configured through chat_template_args. Increasing the reasoning effort makes the model more accurate but also slower. It supports three levels: low, medium, and high.
Below is an example of sending multi-round requests to complete a user query with reasoning and tool calling: Application setup (pseudocode)
# The tool defined by the application
def get_system_health():
for component in system.components:
if not component.health():
return False
return True
# The JSON representation of the declaration in ChatCompletion tool style
tool_choice = '{
"type": "function",
"function": {
"name": "get_system_health",
"description": "Returns the current health status of the LLM runtime—use before critical operations to verify the service is live.",
"parameters": {
"type": "object",
"properties": {}
}
}
}'
# On user query, perform below workflow.
def user_query(app_request):
# first round
# create chat completion with prompt and tool choice
request = ...
response = send(request)
if response["finish_reason"] == "tool_calls":
# second round
function, params = parse_tool_call(response)
function_result = function(params)
# create request with prompt, assistant response, and function result
request = ...
response = send(request)
return app_response(response)First request with tools
curl localhost:8000/v1/chat/completions -H "Content-Type: application/json" -d '
{
"model": "openai/gpt-oss-120b",
"messages": [
{
"role": "user",
"content": "Hey, quick check: is everything up and running?"
}
],
"chat_template_args": {
"reasoning_effort": "low"
},
"tools": [
{
"type": "function",
"function": {
"name": "get_system_health",
"description": "Returns the current health status of the LLM runtime—use before critical operations to verify the service is live.",
"parameters": {
"type": "object",
"properties": {}
}
}
}
],
"response_format": {
"type": "text"
},
"stream": false,
"max_tokens": 300
}'First response with tool choice
{
"id": "chatcmpl-d1c12219-6298-4c83-a6e3-4e7cef16e1a9",
"choices": [
{
"index": 0,
"message": {
"tool_calls": [
{
"id": "call-1",
"type": "function",
"function": {
"name": "get_system_health",
"arguments": "{}"
}
}
],
"role": "assistant",
"reasoning_content": "We need to check system health. Use function."
},
"finish_reason": "tool_calls"
}
],
"created": 1758758741,
"model": "openai/gpt-oss-120b",
"object": "chat.completion",
"usage": null
}Second request with tool calling result
curl localhost:8000/v1/chat/completions -H "Content-Type: application/json" -d '
{
"model": "openai/gpt-oss-120b",
"messages": [
{
"role": "user",
"content": "Hey, quick check: is everything up and running?"
},
{
"role": "assistant",
"tool_calls": [
{
"id": "call-1",
"type": "function",
"function": {
"name": "get_system_health",
"arguments": "{}"
}
}
]
},
{
"role": "tool",
"tool_call_id": "call-1",
"content": "{\"status\":\"ok\",\"uptime_seconds\":372045}"
}
],
"chat_template_args": {
"reasoning_effort": "low"
},
"tools": [
{
"type": "function",
"function": {
"name": "get_system_health",
"description": "Returns the current health status of the LLM runtime—use before critical operations to verify the service is live.",
"parameters": {
"type": "object",
"properties": {}
}
}
}
],
"response_format": {
"type": "text"
},
"stream": false,
"max_tokens": 300
}'Second response with final message
{
"id": "chatcmpl-9ebfe64a-68b9-4c1d-9742-644cf770ad0e",
"choices": [
{
"index": 0,
"message": {
"content": "All systems are green—everything’s up and running smoothly! 🚀 Let me know if you need anything else.",
"role": "assistant",
"reasoning_content": "The user asks: \"Hey, quick check: is everything up and running?\" We have just checked system health, it's ok. Provide friendly response confirming everything's up."
},
"finish_reason": "stop"
}
],
"created": 1758758853,
"model": "openai/gpt-oss-120b",
"object": "chat.completion",
"usage": null
}The Dynamo container includes AIPerf, NVIDIA's tool for benchmarking generative AI models. This tool helps measure throughput, latency, and other performance metrics for your deployment.
Run the following benchmark from inside the container (after completing the deployment steps above):
# Create a directory for benchmark results
mkdir -p /tmp/benchmark-results
# Run the benchmark - this command tests the deployment with high-concurrency synthetic workload
aiperf profile \
--model openai/gpt-oss-120b \
--tokenizer /model \
--endpoint-type chat \
--endpoint /v1/chat/completions \
--streaming \
--url localhost:8000 \
--synthetic-input-tokens-mean 32000 \
--synthetic-input-tokens-stddev 0 \
--output-tokens-mean 256 \
--output-tokens-stddev 0 \
--extra-inputs max_tokens:256 \
--extra-inputs min_tokens:256 \
--extra-inputs ignore_eos:true \
--extra-inputs "{\"nvext\":{\"ignore_eos\":true}}" \
--concurrency 256 \
--request-count 6144 \
--warmup-request-count 1000 \
--num-dataset-entries 8000 \
--random-seed 100 \
--artifact-dir /tmp/benchmark-results \
-v \
-H 'Authorization: Bearer NOT USED' \
-H 'Accept: text/event-stream'This command:
- Tests chat completions with streaming responses against the disaggregated deployment
- Simulates high load with 256 concurrent requests and 6144 total requests
- Uses long context inputs (32K tokens) to test prefill performance
- Generates consistent outputs (256 tokens) to measure decode throughput
- Includes warmup period (1000 requests) to stabilize performance metrics
- Saves detailed results to
/tmp/benchmark-resultsfor analysis
Key parameters you can adjust:
--concurrency: Number of simultaneous requests (impacts GPU utilization)--synthetic-input-tokens-mean: Average input length (tests prefill capacity)--output-tokens-mean: Average output length (tests decode throughput)--request-count: Total number of requests for the benchmark
If you prefer to run benchmarks from outside the container:
# Install AIPerf
pip install aiperf
# Then run the same benchmark command, adjusting the tokenizer path if neededThe disaggregated architecture separates prefill and decode phases:
flowchart TD
Client["Users/Clients<br/>(HTTP)"] --> Frontend["Frontend<br/>Round-Robin Router"]
Frontend --> Prefill["Prefill Worker<br/>(GPUs 0-3)"]
Frontend --> Decode["Decode Worker<br/>(GPUs 4-7)"]
Prefill -.->|KV Cache Transfer<br/>via UCX| Decode
- Disaggregated Serving: Separates compute-intensive prefill from memory-bound decode operations
- Optimized Resource Usage: Different parallelism strategies for prefill vs decode
- Scalable Architecture: Easy to adjust worker counts based on workload
- TensorRT-LLM Optimizations: Leverages TensorRT-LLM's efficient kernels and memory management
-
CUDA Out-of-Memory Errors
- Reduce
--max-num-tokensin the launch commands (currently 20000 for prefill, 16384 for decode) - Lower
--free-gpu-memory-fractionfrom 0.9 to 0.8 or 0.7 - Ensure model checkpoints are compatible with the expected format
- Reduce
-
Workers Not Connecting
- Ensure etcd and NATS services are running:
docker ps | grep -E "(etcd|nats)" - Check network connectivity between containers
- Verify CUDA_VISIBLE_DEVICES settings match your GPU configuration
- Check that no other processes are using the assigned GPUs
- Ensure etcd and NATS services are running:
-
Performance Issues
- Monitor GPU utilization with
nvidia-smiwhile the deployment is running - Check worker logs for bottlenecks or errors
- Ensure that batch sizes in manual commands match those in configuration files
- Adjust chunked prefill settings based on your workload
- For connection issues, ensure port 8000 is not being used by another application
- Monitor GPU utilization with
-
Container Startup Issues
- Verify that the NVIDIA Container Toolkit is properly installed
- Check Docker daemon is running with GPU support
- Ensure sufficient disk space for model weights and container images
- Production Deployment: For multi-node deployments, see the Multi-node Guide
- Advanced Configuration: Explore TensorRT-LLM engine building options for further optimization
- Monitoring: Set up Prometheus and Grafana for production monitoring
- Performance Benchmarking: Use AIPerf to measure and optimize your deployment performance