DP Rank Routing (Attention Data Parallelism)
For general TensorRT-LLM features and configuration, see the Reference Guide.
TensorRT-LLM supports attention data parallelism (attention DP) for models like DeepSeek. When enabled, multiple attention DP ranks run within a single worker, each with its own KV cache. Dynamo can route requests to specific DP ranks based on KV cache state.
Dynamo vs TRT-LLM Internal Routing
- Dynamo DP Rank Routing: The router selects the optimal DP rank based on KV cache overlap and instructs TRT-LLM to use that rank with strict routing (
attention_dp_relax=False). Use this with--router-mode kvfor cache-aware routing. - TRT-LLM Internal Routing: TRT-LLM’s scheduler assigns DP ranks internally. Use this with
--router-mode round-robinorrandomwhen KV-aware routing isn’t needed.
Enabling DP Rank Routing
The --enable-attention-dp flag sets attention_dp_size = tensor_parallel_size and configures Dynamo to publish KV events per DP rank. The router automatically creates routing targets for each (worker_id, dp_rank) combination.
Attention DP requires TRT-LLM’s PyTorch backend. AutoDeploy does not support attention DP.