You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When running an NVTabular workflow with Categorify operations in Triton Inference Server, the performance is significantly slow when dealing with high cardinality data.
The Categorify operation should perform efficiently, with category data being cached between requests, resulting in performance similar to that observed in a Jupyter notebook environment.
Actual Behavior
The Categorify operation is slow, with each request taking as long as the first request, suggesting that category data is not being effectively cached between requests.
Results
Below are the result based on benchmarking script - encode.sh
Cardinality
Ensemble Triton
TransformWorkflow Jupyter
50
30 ms
38 ms
5k
30 ms
43 ms
5M
1270 ms
88.8 ms
50M
15833 ms
550 ms
The text was updated successfully, but these errors were encountered:
Description
When running an NVTabular workflow with Categorify operations in Triton Inference Server, the performance is significantly slow when dealing with high cardinality data.
Environment
Steps to Reproduce
tritonserver --model-repository=./ensemble/
Expected Behavior
The Categorify operation should perform efficiently, with category data being cached between requests, resulting in performance similar to that observed in a Jupyter notebook environment.
Actual Behavior
The Categorify operation is slow, with each request taking as long as the first request, suggesting that category data is not being effectively cached between requests.
Results
Below are the result based on benchmarking script - encode.sh
The text was updated successfully, but these errors were encountered: