The Triton Inference Server provides an optimized cloud and edge inferencing solution.
Stars
10.5k
Forks
1.7k
Watchers
10.5k
Open Issues
869
Overall repository health assessment
No package.json found
This might not be a Node.js project
test: Enforce `max_inflight_requests` as a shared limit across ensemble requests (#8707)
e0a6a1eView on GitHubrefactor: Remove unused variables and arguments in model generation scripts (#8720)
e0f0ae8View on GitHubtest: Fix L0_backend_python model_readiness subtest (#8709)
79877d7View on GitHubfix: UAF in gRPC streaming "triton_grpc_error" mode (#8708)
c1b7cb7View on GitHubfeat: Support explicit model control mode and model management in OpenAI frontend (#8682)
bf2bd08View on GitHubTRI-837: quick fix, installaing defined version of PIP module (#8710)
f642e53View on GitHubtest: Add more test cases to validate model name while loading (#8701)
da9028aView on GitHubfeat: Refactor vLLM build to use upstream container image (#8697)
0177f59View on GitHubtest: Test safe GetElementCount and GetByteSize APIs at common and core repos (#8689)
06afae5View on GitHub