Triton Inference Server Feature Requests
Triton Inference Server Feature Requests
Add a new idea
Filter by status
Already exists
0
Will not implement
0
Planned
0
Shipped
5
Log in / Sign up
Identify yourself with your email address
Email address
Recent
Trending
Popular
2
Vote
Add my vote +1
+2
+3
Unix sockets for pre-processing to inference server
Unix sockets for communication between containers, sending model inputs
Created 08 Feb 22:29
0
1
Vote
Add my vote +1
+2
+3
Pure Python client (no C Library)
TRTIS client for Python includes a C library which is ubuntu specific. Some customers have to build their own gRPC library to TRTIS in Python so that they didn't have a dependency on OS or Linux distribution
Created 27 Aug 17:25
0
1
Vote
Add my vote +1
+2
+3
UVM to avoid over subscribing GPU memory
Want to be able to load more models into TRTIS GPU memory. Concerned about the inability to anticipate GPU memory consumption/fragmentation. Having UVM would solve this
Created 27 Aug 17:23
0
1
Vote
Add my vote +1
+2
+3
C# Client API
In addition to the Python and C++ client libraries provided in TRTIS, also have a C# client to support customers on Windows.
Created 01 May 20:10
0
1
Vote
Add my vote +1
+2
+3
Telegraph Metrics
Telegraph is easier to push metrics to AWS for their SW stack
Created 27 Mar 17:52
0
1
Vote
Add my vote +1
+2
+3
Add Apache Thrift support
Apache Thrift: https://thrift.apache.org/
Created 28 Jan 23:42
0
0
Vote
Add my vote +1
+2
+3
Add support for dynamic PyTorch models (no torchscript needed)
Want to be able to run PyTorch models without having to convert to Torchscript. May be useful for certain customers who have a mix of production models and models not in production and in PyTorch. Would be useful to have both run in the same TRTIS...
Created 21 Nov 21:08
0
0
Vote
Add my vote +1
+2
+3
Add metric for time since last inference
Have TRTIS keep state of which models were deployed where and when they were called last by adding a metric for when a particular model was last called (time stamp of last inference).
Created 21 Nov 21:00
0
0
Vote
Add my vote +1
+2
+3
Model Store Integrations - Azure blob
Have TRTIS be able to pull models from a model repository hosted on Azure in addition to S3, Google Cloud Storage, and a local persistent volume.
Created 21 Nov 20:05
0
0
Vote
Add my vote +1
+2
+3
TRTIS build for embedded
Support for deploying and serving models using TRTIS on embedded GPUs (Jetson, Xavier, etc)
Created 21 Nov 19:58
0
1
2
Next ›
Last »