-
Notifications
You must be signed in to change notification settings - Fork 1.5k
Issues: triton-inference-server/server
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Milestones
Assignee
Sort
Issues list
Thread control options in PyTorch backend
enhancement
New feature or request
#5067
opened Nov 10, 2022 by
kpedro88
allow constant input tensors in (ensemble) models
enhancement
New feature or request
#5037
opened Nov 4, 2022 by
charlesmelby
Add example model for FIL backend
enhancement
New feature or request
#4942
opened Oct 3, 2022 by
wphicks
Add support to "parameters" in Python tritonclient package
enhancement
New feature or request
#4919
opened Sep 23, 2022 by
RafalSkolasinski
Python backend cannot import Tensor
enhancement
New feature or request
#4870
opened Sep 12, 2022 by
Phelan164
Core rebuild is extremely long
enhancement
New feature or request
#4819
opened Aug 29, 2022 by
alxmamaev
Show ensemble stage where error happens
enhancement
New feature or request
#4786
opened Aug 18, 2022 by
alxmamaev
Container images for Jetson devices
enhancement
New feature or request
#4781
opened Aug 17, 2022 by
rgov
Tests for backend examples
enhancement
New feature or request
#4779
opened Aug 17, 2022 by
alxmamaev
Python Backend to support GPU instance
enhancement
New feature or request
#4772
opened Aug 16, 2022 by
MhdKAT
Python Backend complains "triton_python_backend_utils" has no attribute "InferenceRequest"
bug
Something isn't working
#4743
opened Aug 5, 2022 by
Michael-Jing
One click deployment to GKE no longer works as Istio deprecated
enhancement
New feature or request
investigating
The developement team is investigating this issue
#4668
opened Jul 19, 2022 by
samsaam-thg
Add source distribution to Python client package
enhancement
New feature or request
#4661
opened Jul 18, 2022 by
janjagusch
[python backend] Add class_count argument for inference requests with BLS scripting
enhancement
New feature or request
#4647
opened Jul 14, 2022 by
Rusteam
Dynamically loaded models don't work with ensemble
bug
Something isn't working
investigating
The developement team is investigating this issue
#4630
opened Jul 12, 2022 by
fran6co
ONNXRuntime TensorRT cache gets regenerated every time a model is uploaded even with correct settings
investigating
The developement team is investigating this issue
#4587
opened Jul 5, 2022 by
fran6co
Splitting a batch to max_batch_size if the batch size is larger than max_batch_size
enhancement
New feature or request
investigating
The developement team is investigating this issue
#4547
opened Jun 24, 2022 by
omidb
[Question] Customize HTTP response status code for malformed GPU card
bug
Something isn't working
investigating
The developement team is investigating this issue
#4540
opened Jun 22, 2022 by
pragyaarora
Support for the new CUDA virtual memory management functions for shared memory.
enhancement
New feature or request
#4538
opened Jun 22, 2022 by
Reznic
support decoupled mode in perf_analyzer
enhancement
New feature or request
#4530
opened Jun 20, 2022 by
Jackiexiao
Torchscript backend **MUCH** slower only with FP16 on 1650
investigating
The developement team is investigating this issue
#4451
opened May 31, 2022 by
dmenig
ProTip!
What’s not been updated in a month: updated:<2025-02-06.