Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

SHARK Backend integration #4485

Open
powderluv opened this issue Jun 7, 2022 · 1 comment
Open

SHARK Backend integration #4485

powderluv opened this issue Jun 7, 2022 · 1 comment
Labels
enhancement New feature or request

Comments

@powderluv
Copy link

SHARK is a high performance codegen compiler and runtime built on MLIR, IREE and custom RL based tuning infrastructure. Here are some results of using SHARK for the same model across PyTorch, Onnx, TF/XLA and SHARK.

We have an Triton Inference Server integration of SHARK that runs on CPU and CUDA devices here: https://github.com/nod-ai/SHARK/tree/main/inference and we would like to upstream it as an available Triton backend that anyone can build, test and deploy.

@tanmayv25 tanmayv25 added the enhancement New feature or request label Jun 7, 2022
@tanmayv25
Copy link
Contributor

@msalehiNV @dzier for visibility.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Development

No branches or pull requests

2 participants