Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Optimum supports AMDGPU ? #1538

Closed
taikai-zz opened this issue Nov 15, 2023 · 1 comment
Closed

Optimum supports AMDGPU ? #1538

taikai-zz opened this issue Nov 15, 2023 · 1 comment

Comments

@taikai-zz
Copy link

Feature request

Onnxruntime supports AMD-ROCM ,
how to compile on optimum

Motivation

Our company is currently testing amdgpu and has learned that optim can accelerate inference on CUDA. We are not sure if it will support ROCM in the future?

Your contribution

none

@fxmarty
Copy link
Contributor

fxmarty commented Jan 9, 2024

Hi @taikai-zz, the ORTModel classes support ROCm GPUs for inference with ONNX Runtime, following this PR #1546. See as reference https://huggingface.co/docs/optimum/main/en/onnxruntime/usage_guides/amdgpu. However, from internal benchmarks, we did not see very large performance improvements compared to simply using PyTorch.

The status of the support of ROCm GPUs in Hugging Face libraries is summarized in https://huggingface.co/docs/optimum/main/en/amd/amdgpu/overview. Overall, the upstream support in PyTorch makes lots of things easy.

@fxmarty fxmarty closed this as completed Jan 9, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants