Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

undefined symbol: hipGetDevicePropertiesR0600 #31

Open
alain40 opened this issue Jan 2, 2024 · 3 comments
Open

undefined symbol: hipGetDevicePropertiesR0600 #31

alain40 opened this issue Jan 2, 2024 · 3 comments

Comments

@alain40
Copy link

alain40 commented Jan 2, 2024

Getting this error after successful compilation on 7840U.

rocm 6.0 and PyTorch installed and tested.

Here is the error when running exllama which runs fine without flash attention:
Traceback (most recent call last):
File "/home/alain/exllamav2/examples/chat.py", line 5, in
from exllamav2 import(
File "/home/alain/.local/lib/python3.10/site-packages/exllamav2/init.py", line 3, in
from exllamav2.model import ExLlamaV2
File "/home/alain/.local/lib/python3.10/site-packages/exllamav2/model.py", line 21, in
from exllamav2.attn import ExLlamaV2Attention
File "/home/alain/.local/lib/python3.10/site-packages/exllamav2/attn.py", line 19, in
import flash_attn
File "/home/alain/.local/lib/python3.10/site-packages/flash_attn/init.py", line 3, in
from flash_attn.flash_attn_interface import flash_attn_func
File "/home/alain/.local/lib/python3.10/site-packages/flash_attn/flash_attn_interface.py", line 4, in
import flash_attn_2_cuda as flash_attn_cuda
ImportError: /home/alain/.local/lib/python3.10/site-packages/flash_attn_2_cuda.cpython-310-x86_64-linux-gnu.so: undefined symbol: hipGetDevicePropertiesR0600

Otherwise HIP runtime looks fine:
hipconfig.txt

rocm also looks OK (using gfx 11.0.0 override):
rocminfo.txt

@sabreshao sabreshao added the navi hardware label Jan 16, 2024
@TNT3530
Copy link

TNT3530 commented Jan 17, 2024

Also having this issue when attempting to build from source + benchmark on gfx908 (Instinct MI100) and ROCm 6.0.
Running any version I built (including torch 2.0.1, 2.1.1, and 2.2.0) gives the above error.

Running/building the docker images works fine, so I assume its an issue with ROCm 6.0

@nayn99
Copy link

nayn99 commented Feb 28, 2024

Same issue on 4650G.

The compilation is successful, but when loading the library in python, I get ~/coco/llm/lib/python3.11/site-packages/flash_attn_2_cuda.cpython-311-x86_64-linux-gnu.so: undefined symbol: hipGetDevicePropertiesR0600

Otherwise I am able to install xformers, and pytorch works fine. This seems the last ingredient to getting vllm running.

@tcgu-amd
Copy link

Hi @nayn99 @alain40, sorry for the lack the responses. Is there a supported AMD graphics card based on the CDNA2/CDNA3 architectures on your devices? Flash attention is currently only supported on those devices see this issue.

@TNT3530, if running/building in the docker image works fine for you I suspect there is something wrong with how ROCm is configured on your host system. Would you be able to to try uninstalling the current version of docker and re-install the newest version of ROCm following this guide? Thanks!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

6 participants