Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Unable to initialize backend 'gpu' #8

Open
HGX-001 opened this issue Aug 20, 2021 · 7 comments
Open

Unable to initialize backend 'gpu' #8

HGX-001 opened this issue Aug 20, 2021 · 7 comments

Comments

@HGX-001
Copy link

HGX-001 commented Aug 20, 2021

Hello,I have follow you readme to install the anaconda environment,and it can run with cpu ,but it can not run with gpu ,and in my devices have a nvidia RTX TITAN GPU with 24G momeory ,when I whatever use bash python3 run_alphafold.py --fasta_paths=T1050.fasta --max_template_date=2020-05-14# or simply exp/run_local.sh T1050.fasta it warning with I0820 16:00:20.270564 140257858221888 xla_bridge.py:212] Unable to initialize backend 'tpu_driver': Not found: Unable to find driver in registry given worker: local://
I0820 16:00:20.281177 140257858221888 xla_bridge.py:212] Unable to initialize backend 'gpu': Not found: Could not find registered platform with name: "cuda". Available platform names are: Interpreter Host
I0820 16:00:20.281787 140257858221888 xla_bridge.py:212] Unable to initialize backend 'tpu': Invalid argument: TpuPlatform is not available.
W0820 16:00:20.282057 140257858221888 xla_bridge.py:215] No GPU/TPU found, falling back to CPU. (Set TF_CPP_MIN_LOG_LEVEL=0 and rerun for more info.)
I feel it like run only with cpu ,and it only use 315MiB GPU Momory when I try to input nvidia-smi to look how much GPU memory use.

@kuixu
Copy link
Owner

kuixu commented Aug 21, 2021

Does the NVIDIA CUDA driver install successfully?

@tavolivos
Copy link

I have the same issue, even I can run it in CPU.

I installed nvidia-455 and cuda 11.1

@HGX-001
Copy link
Author

HGX-001 commented Aug 27, 2021 via email

@JuergenUniVie
Copy link

same problem...
is it possible to set a enviroment to fix the platform "cuda" issue?

I1108 16:51:57.959832 139635007661120 xla_bridge.py:231] Unable to initialize backend 'tpu_driver': NOT_FOUND: Unable to find driver in registry given worker:
I1108 16:51:57.972085 139635007661120 xla_bridge.py:231] Unable to initialize backend 'gpu': NOT_FOUND: Could not find registered platform with name: "cuda". Available platform names are: Interpreter Host
I1108 16:51:57.972550 139635007661120 xla_bridge.py:231] Unable to initialize backend 'tpu': INVALID_ARGUMENT: TpuPlatform is not available.
W1108 16:51:57.972653 139635007661120 xla_bridge.py:236] No GPU/TPU found, falling back to CPU. (Set TF_CPP_MIN_LOG_LEVEL=0 and rerun for more info.)

+-----------------------------------------------------------------------------+
| NVIDIA-SMI 495.29.05 Driver Version: 495.29.05 CUDA Version: 11.5 |
|-------------------------------+----------------------+----------------------+
| GPU Name Persistence-M| Bus-Id Disp.A | Volatile Uncorr. ECC |
| Fan Temp Perf Pwr:Usage/Cap| Memory-Usage | GPU-Util Compute M. |
| | | MIG M. |
|===============================+======================+======================|
| 0 NVIDIA GeForce ... Off | 00000000:18:00.0 Off | N/A |
| 23% 27C P8 10W / 250W | 15MiB / 11178MiB | 0% Default |
| | | N/A |
+-------------------------------+----------------------+----------------------+
| 1 NVIDIA GeForce ... Off | 00000000:3B:00.0 Off | N/A |
| 23% 28C P8 8W / 250W | 2MiB / 11178MiB | 0% Default |
| | | N/A |
+-------------------------------+----------------------+----------------------+
| 2 NVIDIA GeForce ... Off | 00000000:86:00.0 Off | N/A |
| 23% 30C P8 9W / 250W | 2MiB / 11178MiB | 0% Default |
| | | N/A |
+-------------------------------+----------------------+----------------------+
| 3 NVIDIA GeForce ... Off | 00000000:AF:00.0 Off | N/A |
| 23% 30C P8 8W / 250W | 2MiB / 11178MiB | 0% Default |
| | | N/A |
+-------------------------------+----------------------+----------------------+

+-----------------------------------------------------------------------------+
| Processes: |
| GPU GI CI PID Type Process name GPU Memory |
| ID ID Usage |
|=============================================================================|
| 0 N/A N/A 3626 G /usr/libexec/Xorg 9MiB |
| 0 N/A N/A 4701 G /usr/bin/gnome-shell 3MiB |
+-----------------------------------------------------------------------------+

@mattiasmar
Copy link

Hi guys,
Any insights with regards to the cuda issue?
Ping @JuergenUniVie

@deejy
Copy link

deejy commented Apr 8, 2022

Trying to install alphafold on various platform (including AWS) since 2 months : I always fall on that bug which is frequent but not addressed.
Very sad
JPierre

@coliva92
Copy link

coliva92 commented Jun 2, 2022

I'm facing the same issue. Before running the run_docker.py script I ran the docker run --rm --gpus all nvidia/cuda:11.0-base nvidia-smi command as mentioned in the README file and it correctly displayed my GPU (with no processes running). But I still got the same error messages others are reporting here.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

7 participants