Model Performance Limited by Single Core CPU Speeds - GPU Not Fully Utilized #13583
Unanswered
Offek
asked this question in
Help: Other Questions
Replies: 1 comment
-
We solved it by creating threads around the spacy training (e.g with |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hello!
I've been playing around with spaCy and have created multiple SpanCat, TextCat, and NER models using curated-transformers that run on a GPU.
When either training or running the models, the performance is dictated by the CPU single-core performance.
There is always a single CPU core running at 100% and the GPU utilization hits 40% on average.
I've tested this on multiple machines with different CPUs and the same GPU (RTX 3090), and it proves to be true; the weaker the single-core performance of the CPU, the lower the GPU utilization and thus the slower the performance.
How could this be fixed? Multithreading is currently not a good solution for GPU since the VRAM is scarce. I've gotten a stronger CPU just so the training and execution of the models would be faster, but I believe there is some kind of bottleneck. Solving it could easily result in at least a 2X improvement.
I've tried modifying hyperparameters such as batch_size with no luck.
Please let me know if this is intended, or perhaps there is a known workaround.
Thank you!
Beta Was this translation helpful? Give feedback.
All reactions