spaCy trf model is slow due to torch #11915
kanayer
started this conversation in
Help: Best practices
Replies: 1 comment
-
Transformer models are intended to be run on GPU and will be very slow without one. We consider it important that the code works, from a correctness perspective, but it will be impractical for most use cases. If you do not have a GPU you should be fine with one of the CNN (non-trf) pipelines. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
I have a spacy trf model running on a docker container (no GPU) and in production, it seems to be very slow. It takes up to
17s
to process a single sentence. When I didCprofile
to check which step is taking the most time, the result was this:I tried to increase the number of torch threads by doing:
but it didn't help. Is there any way to speed it up?
Beta Was this translation helpful? Give feedback.
All reactions