spaCy trf model is slow due to torch #11915
kanayer
started this conversation in
Help: Best practices
Replies: 1 comment
-
Transformer models are intended to be run on GPU and will be very slow without one. We consider it important that the code works, from a correctness perspective, but it will be impractical for most use cases. If you do not have a GPU you should be fine with one of the CNN (non-trf) pipelines. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
I have a spacy trf model running on a docker container (no GPU) and in production, it seems to be very slow. It takes up to
17s
to process a single sentence. When I didCprofile
to check which step is taking the most time, the result was this:I tried to increase the number of torch threads by doing:
but it didn't help. Is there any way to speed it up?
Beta Was this translation helpful? Give feedback.
All reactions