Distributed inference for a spaCy custom NER model #11771
dave-espinosa
started this conversation in
Help: Best practices
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
-
Hello everyone. I trained a NER model, and I can get a reasonably fast inference by a applying all the hints explained in the speed FAQ. However, one thing is getting inference from fair amount of texts (<100K), and a whole different world, is getting inference from hundreds of millions of texts (which is what I want to improve). In the past, I have tried the following (chronological) approaches:
nlp.pipe
and its argumentn_process
is the fastest way to improve inference speed, but computational resources are expensive (i.e., number of available cores in your physical or virtual machine).The obvious next step is trying to improve the 4th approach to something more stable (and maybe recurring to some product that I still don't know of? 🤔), but then I think it would be nice to listen to the community. BTW, I am working with Google Cloud Platform related products, but any idea is worth trying at this point. I think I can summarize my queries in the following ones:
Any ideas?
Beta Was this translation helpful? Give feedback.
All reactions