Replies: 2 comments 1 reply
-
Hi, latency depends heavily on the types of recognizers and NLP models you apply, and there's a latency-accuracy tradeoff. Then, going with heavier spacy models ( |
Beta Was this translation helpful? Give feedback.
-
Thanks for your suggestion. The switch to the small models gave marginal improvements but it's not a panacea. Are the built in entity recognizers relying on the ner step in spaCy's pipeline? That still seems the bottleneck for our use case. Is it possible to disable that step? |
Beta Was this translation helpful? Give feedback.
-
Our use case for Presidio is whether PII information is present in the text that is analyzed, without needing to know which entity that is. We use the built-in entities and custom matchers. We're hitting performance issues in terms of latency as we rely on live feedback. The allocated cpu and memory resources are not being maxed out by Presidio.
We suspect that the underlying spaCy pipeline is too heavy for us in the sense that we're not relying on the ner step in our output. Does this analysis make sense? If so, would it be possible to make that pipeline configurable?
Beta Was this translation helpful? Give feedback.
All reactions