Potential bugs with the NER model #1713
Labels
feat / ner
Feature: Named Entity Recognizer
lang / en
English language data and models
perf / accuracy
Performance: accuracy
The parsing result is different for different formatted input
We can see that by removing the space between
Kentucky Fried Chicken
and(KFC)
the parsing result is different, which, in my opinion, is not right, likeKentucky Fried Chicken(
andKFC)
. I guess it's probably the problem with the training data. But not sure.I can't really show the model result here, sorry about that. The problem is that the model is unstable, like by adding some irrelevant words to the begging or the end of the paragraph, the entity being recognized by the self-trained model would be totally changed. And also the entities recognized based on paragraph and sentences is different, which should be the case based on the Transition model described by Hannibal. I'm wondering have that issue happens before.
Thanks!
Info about spaCy
spaCy version: 2.0.3
Python version: 3.6
pre-trained Models: en_core_web_md
The text was updated successfully, but these errors were encountered: