-
Notifications
You must be signed in to change notification settings - Fork 32
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Results in NeuralQA inconsistent with same model running on HF #60
Comments
Could this be related to #39 |
Yes further testing with multiple models does confirm that the results given by NeuralQA are way off the ones returned by HF face model. Hope this can be resolved soon as it's critical to us. Thank you |
Hi @victordibia, just checking in to see if there's any update on this? Seems like a pretty critical issue. Thanks |
@victordibia Is this project still maintained? We have not heard from you for a while. Hope everything is ok. |
@victordibia It's a shame that this is no longer maintained. What are you plans vis-a-vis this project? |
I've tested a model that I've deployed on NeuralQa vs one deployed on HF and noticed that the same inputs are yielding different outputs even though it's using the exact same model. This can of course be attributed to a few things but I can't seem to identify the culprit.
Here's the context:
Question:
Are your handsets locked or unlocked?
Corpus:
['No, all our handsets are unlocked.','Since your SIM isn’t working in your handset while other SIM cards are, it might be an issue with your handset provider; or the mobile phone could be locked, meaning it only accepts SIM cards from a particular service provider. Please contact the handset dealer for more assistance.']
The following returns 'unlocked' which is the correct response:
See Demo on HuggingFace
I've configured the exact same model in NeuralQA (with relsnip disabled) and the result is 'locked' even though I'm feeding exactly the same inputs.
Here my log:
As you can see the 2nd answer gets a higher probability but that doesn't really make sense as it's exactly the same model.
The main difference is that the NeuralQA model is feeding the corpus content independently while in the HF example, we're feeding the entire corpus.
Any ideas on why this is happening?
The text was updated successfully, but these errors were encountered: