-
Notifications
You must be signed in to change notification settings - Fork 235
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
question about DNN learning curve #20
Comments
Thank you for your detailed question. My advice is, In VAD field, I experimentally found that dropout seems not to be necessary as we use already noise corrupted data, which cause the regularization effect. As in your figure 1, the DNN could not stably contain (train) your noisy dataset. Maybe the properties of noise types in your training dataset are too diverge each other. My advice is, try to reduce the batch size (e.g. 256) and do not shuffle your batches . It can make the training more stably. After trying this advice plz let me know. THX. |
Thank you for your sharing. My question is
|
Thank you for patience
|
Hi, Kim, I have tried a few days,
|
Thank you for sharing. Note that there is some additional tricks apart from the training method described in my paper (because of paper limitation). Training noisy data is hard problem because of underfitting (your problem) that I also have experienced. So, One of my way is trying to learn our model with easy noise (high SNR) to hard noise (low SNR). |
Thank you very much! I will try |
Dear @machinelearningisgood |
Hi Kim,
I hope you have a good weekend! When you have a minute, would you please take a look at my questions about your code and let me know what you think of it? Here is the question:
I have run you DNN code with my own timit data and changed the iteration time to get more data as shown in the attachment VAD_DNN.py。However, the valid accuracy is higher than the training accuracy, which is very strange. And the accuracy and cost of training data are fluctuating dramatically. For example, the training accuracy is 47% while the valid is 96% at the last point. Could you explain it?
I also trained my own DNN without using feature extraction and got the opposite result that the accuracy and cost of valid data are fluctuating dramatically. It seems that there is an overfitting problem. The dropout rate is already 0.7. Could you give me some advice
I really appreciate your feedback
The text was updated successfully, but these errors were encountered: