-
Notifications
You must be signed in to change notification settings - Fork 372
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Slow traning #137
Comments
Bump. I'm facing the same issue where the training time is around 5 days on 3 3090s with batch size 192 and size of dataset ~400k images. The number of workers is set to 8. 2024-09-28 05:43:37 | INFO | damo.apis.detector_trainer:368 - epoch: 1/40, iter: 300/1483, mem: 7203Mb, iter_time: 7.779s, model_time: 7.283s, total_loss: 2.5, loss_cls: 0.2, loss_bbox: 1.6, loss_dfl: 0.7, lr: 4.878e-05, size: (640, 640), ETA: 5 days, 10:22:08 |
As a result, I looked at the caching method in YOLOX and tried to apply it in DAMO-YOLO, the ETA was reduced from 2 days 18 hours to 8 hours, but I'm not sure that this did not affect the performance of the model - I'm conducting an experiment for evaluation. It would be great if the authors added caching of the dataset |
just too slow>............why? |
Same question, training is much more slower than using similar models
My cfg:
Device |
Before Asking
I have read the README carefully. 我已经仔细阅读了README上的操作指引。
I want to train my custom dataset, and I have read the tutorials for finetune on your data carefully and organize my dataset correctly; 我想训练自定义数据集,我已经仔细阅读了训练自定义数据的教程,以及按照正确的目录结构存放数据集。
I have pulled the latest code of main branch to run again and the problem still existed. 我已经拉取了主分支上最新的代码,重新运行之后,问题仍不能解决。
Search before asking
Question
the damo-yolo very similar with yolox, but traning on my custome dataset damo-yolo traning time x3 by comparison with yolox, why?
Additional
No response
The text was updated successfully, but these errors were encountered: