-
Notifications
You must be signed in to change notification settings - Fork 11
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Training time #3
Comments
same problem, it seems slow using the code for training BYOL, how can we optim it for fast training? |
@guanfuchen
Second, self-supervised methods usually use a large batch size, e.g., 4096, for better performance. And there are a lot of efficiency challenges in large-scale distributed training.
Besides, there are some points that could be optimized. The most important one is data loading. The dataloader assumes reading raw images and then doing data augmentations, which is rather slow. You can:
|
Thxs, great answer. DALI is not very flexiable, yes, syncbn is important. I am considering the MOCO, it seems more quick than BYOL, and it does not use SyncBN. @yaox12 |
|
|
Thxs @yaox12 , you code is really excelent!!! I will try but the cost is high. I modify the weight decay to 0 using linear evaluation, then the acc is almost same as the origin cv2 implementaion. |
@yaox12 Hi, I notice you actually did not use mixed precision in the BYOL training since |
@youweiliang See #7 |
Thx. I have just tried using |
Hi, thanks for the implementation. Could you provide the (approximate) training time to produce the result in your table in the Readme?
The text was updated successfully, but these errors were encountered: