-
Notifications
You must be signed in to change notification settings - Fork 3.6k
Automatically set sync_batchnorm for training_type_plugin #6536
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
We may want a small test for this, just to make sure we catch this in case! Would you be able to write a small test using the BoringModel using a custom plugin? Just to ensure that sync_batchnorm is set correctly if not present |
|
Hello @amogkam! Thanks for updating this PR. There are currently no PEP 8 issues detected in this Pull Request. Cheers! 🍻 Comment last updated at 2021-03-19 19:49:25 UTC |
justusschock
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
|
Please, do not ignore the Before submitting/PR review headers at the top. |
Co-authored-by: Carlos Mocholí <[email protected]>
Co-authored-by: Roger Shieh <[email protected]>
|
Is this ready to get merged in? |
seems you have failing GPU test |
|
@Borda which job is the failing test on? I’m not able to find it. |
|
I skipped the test on GPU, but the CI is still failing. Any suggestions here? It doesn't look like it's related to this PR. |
|
Though all the required jobs are passing. Is this enough to merge this in? |
Codecov Report
@@ Coverage Diff @@
## master #6536 +/- ##
=======================================
- Coverage 94% 86% -8%
=======================================
Files 166 168 +2
Lines 11634 12205 +571
=======================================
- Hits 10947 10533 -414
- Misses 687 1672 +985 |
kaushikb11
left a comment
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks @amogkam for your contribution! The test was failing for Windows, added a skip for that.
Just wanted to know why is it needed to skip the test on GPU machines?
|
@kaushikb11 it was failing with this error So I thought that |
Co-authored-by: Carlos Mocholí <[email protected]> Co-authored-by: Roger Shieh <[email protected]> Co-authored-by: Kaushik Bokka <[email protected]> (cherry picked from commit 3b72bcc)
Co-authored-by: Carlos Mocholí <[email protected]> Co-authored-by: Roger Shieh <[email protected]> Co-authored-by: Kaushik Bokka <[email protected]> (cherry picked from commit 3b72bcc)
Co-authored-by: Carlos Mocholí <[email protected]> Co-authored-by: Roger Shieh <[email protected]> Co-authored-by: Kaushik Bokka <[email protected]> (cherry picked from commit 3b72bcc)
Co-authored-by: Carlos Mocholí <[email protected]> Co-authored-by: Roger Shieh <[email protected]> Co-authored-by: Kaushik Bokka <[email protected]>
Co-authored-by: Carlos Mocholí <[email protected]> Co-authored-by: Roger Shieh <[email protected]> Co-authored-by: Kaushik Bokka <[email protected]>
Co-authored-by: Carlos Mocholí <[email protected]> Co-authored-by: Roger Shieh <[email protected]> Co-authored-by: Kaushik Bokka <[email protected]>
What does this PR do?
This PR automatically sets the
sync_batchnormattribute for thetraining_type_pluginin theaccelerator_connector. This is useful for custom plugins whensync_batchnormis not known during plugin instantiation.Fixes #<issue_number>
Before submitting
PR review
Anyone in the community is free to review the PR once the tests have passed.
Before you start reviewing make sure you have read Review guidelines. In short, see the following bullet-list:
Did you have fun?
Make sure you had fun coding 🙃