-
Notifications
You must be signed in to change notification settings - Fork 31.2k
TensorFlow 2.0 - Testing with a few Bert architectures #1104
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Codecov Report
@@ Coverage Diff @@
## master #1104 +/- ##
=========================================
- Coverage 79.61% 79.12% -0.5%
=========================================
Files 42 56 +14
Lines 6898 7654 +756
=========================================
+ Hits 5492 6056 +564
- Misses 1406 1598 +192
Continue to review full report at Codecov.
|
qgallouedec
pushed a commit
to qgallouedec/transformers
that referenced
this pull request
May 17, 2025
* Fix token decode in fill-mask pipeline * Add support for ModernBERT * Add modernbert unit tests * Cleanup bert unit tests * Add unit test for `sequence_length > local_attention_window`
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This PR tests how easy it would be to incorporate TF 2.0 models in the current library:
TFBertPreTrainedModel,TFBertModel,TFBertForPretraining,TFBertForMaskedLM,TFBertForNextSentencePrediction,bert-base-uncasedmodel is up on our AWS S3 bucket for the moment),The library is (very) slightly reorganized to allow for this, mostly by spinning configuration classes out of (PyTorch) modeling classes to allow reusability between PyTorch and TF 2.0 models.
With TF 2.0 Keras imperative interface and Eager, the workflow and models are suprisingly similar:
If you want to play with this, you can install from the
tfbranch like this:pip install tensorflow==2.0.0-rc0tfbranch:pip install https://github.com/huggingface/pytorch-transformers/archive/tf.zip