-
Notifications
You must be signed in to change notification settings - Fork 30.7k
Updated Albert model Card #37753
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Updated Albert model Card #37753
Conversation
Hi 👋, thank you for opening this pull request! The pull request is converted to draft by default. The CI will be paused while the PR is in draft mode. When it is ready for review, please click the |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Nice, thanks for your contribution!
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks, just a few more minor details :)
@stevhliu, I have done suggested changes. please review and let me know if it is ok 😄 |
Hey, sorry for the delay! PR looks good to me, feel free to mark it as ready for review whenever you're ready 🙂 |
The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update. |
Hi, would you mind rebasing your branch? |
added the quotes in <hfoption id="Pipeline"> Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
updated checkpoints Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
changed !Tips description Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
updated text Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
updated transformer-cli implementation Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
changed text Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
removed repeated description Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
removed lines
updated pipeline code
updated auto model code, removed quantization as model size is not large, removed the attention visualizer part
updated notes Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
reduced a repeating point in notes
updated transformer-CLI Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
removed extra notes Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
4db5b2d
to
ef41998
Compare
@stevhliu, I guess I made the rebase. Since I am new to this can you please confirm if I have made it? |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks and no problem, this looks good!
- **Cross-layer parameter sharing:** Instead of learning separate parameters for each transformer layer, ALBERT shares parameters across layers, further reducing the number of learnable weights. | ||
|
||
ALBERT uses absolute position embeddings (like BERT) so padding is applied at right. Size of embeddings is 128 While BERT uses 768. ALBERT can processes maximum 512 token at a time. | ||
>>>>>>> 7ba1110083 (Update docs/source/en/model_doc/albert.md) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hey @stevhliu, I found some merge conflict artifacts that weren't removed during the PR merge. These lines are still showing in the docs. Should I open a new PR to clean these up?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I am making a PR for the docs for auto classes, so I could push the fix with that?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Oops, good catch! Can you just raise a separate PR to fix the merge conflict artifacts please?
What does this PR do?
Updates
ALBERT
model card as per #36979Before submitting
Pull Request section?
to it if that's the case.
documentation guidelines, and
here are tips on formatting docstrings.
Who can review?
@stevhliu Please check the PR and see if it's alright 😄