Skip to content

Conversation

souvikchand
Copy link
Contributor

What does this PR do?

Updates ALBERT model card as per #36979

Before submitting

Who can review?

@stevhliu Please check the PR and see if it's alright 😄

@github-actions github-actions bot marked this pull request as draft April 24, 2025 12:44
Copy link
Contributor

Hi 👋, thank you for opening this pull request! The pull request is converted to draft by default. The CI will be paused while the PR is in draft mode. When it is ready for review, please click the Ready for review button (at the bottom of the PR page). This will assign reviewers and trigger CI.

Copy link
Member

@stevhliu stevhliu left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Nice, thanks for your contribution!

Copy link
Member

@stevhliu stevhliu left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks, just a few more minor details :)

@souvikchand
Copy link
Contributor Author

@stevhliu, I have done suggested changes. please review and let me know if it is ok 😄

@stevhliu
Copy link
Member

Hey, sorry for the delay! PR looks good to me, feel free to mark it as ready for review whenever you're ready 🙂

@HuggingFaceDocBuilderDev

The docs for this PR live here. All of your documentation changes will be reflected on that endpoint. The docs are available until 30 days after the last update.

@souvikchand souvikchand marked this pull request as ready for review June 4, 2025 06:10
@stevhliu
Copy link
Member

Hi, would you mind rebasing your branch?

souvikchand and others added 15 commits June 14, 2025 02:29
added the quotes in <hfoption id="Pipeline">

Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
updated checkpoints

Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
changed !Tips description

Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
updated text

Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
updated transformer-cli implementation

Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
changed text

Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
removed repeated description

Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
removed lines
updated pipeline code
updated auto model code, removed quantization as model size is not large, removed the attention visualizer part
updated notes

Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
reduced a  repeating point in notes
updated transformer-CLI

Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
removed extra notes

Co-authored-by: Steven Liu <59462357+stevhliu@users.noreply.github.com>
@souvikchand
Copy link
Contributor Author

@stevhliu, I guess I made the rebase. Since I am new to this can you please confirm if I have made it?
Also sorry for the delay 😓

Copy link
Member

@stevhliu stevhliu left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks and no problem, this looks good!

@stevhliu stevhliu merged commit d5d007a into huggingface:main Jun 13, 2025
10 checks passed
- **Cross-layer parameter sharing:** Instead of learning separate parameters for each transformer layer, ALBERT shares parameters across layers, further reducing the number of learnable weights.

ALBERT uses absolute position embeddings (like BERT) so padding is applied at right. Size of embeddings is 128 While BERT uses 768. ALBERT can processes maximum 512 token at a time.
>>>>>>> 7ba1110083 (Update docs/source/en/model_doc/albert.md)
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Hey @stevhliu, I found some merge conflict artifacts that weren't removed during the PR merge. These lines are still showing in the docs. Should I open a new PR to clean these up?

image

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I am making a PR for the docs for auto classes, so I could push the fix with that?

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Oops, good catch! Can you just raise a separate PR to fix the merge conflict artifacts please?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants