-
Notifications
You must be signed in to change notification settings - Fork 3k
modify transformers.rst #496
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
参数量XX M与HF直接对齐应该问题不大
docs/model_zoo/transformers.rst
Outdated
+--------------------+-------------------------------------+--------------+-----------------------------------------+ | ||
|ERNIE_ |``ernie-1.0`` | Chinese | | | ||
|ERNIE_ |``ernie-1.0`` | Chinese | 12-layer, 768-hidden, | | ||
| | | | 12-heads, _M parameters. | |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
参数量与BERT一致
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
done
docs/model_zoo/transformers.rst
Outdated
| |``gpt-cpm-small-cn-distill`` | Chinese | 12-layer, 768-hidden, | | ||
| | | | 12-heads, 109M parameters. | | ||
| | | | The model distilled from | | ||
| | | | the GPT model ``gpt-cpm-small-cn`` | |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
https://github.com/TsinghuaAI/CPM-Distill
我看这个问的那个,应该是从large(gpt-cpm-large-cn)蒸馏过来的。
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
done
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
表格设置为居中显示体验会更好
@@ -9,145 +9,301 @@ PaddleNLP为用户提供了常用的 ``BERT``、``ERNIE``、``ALBERT``、``RoBER | |||
Transformer预训练模型汇总 | |||
------------------------------------ | |||
|
|||
下表汇总了介绍了目前PaddleNLP支持的各类预训练模型以及对应预训练权重。我们目前提供了 **62** 种预训练的参数权重供用户使用, | |||
下表汇总了介绍了目前PaddleNLP支持的各类预训练模型以及对应预训练权重。我们目前提供了 **67** 种预训练的参数权重供用户使用, | |||
其中包含了 **32** 种中文语言模型的预训练权重。 | |||
|
|||
+--------------------+-------------------------------------+--------------+-----------------------------------------+ | |||
| Model | Pretrained Weight | Language | Details of the model | | |||
+====================+=====================================+==============+=========================================+ |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
表格可能设置为居中模式会更加美观
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
rst语法上不太好设置居中显示,可以通过css端设置
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
像这类文档你觉得用rst组织好还是markdown组织好?这两种方式各有什么利弊?
@@ -158,35 +314,35 @@ Transformer预训练模型适用任务汇总 | |||
+--------------------+-------------------------+----------------------+--------------------+-----------------+ | |||
| Model | Sequence Classification | Token Classification | Question Answering | Text Generation | | |||
+====================+=========================+======================+====================+=================+ | |||
|ALBERT_ |✅ |✅ |✅ |❌ | |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
这些参数使用居中会更美观。
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
rst语法上不太好设置居中显示,可以通过css端设置
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
是否markdown更为方便?
rst语法上不太好设置居中显示,可以通过css端设置 |
HF表格只列了一小部分,很多我们的模型它没有列出来 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
讨论下markdown和rst组织的利弊
@@ -9,145 +9,301 @@ PaddleNLP为用户提供了常用的 ``BERT``、``ERNIE``、``ALBERT``、``RoBER | |||
Transformer预训练模型汇总 | |||
------------------------------------ | |||
|
|||
下表汇总了介绍了目前PaddleNLP支持的各类预训练模型以及对应预训练权重。我们目前提供了 **62** 种预训练的参数权重供用户使用, | |||
下表汇总了介绍了目前PaddleNLP支持的各类预训练模型以及对应预训练权重。我们目前提供了 **67** 种预训练的参数权重供用户使用, | |||
其中包含了 **32** 种中文语言模型的预训练权重。 | |||
|
|||
+--------------------+-------------------------------------+--------------+-----------------------------------------+ | |||
| Model | Pretrained Weight | Language | Details of the model | | |||
+====================+=====================================+==============+=========================================+ |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
像这类文档你觉得用rst组织好还是markdown组织好?这两种方式各有什么利弊?
docs/model_zoo/transformers.rst
Outdated
| +-------------------------------------+--------------+-----------------------------------------+ | ||
| |``skep_roberta_large_en`` | English | | | ||
|SKEP_ |``skep_ernie_1.0_large_ch`` | Chinese | 24-layer, 1024-hidden, | | ||
| | | | 16-heads, _M parameters. | |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
这个模型的size,直接与ernie1.0 large一致即可
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
done
| | | | Trained using the Erine model | | ||
| | | | ``ernie_1.0`` | | ||
| +-------------------------------------+--------------+-----------------------------------------+ | ||
| |``skep_ernie_2.0_large_en`` | English | 24-layer, 1024-hidden, | |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
这个模型的size,与ernie 2.0 large的size一样大即可。也与BERT-large一致
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
done
@@ -158,35 +314,35 @@ Transformer预训练模型适用任务汇总 | |||
+--------------------+-------------------------+----------------------+--------------------+-----------------+ | |||
| Model | Sequence Classification | Token Classification | Question Answering | Text Generation | | |||
+====================+=========================+======================+====================+=================+ | |||
|ALBERT_ |✅ |✅ |✅ |❌ | |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
是否markdown更为方便?
Update MAG240M report
PR types
Function optimization
PR changes
Docs
Description
modify transformers.rst