Skip to content

modify transformers.rst #496

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 12 commits into from
Jun 5, 2021
Merged

modify transformers.rst #496

merged 12 commits into from
Jun 5, 2021

Conversation

yingyibiao
Copy link
Contributor

PR types

Function optimization

PR changes

Docs

Description

modify transformers.rst

@yingyibiao yingyibiao requested a review from ZeyuChen June 5, 2021 04:26
Copy link
Member

@ZeyuChen ZeyuChen left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

参数量XX M与HF直接对齐应该问题不大

+--------------------+-------------------------------------+--------------+-----------------------------------------+
|ERNIE_ |``ernie-1.0`` | Chinese | |
|ERNIE_ |``ernie-1.0`` | Chinese | 12-layer, 768-hidden, |
| | | | 12-heads, _M parameters. |
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

参数量与BERT一致

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

done

| |``gpt-cpm-small-cn-distill`` | Chinese | 12-layer, 768-hidden, |
| | | | 12-heads, 109M parameters. |
| | | | The model distilled from |
| | | | the GPT model ``gpt-cpm-small-cn`` |
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

https://github.com/TsinghuaAI/CPM-Distill
我看这个问的那个,应该是从large(gpt-cpm-large-cn)蒸馏过来的。

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

done

@ZeyuChen ZeyuChen added the documentation Improvements or additions to documentation label Jun 5, 2021
Copy link
Member

@ZeyuChen ZeyuChen left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

表格设置为居中显示体验会更好

@@ -9,145 +9,301 @@ PaddleNLP为用户提供了常用的 ``BERT``、``ERNIE``、``ALBERT``、``RoBER
Transformer预训练模型汇总
------------------------------------

下表汇总了介绍了目前PaddleNLP支持的各类预训练模型以及对应预训练权重。我们目前提供了 **62** 种预训练的参数权重供用户使用,
下表汇总了介绍了目前PaddleNLP支持的各类预训练模型以及对应预训练权重。我们目前提供了 **67** 种预训练的参数权重供用户使用,
其中包含了 **32** 种中文语言模型的预训练权重。

+--------------------+-------------------------------------+--------------+-----------------------------------------+
| Model | Pretrained Weight | Language | Details of the model |
+====================+=====================================+==============+=========================================+
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

表格可能设置为居中模式会更加美观

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

rst语法上不太好设置居中显示,可以通过css端设置

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

像这类文档你觉得用rst组织好还是markdown组织好?这两种方式各有什么利弊?

@@ -158,35 +314,35 @@ Transformer预训练模型适用任务汇总
+--------------------+-------------------------+----------------------+--------------------+-----------------+
| Model | Sequence Classification | Token Classification | Question Answering | Text Generation |
+====================+=========================+======================+====================+=================+
|ALBERT_ |✅ |✅ |✅ |❌ |
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

这些参数使用居中会更美观。

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

rst语法上不太好设置居中显示,可以通过css端设置

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

是否markdown更为方便?

@yingyibiao
Copy link
Contributor Author

表格设置为居中显示体验会更好

rst语法上不太好设置居中显示,可以通过css端设置

@yingyibiao
Copy link
Contributor Author

yingyibiao commented Jun 5, 2021

参数量XX M与HF直接对齐应该问题不大

HF表格只列了一小部分,很多我们的模型它没有列出来
可以先把这版合入,后续再补充缺失的参数量。

@yingyibiao yingyibiao requested a review from ZeyuChen June 5, 2021 06:16
Copy link
Member

@ZeyuChen ZeyuChen left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

讨论下markdown和rst组织的利弊

@@ -9,145 +9,301 @@ PaddleNLP为用户提供了常用的 ``BERT``、``ERNIE``、``ALBERT``、``RoBER
Transformer预训练模型汇总
------------------------------------

下表汇总了介绍了目前PaddleNLP支持的各类预训练模型以及对应预训练权重。我们目前提供了 **62** 种预训练的参数权重供用户使用,
下表汇总了介绍了目前PaddleNLP支持的各类预训练模型以及对应预训练权重。我们目前提供了 **67** 种预训练的参数权重供用户使用,
其中包含了 **32** 种中文语言模型的预训练权重。

+--------------------+-------------------------------------+--------------+-----------------------------------------+
| Model | Pretrained Weight | Language | Details of the model |
+====================+=====================================+==============+=========================================+
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

像这类文档你觉得用rst组织好还是markdown组织好?这两种方式各有什么利弊?

| +-------------------------------------+--------------+-----------------------------------------+
| |``skep_roberta_large_en`` | English | |
|SKEP_ |``skep_ernie_1.0_large_ch`` | Chinese | 24-layer, 1024-hidden, |
| | | | 16-heads, _M parameters. |
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

这个模型的size,直接与ernie1.0 large一致即可

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

done

| | | | Trained using the Erine model |
| | | | ``ernie_1.0`` |
| +-------------------------------------+--------------+-----------------------------------------+
| |``skep_ernie_2.0_large_en`` | English | 24-layer, 1024-hidden, |
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

这个模型的size,与ernie 2.0 large的size一样大即可。也与BERT-large一致

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

done

@@ -158,35 +314,35 @@ Transformer预训练模型适用任务汇总
+--------------------+-------------------------+----------------------+--------------------+-----------------+
| Model | Sequence Classification | Token Classification | Question Answering | Text Generation |
+====================+=========================+======================+====================+=================+
|ALBERT_ |✅ |✅ |✅ |❌ |
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

是否markdown更为方便?

@yingyibiao yingyibiao requested a review from ZeyuChen June 5, 2021 09:58
@ZeyuChen ZeyuChen merged commit 3b02983 into PaddlePaddle:develop Jun 5, 2021
@yingyibiao yingyibiao deleted the docs branch June 7, 2021 03:18
DesmonDay pushed a commit to DesmonDay/PaddleNLP that referenced this pull request Sep 23, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
documentation Improvements or additions to documentation
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants