-
Notifications
You must be signed in to change notification settings - Fork 2.9k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
add CTRL Model #921
add CTRL Model #921
Conversation
前向输出还需要再定位清楚不一致的op! |
这个PR还有效吗?@yingyibiao |
有效 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Things in utils.py is too familiar to HuggingFace. Please change them to PaddleNLP-style. And add more doc. You can see other model for details.
CTRL权重文件 |
numerical representations of tokens that build the input sequence. | ||
Its data type should be `int64` and it has a shape of [batch_size, sequence_length]. | ||
cache (Tuple[Tuple[Tensor]], optional): | ||
Contains pre-computed hidden-states (key and values in the attention blocks) as computed by the model. Can be used to speed up sequential decoding. The `input_ids` which have their past given to this model should not be passed as input ids as they have already been computed. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
这里的文档单行不要太长,最好每85个字符左右换行一次
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
done
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
需要在example/text_generation下面新增一个使用案例 |
飞桨论文复现挑战赛(第四期)CTRL: A Conditional Transformer Language Model for Controllable Generation 论文复现提交。