Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Robert] support 2D attention_mask from tokenizer #1676

Merged
merged 8 commits into from
Mar 4, 2022
Merged

[Robert] support 2D attention_mask from tokenizer #1676

merged 8 commits into from
Mar 4, 2022

Conversation

turkeymz
Copy link
Contributor

PR types

Bug fixes

PR changes

Models

Description

Roberta has the same bug

PR
#1634

support 2D attention_mask from tokenizer
fix 1631 and 1579

Log
ValueError: (InvalidArgument) Broadcast dimension mismatch. Operands could not be broadcast together with the shape of X = [256, 12, 60, 60] and the shape of Y = [256, 60]. Received [60] in X is not equal to [256] in Y at i:2.
[Hint: Expected x_dims_array[i] == y_dims_array[i] || x_dims_array[i] <= 1 || y_dims_array[i] <= 1 == true, but received x_dims_array[i] == y_dims_array[i] || x_dims_array[i] <= 1 || y_dims_array[i] <= 1:0 != true:1.] (at /paddle/paddle/fluid/operators/elementwise/elementwise_op_function.h:240)
[operator < elementwise_add > error]

@CLAassistant
Copy link

CLAassistant commented Feb 15, 2022

CLA assistant check
All committers have signed the CLA.

Copy link
Contributor

@yingyibiao yingyibiao left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@yingyibiao yingyibiao merged commit 7b8dd49 into PaddlePaddle:develop Mar 4, 2022
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

ernie-gram运行添加attention_mask报错
3 participants