Skip to content

question about attention_mask and pad_token_id #5

Open
@zxilin01

Description

@zxilin01
  1. when i run generate.sh of openreview, it outputs:

The attention mask and the pad token id were not set. As a consequence, you may observe unexpected behavior. Please pass your input's attention_mask to obtain reliable results.
Setting pad_token_id to eos_token_id:50256 for open-end generation.

Is this right? Or I need to modify any parameter?

  1. When running generate.sh of openreview, I found that the output of each epoch was like below. Why are there so many instances of 0.0? Is this normal? And the FID is also very large.

Image

Image

If I could receive your reply, I would greatly appreciate it!

Activity

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions