Skip to content

Using (Absolute) Positional Embeddings with Hyena Operators #29

Open
@saberiato

Description

@saberiato

Hi @DanFu09
Hope you're well,

I was reading the source code and the config files, and I realized that use_positional_encodings is True (link). So, the M2-BERT model is using an absolute positional embeddings (link) before feeding the tokens to Hyena operators.

I checked the original Hyena and HyenaDNA source codes, and they haven't used any positional embeddings for their models.
My question is why have you used the positional embeddings? Have you tried not using them? Did it worsen the performance?

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions