Open
Description
Hi @DanFu09
Hope you're well,
I was reading the source code and the config files, and I realized that use_positional_encodings
is True
(link). So, the M2-BERT model is using an absolute positional embeddings (link) before feeding the tokens to Hyena operators.
I checked the original Hyena and HyenaDNA source codes, and they haven't used any positional embeddings for their models.
My question is why have you used the positional embeddings? Have you tried not using them? Did it worsen the performance?
Metadata
Metadata
Assignees
Labels
No labels