Skip to content

Commit

Permalink
Conformer, use RelPosSelfAttention
Browse files Browse the repository at this point in the history
Fix #132
  • Loading branch information
albertz committed Oct 18, 2022
1 parent 6c791b1 commit 328faa6
Showing 1 changed file with 1 addition and 1 deletion.
2 changes: 1 addition & 1 deletion nn/conformer.py
Original file line number Diff line number Diff line change
Expand Up @@ -184,7 +184,7 @@ def __init__(
out_dim=out_dim, kernel_size=conv_kernel_size, norm=conv_norm)
self.conv_layer_norm = nn.LayerNorm(out_dim)

self.self_att = nn.SelfAttention(
self.self_att = nn.RelPosSelfAttention(
out_dim, proj_dim=out_dim,
key_dim_total=out_dim, value_dim_total=out_dim, num_heads=num_heads, att_dropout=att_dropout)
self.self_att_layer_norm = nn.LayerNorm(out_dim)
Expand Down

0 comments on commit 328faa6

Please sign in to comment.