@@ -351,7 +351,7 @@ def forward(
351351 output_hidden_states : Optional [bool ] = None ,
352352 past_key_values : Optional [Cache ] = None ,
353353 use_cache : Optional [bool ] = None ,
354- cache_postion : Optional [torch .Tensor ] = None ,
354+ cache_position : Optional [torch .Tensor ] = None ,
355355 ):
356356 """
357357 Args:
@@ -492,16 +492,16 @@ def _position_bucket(self, relative_position, num_buckets=32, max_distance=128):
492492 relative_position = torch .abs (relative_position )
493493 max_exact = num_buckets // 2
494494 is_small = relative_position < max_exact
495- relative_postion_if_large = max_exact + (
495+ relative_position_if_large = max_exact + (
496496 torch .log (relative_position .float () / max_exact )
497497 / math .log (max_distance / max_exact )
498498 * (num_buckets - max_exact )
499499 ).to (torch .int32 )
500- relative_postion_if_large = torch .min (
501- relative_postion_if_large ,
502- torch .full_like (relative_postion_if_large , num_buckets - 1 ),
500+ relative_position_if_large = torch .min (
501+ relative_position_if_large ,
502+ torch .full_like (relative_position_if_large , num_buckets - 1 ),
503503 )
504- relative_buckets += torch .where (is_small , relative_position .to (torch .int32 ), relative_postion_if_large )
504+ relative_buckets += torch .where (is_small , relative_position .to (torch .int32 ), relative_position_if_large )
505505 return relative_buckets
506506
507507
0 commit comments