@@ -352,7 +352,7 @@ def forward(
352352 output_hidden_states : Optional [bool ] = None ,
353353 past_key_values : Optional [Cache ] = None ,
354354 use_cache : Optional [bool ] = None ,
355- cache_postion : Optional [torch .Tensor ] = None ,
355+ cache_position : Optional [torch .Tensor ] = None ,
356356 ):
357357 """
358358 Args:
@@ -493,16 +493,16 @@ def _position_bucket(self, relative_position, num_buckets=32, max_distance=128):
493493 relative_position = torch .abs (relative_position )
494494 max_exact = num_buckets // 2
495495 is_small = relative_position < max_exact
496- relative_postion_if_large = max_exact + (
496+ relative_position_if_large = max_exact + (
497497 torch .log (relative_position .float () / max_exact )
498498 / math .log (max_distance / max_exact )
499499 * (num_buckets - max_exact )
500500 ).to (torch .int32 )
501- relative_postion_if_large = torch .min (
502- relative_postion_if_large ,
503- torch .full_like (relative_postion_if_large , num_buckets - 1 ),
501+ relative_position_if_large = torch .min (
502+ relative_position_if_large ,
503+ torch .full_like (relative_position_if_large , num_buckets - 1 ),
504504 )
505- relative_buckets += torch .where (is_small , relative_position .to (torch .int32 ), relative_postion_if_large )
505+ relative_buckets += torch .where (is_small , relative_position .to (torch .int32 ), relative_position_if_large )
506506 return relative_buckets
507507
508508
0 commit comments