Skip to content

Commit 3bc301e

Browse files
Removed unused methods
Signed-off-by: Alex-Brooks <Alex.Brooks@ibm.com>
1 parent 2b63731 commit 3bc301e

File tree

1 file changed

+1
-15
lines changed

1 file changed

+1
-15
lines changed

vllm/model_executor/models/granite_speech.py

Lines changed: 1 addition & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,7 @@
3434
from vllm.model_executor.layers.linear import (ColumnParallelLinear,
3535
RowParallelLinear)
3636
from vllm.model_executor.layers.quantization import QuantizationConfig
37-
from vllm.model_executor.layers.sampler import SamplerOutput, get_sampler
37+
from vllm.model_executor.layers.sampler import get_sampler
3838
from vllm.model_executor.models.module_mapping import MultiModelKeys
3939
from vllm.model_executor.sampling_metadata import SamplingMetadata
4040
from vllm.multimodal import MULTIMODAL_REGISTRY
@@ -73,13 +73,6 @@ class GraniteSpeechMultiModalProcessingInfo(BaseProcessingInfo):
7373
def get_supported_mm_limits(self) -> Mapping[str, Optional[int]]:
7474
return {"audio": 1}
7575

76-
def get_mm_max_tokens_per_item(
77-
self,
78-
seq_len: int,
79-
mm_counts: Mapping[str, int],
80-
) -> Mapping[str, int]:
81-
return {"audio": self.get_max_audio_tokens()}
82-
8376
# There is no limit to the maximum number of audio tokens that can be
8477
# encoded as features; we pick ~5000 as a number that is probably higher
8578
# than we would expect to encounter. The sequence of length
@@ -768,13 +761,6 @@ def compute_logits(
768761
sampling_metadata,
769762
)
770763

771-
def sample(
772-
self,
773-
logits: torch.Tensor,
774-
sampling_metadata: SamplingMetadata,
775-
) -> Optional[SamplerOutput]:
776-
return self.language_model.sample(logits, sampling_metadata)
777-
778764
def load_weights(
779765
self,
780766
weights: Iterable[Tuple[str, torch.Tensor]],

0 commit comments

Comments
 (0)