Skip to content

whisper : add Metal support in the Decoder #1047

Closed
@ggerganov

Description

@ggerganov

GPU inference on Apple Silicon via Metal backend was recently added to llama.cpp: ggml-org/llama.cpp#1642

We should port the changes to whisper.cpp and allow the Decoder to run on the GPU in a similar way

Metadata

Metadata

Assignees

Labels

decodingDecoding related issuesperformanceCPU and memory usage - results and comparisons

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions