Skip to content

pawelzeja098/whisper-video-transcription

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 
 
 
 
 

Repository files navigation

Testing OpenAI Whisper for Video Transcription

Project Overview

This project demonstrates how to use OpenAI's Whisper model to transcribe audio from video files into text. The workflow involves extracting audio from video using tools like ffmpeg and then leveraging Whisper's capabilities to generate transcriptions.

Getting Started

Prerequisites

Make sure you have the following installed on your system:

  • Python: 3.8 - 3.11
  • ffmpeg: Required for audio extraction from video files
    • On Windows, example to add ffmpeg/bin to your system's PATH in VSC :
      $env:PATH += ";D:\ffmpeg\bin"
    • On Linux, set the LD_LIBRARY_PATH to include the necessary libraries in VSC:
      export LD_LIBRARY_PATH=/lib/x86_64-linux-gnu:/usr/lib/x86_64-linux-gnu

Dependencies

The project requires the following Python packages:

  • Whisper: 20240930 (latest release at the time of writing)
  • MoviePy: 1.0.3
  • Pydub: 0.25.1
  • NumPy: 1.26.4 (Crashing on latest 2.0.2)
  • PyTorch: 2.5.1

You can install all dependencies using the following command:

pip install openai-whisper moviepy pydub numpy

About

Testing whisper Open-AI to transcribe videos

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages