updated reqs for whisper and cuda

This commit is contained in:
tcsenpai 2025-05-23 11:39:45 +02:00
parent 696cc73e23
commit f7d26a2325
2 changed files with 57 additions and 6 deletions

51
app.py
View File

@ -7,6 +7,8 @@ from typing import List, Tuple, Optional
import youtube_handler as yt
from ollama_handler import OllamaHandler
import logging
import subprocess
import sys
# Configure logging
logging.basicConfig(
@ -487,8 +489,57 @@ def create_interface():
return app
def check_cuda_compatibility():
"""Check if the current CUDA setup is compatible with Whisper."""
logger.info("Checking CUDA compatibility...")
# Check PyTorch CUDA
if not torch.cuda.is_available():
logger.warning("CUDA is not available in PyTorch")
return False
cuda_version = torch.version.cuda
cudnn_version = torch.backends.cudnn.version()
device_name = torch.cuda.get_device_name(0)
logger.info(f"CUDA Version: {cuda_version}")
logger.info(f"cuDNN Version: {cudnn_version}")
logger.info(f"GPU Device: {device_name}")
# Check CUDA version
try:
cuda_major = int(cuda_version.split(".")[0])
if cuda_major > 11:
logger.warning(
f"CUDA {cuda_version} might not be fully compatible with Whisper. Recommended: CUDA 11.x"
)
logger.info(
"Consider creating a new environment with CUDA 11.x if you encounter issues"
)
except Exception as e:
logger.error(f"Error parsing CUDA version: {str(e)}")
# Check if faster-whisper is installed
try:
import faster_whisper
logger.info(f"faster-whisper version: {faster_whisper.__version__}")
except ImportError:
logger.error("faster-whisper is not installed")
return False
return True
if __name__ == "__main__":
logger.info("Starting Whisper Transcription Web App")
# Check CUDA compatibility before starting
if not check_cuda_compatibility():
logger.warning(
"CUDA compatibility check failed. The application might not work as expected."
)
logger.info(f"Server will be available at http://{SERVER_NAME}:{SERVER_PORT}")
app = create_interface()
app.launch(share=SHARE, server_name=SERVER_NAME, server_port=SERVER_PORT)

View File

@ -1,9 +1,9 @@
gradio>=4.0.0
faster-whisper>=0.9.0
torch>=2.0.0,<2.1.0
torchvision>=0.15.0,<0.16.0
torchaudio>=2.0.0,<2.1.0
yt-dlp>=2023.0.0
python-dotenv>=1.0.0
torch>=2.0.0
torchaudio>=2.0.0
yt-dlp>=2023.12.30
pytube>=15.0.0
requests>=2.31.0
ollama>=0.3.0
requests>=2.31.0
ollama>=0.1.0