diff --git a/gpu/modal_deployments/deploy-all.sh b/gpu/modal_deployments/deploy-all.sh index 93303604..02a1d0f9 100755 --- a/gpu/modal_deployments/deploy-all.sh +++ b/gpu/modal_deployments/deploy-all.sh @@ -114,8 +114,8 @@ modal secret create reflector-gpu REFLECTOR_GPU_APIKEY="$API_KEY" # --- Deploy Functions --- echo "" -echo "Deploying transcriber (Whisper)..." -TRANSCRIBER_URL=$(modal deploy reflector_transcriber.py 2>&1 | grep -o 'https://[^ ]*web.modal.run' | head -1) +echo "Deploying transcriber (Parakeet)..." +TRANSCRIBER_URL=$(modal deploy reflector_transcriber_parakeet.py 2>&1 | grep -o 'https://[^ ]*web.modal.run' | head -1) if [ -z "$TRANSCRIBER_URL" ]; then echo "Error: Failed to deploy transcriber. Check Modal dashboard for details." exit 1 diff --git a/gpu/modal_deployments/reflector_diarizer.py b/gpu/modal_deployments/reflector_diarizer.py index 816e17cf..39369509 100644 --- a/gpu/modal_deployments/reflector_diarizer.py +++ b/gpu/modal_deployments/reflector_diarizer.py @@ -113,12 +113,14 @@ def download_pyannote_audio(): diarizer_image = ( - modal.Image.debian_slim(python_version="3.10") + modal.Image.from_registry( + "nvidia/cuda:11.8.0-cudnn8-devel-ubuntu22.04", add_python="3.10" + ) .pip_install( "pyannote.audio==3.1.0", "requests", "onnx", - "torchaudio", + "torchaudio==2.0.1", "onnxruntime-gpu", "torch==2.0.0", "transformers==4.34.0", @@ -133,14 +135,6 @@ diarizer_image = ( secrets=[modal.Secret.from_name("hf_token")], ) .run_function(migrate_cache_llm) - .env( - { - "LD_LIBRARY_PATH": ( - "/usr/local/lib/python3.10/site-packages/nvidia/cudnn/lib/:" - "/opt/conda/lib/python3.10/site-packages/nvidia/cublas/lib/" - ) - } - ) )