mirror of
https://github.com/Monadical-SAS/reflector.git
synced 2025-12-20 12:19:06 +00:00
235 lines
8.1 KiB
Python
235 lines
8.1 KiB
Python
import json
|
|
import os
|
|
from datetime import datetime
|
|
from urllib.parse import unquote
|
|
|
|
import av
|
|
import boto3
|
|
import structlog
|
|
from celery import shared_task
|
|
from celery.utils.log import get_task_logger
|
|
from pydantic import ValidationError
|
|
from reflector.db.meetings import meetings_controller
|
|
from reflector.db.recordings import Recording, recordings_controller
|
|
from reflector.db.rooms import rooms_controller
|
|
from reflector.db.transcripts import SourceKind, transcripts_controller
|
|
from reflector.pipelines.main_live_pipeline import asynctask, task_pipeline_process
|
|
from reflector.settings import settings
|
|
from reflector.whereby import get_room_sessions
|
|
|
|
logger = structlog.wrap_logger(get_task_logger(__name__))
|
|
|
|
|
|
@shared_task
|
|
def process_messages():
|
|
queue_url = settings.AWS_PROCESS_RECORDING_QUEUE_URL
|
|
if not queue_url:
|
|
logger.warning("No process recording queue url")
|
|
return
|
|
try:
|
|
logger.info("Receiving messages from: %s", queue_url)
|
|
sqs = boto3.client(
|
|
"sqs",
|
|
region_name=settings.TRANSCRIPT_STORAGE_AWS_REGION,
|
|
aws_access_key_id=settings.TRANSCRIPT_STORAGE_AWS_ACCESS_KEY_ID,
|
|
aws_secret_access_key=settings.TRANSCRIPT_STORAGE_AWS_SECRET_ACCESS_KEY,
|
|
)
|
|
|
|
response = sqs.receive_message(
|
|
QueueUrl=queue_url,
|
|
AttributeNames=["SentTimestamp"],
|
|
MaxNumberOfMessages=1,
|
|
MessageAttributeNames=["All"],
|
|
VisibilityTimeout=0,
|
|
WaitTimeSeconds=0,
|
|
)
|
|
|
|
for message in response.get("Messages", []):
|
|
receipt_handle = message["ReceiptHandle"]
|
|
body = json.loads(message["Body"])
|
|
|
|
for record in body.get("Records", []):
|
|
if record["eventName"].startswith("ObjectCreated"):
|
|
bucket = record["s3"]["bucket"]["name"]
|
|
key = unquote(record["s3"]["object"]["key"])
|
|
process_recording.delay(bucket, key)
|
|
|
|
sqs.delete_message(QueueUrl=queue_url, ReceiptHandle=receipt_handle)
|
|
logger.info("Processed and deleted message: %s", message)
|
|
|
|
except Exception as e:
|
|
logger.error("process_messages", error=str(e))
|
|
|
|
|
|
@shared_task
|
|
@asynctask
|
|
async def process_recording(bucket_name: str, object_key: str):
|
|
logger.info("Processing recording: %s/%s", bucket_name, object_key)
|
|
|
|
# extract a guid and a datetime from the object key
|
|
room_name = f"/{object_key[:36]}"
|
|
recorded_at = datetime.fromisoformat(object_key[37:57])
|
|
|
|
meeting = await meetings_controller.get_by_room_name(room_name)
|
|
room = await rooms_controller.get_by_id(meeting.room_id)
|
|
|
|
recording = await recordings_controller.get_by_object_key(bucket_name, object_key)
|
|
if not recording:
|
|
recording = await recordings_controller.create(
|
|
Recording(
|
|
bucket_name=bucket_name,
|
|
object_key=object_key,
|
|
recorded_at=recorded_at,
|
|
meeting_id=meeting.id,
|
|
)
|
|
)
|
|
|
|
transcript = await transcripts_controller.get_by_recording_id(recording.id)
|
|
if transcript:
|
|
await transcripts_controller.update(
|
|
transcript,
|
|
{
|
|
"topics": [],
|
|
},
|
|
)
|
|
else:
|
|
transcript = await transcripts_controller.add(
|
|
"",
|
|
source_kind=SourceKind.ROOM,
|
|
source_language="en",
|
|
target_language="en",
|
|
user_id=room.user_id,
|
|
recording_id=recording.id,
|
|
share_mode="public",
|
|
)
|
|
|
|
_, extension = os.path.splitext(object_key)
|
|
upload_filename = transcript.data_path / f"upload{extension}"
|
|
upload_filename.parent.mkdir(parents=True, exist_ok=True)
|
|
|
|
s3 = boto3.client(
|
|
"s3",
|
|
region_name=settings.TRANSCRIPT_STORAGE_AWS_REGION,
|
|
aws_access_key_id=settings.TRANSCRIPT_STORAGE_AWS_ACCESS_KEY_ID,
|
|
aws_secret_access_key=settings.TRANSCRIPT_STORAGE_AWS_SECRET_ACCESS_KEY,
|
|
)
|
|
|
|
with open(upload_filename, "wb") as f:
|
|
s3.download_fileobj(bucket_name, object_key, f)
|
|
|
|
container = av.open(upload_filename.as_posix())
|
|
try:
|
|
if not len(container.streams.audio):
|
|
raise Exception("File has no audio stream")
|
|
except Exception:
|
|
upload_filename.unlink()
|
|
raise
|
|
finally:
|
|
container.close()
|
|
|
|
await transcripts_controller.update(transcript, {"status": "uploaded"})
|
|
|
|
task_pipeline_process.delay(transcript_id=transcript.id)
|
|
|
|
|
|
@shared_task
|
|
@asynctask
|
|
async def process_meetings():
|
|
logger.info("Processing meetings")
|
|
meetings = await meetings_controller.get_all_active()
|
|
for meeting in meetings:
|
|
is_active = False
|
|
if meeting.end_date > datetime.utcnow():
|
|
response = await get_room_sessions(meeting.room_name)
|
|
room_sessions = response.get("results", [])
|
|
is_active = not room_sessions or any(
|
|
rs["endedAt"] is None for rs in room_sessions
|
|
)
|
|
if not is_active:
|
|
await meetings_controller.update_meeting(meeting.id, is_active=False)
|
|
logger.info("Meeting %s is deactivated", meeting.id)
|
|
|
|
logger.info("Processed meetings")
|
|
|
|
|
|
@shared_task
|
|
@asynctask
|
|
async def reprocess_failed_recordings():
|
|
"""
|
|
Find recordings in the S3 bucket and check if they have proper transcriptions.
|
|
If not, requeue them for processing.
|
|
"""
|
|
logger.info("Checking for recordings that need processing or reprocessing")
|
|
|
|
s3 = boto3.client(
|
|
"s3",
|
|
region_name=settings.TRANSCRIPT_STORAGE_AWS_REGION,
|
|
aws_access_key_id=settings.TRANSCRIPT_STORAGE_AWS_ACCESS_KEY_ID,
|
|
aws_secret_access_key=settings.TRANSCRIPT_STORAGE_AWS_SECRET_ACCESS_KEY,
|
|
)
|
|
|
|
reprocessed_count = 0
|
|
try:
|
|
paginator = s3.get_paginator("list_objects_v2")
|
|
bucket_name = settings.AWS_WHEREBY_S3_BUCKET
|
|
pages = paginator.paginate(Bucket=bucket_name)
|
|
|
|
for page in pages:
|
|
if "Contents" not in page:
|
|
continue
|
|
|
|
for obj in page["Contents"]:
|
|
object_key = obj["Key"]
|
|
|
|
if not (object_key.endswith(".mp4")):
|
|
continue
|
|
|
|
recorded_at = datetime.fromisoformat(object_key[37:57])
|
|
recording = await recordings_controller.get_by_object_key(
|
|
bucket_name, object_key
|
|
)
|
|
if not recording:
|
|
recording = await recordings_controller.create(
|
|
Recording(
|
|
bucket_name=bucket_name,
|
|
object_key=object_key,
|
|
recorded_at=recorded_at,
|
|
meeting_id=meeting.id,
|
|
)
|
|
)
|
|
|
|
room_name = f"/{object_key[:36]}"
|
|
meeting = await meetings_controller.get_by_room_name(room_name)
|
|
if not meeting:
|
|
logger.warning(f"No meeting found for recording: {object_key}")
|
|
continue
|
|
|
|
room = await rooms_controller.get_by_id(meeting.room_id)
|
|
if not room:
|
|
logger.warning(f"No room found for meeting: {meeting.id}")
|
|
continue
|
|
|
|
transcript = None
|
|
try:
|
|
transcript = await transcripts_controller.get_by_recording_id(
|
|
recording.id
|
|
)
|
|
except ValidationError:
|
|
await transcripts_controller.remove_by_recording_id(recording.id)
|
|
logger.warning(
|
|
f"Removed invalid transcript for recording: {recording.id}"
|
|
)
|
|
|
|
if transcript is None or transcript.status == "error":
|
|
logger.info(
|
|
f"Queueing recording for processing: {object_key}, meeting {meeting.id}"
|
|
)
|
|
process_recording.delay(bucket_name, object_key)
|
|
reprocessed_count += 1
|
|
|
|
except Exception as e:
|
|
logger.error(f"Error checking S3 bucket: {str(e)}")
|
|
|
|
logger.info(f"Reprocessing complete. Requeued {reprocessed_count} recordings")
|
|
return reprocessed_count
|