mirror of
https://github.com/Monadical-SAS/reflector.git
synced 2025-12-21 12:49:06 +00:00
Compare commits
87 Commits
mathieu/ca
...
v0.17.0
| Author | SHA1 | Date | |
|---|---|---|---|
| 34a3f5618c | |||
|
|
1473fd82dc | ||
| 372202b0e1 | |||
|
|
d20aac66c4 | ||
| dc4b737daa | |||
|
|
0baff7abf7 | ||
|
|
962c40e2b6 | ||
|
|
3c4b9f2103 | ||
|
|
c6c035aacf | ||
| c086b91445 | |||
|
|
9a258abc02 | ||
| af86c47f1d | |||
| 5f6910e513 | |||
| 9a71af145e | |||
| eef6dc3903 | |||
|
|
1dee255fed | ||
| 5d98754305 | |||
|
|
969bd84fcc | ||
|
|
36608849ec | ||
|
|
5bf64b5a41 | ||
| 0aaa42528a | |||
| 565a62900f | |||
|
|
27016e6051 | ||
| 6ddfee0b4e | |||
|
|
47716f6e5d | ||
| 0abcebfc94 | |||
|
|
2b723da08b | ||
| 6566e04300 | |||
| 870e860517 | |||
| 396a95d5ce | |||
| 6f680b5795 | |||
| ab859d65a6 | |||
| fa049e8d06 | |||
| 2ce7479967 | |||
| b42f7cfc60 | |||
| c546e69739 | |||
|
|
3f1fe8c9bf | ||
| 5f143fe364 | |||
|
|
79f161436e | ||
|
|
5cba5d310d | ||
| 43ea9349f5 | |||
|
|
b3a8e9739d | ||
|
|
369ecdff13 | ||
| fc363bd49b | |||
|
|
962038ee3f | ||
|
|
3b85ff3bdf | ||
|
|
cde99ca271 | ||
|
|
f81fe9948a | ||
|
|
5a5b323382 | ||
| 02a3938822 | |||
|
|
7f5a4c9ddc | ||
|
|
08d88ec349 | ||
|
|
c4d2825c81 | ||
| 0663700a61 | |||
| dc82f8bb3b | |||
| 457823e1c1 | |||
|
|
695d1a957d | ||
| ccffdba75b | |||
| 84a381220b | |||
| 5f2f0e9317 | |||
| 88ed7cfa78 | |||
| 6f0c7c1a5e | |||
| 9dfd76996f | |||
| 55cc8637c6 | |||
| f5331a2107 | |||
|
|
124ce03bf8 | ||
| 7030e0f236 | |||
| 37f0110892 | |||
| cf2896a7f4 | |||
| aabf2c2572 | |||
| 6a7b08f016 | |||
| e2736563d9 | |||
| 0f54b7782d | |||
| 359280dd34 | |||
| 9265d201b5 | |||
| 52f9f533d7 | |||
| 0c3878ac3c | |||
|
|
d70beee51b | ||
| bc5b351d2b | |||
|
|
07981e8090 | ||
| 7e366f6338 | |||
| 7592679a35 | |||
| af16178f86 | |||
| 3ea7f6b7b6 | |||
|
|
009590c080 | ||
|
|
fe5d344cff | ||
|
|
86455ce573 |
5
.github/workflows/db_migrations.yml
vendored
5
.github/workflows/db_migrations.yml
vendored
@@ -2,6 +2,8 @@ name: Test Database Migrations
|
|||||||
|
|
||||||
on:
|
on:
|
||||||
push:
|
push:
|
||||||
|
branches:
|
||||||
|
- main
|
||||||
paths:
|
paths:
|
||||||
- "server/migrations/**"
|
- "server/migrations/**"
|
||||||
- "server/reflector/db/**"
|
- "server/reflector/db/**"
|
||||||
@@ -17,6 +19,9 @@ on:
|
|||||||
jobs:
|
jobs:
|
||||||
test-migrations:
|
test-migrations:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
|
concurrency:
|
||||||
|
group: db-ubuntu-latest-${{ github.ref }}
|
||||||
|
cancel-in-progress: true
|
||||||
services:
|
services:
|
||||||
postgres:
|
postgres:
|
||||||
image: postgres:17
|
image: postgres:17
|
||||||
|
|||||||
79
.github/workflows/deploy.yml
vendored
79
.github/workflows/deploy.yml
vendored
@@ -1,4 +1,4 @@
|
|||||||
name: Deploy to Amazon ECS
|
name: Build container/push to container registry
|
||||||
|
|
||||||
on: [workflow_dispatch]
|
on: [workflow_dispatch]
|
||||||
|
|
||||||
@@ -8,18 +8,30 @@ env:
|
|||||||
ECR_REPOSITORY: reflector
|
ECR_REPOSITORY: reflector
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
deploy:
|
build:
|
||||||
runs-on: ubuntu-latest
|
strategy:
|
||||||
|
matrix:
|
||||||
|
include:
|
||||||
|
- platform: linux/amd64
|
||||||
|
runner: linux-amd64
|
||||||
|
arch: amd64
|
||||||
|
- platform: linux/arm64
|
||||||
|
runner: linux-arm64
|
||||||
|
arch: arm64
|
||||||
|
|
||||||
|
runs-on: ${{ matrix.runner }}
|
||||||
|
|
||||||
permissions:
|
permissions:
|
||||||
deployments: write
|
|
||||||
contents: read
|
contents: read
|
||||||
|
|
||||||
|
outputs:
|
||||||
|
registry: ${{ steps.login-ecr.outputs.registry }}
|
||||||
|
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v3
|
- uses: actions/checkout@v4
|
||||||
|
|
||||||
- name: Configure AWS credentials
|
- name: Configure AWS credentials
|
||||||
uses: aws-actions/configure-aws-credentials@0e613a0980cbf65ed5b322eb7a1e075d28913a83
|
uses: aws-actions/configure-aws-credentials@v4
|
||||||
with:
|
with:
|
||||||
aws-access-key-id: ${{ secrets.AWS_ACCESS_KEY_ID }}
|
aws-access-key-id: ${{ secrets.AWS_ACCESS_KEY_ID }}
|
||||||
aws-secret-access-key: ${{ secrets.AWS_SECRET_ACCESS_KEY }}
|
aws-secret-access-key: ${{ secrets.AWS_SECRET_ACCESS_KEY }}
|
||||||
@@ -27,21 +39,52 @@ jobs:
|
|||||||
|
|
||||||
- name: Login to Amazon ECR
|
- name: Login to Amazon ECR
|
||||||
id: login-ecr
|
id: login-ecr
|
||||||
uses: aws-actions/amazon-ecr-login@62f4f872db3836360b72999f4b87f1ff13310f3a
|
uses: aws-actions/amazon-ecr-login@v2
|
||||||
|
|
||||||
- name: Set up QEMU
|
|
||||||
uses: docker/setup-qemu-action@v2
|
|
||||||
|
|
||||||
- name: Set up Docker Buildx
|
- name: Set up Docker Buildx
|
||||||
uses: docker/setup-buildx-action@v2
|
uses: docker/setup-buildx-action@v3
|
||||||
|
|
||||||
- name: Build and push
|
- name: Build and push ${{ matrix.arch }}
|
||||||
id: docker_build
|
uses: docker/build-push-action@v5
|
||||||
uses: docker/build-push-action@v4
|
|
||||||
with:
|
with:
|
||||||
context: server
|
context: server
|
||||||
platforms: linux/amd64,linux/arm64
|
platforms: ${{ matrix.platform }}
|
||||||
push: true
|
push: true
|
||||||
tags: ${{ steps.login-ecr.outputs.registry }}/${{ env.ECR_REPOSITORY }}:latest
|
tags: ${{ steps.login-ecr.outputs.registry }}/${{ env.ECR_REPOSITORY }}:latest-${{ matrix.arch }}
|
||||||
cache-from: type=gha
|
cache-from: type=gha,scope=${{ matrix.arch }}
|
||||||
cache-to: type=gha,mode=max
|
cache-to: type=gha,mode=max,scope=${{ matrix.arch }}
|
||||||
|
github-token: ${{ secrets.GHA_CACHE_TOKEN }}
|
||||||
|
provenance: false
|
||||||
|
|
||||||
|
create-manifest:
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
needs: [build]
|
||||||
|
|
||||||
|
permissions:
|
||||||
|
deployments: write
|
||||||
|
contents: read
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- name: Configure AWS credentials
|
||||||
|
uses: aws-actions/configure-aws-credentials@v4
|
||||||
|
with:
|
||||||
|
aws-access-key-id: ${{ secrets.AWS_ACCESS_KEY_ID }}
|
||||||
|
aws-secret-access-key: ${{ secrets.AWS_SECRET_ACCESS_KEY }}
|
||||||
|
aws-region: ${{ env.AWS_REGION }}
|
||||||
|
|
||||||
|
- name: Login to Amazon ECR
|
||||||
|
uses: aws-actions/amazon-ecr-login@v2
|
||||||
|
|
||||||
|
- name: Create and push multi-arch manifest
|
||||||
|
run: |
|
||||||
|
# Get the registry URL (since we can't easily access job outputs in matrix)
|
||||||
|
ECR_REGISTRY=$(aws ecr describe-registry --query 'registryId' --output text).dkr.ecr.${{ env.AWS_REGION }}.amazonaws.com
|
||||||
|
|
||||||
|
docker manifest create \
|
||||||
|
$ECR_REGISTRY/${{ env.ECR_REPOSITORY }}:latest \
|
||||||
|
$ECR_REGISTRY/${{ env.ECR_REPOSITORY }}:latest-amd64 \
|
||||||
|
$ECR_REGISTRY/${{ env.ECR_REPOSITORY }}:latest-arm64
|
||||||
|
|
||||||
|
docker manifest push $ECR_REGISTRY/${{ env.ECR_REPOSITORY }}:latest
|
||||||
|
|
||||||
|
echo "✅ Multi-arch manifest pushed: $ECR_REGISTRY/${{ env.ECR_REPOSITORY }}:latest"
|
||||||
|
|||||||
57
.github/workflows/docker-frontend.yml
vendored
Normal file
57
.github/workflows/docker-frontend.yml
vendored
Normal file
@@ -0,0 +1,57 @@
|
|||||||
|
name: Build and Push Frontend Docker Image
|
||||||
|
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
branches:
|
||||||
|
- main
|
||||||
|
paths:
|
||||||
|
- 'www/**'
|
||||||
|
- '.github/workflows/docker-frontend.yml'
|
||||||
|
workflow_dispatch:
|
||||||
|
|
||||||
|
env:
|
||||||
|
REGISTRY: ghcr.io
|
||||||
|
IMAGE_NAME: ${{ github.repository }}-frontend
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
build-and-push:
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
permissions:
|
||||||
|
contents: read
|
||||||
|
packages: write
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- name: Checkout repository
|
||||||
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
|
- name: Log in to GitHub Container Registry
|
||||||
|
uses: docker/login-action@v3
|
||||||
|
with:
|
||||||
|
registry: ${{ env.REGISTRY }}
|
||||||
|
username: ${{ github.actor }}
|
||||||
|
password: ${{ secrets.GITHUB_TOKEN }}
|
||||||
|
|
||||||
|
- name: Extract metadata
|
||||||
|
id: meta
|
||||||
|
uses: docker/metadata-action@v5
|
||||||
|
with:
|
||||||
|
images: ${{ env.REGISTRY }}/${{ env.IMAGE_NAME }}
|
||||||
|
tags: |
|
||||||
|
type=ref,event=branch
|
||||||
|
type=sha,prefix={{branch}}-
|
||||||
|
type=raw,value=latest,enable={{is_default_branch}}
|
||||||
|
|
||||||
|
- name: Set up Docker Buildx
|
||||||
|
uses: docker/setup-buildx-action@v3
|
||||||
|
|
||||||
|
- name: Build and push Docker image
|
||||||
|
uses: docker/build-push-action@v5
|
||||||
|
with:
|
||||||
|
context: ./www
|
||||||
|
file: ./www/Dockerfile
|
||||||
|
push: true
|
||||||
|
tags: ${{ steps.meta.outputs.tags }}
|
||||||
|
labels: ${{ steps.meta.outputs.labels }}
|
||||||
|
cache-from: type=gha
|
||||||
|
cache-to: type=gha,mode=max
|
||||||
|
platforms: linux/amd64,linux/arm64
|
||||||
45
.github/workflows/test_next_server.yml
vendored
Normal file
45
.github/workflows/test_next_server.yml
vendored
Normal file
@@ -0,0 +1,45 @@
|
|||||||
|
name: Test Next Server
|
||||||
|
|
||||||
|
on:
|
||||||
|
pull_request:
|
||||||
|
paths:
|
||||||
|
- "www/**"
|
||||||
|
push:
|
||||||
|
branches:
|
||||||
|
- main
|
||||||
|
paths:
|
||||||
|
- "www/**"
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
test-next-server:
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
|
||||||
|
defaults:
|
||||||
|
run:
|
||||||
|
working-directory: ./www
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@v4
|
||||||
|
|
||||||
|
- name: Setup Node.js
|
||||||
|
uses: actions/setup-node@v4
|
||||||
|
with:
|
||||||
|
node-version: '20'
|
||||||
|
|
||||||
|
- name: Install pnpm
|
||||||
|
uses: pnpm/action-setup@v4
|
||||||
|
with:
|
||||||
|
version: 8
|
||||||
|
|
||||||
|
- name: Setup Node.js cache
|
||||||
|
uses: actions/setup-node@v4
|
||||||
|
with:
|
||||||
|
node-version: '20'
|
||||||
|
cache: 'pnpm'
|
||||||
|
cache-dependency-path: './www/pnpm-lock.yaml'
|
||||||
|
|
||||||
|
- name: Install dependencies
|
||||||
|
run: pnpm install
|
||||||
|
|
||||||
|
- name: Run tests
|
||||||
|
run: pnpm test
|
||||||
49
.github/workflows/test_server.yml
vendored
49
.github/workflows/test_server.yml
vendored
@@ -5,12 +5,17 @@ on:
|
|||||||
paths:
|
paths:
|
||||||
- "server/**"
|
- "server/**"
|
||||||
push:
|
push:
|
||||||
|
branches:
|
||||||
|
- main
|
||||||
paths:
|
paths:
|
||||||
- "server/**"
|
- "server/**"
|
||||||
|
|
||||||
jobs:
|
jobs:
|
||||||
pytest:
|
pytest:
|
||||||
runs-on: ubuntu-latest
|
runs-on: ubuntu-latest
|
||||||
|
concurrency:
|
||||||
|
group: pytest-${{ github.ref }}
|
||||||
|
cancel-in-progress: true
|
||||||
services:
|
services:
|
||||||
redis:
|
redis:
|
||||||
image: redis:6
|
image: redis:6
|
||||||
@@ -19,29 +24,47 @@ jobs:
|
|||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v4
|
- uses: actions/checkout@v4
|
||||||
- name: Install uv
|
- name: Install uv
|
||||||
uses: astral-sh/setup-uv@v3
|
uses: astral-sh/setup-uv@v6
|
||||||
with:
|
with:
|
||||||
enable-cache: true
|
enable-cache: true
|
||||||
working-directory: server
|
working-directory: server
|
||||||
|
|
||||||
- name: Tests
|
- name: Tests
|
||||||
run: |
|
run: |
|
||||||
cd server
|
cd server
|
||||||
uv run -m pytest -v tests
|
uv run -m pytest -v tests
|
||||||
|
|
||||||
docker:
|
docker-amd64:
|
||||||
runs-on: ubuntu-latest
|
runs-on: linux-amd64
|
||||||
|
concurrency:
|
||||||
|
group: docker-amd64-${{ github.ref }}
|
||||||
|
cancel-in-progress: true
|
||||||
steps:
|
steps:
|
||||||
- uses: actions/checkout@v4
|
- uses: actions/checkout@v4
|
||||||
- name: Set up QEMU
|
|
||||||
uses: docker/setup-qemu-action@v2
|
|
||||||
- name: Set up Docker Buildx
|
- name: Set up Docker Buildx
|
||||||
uses: docker/setup-buildx-action@v2
|
uses: docker/setup-buildx-action@v3
|
||||||
- name: Build and push
|
- name: Build AMD64
|
||||||
id: docker_build
|
uses: docker/build-push-action@v6
|
||||||
uses: docker/build-push-action@v4
|
|
||||||
with:
|
with:
|
||||||
context: server
|
context: server
|
||||||
platforms: linux/amd64,linux/arm64
|
platforms: linux/amd64
|
||||||
cache-from: type=gha
|
cache-from: type=gha,scope=amd64
|
||||||
cache-to: type=gha,mode=max
|
cache-to: type=gha,mode=max,scope=amd64
|
||||||
|
github-token: ${{ secrets.GHA_CACHE_TOKEN }}
|
||||||
|
|
||||||
|
docker-arm64:
|
||||||
|
runs-on: linux-arm64
|
||||||
|
concurrency:
|
||||||
|
group: docker-arm64-${{ github.ref }}
|
||||||
|
cancel-in-progress: true
|
||||||
|
steps:
|
||||||
|
- uses: actions/checkout@v4
|
||||||
|
- name: Set up Docker Buildx
|
||||||
|
uses: docker/setup-buildx-action@v3
|
||||||
|
- name: Build ARM64
|
||||||
|
uses: docker/build-push-action@v6
|
||||||
|
with:
|
||||||
|
context: server
|
||||||
|
platforms: linux/arm64
|
||||||
|
cache-from: type=gha,scope=arm64
|
||||||
|
cache-to: type=gha,mode=max,scope=arm64
|
||||||
|
github-token: ${{ secrets.GHA_CACHE_TOKEN }}
|
||||||
|
|||||||
3
.gitignore
vendored
3
.gitignore
vendored
@@ -15,3 +15,6 @@ www/REFACTOR.md
|
|||||||
www/reload-frontend
|
www/reload-frontend
|
||||||
server/test.sqlite
|
server/test.sqlite
|
||||||
CLAUDE.local.md
|
CLAUDE.local.md
|
||||||
|
www/.env.development
|
||||||
|
www/.env.production
|
||||||
|
.playwright-mcp
|
||||||
|
|||||||
1
.gitleaksignore
Normal file
1
.gitleaksignore
Normal file
@@ -0,0 +1 @@
|
|||||||
|
b9d891d3424f371642cb032ecfd0e2564470a72c:server/tests/test_transcripts_recording_deletion.py:generic-api-key:15
|
||||||
@@ -27,3 +27,8 @@ repos:
|
|||||||
files: ^server/
|
files: ^server/
|
||||||
- id: ruff-format
|
- id: ruff-format
|
||||||
files: ^server/
|
files: ^server/
|
||||||
|
|
||||||
|
- repo: https://github.com/gitleaks/gitleaks
|
||||||
|
rev: v8.28.0
|
||||||
|
hooks:
|
||||||
|
- id: gitleaks
|
||||||
|
|||||||
192
CHANGELOG.md
192
CHANGELOG.md
@@ -1,5 +1,197 @@
|
|||||||
# Changelog
|
# Changelog
|
||||||
|
|
||||||
|
## [0.17.0](https://github.com/Monadical-SAS/reflector/compare/v0.16.0...v0.17.0) (2025-11-13)
|
||||||
|
|
||||||
|
|
||||||
|
### Features
|
||||||
|
|
||||||
|
* add API key management UI ([#716](https://github.com/Monadical-SAS/reflector/issues/716)) ([372202b](https://github.com/Monadical-SAS/reflector/commit/372202b0e1a86823900b0aa77be1bfbc2893d8a1))
|
||||||
|
* daily.co support as alternative to whereby ([#691](https://github.com/Monadical-SAS/reflector/issues/691)) ([1473fd8](https://github.com/Monadical-SAS/reflector/commit/1473fd82dc472c394cbaa2987212ad662a74bcac))
|
||||||
|
|
||||||
|
## [0.16.0](https://github.com/Monadical-SAS/reflector/compare/v0.15.0...v0.16.0) (2025-10-24)
|
||||||
|
|
||||||
|
|
||||||
|
### Features
|
||||||
|
|
||||||
|
* search date filter ([#710](https://github.com/Monadical-SAS/reflector/issues/710)) ([962c40e](https://github.com/Monadical-SAS/reflector/commit/962c40e2b6428ac42fd10aea926782d7a6f3f902))
|
||||||
|
|
||||||
|
## [0.15.0](https://github.com/Monadical-SAS/reflector/compare/v0.14.0...v0.15.0) (2025-10-20)
|
||||||
|
|
||||||
|
|
||||||
|
### Features
|
||||||
|
|
||||||
|
* api tokens ([#705](https://github.com/Monadical-SAS/reflector/issues/705)) ([9a258ab](https://github.com/Monadical-SAS/reflector/commit/9a258abc0209b0ac3799532a507ea6a9125d703a))
|
||||||
|
|
||||||
|
## [0.14.0](https://github.com/Monadical-SAS/reflector/compare/v0.13.1...v0.14.0) (2025-10-08)
|
||||||
|
|
||||||
|
|
||||||
|
### Features
|
||||||
|
|
||||||
|
* Add calendar event data to transcript webhook payload ([#689](https://github.com/Monadical-SAS/reflector/issues/689)) ([5f6910e](https://github.com/Monadical-SAS/reflector/commit/5f6910e5131b7f28f86c9ecdcc57fed8412ee3cd))
|
||||||
|
* container build for www / github ([#672](https://github.com/Monadical-SAS/reflector/issues/672)) ([969bd84](https://github.com/Monadical-SAS/reflector/commit/969bd84fcc14851d1a101412a0ba115f1b7cde82))
|
||||||
|
* docker-compose for production frontend ([#664](https://github.com/Monadical-SAS/reflector/issues/664)) ([5bf64b5](https://github.com/Monadical-SAS/reflector/commit/5bf64b5a41f64535e22849b4bb11734d4dbb4aae))
|
||||||
|
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
* restore feature boolean logic ([#671](https://github.com/Monadical-SAS/reflector/issues/671)) ([3660884](https://github.com/Monadical-SAS/reflector/commit/36608849ec64e953e3be456172502762e3c33df9))
|
||||||
|
* security review ([#656](https://github.com/Monadical-SAS/reflector/issues/656)) ([5d98754](https://github.com/Monadical-SAS/reflector/commit/5d98754305c6c540dd194dda268544f6d88bfaf8))
|
||||||
|
* update transcript list on reprocess ([#676](https://github.com/Monadical-SAS/reflector/issues/676)) ([9a71af1](https://github.com/Monadical-SAS/reflector/commit/9a71af145ee9b833078c78d0c684590ab12e9f0e))
|
||||||
|
* upgrade nemo toolkit ([#678](https://github.com/Monadical-SAS/reflector/issues/678)) ([eef6dc3](https://github.com/Monadical-SAS/reflector/commit/eef6dc39037329b65804297786d852dddb0557f9))
|
||||||
|
|
||||||
|
## [0.13.1](https://github.com/Monadical-SAS/reflector/compare/v0.13.0...v0.13.1) (2025-09-22)
|
||||||
|
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
* TypeError on not all arguments converted during string formatting in logger ([#667](https://github.com/Monadical-SAS/reflector/issues/667)) ([565a629](https://github.com/Monadical-SAS/reflector/commit/565a62900f5a02fc946b68f9269a42190ed70ab6))
|
||||||
|
|
||||||
|
## [0.13.0](https://github.com/Monadical-SAS/reflector/compare/v0.12.1...v0.13.0) (2025-09-19)
|
||||||
|
|
||||||
|
|
||||||
|
### Features
|
||||||
|
|
||||||
|
* room form edit with enter ([#662](https://github.com/Monadical-SAS/reflector/issues/662)) ([47716f6](https://github.com/Monadical-SAS/reflector/commit/47716f6e5ddee952609d2fa0ffabdfa865286796))
|
||||||
|
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
* invalid cleanup call ([#660](https://github.com/Monadical-SAS/reflector/issues/660)) ([0abcebf](https://github.com/Monadical-SAS/reflector/commit/0abcebfc9491f87f605f21faa3e53996fafedd9a))
|
||||||
|
|
||||||
|
## [0.12.1](https://github.com/Monadical-SAS/reflector/compare/v0.12.0...v0.12.1) (2025-09-17)
|
||||||
|
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
* production blocked because having existing meeting with room_id null ([#657](https://github.com/Monadical-SAS/reflector/issues/657)) ([870e860](https://github.com/Monadical-SAS/reflector/commit/870e8605171a27155a9cbee215eeccb9a8d6c0a2))
|
||||||
|
|
||||||
|
## [0.12.0](https://github.com/Monadical-SAS/reflector/compare/v0.11.0...v0.12.0) (2025-09-17)
|
||||||
|
|
||||||
|
|
||||||
|
### Features
|
||||||
|
|
||||||
|
* calendar integration ([#608](https://github.com/Monadical-SAS/reflector/issues/608)) ([6f680b5](https://github.com/Monadical-SAS/reflector/commit/6f680b57954c688882c4ed49f40f161c52a00a24))
|
||||||
|
* self-hosted gpu api ([#636](https://github.com/Monadical-SAS/reflector/issues/636)) ([ab859d6](https://github.com/Monadical-SAS/reflector/commit/ab859d65a6bded904133a163a081a651b3938d42))
|
||||||
|
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
* ignore player hotkeys for text inputs ([#646](https://github.com/Monadical-SAS/reflector/issues/646)) ([fa049e8](https://github.com/Monadical-SAS/reflector/commit/fa049e8d068190ce7ea015fd9fcccb8543f54a3f))
|
||||||
|
|
||||||
|
## [0.11.0](https://github.com/Monadical-SAS/reflector/compare/v0.10.0...v0.11.0) (2025-09-16)
|
||||||
|
|
||||||
|
|
||||||
|
### Features
|
||||||
|
|
||||||
|
* remove profanity filter that was there for conference ([#652](https://github.com/Monadical-SAS/reflector/issues/652)) ([b42f7cf](https://github.com/Monadical-SAS/reflector/commit/b42f7cfc606783afcee792590efcc78b507468ab))
|
||||||
|
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
* zulip and consent handler on the file pipeline ([#645](https://github.com/Monadical-SAS/reflector/issues/645)) ([5f143fe](https://github.com/Monadical-SAS/reflector/commit/5f143fe3640875dcb56c26694254a93189281d17))
|
||||||
|
* zulip stream and topic selection in share dialog ([#644](https://github.com/Monadical-SAS/reflector/issues/644)) ([c546e69](https://github.com/Monadical-SAS/reflector/commit/c546e69739e68bb74fbc877eb62609928e5b8de6))
|
||||||
|
|
||||||
|
## [0.10.0](https://github.com/Monadical-SAS/reflector/compare/v0.9.0...v0.10.0) (2025-09-11)
|
||||||
|
|
||||||
|
|
||||||
|
### Features
|
||||||
|
|
||||||
|
* replace nextjs-config with environment variables ([#632](https://github.com/Monadical-SAS/reflector/issues/632)) ([369ecdf](https://github.com/Monadical-SAS/reflector/commit/369ecdff13f3862d926a9c0b87df52c9d94c4dde))
|
||||||
|
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
* anonymous users transcript permissions ([#621](https://github.com/Monadical-SAS/reflector/issues/621)) ([f81fe99](https://github.com/Monadical-SAS/reflector/commit/f81fe9948a9237b3e0001b2d8ca84f54d76878f9))
|
||||||
|
* auth post ([#624](https://github.com/Monadical-SAS/reflector/issues/624)) ([cde99ca](https://github.com/Monadical-SAS/reflector/commit/cde99ca2716f84ba26798f289047732f0448742e))
|
||||||
|
* auth post ([#626](https://github.com/Monadical-SAS/reflector/issues/626)) ([3b85ff3](https://github.com/Monadical-SAS/reflector/commit/3b85ff3bdf4fb053b103070646811bc990c0e70a))
|
||||||
|
* auth post ([#627](https://github.com/Monadical-SAS/reflector/issues/627)) ([962038e](https://github.com/Monadical-SAS/reflector/commit/962038ee3f2a555dc3c03856be0e4409456e0996))
|
||||||
|
* missing follow_redirects=True on modal endpoint ([#630](https://github.com/Monadical-SAS/reflector/issues/630)) ([fc363bd](https://github.com/Monadical-SAS/reflector/commit/fc363bd49b17b075e64f9186e5e0185abc325ea7))
|
||||||
|
* sync backend and frontend token refresh logic ([#614](https://github.com/Monadical-SAS/reflector/issues/614)) ([5a5b323](https://github.com/Monadical-SAS/reflector/commit/5a5b3233820df9536da75e87ce6184a983d4713a))
|
||||||
|
|
||||||
|
## [0.9.0](https://github.com/Monadical-SAS/reflector/compare/v0.8.2...v0.9.0) (2025-09-06)
|
||||||
|
|
||||||
|
|
||||||
|
### Features
|
||||||
|
|
||||||
|
* frontend openapi react query ([#606](https://github.com/Monadical-SAS/reflector/issues/606)) ([c4d2825](https://github.com/Monadical-SAS/reflector/commit/c4d2825c81f81ad8835629fbf6ea8c7383f8c31b))
|
||||||
|
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
* align whisper transcriber api with parakeet ([#602](https://github.com/Monadical-SAS/reflector/issues/602)) ([0663700](https://github.com/Monadical-SAS/reflector/commit/0663700a615a4af69a03c96c410f049e23ec9443))
|
||||||
|
* kv use tls explicit ([#610](https://github.com/Monadical-SAS/reflector/issues/610)) ([08d88ec](https://github.com/Monadical-SAS/reflector/commit/08d88ec349f38b0d13e0fa4cb73486c8dfd31836))
|
||||||
|
* source kind for file processing ([#601](https://github.com/Monadical-SAS/reflector/issues/601)) ([dc82f8b](https://github.com/Monadical-SAS/reflector/commit/dc82f8bb3bdf3ab3d4088e592a30fd63907319e1))
|
||||||
|
* token refresh locking ([#613](https://github.com/Monadical-SAS/reflector/issues/613)) ([7f5a4c9](https://github.com/Monadical-SAS/reflector/commit/7f5a4c9ddc7fd098860c8bdda2ca3b57f63ded2f))
|
||||||
|
|
||||||
|
## [0.8.2](https://github.com/Monadical-SAS/reflector/compare/v0.8.1...v0.8.2) (2025-08-29)
|
||||||
|
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
* search-logspam ([#593](https://github.com/Monadical-SAS/reflector/issues/593)) ([695d1a9](https://github.com/Monadical-SAS/reflector/commit/695d1a957d4cd862753049f9beed88836cabd5ab))
|
||||||
|
|
||||||
|
## [0.8.1](https://github.com/Monadical-SAS/reflector/compare/v0.8.0...v0.8.1) (2025-08-29)
|
||||||
|
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
* make webhook secret/url allowing null ([#590](https://github.com/Monadical-SAS/reflector/issues/590)) ([84a3812](https://github.com/Monadical-SAS/reflector/commit/84a381220bc606231d08d6f71d4babc818fa3c75))
|
||||||
|
|
||||||
|
## [0.8.0](https://github.com/Monadical-SAS/reflector/compare/v0.7.3...v0.8.0) (2025-08-29)
|
||||||
|
|
||||||
|
|
||||||
|
### Features
|
||||||
|
|
||||||
|
* **cleanup:** add automatic data retention for public instances ([#574](https://github.com/Monadical-SAS/reflector/issues/574)) ([6f0c7c1](https://github.com/Monadical-SAS/reflector/commit/6f0c7c1a5e751713366886c8e764c2009e12ba72))
|
||||||
|
* **rooms:** add webhook for transcript completion ([#578](https://github.com/Monadical-SAS/reflector/issues/578)) ([88ed7cf](https://github.com/Monadical-SAS/reflector/commit/88ed7cfa7804794b9b54cad4c3facc8a98cf85fd))
|
||||||
|
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
* file pipeline status reporting and websocket updates ([#589](https://github.com/Monadical-SAS/reflector/issues/589)) ([9dfd769](https://github.com/Monadical-SAS/reflector/commit/9dfd76996f851cc52be54feea078adbc0816dc57))
|
||||||
|
* Igor/evaluation ([#575](https://github.com/Monadical-SAS/reflector/issues/575)) ([124ce03](https://github.com/Monadical-SAS/reflector/commit/124ce03bf86044c18313d27228a25da4bc20c9c5))
|
||||||
|
* optimize parakeet transcription batching algorithm ([#577](https://github.com/Monadical-SAS/reflector/issues/577)) ([7030e0f](https://github.com/Monadical-SAS/reflector/commit/7030e0f23649a8cf6c1eb6d5889684a41ce849ec))
|
||||||
|
|
||||||
|
## [0.7.3](https://github.com/Monadical-SAS/reflector/compare/v0.7.2...v0.7.3) (2025-08-22)
|
||||||
|
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
* cleaned repo, and get git-leaks clean ([359280d](https://github.com/Monadical-SAS/reflector/commit/359280dd340433ba4402ed69034094884c825e67))
|
||||||
|
* restore previous behavior on live pipeline + audio downscaler ([#561](https://github.com/Monadical-SAS/reflector/issues/561)) ([9265d20](https://github.com/Monadical-SAS/reflector/commit/9265d201b590d23c628c5f19251b70f473859043))
|
||||||
|
|
||||||
|
## [0.7.2](https://github.com/Monadical-SAS/reflector/compare/v0.7.1...v0.7.2) (2025-08-21)
|
||||||
|
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
* docker image not loading libgomp.so.1 for torch ([#560](https://github.com/Monadical-SAS/reflector/issues/560)) ([773fccd](https://github.com/Monadical-SAS/reflector/commit/773fccd93e887c3493abc2e4a4864dddce610177))
|
||||||
|
* include shared rooms to search ([#558](https://github.com/Monadical-SAS/reflector/issues/558)) ([499eced](https://github.com/Monadical-SAS/reflector/commit/499eced3360b84fb3a90e1c8a3b554290d21adc2))
|
||||||
|
|
||||||
|
## [0.7.1](https://github.com/Monadical-SAS/reflector/compare/v0.7.0...v0.7.1) (2025-08-21)
|
||||||
|
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
* webvtt db null expectation mismatch ([#556](https://github.com/Monadical-SAS/reflector/issues/556)) ([e67ad1a](https://github.com/Monadical-SAS/reflector/commit/e67ad1a4a2054467bfeb1e0258fbac5868aaaf21))
|
||||||
|
|
||||||
|
## [0.7.0](https://github.com/Monadical-SAS/reflector/compare/v0.6.1...v0.7.0) (2025-08-21)
|
||||||
|
|
||||||
|
|
||||||
|
### Features
|
||||||
|
|
||||||
|
* delete recording with transcript ([#547](https://github.com/Monadical-SAS/reflector/issues/547)) ([99cc984](https://github.com/Monadical-SAS/reflector/commit/99cc9840b3f5de01e0adfbfae93234042d706d13))
|
||||||
|
* pipeline improvement with file processing, parakeet, silero-vad ([#540](https://github.com/Monadical-SAS/reflector/issues/540)) ([bcc29c9](https://github.com/Monadical-SAS/reflector/commit/bcc29c9e0050ae215f89d460e9d645aaf6a5e486))
|
||||||
|
* postgresql migration and removal of sqlite in pytest ([#546](https://github.com/Monadical-SAS/reflector/issues/546)) ([cd1990f](https://github.com/Monadical-SAS/reflector/commit/cd1990f8f0fe1503ef5069512f33777a73a93d7f))
|
||||||
|
* search backend ([#537](https://github.com/Monadical-SAS/reflector/issues/537)) ([5f9b892](https://github.com/Monadical-SAS/reflector/commit/5f9b89260c9ef7f3c921319719467df22830453f))
|
||||||
|
* search frontend ([#551](https://github.com/Monadical-SAS/reflector/issues/551)) ([3657242](https://github.com/Monadical-SAS/reflector/commit/365724271ca6e615e3425125a69ae2b46ce39285))
|
||||||
|
|
||||||
|
|
||||||
|
### Bug Fixes
|
||||||
|
|
||||||
|
* evaluation cli event wrap ([#536](https://github.com/Monadical-SAS/reflector/issues/536)) ([941c3db](https://github.com/Monadical-SAS/reflector/commit/941c3db0bdacc7b61fea412f3746cc5a7cb67836))
|
||||||
|
* use structlog not logging ([#550](https://github.com/Monadical-SAS/reflector/issues/550)) ([27e2f81](https://github.com/Monadical-SAS/reflector/commit/27e2f81fda5232e53edc729d3e99c5ef03adbfe9))
|
||||||
|
|
||||||
## [0.6.1](https://github.com/Monadical-SAS/reflector/compare/v0.6.0...v0.6.1) (2025-08-06)
|
## [0.6.1](https://github.com/Monadical-SAS/reflector/compare/v0.6.0...v0.6.1) (2025-08-06)
|
||||||
|
|
||||||
|
|
||||||
|
|||||||
@@ -66,7 +66,6 @@ pnpm install
|
|||||||
|
|
||||||
# Copy configuration templates
|
# Copy configuration templates
|
||||||
cp .env_template .env
|
cp .env_template .env
|
||||||
cp config-template.ts config.ts
|
|
||||||
```
|
```
|
||||||
|
|
||||||
**Development:**
|
**Development:**
|
||||||
@@ -152,7 +151,7 @@ All endpoints prefixed `/v1/`:
|
|||||||
|
|
||||||
**Frontend** (`www/.env`):
|
**Frontend** (`www/.env`):
|
||||||
- `NEXTAUTH_URL`, `NEXTAUTH_SECRET` - Authentication configuration
|
- `NEXTAUTH_URL`, `NEXTAUTH_SECRET` - Authentication configuration
|
||||||
- `NEXT_PUBLIC_REFLECTOR_API_URL` - Backend API endpoint
|
- `REFLECTOR_API_URL` - Backend API endpoint
|
||||||
- `REFLECTOR_DOMAIN_CONFIG` - Feature flags and domain settings
|
- `REFLECTOR_DOMAIN_CONFIG` - Feature flags and domain settings
|
||||||
|
|
||||||
## Testing Strategy
|
## Testing Strategy
|
||||||
|
|||||||
@@ -1,497 +0,0 @@
|
|||||||
# ICS Calendar Integration - Implementation Guide
|
|
||||||
|
|
||||||
## Overview
|
|
||||||
This document provides detailed implementation guidance for integrating ICS calendar feeds with Reflector rooms. Unlike CalDAV which requires complex authentication and protocol handling, ICS integration uses simple HTTP(S) fetching of calendar files.
|
|
||||||
|
|
||||||
## Key Differences from CalDAV Approach
|
|
||||||
|
|
||||||
| Aspect | CalDAV | ICS |
|
|
||||||
|--------|--------|-----|
|
|
||||||
| Protocol | WebDAV extension | HTTP/HTTPS GET |
|
|
||||||
| Authentication | Username/password, OAuth | Tokens embedded in URL |
|
|
||||||
| Data Access | Selective event queries | Full calendar download |
|
|
||||||
| Implementation | Complex (caldav library) | Simple (requests + icalendar) |
|
|
||||||
| Real-time Updates | Supported | Polling only |
|
|
||||||
| Write Access | Yes | No (read-only) |
|
|
||||||
|
|
||||||
## Technical Architecture
|
|
||||||
|
|
||||||
### 1. ICS Fetching Service
|
|
||||||
|
|
||||||
```python
|
|
||||||
# reflector/services/ics_sync.py
|
|
||||||
|
|
||||||
import requests
|
|
||||||
from icalendar import Calendar
|
|
||||||
from typing import List, Optional
|
|
||||||
from datetime import datetime, timedelta
|
|
||||||
|
|
||||||
class ICSFetchService:
|
|
||||||
def __init__(self):
|
|
||||||
self.session = requests.Session()
|
|
||||||
self.session.headers.update({'User-Agent': 'Reflector/1.0'})
|
|
||||||
|
|
||||||
def fetch_ics(self, url: str) -> str:
|
|
||||||
"""Fetch ICS file from URL (authentication via URL token if needed)."""
|
|
||||||
response = self.session.get(url, timeout=30)
|
|
||||||
response.raise_for_status()
|
|
||||||
return response.text
|
|
||||||
|
|
||||||
def parse_ics(self, ics_content: str) -> Calendar:
|
|
||||||
"""Parse ICS content into calendar object."""
|
|
||||||
return Calendar.from_ical(ics_content)
|
|
||||||
|
|
||||||
def extract_room_events(self, calendar: Calendar, room_url: str) -> List[dict]:
|
|
||||||
"""Extract events that match the room URL."""
|
|
||||||
events = []
|
|
||||||
|
|
||||||
for component in calendar.walk():
|
|
||||||
if component.name == "VEVENT":
|
|
||||||
# Check if event matches this room
|
|
||||||
if self._event_matches_room(component, room_url):
|
|
||||||
events.append(self._parse_event(component))
|
|
||||||
|
|
||||||
return events
|
|
||||||
|
|
||||||
def _event_matches_room(self, event, room_url: str) -> bool:
|
|
||||||
"""Check if event location or description contains room URL."""
|
|
||||||
location = str(event.get('LOCATION', ''))
|
|
||||||
description = str(event.get('DESCRIPTION', ''))
|
|
||||||
|
|
||||||
# Support various URL formats
|
|
||||||
patterns = [
|
|
||||||
room_url,
|
|
||||||
room_url.replace('https://', ''),
|
|
||||||
room_url.split('/')[-1], # Just room name
|
|
||||||
]
|
|
||||||
|
|
||||||
for pattern in patterns:
|
|
||||||
if pattern in location or pattern in description:
|
|
||||||
return True
|
|
||||||
|
|
||||||
return False
|
|
||||||
```
|
|
||||||
|
|
||||||
### 2. Database Schema
|
|
||||||
|
|
||||||
```sql
|
|
||||||
-- Modify room table
|
|
||||||
ALTER TABLE room ADD COLUMN ics_url TEXT; -- encrypted to protect embedded tokens
|
|
||||||
ALTER TABLE room ADD COLUMN ics_fetch_interval INTEGER DEFAULT 300; -- seconds
|
|
||||||
ALTER TABLE room ADD COLUMN ics_enabled BOOLEAN DEFAULT FALSE;
|
|
||||||
ALTER TABLE room ADD COLUMN ics_last_sync TIMESTAMP;
|
|
||||||
ALTER TABLE room ADD COLUMN ics_last_etag TEXT; -- for caching
|
|
||||||
|
|
||||||
-- Calendar events table
|
|
||||||
CREATE TABLE calendar_event (
|
|
||||||
id UUID PRIMARY KEY DEFAULT gen_random_uuid(),
|
|
||||||
room_id UUID REFERENCES room(id) ON DELETE CASCADE,
|
|
||||||
external_id TEXT NOT NULL, -- ICS UID
|
|
||||||
title TEXT,
|
|
||||||
description TEXT,
|
|
||||||
start_time TIMESTAMP NOT NULL,
|
|
||||||
end_time TIMESTAMP NOT NULL,
|
|
||||||
attendees JSONB,
|
|
||||||
location TEXT,
|
|
||||||
ics_raw_data TEXT, -- Store raw VEVENT for reference
|
|
||||||
last_synced TIMESTAMP DEFAULT NOW(),
|
|
||||||
is_deleted BOOLEAN DEFAULT FALSE,
|
|
||||||
created_at TIMESTAMP DEFAULT NOW(),
|
|
||||||
updated_at TIMESTAMP DEFAULT NOW(),
|
|
||||||
UNIQUE(room_id, external_id)
|
|
||||||
);
|
|
||||||
|
|
||||||
-- Index for efficient queries
|
|
||||||
CREATE INDEX idx_calendar_event_room_start ON calendar_event(room_id, start_time);
|
|
||||||
CREATE INDEX idx_calendar_event_deleted ON calendar_event(is_deleted) WHERE NOT is_deleted;
|
|
||||||
```
|
|
||||||
|
|
||||||
### 3. Background Tasks
|
|
||||||
|
|
||||||
```python
|
|
||||||
# reflector/worker/tasks/ics_sync.py
|
|
||||||
|
|
||||||
from celery import shared_task
|
|
||||||
from datetime import datetime, timedelta
|
|
||||||
import hashlib
|
|
||||||
|
|
||||||
@shared_task
|
|
||||||
def sync_ics_calendars():
|
|
||||||
"""Sync all enabled ICS calendars based on their fetch intervals."""
|
|
||||||
rooms = Room.query.filter_by(ics_enabled=True).all()
|
|
||||||
|
|
||||||
for room in rooms:
|
|
||||||
# Check if it's time to sync based on fetch interval
|
|
||||||
if should_sync(room):
|
|
||||||
sync_room_calendar.delay(room.id)
|
|
||||||
|
|
||||||
@shared_task
|
|
||||||
def sync_room_calendar(room_id: str):
|
|
||||||
"""Sync calendar for a specific room."""
|
|
||||||
room = Room.query.get(room_id)
|
|
||||||
if not room or not room.ics_enabled:
|
|
||||||
return
|
|
||||||
|
|
||||||
try:
|
|
||||||
# Fetch ICS file (decrypt URL first)
|
|
||||||
service = ICSFetchService()
|
|
||||||
decrypted_url = decrypt_ics_url(room.ics_url)
|
|
||||||
ics_content = service.fetch_ics(decrypted_url)
|
|
||||||
|
|
||||||
# Check if content changed (using ETag or hash)
|
|
||||||
content_hash = hashlib.md5(ics_content.encode()).hexdigest()
|
|
||||||
if room.ics_last_etag == content_hash:
|
|
||||||
logger.info(f"No changes in ICS for room {room_id}")
|
|
||||||
return
|
|
||||||
|
|
||||||
# Parse and extract events
|
|
||||||
calendar = service.parse_ics(ics_content)
|
|
||||||
events = service.extract_room_events(calendar, room.url)
|
|
||||||
|
|
||||||
# Update database
|
|
||||||
sync_events_to_database(room_id, events)
|
|
||||||
|
|
||||||
# Update sync metadata
|
|
||||||
room.ics_last_sync = datetime.utcnow()
|
|
||||||
room.ics_last_etag = content_hash
|
|
||||||
db.session.commit()
|
|
||||||
|
|
||||||
except Exception as e:
|
|
||||||
logger.error(f"Failed to sync ICS for room {room_id}: {e}")
|
|
||||||
|
|
||||||
def should_sync(room) -> bool:
|
|
||||||
"""Check if room calendar should be synced."""
|
|
||||||
if not room.ics_last_sync:
|
|
||||||
return True
|
|
||||||
|
|
||||||
time_since_sync = datetime.utcnow() - room.ics_last_sync
|
|
||||||
return time_since_sync.total_seconds() >= room.ics_fetch_interval
|
|
||||||
```
|
|
||||||
|
|
||||||
### 4. Celery Beat Schedule
|
|
||||||
|
|
||||||
```python
|
|
||||||
# reflector/worker/celeryconfig.py
|
|
||||||
|
|
||||||
from celery.schedules import crontab
|
|
||||||
|
|
||||||
beat_schedule = {
|
|
||||||
'sync-ics-calendars': {
|
|
||||||
'task': 'reflector.worker.tasks.ics_sync.sync_ics_calendars',
|
|
||||||
'schedule': 60.0, # Check every minute which calendars need syncing
|
|
||||||
},
|
|
||||||
'pre-create-meetings': {
|
|
||||||
'task': 'reflector.worker.tasks.ics_sync.pre_create_calendar_meetings',
|
|
||||||
'schedule': 60.0, # Check every minute for upcoming meetings
|
|
||||||
},
|
|
||||||
}
|
|
||||||
```
|
|
||||||
|
|
||||||
## API Endpoints
|
|
||||||
|
|
||||||
### Room ICS Configuration
|
|
||||||
|
|
||||||
```python
|
|
||||||
# PATCH /v1/rooms/{room_id}
|
|
||||||
{
|
|
||||||
"ics_url": "https://calendar.google.com/calendar/ical/.../private-token/basic.ics",
|
|
||||||
"ics_fetch_interval": 300, # seconds
|
|
||||||
"ics_enabled": true
|
|
||||||
# URL will be encrypted in database to protect embedded tokens
|
|
||||||
}
|
|
||||||
```
|
|
||||||
|
|
||||||
### Manual Sync Trigger
|
|
||||||
|
|
||||||
```python
|
|
||||||
# POST /v1/rooms/{room_name}/ics/sync
|
|
||||||
# Response:
|
|
||||||
{
|
|
||||||
"status": "syncing",
|
|
||||||
"last_sync": "2024-01-15T10:30:00Z",
|
|
||||||
"events_found": 5
|
|
||||||
}
|
|
||||||
```
|
|
||||||
|
|
||||||
### ICS Status
|
|
||||||
|
|
||||||
```python
|
|
||||||
# GET /v1/rooms/{room_name}/ics/status
|
|
||||||
# Response:
|
|
||||||
{
|
|
||||||
"enabled": true,
|
|
||||||
"last_sync": "2024-01-15T10:30:00Z",
|
|
||||||
"next_sync": "2024-01-15T10:35:00Z",
|
|
||||||
"fetch_interval": 300,
|
|
||||||
"events_count": 12,
|
|
||||||
"upcoming_events": 3
|
|
||||||
}
|
|
||||||
```
|
|
||||||
|
|
||||||
## ICS Parsing Details
|
|
||||||
|
|
||||||
### Event Field Mapping
|
|
||||||
|
|
||||||
| ICS Field | Database Field | Notes |
|
|
||||||
|-----------|---------------|-------|
|
|
||||||
| UID | external_id | Unique identifier |
|
|
||||||
| SUMMARY | title | Event title |
|
|
||||||
| DESCRIPTION | description | Full description |
|
|
||||||
| DTSTART | start_time | Convert to UTC |
|
|
||||||
| DTEND | end_time | Convert to UTC |
|
|
||||||
| LOCATION | location | Check for room URL |
|
|
||||||
| ATTENDEE | attendees | Parse into JSON |
|
|
||||||
| ORGANIZER | attendees | Add as organizer |
|
|
||||||
| STATUS | (internal) | Filter cancelled events |
|
|
||||||
|
|
||||||
### Handling Recurring Events
|
|
||||||
|
|
||||||
```python
|
|
||||||
def expand_recurring_events(event, start_date, end_date):
|
|
||||||
"""Expand recurring events into individual occurrences."""
|
|
||||||
from dateutil.rrule import rrulestr
|
|
||||||
|
|
||||||
if 'RRULE' not in event:
|
|
||||||
return [event]
|
|
||||||
|
|
||||||
# Parse recurrence rule
|
|
||||||
rrule_str = event['RRULE'].to_ical().decode()
|
|
||||||
dtstart = event['DTSTART'].dt
|
|
||||||
|
|
||||||
# Generate occurrences
|
|
||||||
rrule = rrulestr(rrule_str, dtstart=dtstart)
|
|
||||||
occurrences = []
|
|
||||||
|
|
||||||
for dt in rrule.between(start_date, end_date):
|
|
||||||
# Clone event with new date
|
|
||||||
occurrence = event.copy()
|
|
||||||
occurrence['DTSTART'].dt = dt
|
|
||||||
if 'DTEND' in event:
|
|
||||||
duration = event['DTEND'].dt - event['DTSTART'].dt
|
|
||||||
occurrence['DTEND'].dt = dt + duration
|
|
||||||
|
|
||||||
# Unique ID for each occurrence
|
|
||||||
occurrence['UID'] = f"{event['UID']}_{dt.isoformat()}"
|
|
||||||
occurrences.append(occurrence)
|
|
||||||
|
|
||||||
return occurrences
|
|
||||||
```
|
|
||||||
|
|
||||||
### Timezone Handling
|
|
||||||
|
|
||||||
```python
|
|
||||||
def normalize_datetime(dt):
|
|
||||||
"""Convert various datetime formats to UTC."""
|
|
||||||
import pytz
|
|
||||||
from datetime import datetime
|
|
||||||
|
|
||||||
if hasattr(dt, 'dt'): # icalendar property
|
|
||||||
dt = dt.dt
|
|
||||||
|
|
||||||
if isinstance(dt, datetime):
|
|
||||||
if dt.tzinfo is None:
|
|
||||||
# Assume local timezone if naive
|
|
||||||
dt = pytz.timezone('UTC').localize(dt)
|
|
||||||
else:
|
|
||||||
# Convert to UTC
|
|
||||||
dt = dt.astimezone(pytz.UTC)
|
|
||||||
|
|
||||||
return dt
|
|
||||||
```
|
|
||||||
|
|
||||||
## Security Considerations
|
|
||||||
|
|
||||||
### 1. URL Validation
|
|
||||||
|
|
||||||
```python
|
|
||||||
def validate_ics_url(url: str) -> bool:
|
|
||||||
"""Validate ICS URL for security."""
|
|
||||||
from urllib.parse import urlparse
|
|
||||||
|
|
||||||
parsed = urlparse(url)
|
|
||||||
|
|
||||||
# Must be HTTPS in production
|
|
||||||
if not settings.DEBUG and parsed.scheme != 'https':
|
|
||||||
return False
|
|
||||||
|
|
||||||
# Prevent local file access
|
|
||||||
if parsed.scheme in ('file', 'ftp'):
|
|
||||||
return False
|
|
||||||
|
|
||||||
# Prevent internal network access
|
|
||||||
if is_internal_ip(parsed.hostname):
|
|
||||||
return False
|
|
||||||
|
|
||||||
return True
|
|
||||||
```
|
|
||||||
|
|
||||||
### 2. Rate Limiting
|
|
||||||
|
|
||||||
```python
|
|
||||||
# Implement per-room rate limiting
|
|
||||||
RATE_LIMITS = {
|
|
||||||
'min_fetch_interval': 60, # Minimum 1 minute between fetches
|
|
||||||
'max_requests_per_hour': 60, # Max 60 requests per hour per room
|
|
||||||
'max_file_size': 10 * 1024 * 1024, # Max 10MB ICS file
|
|
||||||
}
|
|
||||||
```
|
|
||||||
|
|
||||||
### 3. ICS URL Encryption
|
|
||||||
|
|
||||||
```python
|
|
||||||
from cryptography.fernet import Fernet
|
|
||||||
|
|
||||||
class URLEncryption:
|
|
||||||
def __init__(self):
|
|
||||||
self.cipher = Fernet(settings.ENCRYPTION_KEY)
|
|
||||||
|
|
||||||
def encrypt_url(self, url: str) -> str:
|
|
||||||
"""Encrypt ICS URL to protect embedded tokens."""
|
|
||||||
return self.cipher.encrypt(url.encode()).decode()
|
|
||||||
|
|
||||||
def decrypt_url(self, encrypted: str) -> str:
|
|
||||||
"""Decrypt ICS URL for fetching."""
|
|
||||||
return self.cipher.decrypt(encrypted.encode()).decode()
|
|
||||||
|
|
||||||
def mask_url(self, url: str) -> str:
|
|
||||||
"""Mask sensitive parts of URL for display."""
|
|
||||||
from urllib.parse import urlparse, urlunparse
|
|
||||||
|
|
||||||
parsed = urlparse(url)
|
|
||||||
# Keep scheme, host, and path structure but mask tokens
|
|
||||||
if '/private-' in parsed.path:
|
|
||||||
# Google Calendar format
|
|
||||||
parts = parsed.path.split('/private-')
|
|
||||||
masked_path = parts[0] + '/private-***' + parts[1].split('/')[-1]
|
|
||||||
elif 'token=' in url:
|
|
||||||
# Query parameter token
|
|
||||||
masked_path = parsed.path
|
|
||||||
parsed = parsed._replace(query='token=***')
|
|
||||||
else:
|
|
||||||
# Generic masking of path segments that look like tokens
|
|
||||||
import re
|
|
||||||
masked_path = re.sub(r'/[a-zA-Z0-9]{20,}/', '/***/', parsed.path)
|
|
||||||
|
|
||||||
return urlunparse(parsed._replace(path=masked_path))
|
|
||||||
```
|
|
||||||
|
|
||||||
## Testing Strategy
|
|
||||||
|
|
||||||
### 1. Unit Tests
|
|
||||||
|
|
||||||
```python
|
|
||||||
# tests/test_ics_sync.py
|
|
||||||
|
|
||||||
def test_ics_parsing():
|
|
||||||
"""Test ICS file parsing."""
|
|
||||||
ics_content = """BEGIN:VCALENDAR
|
|
||||||
VERSION:2.0
|
|
||||||
BEGIN:VEVENT
|
|
||||||
UID:test-123
|
|
||||||
SUMMARY:Team Meeting
|
|
||||||
LOCATION:https://reflector.monadical.com/engineering
|
|
||||||
DTSTART:20240115T100000Z
|
|
||||||
DTEND:20240115T110000Z
|
|
||||||
END:VEVENT
|
|
||||||
END:VCALENDAR"""
|
|
||||||
|
|
||||||
service = ICSFetchService()
|
|
||||||
calendar = service.parse_ics(ics_content)
|
|
||||||
events = service.extract_room_events(
|
|
||||||
calendar,
|
|
||||||
"https://reflector.monadical.com/engineering"
|
|
||||||
)
|
|
||||||
|
|
||||||
assert len(events) == 1
|
|
||||||
assert events[0]['title'] == 'Team Meeting'
|
|
||||||
```
|
|
||||||
|
|
||||||
### 2. Integration Tests
|
|
||||||
|
|
||||||
```python
|
|
||||||
def test_full_sync_flow():
|
|
||||||
"""Test complete sync workflow."""
|
|
||||||
# Create room with ICS URL (encrypt URL to protect tokens)
|
|
||||||
encryption = URLEncryption()
|
|
||||||
room = Room(
|
|
||||||
name="test-room",
|
|
||||||
ics_url=encryption.encrypt_url("https://example.com/calendar.ics?token=secret"),
|
|
||||||
ics_enabled=True
|
|
||||||
)
|
|
||||||
|
|
||||||
# Mock ICS fetch
|
|
||||||
with patch('requests.get') as mock_get:
|
|
||||||
mock_get.return_value.text = sample_ics_content
|
|
||||||
|
|
||||||
# Run sync
|
|
||||||
sync_room_calendar(room.id)
|
|
||||||
|
|
||||||
# Verify events created
|
|
||||||
events = CalendarEvent.query.filter_by(room_id=room.id).all()
|
|
||||||
assert len(events) > 0
|
|
||||||
```
|
|
||||||
|
|
||||||
## Common ICS Provider Configurations
|
|
||||||
|
|
||||||
### Google Calendar
|
|
||||||
- URL Format: `https://calendar.google.com/calendar/ical/{calendar_id}/private-{token}/basic.ics`
|
|
||||||
- Authentication via token embedded in URL
|
|
||||||
- Updates every 3-8 hours by default
|
|
||||||
|
|
||||||
### Outlook/Office 365
|
|
||||||
- URL Format: `https://outlook.office365.com/owa/calendar/{id}/calendar.ics`
|
|
||||||
- May include token in URL path or query parameters
|
|
||||||
- Real-time updates
|
|
||||||
|
|
||||||
### Apple iCloud
|
|
||||||
- URL Format: `webcal://p{XX}-caldav.icloud.com/published/2/{token}`
|
|
||||||
- Convert webcal:// to https://
|
|
||||||
- Token embedded in URL path
|
|
||||||
- Public calendars only
|
|
||||||
|
|
||||||
### Nextcloud/ownCloud
|
|
||||||
- URL Format: `https://cloud.example.com/remote.php/dav/public-calendars/{token}`
|
|
||||||
- Token embedded in URL path
|
|
||||||
- Configurable update frequency
|
|
||||||
|
|
||||||
## Migration from CalDAV
|
|
||||||
|
|
||||||
If migrating from an existing CalDAV implementation:
|
|
||||||
|
|
||||||
1. **Database Migration**: Rename fields from `caldav_*` to `ics_*`
|
|
||||||
2. **URL Conversion**: Most CalDAV servers provide ICS export endpoints
|
|
||||||
3. **Authentication**: Convert from username/password to URL-embedded tokens
|
|
||||||
4. **Remove Dependencies**: Uninstall caldav library, add icalendar
|
|
||||||
5. **Update Background Tasks**: Replace CalDAV sync with ICS fetch
|
|
||||||
|
|
||||||
## Performance Optimizations
|
|
||||||
|
|
||||||
1. **Caching**: Use ETag/Last-Modified headers to avoid refetching unchanged calendars
|
|
||||||
2. **Incremental Sync**: Store last sync timestamp, only process new/modified events
|
|
||||||
3. **Batch Processing**: Process multiple room calendars in parallel
|
|
||||||
4. **Connection Pooling**: Reuse HTTP connections for multiple requests
|
|
||||||
5. **Compression**: Support gzip encoding for large ICS files
|
|
||||||
|
|
||||||
## Monitoring and Debugging
|
|
||||||
|
|
||||||
### Metrics to Track
|
|
||||||
- Sync success/failure rate per room
|
|
||||||
- Average sync duration
|
|
||||||
- ICS file sizes
|
|
||||||
- Number of events processed
|
|
||||||
- Failed event matches
|
|
||||||
|
|
||||||
### Debug Logging
|
|
||||||
```python
|
|
||||||
logger.debug(f"Fetching ICS from {room.ics_url}")
|
|
||||||
logger.debug(f"ICS content size: {len(ics_content)} bytes")
|
|
||||||
logger.debug(f"Found {len(events)} matching events")
|
|
||||||
logger.debug(f"Event UIDs: {[e['external_id'] for e in events]}")
|
|
||||||
```
|
|
||||||
|
|
||||||
### Common Issues
|
|
||||||
1. **SSL Certificate Errors**: Add certificate validation options
|
|
||||||
2. **Timeout Issues**: Increase timeout for large calendars
|
|
||||||
3. **Encoding Problems**: Handle various character encodings
|
|
||||||
4. **Timezone Mismatches**: Always convert to UTC
|
|
||||||
5. **Memory Issues**: Stream large ICS files instead of loading entirely
|
|
||||||
337
PLAN.md
337
PLAN.md
@@ -1,337 +0,0 @@
|
|||||||
# ICS Calendar Integration Plan
|
|
||||||
|
|
||||||
## Core Concept
|
|
||||||
ICS calendar URLs are attached to rooms (not users) to enable automatic meeting tracking and management through periodic fetching of calendar data.
|
|
||||||
|
|
||||||
## Database Schema Updates
|
|
||||||
|
|
||||||
### 1. Add ICS configuration to rooms
|
|
||||||
- Add `ics_url` field to room table (URL to .ics file, may include auth token)
|
|
||||||
- Add `ics_fetch_interval` field to room table (default: 5 minutes, configurable)
|
|
||||||
- Add `ics_enabled` boolean field to room table
|
|
||||||
- Add `ics_last_sync` timestamp field to room table
|
|
||||||
|
|
||||||
### 2. Create calendar_events table
|
|
||||||
- `id` - UUID primary key
|
|
||||||
- `room_id` - Foreign key to room
|
|
||||||
- `external_id` - ICS event UID
|
|
||||||
- `title` - Event title
|
|
||||||
- `description` - Event description
|
|
||||||
- `start_time` - Event start timestamp
|
|
||||||
- `end_time` - Event end timestamp
|
|
||||||
- `attendees` - JSON field with attendee list and status
|
|
||||||
- `location` - Meeting location (should contain room name)
|
|
||||||
- `last_synced` - Last sync timestamp
|
|
||||||
- `is_deleted` - Boolean flag for soft delete (preserve past events)
|
|
||||||
- `ics_raw_data` - TEXT field to store raw VEVENT data for reference
|
|
||||||
|
|
||||||
### 3. Update meeting table
|
|
||||||
- Add `calendar_event_id` - Foreign key to calendar_events
|
|
||||||
- Add `calendar_metadata` - JSON field for additional calendar data
|
|
||||||
- Remove unique constraint on room_id + active status (allow multiple active meetings per room)
|
|
||||||
|
|
||||||
## Backend Implementation
|
|
||||||
|
|
||||||
### 1. ICS Sync Service
|
|
||||||
- Create background task that runs based on room's `ics_fetch_interval` (default: 5 minutes)
|
|
||||||
- For each room with ICS enabled, fetch the .ics file via HTTP/HTTPS
|
|
||||||
- Parse ICS file using icalendar library
|
|
||||||
- Extract VEVENT components and filter events looking for room URL (e.g., "https://reflector.monadical.com/max")
|
|
||||||
- Store matching events in calendar_events table
|
|
||||||
- Mark events as "upcoming" if start_time is within next 30 minutes
|
|
||||||
- Pre-create Whereby meetings 1 minute before start (ensures no delay when users join)
|
|
||||||
- Soft-delete future events that were removed from calendar (set is_deleted=true)
|
|
||||||
- Never delete past events (preserve for historical record)
|
|
||||||
- Support authenticated ICS feeds via tokens embedded in URL
|
|
||||||
|
|
||||||
### 2. Meeting Management Updates
|
|
||||||
- Allow multiple active meetings per room
|
|
||||||
- Pre-create meeting record 1 minute before calendar event starts (ensures meeting is ready)
|
|
||||||
- Link meeting to calendar_event for metadata
|
|
||||||
- Keep meeting active for 15 minutes after last participant leaves (grace period)
|
|
||||||
- Don't auto-close if new participant joins within grace period
|
|
||||||
|
|
||||||
### 3. API Endpoints
|
|
||||||
- `GET /v1/rooms/{room_name}/meetings` - List all active and upcoming meetings for a room
|
|
||||||
- Returns filtered data based on user role (owner vs participant)
|
|
||||||
- `GET /v1/rooms/{room_name}/meetings/upcoming` - List upcoming meetings (next 30 min)
|
|
||||||
- Returns filtered data based on user role
|
|
||||||
- `POST /v1/rooms/{room_name}/meetings/{meeting_id}/join` - Join specific meeting
|
|
||||||
- `PATCH /v1/rooms/{room_id}` - Update room settings (including ICS configuration)
|
|
||||||
- ICS fields only visible/editable by room owner
|
|
||||||
- `POST /v1/rooms/{room_name}/ics/sync` - Trigger manual ICS sync
|
|
||||||
- Only accessible by room owner
|
|
||||||
- `GET /v1/rooms/{room_name}/ics/status` - Get ICS sync status and last fetch time
|
|
||||||
- Only accessible by room owner
|
|
||||||
|
|
||||||
## Frontend Implementation
|
|
||||||
|
|
||||||
### 1. Room Settings Page
|
|
||||||
- Add ICS configuration section
|
|
||||||
- Field for ICS URL (e.g., Google Calendar private URL, Outlook ICS export)
|
|
||||||
- Field for fetch interval (dropdown: 1 min, 5 min, 10 min, 30 min, 1 hour)
|
|
||||||
- Test connection button (validates ICS file can be fetched and parsed)
|
|
||||||
- Manual sync button
|
|
||||||
- Show last sync time and next scheduled sync
|
|
||||||
|
|
||||||
### 2. Meeting Selection Page (New)
|
|
||||||
- Show when accessing `/room/{room_name}`
|
|
||||||
- **Host view** (room owner):
|
|
||||||
- Full calendar event details
|
|
||||||
- Meeting title and description
|
|
||||||
- Complete attendee list with RSVP status
|
|
||||||
- Number of current participants
|
|
||||||
- Duration (how long it's been running)
|
|
||||||
- **Participant view** (non-owners):
|
|
||||||
- Meeting title only
|
|
||||||
- Date and time
|
|
||||||
- Number of current participants
|
|
||||||
- Duration (how long it's been running)
|
|
||||||
- No attendee list or description (privacy)
|
|
||||||
- Display upcoming meetings (visible 30min before):
|
|
||||||
- Show countdown to start
|
|
||||||
- Can click to join early → redirected to waiting page
|
|
||||||
- Waiting page shows countdown until meeting starts
|
|
||||||
- Meeting pre-created by background task (ready when users arrive)
|
|
||||||
- Option to create unscheduled meeting (uses existing flow)
|
|
||||||
|
|
||||||
### 3. Meeting Room Updates
|
|
||||||
- Show calendar metadata in meeting info
|
|
||||||
- Display invited attendees vs actual participants
|
|
||||||
- Show meeting title from calendar event
|
|
||||||
|
|
||||||
## Meeting Lifecycle
|
|
||||||
|
|
||||||
### 1. Meeting Creation
|
|
||||||
- Automatic: Pre-created 1 minute before calendar event starts (ensures Whereby room is ready)
|
|
||||||
- Manual: User creates unscheduled meeting (existing `/rooms/{room_name}/meeting` endpoint)
|
|
||||||
- Background task handles pre-creation to avoid delays when users join
|
|
||||||
|
|
||||||
### 2. Meeting Join Rules
|
|
||||||
- Can join active meetings immediately
|
|
||||||
- Can see upcoming meetings 30 minutes before start
|
|
||||||
- Can click to join upcoming meetings early → sent to waiting page
|
|
||||||
- Waiting page automatically transitions to meeting at scheduled time
|
|
||||||
- Unscheduled meetings always joinable (current behavior)
|
|
||||||
|
|
||||||
### 3. Meeting Closure Rules
|
|
||||||
- All meetings: 15-minute grace period after last participant leaves
|
|
||||||
- If participant rejoins within grace period, keep meeting active
|
|
||||||
- Calendar meetings: Force close 30 minutes after scheduled end time
|
|
||||||
- Unscheduled meetings: Keep active for 8 hours (current behavior)
|
|
||||||
|
|
||||||
## ICS Parsing Logic
|
|
||||||
|
|
||||||
### 1. Event Matching
|
|
||||||
- Parse ICS file using Python icalendar library
|
|
||||||
- Iterate through VEVENT components
|
|
||||||
- Check LOCATION field for full FQDN URL (e.g., "https://reflector.monadical.com/max")
|
|
||||||
- Check DESCRIPTION for room URL or mention
|
|
||||||
- Support multiple formats:
|
|
||||||
- Full URL: "https://reflector.monadical.com/max"
|
|
||||||
- With /room path: "https://reflector.monadical.com/room/max"
|
|
||||||
- Partial paths: "room/max", "/max room"
|
|
||||||
|
|
||||||
### 2. Attendee Extraction
|
|
||||||
- Parse ATTENDEE properties from VEVENT
|
|
||||||
- Extract email (MAILTO), name (CN parameter), and RSVP status (PARTSTAT)
|
|
||||||
- Store as JSON in calendar_events.attendees
|
|
||||||
|
|
||||||
### 3. Sync Strategy
|
|
||||||
- Fetch complete ICS file (contains all events)
|
|
||||||
- Filter events from (now - 1 hour) to (now + 24 hours) for processing
|
|
||||||
- Update existing events if LAST-MODIFIED or SEQUENCE changed
|
|
||||||
- Delete future events that no longer exist in ICS (start_time > now)
|
|
||||||
- Keep past events for historical record (never delete if start_time < now)
|
|
||||||
- Handle recurring events (RRULE) - expand to individual instances
|
|
||||||
- Track deleted calendar events to clean up future meetings
|
|
||||||
- Cache ICS file hash to detect changes and skip unnecessary processing
|
|
||||||
|
|
||||||
## Security Considerations
|
|
||||||
|
|
||||||
### 1. ICS URL Security
|
|
||||||
- ICS URLs may contain authentication tokens (e.g., Google Calendar private URLs)
|
|
||||||
- Store full ICS URLs encrypted using Fernet to protect embedded tokens
|
|
||||||
- Validate ICS URLs (must be HTTPS for production)
|
|
||||||
- Never expose full ICS URLs in API responses (return masked version)
|
|
||||||
- Rate limit ICS fetching to prevent abuse
|
|
||||||
|
|
||||||
### 2. Room Access
|
|
||||||
- Only room owner can configure ICS URL
|
|
||||||
- ICS URL shown as masked version to room owner (hides embedded tokens)
|
|
||||||
- ICS settings not visible to other users
|
|
||||||
- Meeting list visible to all room participants
|
|
||||||
- ICS fetch logs only visible to room owner
|
|
||||||
|
|
||||||
### 3. Meeting Privacy
|
|
||||||
- Full calendar details visible only to room owner
|
|
||||||
- Participants see limited info: title, date/time only
|
|
||||||
- Attendee list and description hidden from non-owners
|
|
||||||
- Meeting titles visible in room listing to all
|
|
||||||
|
|
||||||
## Implementation Phases
|
|
||||||
|
|
||||||
### Phase 1: Database and ICS Setup (Week 1) ✅ COMPLETED (2025-08-18)
|
|
||||||
1. ✅ Created database migrations for ICS fields and calendar_events table
|
|
||||||
- Added ics_url, ics_fetch_interval, ics_enabled, ics_last_sync, ics_last_etag to room table
|
|
||||||
- Created calendar_event table with ics_uid (instead of external_id) and proper typing
|
|
||||||
- Added calendar_event_id and calendar_metadata (JSONB) to meeting table
|
|
||||||
- Removed server_default from datetime fields for consistency
|
|
||||||
2. ✅ Installed icalendar Python library for ICS parsing
|
|
||||||
- Added icalendar>=6.0.0 to dependencies
|
|
||||||
- No encryption needed - ICS URLs are read-only
|
|
||||||
3. ✅ Built ICS fetch and sync service
|
|
||||||
- Simple HTTP fetching without unnecessary validation
|
|
||||||
- Proper TypedDict typing for event data structures
|
|
||||||
- Supports any standard ICS format
|
|
||||||
- Event matching on full room URL only
|
|
||||||
4. ✅ API endpoints for ICS configuration
|
|
||||||
- Room model updated to support ICS fields via existing PATCH endpoint
|
|
||||||
- POST /v1/rooms/{room_name}/ics/sync - Trigger manual sync (owner only)
|
|
||||||
- GET /v1/rooms/{room_name}/ics/status - Get sync status (owner only)
|
|
||||||
- GET /v1/rooms/{room_name}/meetings - List meetings with privacy controls
|
|
||||||
- GET /v1/rooms/{room_name}/meetings/upcoming - List upcoming meetings
|
|
||||||
5. ✅ Celery background tasks for periodic sync
|
|
||||||
- sync_room_ics - Sync individual room calendar
|
|
||||||
- sync_all_ics_calendars - Check all rooms and queue sync based on fetch intervals
|
|
||||||
- pre_create_upcoming_meetings - Pre-create Whereby meetings 1 minute before start
|
|
||||||
- Tasks scheduled in beat schedule (every minute for checking, respects individual intervals)
|
|
||||||
6. ✅ Tests written and passing
|
|
||||||
- 6 tests for Room ICS fields
|
|
||||||
- 7 tests for CalendarEvent model
|
|
||||||
- 7 tests for ICS sync service
|
|
||||||
- 11 tests for API endpoints
|
|
||||||
- 6 tests for background tasks
|
|
||||||
- All 31 ICS-related tests passing
|
|
||||||
|
|
||||||
### Phase 2: Meeting Management (Week 2) ✅ COMPLETED (2025-08-19)
|
|
||||||
1. ✅ Updated meeting lifecycle logic with grace period support
|
|
||||||
- 15-minute grace period after last participant leaves
|
|
||||||
- Automatic reactivation when participants rejoin
|
|
||||||
- Force close calendar meetings 30 minutes after scheduled end
|
|
||||||
2. ✅ Support multiple active meetings per room
|
|
||||||
- Removed unique constraint on active meetings
|
|
||||||
- Added get_all_active_for_room() method
|
|
||||||
- Added get_active_by_calendar_event() method
|
|
||||||
3. ✅ Implemented grace period logic
|
|
||||||
- Added last_participant_left_at and grace_period_minutes fields
|
|
||||||
- Process meetings task handles grace period checking
|
|
||||||
- Whereby webhooks clear grace period on participant join
|
|
||||||
4. ✅ Link meetings to calendar events
|
|
||||||
- Pre-created meetings properly linked via calendar_event_id
|
|
||||||
- Calendar metadata stored with meeting
|
|
||||||
- API endpoints for listing and joining specific meetings
|
|
||||||
|
|
||||||
### Phase 3: Frontend Meeting Selection (Week 3)
|
|
||||||
1. Build meeting selection page
|
|
||||||
2. Show active and upcoming meetings
|
|
||||||
3. Implement waiting page for early joiners
|
|
||||||
4. Add automatic transition from waiting to meeting
|
|
||||||
5. Support unscheduled meeting creation
|
|
||||||
|
|
||||||
### Phase 4: Calendar Integration UI (Week 4)
|
|
||||||
1. Add ICS settings to room configuration
|
|
||||||
2. Display calendar metadata in meetings
|
|
||||||
3. Show attendee information
|
|
||||||
4. Add sync status indicators
|
|
||||||
5. Show fetch interval and next sync time
|
|
||||||
|
|
||||||
## Success Metrics
|
|
||||||
- Zero merged meetings from consecutive calendar events
|
|
||||||
- Successful ICS sync from major providers (Google Calendar, Outlook, Apple Calendar, Nextcloud)
|
|
||||||
- Meeting join accuracy: correct meeting 100% of the time
|
|
||||||
- Grace period prevents 90% of accidental meeting closures
|
|
||||||
- Configurable fetch intervals reduce unnecessary API calls
|
|
||||||
|
|
||||||
## Design Decisions
|
|
||||||
1. **ICS attached to room, not user** - Prevents duplicate meetings from multiple calendars
|
|
||||||
2. **Multiple active meetings per room** - Supported with meeting selection page
|
|
||||||
3. **Grace period for rejoining** - 15 minutes after last participant leaves
|
|
||||||
4. **Upcoming meeting visibility** - Show 30 minutes before, join only on time
|
|
||||||
5. **Calendar data storage** - Attached to meeting record for full context
|
|
||||||
6. **No "ad-hoc" meetings** - Use existing meeting creation flow (unscheduled meetings)
|
|
||||||
7. **ICS configuration via room PATCH** - Reuse existing room configuration endpoint
|
|
||||||
8. **Event deletion handling** - Soft-delete future events, preserve past meetings
|
|
||||||
9. **Configurable fetch interval** - Balance between freshness and server load
|
|
||||||
10. **ICS over CalDAV** - Simpler implementation, wider compatibility, no complex auth
|
|
||||||
|
|
||||||
## Phase 2 Implementation Files
|
|
||||||
|
|
||||||
### Database Migrations
|
|
||||||
- `/server/migrations/versions/6025e9b2bef2_remove_one_active_meeting_per_room_.py` - Remove unique constraint
|
|
||||||
- `/server/migrations/versions/d4a1c446458c_add_grace_period_fields_to_meeting.py` - Add grace period fields
|
|
||||||
|
|
||||||
### Updated Models
|
|
||||||
- `/server/reflector/db/meetings.py` - Added grace period fields and new query methods
|
|
||||||
|
|
||||||
### Updated Services
|
|
||||||
- `/server/reflector/worker/process.py` - Enhanced with grace period logic and multiple meeting support
|
|
||||||
|
|
||||||
### Updated API
|
|
||||||
- `/server/reflector/views/rooms.py` - Added endpoints for listing active meetings and joining specific meetings
|
|
||||||
- `/server/reflector/views/whereby.py` - Clear grace period on participant join
|
|
||||||
|
|
||||||
### Tests
|
|
||||||
- `/server/tests/test_multiple_active_meetings.py` - Comprehensive tests for Phase 2 features (5 tests)
|
|
||||||
|
|
||||||
## Phase 1 Implementation Files Created
|
|
||||||
|
|
||||||
### Database Models
|
|
||||||
- `/server/reflector/db/rooms.py` - Updated with ICS fields (url, fetch_interval, enabled, last_sync, etag)
|
|
||||||
- `/server/reflector/db/calendar_events.py` - New CalendarEvent model with ics_uid and proper typing
|
|
||||||
- `/server/reflector/db/meetings.py` - Updated with calendar_event_id and calendar_metadata (JSONB)
|
|
||||||
|
|
||||||
### Services
|
|
||||||
- `/server/reflector/services/ics_sync.py` - ICS fetching and parsing with TypedDict for proper typing
|
|
||||||
|
|
||||||
### API Endpoints
|
|
||||||
- `/server/reflector/views/rooms.py` - Added ICS management endpoints with privacy controls
|
|
||||||
|
|
||||||
### Background Tasks
|
|
||||||
- `/server/reflector/worker/ics_sync.py` - Celery tasks for automatic periodic sync
|
|
||||||
- `/server/reflector/worker/app.py` - Updated beat schedule for ICS tasks
|
|
||||||
|
|
||||||
### Tests
|
|
||||||
- `/server/tests/test_room_ics.py` - Room model ICS fields tests (6 tests)
|
|
||||||
- `/server/tests/test_calendar_event.py` - CalendarEvent model tests (7 tests)
|
|
||||||
- `/server/tests/test_ics_sync.py` - ICS sync service tests (7 tests)
|
|
||||||
- `/server/tests/test_room_ics_api.py` - API endpoint tests (11 tests)
|
|
||||||
- `/server/tests/test_ics_background_tasks.py` - Background task tests (6 tests)
|
|
||||||
|
|
||||||
### Key Design Decisions
|
|
||||||
- No encryption needed - ICS URLs are read-only access
|
|
||||||
- Using ics_uid instead of external_id for clarity
|
|
||||||
- Proper TypedDict typing for event data structures
|
|
||||||
- Removed unnecessary URL validation and webcal handling
|
|
||||||
- calendar_metadata in meetings stores flexible calendar data (organizer, recurrence, etc)
|
|
||||||
- Background tasks query all rooms directly to avoid filtering issues
|
|
||||||
- Sync intervals respected per-room configuration
|
|
||||||
|
|
||||||
## Implementation Approach
|
|
||||||
|
|
||||||
### ICS Fetching vs CalDAV
|
|
||||||
- **ICS Benefits**:
|
|
||||||
- Simpler implementation (HTTP GET vs CalDAV protocol)
|
|
||||||
- Wider compatibility (all calendar apps can export ICS)
|
|
||||||
- No authentication complexity (simple URL with optional token)
|
|
||||||
- Easier debugging (ICS is plain text)
|
|
||||||
- Lower server requirements (no CalDAV library dependencies)
|
|
||||||
|
|
||||||
### Supported Calendar Providers
|
|
||||||
1. **Google Calendar**: Private ICS URL from calendar settings
|
|
||||||
2. **Outlook/Office 365**: ICS export URL from calendar sharing
|
|
||||||
3. **Apple Calendar**: Published calendar ICS URL
|
|
||||||
4. **Nextcloud**: Public/private calendar ICS export
|
|
||||||
5. **Any CalDAV server**: Via ICS export endpoint
|
|
||||||
|
|
||||||
### ICS URL Examples
|
|
||||||
- Google: `https://calendar.google.com/calendar/ical/{calendar_id}/private-{token}/basic.ics`
|
|
||||||
- Outlook: `https://outlook.live.com/owa/calendar/{id}/calendar.ics`
|
|
||||||
- Custom: `https://example.com/calendars/room-schedule.ics`
|
|
||||||
|
|
||||||
### Fetch Interval Configuration
|
|
||||||
- 1 minute: For critical/high-activity rooms
|
|
||||||
- 5 minutes (default): Balance of freshness and efficiency
|
|
||||||
- 10 minutes: Standard meeting rooms
|
|
||||||
- 30 minutes: Low-activity rooms
|
|
||||||
- 1 hour: Rarely-used rooms or stable schedules
|
|
||||||
88
README.md
88
README.md
@@ -1,43 +1,60 @@
|
|||||||
<div align="center">
|
<div align="center">
|
||||||
|
<img width="100" alt="image" src="https://github.com/user-attachments/assets/66fb367b-2c89-4516-9912-f47ac59c6a7f"/>
|
||||||
|
|
||||||
# Reflector
|
# Reflector
|
||||||
|
|
||||||
Reflector Audio Management and Analysis is a cutting-edge web application under development by Monadical. It utilizes AI to record meetings, providing a permanent record with transcripts, translations, and automated summaries.
|
Reflector is an AI-powered audio transcription and meeting analysis platform that provides real-time transcription, speaker diarization, translation and summarization for audio content and live meetings. It works 100% with local models (whisper/parakeet, pyannote, seamless-m4t, and your local llm like phi-4).
|
||||||
|
|
||||||
[](https://github.com/monadical-sas/reflector/actions/workflows/pytests.yml)
|
[](https://github.com/monadical-sas/reflector/actions/workflows/test_server.yml)
|
||||||
[](https://opensource.org/licenses/MIT)
|
[](https://opensource.org/licenses/MIT)
|
||||||
</div>
|
</div>
|
||||||
|
</div>
|
||||||
## Screenshots
|
|
||||||
<table>
|
<table>
|
||||||
<tr>
|
<tr>
|
||||||
<td>
|
<td>
|
||||||
<a href="https://github.com/user-attachments/assets/3a976930-56c1-47ef-8c76-55d3864309e3">
|
<a href="https://github.com/user-attachments/assets/21f5597c-2930-4899-a154-f7bd61a59e97">
|
||||||
<img width="700" alt="image" src="https://github.com/user-attachments/assets/3a976930-56c1-47ef-8c76-55d3864309e3" />
|
<img width="700" alt="image" src="https://github.com/user-attachments/assets/21f5597c-2930-4899-a154-f7bd61a59e97" />
|
||||||
</a>
|
</a>
|
||||||
</td>
|
</td>
|
||||||
<td>
|
<td>
|
||||||
<a href="https://github.com/user-attachments/assets/bfe3bde3-08af-4426-a9a1-11ad5cd63b33">
|
<a href="https://github.com/user-attachments/assets/f6b9399a-5e51-4bae-b807-59128d0a940c">
|
||||||
<img width="700" alt="image" src="https://github.com/user-attachments/assets/bfe3bde3-08af-4426-a9a1-11ad5cd63b33" />
|
<img width="700" alt="image" src="https://github.com/user-attachments/assets/f6b9399a-5e51-4bae-b807-59128d0a940c" />
|
||||||
</a>
|
</a>
|
||||||
</td>
|
</td>
|
||||||
<td>
|
<td>
|
||||||
<a href="https://github.com/user-attachments/assets/7b60c9d0-efe4-474f-a27b-ea13bd0fabdc">
|
<a href="https://github.com/user-attachments/assets/a42ce460-c1fd-4489-a995-270516193897">
|
||||||
<img width="700" alt="image" src="https://github.com/user-attachments/assets/7b60c9d0-efe4-474f-a27b-ea13bd0fabdc" />
|
<img width="700" alt="image" src="https://github.com/user-attachments/assets/a42ce460-c1fd-4489-a995-270516193897" />
|
||||||
|
</a>
|
||||||
|
</td>
|
||||||
|
<td>
|
||||||
|
<a href="https://github.com/user-attachments/assets/21929f6d-c309-42fe-9c11-f1299e50fbd4">
|
||||||
|
<img width="700" alt="image" src="https://github.com/user-attachments/assets/21929f6d-c309-42fe-9c11-f1299e50fbd4" />
|
||||||
</a>
|
</a>
|
||||||
</td>
|
</td>
|
||||||
</tr>
|
</tr>
|
||||||
</table>
|
</table>
|
||||||
|
|
||||||
|
## What is Reflector?
|
||||||
|
|
||||||
|
Reflector is a web application that utilizes local models to process audio content, providing:
|
||||||
|
|
||||||
|
- **Real-time Transcription**: Convert speech to text using [Whisper](https://github.com/openai/whisper) (multi-language) or [Parakeet](https://huggingface.co/nvidia/parakeet-tdt-0.6b-v2) (English) models
|
||||||
|
- **Speaker Diarization**: Identify and label different speakers using [Pyannote](https://github.com/pyannote/pyannote-audio) 3.1
|
||||||
|
- **Live Translation**: Translate audio content in real-time to many languages with [Facebook Seamless-M4T](https://github.com/facebookresearch/seamless_communication)
|
||||||
|
- **Topic Detection & Summarization**: Extract key topics and generate concise summaries using LLMs
|
||||||
|
- **Meeting Recording**: Create permanent records of meetings with searchable transcripts
|
||||||
|
|
||||||
|
Currently we provide [modal.com](https://modal.com/) gpu template to deploy.
|
||||||
|
|
||||||
## Background
|
## Background
|
||||||
|
|
||||||
The project architecture consists of three primary components:
|
The project architecture consists of three primary components:
|
||||||
|
|
||||||
- **Front-End**: NextJS React project hosted on Vercel, located in `www/`.
|
|
||||||
- **Back-End**: Python server that offers an API and data persistence, found in `server/`.
|
- **Back-End**: Python server that offers an API and data persistence, found in `server/`.
|
||||||
- **GPU implementation**: Providing services such as speech-to-text transcription, topic generation, automated summaries, and translations. Most reliable option is Modal deployment
|
- **Front-End**: NextJS React project hosted on Vercel, located in `www/`.
|
||||||
|
- **GPU implementation**: Providing services such as speech-to-text transcription, topic generation, automated summaries, and translations.
|
||||||
|
|
||||||
It also uses authentik for authentication if activated, and Vercel for deployment and configuration of the front-end.
|
It also uses authentik for authentication if activated.
|
||||||
|
|
||||||
## Contribution Guidelines
|
## Contribution Guidelines
|
||||||
|
|
||||||
@@ -72,6 +89,8 @@ Note: We currently do not have instructions for Windows users.
|
|||||||
|
|
||||||
## Installation
|
## Installation
|
||||||
|
|
||||||
|
*Note: we're working toward better installation, theses instructions are not accurate for now*
|
||||||
|
|
||||||
### Frontend
|
### Frontend
|
||||||
|
|
||||||
Start with `cd www`.
|
Start with `cd www`.
|
||||||
@@ -80,11 +99,10 @@ Start with `cd www`.
|
|||||||
|
|
||||||
```bash
|
```bash
|
||||||
pnpm install
|
pnpm install
|
||||||
cp .env_template .env
|
cp .env.example .env
|
||||||
cp config-template.ts config.ts
|
|
||||||
```
|
```
|
||||||
|
|
||||||
Then, fill in the environment variables in `.env` and the configuration in `config.ts` as needed. If you are unsure on how to proceed, ask in Zulip.
|
Then, fill in the environment variables in `.env` as needed. If you are unsure on how to proceed, ask in Zulip.
|
||||||
|
|
||||||
**Run in development mode**
|
**Run in development mode**
|
||||||
|
|
||||||
@@ -149,3 +167,41 @@ You can manually process an audio file by calling the process tool:
|
|||||||
```bash
|
```bash
|
||||||
uv run python -m reflector.tools.process path/to/audio.wav
|
uv run python -m reflector.tools.process path/to/audio.wav
|
||||||
```
|
```
|
||||||
|
|
||||||
|
## Build-time env variables
|
||||||
|
|
||||||
|
Next.js projects are more used to NEXT_PUBLIC_ prefixed buildtime vars. We don't have those for the reason we need to serve a ccustomizable prebuild docker container.
|
||||||
|
|
||||||
|
Instead, all the variables are runtime. Variables needed to the frontend are served to the frontend app at initial render.
|
||||||
|
|
||||||
|
It also means there's no static prebuild and no static files to serve for js/html.
|
||||||
|
|
||||||
|
## Feature Flags
|
||||||
|
|
||||||
|
Reflector uses environment variable-based feature flags to control application functionality. These flags allow you to enable or disable features without code changes.
|
||||||
|
|
||||||
|
### Available Feature Flags
|
||||||
|
|
||||||
|
| Feature Flag | Environment Variable |
|
||||||
|
|-------------|---------------------|
|
||||||
|
| `requireLogin` | `FEATURE_REQUIRE_LOGIN` |
|
||||||
|
| `privacy` | `FEATURE_PRIVACY` |
|
||||||
|
| `browse` | `FEATURE_BROWSE` |
|
||||||
|
| `sendToZulip` | `FEATURE_SEND_TO_ZULIP` |
|
||||||
|
| `rooms` | `FEATURE_ROOMS` |
|
||||||
|
|
||||||
|
### Setting Feature Flags
|
||||||
|
|
||||||
|
Feature flags are controlled via environment variables using the pattern `FEATURE_{FEATURE_NAME}` where `{FEATURE_NAME}` is the SCREAMING_SNAKE_CASE version of the feature name.
|
||||||
|
|
||||||
|
**Examples:**
|
||||||
|
```bash
|
||||||
|
# Enable user authentication requirement
|
||||||
|
FEATURE_REQUIRE_LOGIN=true
|
||||||
|
|
||||||
|
# Disable browse functionality
|
||||||
|
FEATURE_BROWSE=false
|
||||||
|
|
||||||
|
# Enable Zulip integration
|
||||||
|
FEATURE_SEND_TO_ZULIP=true
|
||||||
|
```
|
||||||
|
|||||||
39
docker-compose.prod.yml
Normal file
39
docker-compose.prod.yml
Normal file
@@ -0,0 +1,39 @@
|
|||||||
|
# Production Docker Compose configuration for Frontend
|
||||||
|
# Usage: docker compose -f docker-compose.prod.yml up -d
|
||||||
|
|
||||||
|
services:
|
||||||
|
web:
|
||||||
|
build:
|
||||||
|
context: ./www
|
||||||
|
dockerfile: Dockerfile
|
||||||
|
image: reflector-frontend:latest
|
||||||
|
environment:
|
||||||
|
- KV_URL=${KV_URL:-redis://redis:6379}
|
||||||
|
- SITE_URL=${SITE_URL}
|
||||||
|
- API_URL=${API_URL}
|
||||||
|
- WEBSOCKET_URL=${WEBSOCKET_URL}
|
||||||
|
- NEXTAUTH_URL=${NEXTAUTH_URL:-http://localhost:3000}
|
||||||
|
- NEXTAUTH_SECRET=${NEXTAUTH_SECRET:-changeme-in-production}
|
||||||
|
- AUTHENTIK_ISSUER=${AUTHENTIK_ISSUER}
|
||||||
|
- AUTHENTIK_CLIENT_ID=${AUTHENTIK_CLIENT_ID}
|
||||||
|
- AUTHENTIK_CLIENT_SECRET=${AUTHENTIK_CLIENT_SECRET}
|
||||||
|
- AUTHENTIK_REFRESH_TOKEN_URL=${AUTHENTIK_REFRESH_TOKEN_URL}
|
||||||
|
- SENTRY_DSN=${SENTRY_DSN}
|
||||||
|
- SENTRY_IGNORE_API_RESOLUTION_ERROR=${SENTRY_IGNORE_API_RESOLUTION_ERROR:-1}
|
||||||
|
depends_on:
|
||||||
|
- redis
|
||||||
|
restart: unless-stopped
|
||||||
|
|
||||||
|
redis:
|
||||||
|
image: redis:7.2-alpine
|
||||||
|
restart: unless-stopped
|
||||||
|
healthcheck:
|
||||||
|
test: ["CMD", "redis-cli", "ping"]
|
||||||
|
interval: 30s
|
||||||
|
timeout: 3s
|
||||||
|
retries: 3
|
||||||
|
volumes:
|
||||||
|
- redis_data:/data
|
||||||
|
|
||||||
|
volumes:
|
||||||
|
redis_data:
|
||||||
@@ -6,6 +6,7 @@ services:
|
|||||||
- 1250:1250
|
- 1250:1250
|
||||||
volumes:
|
volumes:
|
||||||
- ./server/:/app/
|
- ./server/:/app/
|
||||||
|
- /app/.venv
|
||||||
env_file:
|
env_file:
|
||||||
- ./server/.env
|
- ./server/.env
|
||||||
environment:
|
environment:
|
||||||
@@ -16,6 +17,7 @@ services:
|
|||||||
context: server
|
context: server
|
||||||
volumes:
|
volumes:
|
||||||
- ./server/:/app/
|
- ./server/:/app/
|
||||||
|
- /app/.venv
|
||||||
env_file:
|
env_file:
|
||||||
- ./server/.env
|
- ./server/.env
|
||||||
environment:
|
environment:
|
||||||
@@ -26,6 +28,7 @@ services:
|
|||||||
context: server
|
context: server
|
||||||
volumes:
|
volumes:
|
||||||
- ./server/:/app/
|
- ./server/:/app/
|
||||||
|
- /app/.venv
|
||||||
env_file:
|
env_file:
|
||||||
- ./server/.env
|
- ./server/.env
|
||||||
environment:
|
environment:
|
||||||
@@ -36,7 +39,7 @@ services:
|
|||||||
ports:
|
ports:
|
||||||
- 6379:6379
|
- 6379:6379
|
||||||
web:
|
web:
|
||||||
image: node:18
|
image: node:22-alpine
|
||||||
ports:
|
ports:
|
||||||
- "3000:3000"
|
- "3000:3000"
|
||||||
command: sh -c "corepack enable && pnpm install && pnpm dev"
|
command: sh -c "corepack enable && pnpm install && pnpm dev"
|
||||||
@@ -47,6 +50,8 @@ services:
|
|||||||
- /app/node_modules
|
- /app/node_modules
|
||||||
env_file:
|
env_file:
|
||||||
- ./www/.env.local
|
- ./www/.env.local
|
||||||
|
environment:
|
||||||
|
- NODE_ENV=development
|
||||||
|
|
||||||
postgres:
|
postgres:
|
||||||
image: postgres:17
|
image: postgres:17
|
||||||
33
gpu/modal_deployments/.gitignore
vendored
Normal file
33
gpu/modal_deployments/.gitignore
vendored
Normal file
@@ -0,0 +1,33 @@
|
|||||||
|
# OS / Editor
|
||||||
|
.DS_Store
|
||||||
|
.vscode/
|
||||||
|
.idea/
|
||||||
|
|
||||||
|
# Python
|
||||||
|
__pycache__/
|
||||||
|
*.py[cod]
|
||||||
|
*$py.class
|
||||||
|
|
||||||
|
# Logs
|
||||||
|
*.log
|
||||||
|
|
||||||
|
# Env and secrets
|
||||||
|
.env
|
||||||
|
.env.*
|
||||||
|
*.env
|
||||||
|
*.secret
|
||||||
|
|
||||||
|
# Build / dist
|
||||||
|
build/
|
||||||
|
dist/
|
||||||
|
.eggs/
|
||||||
|
*.egg-info/
|
||||||
|
|
||||||
|
# Coverage / test
|
||||||
|
.pytest_cache/
|
||||||
|
.coverage*
|
||||||
|
htmlcov/
|
||||||
|
|
||||||
|
# Modal local state (if any)
|
||||||
|
modal_mounts/
|
||||||
|
.modal_cache/
|
||||||
171
gpu/modal_deployments/README.md
Normal file
171
gpu/modal_deployments/README.md
Normal file
@@ -0,0 +1,171 @@
|
|||||||
|
# Reflector GPU implementation - Transcription and LLM
|
||||||
|
|
||||||
|
This repository hold an API for the GPU implementation of the Reflector API service,
|
||||||
|
and use [Modal.com](https://modal.com)
|
||||||
|
|
||||||
|
- `reflector_diarizer.py` - Diarization API
|
||||||
|
- `reflector_transcriber.py` - Transcription API (Whisper)
|
||||||
|
- `reflector_transcriber_parakeet.py` - Transcription API (NVIDIA Parakeet)
|
||||||
|
- `reflector_translator.py` - Translation API
|
||||||
|
|
||||||
|
## Modal.com deployment
|
||||||
|
|
||||||
|
Create a modal secret, and name it `reflector-gpu`.
|
||||||
|
It should contain an `REFLECTOR_APIKEY` environment variable with a value.
|
||||||
|
|
||||||
|
The deployment is done using [Modal.com](https://modal.com) service.
|
||||||
|
|
||||||
|
```
|
||||||
|
$ modal deploy reflector_transcriber.py
|
||||||
|
...
|
||||||
|
└── 🔨 Created web => https://xxxx--reflector-transcriber-web.modal.run
|
||||||
|
|
||||||
|
$ modal deploy reflector_transcriber_parakeet.py
|
||||||
|
...
|
||||||
|
└── 🔨 Created web => https://xxxx--reflector-transcriber-parakeet-web.modal.run
|
||||||
|
|
||||||
|
$ modal deploy reflector_llm.py
|
||||||
|
...
|
||||||
|
└── 🔨 Created web => https://xxxx--reflector-llm-web.modal.run
|
||||||
|
```
|
||||||
|
|
||||||
|
Then in your reflector api configuration `.env`, you can set these keys:
|
||||||
|
|
||||||
|
```
|
||||||
|
TRANSCRIPT_BACKEND=modal
|
||||||
|
TRANSCRIPT_URL=https://xxxx--reflector-transcriber-web.modal.run
|
||||||
|
TRANSCRIPT_MODAL_API_KEY=REFLECTOR_APIKEY
|
||||||
|
|
||||||
|
DIARIZATION_BACKEND=modal
|
||||||
|
DIARIZATION_URL=https://xxxx--reflector-diarizer-web.modal.run
|
||||||
|
DIARIZATION_MODAL_API_KEY=REFLECTOR_APIKEY
|
||||||
|
|
||||||
|
TRANSLATION_BACKEND=modal
|
||||||
|
TRANSLATION_URL=https://xxxx--reflector-translator-web.modal.run
|
||||||
|
TRANSLATION_MODAL_API_KEY=REFLECTOR_APIKEY
|
||||||
|
```
|
||||||
|
|
||||||
|
## API
|
||||||
|
|
||||||
|
Authentication must be passed with the `Authorization` header, using the `bearer` scheme.
|
||||||
|
|
||||||
|
```
|
||||||
|
Authorization: bearer <REFLECTOR_APIKEY>
|
||||||
|
```
|
||||||
|
|
||||||
|
### LLM
|
||||||
|
|
||||||
|
`POST /llm`
|
||||||
|
|
||||||
|
**request**
|
||||||
|
```
|
||||||
|
{
|
||||||
|
"prompt": "xxx"
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
**response**
|
||||||
|
```
|
||||||
|
{
|
||||||
|
"text": "xxx completed"
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
### Transcription
|
||||||
|
|
||||||
|
#### Parakeet Transcriber (`reflector_transcriber_parakeet.py`)
|
||||||
|
|
||||||
|
NVIDIA Parakeet is a state-of-the-art ASR model optimized for real-time transcription with superior word-level timestamps.
|
||||||
|
|
||||||
|
**GPU Configuration:**
|
||||||
|
- **A10G GPU** - Used for `/v1/audio/transcriptions` endpoint (small files, live transcription)
|
||||||
|
- Higher concurrency (max_inputs=10)
|
||||||
|
- Optimized for multiple small audio files
|
||||||
|
- Supports batch processing for efficiency
|
||||||
|
|
||||||
|
- **L40S GPU** - Used for `/v1/audio/transcriptions-from-url` endpoint (large files)
|
||||||
|
- Lower concurrency but more powerful processing
|
||||||
|
- Optimized for single large audio files
|
||||||
|
- VAD-based chunking for long-form audio
|
||||||
|
|
||||||
|
##### `/v1/audio/transcriptions` - Small file transcription
|
||||||
|
|
||||||
|
**request** (multipart/form-data)
|
||||||
|
- `file` or `files[]` - audio file(s) to transcribe
|
||||||
|
- `model` - model name (default: `nvidia/parakeet-tdt-0.6b-v2`)
|
||||||
|
- `language` - language code (default: `en`)
|
||||||
|
- `batch` - whether to use batch processing for multiple files (default: `true`)
|
||||||
|
|
||||||
|
**response**
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"text": "transcribed text",
|
||||||
|
"words": [
|
||||||
|
{"word": "hello", "start": 0.0, "end": 0.5},
|
||||||
|
{"word": "world", "start": 0.5, "end": 1.0}
|
||||||
|
],
|
||||||
|
"filename": "audio.mp3"
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
For multiple files with batch=true:
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"results": [
|
||||||
|
{
|
||||||
|
"filename": "audio1.mp3",
|
||||||
|
"text": "transcribed text",
|
||||||
|
"words": [...]
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"filename": "audio2.mp3",
|
||||||
|
"text": "transcribed text",
|
||||||
|
"words": [...]
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
##### `/v1/audio/transcriptions-from-url` - Large file transcription
|
||||||
|
|
||||||
|
**request** (application/json)
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"audio_file_url": "https://example.com/audio.mp3",
|
||||||
|
"model": "nvidia/parakeet-tdt-0.6b-v2",
|
||||||
|
"language": "en",
|
||||||
|
"timestamp_offset": 0.0
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
**response**
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"text": "transcribed text from large file",
|
||||||
|
"words": [
|
||||||
|
{"word": "hello", "start": 0.0, "end": 0.5},
|
||||||
|
{"word": "world", "start": 0.5, "end": 1.0}
|
||||||
|
]
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
**Supported file types:** mp3, mp4, mpeg, mpga, m4a, wav, webm
|
||||||
|
|
||||||
|
#### Whisper Transcriber (`reflector_transcriber.py`)
|
||||||
|
|
||||||
|
`POST /transcribe`
|
||||||
|
|
||||||
|
**request** (multipart/form-data)
|
||||||
|
|
||||||
|
- `file` - audio file
|
||||||
|
- `language` - language code (e.g. `en`)
|
||||||
|
|
||||||
|
**response**
|
||||||
|
```
|
||||||
|
{
|
||||||
|
"text": "xxx",
|
||||||
|
"words": [
|
||||||
|
{"text": "xxx", "start": 0.0, "end": 1.0}
|
||||||
|
]
|
||||||
|
}
|
||||||
|
```
|
||||||
253
gpu/modal_deployments/reflector_diarizer.py
Normal file
253
gpu/modal_deployments/reflector_diarizer.py
Normal file
@@ -0,0 +1,253 @@
|
|||||||
|
"""
|
||||||
|
Reflector GPU backend - diarizer
|
||||||
|
===================================
|
||||||
|
"""
|
||||||
|
|
||||||
|
import os
|
||||||
|
import uuid
|
||||||
|
from typing import Mapping, NewType
|
||||||
|
from urllib.parse import urlparse
|
||||||
|
|
||||||
|
import modal
|
||||||
|
|
||||||
|
PYANNOTE_MODEL_NAME: str = "pyannote/speaker-diarization-3.1"
|
||||||
|
MODEL_DIR = "/root/diarization_models"
|
||||||
|
UPLOADS_PATH = "/uploads"
|
||||||
|
SUPPORTED_FILE_EXTENSIONS = ["mp3", "mp4", "mpeg", "mpga", "m4a", "wav", "webm"]
|
||||||
|
|
||||||
|
DiarizerUniqFilename = NewType("DiarizerUniqFilename", str)
|
||||||
|
AudioFileExtension = NewType("AudioFileExtension", str)
|
||||||
|
|
||||||
|
app = modal.App(name="reflector-diarizer")
|
||||||
|
|
||||||
|
# Volume for temporary file uploads
|
||||||
|
upload_volume = modal.Volume.from_name("diarizer-uploads", create_if_missing=True)
|
||||||
|
|
||||||
|
|
||||||
|
def detect_audio_format(url: str, headers: Mapping[str, str]) -> AudioFileExtension:
|
||||||
|
parsed_url = urlparse(url)
|
||||||
|
url_path = parsed_url.path
|
||||||
|
|
||||||
|
for ext in SUPPORTED_FILE_EXTENSIONS:
|
||||||
|
if url_path.lower().endswith(f".{ext}"):
|
||||||
|
return AudioFileExtension(ext)
|
||||||
|
|
||||||
|
content_type = headers.get("content-type", "").lower()
|
||||||
|
if "audio/mpeg" in content_type or "audio/mp3" in content_type:
|
||||||
|
return AudioFileExtension("mp3")
|
||||||
|
if "audio/wav" in content_type:
|
||||||
|
return AudioFileExtension("wav")
|
||||||
|
if "audio/mp4" in content_type:
|
||||||
|
return AudioFileExtension("mp4")
|
||||||
|
|
||||||
|
raise ValueError(
|
||||||
|
f"Unsupported audio format for URL: {url}. "
|
||||||
|
f"Supported extensions: {', '.join(SUPPORTED_FILE_EXTENSIONS)}"
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def download_audio_to_volume(
|
||||||
|
audio_file_url: str,
|
||||||
|
) -> tuple[DiarizerUniqFilename, AudioFileExtension]:
|
||||||
|
import requests
|
||||||
|
from fastapi import HTTPException
|
||||||
|
|
||||||
|
print(f"Checking audio file at: {audio_file_url}")
|
||||||
|
response = requests.head(audio_file_url, allow_redirects=True)
|
||||||
|
if response.status_code == 404:
|
||||||
|
raise HTTPException(status_code=404, detail="Audio file not found")
|
||||||
|
|
||||||
|
print(f"Downloading audio file from: {audio_file_url}")
|
||||||
|
response = requests.get(audio_file_url, allow_redirects=True)
|
||||||
|
|
||||||
|
if response.status_code != 200:
|
||||||
|
print(f"Download failed with status {response.status_code}: {response.text}")
|
||||||
|
raise HTTPException(
|
||||||
|
status_code=response.status_code,
|
||||||
|
detail=f"Failed to download audio file: {response.status_code}",
|
||||||
|
)
|
||||||
|
|
||||||
|
audio_suffix = detect_audio_format(audio_file_url, response.headers)
|
||||||
|
unique_filename = DiarizerUniqFilename(f"{uuid.uuid4()}.{audio_suffix}")
|
||||||
|
file_path = f"{UPLOADS_PATH}/{unique_filename}"
|
||||||
|
|
||||||
|
print(f"Writing file to: {file_path} (size: {len(response.content)} bytes)")
|
||||||
|
with open(file_path, "wb") as f:
|
||||||
|
f.write(response.content)
|
||||||
|
|
||||||
|
upload_volume.commit()
|
||||||
|
print(f"File saved as: {unique_filename}")
|
||||||
|
return unique_filename, audio_suffix
|
||||||
|
|
||||||
|
|
||||||
|
def migrate_cache_llm():
|
||||||
|
"""
|
||||||
|
XXX The cache for model files in Transformers v4.22.0 has been updated.
|
||||||
|
Migrating your old cache. This is a one-time only operation. You can
|
||||||
|
interrupt this and resume the migration later on by calling
|
||||||
|
`transformers.utils.move_cache()`.
|
||||||
|
"""
|
||||||
|
from transformers.utils.hub import move_cache
|
||||||
|
|
||||||
|
print("Moving LLM cache")
|
||||||
|
move_cache(cache_dir=MODEL_DIR, new_cache_dir=MODEL_DIR)
|
||||||
|
print("LLM cache moved")
|
||||||
|
|
||||||
|
|
||||||
|
def download_pyannote_audio():
|
||||||
|
from pyannote.audio import Pipeline
|
||||||
|
|
||||||
|
Pipeline.from_pretrained(
|
||||||
|
PYANNOTE_MODEL_NAME,
|
||||||
|
cache_dir=MODEL_DIR,
|
||||||
|
use_auth_token=os.environ["HF_TOKEN"],
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
diarizer_image = (
|
||||||
|
modal.Image.debian_slim(python_version="3.10.8")
|
||||||
|
.pip_install(
|
||||||
|
"pyannote.audio==3.1.0",
|
||||||
|
"requests",
|
||||||
|
"onnx",
|
||||||
|
"torchaudio",
|
||||||
|
"onnxruntime-gpu",
|
||||||
|
"torch==2.0.0",
|
||||||
|
"transformers==4.34.0",
|
||||||
|
"sentencepiece",
|
||||||
|
"protobuf",
|
||||||
|
"numpy",
|
||||||
|
"huggingface_hub",
|
||||||
|
"hf-transfer",
|
||||||
|
)
|
||||||
|
.run_function(
|
||||||
|
download_pyannote_audio,
|
||||||
|
secrets=[modal.Secret.from_name("hf_token")],
|
||||||
|
)
|
||||||
|
.run_function(migrate_cache_llm)
|
||||||
|
.env(
|
||||||
|
{
|
||||||
|
"LD_LIBRARY_PATH": (
|
||||||
|
"/usr/local/lib/python3.10/site-packages/nvidia/cudnn/lib/:"
|
||||||
|
"/opt/conda/lib/python3.10/site-packages/nvidia/cublas/lib/"
|
||||||
|
)
|
||||||
|
}
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@app.cls(
|
||||||
|
gpu="A100",
|
||||||
|
timeout=60 * 30,
|
||||||
|
image=diarizer_image,
|
||||||
|
volumes={UPLOADS_PATH: upload_volume},
|
||||||
|
enable_memory_snapshot=True,
|
||||||
|
experimental_options={"enable_gpu_snapshot": True},
|
||||||
|
secrets=[
|
||||||
|
modal.Secret.from_name("hf_token"),
|
||||||
|
],
|
||||||
|
)
|
||||||
|
@modal.concurrent(max_inputs=1)
|
||||||
|
class Diarizer:
|
||||||
|
@modal.enter(snap=True)
|
||||||
|
def enter(self):
|
||||||
|
import torch
|
||||||
|
from pyannote.audio import Pipeline
|
||||||
|
|
||||||
|
self.use_gpu = torch.cuda.is_available()
|
||||||
|
self.device = "cuda" if self.use_gpu else "cpu"
|
||||||
|
print(f"Using device: {self.device}")
|
||||||
|
self.diarization_pipeline = Pipeline.from_pretrained(
|
||||||
|
PYANNOTE_MODEL_NAME,
|
||||||
|
cache_dir=MODEL_DIR,
|
||||||
|
use_auth_token=os.environ["HF_TOKEN"],
|
||||||
|
)
|
||||||
|
self.diarization_pipeline.to(torch.device(self.device))
|
||||||
|
|
||||||
|
@modal.method()
|
||||||
|
def diarize(self, filename: str, timestamp: float = 0.0):
|
||||||
|
import torchaudio
|
||||||
|
|
||||||
|
upload_volume.reload()
|
||||||
|
|
||||||
|
file_path = f"{UPLOADS_PATH}/{filename}"
|
||||||
|
if not os.path.exists(file_path):
|
||||||
|
raise FileNotFoundError(f"File not found: {file_path}")
|
||||||
|
|
||||||
|
print(f"Diarizing audio from: {file_path}")
|
||||||
|
waveform, sample_rate = torchaudio.load(file_path)
|
||||||
|
diarization = self.diarization_pipeline(
|
||||||
|
{"waveform": waveform, "sample_rate": sample_rate}
|
||||||
|
)
|
||||||
|
|
||||||
|
words = []
|
||||||
|
for diarization_segment, _, speaker in diarization.itertracks(yield_label=True):
|
||||||
|
words.append(
|
||||||
|
{
|
||||||
|
"start": round(timestamp + diarization_segment.start, 3),
|
||||||
|
"end": round(timestamp + diarization_segment.end, 3),
|
||||||
|
"speaker": int(speaker[-2:]),
|
||||||
|
}
|
||||||
|
)
|
||||||
|
print("Diarization complete")
|
||||||
|
return {"diarization": words}
|
||||||
|
|
||||||
|
|
||||||
|
# -------------------------------------------------------------------
|
||||||
|
# Web API
|
||||||
|
# -------------------------------------------------------------------
|
||||||
|
|
||||||
|
|
||||||
|
@app.function(
|
||||||
|
timeout=60 * 10,
|
||||||
|
scaledown_window=60 * 3,
|
||||||
|
secrets=[
|
||||||
|
modal.Secret.from_name("reflector-gpu"),
|
||||||
|
],
|
||||||
|
volumes={UPLOADS_PATH: upload_volume},
|
||||||
|
image=diarizer_image,
|
||||||
|
)
|
||||||
|
@modal.concurrent(max_inputs=40)
|
||||||
|
@modal.asgi_app()
|
||||||
|
def web():
|
||||||
|
from fastapi import Depends, FastAPI, HTTPException, status
|
||||||
|
from fastapi.security import OAuth2PasswordBearer
|
||||||
|
from pydantic import BaseModel
|
||||||
|
|
||||||
|
diarizerstub = Diarizer()
|
||||||
|
|
||||||
|
app = FastAPI()
|
||||||
|
|
||||||
|
oauth2_scheme = OAuth2PasswordBearer(tokenUrl="token")
|
||||||
|
|
||||||
|
def apikey_auth(apikey: str = Depends(oauth2_scheme)):
|
||||||
|
if apikey != os.environ["REFLECTOR_GPU_APIKEY"]:
|
||||||
|
raise HTTPException(
|
||||||
|
status_code=status.HTTP_401_UNAUTHORIZED,
|
||||||
|
detail="Invalid API key",
|
||||||
|
headers={"WWW-Authenticate": "Bearer"},
|
||||||
|
)
|
||||||
|
|
||||||
|
class DiarizationResponse(BaseModel):
|
||||||
|
result: dict
|
||||||
|
|
||||||
|
@app.post("/diarize", dependencies=[Depends(apikey_auth)])
|
||||||
|
def diarize(audio_file_url: str, timestamp: float = 0.0) -> DiarizationResponse:
|
||||||
|
unique_filename, audio_suffix = download_audio_to_volume(audio_file_url)
|
||||||
|
|
||||||
|
try:
|
||||||
|
func = diarizerstub.diarize.spawn(
|
||||||
|
filename=unique_filename, timestamp=timestamp
|
||||||
|
)
|
||||||
|
result = func.get()
|
||||||
|
return result
|
||||||
|
finally:
|
||||||
|
try:
|
||||||
|
file_path = f"{UPLOADS_PATH}/{unique_filename}"
|
||||||
|
print(f"Deleting file: {file_path}")
|
||||||
|
os.remove(file_path)
|
||||||
|
upload_volume.commit()
|
||||||
|
except Exception as e:
|
||||||
|
print(f"Error cleaning up {unique_filename}: {e}")
|
||||||
|
|
||||||
|
return app
|
||||||
608
gpu/modal_deployments/reflector_transcriber.py
Normal file
608
gpu/modal_deployments/reflector_transcriber.py
Normal file
@@ -0,0 +1,608 @@
|
|||||||
|
import os
|
||||||
|
import sys
|
||||||
|
import threading
|
||||||
|
import uuid
|
||||||
|
from typing import Generator, Mapping, NamedTuple, NewType, TypedDict
|
||||||
|
from urllib.parse import urlparse
|
||||||
|
|
||||||
|
import modal
|
||||||
|
|
||||||
|
MODEL_NAME = "large-v2"
|
||||||
|
MODEL_COMPUTE_TYPE: str = "float16"
|
||||||
|
MODEL_NUM_WORKERS: int = 1
|
||||||
|
MINUTES = 60 # seconds
|
||||||
|
SAMPLERATE = 16000
|
||||||
|
UPLOADS_PATH = "/uploads"
|
||||||
|
CACHE_PATH = "/models"
|
||||||
|
SUPPORTED_FILE_EXTENSIONS = ["mp3", "mp4", "mpeg", "mpga", "m4a", "wav", "webm"]
|
||||||
|
VAD_CONFIG = {
|
||||||
|
"batch_max_duration": 30.0,
|
||||||
|
"silence_padding": 0.5,
|
||||||
|
"window_size": 512,
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
|
WhisperUniqFilename = NewType("WhisperUniqFilename", str)
|
||||||
|
AudioFileExtension = NewType("AudioFileExtension", str)
|
||||||
|
|
||||||
|
app = modal.App("reflector-transcriber")
|
||||||
|
|
||||||
|
model_cache = modal.Volume.from_name("models", create_if_missing=True)
|
||||||
|
upload_volume = modal.Volume.from_name("whisper-uploads", create_if_missing=True)
|
||||||
|
|
||||||
|
|
||||||
|
class TimeSegment(NamedTuple):
|
||||||
|
"""Represents a time segment with start and end times."""
|
||||||
|
|
||||||
|
start: float
|
||||||
|
end: float
|
||||||
|
|
||||||
|
|
||||||
|
class AudioSegment(NamedTuple):
|
||||||
|
"""Represents an audio segment with timing and audio data."""
|
||||||
|
|
||||||
|
start: float
|
||||||
|
end: float
|
||||||
|
audio: any
|
||||||
|
|
||||||
|
|
||||||
|
class TranscriptResult(NamedTuple):
|
||||||
|
"""Represents a transcription result with text and word timings."""
|
||||||
|
|
||||||
|
text: str
|
||||||
|
words: list["WordTiming"]
|
||||||
|
|
||||||
|
|
||||||
|
class WordTiming(TypedDict):
|
||||||
|
"""Represents a word with its timing information."""
|
||||||
|
|
||||||
|
word: str
|
||||||
|
start: float
|
||||||
|
end: float
|
||||||
|
|
||||||
|
|
||||||
|
def download_model():
|
||||||
|
from faster_whisper import download_model
|
||||||
|
|
||||||
|
model_cache.reload()
|
||||||
|
|
||||||
|
download_model(MODEL_NAME, cache_dir=CACHE_PATH)
|
||||||
|
|
||||||
|
model_cache.commit()
|
||||||
|
|
||||||
|
|
||||||
|
image = (
|
||||||
|
modal.Image.debian_slim(python_version="3.12")
|
||||||
|
.env(
|
||||||
|
{
|
||||||
|
"HF_HUB_ENABLE_HF_TRANSFER": "1",
|
||||||
|
"LD_LIBRARY_PATH": (
|
||||||
|
"/usr/local/lib/python3.12/site-packages/nvidia/cudnn/lib/:"
|
||||||
|
"/opt/conda/lib/python3.12/site-packages/nvidia/cublas/lib/"
|
||||||
|
),
|
||||||
|
}
|
||||||
|
)
|
||||||
|
.apt_install("ffmpeg")
|
||||||
|
.pip_install(
|
||||||
|
"huggingface_hub==0.27.1",
|
||||||
|
"hf-transfer==0.1.9",
|
||||||
|
"torch==2.5.1",
|
||||||
|
"faster-whisper==1.1.1",
|
||||||
|
"fastapi==0.115.12",
|
||||||
|
"requests",
|
||||||
|
"librosa==0.10.1",
|
||||||
|
"numpy<2",
|
||||||
|
"silero-vad==5.1.0",
|
||||||
|
)
|
||||||
|
.run_function(download_model, volumes={CACHE_PATH: model_cache})
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def detect_audio_format(url: str, headers: Mapping[str, str]) -> AudioFileExtension:
|
||||||
|
parsed_url = urlparse(url)
|
||||||
|
url_path = parsed_url.path
|
||||||
|
|
||||||
|
for ext in SUPPORTED_FILE_EXTENSIONS:
|
||||||
|
if url_path.lower().endswith(f".{ext}"):
|
||||||
|
return AudioFileExtension(ext)
|
||||||
|
|
||||||
|
content_type = headers.get("content-type", "").lower()
|
||||||
|
if "audio/mpeg" in content_type or "audio/mp3" in content_type:
|
||||||
|
return AudioFileExtension("mp3")
|
||||||
|
if "audio/wav" in content_type:
|
||||||
|
return AudioFileExtension("wav")
|
||||||
|
if "audio/mp4" in content_type:
|
||||||
|
return AudioFileExtension("mp4")
|
||||||
|
|
||||||
|
raise ValueError(
|
||||||
|
f"Unsupported audio format for URL: {url}. "
|
||||||
|
f"Supported extensions: {', '.join(SUPPORTED_FILE_EXTENSIONS)}"
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def download_audio_to_volume(
|
||||||
|
audio_file_url: str,
|
||||||
|
) -> tuple[WhisperUniqFilename, AudioFileExtension]:
|
||||||
|
import requests
|
||||||
|
from fastapi import HTTPException
|
||||||
|
|
||||||
|
response = requests.head(audio_file_url, allow_redirects=True)
|
||||||
|
if response.status_code == 404:
|
||||||
|
raise HTTPException(status_code=404, detail="Audio file not found")
|
||||||
|
|
||||||
|
response = requests.get(audio_file_url, allow_redirects=True)
|
||||||
|
response.raise_for_status()
|
||||||
|
|
||||||
|
audio_suffix = detect_audio_format(audio_file_url, response.headers)
|
||||||
|
unique_filename = WhisperUniqFilename(f"{uuid.uuid4()}.{audio_suffix}")
|
||||||
|
file_path = f"{UPLOADS_PATH}/{unique_filename}"
|
||||||
|
|
||||||
|
with open(file_path, "wb") as f:
|
||||||
|
f.write(response.content)
|
||||||
|
|
||||||
|
upload_volume.commit()
|
||||||
|
return unique_filename, audio_suffix
|
||||||
|
|
||||||
|
|
||||||
|
def pad_audio(audio_array, sample_rate: int = SAMPLERATE):
|
||||||
|
"""Add 0.5s of silence if audio is shorter than the silence_padding window.
|
||||||
|
|
||||||
|
Whisper does not require this strictly, but aligning behavior with Parakeet
|
||||||
|
avoids edge-case crashes on extremely short inputs and makes comparisons easier.
|
||||||
|
"""
|
||||||
|
import numpy as np
|
||||||
|
|
||||||
|
audio_duration = len(audio_array) / sample_rate
|
||||||
|
if audio_duration < VAD_CONFIG["silence_padding"]:
|
||||||
|
silence_samples = int(sample_rate * VAD_CONFIG["silence_padding"])
|
||||||
|
silence = np.zeros(silence_samples, dtype=np.float32)
|
||||||
|
return np.concatenate([audio_array, silence])
|
||||||
|
return audio_array
|
||||||
|
|
||||||
|
|
||||||
|
@app.cls(
|
||||||
|
gpu="A10G",
|
||||||
|
timeout=5 * MINUTES,
|
||||||
|
scaledown_window=5 * MINUTES,
|
||||||
|
image=image,
|
||||||
|
volumes={CACHE_PATH: model_cache, UPLOADS_PATH: upload_volume},
|
||||||
|
)
|
||||||
|
@modal.concurrent(max_inputs=10)
|
||||||
|
class TranscriberWhisperLive:
|
||||||
|
"""Live transcriber class for small audio segments (A10G).
|
||||||
|
|
||||||
|
Mirrors the Parakeet live class API but uses Faster-Whisper under the hood.
|
||||||
|
"""
|
||||||
|
|
||||||
|
@modal.enter()
|
||||||
|
def enter(self):
|
||||||
|
import faster_whisper
|
||||||
|
import torch
|
||||||
|
|
||||||
|
self.lock = threading.Lock()
|
||||||
|
self.use_gpu = torch.cuda.is_available()
|
||||||
|
self.device = "cuda" if self.use_gpu else "cpu"
|
||||||
|
self.model = faster_whisper.WhisperModel(
|
||||||
|
MODEL_NAME,
|
||||||
|
device=self.device,
|
||||||
|
compute_type=MODEL_COMPUTE_TYPE,
|
||||||
|
num_workers=MODEL_NUM_WORKERS,
|
||||||
|
download_root=CACHE_PATH,
|
||||||
|
local_files_only=True,
|
||||||
|
)
|
||||||
|
print(f"Model is on device: {self.device}")
|
||||||
|
|
||||||
|
@modal.method()
|
||||||
|
def transcribe_segment(
|
||||||
|
self,
|
||||||
|
filename: str,
|
||||||
|
language: str = "en",
|
||||||
|
):
|
||||||
|
"""Transcribe a single uploaded audio file by filename."""
|
||||||
|
upload_volume.reload()
|
||||||
|
|
||||||
|
file_path = f"{UPLOADS_PATH}/{filename}"
|
||||||
|
if not os.path.exists(file_path):
|
||||||
|
raise FileNotFoundError(f"File not found: {file_path}")
|
||||||
|
|
||||||
|
with self.lock:
|
||||||
|
with NoStdStreams():
|
||||||
|
segments, _ = self.model.transcribe(
|
||||||
|
file_path,
|
||||||
|
language=language,
|
||||||
|
beam_size=5,
|
||||||
|
word_timestamps=True,
|
||||||
|
vad_filter=True,
|
||||||
|
vad_parameters={"min_silence_duration_ms": 500},
|
||||||
|
)
|
||||||
|
|
||||||
|
segments = list(segments)
|
||||||
|
text = "".join(segment.text for segment in segments).strip()
|
||||||
|
words = [
|
||||||
|
{
|
||||||
|
"word": word.word,
|
||||||
|
"start": round(float(word.start), 2),
|
||||||
|
"end": round(float(word.end), 2),
|
||||||
|
}
|
||||||
|
for segment in segments
|
||||||
|
for word in segment.words
|
||||||
|
]
|
||||||
|
|
||||||
|
return {"text": text, "words": words}
|
||||||
|
|
||||||
|
@modal.method()
|
||||||
|
def transcribe_batch(
|
||||||
|
self,
|
||||||
|
filenames: list[str],
|
||||||
|
language: str = "en",
|
||||||
|
):
|
||||||
|
"""Transcribe multiple uploaded audio files and return per-file results."""
|
||||||
|
upload_volume.reload()
|
||||||
|
|
||||||
|
results = []
|
||||||
|
for filename in filenames:
|
||||||
|
file_path = f"{UPLOADS_PATH}/{filename}"
|
||||||
|
if not os.path.exists(file_path):
|
||||||
|
raise FileNotFoundError(f"Batch file not found: {file_path}")
|
||||||
|
|
||||||
|
with self.lock:
|
||||||
|
with NoStdStreams():
|
||||||
|
segments, _ = self.model.transcribe(
|
||||||
|
file_path,
|
||||||
|
language=language,
|
||||||
|
beam_size=5,
|
||||||
|
word_timestamps=True,
|
||||||
|
vad_filter=True,
|
||||||
|
vad_parameters={"min_silence_duration_ms": 500},
|
||||||
|
)
|
||||||
|
|
||||||
|
segments = list(segments)
|
||||||
|
text = "".join(seg.text for seg in segments).strip()
|
||||||
|
words = [
|
||||||
|
{
|
||||||
|
"word": w.word,
|
||||||
|
"start": round(float(w.start), 2),
|
||||||
|
"end": round(float(w.end), 2),
|
||||||
|
}
|
||||||
|
for seg in segments
|
||||||
|
for w in seg.words
|
||||||
|
]
|
||||||
|
|
||||||
|
results.append(
|
||||||
|
{
|
||||||
|
"filename": filename,
|
||||||
|
"text": text,
|
||||||
|
"words": words,
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
return results
|
||||||
|
|
||||||
|
|
||||||
|
@app.cls(
|
||||||
|
gpu="L40S",
|
||||||
|
timeout=15 * MINUTES,
|
||||||
|
image=image,
|
||||||
|
volumes={CACHE_PATH: model_cache, UPLOADS_PATH: upload_volume},
|
||||||
|
)
|
||||||
|
class TranscriberWhisperFile:
|
||||||
|
"""File transcriber for larger/longer audio, using VAD-driven batching (L40S)."""
|
||||||
|
|
||||||
|
@modal.enter()
|
||||||
|
def enter(self):
|
||||||
|
import faster_whisper
|
||||||
|
import torch
|
||||||
|
from silero_vad import load_silero_vad
|
||||||
|
|
||||||
|
self.lock = threading.Lock()
|
||||||
|
self.use_gpu = torch.cuda.is_available()
|
||||||
|
self.device = "cuda" if self.use_gpu else "cpu"
|
||||||
|
self.model = faster_whisper.WhisperModel(
|
||||||
|
MODEL_NAME,
|
||||||
|
device=self.device,
|
||||||
|
compute_type=MODEL_COMPUTE_TYPE,
|
||||||
|
num_workers=MODEL_NUM_WORKERS,
|
||||||
|
download_root=CACHE_PATH,
|
||||||
|
local_files_only=True,
|
||||||
|
)
|
||||||
|
self.vad_model = load_silero_vad(onnx=False)
|
||||||
|
|
||||||
|
@modal.method()
|
||||||
|
def transcribe_segment(
|
||||||
|
self, filename: str, timestamp_offset: float = 0.0, language: str = "en"
|
||||||
|
):
|
||||||
|
import librosa
|
||||||
|
import numpy as np
|
||||||
|
from silero_vad import VADIterator
|
||||||
|
|
||||||
|
def vad_segments(
|
||||||
|
audio_array,
|
||||||
|
sample_rate: int = SAMPLERATE,
|
||||||
|
window_size: int = VAD_CONFIG["window_size"],
|
||||||
|
) -> Generator[TimeSegment, None, None]:
|
||||||
|
"""Generate speech segments as TimeSegment using Silero VAD."""
|
||||||
|
iterator = VADIterator(self.vad_model, sampling_rate=sample_rate)
|
||||||
|
start = None
|
||||||
|
for i in range(0, len(audio_array), window_size):
|
||||||
|
chunk = audio_array[i : i + window_size]
|
||||||
|
if len(chunk) < window_size:
|
||||||
|
chunk = np.pad(
|
||||||
|
chunk, (0, window_size - len(chunk)), mode="constant"
|
||||||
|
)
|
||||||
|
speech = iterator(chunk)
|
||||||
|
if not speech:
|
||||||
|
continue
|
||||||
|
if "start" in speech:
|
||||||
|
start = speech["start"]
|
||||||
|
continue
|
||||||
|
if "end" in speech and start is not None:
|
||||||
|
end = speech["end"]
|
||||||
|
yield TimeSegment(
|
||||||
|
start / float(SAMPLERATE), end / float(SAMPLERATE)
|
||||||
|
)
|
||||||
|
start = None
|
||||||
|
iterator.reset_states()
|
||||||
|
|
||||||
|
upload_volume.reload()
|
||||||
|
file_path = f"{UPLOADS_PATH}/{filename}"
|
||||||
|
if not os.path.exists(file_path):
|
||||||
|
raise FileNotFoundError(f"File not found: {file_path}")
|
||||||
|
|
||||||
|
audio_array, _sr = librosa.load(file_path, sr=SAMPLERATE, mono=True)
|
||||||
|
|
||||||
|
# Batch segments up to ~30s windows by merging contiguous VAD segments
|
||||||
|
merged_batches: list[TimeSegment] = []
|
||||||
|
batch_start = None
|
||||||
|
batch_end = None
|
||||||
|
max_duration = VAD_CONFIG["batch_max_duration"]
|
||||||
|
for segment in vad_segments(audio_array):
|
||||||
|
seg_start, seg_end = segment.start, segment.end
|
||||||
|
if batch_start is None:
|
||||||
|
batch_start, batch_end = seg_start, seg_end
|
||||||
|
continue
|
||||||
|
if seg_end - batch_start <= max_duration:
|
||||||
|
batch_end = seg_end
|
||||||
|
else:
|
||||||
|
merged_batches.append(TimeSegment(batch_start, batch_end))
|
||||||
|
batch_start, batch_end = seg_start, seg_end
|
||||||
|
if batch_start is not None and batch_end is not None:
|
||||||
|
merged_batches.append(TimeSegment(batch_start, batch_end))
|
||||||
|
|
||||||
|
all_text = []
|
||||||
|
all_words = []
|
||||||
|
|
||||||
|
for segment in merged_batches:
|
||||||
|
start_time, end_time = segment.start, segment.end
|
||||||
|
s_idx = int(start_time * SAMPLERATE)
|
||||||
|
e_idx = int(end_time * SAMPLERATE)
|
||||||
|
segment = audio_array[s_idx:e_idx]
|
||||||
|
segment = pad_audio(segment, SAMPLERATE)
|
||||||
|
|
||||||
|
with self.lock:
|
||||||
|
segments, _ = self.model.transcribe(
|
||||||
|
segment,
|
||||||
|
language=language,
|
||||||
|
beam_size=5,
|
||||||
|
word_timestamps=True,
|
||||||
|
vad_filter=True,
|
||||||
|
vad_parameters={"min_silence_duration_ms": 500},
|
||||||
|
)
|
||||||
|
|
||||||
|
segments = list(segments)
|
||||||
|
text = "".join(seg.text for seg in segments).strip()
|
||||||
|
words = [
|
||||||
|
{
|
||||||
|
"word": w.word,
|
||||||
|
"start": round(float(w.start) + start_time + timestamp_offset, 2),
|
||||||
|
"end": round(float(w.end) + start_time + timestamp_offset, 2),
|
||||||
|
}
|
||||||
|
for seg in segments
|
||||||
|
for w in seg.words
|
||||||
|
]
|
||||||
|
if text:
|
||||||
|
all_text.append(text)
|
||||||
|
all_words.extend(words)
|
||||||
|
|
||||||
|
return {"text": " ".join(all_text), "words": all_words}
|
||||||
|
|
||||||
|
|
||||||
|
def detect_audio_format(url: str, headers: dict) -> str:
|
||||||
|
from urllib.parse import urlparse
|
||||||
|
|
||||||
|
from fastapi import HTTPException
|
||||||
|
|
||||||
|
url_path = urlparse(url).path
|
||||||
|
for ext in SUPPORTED_FILE_EXTENSIONS:
|
||||||
|
if url_path.lower().endswith(f".{ext}"):
|
||||||
|
return ext
|
||||||
|
|
||||||
|
content_type = headers.get("content-type", "").lower()
|
||||||
|
if "audio/mpeg" in content_type or "audio/mp3" in content_type:
|
||||||
|
return "mp3"
|
||||||
|
if "audio/wav" in content_type:
|
||||||
|
return "wav"
|
||||||
|
if "audio/mp4" in content_type:
|
||||||
|
return "mp4"
|
||||||
|
|
||||||
|
raise HTTPException(
|
||||||
|
status_code=400,
|
||||||
|
detail=(
|
||||||
|
f"Unsupported audio format for URL. Supported extensions: {', '.join(SUPPORTED_FILE_EXTENSIONS)}"
|
||||||
|
),
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def download_audio_to_volume(audio_file_url: str) -> tuple[str, str]:
|
||||||
|
import requests
|
||||||
|
from fastapi import HTTPException
|
||||||
|
|
||||||
|
response = requests.head(audio_file_url, allow_redirects=True)
|
||||||
|
if response.status_code == 404:
|
||||||
|
raise HTTPException(status_code=404, detail="Audio file not found")
|
||||||
|
|
||||||
|
response = requests.get(audio_file_url, allow_redirects=True)
|
||||||
|
response.raise_for_status()
|
||||||
|
|
||||||
|
audio_suffix = detect_audio_format(audio_file_url, response.headers)
|
||||||
|
unique_filename = f"{uuid.uuid4()}.{audio_suffix}"
|
||||||
|
file_path = f"{UPLOADS_PATH}/{unique_filename}"
|
||||||
|
|
||||||
|
with open(file_path, "wb") as f:
|
||||||
|
f.write(response.content)
|
||||||
|
|
||||||
|
upload_volume.commit()
|
||||||
|
return unique_filename, audio_suffix
|
||||||
|
|
||||||
|
|
||||||
|
@app.function(
|
||||||
|
scaledown_window=60,
|
||||||
|
timeout=600,
|
||||||
|
secrets=[
|
||||||
|
modal.Secret.from_name("reflector-gpu"),
|
||||||
|
],
|
||||||
|
volumes={CACHE_PATH: model_cache, UPLOADS_PATH: upload_volume},
|
||||||
|
image=image,
|
||||||
|
)
|
||||||
|
@modal.concurrent(max_inputs=40)
|
||||||
|
@modal.asgi_app()
|
||||||
|
def web():
|
||||||
|
from fastapi import (
|
||||||
|
Body,
|
||||||
|
Depends,
|
||||||
|
FastAPI,
|
||||||
|
Form,
|
||||||
|
HTTPException,
|
||||||
|
UploadFile,
|
||||||
|
status,
|
||||||
|
)
|
||||||
|
from fastapi.security import OAuth2PasswordBearer
|
||||||
|
|
||||||
|
transcriber_live = TranscriberWhisperLive()
|
||||||
|
transcriber_file = TranscriberWhisperFile()
|
||||||
|
|
||||||
|
app = FastAPI()
|
||||||
|
|
||||||
|
oauth2_scheme = OAuth2PasswordBearer(tokenUrl="token")
|
||||||
|
|
||||||
|
def apikey_auth(apikey: str = Depends(oauth2_scheme)):
|
||||||
|
if apikey == os.environ["REFLECTOR_GPU_APIKEY"]:
|
||||||
|
return
|
||||||
|
raise HTTPException(
|
||||||
|
status_code=status.HTTP_401_UNAUTHORIZED,
|
||||||
|
detail="Invalid API key",
|
||||||
|
headers={"WWW-Authenticate": "Bearer"},
|
||||||
|
)
|
||||||
|
|
||||||
|
class TranscriptResponse(dict):
|
||||||
|
pass
|
||||||
|
|
||||||
|
@app.post("/v1/audio/transcriptions", dependencies=[Depends(apikey_auth)])
|
||||||
|
def transcribe(
|
||||||
|
file: UploadFile = None,
|
||||||
|
files: list[UploadFile] | None = None,
|
||||||
|
model: str = Form(MODEL_NAME),
|
||||||
|
language: str = Form("en"),
|
||||||
|
batch: bool = Form(False),
|
||||||
|
):
|
||||||
|
if not file and not files:
|
||||||
|
raise HTTPException(
|
||||||
|
status_code=400, detail="Either 'file' or 'files' parameter is required"
|
||||||
|
)
|
||||||
|
if batch and not files:
|
||||||
|
raise HTTPException(
|
||||||
|
status_code=400, detail="Batch transcription requires 'files'"
|
||||||
|
)
|
||||||
|
|
||||||
|
upload_files = [file] if file else files
|
||||||
|
|
||||||
|
uploaded_filenames: list[str] = []
|
||||||
|
for upload_file in upload_files:
|
||||||
|
audio_suffix = upload_file.filename.split(".")[-1]
|
||||||
|
if audio_suffix not in SUPPORTED_FILE_EXTENSIONS:
|
||||||
|
raise HTTPException(
|
||||||
|
status_code=400,
|
||||||
|
detail=(
|
||||||
|
f"Unsupported audio format. Supported extensions: {', '.join(SUPPORTED_FILE_EXTENSIONS)}"
|
||||||
|
),
|
||||||
|
)
|
||||||
|
|
||||||
|
unique_filename = f"{uuid.uuid4()}.{audio_suffix}"
|
||||||
|
file_path = f"{UPLOADS_PATH}/{unique_filename}"
|
||||||
|
with open(file_path, "wb") as f:
|
||||||
|
content = upload_file.file.read()
|
||||||
|
f.write(content)
|
||||||
|
uploaded_filenames.append(unique_filename)
|
||||||
|
|
||||||
|
upload_volume.commit()
|
||||||
|
|
||||||
|
try:
|
||||||
|
if batch and len(upload_files) > 1:
|
||||||
|
func = transcriber_live.transcribe_batch.spawn(
|
||||||
|
filenames=uploaded_filenames,
|
||||||
|
language=language,
|
||||||
|
)
|
||||||
|
results = func.get()
|
||||||
|
return {"results": results}
|
||||||
|
|
||||||
|
results = []
|
||||||
|
for filename in uploaded_filenames:
|
||||||
|
func = transcriber_live.transcribe_segment.spawn(
|
||||||
|
filename=filename,
|
||||||
|
language=language,
|
||||||
|
)
|
||||||
|
result = func.get()
|
||||||
|
result["filename"] = filename
|
||||||
|
results.append(result)
|
||||||
|
|
||||||
|
return {"results": results} if len(results) > 1 else results[0]
|
||||||
|
finally:
|
||||||
|
for filename in uploaded_filenames:
|
||||||
|
try:
|
||||||
|
file_path = f"{UPLOADS_PATH}/{filename}"
|
||||||
|
os.remove(file_path)
|
||||||
|
except Exception:
|
||||||
|
pass
|
||||||
|
upload_volume.commit()
|
||||||
|
|
||||||
|
@app.post("/v1/audio/transcriptions-from-url", dependencies=[Depends(apikey_auth)])
|
||||||
|
def transcribe_from_url(
|
||||||
|
audio_file_url: str = Body(
|
||||||
|
..., description="URL of the audio file to transcribe"
|
||||||
|
),
|
||||||
|
model: str = Body(MODEL_NAME),
|
||||||
|
language: str = Body("en"),
|
||||||
|
timestamp_offset: float = Body(0.0),
|
||||||
|
):
|
||||||
|
unique_filename, _audio_suffix = download_audio_to_volume(audio_file_url)
|
||||||
|
try:
|
||||||
|
func = transcriber_file.transcribe_segment.spawn(
|
||||||
|
filename=unique_filename,
|
||||||
|
timestamp_offset=timestamp_offset,
|
||||||
|
language=language,
|
||||||
|
)
|
||||||
|
result = func.get()
|
||||||
|
return result
|
||||||
|
finally:
|
||||||
|
try:
|
||||||
|
file_path = f"{UPLOADS_PATH}/{unique_filename}"
|
||||||
|
os.remove(file_path)
|
||||||
|
upload_volume.commit()
|
||||||
|
except Exception:
|
||||||
|
pass
|
||||||
|
|
||||||
|
return app
|
||||||
|
|
||||||
|
|
||||||
|
class NoStdStreams:
|
||||||
|
def __init__(self):
|
||||||
|
self.devnull = open(os.devnull, "w")
|
||||||
|
|
||||||
|
def __enter__(self):
|
||||||
|
self._stdout, self._stderr = sys.stdout, sys.stderr
|
||||||
|
self._stdout.flush()
|
||||||
|
self._stderr.flush()
|
||||||
|
sys.stdout, sys.stderr = self.devnull, self.devnull
|
||||||
|
|
||||||
|
def __exit__(self, exc_type, exc_value, traceback):
|
||||||
|
sys.stdout, sys.stderr = self._stdout, self._stderr
|
||||||
|
self.devnull.close()
|
||||||
658
gpu/modal_deployments/reflector_transcriber_parakeet.py
Normal file
658
gpu/modal_deployments/reflector_transcriber_parakeet.py
Normal file
@@ -0,0 +1,658 @@
|
|||||||
|
import logging
|
||||||
|
import os
|
||||||
|
import sys
|
||||||
|
import threading
|
||||||
|
import uuid
|
||||||
|
from typing import Generator, Mapping, NamedTuple, NewType, TypedDict
|
||||||
|
from urllib.parse import urlparse
|
||||||
|
|
||||||
|
import modal
|
||||||
|
|
||||||
|
MODEL_NAME = "nvidia/parakeet-tdt-0.6b-v2"
|
||||||
|
SUPPORTED_FILE_EXTENSIONS = ["mp3", "mp4", "mpeg", "mpga", "m4a", "wav", "webm"]
|
||||||
|
SAMPLERATE = 16000
|
||||||
|
UPLOADS_PATH = "/uploads"
|
||||||
|
CACHE_PATH = "/cache"
|
||||||
|
VAD_CONFIG = {
|
||||||
|
"batch_max_duration": 30.0,
|
||||||
|
"silence_padding": 0.5,
|
||||||
|
"window_size": 512,
|
||||||
|
}
|
||||||
|
|
||||||
|
ParakeetUniqFilename = NewType("ParakeetUniqFilename", str)
|
||||||
|
AudioFileExtension = NewType("AudioFileExtension", str)
|
||||||
|
|
||||||
|
|
||||||
|
class TimeSegment(NamedTuple):
|
||||||
|
"""Represents a time segment with start and end times."""
|
||||||
|
|
||||||
|
start: float
|
||||||
|
end: float
|
||||||
|
|
||||||
|
|
||||||
|
class AudioSegment(NamedTuple):
|
||||||
|
"""Represents an audio segment with timing and audio data."""
|
||||||
|
|
||||||
|
start: float
|
||||||
|
end: float
|
||||||
|
audio: any
|
||||||
|
|
||||||
|
|
||||||
|
class TranscriptResult(NamedTuple):
|
||||||
|
"""Represents a transcription result with text and word timings."""
|
||||||
|
|
||||||
|
text: str
|
||||||
|
words: list["WordTiming"]
|
||||||
|
|
||||||
|
|
||||||
|
class WordTiming(TypedDict):
|
||||||
|
"""Represents a word with its timing information."""
|
||||||
|
|
||||||
|
word: str
|
||||||
|
start: float
|
||||||
|
end: float
|
||||||
|
|
||||||
|
|
||||||
|
app = modal.App("reflector-transcriber-parakeet")
|
||||||
|
|
||||||
|
# Volume for caching model weights
|
||||||
|
model_cache = modal.Volume.from_name("parakeet-model-cache", create_if_missing=True)
|
||||||
|
# Volume for temporary file uploads
|
||||||
|
upload_volume = modal.Volume.from_name("parakeet-uploads", create_if_missing=True)
|
||||||
|
|
||||||
|
image = (
|
||||||
|
modal.Image.from_registry(
|
||||||
|
"nvidia/cuda:12.8.0-cudnn-devel-ubuntu22.04", add_python="3.12"
|
||||||
|
)
|
||||||
|
.env(
|
||||||
|
{
|
||||||
|
"HF_HUB_ENABLE_HF_TRANSFER": "1",
|
||||||
|
"HF_HOME": "/cache",
|
||||||
|
"DEBIAN_FRONTEND": "noninteractive",
|
||||||
|
"CXX": "g++",
|
||||||
|
"CC": "g++",
|
||||||
|
}
|
||||||
|
)
|
||||||
|
.apt_install("ffmpeg")
|
||||||
|
.pip_install(
|
||||||
|
"hf_transfer==0.1.9",
|
||||||
|
"huggingface_hub[hf-xet]==0.31.2",
|
||||||
|
"nemo_toolkit[asr]==2.5.0",
|
||||||
|
"cuda-python==12.8.0",
|
||||||
|
"fastapi==0.115.12",
|
||||||
|
"numpy<2",
|
||||||
|
"librosa==0.10.1",
|
||||||
|
"requests",
|
||||||
|
"silero-vad==5.1.0",
|
||||||
|
"torch",
|
||||||
|
)
|
||||||
|
.entrypoint([]) # silence chatty logs by container on start
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def detect_audio_format(url: str, headers: Mapping[str, str]) -> AudioFileExtension:
|
||||||
|
parsed_url = urlparse(url)
|
||||||
|
url_path = parsed_url.path
|
||||||
|
|
||||||
|
for ext in SUPPORTED_FILE_EXTENSIONS:
|
||||||
|
if url_path.lower().endswith(f".{ext}"):
|
||||||
|
return AudioFileExtension(ext)
|
||||||
|
|
||||||
|
content_type = headers.get("content-type", "").lower()
|
||||||
|
if "audio/mpeg" in content_type or "audio/mp3" in content_type:
|
||||||
|
return AudioFileExtension("mp3")
|
||||||
|
if "audio/wav" in content_type:
|
||||||
|
return AudioFileExtension("wav")
|
||||||
|
if "audio/mp4" in content_type:
|
||||||
|
return AudioFileExtension("mp4")
|
||||||
|
|
||||||
|
raise ValueError(
|
||||||
|
f"Unsupported audio format for URL: {url}. "
|
||||||
|
f"Supported extensions: {', '.join(SUPPORTED_FILE_EXTENSIONS)}"
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def download_audio_to_volume(
|
||||||
|
audio_file_url: str,
|
||||||
|
) -> tuple[ParakeetUniqFilename, AudioFileExtension]:
|
||||||
|
import requests
|
||||||
|
from fastapi import HTTPException
|
||||||
|
|
||||||
|
response = requests.head(audio_file_url, allow_redirects=True)
|
||||||
|
if response.status_code == 404:
|
||||||
|
raise HTTPException(status_code=404, detail="Audio file not found")
|
||||||
|
|
||||||
|
response = requests.get(audio_file_url, allow_redirects=True)
|
||||||
|
response.raise_for_status()
|
||||||
|
|
||||||
|
audio_suffix = detect_audio_format(audio_file_url, response.headers)
|
||||||
|
unique_filename = ParakeetUniqFilename(f"{uuid.uuid4()}.{audio_suffix}")
|
||||||
|
file_path = f"{UPLOADS_PATH}/{unique_filename}"
|
||||||
|
|
||||||
|
with open(file_path, "wb") as f:
|
||||||
|
f.write(response.content)
|
||||||
|
|
||||||
|
upload_volume.commit()
|
||||||
|
return unique_filename, audio_suffix
|
||||||
|
|
||||||
|
|
||||||
|
def pad_audio(audio_array, sample_rate: int = SAMPLERATE):
|
||||||
|
"""Add 0.5 seconds of silence if audio is less than 500ms.
|
||||||
|
|
||||||
|
This is a workaround for a Parakeet bug where very short audio (<500ms) causes:
|
||||||
|
ValueError: `char_offsets`: [] and `processed_tokens`: [157, 834, 834, 841]
|
||||||
|
have to be of the same length
|
||||||
|
|
||||||
|
See: https://github.com/NVIDIA/NeMo/issues/8451
|
||||||
|
"""
|
||||||
|
import numpy as np
|
||||||
|
|
||||||
|
audio_duration = len(audio_array) / sample_rate
|
||||||
|
if audio_duration < 0.5:
|
||||||
|
silence_samples = int(sample_rate * 0.5)
|
||||||
|
silence = np.zeros(silence_samples, dtype=np.float32)
|
||||||
|
return np.concatenate([audio_array, silence])
|
||||||
|
return audio_array
|
||||||
|
|
||||||
|
|
||||||
|
@app.cls(
|
||||||
|
gpu="A10G",
|
||||||
|
timeout=600,
|
||||||
|
scaledown_window=300,
|
||||||
|
image=image,
|
||||||
|
volumes={CACHE_PATH: model_cache, UPLOADS_PATH: upload_volume},
|
||||||
|
enable_memory_snapshot=True,
|
||||||
|
experimental_options={"enable_gpu_snapshot": True},
|
||||||
|
)
|
||||||
|
@modal.concurrent(max_inputs=10)
|
||||||
|
class TranscriberParakeetLive:
|
||||||
|
@modal.enter(snap=True)
|
||||||
|
def enter(self):
|
||||||
|
import nemo.collections.asr as nemo_asr
|
||||||
|
|
||||||
|
logging.getLogger("nemo_logger").setLevel(logging.CRITICAL)
|
||||||
|
|
||||||
|
self.lock = threading.Lock()
|
||||||
|
self.model = nemo_asr.models.ASRModel.from_pretrained(model_name=MODEL_NAME)
|
||||||
|
device = next(self.model.parameters()).device
|
||||||
|
print(f"Model is on device: {device}")
|
||||||
|
|
||||||
|
@modal.method()
|
||||||
|
def transcribe_segment(
|
||||||
|
self,
|
||||||
|
filename: str,
|
||||||
|
):
|
||||||
|
import librosa
|
||||||
|
|
||||||
|
upload_volume.reload()
|
||||||
|
|
||||||
|
file_path = f"{UPLOADS_PATH}/{filename}"
|
||||||
|
if not os.path.exists(file_path):
|
||||||
|
raise FileNotFoundError(f"File not found: {file_path}")
|
||||||
|
|
||||||
|
audio_array, sample_rate = librosa.load(file_path, sr=SAMPLERATE, mono=True)
|
||||||
|
padded_audio = pad_audio(audio_array, sample_rate)
|
||||||
|
|
||||||
|
with self.lock:
|
||||||
|
with NoStdStreams():
|
||||||
|
(output,) = self.model.transcribe([padded_audio], timestamps=True)
|
||||||
|
|
||||||
|
text = output.text.strip()
|
||||||
|
words: list[WordTiming] = [
|
||||||
|
WordTiming(
|
||||||
|
# XXX the space added here is to match the output of whisper
|
||||||
|
# whisper add space to each words, while parakeet don't
|
||||||
|
word=word_info["word"] + " ",
|
||||||
|
start=round(word_info["start"], 2),
|
||||||
|
end=round(word_info["end"], 2),
|
||||||
|
)
|
||||||
|
for word_info in output.timestamp["word"]
|
||||||
|
]
|
||||||
|
|
||||||
|
return {"text": text, "words": words}
|
||||||
|
|
||||||
|
@modal.method()
|
||||||
|
def transcribe_batch(
|
||||||
|
self,
|
||||||
|
filenames: list[str],
|
||||||
|
):
|
||||||
|
import librosa
|
||||||
|
|
||||||
|
upload_volume.reload()
|
||||||
|
|
||||||
|
results = []
|
||||||
|
audio_arrays = []
|
||||||
|
|
||||||
|
# Load all audio files with padding
|
||||||
|
for filename in filenames:
|
||||||
|
file_path = f"{UPLOADS_PATH}/{filename}"
|
||||||
|
if not os.path.exists(file_path):
|
||||||
|
raise FileNotFoundError(f"Batch file not found: {file_path}")
|
||||||
|
|
||||||
|
audio_array, sample_rate = librosa.load(file_path, sr=SAMPLERATE, mono=True)
|
||||||
|
padded_audio = pad_audio(audio_array, sample_rate)
|
||||||
|
audio_arrays.append(padded_audio)
|
||||||
|
|
||||||
|
with self.lock:
|
||||||
|
with NoStdStreams():
|
||||||
|
outputs = self.model.transcribe(audio_arrays, timestamps=True)
|
||||||
|
|
||||||
|
# Process results for each file
|
||||||
|
for i, (filename, output) in enumerate(zip(filenames, outputs)):
|
||||||
|
text = output.text.strip()
|
||||||
|
|
||||||
|
words: list[WordTiming] = [
|
||||||
|
WordTiming(
|
||||||
|
word=word_info["word"] + " ",
|
||||||
|
start=round(word_info["start"], 2),
|
||||||
|
end=round(word_info["end"], 2),
|
||||||
|
)
|
||||||
|
for word_info in output.timestamp["word"]
|
||||||
|
]
|
||||||
|
|
||||||
|
results.append(
|
||||||
|
{
|
||||||
|
"filename": filename,
|
||||||
|
"text": text,
|
||||||
|
"words": words,
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
return results
|
||||||
|
|
||||||
|
|
||||||
|
# L40S class for file transcription (bigger files)
|
||||||
|
@app.cls(
|
||||||
|
gpu="L40S",
|
||||||
|
timeout=900,
|
||||||
|
image=image,
|
||||||
|
volumes={CACHE_PATH: model_cache, UPLOADS_PATH: upload_volume},
|
||||||
|
enable_memory_snapshot=True,
|
||||||
|
experimental_options={"enable_gpu_snapshot": True},
|
||||||
|
)
|
||||||
|
class TranscriberParakeetFile:
|
||||||
|
@modal.enter(snap=True)
|
||||||
|
def enter(self):
|
||||||
|
import nemo.collections.asr as nemo_asr
|
||||||
|
import torch
|
||||||
|
from silero_vad import load_silero_vad
|
||||||
|
|
||||||
|
logging.getLogger("nemo_logger").setLevel(logging.CRITICAL)
|
||||||
|
|
||||||
|
self.model = nemo_asr.models.ASRModel.from_pretrained(model_name=MODEL_NAME)
|
||||||
|
device = next(self.model.parameters()).device
|
||||||
|
print(f"Model is on device: {device}")
|
||||||
|
|
||||||
|
torch.set_num_threads(1)
|
||||||
|
self.vad_model = load_silero_vad(onnx=False)
|
||||||
|
print("Silero VAD initialized")
|
||||||
|
|
||||||
|
@modal.method()
|
||||||
|
def transcribe_segment(
|
||||||
|
self,
|
||||||
|
filename: str,
|
||||||
|
timestamp_offset: float = 0.0,
|
||||||
|
):
|
||||||
|
import librosa
|
||||||
|
import numpy as np
|
||||||
|
from silero_vad import VADIterator
|
||||||
|
|
||||||
|
def load_and_convert_audio(file_path):
|
||||||
|
audio_array, sample_rate = librosa.load(file_path, sr=SAMPLERATE, mono=True)
|
||||||
|
return audio_array
|
||||||
|
|
||||||
|
def vad_segment_generator(
|
||||||
|
audio_array,
|
||||||
|
) -> Generator[TimeSegment, None, None]:
|
||||||
|
"""Generate speech segments using VAD with start/end sample indices"""
|
||||||
|
vad_iterator = VADIterator(self.vad_model, sampling_rate=SAMPLERATE)
|
||||||
|
window_size = VAD_CONFIG["window_size"]
|
||||||
|
start = None
|
||||||
|
|
||||||
|
for i in range(0, len(audio_array), window_size):
|
||||||
|
chunk = audio_array[i : i + window_size]
|
||||||
|
if len(chunk) < window_size:
|
||||||
|
chunk = np.pad(
|
||||||
|
chunk, (0, window_size - len(chunk)), mode="constant"
|
||||||
|
)
|
||||||
|
|
||||||
|
speech_dict = vad_iterator(chunk)
|
||||||
|
if not speech_dict:
|
||||||
|
continue
|
||||||
|
|
||||||
|
if "start" in speech_dict:
|
||||||
|
start = speech_dict["start"]
|
||||||
|
continue
|
||||||
|
|
||||||
|
if "end" in speech_dict and start is not None:
|
||||||
|
end = speech_dict["end"]
|
||||||
|
start_time = start / float(SAMPLERATE)
|
||||||
|
end_time = end / float(SAMPLERATE)
|
||||||
|
|
||||||
|
yield TimeSegment(start_time, end_time)
|
||||||
|
start = None
|
||||||
|
|
||||||
|
vad_iterator.reset_states()
|
||||||
|
|
||||||
|
def batch_speech_segments(
|
||||||
|
segments: Generator[TimeSegment, None, None], max_duration: int
|
||||||
|
) -> Generator[TimeSegment, None, None]:
|
||||||
|
"""
|
||||||
|
Input segments:
|
||||||
|
[0-2] [3-5] [6-8] [10-11] [12-15] [17-19] [20-22]
|
||||||
|
|
||||||
|
↓ (max_duration=10)
|
||||||
|
|
||||||
|
Output batches:
|
||||||
|
[0-8] [10-19] [20-22]
|
||||||
|
|
||||||
|
Note: silences are kept for better transcription, previous implementation was
|
||||||
|
passing segments separatly, but the output was less accurate.
|
||||||
|
"""
|
||||||
|
batch_start_time = None
|
||||||
|
batch_end_time = None
|
||||||
|
|
||||||
|
for segment in segments:
|
||||||
|
start_time, end_time = segment.start, segment.end
|
||||||
|
if batch_start_time is None or batch_end_time is None:
|
||||||
|
batch_start_time = start_time
|
||||||
|
batch_end_time = end_time
|
||||||
|
continue
|
||||||
|
|
||||||
|
total_duration = end_time - batch_start_time
|
||||||
|
|
||||||
|
if total_duration <= max_duration:
|
||||||
|
batch_end_time = end_time
|
||||||
|
continue
|
||||||
|
|
||||||
|
yield TimeSegment(batch_start_time, batch_end_time)
|
||||||
|
batch_start_time = start_time
|
||||||
|
batch_end_time = end_time
|
||||||
|
|
||||||
|
if batch_start_time is None or batch_end_time is None:
|
||||||
|
return
|
||||||
|
|
||||||
|
yield TimeSegment(batch_start_time, batch_end_time)
|
||||||
|
|
||||||
|
def batch_segment_to_audio_segment(
|
||||||
|
segments: Generator[TimeSegment, None, None],
|
||||||
|
audio_array,
|
||||||
|
) -> Generator[AudioSegment, None, None]:
|
||||||
|
"""Extract audio segments and apply padding for Parakeet compatibility.
|
||||||
|
|
||||||
|
Uses pad_audio to ensure segments are at least 0.5s long, preventing
|
||||||
|
Parakeet crashes. This padding may cause slight timing overlaps between
|
||||||
|
segments, which are corrected by enforce_word_timing_constraints.
|
||||||
|
"""
|
||||||
|
for segment in segments:
|
||||||
|
start_time, end_time = segment.start, segment.end
|
||||||
|
start_sample = int(start_time * SAMPLERATE)
|
||||||
|
end_sample = int(end_time * SAMPLERATE)
|
||||||
|
audio_segment = audio_array[start_sample:end_sample]
|
||||||
|
|
||||||
|
padded_segment = pad_audio(audio_segment, SAMPLERATE)
|
||||||
|
|
||||||
|
yield AudioSegment(start_time, end_time, padded_segment)
|
||||||
|
|
||||||
|
def transcribe_batch(model, audio_segments: list) -> list:
|
||||||
|
with NoStdStreams():
|
||||||
|
outputs = model.transcribe(audio_segments, timestamps=True)
|
||||||
|
return outputs
|
||||||
|
|
||||||
|
def enforce_word_timing_constraints(
|
||||||
|
words: list[WordTiming],
|
||||||
|
) -> list[WordTiming]:
|
||||||
|
"""Enforce that word end times don't exceed the start time of the next word.
|
||||||
|
|
||||||
|
Due to silence padding added in batch_segment_to_audio_segment for better
|
||||||
|
transcription accuracy, word timings from different segments may overlap.
|
||||||
|
This function ensures there are no overlaps by adjusting end times.
|
||||||
|
"""
|
||||||
|
if len(words) <= 1:
|
||||||
|
return words
|
||||||
|
|
||||||
|
enforced_words = []
|
||||||
|
for i, word in enumerate(words):
|
||||||
|
enforced_word = word.copy()
|
||||||
|
|
||||||
|
if i < len(words) - 1:
|
||||||
|
next_start = words[i + 1]["start"]
|
||||||
|
if enforced_word["end"] > next_start:
|
||||||
|
enforced_word["end"] = next_start
|
||||||
|
|
||||||
|
enforced_words.append(enforced_word)
|
||||||
|
|
||||||
|
return enforced_words
|
||||||
|
|
||||||
|
def emit_results(
|
||||||
|
results: list,
|
||||||
|
segments_info: list[AudioSegment],
|
||||||
|
) -> Generator[TranscriptResult, None, None]:
|
||||||
|
"""Yield transcribed text and word timings from model output, adjusting timestamps to absolute positions."""
|
||||||
|
for i, (output, segment) in enumerate(zip(results, segments_info)):
|
||||||
|
start_time, end_time = segment.start, segment.end
|
||||||
|
text = output.text.strip()
|
||||||
|
words: list[WordTiming] = [
|
||||||
|
WordTiming(
|
||||||
|
word=word_info["word"] + " ",
|
||||||
|
start=round(
|
||||||
|
word_info["start"] + start_time + timestamp_offset, 2
|
||||||
|
),
|
||||||
|
end=round(word_info["end"] + start_time + timestamp_offset, 2),
|
||||||
|
)
|
||||||
|
for word_info in output.timestamp["word"]
|
||||||
|
]
|
||||||
|
|
||||||
|
yield TranscriptResult(text, words)
|
||||||
|
|
||||||
|
upload_volume.reload()
|
||||||
|
|
||||||
|
file_path = f"{UPLOADS_PATH}/{filename}"
|
||||||
|
if not os.path.exists(file_path):
|
||||||
|
raise FileNotFoundError(f"File not found: {file_path}")
|
||||||
|
|
||||||
|
audio_array = load_and_convert_audio(file_path)
|
||||||
|
total_duration = len(audio_array) / float(SAMPLERATE)
|
||||||
|
|
||||||
|
all_text_parts: list[str] = []
|
||||||
|
all_words: list[WordTiming] = []
|
||||||
|
|
||||||
|
raw_segments = vad_segment_generator(audio_array)
|
||||||
|
speech_segments = batch_speech_segments(
|
||||||
|
raw_segments,
|
||||||
|
VAD_CONFIG["batch_max_duration"],
|
||||||
|
)
|
||||||
|
audio_segments = batch_segment_to_audio_segment(speech_segments, audio_array)
|
||||||
|
|
||||||
|
for batch in audio_segments:
|
||||||
|
audio_segment = batch.audio
|
||||||
|
results = transcribe_batch(self.model, [audio_segment])
|
||||||
|
|
||||||
|
for result in emit_results(
|
||||||
|
results,
|
||||||
|
[batch],
|
||||||
|
):
|
||||||
|
if not result.text:
|
||||||
|
continue
|
||||||
|
all_text_parts.append(result.text)
|
||||||
|
all_words.extend(result.words)
|
||||||
|
|
||||||
|
all_words = enforce_word_timing_constraints(all_words)
|
||||||
|
|
||||||
|
combined_text = " ".join(all_text_parts)
|
||||||
|
return {"text": combined_text, "words": all_words}
|
||||||
|
|
||||||
|
|
||||||
|
@app.function(
|
||||||
|
scaledown_window=60,
|
||||||
|
timeout=600,
|
||||||
|
secrets=[
|
||||||
|
modal.Secret.from_name("reflector-gpu"),
|
||||||
|
],
|
||||||
|
volumes={CACHE_PATH: model_cache, UPLOADS_PATH: upload_volume},
|
||||||
|
image=image,
|
||||||
|
)
|
||||||
|
@modal.concurrent(max_inputs=40)
|
||||||
|
@modal.asgi_app()
|
||||||
|
def web():
|
||||||
|
import os
|
||||||
|
import uuid
|
||||||
|
|
||||||
|
from fastapi import (
|
||||||
|
Body,
|
||||||
|
Depends,
|
||||||
|
FastAPI,
|
||||||
|
Form,
|
||||||
|
HTTPException,
|
||||||
|
UploadFile,
|
||||||
|
status,
|
||||||
|
)
|
||||||
|
from fastapi.security import OAuth2PasswordBearer
|
||||||
|
from pydantic import BaseModel
|
||||||
|
|
||||||
|
transcriber_live = TranscriberParakeetLive()
|
||||||
|
transcriber_file = TranscriberParakeetFile()
|
||||||
|
|
||||||
|
app = FastAPI()
|
||||||
|
|
||||||
|
oauth2_scheme = OAuth2PasswordBearer(tokenUrl="token")
|
||||||
|
|
||||||
|
def apikey_auth(apikey: str = Depends(oauth2_scheme)):
|
||||||
|
if apikey == os.environ["REFLECTOR_GPU_APIKEY"]:
|
||||||
|
return
|
||||||
|
raise HTTPException(
|
||||||
|
status_code=status.HTTP_401_UNAUTHORIZED,
|
||||||
|
detail="Invalid API key",
|
||||||
|
headers={"WWW-Authenticate": "Bearer"},
|
||||||
|
)
|
||||||
|
|
||||||
|
class TranscriptResponse(BaseModel):
|
||||||
|
result: dict
|
||||||
|
|
||||||
|
@app.post("/v1/audio/transcriptions", dependencies=[Depends(apikey_auth)])
|
||||||
|
def transcribe(
|
||||||
|
file: UploadFile = None,
|
||||||
|
files: list[UploadFile] | None = None,
|
||||||
|
model: str = Form(MODEL_NAME),
|
||||||
|
language: str = Form("en"),
|
||||||
|
batch: bool = Form(False),
|
||||||
|
):
|
||||||
|
# Parakeet only supports English
|
||||||
|
if language != "en":
|
||||||
|
raise HTTPException(
|
||||||
|
status_code=400,
|
||||||
|
detail=f"Parakeet model only supports English. Got language='{language}'",
|
||||||
|
)
|
||||||
|
# Handle both single file and multiple files
|
||||||
|
if not file and not files:
|
||||||
|
raise HTTPException(
|
||||||
|
status_code=400, detail="Either 'file' or 'files' parameter is required"
|
||||||
|
)
|
||||||
|
if batch and not files:
|
||||||
|
raise HTTPException(
|
||||||
|
status_code=400, detail="Batch transcription requires 'files'"
|
||||||
|
)
|
||||||
|
|
||||||
|
upload_files = [file] if file else files
|
||||||
|
|
||||||
|
# Upload files to volume
|
||||||
|
uploaded_filenames = []
|
||||||
|
for upload_file in upload_files:
|
||||||
|
audio_suffix = upload_file.filename.split(".")[-1]
|
||||||
|
assert audio_suffix in SUPPORTED_FILE_EXTENSIONS
|
||||||
|
|
||||||
|
# Generate unique filename
|
||||||
|
unique_filename = f"{uuid.uuid4()}.{audio_suffix}"
|
||||||
|
file_path = f"{UPLOADS_PATH}/{unique_filename}"
|
||||||
|
|
||||||
|
print(f"Writing file to: {file_path}")
|
||||||
|
with open(file_path, "wb") as f:
|
||||||
|
content = upload_file.file.read()
|
||||||
|
f.write(content)
|
||||||
|
|
||||||
|
uploaded_filenames.append(unique_filename)
|
||||||
|
|
||||||
|
upload_volume.commit()
|
||||||
|
|
||||||
|
try:
|
||||||
|
# Use A10G live transcriber for per-file transcription
|
||||||
|
if batch and len(upload_files) > 1:
|
||||||
|
# Use batch transcription
|
||||||
|
func = transcriber_live.transcribe_batch.spawn(
|
||||||
|
filenames=uploaded_filenames,
|
||||||
|
)
|
||||||
|
results = func.get()
|
||||||
|
return {"results": results}
|
||||||
|
|
||||||
|
# Per-file transcription
|
||||||
|
results = []
|
||||||
|
for filename in uploaded_filenames:
|
||||||
|
func = transcriber_live.transcribe_segment.spawn(
|
||||||
|
filename=filename,
|
||||||
|
)
|
||||||
|
result = func.get()
|
||||||
|
result["filename"] = filename
|
||||||
|
results.append(result)
|
||||||
|
|
||||||
|
return {"results": results} if len(results) > 1 else results[0]
|
||||||
|
|
||||||
|
finally:
|
||||||
|
for filename in uploaded_filenames:
|
||||||
|
try:
|
||||||
|
file_path = f"{UPLOADS_PATH}/{filename}"
|
||||||
|
print(f"Deleting file: {file_path}")
|
||||||
|
os.remove(file_path)
|
||||||
|
except Exception as e:
|
||||||
|
print(f"Error deleting {filename}: {e}")
|
||||||
|
|
||||||
|
upload_volume.commit()
|
||||||
|
|
||||||
|
@app.post("/v1/audio/transcriptions-from-url", dependencies=[Depends(apikey_auth)])
|
||||||
|
def transcribe_from_url(
|
||||||
|
audio_file_url: str = Body(
|
||||||
|
..., description="URL of the audio file to transcribe"
|
||||||
|
),
|
||||||
|
model: str = Body(MODEL_NAME),
|
||||||
|
language: str = Body("en", description="Language code (only 'en' supported)"),
|
||||||
|
timestamp_offset: float = Body(0.0),
|
||||||
|
):
|
||||||
|
# Parakeet only supports English
|
||||||
|
if language != "en":
|
||||||
|
raise HTTPException(
|
||||||
|
status_code=400,
|
||||||
|
detail=f"Parakeet model only supports English. Got language='{language}'",
|
||||||
|
)
|
||||||
|
unique_filename, audio_suffix = download_audio_to_volume(audio_file_url)
|
||||||
|
|
||||||
|
try:
|
||||||
|
func = transcriber_file.transcribe_segment.spawn(
|
||||||
|
filename=unique_filename,
|
||||||
|
timestamp_offset=timestamp_offset,
|
||||||
|
)
|
||||||
|
result = func.get()
|
||||||
|
return result
|
||||||
|
finally:
|
||||||
|
try:
|
||||||
|
file_path = f"{UPLOADS_PATH}/{unique_filename}"
|
||||||
|
print(f"Deleting file: {file_path}")
|
||||||
|
os.remove(file_path)
|
||||||
|
upload_volume.commit()
|
||||||
|
except Exception as e:
|
||||||
|
print(f"Error cleaning up {unique_filename}: {e}")
|
||||||
|
|
||||||
|
return app
|
||||||
|
|
||||||
|
|
||||||
|
class NoStdStreams:
|
||||||
|
def __init__(self):
|
||||||
|
self.devnull = open(os.devnull, "w")
|
||||||
|
|
||||||
|
def __enter__(self):
|
||||||
|
self._stdout, self._stderr = sys.stdout, sys.stderr
|
||||||
|
self._stdout.flush()
|
||||||
|
self._stderr.flush()
|
||||||
|
sys.stdout, sys.stderr = self.devnull, self.devnull
|
||||||
|
|
||||||
|
def __exit__(self, exc_type, exc_value, traceback):
|
||||||
|
sys.stdout, sys.stderr = self._stdout, self._stderr
|
||||||
|
self.devnull.close()
|
||||||
2
gpu/self_hosted/.env.example
Normal file
2
gpu/self_hosted/.env.example
Normal file
@@ -0,0 +1,2 @@
|
|||||||
|
REFLECTOR_GPU_APIKEY=
|
||||||
|
HF_TOKEN=
|
||||||
38
gpu/self_hosted/.gitignore
vendored
Normal file
38
gpu/self_hosted/.gitignore
vendored
Normal file
@@ -0,0 +1,38 @@
|
|||||||
|
cache/
|
||||||
|
|
||||||
|
# OS / Editor
|
||||||
|
.DS_Store
|
||||||
|
.vscode/
|
||||||
|
.idea/
|
||||||
|
|
||||||
|
# Python
|
||||||
|
__pycache__/
|
||||||
|
*.py[cod]
|
||||||
|
*$py.class
|
||||||
|
|
||||||
|
# Env and secrets
|
||||||
|
.env
|
||||||
|
*.env
|
||||||
|
*.secret
|
||||||
|
HF_TOKEN
|
||||||
|
REFLECTOR_GPU_APIKEY
|
||||||
|
|
||||||
|
# Virtual env / uv
|
||||||
|
.venv/
|
||||||
|
venv/
|
||||||
|
ENV/
|
||||||
|
uv/
|
||||||
|
|
||||||
|
# Build / dist
|
||||||
|
build/
|
||||||
|
dist/
|
||||||
|
.eggs/
|
||||||
|
*.egg-info/
|
||||||
|
|
||||||
|
# Coverage / test
|
||||||
|
.pytest_cache/
|
||||||
|
.coverage*
|
||||||
|
htmlcov/
|
||||||
|
|
||||||
|
# Logs
|
||||||
|
*.log
|
||||||
46
gpu/self_hosted/Dockerfile
Normal file
46
gpu/self_hosted/Dockerfile
Normal file
@@ -0,0 +1,46 @@
|
|||||||
|
FROM python:3.12-slim
|
||||||
|
|
||||||
|
ENV PYTHONUNBUFFERED=1 \
|
||||||
|
UV_LINK_MODE=copy \
|
||||||
|
UV_NO_CACHE=1
|
||||||
|
|
||||||
|
WORKDIR /tmp
|
||||||
|
RUN apt-get update \
|
||||||
|
&& apt-get install -y \
|
||||||
|
ffmpeg \
|
||||||
|
curl \
|
||||||
|
ca-certificates \
|
||||||
|
gnupg \
|
||||||
|
wget \
|
||||||
|
&& apt-get clean
|
||||||
|
# Add NVIDIA CUDA repo for Debian 12 (bookworm) and install cuDNN 9 for CUDA 12
|
||||||
|
ADD https://developer.download.nvidia.com/compute/cuda/repos/debian12/x86_64/cuda-keyring_1.1-1_all.deb /cuda-keyring.deb
|
||||||
|
RUN dpkg -i /cuda-keyring.deb \
|
||||||
|
&& rm /cuda-keyring.deb \
|
||||||
|
&& apt-get update \
|
||||||
|
&& apt-get install -y --no-install-recommends \
|
||||||
|
cuda-cudart-12-6 \
|
||||||
|
libcublas-12-6 \
|
||||||
|
libcudnn9-cuda-12 \
|
||||||
|
libcudnn9-dev-cuda-12 \
|
||||||
|
&& apt-get clean \
|
||||||
|
&& rm -rf /var/lib/apt/lists/*
|
||||||
|
ADD https://astral.sh/uv/install.sh /uv-installer.sh
|
||||||
|
RUN sh /uv-installer.sh && rm /uv-installer.sh
|
||||||
|
ENV PATH="/root/.local/bin/:$PATH"
|
||||||
|
ENV LD_LIBRARY_PATH="/usr/local/cuda/lib64:/usr/lib/x86_64-linux-gnu:$LD_LIBRARY_PATH"
|
||||||
|
|
||||||
|
RUN mkdir -p /app
|
||||||
|
WORKDIR /app
|
||||||
|
COPY pyproject.toml uv.lock /app/
|
||||||
|
|
||||||
|
|
||||||
|
COPY ./app /app/app
|
||||||
|
COPY ./main.py /app/
|
||||||
|
COPY ./runserver.sh /app/
|
||||||
|
|
||||||
|
EXPOSE 8000
|
||||||
|
|
||||||
|
CMD ["sh", "/app/runserver.sh"]
|
||||||
|
|
||||||
|
|
||||||
73
gpu/self_hosted/README.md
Normal file
73
gpu/self_hosted/README.md
Normal file
@@ -0,0 +1,73 @@
|
|||||||
|
# Self-hosted Model API
|
||||||
|
|
||||||
|
Run transcription, translation, and diarization services compatible with Reflector's GPU Model API. Works on CPU or GPU.
|
||||||
|
|
||||||
|
Environment variables
|
||||||
|
|
||||||
|
- REFLECTOR_GPU_APIKEY: Optional Bearer token. If unset, auth is disabled.
|
||||||
|
- HF_TOKEN: Optional. Required for diarization to download pyannote pipelines
|
||||||
|
|
||||||
|
Requirements
|
||||||
|
|
||||||
|
- FFmpeg must be installed and on PATH (used for URL-based and segmented transcription)
|
||||||
|
- Python 3.12+
|
||||||
|
- NVIDIA GPU optional. If available, it will be used automatically
|
||||||
|
|
||||||
|
Local run
|
||||||
|
Set env vars in self_hosted/.env file
|
||||||
|
uv sync
|
||||||
|
|
||||||
|
uv run uvicorn main:app --host 0.0.0.0 --port 8000
|
||||||
|
|
||||||
|
Authentication
|
||||||
|
|
||||||
|
- If REFLECTOR_GPU_APIKEY is set, include header: Authorization: Bearer <key>
|
||||||
|
|
||||||
|
Endpoints
|
||||||
|
|
||||||
|
- POST /v1/audio/transcriptions
|
||||||
|
|
||||||
|
- multipart/form-data
|
||||||
|
- fields: file (single file) OR files[] (multiple files), language, batch (true/false)
|
||||||
|
- response: single { text, words, filename } or { results: [ ... ] }
|
||||||
|
|
||||||
|
- POST /v1/audio/transcriptions-from-url
|
||||||
|
|
||||||
|
- application/json
|
||||||
|
- body: { audio_file_url, language, timestamp_offset }
|
||||||
|
- response: { text, words }
|
||||||
|
|
||||||
|
- POST /translate
|
||||||
|
|
||||||
|
- text: query parameter
|
||||||
|
- body (application/json): { source_language, target_language }
|
||||||
|
- response: { text: { <src>: original, <tgt>: translated } }
|
||||||
|
|
||||||
|
- POST /diarize
|
||||||
|
- query parameters: audio_file_url, timestamp (optional)
|
||||||
|
- requires HF_TOKEN to be set (for pyannote)
|
||||||
|
- response: { diarization: [ { start, end, speaker } ] }
|
||||||
|
|
||||||
|
OpenAPI docs
|
||||||
|
|
||||||
|
- Visit /docs when the server is running
|
||||||
|
|
||||||
|
Docker
|
||||||
|
|
||||||
|
- Not yet provided in this directory. A Dockerfile will be added later. For now, use Local run above
|
||||||
|
|
||||||
|
Conformance tests
|
||||||
|
|
||||||
|
# From this directory
|
||||||
|
|
||||||
|
TRANSCRIPT_URL=http://localhost:8000 \
|
||||||
|
TRANSCRIPT_API_KEY=dev-key \
|
||||||
|
uv run -m pytest -m model_api --no-cov ../../server/tests/test_model_api_transcript.py
|
||||||
|
|
||||||
|
TRANSLATION_URL=http://localhost:8000 \
|
||||||
|
TRANSLATION_API_KEY=dev-key \
|
||||||
|
uv run -m pytest -m model_api --no-cov ../../server/tests/test_model_api_translation.py
|
||||||
|
|
||||||
|
DIARIZATION_URL=http://localhost:8000 \
|
||||||
|
DIARIZATION_API_KEY=dev-key \
|
||||||
|
uv run -m pytest -m model_api --no-cov ../../server/tests/test_model_api_diarization.py
|
||||||
19
gpu/self_hosted/app/auth.py
Normal file
19
gpu/self_hosted/app/auth.py
Normal file
@@ -0,0 +1,19 @@
|
|||||||
|
import os
|
||||||
|
|
||||||
|
from fastapi import Depends, HTTPException, status
|
||||||
|
from fastapi.security import OAuth2PasswordBearer
|
||||||
|
|
||||||
|
oauth2_scheme = OAuth2PasswordBearer(tokenUrl="token")
|
||||||
|
|
||||||
|
|
||||||
|
def apikey_auth(apikey: str = Depends(oauth2_scheme)):
|
||||||
|
required_key = os.environ.get("REFLECTOR_GPU_APIKEY")
|
||||||
|
if not required_key:
|
||||||
|
return
|
||||||
|
if apikey == required_key:
|
||||||
|
return
|
||||||
|
raise HTTPException(
|
||||||
|
status_code=status.HTTP_401_UNAUTHORIZED,
|
||||||
|
detail="Invalid API key",
|
||||||
|
headers={"WWW-Authenticate": "Bearer"},
|
||||||
|
)
|
||||||
12
gpu/self_hosted/app/config.py
Normal file
12
gpu/self_hosted/app/config.py
Normal file
@@ -0,0 +1,12 @@
|
|||||||
|
from pathlib import Path
|
||||||
|
|
||||||
|
SUPPORTED_FILE_EXTENSIONS = ["mp3", "mp4", "mpeg", "mpga", "m4a", "wav", "webm"]
|
||||||
|
SAMPLE_RATE = 16000
|
||||||
|
VAD_CONFIG = {
|
||||||
|
"batch_max_duration": 30.0,
|
||||||
|
"silence_padding": 0.5,
|
||||||
|
"window_size": 512,
|
||||||
|
}
|
||||||
|
|
||||||
|
# App-level paths
|
||||||
|
UPLOADS_PATH = Path("/tmp/whisper-uploads")
|
||||||
30
gpu/self_hosted/app/factory.py
Normal file
30
gpu/self_hosted/app/factory.py
Normal file
@@ -0,0 +1,30 @@
|
|||||||
|
from contextlib import asynccontextmanager
|
||||||
|
|
||||||
|
from fastapi import FastAPI
|
||||||
|
|
||||||
|
from .routers.diarization import router as diarization_router
|
||||||
|
from .routers.transcription import router as transcription_router
|
||||||
|
from .routers.translation import router as translation_router
|
||||||
|
from .services.transcriber import WhisperService
|
||||||
|
from .services.diarizer import PyannoteDiarizationService
|
||||||
|
from .utils import ensure_dirs
|
||||||
|
|
||||||
|
|
||||||
|
@asynccontextmanager
|
||||||
|
async def lifespan(app: FastAPI):
|
||||||
|
ensure_dirs()
|
||||||
|
whisper_service = WhisperService()
|
||||||
|
whisper_service.load()
|
||||||
|
app.state.whisper = whisper_service
|
||||||
|
diarization_service = PyannoteDiarizationService()
|
||||||
|
diarization_service.load()
|
||||||
|
app.state.diarizer = diarization_service
|
||||||
|
yield
|
||||||
|
|
||||||
|
|
||||||
|
def create_app() -> FastAPI:
|
||||||
|
app = FastAPI(lifespan=lifespan)
|
||||||
|
app.include_router(transcription_router)
|
||||||
|
app.include_router(translation_router)
|
||||||
|
app.include_router(diarization_router)
|
||||||
|
return app
|
||||||
30
gpu/self_hosted/app/routers/diarization.py
Normal file
30
gpu/self_hosted/app/routers/diarization.py
Normal file
@@ -0,0 +1,30 @@
|
|||||||
|
from typing import List
|
||||||
|
|
||||||
|
from fastapi import APIRouter, Depends, Request
|
||||||
|
from pydantic import BaseModel
|
||||||
|
|
||||||
|
from ..auth import apikey_auth
|
||||||
|
from ..services.diarizer import PyannoteDiarizationService
|
||||||
|
from ..utils import download_audio_file
|
||||||
|
|
||||||
|
router = APIRouter(tags=["diarization"])
|
||||||
|
|
||||||
|
|
||||||
|
class DiarizationSegment(BaseModel):
|
||||||
|
start: float
|
||||||
|
end: float
|
||||||
|
speaker: int
|
||||||
|
|
||||||
|
|
||||||
|
class DiarizationResponse(BaseModel):
|
||||||
|
diarization: List[DiarizationSegment]
|
||||||
|
|
||||||
|
|
||||||
|
@router.post(
|
||||||
|
"/diarize", dependencies=[Depends(apikey_auth)], response_model=DiarizationResponse
|
||||||
|
)
|
||||||
|
def diarize(request: Request, audio_file_url: str, timestamp: float = 0.0):
|
||||||
|
with download_audio_file(audio_file_url) as (file_path, _ext):
|
||||||
|
file_path = str(file_path)
|
||||||
|
diarizer: PyannoteDiarizationService = request.app.state.diarizer
|
||||||
|
return diarizer.diarize_file(file_path, timestamp=timestamp)
|
||||||
109
gpu/self_hosted/app/routers/transcription.py
Normal file
109
gpu/self_hosted/app/routers/transcription.py
Normal file
@@ -0,0 +1,109 @@
|
|||||||
|
import uuid
|
||||||
|
from typing import Optional, Union
|
||||||
|
|
||||||
|
from fastapi import APIRouter, Body, Depends, Form, HTTPException, Request, UploadFile
|
||||||
|
from pydantic import BaseModel
|
||||||
|
from pathlib import Path
|
||||||
|
from ..auth import apikey_auth
|
||||||
|
from ..config import SUPPORTED_FILE_EXTENSIONS, UPLOADS_PATH
|
||||||
|
from ..services.transcriber import MODEL_NAME
|
||||||
|
from ..utils import cleanup_uploaded_files, download_audio_file
|
||||||
|
|
||||||
|
router = APIRouter(prefix="/v1/audio", tags=["transcription"])
|
||||||
|
|
||||||
|
|
||||||
|
class WordTiming(BaseModel):
|
||||||
|
word: str
|
||||||
|
start: float
|
||||||
|
end: float
|
||||||
|
|
||||||
|
|
||||||
|
class TranscriptResult(BaseModel):
|
||||||
|
text: str
|
||||||
|
words: list[WordTiming]
|
||||||
|
filename: Optional[str] = None
|
||||||
|
|
||||||
|
|
||||||
|
class TranscriptBatchResponse(BaseModel):
|
||||||
|
results: list[TranscriptResult]
|
||||||
|
|
||||||
|
|
||||||
|
@router.post(
|
||||||
|
"/transcriptions",
|
||||||
|
dependencies=[Depends(apikey_auth)],
|
||||||
|
response_model=Union[TranscriptResult, TranscriptBatchResponse],
|
||||||
|
)
|
||||||
|
def transcribe(
|
||||||
|
request: Request,
|
||||||
|
file: UploadFile = None,
|
||||||
|
files: list[UploadFile] | None = None,
|
||||||
|
model: str = Form(MODEL_NAME),
|
||||||
|
language: str = Form("en"),
|
||||||
|
batch: bool = Form(False),
|
||||||
|
):
|
||||||
|
service = request.app.state.whisper
|
||||||
|
if not file and not files:
|
||||||
|
raise HTTPException(
|
||||||
|
status_code=400, detail="Either 'file' or 'files' parameter is required"
|
||||||
|
)
|
||||||
|
if batch and not files:
|
||||||
|
raise HTTPException(
|
||||||
|
status_code=400, detail="Batch transcription requires 'files'"
|
||||||
|
)
|
||||||
|
|
||||||
|
upload_files = [file] if file else files
|
||||||
|
|
||||||
|
uploaded_paths: list[Path] = []
|
||||||
|
with cleanup_uploaded_files(uploaded_paths):
|
||||||
|
for upload_file in upload_files:
|
||||||
|
audio_suffix = upload_file.filename.split(".")[-1].lower()
|
||||||
|
if audio_suffix not in SUPPORTED_FILE_EXTENSIONS:
|
||||||
|
raise HTTPException(
|
||||||
|
status_code=400,
|
||||||
|
detail=(
|
||||||
|
f"Unsupported audio format. Supported extensions: {', '.join(SUPPORTED_FILE_EXTENSIONS)}"
|
||||||
|
),
|
||||||
|
)
|
||||||
|
unique_filename = f"{uuid.uuid4()}.{audio_suffix}"
|
||||||
|
file_path = UPLOADS_PATH / unique_filename
|
||||||
|
with open(file_path, "wb") as f:
|
||||||
|
content = upload_file.file.read()
|
||||||
|
f.write(content)
|
||||||
|
uploaded_paths.append(file_path)
|
||||||
|
|
||||||
|
if batch and len(upload_files) > 1:
|
||||||
|
results = []
|
||||||
|
for path in uploaded_paths:
|
||||||
|
result = service.transcribe_file(str(path), language=language)
|
||||||
|
result["filename"] = path.name
|
||||||
|
results.append(result)
|
||||||
|
return {"results": results}
|
||||||
|
|
||||||
|
results = []
|
||||||
|
for path in uploaded_paths:
|
||||||
|
result = service.transcribe_file(str(path), language=language)
|
||||||
|
result["filename"] = path.name
|
||||||
|
results.append(result)
|
||||||
|
|
||||||
|
return {"results": results} if len(results) > 1 else results[0]
|
||||||
|
|
||||||
|
|
||||||
|
@router.post(
|
||||||
|
"/transcriptions-from-url",
|
||||||
|
dependencies=[Depends(apikey_auth)],
|
||||||
|
response_model=TranscriptResult,
|
||||||
|
)
|
||||||
|
def transcribe_from_url(
|
||||||
|
request: Request,
|
||||||
|
audio_file_url: str = Body(..., description="URL of the audio file to transcribe"),
|
||||||
|
model: str = Body(MODEL_NAME),
|
||||||
|
language: str = Body("en"),
|
||||||
|
timestamp_offset: float = Body(0.0),
|
||||||
|
):
|
||||||
|
service = request.app.state.whisper
|
||||||
|
with download_audio_file(audio_file_url) as (file_path, _ext):
|
||||||
|
file_path = str(file_path)
|
||||||
|
result = service.transcribe_vad_url_segment(
|
||||||
|
file_path=file_path, timestamp_offset=timestamp_offset, language=language
|
||||||
|
)
|
||||||
|
return result
|
||||||
28
gpu/self_hosted/app/routers/translation.py
Normal file
28
gpu/self_hosted/app/routers/translation.py
Normal file
@@ -0,0 +1,28 @@
|
|||||||
|
from typing import Dict
|
||||||
|
|
||||||
|
from fastapi import APIRouter, Body, Depends
|
||||||
|
from pydantic import BaseModel
|
||||||
|
|
||||||
|
from ..auth import apikey_auth
|
||||||
|
from ..services.translator import TextTranslatorService
|
||||||
|
|
||||||
|
router = APIRouter(tags=["translation"])
|
||||||
|
|
||||||
|
translator = TextTranslatorService()
|
||||||
|
|
||||||
|
|
||||||
|
class TranslationResponse(BaseModel):
|
||||||
|
text: Dict[str, str]
|
||||||
|
|
||||||
|
|
||||||
|
@router.post(
|
||||||
|
"/translate",
|
||||||
|
dependencies=[Depends(apikey_auth)],
|
||||||
|
response_model=TranslationResponse,
|
||||||
|
)
|
||||||
|
def translate(
|
||||||
|
text: str,
|
||||||
|
source_language: str = Body("en"),
|
||||||
|
target_language: str = Body("fr"),
|
||||||
|
):
|
||||||
|
return translator.translate(text, source_language, target_language)
|
||||||
42
gpu/self_hosted/app/services/diarizer.py
Normal file
42
gpu/self_hosted/app/services/diarizer.py
Normal file
@@ -0,0 +1,42 @@
|
|||||||
|
import os
|
||||||
|
import threading
|
||||||
|
|
||||||
|
import torch
|
||||||
|
import torchaudio
|
||||||
|
from pyannote.audio import Pipeline
|
||||||
|
|
||||||
|
|
||||||
|
class PyannoteDiarizationService:
|
||||||
|
def __init__(self):
|
||||||
|
self._pipeline = None
|
||||||
|
self._device = "cpu"
|
||||||
|
self._lock = threading.Lock()
|
||||||
|
|
||||||
|
def load(self):
|
||||||
|
self._device = "cuda" if torch.cuda.is_available() else "cpu"
|
||||||
|
self._pipeline = Pipeline.from_pretrained(
|
||||||
|
"pyannote/speaker-diarization-3.1",
|
||||||
|
use_auth_token=os.environ.get("HF_TOKEN"),
|
||||||
|
)
|
||||||
|
self._pipeline.to(torch.device(self._device))
|
||||||
|
|
||||||
|
def diarize_file(self, file_path: str, timestamp: float = 0.0) -> dict:
|
||||||
|
if self._pipeline is None:
|
||||||
|
self.load()
|
||||||
|
waveform, sample_rate = torchaudio.load(file_path)
|
||||||
|
with self._lock:
|
||||||
|
diarization = self._pipeline(
|
||||||
|
{"waveform": waveform, "sample_rate": sample_rate}
|
||||||
|
)
|
||||||
|
words = []
|
||||||
|
for diarization_segment, _, speaker in diarization.itertracks(yield_label=True):
|
||||||
|
words.append(
|
||||||
|
{
|
||||||
|
"start": round(timestamp + diarization_segment.start, 3),
|
||||||
|
"end": round(timestamp + diarization_segment.end, 3),
|
||||||
|
"speaker": int(speaker[-2:])
|
||||||
|
if speaker and speaker[-2:].isdigit()
|
||||||
|
else 0,
|
||||||
|
}
|
||||||
|
)
|
||||||
|
return {"diarization": words}
|
||||||
208
gpu/self_hosted/app/services/transcriber.py
Normal file
208
gpu/self_hosted/app/services/transcriber.py
Normal file
@@ -0,0 +1,208 @@
|
|||||||
|
import os
|
||||||
|
import shutil
|
||||||
|
import subprocess
|
||||||
|
import threading
|
||||||
|
from typing import Generator
|
||||||
|
|
||||||
|
import faster_whisper
|
||||||
|
import librosa
|
||||||
|
import numpy as np
|
||||||
|
import torch
|
||||||
|
from fastapi import HTTPException
|
||||||
|
from silero_vad import VADIterator, load_silero_vad
|
||||||
|
|
||||||
|
from ..config import SAMPLE_RATE, VAD_CONFIG
|
||||||
|
|
||||||
|
# Whisper configuration (service-local defaults)
|
||||||
|
MODEL_NAME = "large-v2"
|
||||||
|
# None delegates compute type to runtime: float16 on CUDA, int8 on CPU
|
||||||
|
MODEL_COMPUTE_TYPE = None
|
||||||
|
MODEL_NUM_WORKERS = 1
|
||||||
|
CACHE_PATH = os.path.join(os.path.expanduser("~"), ".cache", "reflector-whisper")
|
||||||
|
from ..utils import NoStdStreams
|
||||||
|
|
||||||
|
|
||||||
|
class WhisperService:
|
||||||
|
def __init__(self):
|
||||||
|
self.model = None
|
||||||
|
self.device = "cpu"
|
||||||
|
self.lock = threading.Lock()
|
||||||
|
|
||||||
|
def load(self):
|
||||||
|
self.device = "cuda" if torch.cuda.is_available() else "cpu"
|
||||||
|
compute_type = MODEL_COMPUTE_TYPE or (
|
||||||
|
"float16" if self.device == "cuda" else "int8"
|
||||||
|
)
|
||||||
|
self.model = faster_whisper.WhisperModel(
|
||||||
|
MODEL_NAME,
|
||||||
|
device=self.device,
|
||||||
|
compute_type=compute_type,
|
||||||
|
num_workers=MODEL_NUM_WORKERS,
|
||||||
|
download_root=CACHE_PATH,
|
||||||
|
)
|
||||||
|
|
||||||
|
def pad_audio(self, audio_array, sample_rate: int = SAMPLE_RATE):
|
||||||
|
audio_duration = len(audio_array) / sample_rate
|
||||||
|
if audio_duration < VAD_CONFIG["silence_padding"]:
|
||||||
|
silence_samples = int(sample_rate * VAD_CONFIG["silence_padding"])
|
||||||
|
silence = np.zeros(silence_samples, dtype=np.float32)
|
||||||
|
return np.concatenate([audio_array, silence])
|
||||||
|
return audio_array
|
||||||
|
|
||||||
|
def enforce_word_timing_constraints(self, words: list[dict]) -> list[dict]:
|
||||||
|
if len(words) <= 1:
|
||||||
|
return words
|
||||||
|
enforced: list[dict] = []
|
||||||
|
for i, word in enumerate(words):
|
||||||
|
current = dict(word)
|
||||||
|
if i < len(words) - 1:
|
||||||
|
next_start = words[i + 1]["start"]
|
||||||
|
if current["end"] > next_start:
|
||||||
|
current["end"] = next_start
|
||||||
|
enforced.append(current)
|
||||||
|
return enforced
|
||||||
|
|
||||||
|
def transcribe_file(self, file_path: str, language: str = "en") -> dict:
|
||||||
|
input_for_model: str | "object" = file_path
|
||||||
|
try:
|
||||||
|
audio_array, _sample_rate = librosa.load(
|
||||||
|
file_path, sr=SAMPLE_RATE, mono=True
|
||||||
|
)
|
||||||
|
if len(audio_array) / float(SAMPLE_RATE) < VAD_CONFIG["silence_padding"]:
|
||||||
|
input_for_model = self.pad_audio(audio_array, SAMPLE_RATE)
|
||||||
|
except Exception:
|
||||||
|
pass
|
||||||
|
|
||||||
|
with self.lock:
|
||||||
|
with NoStdStreams():
|
||||||
|
segments, _ = self.model.transcribe(
|
||||||
|
input_for_model,
|
||||||
|
language=language,
|
||||||
|
beam_size=5,
|
||||||
|
word_timestamps=True,
|
||||||
|
vad_filter=True,
|
||||||
|
vad_parameters={"min_silence_duration_ms": 500},
|
||||||
|
)
|
||||||
|
|
||||||
|
segments = list(segments)
|
||||||
|
text = "".join(segment.text for segment in segments).strip()
|
||||||
|
words = [
|
||||||
|
{
|
||||||
|
"word": word.word,
|
||||||
|
"start": round(float(word.start), 2),
|
||||||
|
"end": round(float(word.end), 2),
|
||||||
|
}
|
||||||
|
for segment in segments
|
||||||
|
for word in segment.words
|
||||||
|
]
|
||||||
|
words = self.enforce_word_timing_constraints(words)
|
||||||
|
return {"text": text, "words": words}
|
||||||
|
|
||||||
|
def transcribe_vad_url_segment(
|
||||||
|
self, file_path: str, timestamp_offset: float = 0.0, language: str = "en"
|
||||||
|
) -> dict:
|
||||||
|
def load_audio_via_ffmpeg(input_path: str, sample_rate: int) -> np.ndarray:
|
||||||
|
ffmpeg_bin = shutil.which("ffmpeg") or "ffmpeg"
|
||||||
|
cmd = [
|
||||||
|
ffmpeg_bin,
|
||||||
|
"-nostdin",
|
||||||
|
"-threads",
|
||||||
|
"1",
|
||||||
|
"-i",
|
||||||
|
input_path,
|
||||||
|
"-f",
|
||||||
|
"f32le",
|
||||||
|
"-acodec",
|
||||||
|
"pcm_f32le",
|
||||||
|
"-ac",
|
||||||
|
"1",
|
||||||
|
"-ar",
|
||||||
|
str(sample_rate),
|
||||||
|
"pipe:1",
|
||||||
|
]
|
||||||
|
try:
|
||||||
|
proc = subprocess.run(
|
||||||
|
cmd, stdout=subprocess.PIPE, stderr=subprocess.PIPE, check=True
|
||||||
|
)
|
||||||
|
except Exception as e:
|
||||||
|
raise HTTPException(status_code=400, detail=f"ffmpeg failed: {e}")
|
||||||
|
audio = np.frombuffer(proc.stdout, dtype=np.float32)
|
||||||
|
return audio
|
||||||
|
|
||||||
|
def vad_segments(
|
||||||
|
audio_array,
|
||||||
|
sample_rate: int = SAMPLE_RATE,
|
||||||
|
window_size: int = VAD_CONFIG["window_size"],
|
||||||
|
) -> Generator[tuple[float, float], None, None]:
|
||||||
|
vad_model = load_silero_vad(onnx=False)
|
||||||
|
iterator = VADIterator(vad_model, sampling_rate=sample_rate)
|
||||||
|
start = None
|
||||||
|
for i in range(0, len(audio_array), window_size):
|
||||||
|
chunk = audio_array[i : i + window_size]
|
||||||
|
if len(chunk) < window_size:
|
||||||
|
chunk = np.pad(
|
||||||
|
chunk, (0, window_size - len(chunk)), mode="constant"
|
||||||
|
)
|
||||||
|
speech = iterator(chunk)
|
||||||
|
if not speech:
|
||||||
|
continue
|
||||||
|
if "start" in speech:
|
||||||
|
start = speech["start"]
|
||||||
|
continue
|
||||||
|
if "end" in speech and start is not None:
|
||||||
|
end = speech["end"]
|
||||||
|
yield (start / float(SAMPLE_RATE), end / float(SAMPLE_RATE))
|
||||||
|
start = None
|
||||||
|
iterator.reset_states()
|
||||||
|
|
||||||
|
audio_array = load_audio_via_ffmpeg(file_path, SAMPLE_RATE)
|
||||||
|
|
||||||
|
merged_batches: list[tuple[float, float]] = []
|
||||||
|
batch_start = None
|
||||||
|
batch_end = None
|
||||||
|
max_duration = VAD_CONFIG["batch_max_duration"]
|
||||||
|
for seg_start, seg_end in vad_segments(audio_array):
|
||||||
|
if batch_start is None:
|
||||||
|
batch_start, batch_end = seg_start, seg_end
|
||||||
|
continue
|
||||||
|
if seg_end - batch_start <= max_duration:
|
||||||
|
batch_end = seg_end
|
||||||
|
else:
|
||||||
|
merged_batches.append((batch_start, batch_end))
|
||||||
|
batch_start, batch_end = seg_start, seg_end
|
||||||
|
if batch_start is not None and batch_end is not None:
|
||||||
|
merged_batches.append((batch_start, batch_end))
|
||||||
|
|
||||||
|
all_text = []
|
||||||
|
all_words = []
|
||||||
|
for start_time, end_time in merged_batches:
|
||||||
|
s_idx = int(start_time * SAMPLE_RATE)
|
||||||
|
e_idx = int(end_time * SAMPLE_RATE)
|
||||||
|
segment = audio_array[s_idx:e_idx]
|
||||||
|
segment = self.pad_audio(segment, SAMPLE_RATE)
|
||||||
|
with self.lock:
|
||||||
|
segments, _ = self.model.transcribe(
|
||||||
|
segment,
|
||||||
|
language=language,
|
||||||
|
beam_size=5,
|
||||||
|
word_timestamps=True,
|
||||||
|
vad_filter=True,
|
||||||
|
vad_parameters={"min_silence_duration_ms": 500},
|
||||||
|
)
|
||||||
|
segments = list(segments)
|
||||||
|
text = "".join(seg.text for seg in segments).strip()
|
||||||
|
words = [
|
||||||
|
{
|
||||||
|
"word": w.word,
|
||||||
|
"start": round(float(w.start) + start_time + timestamp_offset, 2),
|
||||||
|
"end": round(float(w.end) + start_time + timestamp_offset, 2),
|
||||||
|
}
|
||||||
|
for seg in segments
|
||||||
|
for w in seg.words
|
||||||
|
]
|
||||||
|
if text:
|
||||||
|
all_text.append(text)
|
||||||
|
all_words.extend(words)
|
||||||
|
|
||||||
|
all_words = self.enforce_word_timing_constraints(all_words)
|
||||||
|
return {"text": " ".join(all_text), "words": all_words}
|
||||||
44
gpu/self_hosted/app/services/translator.py
Normal file
44
gpu/self_hosted/app/services/translator.py
Normal file
@@ -0,0 +1,44 @@
|
|||||||
|
import threading
|
||||||
|
|
||||||
|
from transformers import MarianMTModel, MarianTokenizer, pipeline
|
||||||
|
|
||||||
|
|
||||||
|
class TextTranslatorService:
|
||||||
|
"""Simple text-to-text translator using HuggingFace MarianMT models.
|
||||||
|
|
||||||
|
This mirrors the modal translator API shape but uses text translation only.
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(self):
|
||||||
|
self._pipeline = None
|
||||||
|
self._lock = threading.Lock()
|
||||||
|
|
||||||
|
def load(self, source_language: str = "en", target_language: str = "fr"):
|
||||||
|
# Pick a default MarianMT model pair if available; fall back to Helsinki-NLP en->fr
|
||||||
|
model_name = self._resolve_model_name(source_language, target_language)
|
||||||
|
tokenizer = MarianTokenizer.from_pretrained(model_name)
|
||||||
|
model = MarianMTModel.from_pretrained(model_name)
|
||||||
|
self._pipeline = pipeline("translation", model=model, tokenizer=tokenizer)
|
||||||
|
|
||||||
|
def _resolve_model_name(self, src: str, tgt: str) -> str:
|
||||||
|
# Minimal mapping; extend as needed
|
||||||
|
pair = (src.lower(), tgt.lower())
|
||||||
|
mapping = {
|
||||||
|
("en", "fr"): "Helsinki-NLP/opus-mt-en-fr",
|
||||||
|
("fr", "en"): "Helsinki-NLP/opus-mt-fr-en",
|
||||||
|
("en", "es"): "Helsinki-NLP/opus-mt-en-es",
|
||||||
|
("es", "en"): "Helsinki-NLP/opus-mt-es-en",
|
||||||
|
("en", "de"): "Helsinki-NLP/opus-mt-en-de",
|
||||||
|
("de", "en"): "Helsinki-NLP/opus-mt-de-en",
|
||||||
|
}
|
||||||
|
return mapping.get(pair, "Helsinki-NLP/opus-mt-en-fr")
|
||||||
|
|
||||||
|
def translate(self, text: str, source_language: str, target_language: str) -> dict:
|
||||||
|
if self._pipeline is None:
|
||||||
|
self.load(source_language, target_language)
|
||||||
|
with self._lock:
|
||||||
|
results = self._pipeline(
|
||||||
|
text, src_lang=source_language, tgt_lang=target_language
|
||||||
|
)
|
||||||
|
translated = results[0]["translation_text"] if results else ""
|
||||||
|
return {"text": {source_language: text, target_language: translated}}
|
||||||
107
gpu/self_hosted/app/utils.py
Normal file
107
gpu/self_hosted/app/utils.py
Normal file
@@ -0,0 +1,107 @@
|
|||||||
|
import logging
|
||||||
|
import os
|
||||||
|
import sys
|
||||||
|
import uuid
|
||||||
|
from contextlib import contextmanager
|
||||||
|
from typing import Mapping
|
||||||
|
from urllib.parse import urlparse
|
||||||
|
from pathlib import Path
|
||||||
|
|
||||||
|
import requests
|
||||||
|
from fastapi import HTTPException
|
||||||
|
|
||||||
|
from .config import SUPPORTED_FILE_EXTENSIONS, UPLOADS_PATH
|
||||||
|
|
||||||
|
logger = logging.getLogger(__name__)
|
||||||
|
|
||||||
|
|
||||||
|
class NoStdStreams:
|
||||||
|
def __init__(self):
|
||||||
|
self.devnull = open(os.devnull, "w")
|
||||||
|
|
||||||
|
def __enter__(self):
|
||||||
|
self._stdout, self._stderr = sys.stdout, sys.stderr
|
||||||
|
self._stdout.flush()
|
||||||
|
self._stderr.flush()
|
||||||
|
sys.stdout, sys.stderr = self.devnull, self.devnull
|
||||||
|
|
||||||
|
def __exit__(self, exc_type, exc_value, traceback):
|
||||||
|
sys.stdout, sys.stderr = self._stdout, self._stderr
|
||||||
|
self.devnull.close()
|
||||||
|
|
||||||
|
|
||||||
|
def ensure_dirs():
|
||||||
|
UPLOADS_PATH.mkdir(parents=True, exist_ok=True)
|
||||||
|
|
||||||
|
|
||||||
|
def detect_audio_format(url: str, headers: Mapping[str, str]) -> str:
|
||||||
|
url_path = urlparse(url).path
|
||||||
|
for ext in SUPPORTED_FILE_EXTENSIONS:
|
||||||
|
if url_path.lower().endswith(f".{ext}"):
|
||||||
|
return ext
|
||||||
|
|
||||||
|
content_type = headers.get("content-type", "").lower()
|
||||||
|
if "audio/mpeg" in content_type or "audio/mp3" in content_type:
|
||||||
|
return "mp3"
|
||||||
|
if "audio/wav" in content_type:
|
||||||
|
return "wav"
|
||||||
|
if "audio/mp4" in content_type:
|
||||||
|
return "mp4"
|
||||||
|
|
||||||
|
raise HTTPException(
|
||||||
|
status_code=400,
|
||||||
|
detail=(
|
||||||
|
f"Unsupported audio format for URL. Supported extensions: {', '.join(SUPPORTED_FILE_EXTENSIONS)}"
|
||||||
|
),
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def download_audio_to_uploads(audio_file_url: str) -> tuple[Path, str]:
|
||||||
|
response = requests.head(audio_file_url, allow_redirects=True)
|
||||||
|
if response.status_code == 404:
|
||||||
|
raise HTTPException(status_code=404, detail="Audio file not found")
|
||||||
|
|
||||||
|
response = requests.get(audio_file_url, allow_redirects=True)
|
||||||
|
response.raise_for_status()
|
||||||
|
|
||||||
|
audio_suffix = detect_audio_format(audio_file_url, response.headers)
|
||||||
|
unique_filename = f"{uuid.uuid4()}.{audio_suffix}"
|
||||||
|
file_path: Path = UPLOADS_PATH / unique_filename
|
||||||
|
|
||||||
|
with open(file_path, "wb") as f:
|
||||||
|
f.write(response.content)
|
||||||
|
|
||||||
|
return file_path, audio_suffix
|
||||||
|
|
||||||
|
|
||||||
|
@contextmanager
|
||||||
|
def download_audio_file(audio_file_url: str):
|
||||||
|
"""Download an audio file to UPLOADS_PATH and remove it after use.
|
||||||
|
|
||||||
|
Yields (file_path: Path, audio_suffix: str).
|
||||||
|
"""
|
||||||
|
file_path, audio_suffix = download_audio_to_uploads(audio_file_url)
|
||||||
|
try:
|
||||||
|
yield file_path, audio_suffix
|
||||||
|
finally:
|
||||||
|
try:
|
||||||
|
file_path.unlink(missing_ok=True)
|
||||||
|
except Exception as e:
|
||||||
|
logger.error("Error deleting temporary file %s: %s", file_path, e)
|
||||||
|
|
||||||
|
|
||||||
|
@contextmanager
|
||||||
|
def cleanup_uploaded_files(file_paths: list[Path]):
|
||||||
|
"""Ensure provided file paths are removed after use.
|
||||||
|
|
||||||
|
The provided list can be populated inside the context; all present entries
|
||||||
|
at exit will be deleted.
|
||||||
|
"""
|
||||||
|
try:
|
||||||
|
yield file_paths
|
||||||
|
finally:
|
||||||
|
for path in list(file_paths):
|
||||||
|
try:
|
||||||
|
path.unlink(missing_ok=True)
|
||||||
|
except Exception as e:
|
||||||
|
logger.error("Error deleting temporary file %s: %s", path, e)
|
||||||
10
gpu/self_hosted/compose.yml
Normal file
10
gpu/self_hosted/compose.yml
Normal file
@@ -0,0 +1,10 @@
|
|||||||
|
services:
|
||||||
|
reflector_gpu:
|
||||||
|
build:
|
||||||
|
context: .
|
||||||
|
ports:
|
||||||
|
- "8000:8000"
|
||||||
|
env_file:
|
||||||
|
- .env
|
||||||
|
volumes:
|
||||||
|
- ./cache:/root/.cache
|
||||||
3
gpu/self_hosted/main.py
Normal file
3
gpu/self_hosted/main.py
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
from app.factory import create_app
|
||||||
|
|
||||||
|
app = create_app()
|
||||||
19
gpu/self_hosted/pyproject.toml
Normal file
19
gpu/self_hosted/pyproject.toml
Normal file
@@ -0,0 +1,19 @@
|
|||||||
|
[project]
|
||||||
|
name = "reflector-gpu"
|
||||||
|
version = "0.1.0"
|
||||||
|
description = "Self-hosted GPU service for speech transcription, diarization, and translation via FastAPI."
|
||||||
|
readme = "README.md"
|
||||||
|
requires-python = ">=3.12"
|
||||||
|
dependencies = [
|
||||||
|
"fastapi[standard]>=0.116.1",
|
||||||
|
"uvicorn[standard]>=0.30.0",
|
||||||
|
"torch>=2.3.0",
|
||||||
|
"faster-whisper>=1.1.0",
|
||||||
|
"librosa==0.10.1",
|
||||||
|
"numpy<2",
|
||||||
|
"silero-vad==5.1.0",
|
||||||
|
"transformers>=4.35.0",
|
||||||
|
"sentencepiece",
|
||||||
|
"pyannote.audio==3.1.0",
|
||||||
|
"torchaudio>=2.3.0",
|
||||||
|
]
|
||||||
17
gpu/self_hosted/runserver.sh
Normal file
17
gpu/self_hosted/runserver.sh
Normal file
@@ -0,0 +1,17 @@
|
|||||||
|
#!/bin/sh
|
||||||
|
set -e
|
||||||
|
|
||||||
|
export PATH="/root/.local/bin:$PATH"
|
||||||
|
cd /app
|
||||||
|
|
||||||
|
# Install Python dependencies at runtime (first run or when FORCE_SYNC=1)
|
||||||
|
if [ ! -d "/app/.venv" ] || [ "$FORCE_SYNC" = "1" ]; then
|
||||||
|
echo "[startup] Installing Python dependencies with uv..."
|
||||||
|
uv sync --compile-bytecode --locked
|
||||||
|
else
|
||||||
|
echo "[startup] Using existing virtual environment at /app/.venv"
|
||||||
|
fi
|
||||||
|
|
||||||
|
exec uv run uvicorn main:app --host 0.0.0.0 --port 8000
|
||||||
|
|
||||||
|
|
||||||
3013
gpu/self_hosted/uv.lock
generated
Normal file
3013
gpu/self_hosted/uv.lock
generated
Normal file
File diff suppressed because it is too large
Load Diff
3
server/.gitignore
vendored
3
server/.gitignore
vendored
@@ -176,7 +176,8 @@ artefacts/
|
|||||||
audio_*.wav
|
audio_*.wav
|
||||||
|
|
||||||
# ignore local database
|
# ignore local database
|
||||||
reflector.sqlite3
|
*.sqlite3
|
||||||
|
*.db
|
||||||
data/
|
data/
|
||||||
|
|
||||||
dump.rdb
|
dump.rdb
|
||||||
|
|||||||
@@ -1,7 +1,8 @@
|
|||||||
FROM python:3.12-slim
|
FROM python:3.12-slim
|
||||||
|
|
||||||
ENV PYTHONUNBUFFERED=1 \
|
ENV PYTHONUNBUFFERED=1 \
|
||||||
UV_LINK_MODE=copy
|
UV_LINK_MODE=copy \
|
||||||
|
UV_NO_CACHE=1
|
||||||
|
|
||||||
# builder install base dependencies
|
# builder install base dependencies
|
||||||
WORKDIR /tmp
|
WORKDIR /tmp
|
||||||
@@ -13,8 +14,8 @@ ENV PATH="/root/.local/bin/:$PATH"
|
|||||||
# install application dependencies
|
# install application dependencies
|
||||||
RUN mkdir -p /app
|
RUN mkdir -p /app
|
||||||
WORKDIR /app
|
WORKDIR /app
|
||||||
COPY pyproject.toml uv.lock /app/
|
COPY pyproject.toml uv.lock README.md /app/
|
||||||
RUN touch README.md && env uv sync --compile-bytecode --locked
|
RUN uv sync --compile-bytecode --locked
|
||||||
|
|
||||||
# pre-download nltk packages
|
# pre-download nltk packages
|
||||||
RUN uv run python -c "import nltk; nltk.download('punkt_tab'); nltk.download('averaged_perceptron_tagger_eng')"
|
RUN uv run python -c "import nltk; nltk.download('punkt_tab'); nltk.download('averaged_perceptron_tagger_eng')"
|
||||||
@@ -26,4 +27,15 @@ COPY migrations /app/migrations
|
|||||||
COPY reflector /app/reflector
|
COPY reflector /app/reflector
|
||||||
WORKDIR /app
|
WORKDIR /app
|
||||||
|
|
||||||
|
# Create symlink for libgomp if it doesn't exist (for ARM64 compatibility)
|
||||||
|
RUN if [ "$(uname -m)" = "aarch64" ] && [ ! -f /usr/lib/libgomp.so.1 ]; then \
|
||||||
|
LIBGOMP_PATH=$(find /app/.venv/lib -path "*/torch.libs/libgomp*.so.*" 2>/dev/null | head -n1); \
|
||||||
|
if [ -n "$LIBGOMP_PATH" ]; then \
|
||||||
|
ln -sf "$LIBGOMP_PATH" /usr/lib/libgomp.so.1; \
|
||||||
|
fi \
|
||||||
|
fi
|
||||||
|
|
||||||
|
# Pre-check just to make sure the image will not fail
|
||||||
|
RUN uv run python -c "import silero_vad.model"
|
||||||
|
|
||||||
CMD ["./runserver.sh"]
|
CMD ["./runserver.sh"]
|
||||||
|
|||||||
@@ -1,3 +1,29 @@
|
|||||||
|
## API Key Management
|
||||||
|
|
||||||
|
### Finding Your User ID
|
||||||
|
|
||||||
|
```bash
|
||||||
|
# Get your OAuth sub (user ID) - requires authentication
|
||||||
|
curl -H "Authorization: Bearer <your_jwt>" http://localhost:1250/v1/me
|
||||||
|
# Returns: {"sub": "your-oauth-sub-here", "email": "...", ...}
|
||||||
|
```
|
||||||
|
|
||||||
|
### Creating API Keys
|
||||||
|
|
||||||
|
```bash
|
||||||
|
curl -X POST http://localhost:1250/v1/user/api-keys \
|
||||||
|
-H "Authorization: Bearer <your_jwt>" \
|
||||||
|
-H "Content-Type: application/json" \
|
||||||
|
-d '{"name": "My API Key"}'
|
||||||
|
```
|
||||||
|
|
||||||
|
### Using API Keys
|
||||||
|
|
||||||
|
```bash
|
||||||
|
# Use X-API-Key header instead of Authorization
|
||||||
|
curl -H "X-API-Key: <your_api_key>" http://localhost:1250/v1/transcripts
|
||||||
|
```
|
||||||
|
|
||||||
## AWS S3/SQS usage clarification
|
## AWS S3/SQS usage clarification
|
||||||
|
|
||||||
Whereby.com uploads recordings directly to our S3 bucket when meetings end.
|
Whereby.com uploads recordings directly to our S3 bucket when meetings end.
|
||||||
@@ -40,3 +66,5 @@ uv run python -c "from reflector.pipelines.main_live_pipeline import task_pipeli
|
|||||||
```bash
|
```bash
|
||||||
uv run python -c "from reflector.pipelines.main_live_pipeline import pipeline_post; pipeline_post(transcript_id='TRANSCRIPT_ID')"
|
uv run python -c "from reflector.pipelines.main_live_pipeline import pipeline_post; pipeline_post(transcript_id='TRANSCRIPT_ID')"
|
||||||
```
|
```
|
||||||
|
|
||||||
|
.
|
||||||
|
|||||||
95
server/docs/data_retention.md
Normal file
95
server/docs/data_retention.md
Normal file
@@ -0,0 +1,95 @@
|
|||||||
|
# Data Retention and Cleanup
|
||||||
|
|
||||||
|
## Overview
|
||||||
|
|
||||||
|
For public instances of Reflector, a data retention policy is automatically enforced to delete anonymous user data after a configurable period (default: 7 days). This ensures compliance with privacy expectations and prevents unbounded storage growth.
|
||||||
|
|
||||||
|
## Configuration
|
||||||
|
|
||||||
|
### Environment Variables
|
||||||
|
|
||||||
|
- `PUBLIC_MODE` (bool): Must be set to `true` to enable automatic cleanup
|
||||||
|
- `PUBLIC_DATA_RETENTION_DAYS` (int): Number of days to retain anonymous data (default: 7)
|
||||||
|
|
||||||
|
### What Gets Deleted
|
||||||
|
|
||||||
|
When data reaches the retention period, the following items are automatically removed:
|
||||||
|
|
||||||
|
1. **Transcripts** from anonymous users (where `user_id` is NULL):
|
||||||
|
- Database records
|
||||||
|
- Local files (audio.wav, audio.mp3, audio.json waveform)
|
||||||
|
- Storage files (cloud storage if configured)
|
||||||
|
|
||||||
|
## Automatic Cleanup
|
||||||
|
|
||||||
|
### Celery Beat Schedule
|
||||||
|
|
||||||
|
When `PUBLIC_MODE=true`, a Celery beat task runs daily at 3 AM to clean up old data:
|
||||||
|
|
||||||
|
```python
|
||||||
|
# Automatically scheduled when PUBLIC_MODE=true
|
||||||
|
"cleanup_old_public_data": {
|
||||||
|
"task": "reflector.worker.cleanup.cleanup_old_public_data",
|
||||||
|
"schedule": crontab(hour=3, minute=0), # Daily at 3 AM
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
### Running the Worker
|
||||||
|
|
||||||
|
Ensure both Celery worker and beat scheduler are running:
|
||||||
|
|
||||||
|
```bash
|
||||||
|
# Start Celery worker
|
||||||
|
uv run celery -A reflector.worker.app worker --loglevel=info
|
||||||
|
|
||||||
|
# Start Celery beat scheduler (in another terminal)
|
||||||
|
uv run celery -A reflector.worker.app beat
|
||||||
|
```
|
||||||
|
|
||||||
|
## Manual Cleanup
|
||||||
|
|
||||||
|
For testing or manual intervention, use the cleanup tool:
|
||||||
|
|
||||||
|
```bash
|
||||||
|
# Delete data older than 7 days (default)
|
||||||
|
uv run python -m reflector.tools.cleanup_old_data
|
||||||
|
|
||||||
|
# Delete data older than 30 days
|
||||||
|
uv run python -m reflector.tools.cleanup_old_data --days 30
|
||||||
|
```
|
||||||
|
|
||||||
|
Note: The manual tool uses the same implementation as the Celery worker task to ensure consistency.
|
||||||
|
|
||||||
|
## Important Notes
|
||||||
|
|
||||||
|
1. **User Data Deletion**: Only anonymous data (where `user_id` is NULL) is deleted. Authenticated user data is preserved.
|
||||||
|
|
||||||
|
2. **Storage Cleanup**: The system properly cleans up both local files and cloud storage when configured.
|
||||||
|
|
||||||
|
3. **Error Handling**: If individual deletions fail, the cleanup continues and logs errors. Failed deletions are reported in the task output.
|
||||||
|
|
||||||
|
4. **Public Instance Only**: The automatic cleanup task only runs when `PUBLIC_MODE=true` to prevent accidental data loss in private deployments.
|
||||||
|
|
||||||
|
## Testing
|
||||||
|
|
||||||
|
Run the cleanup tests:
|
||||||
|
|
||||||
|
```bash
|
||||||
|
uv run pytest tests/test_cleanup.py -v
|
||||||
|
```
|
||||||
|
|
||||||
|
## Monitoring
|
||||||
|
|
||||||
|
Check Celery logs for cleanup task execution:
|
||||||
|
|
||||||
|
```bash
|
||||||
|
# Look for cleanup task logs
|
||||||
|
grep "cleanup_old_public_data" celery.log
|
||||||
|
grep "Starting cleanup of old public data" celery.log
|
||||||
|
```
|
||||||
|
|
||||||
|
Task statistics are logged after each run:
|
||||||
|
- Number of transcripts deleted
|
||||||
|
- Number of meetings deleted
|
||||||
|
- Number of orphaned recordings deleted
|
||||||
|
- Any errors encountered
|
||||||
194
server/docs/gpu/api-transcription.md
Normal file
194
server/docs/gpu/api-transcription.md
Normal file
@@ -0,0 +1,194 @@
|
|||||||
|
## Reflector GPU Transcription API (Specification)
|
||||||
|
|
||||||
|
This document defines the Reflector GPU transcription API that all implementations must adhere to. Current implementations include NVIDIA Parakeet (NeMo) and Whisper (faster-whisper), both deployed on Modal.com. The API surface and response shapes are OpenAI/Whisper-compatible, so clients can switch implementations by changing only the base URL.
|
||||||
|
|
||||||
|
### Base URL and Authentication
|
||||||
|
|
||||||
|
- Example base URLs (Modal web endpoints):
|
||||||
|
|
||||||
|
- Parakeet: `https://<account>--reflector-transcriber-parakeet-web.modal.run`
|
||||||
|
- Whisper: `https://<account>--reflector-transcriber-web.modal.run`
|
||||||
|
|
||||||
|
- All endpoints are served under `/v1` and require a Bearer token:
|
||||||
|
|
||||||
|
```
|
||||||
|
Authorization: Bearer <REFLECTOR_GPU_APIKEY>
|
||||||
|
```
|
||||||
|
|
||||||
|
Note: To switch implementations, deploy the desired variant and point `TRANSCRIPT_URL` to its base URL. The API is identical.
|
||||||
|
|
||||||
|
### Supported file types
|
||||||
|
|
||||||
|
`mp3, mp4, mpeg, mpga, m4a, wav, webm`
|
||||||
|
|
||||||
|
### Models and languages
|
||||||
|
|
||||||
|
- Parakeet (NVIDIA NeMo): default `nvidia/parakeet-tdt-0.6b-v2`
|
||||||
|
- Language support: only `en`. Other languages return HTTP 400.
|
||||||
|
- Whisper (faster-whisper): default `large-v2` (or deployment-specific)
|
||||||
|
- Language support: multilingual (per Whisper model capabilities).
|
||||||
|
|
||||||
|
Note: The `model` parameter is accepted by all implementations for interface parity. Some backends may treat it as informational.
|
||||||
|
|
||||||
|
### Endpoints
|
||||||
|
|
||||||
|
#### POST /v1/audio/transcriptions
|
||||||
|
|
||||||
|
Transcribe one or more uploaded audio files.
|
||||||
|
|
||||||
|
Request: multipart/form-data
|
||||||
|
|
||||||
|
- `file` (File) — optional. Single file to transcribe.
|
||||||
|
- `files` (File[]) — optional. One or more files to transcribe.
|
||||||
|
- `model` (string) — optional. Defaults to the implementation-specific model (see above).
|
||||||
|
- `language` (string) — optional, defaults to `en`.
|
||||||
|
- Parakeet: only `en` is accepted; other values return HTTP 400
|
||||||
|
- Whisper: model-dependent; typically multilingual
|
||||||
|
- `batch` (boolean) — optional, defaults to `false`.
|
||||||
|
|
||||||
|
Notes:
|
||||||
|
|
||||||
|
- Provide either `file` or `files`, not both. If neither is provided, HTTP 400.
|
||||||
|
- `batch` requires `files`; using `batch=true` without `files` returns HTTP 400.
|
||||||
|
- Response shape for multiple files is the same regardless of `batch`.
|
||||||
|
- Files sent to this endpoint are processed in a single pass (no VAD/chunking). This is intended for short clips (roughly ≤ 30s; depends on GPU memory/model). For longer audio, prefer `/v1/audio/transcriptions-from-url` which supports VAD-based chunking.
|
||||||
|
|
||||||
|
Responses
|
||||||
|
|
||||||
|
Single file response:
|
||||||
|
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"text": "transcribed text",
|
||||||
|
"words": [
|
||||||
|
{ "word": "hello", "start": 0.0, "end": 0.5 },
|
||||||
|
{ "word": "world", "start": 0.5, "end": 1.0 }
|
||||||
|
],
|
||||||
|
"filename": "audio.mp3"
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
Multiple files response:
|
||||||
|
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"results": [
|
||||||
|
{"filename": "a1.mp3", "text": "...", "words": [...]},
|
||||||
|
{"filename": "a2.mp3", "text": "...", "words": [...]}]
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
Notes:
|
||||||
|
|
||||||
|
- Word objects always include keys: `word`, `start`, `end`.
|
||||||
|
- Some implementations may include a trailing space in `word` to match Whisper tokenization behavior; clients should trim if needed.
|
||||||
|
|
||||||
|
Example curl (single file):
|
||||||
|
|
||||||
|
```bash
|
||||||
|
curl -X POST \
|
||||||
|
-H "Authorization: Bearer $REFLECTOR_GPU_APIKEY" \
|
||||||
|
-F "file=@/path/to/audio.mp3" \
|
||||||
|
-F "language=en" \
|
||||||
|
"$BASE_URL/v1/audio/transcriptions"
|
||||||
|
```
|
||||||
|
|
||||||
|
Example curl (multiple files, batch):
|
||||||
|
|
||||||
|
```bash
|
||||||
|
curl -X POST \
|
||||||
|
-H "Authorization: Bearer $REFLECTOR_GPU_APIKEY" \
|
||||||
|
-F "files=@/path/a1.mp3" -F "files=@/path/a2.mp3" \
|
||||||
|
-F "batch=true" -F "language=en" \
|
||||||
|
"$BASE_URL/v1/audio/transcriptions"
|
||||||
|
```
|
||||||
|
|
||||||
|
#### POST /v1/audio/transcriptions-from-url
|
||||||
|
|
||||||
|
Transcribe a single remote audio file by URL.
|
||||||
|
|
||||||
|
Request: application/json
|
||||||
|
|
||||||
|
Body parameters:
|
||||||
|
|
||||||
|
- `audio_file_url` (string) — required. URL of the audio file to transcribe.
|
||||||
|
- `model` (string) — optional. Defaults to the implementation-specific model (see above).
|
||||||
|
- `language` (string) — optional, defaults to `en`. Parakeet only accepts `en`.
|
||||||
|
- `timestamp_offset` (number) — optional, defaults to `0.0`. Added to each word's `start`/`end` in the response.
|
||||||
|
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"audio_file_url": "https://example.com/audio.mp3",
|
||||||
|
"model": "nvidia/parakeet-tdt-0.6b-v2",
|
||||||
|
"language": "en",
|
||||||
|
"timestamp_offset": 0.0
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
Response:
|
||||||
|
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"text": "transcribed text",
|
||||||
|
"words": [
|
||||||
|
{ "word": "hello", "start": 10.0, "end": 10.5 },
|
||||||
|
{ "word": "world", "start": 10.5, "end": 11.0 }
|
||||||
|
]
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
Notes:
|
||||||
|
|
||||||
|
- `timestamp_offset` is added to each word’s `start`/`end` in the response.
|
||||||
|
- Implementations may perform VAD-based chunking and batching for long-form audio; word timings are adjusted accordingly.
|
||||||
|
|
||||||
|
Example curl:
|
||||||
|
|
||||||
|
```bash
|
||||||
|
curl -X POST \
|
||||||
|
-H "Authorization: Bearer $REFLECTOR_GPU_APIKEY" \
|
||||||
|
-H "Content-Type: application/json" \
|
||||||
|
-d '{
|
||||||
|
"audio_file_url": "https://example.com/audio.mp3",
|
||||||
|
"language": "en",
|
||||||
|
"timestamp_offset": 0
|
||||||
|
}' \
|
||||||
|
"$BASE_URL/v1/audio/transcriptions-from-url"
|
||||||
|
```
|
||||||
|
|
||||||
|
### Error handling
|
||||||
|
|
||||||
|
- 400 Bad Request
|
||||||
|
- Parakeet: `language` other than `en`
|
||||||
|
- Missing required parameters (`file`/`files` for upload; `audio_file_url` for URL endpoint)
|
||||||
|
- Unsupported file extension
|
||||||
|
- 401 Unauthorized
|
||||||
|
- Missing or invalid Bearer token
|
||||||
|
- 404 Not Found
|
||||||
|
- `audio_file_url` does not exist
|
||||||
|
|
||||||
|
### Implementation details
|
||||||
|
|
||||||
|
- GPUs: A10G for small-file/live, L40S for large-file URL transcription (subject to deployment)
|
||||||
|
- VAD chunking and segment batching; word timings adjusted and overlapping ends constrained
|
||||||
|
- Pads very short segments (< 0.5s) to avoid model crashes on some backends
|
||||||
|
|
||||||
|
### Server configuration (Reflector API)
|
||||||
|
|
||||||
|
Set the Reflector server to use the Modal backend and point `TRANSCRIPT_URL` to your chosen deployment:
|
||||||
|
|
||||||
|
```
|
||||||
|
TRANSCRIPT_BACKEND=modal
|
||||||
|
TRANSCRIPT_URL=https://<account>--reflector-transcriber-parakeet-web.modal.run
|
||||||
|
TRANSCRIPT_MODAL_API_KEY=<REFLECTOR_GPU_APIKEY>
|
||||||
|
```
|
||||||
|
|
||||||
|
### Conformance tests
|
||||||
|
|
||||||
|
Use the pytest-based conformance tests to validate any new implementation (including self-hosted) against this spec:
|
||||||
|
|
||||||
|
```
|
||||||
|
TRANSCRIPT_URL=https://<your-deployment-base> \
|
||||||
|
TRANSCRIPT_MODAL_API_KEY=your-api-key \
|
||||||
|
uv run -m pytest -m model_api --no-cov server/tests/test_model_api_transcript.py
|
||||||
|
```
|
||||||
234
server/docs/video-platforms/README.md
Normal file
234
server/docs/video-platforms/README.md
Normal file
@@ -0,0 +1,234 @@
|
|||||||
|
# Reflector Architecture: Whereby + Daily.co Recording Storage
|
||||||
|
|
||||||
|
## System Overview
|
||||||
|
|
||||||
|
```mermaid
|
||||||
|
graph TB
|
||||||
|
subgraph "Actors"
|
||||||
|
APP[Our App<br/>Reflector]
|
||||||
|
WHEREBY[Whereby Service<br/>External]
|
||||||
|
DAILY[Daily.co Service<br/>External]
|
||||||
|
end
|
||||||
|
|
||||||
|
subgraph "AWS S3 Buckets"
|
||||||
|
TRANSCRIPT_BUCKET[Transcript Bucket<br/>reflector-transcripts<br/>Output: Processed MP3s]
|
||||||
|
WHEREBY_BUCKET[Whereby Bucket<br/>reflector-whereby-recordings<br/>Input: Raw MP4s]
|
||||||
|
DAILY_BUCKET[Daily.co Bucket<br/>reflector-dailyco-recordings<br/>Input: Raw WebM tracks]
|
||||||
|
end
|
||||||
|
|
||||||
|
subgraph "AWS Infrastructure"
|
||||||
|
SQS[SQS Queue<br/>Whereby notifications]
|
||||||
|
end
|
||||||
|
|
||||||
|
subgraph "Database"
|
||||||
|
DB[(PostgreSQL<br/>Recordings, Transcripts, Meetings)]
|
||||||
|
end
|
||||||
|
|
||||||
|
APP -->|Write processed| TRANSCRIPT_BUCKET
|
||||||
|
APP -->|Read/Delete| WHEREBY_BUCKET
|
||||||
|
APP -->|Read/Delete| DAILY_BUCKET
|
||||||
|
APP -->|Poll| SQS
|
||||||
|
APP -->|Store metadata| DB
|
||||||
|
|
||||||
|
WHEREBY -->|Write recordings| WHEREBY_BUCKET
|
||||||
|
WHEREBY_BUCKET -->|S3 Event| SQS
|
||||||
|
WHEREBY -->|Participant webhooks<br/>room.client.joined/left| APP
|
||||||
|
|
||||||
|
DAILY -->|Write recordings| DAILY_BUCKET
|
||||||
|
DAILY -->|Recording webhook<br/>recording.ready-to-download| APP
|
||||||
|
```
|
||||||
|
|
||||||
|
**Note on Webhook vs S3 Event for Recording Processing:**
|
||||||
|
- **Whereby**: Uses S3 Events → SQS for recording availability (S3 as source of truth, no race conditions)
|
||||||
|
- **Daily.co**: Uses webhooks for recording availability (more immediate, built-in reliability)
|
||||||
|
- **Both**: Use webhooks for participant tracking (real-time updates)
|
||||||
|
|
||||||
|
## Credentials & Permissions
|
||||||
|
|
||||||
|
```mermaid
|
||||||
|
graph LR
|
||||||
|
subgraph "Master Credentials"
|
||||||
|
MASTER[TRANSCRIPT_STORAGE_AWS_*<br/>Access Key ID + Secret]
|
||||||
|
end
|
||||||
|
|
||||||
|
subgraph "Whereby Upload Credentials"
|
||||||
|
WHEREBY_CREDS[AWS_WHEREBY_ACCESS_KEY_*<br/>Access Key ID + Secret]
|
||||||
|
end
|
||||||
|
|
||||||
|
subgraph "Daily.co Upload Role"
|
||||||
|
DAILY_ROLE[DAILY_STORAGE_AWS_ROLE_ARN<br/>IAM Role ARN]
|
||||||
|
end
|
||||||
|
|
||||||
|
subgraph "Our App Uses"
|
||||||
|
MASTER -->|Read/Write/Delete| TRANSCRIPT_BUCKET[Transcript Bucket]
|
||||||
|
MASTER -->|Read/Delete| WHEREBY_BUCKET[Whereby Bucket]
|
||||||
|
MASTER -->|Read/Delete| DAILY_BUCKET[Daily.co Bucket]
|
||||||
|
MASTER -->|Poll/Delete| SQS[SQS Queue]
|
||||||
|
end
|
||||||
|
|
||||||
|
subgraph "We Give To Services"
|
||||||
|
WHEREBY_CREDS -->|Passed in API call| WHEREBY_SERVICE[Whereby Service]
|
||||||
|
WHEREBY_SERVICE -->|Write Only| WHEREBY_BUCKET
|
||||||
|
|
||||||
|
DAILY_ROLE -->|Passed in API call| DAILY_SERVICE[Daily.co Service]
|
||||||
|
DAILY_SERVICE -->|Assume Role| DAILY_ROLE
|
||||||
|
DAILY_SERVICE -->|Write Only| DAILY_BUCKET
|
||||||
|
end
|
||||||
|
```
|
||||||
|
|
||||||
|
# Video Platform Recording Integration
|
||||||
|
|
||||||
|
This document explains how Reflector receives and identifies multitrack audio recordings from different video platforms.
|
||||||
|
|
||||||
|
## Platform Comparison
|
||||||
|
|
||||||
|
| Platform | Delivery Method | Track Identification |
|
||||||
|
|----------|----------------|---------------------|
|
||||||
|
| **Daily.co** | Webhook | Explicit track list in payload |
|
||||||
|
| **Whereby** | SQS (S3 notifications) | Single file per notification |
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## Daily.co (Webhook-based)
|
||||||
|
|
||||||
|
Daily.co uses **webhooks** to notify Reflector when recordings are ready.
|
||||||
|
|
||||||
|
### How It Works
|
||||||
|
|
||||||
|
1. **Daily.co sends webhook** when recording is ready
|
||||||
|
- Event type: `recording.ready-to-download`
|
||||||
|
- Endpoint: `/v1/daily/webhook` (`reflector/views/daily.py:46-102`)
|
||||||
|
|
||||||
|
2. **Webhook payload explicitly includes track list**:
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"recording_id": "7443ee0a-dab1-40eb-b316-33d6c0d5ff88",
|
||||||
|
"room_name": "daily-20251020193458",
|
||||||
|
"tracks": [
|
||||||
|
{
|
||||||
|
"type": "audio",
|
||||||
|
"s3Key": "monadical/daily-20251020193458/1760988935484-52f7f48b-fbab-431f-9a50-87b9abfc8255-cam-audio-1760988935922",
|
||||||
|
"size": 831843
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"type": "audio",
|
||||||
|
"s3Key": "monadical/daily-20251020193458/1760988935484-a37c35e3-6f8e-4274-a482-e9d0f102a732-cam-audio-1760988943823",
|
||||||
|
"size": 408438
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"type": "video",
|
||||||
|
"s3Key": "monadical/daily-20251020193458/...-video.webm",
|
||||||
|
"size": 30000000
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
3. **System extracts audio tracks** (`daily.py:211`):
|
||||||
|
```python
|
||||||
|
track_keys = [t.s3Key for t in tracks if t.type == "audio"]
|
||||||
|
```
|
||||||
|
|
||||||
|
4. **Triggers multitrack processing** (`daily.py:213-218`):
|
||||||
|
```python
|
||||||
|
process_multitrack_recording.delay(
|
||||||
|
bucket_name=bucket_name, # reflector-dailyco-local
|
||||||
|
room_name=room_name, # daily-20251020193458
|
||||||
|
recording_id=recording_id, # 7443ee0a-dab1-40eb-b316-33d6c0d5ff88
|
||||||
|
track_keys=track_keys # Only audio s3Keys
|
||||||
|
)
|
||||||
|
```
|
||||||
|
|
||||||
|
### Key Advantage: No Ambiguity
|
||||||
|
|
||||||
|
Even though multiple meetings may share the same S3 bucket/folder (`monadical/`), **there's no ambiguity** because:
|
||||||
|
- Each webhook payload contains the exact `s3Key` list for that specific `recording_id`
|
||||||
|
- No need to scan folders or guess which files belong together
|
||||||
|
- Each track's s3Key includes the room timestamp subfolder (e.g., `daily-20251020193458/`)
|
||||||
|
|
||||||
|
The room name includes timestamp (`daily-20251020193458`) to keep recordings organized, but **the webhook's explicit track list is what prevents mixing files from different meetings**.
|
||||||
|
|
||||||
|
### Track Timeline Extraction
|
||||||
|
|
||||||
|
Daily.co provides timing information in two places:
|
||||||
|
|
||||||
|
**1. PyAV WebM Metadata (current approach)**:
|
||||||
|
```python
|
||||||
|
# Read from WebM container stream metadata
|
||||||
|
stream.start_time = 8.130s # Meeting-relative timing
|
||||||
|
```
|
||||||
|
|
||||||
|
**2. Filename Timestamps (alternative approach, commit 3bae9076)**:
|
||||||
|
```
|
||||||
|
Filename format: {recording_start_ts}-{uuid}-cam-audio-{track_start_ts}.webm
|
||||||
|
Example: 1760988935484-52f7f48b-fbab-431f-9a50-87b9abfc8255-cam-audio-1760988935922.webm
|
||||||
|
|
||||||
|
Parse timestamps:
|
||||||
|
- recording_start_ts: 1760988935484 (Unix ms)
|
||||||
|
- track_start_ts: 1760988935922 (Unix ms)
|
||||||
|
- offset: (1760988935922 - 1760988935484) / 1000 = 0.438s
|
||||||
|
```
|
||||||
|
|
||||||
|
**Time Difference (PyAV vs Filename)**:
|
||||||
|
```
|
||||||
|
Track 0:
|
||||||
|
Filename offset: 438ms
|
||||||
|
PyAV metadata: 229ms
|
||||||
|
Difference: 209ms
|
||||||
|
|
||||||
|
Track 1:
|
||||||
|
Filename offset: 8339ms
|
||||||
|
PyAV metadata: 8130ms
|
||||||
|
Difference: 209ms
|
||||||
|
```
|
||||||
|
|
||||||
|
**Consistent 209ms delta** suggests network/encoding delay between file upload initiation (filename) and actual audio stream start (metadata).
|
||||||
|
|
||||||
|
**Current implementation uses PyAV metadata** because:
|
||||||
|
- More accurate (represents when audio actually started)
|
||||||
|
- Padding BEFORE transcription produces correct Whisper timestamps automatically
|
||||||
|
- No manual offset adjustment needed during transcript merge
|
||||||
|
|
||||||
|
### Why Re-encoding During Padding
|
||||||
|
|
||||||
|
Padding coincidentally involves re-encoding, which is important for Daily.co + Whisper:
|
||||||
|
|
||||||
|
**Problem:** Daily.co skips frames in recordings when microphone is muted or paused
|
||||||
|
- WebM containers have gaps where audio frames should be
|
||||||
|
- Whisper doesn't understand these gaps and produces incorrect timestamps
|
||||||
|
- Example: 5s of audio with 2s muted → file has frames only for 3s, Whisper thinks duration is 3s
|
||||||
|
|
||||||
|
**Solution:** Re-encoding via PyAV filter graph (`adelay` + `aresample`)
|
||||||
|
- Restores missing frames as silence
|
||||||
|
- Produces continuous audio stream without gaps
|
||||||
|
- Whisper now sees correct duration and produces accurate timestamps
|
||||||
|
|
||||||
|
**Why combined with padding:**
|
||||||
|
- Already re-encoding for padding (adding initial silence)
|
||||||
|
- More performant to do both operations in single PyAV pipeline
|
||||||
|
- Padded values needed for mixdown anyway (creating final MP3)
|
||||||
|
|
||||||
|
Implementation: `main_multitrack_pipeline.py:_apply_audio_padding_streaming()`
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
## Whereby (SQS-based)
|
||||||
|
|
||||||
|
Whereby uses **AWS SQS** (via S3 notifications) to notify Reflector when files are uploaded.
|
||||||
|
|
||||||
|
### How It Works
|
||||||
|
|
||||||
|
1. **Whereby uploads recording** to S3
|
||||||
|
2. **S3 sends notification** to SQS queue (one notification per file)
|
||||||
|
3. **Reflector polls SQS queue** (`worker/process.py:process_messages()`)
|
||||||
|
4. **System processes single file** (`worker/process.py:process_recording()`)
|
||||||
|
|
||||||
|
### Key Difference from Daily.co
|
||||||
|
|
||||||
|
**Whereby (SQS):** System receives S3 notification "file X was created" - only knows about one file at a time, would need to scan folder to find related files
|
||||||
|
|
||||||
|
**Daily.co (Webhook):** Daily explicitly tells system which files belong together in the webhook payload
|
||||||
|
|
||||||
|
---
|
||||||
|
|
||||||
|
|
||||||
233
server/docs/webhook.md
Normal file
233
server/docs/webhook.md
Normal file
@@ -0,0 +1,233 @@
|
|||||||
|
# Reflector Webhook Documentation
|
||||||
|
|
||||||
|
## Overview
|
||||||
|
|
||||||
|
Reflector supports webhook notifications to notify external systems when transcript processing is completed. Webhooks can be configured per room and are triggered automatically after a transcript is successfully processed.
|
||||||
|
|
||||||
|
## Configuration
|
||||||
|
|
||||||
|
Webhooks are configured at the room level with two fields:
|
||||||
|
- `webhook_url`: The HTTPS endpoint to receive webhook notifications
|
||||||
|
- `webhook_secret`: Optional secret key for HMAC signature verification (auto-generated if not provided)
|
||||||
|
|
||||||
|
## Events
|
||||||
|
|
||||||
|
### `transcript.completed`
|
||||||
|
|
||||||
|
Triggered when a transcript has been fully processed, including transcription, diarization, summarization, topic detection and calendar event integration.
|
||||||
|
|
||||||
|
### `test`
|
||||||
|
|
||||||
|
A test event that can be triggered manually to verify webhook configuration.
|
||||||
|
|
||||||
|
## Webhook Request Format
|
||||||
|
|
||||||
|
### Headers
|
||||||
|
|
||||||
|
All webhook requests include the following headers:
|
||||||
|
|
||||||
|
| Header | Description | Example |
|
||||||
|
|--------|-------------|---------|
|
||||||
|
| `Content-Type` | Always `application/json` | `application/json` |
|
||||||
|
| `User-Agent` | Identifies Reflector as the source | `Reflector-Webhook/1.0` |
|
||||||
|
| `X-Webhook-Event` | The event type | `transcript.completed` or `test` |
|
||||||
|
| `X-Webhook-Retry` | Current retry attempt number | `0`, `1`, `2`... |
|
||||||
|
| `X-Webhook-Signature` | HMAC signature (if secret configured) | `t=1735306800,v1=abc123...` |
|
||||||
|
|
||||||
|
### Signature Verification
|
||||||
|
|
||||||
|
If a webhook secret is configured, Reflector includes an HMAC-SHA256 signature in the `X-Webhook-Signature` header to verify the webhook authenticity.
|
||||||
|
|
||||||
|
The signature format is: `t={timestamp},v1={signature}`
|
||||||
|
|
||||||
|
To verify the signature:
|
||||||
|
1. Extract the timestamp and signature from the header
|
||||||
|
2. Create the signed payload: `{timestamp}.{request_body}`
|
||||||
|
3. Compute HMAC-SHA256 of the signed payload using your webhook secret
|
||||||
|
4. Compare the computed signature with the received signature
|
||||||
|
|
||||||
|
Example verification (Python):
|
||||||
|
```python
|
||||||
|
import hmac
|
||||||
|
import hashlib
|
||||||
|
|
||||||
|
def verify_webhook_signature(payload: bytes, signature_header: str, secret: str) -> bool:
|
||||||
|
# Parse header: "t=1735306800,v1=abc123..."
|
||||||
|
parts = dict(part.split("=") for part in signature_header.split(","))
|
||||||
|
timestamp = parts["t"]
|
||||||
|
received_signature = parts["v1"]
|
||||||
|
|
||||||
|
# Create signed payload
|
||||||
|
signed_payload = f"{timestamp}.{payload.decode('utf-8')}"
|
||||||
|
|
||||||
|
# Compute expected signature
|
||||||
|
expected_signature = hmac.new(
|
||||||
|
secret.encode("utf-8"),
|
||||||
|
signed_payload.encode("utf-8"),
|
||||||
|
hashlib.sha256
|
||||||
|
).hexdigest()
|
||||||
|
|
||||||
|
# Compare signatures
|
||||||
|
return hmac.compare_digest(expected_signature, received_signature)
|
||||||
|
```
|
||||||
|
|
||||||
|
## Event Payloads
|
||||||
|
|
||||||
|
### `transcript.completed` Event
|
||||||
|
|
||||||
|
This event includes a convenient URL for accessing the transcript:
|
||||||
|
- `frontend_url`: Direct link to view the transcript in the web interface
|
||||||
|
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"event": "transcript.completed",
|
||||||
|
"event_id": "transcript.completed-abc-123-def-456",
|
||||||
|
"timestamp": "2025-08-27T12:34:56.789012Z",
|
||||||
|
"transcript": {
|
||||||
|
"id": "abc-123-def-456",
|
||||||
|
"room_id": "room-789",
|
||||||
|
"created_at": "2025-08-27T12:00:00Z",
|
||||||
|
"duration": 1800.5,
|
||||||
|
"title": "Q3 Product Planning Meeting",
|
||||||
|
"short_summary": "Team discussed Q3 product roadmap, prioritizing mobile app features and API improvements.",
|
||||||
|
"long_summary": "The product team met to finalize the Q3 roadmap. Key decisions included...",
|
||||||
|
"webvtt": "WEBVTT\n\n00:00:00.000 --> 00:00:05.000\n<v Speaker 1>Welcome everyone to today's meeting...",
|
||||||
|
"topics": [
|
||||||
|
{
|
||||||
|
"title": "Introduction and Agenda",
|
||||||
|
"summary": "Meeting kickoff with agenda review",
|
||||||
|
"timestamp": 0.0,
|
||||||
|
"duration": 120.0,
|
||||||
|
"webvtt": "WEBVTT\n\n00:00:00.000 --> 00:00:05.000\n<v Speaker 1>Welcome everyone..."
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"title": "Mobile App Features Discussion",
|
||||||
|
"summary": "Team reviewed proposed mobile app features for Q3",
|
||||||
|
"timestamp": 120.0,
|
||||||
|
"duration": 600.0,
|
||||||
|
"webvtt": "WEBVTT\n\n00:02:00.000 --> 00:02:10.000\n<v Speaker 2>Let's talk about the mobile app..."
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"participants": [
|
||||||
|
{
|
||||||
|
"id": "participant-1",
|
||||||
|
"name": "John Doe",
|
||||||
|
"speaker": "Speaker 1"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"id": "participant-2",
|
||||||
|
"name": "Jane Smith",
|
||||||
|
"speaker": "Speaker 2"
|
||||||
|
}
|
||||||
|
],
|
||||||
|
"source_language": "en",
|
||||||
|
"target_language": "en",
|
||||||
|
"status": "completed",
|
||||||
|
"frontend_url": "https://app.reflector.com/transcripts/abc-123-def-456"
|
||||||
|
},
|
||||||
|
"room": {
|
||||||
|
"id": "room-789",
|
||||||
|
"name": "Product Team Room"
|
||||||
|
},
|
||||||
|
"calendar_event": {
|
||||||
|
"id": "calendar-event-123",
|
||||||
|
"ics_uid": "event-123",
|
||||||
|
"title": "Q3 Product Planning Meeting",
|
||||||
|
"start_time": "2025-08-27T12:00:00Z",
|
||||||
|
"end_time": "2025-08-27T12:30:00Z",
|
||||||
|
"description": "Team discussed Q3 product roadmap, prioritizing mobile app features and API improvements.",
|
||||||
|
"location": "Conference Room 1",
|
||||||
|
"attendees": [
|
||||||
|
{
|
||||||
|
"id": "participant-1",
|
||||||
|
"name": "John Doe",
|
||||||
|
"speaker": "Speaker 1"
|
||||||
|
},
|
||||||
|
{
|
||||||
|
"id": "participant-2",
|
||||||
|
"name": "Jane Smith",
|
||||||
|
"speaker": "Speaker 2"
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
### `test` Event
|
||||||
|
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"event": "test",
|
||||||
|
"event_id": "test.2025-08-27T12:34:56.789012Z",
|
||||||
|
"timestamp": "2025-08-27T12:34:56.789012Z",
|
||||||
|
"message": "This is a test webhook from Reflector",
|
||||||
|
"room": {
|
||||||
|
"id": "room-789",
|
||||||
|
"name": "Product Team Room"
|
||||||
|
}
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
## Retry Policy
|
||||||
|
|
||||||
|
Webhooks are delivered with automatic retry logic to handle transient failures. When a webhook delivery fails due to server errors or network issues, Reflector will automatically retry the delivery multiple times over an extended period.
|
||||||
|
|
||||||
|
### Retry Mechanism
|
||||||
|
|
||||||
|
Reflector implements an exponential backoff strategy for webhook retries:
|
||||||
|
|
||||||
|
- **Initial retry delay**: 60 seconds after the first failure
|
||||||
|
- **Exponential backoff**: Each subsequent retry waits approximately twice as long as the previous one
|
||||||
|
- **Maximum retry interval**: 1 hour (backoff is capped at this duration)
|
||||||
|
- **Maximum retry attempts**: 30 attempts total
|
||||||
|
- **Total retry duration**: Retries continue for approximately 24 hours
|
||||||
|
|
||||||
|
### How Retries Work
|
||||||
|
|
||||||
|
When a webhook fails, Reflector will:
|
||||||
|
1. Wait 60 seconds, then retry (attempt #1)
|
||||||
|
2. If it fails again, wait ~2 minutes, then retry (attempt #2)
|
||||||
|
3. Continue doubling the wait time up to a maximum of 1 hour between attempts
|
||||||
|
4. Keep retrying at 1-hour intervals until successful or 30 attempts are exhausted
|
||||||
|
|
||||||
|
The `X-Webhook-Retry` header indicates the current retry attempt number (0 for the initial attempt, 1 for first retry, etc.), allowing your endpoint to track retry attempts.
|
||||||
|
|
||||||
|
### Retry Behavior by HTTP Status Code
|
||||||
|
|
||||||
|
| Status Code | Behavior |
|
||||||
|
|-------------|----------|
|
||||||
|
| 2xx (Success) | No retry, webhook marked as delivered |
|
||||||
|
| 4xx (Client Error) | No retry, request is considered permanently failed |
|
||||||
|
| 5xx (Server Error) | Automatic retry with exponential backoff |
|
||||||
|
| Network/Timeout Error | Automatic retry with exponential backoff |
|
||||||
|
|
||||||
|
**Important Notes:**
|
||||||
|
- Webhooks timeout after 30 seconds. If your endpoint takes longer to respond, it will be considered a timeout error and retried.
|
||||||
|
- During the retry period (~24 hours), you may receive the same webhook multiple times if your endpoint experiences intermittent failures.
|
||||||
|
- There is no mechanism to manually retry failed webhooks after the retry period expires.
|
||||||
|
|
||||||
|
## Testing Webhooks
|
||||||
|
|
||||||
|
You can test your webhook configuration before processing transcripts:
|
||||||
|
|
||||||
|
```http
|
||||||
|
POST /v1/rooms/{room_id}/webhook/test
|
||||||
|
```
|
||||||
|
|
||||||
|
Response:
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"success": true,
|
||||||
|
"status_code": 200,
|
||||||
|
"message": "Webhook test successful",
|
||||||
|
"response_preview": "OK"
|
||||||
|
}
|
||||||
|
```
|
||||||
|
|
||||||
|
Or in case of failure:
|
||||||
|
```json
|
||||||
|
{
|
||||||
|
"success": false,
|
||||||
|
"error": "Webhook request timed out (10 seconds)"
|
||||||
|
}
|
||||||
|
```
|
||||||
@@ -27,7 +27,7 @@ AUTH_JWT_AUDIENCE=
|
|||||||
#TRANSCRIPT_MODAL_API_KEY=xxxxx
|
#TRANSCRIPT_MODAL_API_KEY=xxxxx
|
||||||
|
|
||||||
TRANSCRIPT_BACKEND=modal
|
TRANSCRIPT_BACKEND=modal
|
||||||
TRANSCRIPT_URL=https://monadical-sas--reflector-transcriber-web.modal.run
|
TRANSCRIPT_URL=https://monadical-sas--reflector-transcriber-parakeet-web.modal.run
|
||||||
TRANSCRIPT_MODAL_API_KEY=
|
TRANSCRIPT_MODAL_API_KEY=
|
||||||
|
|
||||||
## =======================================================
|
## =======================================================
|
||||||
@@ -71,3 +71,30 @@ DIARIZATION_URL=https://monadical-sas--reflector-diarizer-web.modal.run
|
|||||||
|
|
||||||
## Sentry DSN configuration
|
## Sentry DSN configuration
|
||||||
#SENTRY_DSN=
|
#SENTRY_DSN=
|
||||||
|
|
||||||
|
## =======================================================
|
||||||
|
## Video Platform Configuration
|
||||||
|
## =======================================================
|
||||||
|
|
||||||
|
## Whereby
|
||||||
|
#WHEREBY_API_KEY=your-whereby-api-key
|
||||||
|
#WHEREBY_WEBHOOK_SECRET=your-whereby-webhook-secret
|
||||||
|
#WHEREBY_STORAGE_AWS_ACCESS_KEY_ID=your-aws-key
|
||||||
|
#WHEREBY_STORAGE_AWS_SECRET_ACCESS_KEY=your-aws-secret
|
||||||
|
#AWS_PROCESS_RECORDING_QUEUE_URL=https://sqs.us-west-2.amazonaws.com/...
|
||||||
|
|
||||||
|
## Daily.co
|
||||||
|
#DAILY_API_KEY=your-daily-api-key
|
||||||
|
#DAILY_WEBHOOK_SECRET=your-daily-webhook-secret
|
||||||
|
#DAILY_SUBDOMAIN=your-subdomain
|
||||||
|
#DAILY_WEBHOOK_UUID= # Auto-populated by recreate_daily_webhook.py script
|
||||||
|
#DAILYCO_STORAGE_AWS_ROLE_ARN=... # IAM role ARN for Daily.co S3 access
|
||||||
|
#DAILYCO_STORAGE_AWS_BUCKET_NAME=reflector-dailyco
|
||||||
|
#DAILYCO_STORAGE_AWS_REGION=us-west-2
|
||||||
|
|
||||||
|
## Whereby (optional separate bucket)
|
||||||
|
#WHEREBY_STORAGE_AWS_BUCKET_NAME=reflector-whereby
|
||||||
|
#WHEREBY_STORAGE_AWS_REGION=us-east-1
|
||||||
|
|
||||||
|
## Platform Configuration
|
||||||
|
#DEFAULT_VIDEO_PLATFORM=whereby # Default platform for new rooms
|
||||||
|
|||||||
@@ -1,86 +0,0 @@
|
|||||||
# Reflector GPU implementation - Transcription and LLM
|
|
||||||
|
|
||||||
This repository hold an API for the GPU implementation of the Reflector API service,
|
|
||||||
and use [Modal.com](https://modal.com)
|
|
||||||
|
|
||||||
- `reflector_diarizer.py` - Diarization API
|
|
||||||
- `reflector_transcriber.py` - Transcription API
|
|
||||||
- `reflector_translator.py` - Translation API
|
|
||||||
|
|
||||||
## Modal.com deployment
|
|
||||||
|
|
||||||
Create a modal secret, and name it `reflector-gpu`.
|
|
||||||
It should contain an `REFLECTOR_APIKEY` environment variable with a value.
|
|
||||||
|
|
||||||
The deployment is done using [Modal.com](https://modal.com) service.
|
|
||||||
|
|
||||||
```
|
|
||||||
$ modal deploy reflector_transcriber.py
|
|
||||||
...
|
|
||||||
└── 🔨 Created web => https://xxxx--reflector-transcriber-web.modal.run
|
|
||||||
|
|
||||||
$ modal deploy reflector_llm.py
|
|
||||||
...
|
|
||||||
└── 🔨 Created web => https://xxxx--reflector-llm-web.modal.run
|
|
||||||
```
|
|
||||||
|
|
||||||
Then in your reflector api configuration `.env`, you can set these keys:
|
|
||||||
|
|
||||||
```
|
|
||||||
TRANSCRIPT_BACKEND=modal
|
|
||||||
TRANSCRIPT_URL=https://xxxx--reflector-transcriber-web.modal.run
|
|
||||||
TRANSCRIPT_MODAL_API_KEY=REFLECTOR_APIKEY
|
|
||||||
|
|
||||||
DIARIZATION_BACKEND=modal
|
|
||||||
DIARIZATION_URL=https://xxxx--reflector-diarizer-web.modal.run
|
|
||||||
DIARIZATION_MODAL_API_KEY=REFLECTOR_APIKEY
|
|
||||||
|
|
||||||
TRANSLATION_BACKEND=modal
|
|
||||||
TRANSLATION_URL=https://xxxx--reflector-translator-web.modal.run
|
|
||||||
TRANSLATION_MODAL_API_KEY=REFLECTOR_APIKEY
|
|
||||||
```
|
|
||||||
|
|
||||||
## API
|
|
||||||
|
|
||||||
Authentication must be passed with the `Authorization` header, using the `bearer` scheme.
|
|
||||||
|
|
||||||
```
|
|
||||||
Authorization: bearer <REFLECTOR_APIKEY>
|
|
||||||
```
|
|
||||||
|
|
||||||
### LLM
|
|
||||||
|
|
||||||
`POST /llm`
|
|
||||||
|
|
||||||
**request**
|
|
||||||
```
|
|
||||||
{
|
|
||||||
"prompt": "xxx"
|
|
||||||
}
|
|
||||||
```
|
|
||||||
|
|
||||||
**response**
|
|
||||||
```
|
|
||||||
{
|
|
||||||
"text": "xxx completed"
|
|
||||||
}
|
|
||||||
```
|
|
||||||
|
|
||||||
### Transcription
|
|
||||||
|
|
||||||
`POST /transcribe`
|
|
||||||
|
|
||||||
**request** (multipart/form-data)
|
|
||||||
|
|
||||||
- `file` - audio file
|
|
||||||
- `language` - language code (e.g. `en`)
|
|
||||||
|
|
||||||
**response**
|
|
||||||
```
|
|
||||||
{
|
|
||||||
"text": "xxx",
|
|
||||||
"words": [
|
|
||||||
{"text": "xxx", "start": 0.0, "end": 1.0}
|
|
||||||
]
|
|
||||||
}
|
|
||||||
```
|
|
||||||
@@ -1,187 +0,0 @@
|
|||||||
"""
|
|
||||||
Reflector GPU backend - diarizer
|
|
||||||
===================================
|
|
||||||
"""
|
|
||||||
|
|
||||||
import os
|
|
||||||
|
|
||||||
import modal.gpu
|
|
||||||
from modal import App, Image, Secret, asgi_app, enter, method
|
|
||||||
from pydantic import BaseModel
|
|
||||||
|
|
||||||
PYANNOTE_MODEL_NAME: str = "pyannote/speaker-diarization-3.1"
|
|
||||||
MODEL_DIR = "/root/diarization_models"
|
|
||||||
app = App(name="reflector-diarizer")
|
|
||||||
|
|
||||||
|
|
||||||
def migrate_cache_llm():
|
|
||||||
"""
|
|
||||||
XXX The cache for model files in Transformers v4.22.0 has been updated.
|
|
||||||
Migrating your old cache. This is a one-time only operation. You can
|
|
||||||
interrupt this and resume the migration later on by calling
|
|
||||||
`transformers.utils.move_cache()`.
|
|
||||||
"""
|
|
||||||
from transformers.utils.hub import move_cache
|
|
||||||
|
|
||||||
print("Moving LLM cache")
|
|
||||||
move_cache(cache_dir=MODEL_DIR, new_cache_dir=MODEL_DIR)
|
|
||||||
print("LLM cache moved")
|
|
||||||
|
|
||||||
|
|
||||||
def download_pyannote_audio():
|
|
||||||
from pyannote.audio import Pipeline
|
|
||||||
|
|
||||||
Pipeline.from_pretrained(
|
|
||||||
PYANNOTE_MODEL_NAME,
|
|
||||||
cache_dir=MODEL_DIR,
|
|
||||||
use_auth_token=os.environ["HF_TOKEN"],
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
diarizer_image = (
|
|
||||||
Image.debian_slim(python_version="3.10.8")
|
|
||||||
.pip_install(
|
|
||||||
"pyannote.audio==3.1.0",
|
|
||||||
"requests",
|
|
||||||
"onnx",
|
|
||||||
"torchaudio",
|
|
||||||
"onnxruntime-gpu",
|
|
||||||
"torch==2.0.0",
|
|
||||||
"transformers==4.34.0",
|
|
||||||
"sentencepiece",
|
|
||||||
"protobuf",
|
|
||||||
"numpy",
|
|
||||||
"huggingface_hub",
|
|
||||||
"hf-transfer",
|
|
||||||
)
|
|
||||||
.run_function(
|
|
||||||
download_pyannote_audio, secrets=[Secret.from_name("my-huggingface-secret")]
|
|
||||||
)
|
|
||||||
.run_function(migrate_cache_llm)
|
|
||||||
.env(
|
|
||||||
{
|
|
||||||
"LD_LIBRARY_PATH": (
|
|
||||||
"/usr/local/lib/python3.10/site-packages/nvidia/cudnn/lib/:"
|
|
||||||
"/opt/conda/lib/python3.10/site-packages/nvidia/cublas/lib/"
|
|
||||||
)
|
|
||||||
}
|
|
||||||
)
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
@app.cls(
|
|
||||||
gpu=modal.gpu.A100(size="40GB"),
|
|
||||||
timeout=60 * 30,
|
|
||||||
scaledown_window=60,
|
|
||||||
allow_concurrent_inputs=1,
|
|
||||||
image=diarizer_image,
|
|
||||||
)
|
|
||||||
class Diarizer:
|
|
||||||
@enter()
|
|
||||||
def enter(self):
|
|
||||||
import torch
|
|
||||||
from pyannote.audio import Pipeline
|
|
||||||
|
|
||||||
self.use_gpu = torch.cuda.is_available()
|
|
||||||
self.device = "cuda" if self.use_gpu else "cpu"
|
|
||||||
self.diarization_pipeline = Pipeline.from_pretrained(
|
|
||||||
PYANNOTE_MODEL_NAME, cache_dir=MODEL_DIR
|
|
||||||
)
|
|
||||||
self.diarization_pipeline.to(torch.device(self.device))
|
|
||||||
|
|
||||||
@method()
|
|
||||||
def diarize(self, audio_data: str, audio_suffix: str, timestamp: float):
|
|
||||||
import tempfile
|
|
||||||
|
|
||||||
import torchaudio
|
|
||||||
|
|
||||||
with tempfile.NamedTemporaryFile("wb+", suffix=f".{audio_suffix}") as fp:
|
|
||||||
fp.write(audio_data)
|
|
||||||
|
|
||||||
print("Diarizing audio")
|
|
||||||
waveform, sample_rate = torchaudio.load(fp.name)
|
|
||||||
diarization = self.diarization_pipeline(
|
|
||||||
{"waveform": waveform, "sample_rate": sample_rate}
|
|
||||||
)
|
|
||||||
|
|
||||||
words = []
|
|
||||||
for diarization_segment, _, speaker in diarization.itertracks(
|
|
||||||
yield_label=True
|
|
||||||
):
|
|
||||||
words.append(
|
|
||||||
{
|
|
||||||
"start": round(timestamp + diarization_segment.start, 3),
|
|
||||||
"end": round(timestamp + diarization_segment.end, 3),
|
|
||||||
"speaker": int(speaker[-2:]),
|
|
||||||
}
|
|
||||||
)
|
|
||||||
print("Diarization complete")
|
|
||||||
return {"diarization": words}
|
|
||||||
|
|
||||||
|
|
||||||
# -------------------------------------------------------------------
|
|
||||||
# Web API
|
|
||||||
# -------------------------------------------------------------------
|
|
||||||
|
|
||||||
|
|
||||||
@app.function(
|
|
||||||
timeout=60 * 10,
|
|
||||||
scaledown_window=60 * 3,
|
|
||||||
allow_concurrent_inputs=40,
|
|
||||||
secrets=[
|
|
||||||
Secret.from_name("reflector-gpu"),
|
|
||||||
],
|
|
||||||
image=diarizer_image,
|
|
||||||
)
|
|
||||||
@asgi_app()
|
|
||||||
def web():
|
|
||||||
import requests
|
|
||||||
from fastapi import Depends, FastAPI, HTTPException, status
|
|
||||||
from fastapi.security import OAuth2PasswordBearer
|
|
||||||
|
|
||||||
diarizerstub = Diarizer()
|
|
||||||
|
|
||||||
app = FastAPI()
|
|
||||||
|
|
||||||
oauth2_scheme = OAuth2PasswordBearer(tokenUrl="token")
|
|
||||||
|
|
||||||
def apikey_auth(apikey: str = Depends(oauth2_scheme)):
|
|
||||||
if apikey != os.environ["REFLECTOR_GPU_APIKEY"]:
|
|
||||||
raise HTTPException(
|
|
||||||
status_code=status.HTTP_401_UNAUTHORIZED,
|
|
||||||
detail="Invalid API key",
|
|
||||||
headers={"WWW-Authenticate": "Bearer"},
|
|
||||||
)
|
|
||||||
|
|
||||||
def validate_audio_file(audio_file_url: str):
|
|
||||||
# Check if the audio file exists
|
|
||||||
response = requests.head(audio_file_url, allow_redirects=True)
|
|
||||||
if response.status_code == 404:
|
|
||||||
raise HTTPException(
|
|
||||||
status_code=response.status_code,
|
|
||||||
detail="The audio file does not exist.",
|
|
||||||
)
|
|
||||||
|
|
||||||
class DiarizationResponse(BaseModel):
|
|
||||||
result: dict
|
|
||||||
|
|
||||||
@app.post(
|
|
||||||
"/diarize", dependencies=[Depends(apikey_auth), Depends(validate_audio_file)]
|
|
||||||
)
|
|
||||||
def diarize(
|
|
||||||
audio_file_url: str, timestamp: float = 0.0
|
|
||||||
) -> HTTPException | DiarizationResponse:
|
|
||||||
# Currently the uploaded files are in mp3 format
|
|
||||||
audio_suffix = "mp3"
|
|
||||||
|
|
||||||
print("Downloading audio file")
|
|
||||||
response = requests.get(audio_file_url, allow_redirects=True)
|
|
||||||
print("Audio file downloaded successfully")
|
|
||||||
|
|
||||||
func = diarizerstub.diarize.spawn(
|
|
||||||
audio_data=response.content, audio_suffix=audio_suffix, timestamp=timestamp
|
|
||||||
)
|
|
||||||
result = func.get()
|
|
||||||
return result
|
|
||||||
|
|
||||||
return app
|
|
||||||
@@ -1,161 +0,0 @@
|
|||||||
import os
|
|
||||||
import tempfile
|
|
||||||
import threading
|
|
||||||
|
|
||||||
import modal
|
|
||||||
from pydantic import BaseModel
|
|
||||||
|
|
||||||
MODELS_DIR = "/models"
|
|
||||||
|
|
||||||
MODEL_NAME = "large-v2"
|
|
||||||
MODEL_COMPUTE_TYPE: str = "float16"
|
|
||||||
MODEL_NUM_WORKERS: int = 1
|
|
||||||
|
|
||||||
MINUTES = 60 # seconds
|
|
||||||
|
|
||||||
volume = modal.Volume.from_name("models", create_if_missing=True)
|
|
||||||
|
|
||||||
app = modal.App("reflector-transcriber")
|
|
||||||
|
|
||||||
|
|
||||||
def download_model():
|
|
||||||
from faster_whisper import download_model
|
|
||||||
|
|
||||||
volume.reload()
|
|
||||||
|
|
||||||
download_model(MODEL_NAME, cache_dir=MODELS_DIR)
|
|
||||||
|
|
||||||
volume.commit()
|
|
||||||
|
|
||||||
|
|
||||||
image = (
|
|
||||||
modal.Image.debian_slim(python_version="3.12")
|
|
||||||
.pip_install(
|
|
||||||
"huggingface_hub==0.27.1",
|
|
||||||
"hf-transfer==0.1.9",
|
|
||||||
"torch==2.5.1",
|
|
||||||
"faster-whisper==1.1.1",
|
|
||||||
)
|
|
||||||
.env(
|
|
||||||
{
|
|
||||||
"HF_HUB_ENABLE_HF_TRANSFER": "1",
|
|
||||||
"LD_LIBRARY_PATH": (
|
|
||||||
"/usr/local/lib/python3.12/site-packages/nvidia/cudnn/lib/:"
|
|
||||||
"/opt/conda/lib/python3.12/site-packages/nvidia/cublas/lib/"
|
|
||||||
),
|
|
||||||
}
|
|
||||||
)
|
|
||||||
.run_function(download_model, volumes={MODELS_DIR: volume})
|
|
||||||
)
|
|
||||||
|
|
||||||
|
|
||||||
@app.cls(
|
|
||||||
gpu="A10G",
|
|
||||||
timeout=5 * MINUTES,
|
|
||||||
scaledown_window=5 * MINUTES,
|
|
||||||
allow_concurrent_inputs=6,
|
|
||||||
image=image,
|
|
||||||
volumes={MODELS_DIR: volume},
|
|
||||||
)
|
|
||||||
class Transcriber:
|
|
||||||
@modal.enter()
|
|
||||||
def enter(self):
|
|
||||||
import faster_whisper
|
|
||||||
import torch
|
|
||||||
|
|
||||||
self.lock = threading.Lock()
|
|
||||||
self.use_gpu = torch.cuda.is_available()
|
|
||||||
self.device = "cuda" if self.use_gpu else "cpu"
|
|
||||||
self.model = faster_whisper.WhisperModel(
|
|
||||||
MODEL_NAME,
|
|
||||||
device=self.device,
|
|
||||||
compute_type=MODEL_COMPUTE_TYPE,
|
|
||||||
num_workers=MODEL_NUM_WORKERS,
|
|
||||||
download_root=MODELS_DIR,
|
|
||||||
local_files_only=True,
|
|
||||||
)
|
|
||||||
|
|
||||||
@modal.method()
|
|
||||||
def transcribe_segment(
|
|
||||||
self,
|
|
||||||
audio_data: str,
|
|
||||||
audio_suffix: str,
|
|
||||||
language: str,
|
|
||||||
):
|
|
||||||
with tempfile.NamedTemporaryFile("wb+", suffix=f".{audio_suffix}") as fp:
|
|
||||||
fp.write(audio_data)
|
|
||||||
|
|
||||||
with self.lock:
|
|
||||||
segments, _ = self.model.transcribe(
|
|
||||||
fp.name,
|
|
||||||
language=language,
|
|
||||||
beam_size=5,
|
|
||||||
word_timestamps=True,
|
|
||||||
vad_filter=True,
|
|
||||||
vad_parameters={"min_silence_duration_ms": 500},
|
|
||||||
)
|
|
||||||
|
|
||||||
segments = list(segments)
|
|
||||||
text = "".join(segment.text for segment in segments)
|
|
||||||
words = [
|
|
||||||
{"word": word.word, "start": word.start, "end": word.end}
|
|
||||||
for segment in segments
|
|
||||||
for word in segment.words
|
|
||||||
]
|
|
||||||
|
|
||||||
return {"text": text, "words": words}
|
|
||||||
|
|
||||||
|
|
||||||
@app.function(
|
|
||||||
scaledown_window=60,
|
|
||||||
timeout=60,
|
|
||||||
allow_concurrent_inputs=40,
|
|
||||||
secrets=[
|
|
||||||
modal.Secret.from_name("reflector-gpu"),
|
|
||||||
],
|
|
||||||
volumes={MODELS_DIR: volume},
|
|
||||||
)
|
|
||||||
@modal.asgi_app()
|
|
||||||
def web():
|
|
||||||
from fastapi import Body, Depends, FastAPI, HTTPException, UploadFile, status
|
|
||||||
from fastapi.security import OAuth2PasswordBearer
|
|
||||||
from typing_extensions import Annotated
|
|
||||||
|
|
||||||
transcriber = Transcriber()
|
|
||||||
|
|
||||||
app = FastAPI()
|
|
||||||
|
|
||||||
oauth2_scheme = OAuth2PasswordBearer(tokenUrl="token")
|
|
||||||
|
|
||||||
supported_file_types = ["mp3", "mp4", "mpeg", "mpga", "m4a", "wav", "webm"]
|
|
||||||
|
|
||||||
def apikey_auth(apikey: str = Depends(oauth2_scheme)):
|
|
||||||
if apikey != os.environ["REFLECTOR_GPU_APIKEY"]:
|
|
||||||
raise HTTPException(
|
|
||||||
status_code=status.HTTP_401_UNAUTHORIZED,
|
|
||||||
detail="Invalid API key",
|
|
||||||
headers={"WWW-Authenticate": "Bearer"},
|
|
||||||
)
|
|
||||||
|
|
||||||
class TranscriptResponse(BaseModel):
|
|
||||||
result: dict
|
|
||||||
|
|
||||||
@app.post("/v1/audio/transcriptions", dependencies=[Depends(apikey_auth)])
|
|
||||||
def transcribe(
|
|
||||||
file: UploadFile,
|
|
||||||
model: str = "whisper-1",
|
|
||||||
language: Annotated[str, Body(...)] = "en",
|
|
||||||
) -> TranscriptResponse:
|
|
||||||
audio_data = file.file.read()
|
|
||||||
audio_suffix = file.filename.split(".")[-1]
|
|
||||||
assert audio_suffix in supported_file_types
|
|
||||||
|
|
||||||
func = transcriber.transcribe_segment.spawn(
|
|
||||||
audio_data=audio_data,
|
|
||||||
audio_suffix=audio_suffix,
|
|
||||||
language=language,
|
|
||||||
)
|
|
||||||
result = func.get()
|
|
||||||
return result
|
|
||||||
|
|
||||||
return app
|
|
||||||
@@ -0,0 +1,36 @@
|
|||||||
|
"""Add webhook fields to rooms
|
||||||
|
|
||||||
|
Revision ID: 0194f65cd6d3
|
||||||
|
Revises: 5a8907fd1d78
|
||||||
|
Create Date: 2025-08-27 09:03:19.610995
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from typing import Sequence, Union
|
||||||
|
|
||||||
|
import sqlalchemy as sa
|
||||||
|
from alembic import op
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "0194f65cd6d3"
|
||||||
|
down_revision: Union[str, None] = "5a8907fd1d78"
|
||||||
|
branch_labels: Union[str, Sequence[str], None] = None
|
||||||
|
depends_on: Union[str, Sequence[str], None] = None
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
with op.batch_alter_table("room", schema=None) as batch_op:
|
||||||
|
batch_op.add_column(sa.Column("webhook_url", sa.String(), nullable=True))
|
||||||
|
batch_op.add_column(sa.Column("webhook_secret", sa.String(), nullable=True))
|
||||||
|
|
||||||
|
# ### end Alembic commands ###
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
with op.batch_alter_table("room", schema=None) as batch_op:
|
||||||
|
batch_op.drop_column("webhook_secret")
|
||||||
|
batch_op.drop_column("webhook_url")
|
||||||
|
|
||||||
|
# ### end Alembic commands ###
|
||||||
@@ -0,0 +1,64 @@
|
|||||||
|
"""add_long_summary_to_search_vector
|
||||||
|
|
||||||
|
Revision ID: 0ab2d7ffaa16
|
||||||
|
Revises: b1c33bd09963
|
||||||
|
Create Date: 2025-08-15 13:27:52.680211
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from typing import Sequence, Union
|
||||||
|
|
||||||
|
from alembic import op
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "0ab2d7ffaa16"
|
||||||
|
down_revision: Union[str, None] = "b1c33bd09963"
|
||||||
|
branch_labels: Union[str, Sequence[str], None] = None
|
||||||
|
depends_on: Union[str, Sequence[str], None] = None
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
# Drop the existing search vector column and index
|
||||||
|
op.drop_index("idx_transcript_search_vector_en", table_name="transcript")
|
||||||
|
op.drop_column("transcript", "search_vector_en")
|
||||||
|
|
||||||
|
# Recreate the search vector column with long_summary included
|
||||||
|
op.execute("""
|
||||||
|
ALTER TABLE transcript ADD COLUMN search_vector_en tsvector
|
||||||
|
GENERATED ALWAYS AS (
|
||||||
|
setweight(to_tsvector('english', coalesce(title, '')), 'A') ||
|
||||||
|
setweight(to_tsvector('english', coalesce(long_summary, '')), 'B') ||
|
||||||
|
setweight(to_tsvector('english', coalesce(webvtt, '')), 'C')
|
||||||
|
) STORED
|
||||||
|
""")
|
||||||
|
|
||||||
|
# Recreate the GIN index for the search vector
|
||||||
|
op.create_index(
|
||||||
|
"idx_transcript_search_vector_en",
|
||||||
|
"transcript",
|
||||||
|
["search_vector_en"],
|
||||||
|
postgresql_using="gin",
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
# Drop the updated search vector column and index
|
||||||
|
op.drop_index("idx_transcript_search_vector_en", table_name="transcript")
|
||||||
|
op.drop_column("transcript", "search_vector_en")
|
||||||
|
|
||||||
|
# Recreate the original search vector column without long_summary
|
||||||
|
op.execute("""
|
||||||
|
ALTER TABLE transcript ADD COLUMN search_vector_en tsvector
|
||||||
|
GENERATED ALWAYS AS (
|
||||||
|
setweight(to_tsvector('english', coalesce(title, '')), 'A') ||
|
||||||
|
setweight(to_tsvector('english', coalesce(webvtt, '')), 'B')
|
||||||
|
) STORED
|
||||||
|
""")
|
||||||
|
|
||||||
|
# Recreate the GIN index for the search vector
|
||||||
|
op.create_index(
|
||||||
|
"idx_transcript_search_vector_en",
|
||||||
|
"transcript",
|
||||||
|
["search_vector_en"],
|
||||||
|
postgresql_using="gin",
|
||||||
|
)
|
||||||
@@ -0,0 +1,36 @@
|
|||||||
|
"""remove user_id from meeting table
|
||||||
|
|
||||||
|
Revision ID: 0ce521cda2ee
|
||||||
|
Revises: 6dec9fb5b46c
|
||||||
|
Create Date: 2025-09-10 12:40:55.688899
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from typing import Sequence, Union
|
||||||
|
|
||||||
|
import sqlalchemy as sa
|
||||||
|
from alembic import op
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "0ce521cda2ee"
|
||||||
|
down_revision: Union[str, None] = "6dec9fb5b46c"
|
||||||
|
branch_labels: Union[str, Sequence[str], None] = None
|
||||||
|
depends_on: Union[str, Sequence[str], None] = None
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
with op.batch_alter_table("meeting", schema=None) as batch_op:
|
||||||
|
batch_op.drop_column("user_id")
|
||||||
|
|
||||||
|
# ### end Alembic commands ###
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
with op.batch_alter_table("meeting", schema=None) as batch_op:
|
||||||
|
batch_op.add_column(
|
||||||
|
sa.Column("user_id", sa.VARCHAR(), autoincrement=False, nullable=True)
|
||||||
|
)
|
||||||
|
|
||||||
|
# ### end Alembic commands ###
|
||||||
@@ -0,0 +1,50 @@
|
|||||||
|
"""add_platform_support
|
||||||
|
|
||||||
|
Revision ID: 1e49625677e4
|
||||||
|
Revises: 9e3f7b2a4c8e
|
||||||
|
Create Date: 2025-10-08 13:17:29.943612
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from typing import Sequence, Union
|
||||||
|
|
||||||
|
import sqlalchemy as sa
|
||||||
|
from alembic import op
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "1e49625677e4"
|
||||||
|
down_revision: Union[str, None] = "9e3f7b2a4c8e"
|
||||||
|
branch_labels: Union[str, Sequence[str], None] = None
|
||||||
|
depends_on: Union[str, Sequence[str], None] = None
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
"""Add platform field with default 'whereby' for backward compatibility."""
|
||||||
|
with op.batch_alter_table("room", schema=None) as batch_op:
|
||||||
|
batch_op.add_column(
|
||||||
|
sa.Column(
|
||||||
|
"platform",
|
||||||
|
sa.String(),
|
||||||
|
nullable=True,
|
||||||
|
server_default=None,
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
|
with op.batch_alter_table("meeting", schema=None) as batch_op:
|
||||||
|
batch_op.add_column(
|
||||||
|
sa.Column(
|
||||||
|
"platform",
|
||||||
|
sa.String(),
|
||||||
|
nullable=False,
|
||||||
|
server_default="whereby",
|
||||||
|
)
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
"""Remove platform field."""
|
||||||
|
with op.batch_alter_table("meeting", schema=None) as batch_op:
|
||||||
|
batch_op.drop_column("platform")
|
||||||
|
|
||||||
|
with op.batch_alter_table("room", schema=None) as batch_op:
|
||||||
|
batch_op.drop_column("platform")
|
||||||
@@ -0,0 +1,32 @@
|
|||||||
|
"""clean up orphaned room_id references in meeting table
|
||||||
|
|
||||||
|
Revision ID: 2ae3db106d4e
|
||||||
|
Revises: def1b5867d4c
|
||||||
|
Create Date: 2025-09-11 10:35:15.759967
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from typing import Sequence, Union
|
||||||
|
|
||||||
|
from alembic import op
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "2ae3db106d4e"
|
||||||
|
down_revision: Union[str, None] = "def1b5867d4c"
|
||||||
|
branch_labels: Union[str, Sequence[str], None] = None
|
||||||
|
depends_on: Union[str, Sequence[str], None] = None
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
# Set room_id to NULL for meetings that reference non-existent rooms
|
||||||
|
op.execute("""
|
||||||
|
UPDATE meeting
|
||||||
|
SET room_id = NULL
|
||||||
|
WHERE room_id IS NOT NULL
|
||||||
|
AND room_id NOT IN (SELECT id FROM room WHERE id IS NOT NULL)
|
||||||
|
""")
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
# Cannot restore orphaned references - no operation needed
|
||||||
|
pass
|
||||||
@@ -0,0 +1,50 @@
|
|||||||
|
"""add cascade delete to meeting consent foreign key
|
||||||
|
|
||||||
|
Revision ID: 5a8907fd1d78
|
||||||
|
Revises: 0ab2d7ffaa16
|
||||||
|
Create Date: 2025-08-26 17:26:50.945491
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from typing import Sequence, Union
|
||||||
|
|
||||||
|
from alembic import op
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "5a8907fd1d78"
|
||||||
|
down_revision: Union[str, None] = "0ab2d7ffaa16"
|
||||||
|
branch_labels: Union[str, Sequence[str], None] = None
|
||||||
|
depends_on: Union[str, Sequence[str], None] = None
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
with op.batch_alter_table("meeting_consent", schema=None) as batch_op:
|
||||||
|
batch_op.drop_constraint(
|
||||||
|
batch_op.f("meeting_consent_meeting_id_fkey"), type_="foreignkey"
|
||||||
|
)
|
||||||
|
batch_op.create_foreign_key(
|
||||||
|
batch_op.f("meeting_consent_meeting_id_fkey"),
|
||||||
|
"meeting",
|
||||||
|
["meeting_id"],
|
||||||
|
["id"],
|
||||||
|
ondelete="CASCADE",
|
||||||
|
)
|
||||||
|
|
||||||
|
# ### end Alembic commands ###
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
with op.batch_alter_table("meeting_consent", schema=None) as batch_op:
|
||||||
|
batch_op.drop_constraint(
|
||||||
|
batch_op.f("meeting_consent_meeting_id_fkey"), type_="foreignkey"
|
||||||
|
)
|
||||||
|
batch_op.create_foreign_key(
|
||||||
|
batch_op.f("meeting_consent_meeting_id_fkey"),
|
||||||
|
"meeting",
|
||||||
|
["meeting_id"],
|
||||||
|
["id"],
|
||||||
|
)
|
||||||
|
|
||||||
|
# ### end Alembic commands ###
|
||||||
@@ -1,7 +1,7 @@
|
|||||||
"""remove_one_active_meeting_per_room_constraint
|
"""remove_one_active_meeting_per_room_constraint
|
||||||
|
|
||||||
Revision ID: 6025e9b2bef2
|
Revision ID: 6025e9b2bef2
|
||||||
Revises: 9f5c78d352d6
|
Revises: 2ae3db106d4e
|
||||||
Create Date: 2025-08-18 18:45:44.418392
|
Create Date: 2025-08-18 18:45:44.418392
|
||||||
|
|
||||||
"""
|
"""
|
||||||
@@ -13,7 +13,7 @@ from alembic import op
|
|||||||
|
|
||||||
# revision identifiers, used by Alembic.
|
# revision identifiers, used by Alembic.
|
||||||
revision: str = "6025e9b2bef2"
|
revision: str = "6025e9b2bef2"
|
||||||
down_revision: Union[str, None] = "9f5c78d352d6"
|
down_revision: Union[str, None] = "2ae3db106d4e"
|
||||||
branch_labels: Union[str, Sequence[str], None] = None
|
branch_labels: Union[str, Sequence[str], None] = None
|
||||||
depends_on: Union[str, Sequence[str], None] = None
|
depends_on: Union[str, Sequence[str], None] = None
|
||||||
|
|
||||||
|
|||||||
@@ -0,0 +1,28 @@
|
|||||||
|
"""webhook url and secret null by default
|
||||||
|
|
||||||
|
|
||||||
|
Revision ID: 61882a919591
|
||||||
|
Revises: 0194f65cd6d3
|
||||||
|
Create Date: 2025-08-29 11:46:36.738091
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from typing import Sequence, Union
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "61882a919591"
|
||||||
|
down_revision: Union[str, None] = "0194f65cd6d3"
|
||||||
|
branch_labels: Union[str, Sequence[str], None] = None
|
||||||
|
depends_on: Union[str, Sequence[str], None] = None
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
pass
|
||||||
|
# ### end Alembic commands ###
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
pass
|
||||||
|
# ### end Alembic commands ###
|
||||||
@@ -0,0 +1,35 @@
|
|||||||
|
"""make meeting room_id required and add foreign key
|
||||||
|
|
||||||
|
Revision ID: 6dec9fb5b46c
|
||||||
|
Revises: 61882a919591
|
||||||
|
Create Date: 2025-09-10 10:47:06.006819
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from typing import Sequence, Union
|
||||||
|
|
||||||
|
from alembic import op
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "6dec9fb5b46c"
|
||||||
|
down_revision: Union[str, None] = "61882a919591"
|
||||||
|
branch_labels: Union[str, Sequence[str], None] = None
|
||||||
|
depends_on: Union[str, Sequence[str], None] = None
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
with op.batch_alter_table("meeting", schema=None) as batch_op:
|
||||||
|
batch_op.create_foreign_key(
|
||||||
|
None, "room", ["room_id"], ["id"], ondelete="CASCADE"
|
||||||
|
)
|
||||||
|
|
||||||
|
# ### end Alembic commands ###
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
with op.batch_alter_table("meeting", schema=None) as batch_op:
|
||||||
|
batch_op.drop_constraint("meeting_room_id_fkey", type_="foreignkey")
|
||||||
|
|
||||||
|
# ### end Alembic commands ###
|
||||||
38
server/migrations/versions/9e3f7b2a4c8e_add_user_api_keys.py
Normal file
38
server/migrations/versions/9e3f7b2a4c8e_add_user_api_keys.py
Normal file
@@ -0,0 +1,38 @@
|
|||||||
|
"""add user api keys
|
||||||
|
|
||||||
|
Revision ID: 9e3f7b2a4c8e
|
||||||
|
Revises: dc035ff72fd5
|
||||||
|
Create Date: 2025-10-17 00:00:00.000000
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from typing import Sequence, Union
|
||||||
|
|
||||||
|
import sqlalchemy as sa
|
||||||
|
from alembic import op
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "9e3f7b2a4c8e"
|
||||||
|
down_revision: Union[str, None] = "dc035ff72fd5"
|
||||||
|
branch_labels: Union[str, Sequence[str], None] = None
|
||||||
|
depends_on: Union[str, Sequence[str], None] = None
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
op.create_table(
|
||||||
|
"user_api_key",
|
||||||
|
sa.Column("id", sa.String(), nullable=False),
|
||||||
|
sa.Column("user_id", sa.String(), nullable=False),
|
||||||
|
sa.Column("key_hash", sa.String(), nullable=False),
|
||||||
|
sa.Column("name", sa.String(), nullable=True),
|
||||||
|
sa.Column("created_at", sa.DateTime(timezone=True), nullable=False),
|
||||||
|
sa.PrimaryKeyConstraint("id"),
|
||||||
|
)
|
||||||
|
|
||||||
|
with op.batch_alter_table("user_api_key", schema=None) as batch_op:
|
||||||
|
batch_op.create_index("idx_user_api_key_hash", ["key_hash"], unique=True)
|
||||||
|
batch_op.create_index("idx_user_api_key_user_id", ["user_id"], unique=False)
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
op.drop_table("user_api_key")
|
||||||
@@ -0,0 +1,41 @@
|
|||||||
|
"""add_search_optimization_indexes
|
||||||
|
|
||||||
|
Revision ID: b1c33bd09963
|
||||||
|
Revises: 9f5c78d352d6
|
||||||
|
Create Date: 2025-08-14 17:26:02.117408
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from typing import Sequence, Union
|
||||||
|
|
||||||
|
from alembic import op
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "b1c33bd09963"
|
||||||
|
down_revision: Union[str, None] = "9f5c78d352d6"
|
||||||
|
branch_labels: Union[str, Sequence[str], None] = None
|
||||||
|
depends_on: Union[str, Sequence[str], None] = None
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
# Add indexes for actual search filtering patterns used in frontend
|
||||||
|
# Based on /browse page filters: room_id and source_kind
|
||||||
|
|
||||||
|
# Index for room_id + created_at (for room-specific searches with date ordering)
|
||||||
|
op.create_index(
|
||||||
|
"idx_transcript_room_id_created_at",
|
||||||
|
"transcript",
|
||||||
|
["room_id", "created_at"],
|
||||||
|
if_not_exists=True,
|
||||||
|
)
|
||||||
|
|
||||||
|
# Index for source_kind alone (actively used filter in frontend)
|
||||||
|
op.create_index(
|
||||||
|
"idx_transcript_source_kind", "transcript", ["source_kind"], if_not_exists=True
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
# Remove the indexes in reverse order
|
||||||
|
op.drop_index("idx_transcript_source_kind", "transcript", if_exists=True)
|
||||||
|
op.drop_index("idx_transcript_room_id_created_at", "transcript", if_exists=True)
|
||||||
129
server/migrations/versions/d8e204bbf615_add_calendar.py
Normal file
129
server/migrations/versions/d8e204bbf615_add_calendar.py
Normal file
@@ -0,0 +1,129 @@
|
|||||||
|
"""add calendar
|
||||||
|
|
||||||
|
Revision ID: d8e204bbf615
|
||||||
|
Revises: d4a1c446458c
|
||||||
|
Create Date: 2025-09-10 19:56:22.295756
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from typing import Sequence, Union
|
||||||
|
|
||||||
|
import sqlalchemy as sa
|
||||||
|
from alembic import op
|
||||||
|
from sqlalchemy.dialects import postgresql
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "d8e204bbf615"
|
||||||
|
down_revision: Union[str, None] = "d4a1c446458c"
|
||||||
|
branch_labels: Union[str, Sequence[str], None] = None
|
||||||
|
depends_on: Union[str, Sequence[str], None] = None
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
op.create_table(
|
||||||
|
"calendar_event",
|
||||||
|
sa.Column("id", sa.String(), nullable=False),
|
||||||
|
sa.Column("room_id", sa.String(), nullable=False),
|
||||||
|
sa.Column("ics_uid", sa.Text(), nullable=False),
|
||||||
|
sa.Column("title", sa.Text(), nullable=True),
|
||||||
|
sa.Column("description", sa.Text(), nullable=True),
|
||||||
|
sa.Column("start_time", sa.DateTime(timezone=True), nullable=False),
|
||||||
|
sa.Column("end_time", sa.DateTime(timezone=True), nullable=False),
|
||||||
|
sa.Column("attendees", postgresql.JSONB(astext_type=sa.Text()), nullable=True),
|
||||||
|
sa.Column("location", sa.Text(), nullable=True),
|
||||||
|
sa.Column("ics_raw_data", sa.Text(), nullable=True),
|
||||||
|
sa.Column("last_synced", sa.DateTime(timezone=True), nullable=False),
|
||||||
|
sa.Column(
|
||||||
|
"is_deleted", sa.Boolean(), server_default=sa.text("false"), nullable=False
|
||||||
|
),
|
||||||
|
sa.Column("created_at", sa.DateTime(timezone=True), nullable=False),
|
||||||
|
sa.Column("updated_at", sa.DateTime(timezone=True), nullable=False),
|
||||||
|
sa.ForeignKeyConstraint(
|
||||||
|
["room_id"],
|
||||||
|
["room.id"],
|
||||||
|
name="fk_calendar_event_room_id",
|
||||||
|
ondelete="CASCADE",
|
||||||
|
),
|
||||||
|
sa.PrimaryKeyConstraint("id"),
|
||||||
|
sa.UniqueConstraint("room_id", "ics_uid", name="uq_room_calendar_event"),
|
||||||
|
)
|
||||||
|
with op.batch_alter_table("calendar_event", schema=None) as batch_op:
|
||||||
|
batch_op.create_index(
|
||||||
|
"idx_calendar_event_deleted",
|
||||||
|
["is_deleted"],
|
||||||
|
unique=False,
|
||||||
|
postgresql_where=sa.text("NOT is_deleted"),
|
||||||
|
)
|
||||||
|
batch_op.create_index(
|
||||||
|
"idx_calendar_event_room_start", ["room_id", "start_time"], unique=False
|
||||||
|
)
|
||||||
|
|
||||||
|
with op.batch_alter_table("meeting", schema=None) as batch_op:
|
||||||
|
batch_op.add_column(sa.Column("calendar_event_id", sa.String(), nullable=True))
|
||||||
|
batch_op.add_column(
|
||||||
|
sa.Column(
|
||||||
|
"calendar_metadata",
|
||||||
|
postgresql.JSONB(astext_type=sa.Text()),
|
||||||
|
nullable=True,
|
||||||
|
)
|
||||||
|
)
|
||||||
|
batch_op.create_index(
|
||||||
|
"idx_meeting_calendar_event", ["calendar_event_id"], unique=False
|
||||||
|
)
|
||||||
|
batch_op.create_foreign_key(
|
||||||
|
"fk_meeting_calendar_event_id",
|
||||||
|
"calendar_event",
|
||||||
|
["calendar_event_id"],
|
||||||
|
["id"],
|
||||||
|
ondelete="SET NULL",
|
||||||
|
)
|
||||||
|
|
||||||
|
with op.batch_alter_table("room", schema=None) as batch_op:
|
||||||
|
batch_op.add_column(sa.Column("ics_url", sa.Text(), nullable=True))
|
||||||
|
batch_op.add_column(
|
||||||
|
sa.Column(
|
||||||
|
"ics_fetch_interval", sa.Integer(), server_default="300", nullable=True
|
||||||
|
)
|
||||||
|
)
|
||||||
|
batch_op.add_column(
|
||||||
|
sa.Column(
|
||||||
|
"ics_enabled",
|
||||||
|
sa.Boolean(),
|
||||||
|
server_default=sa.text("false"),
|
||||||
|
nullable=False,
|
||||||
|
)
|
||||||
|
)
|
||||||
|
batch_op.add_column(
|
||||||
|
sa.Column("ics_last_sync", sa.DateTime(timezone=True), nullable=True)
|
||||||
|
)
|
||||||
|
batch_op.add_column(sa.Column("ics_last_etag", sa.Text(), nullable=True))
|
||||||
|
batch_op.create_index("idx_room_ics_enabled", ["ics_enabled"], unique=False)
|
||||||
|
|
||||||
|
# ### end Alembic commands ###
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
with op.batch_alter_table("room", schema=None) as batch_op:
|
||||||
|
batch_op.drop_index("idx_room_ics_enabled")
|
||||||
|
batch_op.drop_column("ics_last_etag")
|
||||||
|
batch_op.drop_column("ics_last_sync")
|
||||||
|
batch_op.drop_column("ics_enabled")
|
||||||
|
batch_op.drop_column("ics_fetch_interval")
|
||||||
|
batch_op.drop_column("ics_url")
|
||||||
|
|
||||||
|
with op.batch_alter_table("meeting", schema=None) as batch_op:
|
||||||
|
batch_op.drop_constraint("fk_meeting_calendar_event_id", type_="foreignkey")
|
||||||
|
batch_op.drop_index("idx_meeting_calendar_event")
|
||||||
|
batch_op.drop_column("calendar_metadata")
|
||||||
|
batch_op.drop_column("calendar_event_id")
|
||||||
|
|
||||||
|
with op.batch_alter_table("calendar_event", schema=None) as batch_op:
|
||||||
|
batch_op.drop_index("idx_calendar_event_room_start")
|
||||||
|
batch_op.drop_index(
|
||||||
|
"idx_calendar_event_deleted", postgresql_where=sa.text("NOT is_deleted")
|
||||||
|
)
|
||||||
|
|
||||||
|
op.drop_table("calendar_event")
|
||||||
|
# ### end Alembic commands ###
|
||||||
@@ -0,0 +1,43 @@
|
|||||||
|
"""remove_grace_period_fields
|
||||||
|
|
||||||
|
Revision ID: dc035ff72fd5
|
||||||
|
Revises: d8e204bbf615
|
||||||
|
Create Date: 2025-09-11 10:36:45.197588
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from typing import Sequence, Union
|
||||||
|
|
||||||
|
import sqlalchemy as sa
|
||||||
|
from alembic import op
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "dc035ff72fd5"
|
||||||
|
down_revision: Union[str, None] = "d8e204bbf615"
|
||||||
|
branch_labels: Union[str, Sequence[str], None] = None
|
||||||
|
depends_on: Union[str, Sequence[str], None] = None
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
# Remove grace period columns from meeting table
|
||||||
|
op.drop_column("meeting", "last_participant_left_at")
|
||||||
|
op.drop_column("meeting", "grace_period_minutes")
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
# Add back grace period columns to meeting table
|
||||||
|
op.add_column(
|
||||||
|
"meeting",
|
||||||
|
sa.Column(
|
||||||
|
"last_participant_left_at", sa.DateTime(timezone=True), nullable=True
|
||||||
|
),
|
||||||
|
)
|
||||||
|
op.add_column(
|
||||||
|
"meeting",
|
||||||
|
sa.Column(
|
||||||
|
"grace_period_minutes",
|
||||||
|
sa.Integer(),
|
||||||
|
server_default=sa.text("15"),
|
||||||
|
nullable=True,
|
||||||
|
),
|
||||||
|
)
|
||||||
@@ -0,0 +1,34 @@
|
|||||||
|
"""make meeting room_id nullable but keep foreign key
|
||||||
|
|
||||||
|
Revision ID: def1b5867d4c
|
||||||
|
Revises: 0ce521cda2ee
|
||||||
|
Create Date: 2025-09-11 09:42:18.697264
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from typing import Sequence, Union
|
||||||
|
|
||||||
|
import sqlalchemy as sa
|
||||||
|
from alembic import op
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "def1b5867d4c"
|
||||||
|
down_revision: Union[str, None] = "0ce521cda2ee"
|
||||||
|
branch_labels: Union[str, Sequence[str], None] = None
|
||||||
|
depends_on: Union[str, Sequence[str], None] = None
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
with op.batch_alter_table("meeting", schema=None) as batch_op:
|
||||||
|
batch_op.alter_column("room_id", existing_type=sa.VARCHAR(), nullable=True)
|
||||||
|
|
||||||
|
# ### end Alembic commands ###
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
# ### commands auto generated by Alembic - please adjust! ###
|
||||||
|
with op.batch_alter_table("meeting", schema=None) as batch_op:
|
||||||
|
batch_op.alter_column("room_id", existing_type=sa.VARCHAR(), nullable=False)
|
||||||
|
|
||||||
|
# ### end Alembic commands ###
|
||||||
28
server/migrations/versions/f8294b31f022_add_track_keys.py
Normal file
28
server/migrations/versions/f8294b31f022_add_track_keys.py
Normal file
@@ -0,0 +1,28 @@
|
|||||||
|
"""add_track_keys
|
||||||
|
|
||||||
|
Revision ID: f8294b31f022
|
||||||
|
Revises: 1e49625677e4
|
||||||
|
Create Date: 2025-10-27 18:52:17.589167
|
||||||
|
|
||||||
|
"""
|
||||||
|
|
||||||
|
from typing import Sequence, Union
|
||||||
|
|
||||||
|
import sqlalchemy as sa
|
||||||
|
from alembic import op
|
||||||
|
|
||||||
|
# revision identifiers, used by Alembic.
|
||||||
|
revision: str = "f8294b31f022"
|
||||||
|
down_revision: Union[str, None] = "1e49625677e4"
|
||||||
|
branch_labels: Union[str, Sequence[str], None] = None
|
||||||
|
depends_on: Union[str, Sequence[str], None] = None
|
||||||
|
|
||||||
|
|
||||||
|
def upgrade() -> None:
|
||||||
|
with op.batch_alter_table("recording", schema=None) as batch_op:
|
||||||
|
batch_op.add_column(sa.Column("track_keys", sa.JSON(), nullable=True))
|
||||||
|
|
||||||
|
|
||||||
|
def downgrade() -> None:
|
||||||
|
with op.batch_alter_table("recording", schema=None) as batch_op:
|
||||||
|
batch_op.drop_column("track_keys")
|
||||||
@@ -12,7 +12,6 @@ dependencies = [
|
|||||||
"requests>=2.31.0",
|
"requests>=2.31.0",
|
||||||
"aiortc>=1.5.0",
|
"aiortc>=1.5.0",
|
||||||
"sortedcontainers>=2.4.0",
|
"sortedcontainers>=2.4.0",
|
||||||
"loguru>=0.7.0",
|
|
||||||
"pydantic-settings>=2.0.2",
|
"pydantic-settings>=2.0.2",
|
||||||
"structlog>=23.1.0",
|
"structlog>=23.1.0",
|
||||||
"uvicorn[standard]>=0.23.1",
|
"uvicorn[standard]>=0.23.1",
|
||||||
@@ -27,12 +26,10 @@ dependencies = [
|
|||||||
"prometheus-fastapi-instrumentator>=6.1.0",
|
"prometheus-fastapi-instrumentator>=6.1.0",
|
||||||
"sentencepiece>=0.1.99",
|
"sentencepiece>=0.1.99",
|
||||||
"protobuf>=4.24.3",
|
"protobuf>=4.24.3",
|
||||||
"profanityfilter>=2.0.6",
|
|
||||||
"celery>=5.3.4",
|
"celery>=5.3.4",
|
||||||
"redis>=5.0.1",
|
"redis>=5.0.1",
|
||||||
"python-jose[cryptography]>=3.3.0",
|
"python-jose[cryptography]>=3.3.0",
|
||||||
"python-multipart>=0.0.6",
|
"python-multipart>=0.0.6",
|
||||||
"faster-whisper>=0.10.0",
|
|
||||||
"transformers>=4.36.2",
|
"transformers>=4.36.2",
|
||||||
"jsonschema>=4.23.0",
|
"jsonschema>=4.23.0",
|
||||||
"openai>=1.59.7",
|
"openai>=1.59.7",
|
||||||
@@ -58,6 +55,7 @@ tests = [
|
|||||||
"httpx-ws>=0.4.1",
|
"httpx-ws>=0.4.1",
|
||||||
"pytest-httpx>=0.23.1",
|
"pytest-httpx>=0.23.1",
|
||||||
"pytest-celery>=0.0.0",
|
"pytest-celery>=0.0.0",
|
||||||
|
"pytest-recording>=0.13.4",
|
||||||
"pytest-docker>=3.2.3",
|
"pytest-docker>=3.2.3",
|
||||||
"asgi-lifespan>=2.1.0",
|
"asgi-lifespan>=2.1.0",
|
||||||
]
|
]
|
||||||
@@ -68,6 +66,15 @@ evaluation = [
|
|||||||
"tqdm>=4.66.0",
|
"tqdm>=4.66.0",
|
||||||
"pydantic>=2.1.1",
|
"pydantic>=2.1.1",
|
||||||
]
|
]
|
||||||
|
local = [
|
||||||
|
"pyannote-audio>=3.3.2",
|
||||||
|
"faster-whisper>=0.10.0",
|
||||||
|
]
|
||||||
|
silero-vad = [
|
||||||
|
"silero-vad>=5.1.2",
|
||||||
|
"torch>=2.8.0",
|
||||||
|
"torchaudio>=2.8.0",
|
||||||
|
]
|
||||||
|
|
||||||
[tool.uv]
|
[tool.uv]
|
||||||
default-groups = [
|
default-groups = [
|
||||||
@@ -75,6 +82,21 @@ default-groups = [
|
|||||||
"tests",
|
"tests",
|
||||||
"aws",
|
"aws",
|
||||||
"evaluation",
|
"evaluation",
|
||||||
|
"local",
|
||||||
|
"silero-vad"
|
||||||
|
]
|
||||||
|
|
||||||
|
[[tool.uv.index]]
|
||||||
|
name = "pytorch-cpu"
|
||||||
|
url = "https://download.pytorch.org/whl/cpu"
|
||||||
|
explicit = true
|
||||||
|
|
||||||
|
[tool.uv.sources]
|
||||||
|
torch = [
|
||||||
|
{ index = "pytorch-cpu" },
|
||||||
|
]
|
||||||
|
torchaudio = [
|
||||||
|
{ index = "pytorch-cpu" },
|
||||||
]
|
]
|
||||||
|
|
||||||
[build-system]
|
[build-system]
|
||||||
@@ -90,11 +112,15 @@ source = ["reflector"]
|
|||||||
[tool.pytest_env]
|
[tool.pytest_env]
|
||||||
ENVIRONMENT = "pytest"
|
ENVIRONMENT = "pytest"
|
||||||
DATABASE_URL = "postgresql://test_user:test_password@localhost:15432/reflector_test"
|
DATABASE_URL = "postgresql://test_user:test_password@localhost:15432/reflector_test"
|
||||||
|
AUTH_BACKEND = "jwt"
|
||||||
|
|
||||||
[tool.pytest.ini_options]
|
[tool.pytest.ini_options]
|
||||||
addopts = "-ra -q --disable-pytest-warnings --cov --cov-report html -v"
|
addopts = "-ra -q --disable-pytest-warnings --cov --cov-report html -v"
|
||||||
testpaths = ["tests"]
|
testpaths = ["tests"]
|
||||||
asyncio_mode = "auto"
|
asyncio_mode = "auto"
|
||||||
|
markers = [
|
||||||
|
"model_api: tests for the unified model-serving HTTP API (backend- and hardware-agnostic)",
|
||||||
|
]
|
||||||
|
|
||||||
[tool.ruff.lint]
|
[tool.ruff.lint]
|
||||||
select = [
|
select = [
|
||||||
@@ -105,7 +131,7 @@ select = [
|
|||||||
|
|
||||||
[tool.ruff.lint.per-file-ignores]
|
[tool.ruff.lint.per-file-ignores]
|
||||||
"reflector/processors/summary/summary_builder.py" = ["E501"]
|
"reflector/processors/summary/summary_builder.py" = ["E501"]
|
||||||
"gpu/**.py" = ["PLC0415"]
|
"gpu/modal_deployments/**.py" = ["PLC0415"]
|
||||||
"reflector/tools/**.py" = ["PLC0415"]
|
"reflector/tools/**.py" = ["PLC0415"]
|
||||||
"migrations/versions/**.py" = ["PLC0415"]
|
"migrations/versions/**.py" = ["PLC0415"]
|
||||||
"tests/**.py" = ["PLC0415"]
|
"tests/**.py" = ["PLC0415"]
|
||||||
|
|||||||
@@ -12,6 +12,7 @@ from reflector.events import subscribers_shutdown, subscribers_startup
|
|||||||
from reflector.logger import logger
|
from reflector.logger import logger
|
||||||
from reflector.metrics import metrics_init
|
from reflector.metrics import metrics_init
|
||||||
from reflector.settings import settings
|
from reflector.settings import settings
|
||||||
|
from reflector.views.daily import router as daily_router
|
||||||
from reflector.views.meetings import router as meetings_router
|
from reflector.views.meetings import router as meetings_router
|
||||||
from reflector.views.rooms import router as rooms_router
|
from reflector.views.rooms import router as rooms_router
|
||||||
from reflector.views.rtc_offer import router as rtc_offer_router
|
from reflector.views.rtc_offer import router as rtc_offer_router
|
||||||
@@ -26,6 +27,8 @@ from reflector.views.transcripts_upload import router as transcripts_upload_rout
|
|||||||
from reflector.views.transcripts_webrtc import router as transcripts_webrtc_router
|
from reflector.views.transcripts_webrtc import router as transcripts_webrtc_router
|
||||||
from reflector.views.transcripts_websocket import router as transcripts_websocket_router
|
from reflector.views.transcripts_websocket import router as transcripts_websocket_router
|
||||||
from reflector.views.user import router as user_router
|
from reflector.views.user import router as user_router
|
||||||
|
from reflector.views.user_api_keys import router as user_api_keys_router
|
||||||
|
from reflector.views.user_websocket import router as user_ws_router
|
||||||
from reflector.views.whereby import router as whereby_router
|
from reflector.views.whereby import router as whereby_router
|
||||||
from reflector.views.zulip import router as zulip_router
|
from reflector.views.zulip import router as zulip_router
|
||||||
|
|
||||||
@@ -65,6 +68,12 @@ app.add_middleware(
|
|||||||
allow_headers=["*"],
|
allow_headers=["*"],
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@app.get("/health")
|
||||||
|
async def health():
|
||||||
|
return {"status": "healthy"}
|
||||||
|
|
||||||
|
|
||||||
# metrics
|
# metrics
|
||||||
instrumentator = Instrumentator(
|
instrumentator = Instrumentator(
|
||||||
excluded_handlers=["/docs", "/metrics"],
|
excluded_handlers=["/docs", "/metrics"],
|
||||||
@@ -84,8 +93,11 @@ app.include_router(transcripts_websocket_router, prefix="/v1")
|
|||||||
app.include_router(transcripts_webrtc_router, prefix="/v1")
|
app.include_router(transcripts_webrtc_router, prefix="/v1")
|
||||||
app.include_router(transcripts_process_router, prefix="/v1")
|
app.include_router(transcripts_process_router, prefix="/v1")
|
||||||
app.include_router(user_router, prefix="/v1")
|
app.include_router(user_router, prefix="/v1")
|
||||||
|
app.include_router(user_api_keys_router, prefix="/v1")
|
||||||
|
app.include_router(user_ws_router, prefix="/v1")
|
||||||
app.include_router(zulip_router, prefix="/v1")
|
app.include_router(zulip_router, prefix="/v1")
|
||||||
app.include_router(whereby_router, prefix="/v1")
|
app.include_router(whereby_router, prefix="/v1")
|
||||||
|
app.include_router(daily_router, prefix="/v1/daily")
|
||||||
add_pagination(app)
|
add_pagination(app)
|
||||||
|
|
||||||
# prepare celery
|
# prepare celery
|
||||||
|
|||||||
27
server/reflector/asynctask.py
Normal file
27
server/reflector/asynctask.py
Normal file
@@ -0,0 +1,27 @@
|
|||||||
|
import asyncio
|
||||||
|
import functools
|
||||||
|
|
||||||
|
from reflector.db import get_database
|
||||||
|
|
||||||
|
|
||||||
|
def asynctask(f):
|
||||||
|
@functools.wraps(f)
|
||||||
|
def wrapper(*args, **kwargs):
|
||||||
|
async def run_with_db():
|
||||||
|
database = get_database()
|
||||||
|
await database.connect()
|
||||||
|
try:
|
||||||
|
return await f(*args, **kwargs)
|
||||||
|
finally:
|
||||||
|
await database.disconnect()
|
||||||
|
|
||||||
|
coro = run_with_db()
|
||||||
|
try:
|
||||||
|
loop = asyncio.get_running_loop()
|
||||||
|
except RuntimeError:
|
||||||
|
loop = None
|
||||||
|
if loop and loop.is_running():
|
||||||
|
return loop.run_until_complete(coro)
|
||||||
|
return asyncio.run(coro)
|
||||||
|
|
||||||
|
return wrapper
|
||||||
@@ -1,14 +1,16 @@
|
|||||||
from typing import Annotated, Optional
|
from typing import Annotated, List, Optional
|
||||||
|
|
||||||
from fastapi import Depends, HTTPException
|
from fastapi import Depends, HTTPException
|
||||||
from fastapi.security import OAuth2PasswordBearer
|
from fastapi.security import APIKeyHeader, OAuth2PasswordBearer
|
||||||
from jose import JWTError, jwt
|
from jose import JWTError, jwt
|
||||||
from pydantic import BaseModel
|
from pydantic import BaseModel
|
||||||
|
|
||||||
|
from reflector.db.user_api_keys import user_api_keys_controller
|
||||||
from reflector.logger import logger
|
from reflector.logger import logger
|
||||||
from reflector.settings import settings
|
from reflector.settings import settings
|
||||||
|
|
||||||
oauth2_scheme = OAuth2PasswordBearer(tokenUrl="token", auto_error=False)
|
oauth2_scheme = OAuth2PasswordBearer(tokenUrl="token", auto_error=False)
|
||||||
|
api_key_header = APIKeyHeader(name="X-API-Key", auto_error=False)
|
||||||
|
|
||||||
jwt_public_key = open(f"reflector/auth/jwt/keys/{settings.AUTH_JWT_PUBLIC_KEY}").read()
|
jwt_public_key = open(f"reflector/auth/jwt/keys/{settings.AUTH_JWT_PUBLIC_KEY}").read()
|
||||||
jwt_algorithm = settings.AUTH_JWT_ALGORITHM
|
jwt_algorithm = settings.AUTH_JWT_ALGORITHM
|
||||||
@@ -26,7 +28,7 @@ class JWTException(Exception):
|
|||||||
|
|
||||||
class UserInfo(BaseModel):
|
class UserInfo(BaseModel):
|
||||||
sub: str
|
sub: str
|
||||||
email: str
|
email: Optional[str] = None
|
||||||
|
|
||||||
def __getitem__(self, key):
|
def __getitem__(self, key):
|
||||||
return getattr(self, key)
|
return getattr(self, key)
|
||||||
@@ -58,33 +60,53 @@ def authenticated(token: Annotated[str, Depends(oauth2_scheme)]):
|
|||||||
return None
|
return None
|
||||||
|
|
||||||
|
|
||||||
def current_user(
|
async def _authenticate_user(
|
||||||
token: Annotated[Optional[str], Depends(oauth2_scheme)],
|
jwt_token: Optional[str],
|
||||||
jwtauth: JWTAuth = Depends(),
|
api_key: Optional[str],
|
||||||
):
|
jwtauth: JWTAuth,
|
||||||
if token is None:
|
) -> UserInfo | None:
|
||||||
raise HTTPException(status_code=401, detail="Not authenticated")
|
user_infos: List[UserInfo] = []
|
||||||
try:
|
if api_key:
|
||||||
payload = jwtauth.verify_token(token)
|
user_api_key = await user_api_keys_controller.verify_key(api_key)
|
||||||
sub = payload["sub"]
|
if user_api_key:
|
||||||
return UserInfo(sub=sub)
|
user_infos.append(UserInfo(sub=user_api_key.user_id, email=None))
|
||||||
except JWTError as e:
|
|
||||||
logger.error(f"JWT error: {e}")
|
|
||||||
raise HTTPException(status_code=401, detail="Invalid authentication")
|
|
||||||
|
|
||||||
|
if jwt_token:
|
||||||
def current_user_optional(
|
|
||||||
token: Annotated[Optional[str], Depends(oauth2_scheme)],
|
|
||||||
jwtauth: JWTAuth = Depends(),
|
|
||||||
):
|
|
||||||
# we accept no token, but if one is provided, it must be a valid one.
|
|
||||||
if token is None:
|
|
||||||
return None
|
|
||||||
try:
|
try:
|
||||||
payload = jwtauth.verify_token(token)
|
payload = jwtauth.verify_token(jwt_token)
|
||||||
sub = payload["sub"]
|
sub = payload["sub"]
|
||||||
email = payload["email"]
|
email = payload["email"]
|
||||||
return UserInfo(sub=sub, email=email)
|
user_infos.append(UserInfo(sub=sub, email=email))
|
||||||
except JWTError as e:
|
except JWTError as e:
|
||||||
logger.error(f"JWT error: {e}")
|
logger.error(f"JWT error: {e}")
|
||||||
raise HTTPException(status_code=401, detail="Invalid authentication")
|
raise HTTPException(status_code=401, detail="Invalid authentication")
|
||||||
|
|
||||||
|
if len(user_infos) == 0:
|
||||||
|
return None
|
||||||
|
|
||||||
|
if len(set([x.sub for x in user_infos])) > 1:
|
||||||
|
raise JWTException(
|
||||||
|
status_code=401,
|
||||||
|
detail="Invalid authentication: more than one user provided",
|
||||||
|
)
|
||||||
|
|
||||||
|
return user_infos[0]
|
||||||
|
|
||||||
|
|
||||||
|
async def current_user(
|
||||||
|
jwt_token: Annotated[Optional[str], Depends(oauth2_scheme)],
|
||||||
|
api_key: Annotated[Optional[str], Depends(api_key_header)],
|
||||||
|
jwtauth: JWTAuth = Depends(),
|
||||||
|
):
|
||||||
|
user = await _authenticate_user(jwt_token, api_key, jwtauth)
|
||||||
|
if user is None:
|
||||||
|
raise HTTPException(status_code=401, detail="Not authenticated")
|
||||||
|
return user
|
||||||
|
|
||||||
|
|
||||||
|
async def current_user_optional(
|
||||||
|
jwt_token: Annotated[Optional[str], Depends(oauth2_scheme)],
|
||||||
|
api_key: Annotated[Optional[str], Depends(api_key_header)],
|
||||||
|
jwtauth: JWTAuth = Depends(),
|
||||||
|
):
|
||||||
|
return await _authenticate_user(jwt_token, api_key, jwtauth)
|
||||||
|
|||||||
@@ -29,6 +29,7 @@ import reflector.db.meetings # noqa
|
|||||||
import reflector.db.recordings # noqa
|
import reflector.db.recordings # noqa
|
||||||
import reflector.db.rooms # noqa
|
import reflector.db.rooms # noqa
|
||||||
import reflector.db.transcripts # noqa
|
import reflector.db.transcripts # noqa
|
||||||
|
import reflector.db.user_api_keys # noqa
|
||||||
|
|
||||||
kwargs = {}
|
kwargs = {}
|
||||||
if "postgres" not in settings.DATABASE_URL:
|
if "postgres" not in settings.DATABASE_URL:
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
from datetime import datetime, timezone
|
from datetime import datetime, timedelta, timezone
|
||||||
from typing import Any
|
from typing import Any
|
||||||
|
|
||||||
import sqlalchemy as sa
|
import sqlalchemy as sa
|
||||||
@@ -15,7 +15,7 @@ calendar_events = sa.Table(
|
|||||||
sa.Column(
|
sa.Column(
|
||||||
"room_id",
|
"room_id",
|
||||||
sa.String,
|
sa.String,
|
||||||
sa.ForeignKey("room.id", ondelete="CASCADE"),
|
sa.ForeignKey("room.id", ondelete="CASCADE", name="fk_calendar_event_room_id"),
|
||||||
nullable=False,
|
nullable=False,
|
||||||
),
|
),
|
||||||
sa.Column("ics_uid", sa.Text, nullable=False),
|
sa.Column("ics_uid", sa.Text, nullable=False),
|
||||||
@@ -65,7 +65,6 @@ class CalendarEventController:
|
|||||||
start_after: datetime | None = None,
|
start_after: datetime | None = None,
|
||||||
end_before: datetime | None = None,
|
end_before: datetime | None = None,
|
||||||
) -> list[CalendarEvent]:
|
) -> list[CalendarEvent]:
|
||||||
"""Get calendar events for a room."""
|
|
||||||
query = calendar_events.select().where(calendar_events.c.room_id == room_id)
|
query = calendar_events.select().where(calendar_events.c.room_id == room_id)
|
||||||
|
|
||||||
if not include_deleted:
|
if not include_deleted:
|
||||||
@@ -83,9 +82,9 @@ class CalendarEventController:
|
|||||||
return [CalendarEvent(**result) for result in results]
|
return [CalendarEvent(**result) for result in results]
|
||||||
|
|
||||||
async def get_upcoming(
|
async def get_upcoming(
|
||||||
self, room_id: str, minutes_ahead: int = 30
|
self, room_id: str, minutes_ahead: int = 120
|
||||||
) -> list[CalendarEvent]:
|
) -> list[CalendarEvent]:
|
||||||
"""Get upcoming events for a room within the specified minutes."""
|
"""Get upcoming events for a room within the specified minutes, including currently happening events."""
|
||||||
now = datetime.now(timezone.utc)
|
now = datetime.now(timezone.utc)
|
||||||
future_time = now + timedelta(minutes=minutes_ahead)
|
future_time = now + timedelta(minutes=minutes_ahead)
|
||||||
|
|
||||||
@@ -95,8 +94,8 @@ class CalendarEventController:
|
|||||||
sa.and_(
|
sa.and_(
|
||||||
calendar_events.c.room_id == room_id,
|
calendar_events.c.room_id == room_id,
|
||||||
calendar_events.c.is_deleted == False,
|
calendar_events.c.is_deleted == False,
|
||||||
calendar_events.c.start_time >= now,
|
|
||||||
calendar_events.c.start_time <= future_time,
|
calendar_events.c.start_time <= future_time,
|
||||||
|
calendar_events.c.end_time >= now,
|
||||||
)
|
)
|
||||||
)
|
)
|
||||||
.order_by(calendar_events.c.start_time.asc())
|
.order_by(calendar_events.c.start_time.asc())
|
||||||
@@ -105,8 +104,12 @@ class CalendarEventController:
|
|||||||
results = await get_database().fetch_all(query)
|
results = await get_database().fetch_all(query)
|
||||||
return [CalendarEvent(**result) for result in results]
|
return [CalendarEvent(**result) for result in results]
|
||||||
|
|
||||||
|
async def get_by_id(self, event_id: str) -> CalendarEvent | None:
|
||||||
|
query = calendar_events.select().where(calendar_events.c.id == event_id)
|
||||||
|
result = await get_database().fetch_one(query)
|
||||||
|
return CalendarEvent(**result) if result else None
|
||||||
|
|
||||||
async def get_by_ics_uid(self, room_id: str, ics_uid: str) -> CalendarEvent | None:
|
async def get_by_ics_uid(self, room_id: str, ics_uid: str) -> CalendarEvent | None:
|
||||||
"""Get a calendar event by its ICS UID."""
|
|
||||||
query = calendar_events.select().where(
|
query = calendar_events.select().where(
|
||||||
sa.and_(
|
sa.and_(
|
||||||
calendar_events.c.room_id == room_id,
|
calendar_events.c.room_id == room_id,
|
||||||
@@ -117,11 +120,9 @@ class CalendarEventController:
|
|||||||
return CalendarEvent(**result) if result else None
|
return CalendarEvent(**result) if result else None
|
||||||
|
|
||||||
async def upsert(self, event: CalendarEvent) -> CalendarEvent:
|
async def upsert(self, event: CalendarEvent) -> CalendarEvent:
|
||||||
"""Create or update a calendar event."""
|
|
||||||
existing = await self.get_by_ics_uid(event.room_id, event.ics_uid)
|
existing = await self.get_by_ics_uid(event.room_id, event.ics_uid)
|
||||||
|
|
||||||
if existing:
|
if existing:
|
||||||
# Update existing event
|
|
||||||
event.id = existing.id
|
event.id = existing.id
|
||||||
event.created_at = existing.created_at
|
event.created_at = existing.created_at
|
||||||
event.updated_at = datetime.now(timezone.utc)
|
event.updated_at = datetime.now(timezone.utc)
|
||||||
@@ -132,7 +133,6 @@ class CalendarEventController:
|
|||||||
.values(**event.model_dump())
|
.values(**event.model_dump())
|
||||||
)
|
)
|
||||||
else:
|
else:
|
||||||
# Insert new event
|
|
||||||
query = calendar_events.insert().values(**event.model_dump())
|
query = calendar_events.insert().values(**event.model_dump())
|
||||||
|
|
||||||
await get_database().execute(query)
|
await get_database().execute(query)
|
||||||
@@ -144,7 +144,6 @@ class CalendarEventController:
|
|||||||
"""Soft delete future events that are no longer in the calendar."""
|
"""Soft delete future events that are no longer in the calendar."""
|
||||||
now = datetime.now(timezone.utc)
|
now = datetime.now(timezone.utc)
|
||||||
|
|
||||||
# First, get the IDs of events to delete
|
|
||||||
select_query = calendar_events.select().where(
|
select_query = calendar_events.select().where(
|
||||||
sa.and_(
|
sa.and_(
|
||||||
calendar_events.c.room_id == room_id,
|
calendar_events.c.room_id == room_id,
|
||||||
@@ -160,7 +159,6 @@ class CalendarEventController:
|
|||||||
delete_count = len(to_delete)
|
delete_count = len(to_delete)
|
||||||
|
|
||||||
if delete_count > 0:
|
if delete_count > 0:
|
||||||
# Now update them
|
|
||||||
update_query = (
|
update_query = (
|
||||||
calendar_events.update()
|
calendar_events.update()
|
||||||
.where(
|
.where(
|
||||||
@@ -181,13 +179,9 @@ class CalendarEventController:
|
|||||||
return delete_count
|
return delete_count
|
||||||
|
|
||||||
async def delete_by_room(self, room_id: str) -> int:
|
async def delete_by_room(self, room_id: str) -> int:
|
||||||
"""Hard delete all events for a room (used when room is deleted)."""
|
|
||||||
query = calendar_events.delete().where(calendar_events.c.room_id == room_id)
|
query = calendar_events.delete().where(calendar_events.c.room_id == room_id)
|
||||||
result = await get_database().execute(query)
|
result = await get_database().execute(query)
|
||||||
return result.rowcount
|
return result.rowcount
|
||||||
|
|
||||||
|
|
||||||
# Add missing import
|
|
||||||
from datetime import timedelta
|
|
||||||
|
|
||||||
calendar_events_controller = CalendarEventController()
|
calendar_events_controller = CalendarEventController()
|
||||||
|
|||||||
@@ -2,13 +2,15 @@ from datetime import datetime
|
|||||||
from typing import Any, Literal
|
from typing import Any, Literal
|
||||||
|
|
||||||
import sqlalchemy as sa
|
import sqlalchemy as sa
|
||||||
from fastapi import HTTPException
|
|
||||||
from pydantic import BaseModel, Field
|
from pydantic import BaseModel, Field
|
||||||
from sqlalchemy.dialects.postgresql import JSONB
|
from sqlalchemy.dialects.postgresql import JSONB
|
||||||
|
|
||||||
from reflector.db import get_database, metadata
|
from reflector.db import get_database, metadata
|
||||||
from reflector.db.rooms import Room
|
from reflector.db.rooms import Room
|
||||||
|
from reflector.schemas.platform import WHEREBY_PLATFORM, Platform
|
||||||
from reflector.utils import generate_uuid4
|
from reflector.utils import generate_uuid4
|
||||||
|
from reflector.utils.string import assert_equal
|
||||||
|
from reflector.video_platforms.factory import get_platform
|
||||||
|
|
||||||
meetings = sa.Table(
|
meetings = sa.Table(
|
||||||
"meeting",
|
"meeting",
|
||||||
@@ -19,8 +21,12 @@ meetings = sa.Table(
|
|||||||
sa.Column("host_room_url", sa.String),
|
sa.Column("host_room_url", sa.String),
|
||||||
sa.Column("start_date", sa.DateTime(timezone=True)),
|
sa.Column("start_date", sa.DateTime(timezone=True)),
|
||||||
sa.Column("end_date", sa.DateTime(timezone=True)),
|
sa.Column("end_date", sa.DateTime(timezone=True)),
|
||||||
sa.Column("user_id", sa.String),
|
sa.Column(
|
||||||
sa.Column("room_id", sa.String),
|
"room_id",
|
||||||
|
sa.String,
|
||||||
|
sa.ForeignKey("room.id", ondelete="CASCADE"),
|
||||||
|
nullable=True,
|
||||||
|
),
|
||||||
sa.Column("is_locked", sa.Boolean, nullable=False, server_default=sa.false()),
|
sa.Column("is_locked", sa.Boolean, nullable=False, server_default=sa.false()),
|
||||||
sa.Column("room_mode", sa.String, nullable=False, server_default="normal"),
|
sa.Column("room_mode", sa.String, nullable=False, server_default="normal"),
|
||||||
sa.Column("recording_type", sa.String, nullable=False, server_default="cloud"),
|
sa.Column("recording_type", sa.String, nullable=False, server_default="cloud"),
|
||||||
@@ -45,11 +51,19 @@ meetings = sa.Table(
|
|||||||
sa.Column(
|
sa.Column(
|
||||||
"calendar_event_id",
|
"calendar_event_id",
|
||||||
sa.String,
|
sa.String,
|
||||||
sa.ForeignKey("calendar_event.id", ondelete="SET NULL"),
|
sa.ForeignKey(
|
||||||
|
"calendar_event.id",
|
||||||
|
ondelete="SET NULL",
|
||||||
|
name="fk_meeting_calendar_event_id",
|
||||||
|
),
|
||||||
),
|
),
|
||||||
sa.Column("calendar_metadata", JSONB),
|
sa.Column("calendar_metadata", JSONB),
|
||||||
sa.Column("last_participant_left_at", sa.DateTime(timezone=True)),
|
sa.Column(
|
||||||
sa.Column("grace_period_minutes", sa.Integer, server_default=sa.text("15")),
|
"platform",
|
||||||
|
sa.String,
|
||||||
|
nullable=False,
|
||||||
|
server_default=assert_equal(WHEREBY_PLATFORM, "whereby"),
|
||||||
|
),
|
||||||
sa.Index("idx_meeting_room_id", "room_id"),
|
sa.Index("idx_meeting_room_id", "room_id"),
|
||||||
sa.Index("idx_meeting_calendar_event", "calendar_event_id"),
|
sa.Index("idx_meeting_calendar_event", "calendar_event_id"),
|
||||||
)
|
)
|
||||||
@@ -58,7 +72,12 @@ meeting_consent = sa.Table(
|
|||||||
"meeting_consent",
|
"meeting_consent",
|
||||||
metadata,
|
metadata,
|
||||||
sa.Column("id", sa.String, primary_key=True),
|
sa.Column("id", sa.String, primary_key=True),
|
||||||
sa.Column("meeting_id", sa.String, sa.ForeignKey("meeting.id"), nullable=False),
|
sa.Column(
|
||||||
|
"meeting_id",
|
||||||
|
sa.String,
|
||||||
|
sa.ForeignKey("meeting.id", ondelete="CASCADE"),
|
||||||
|
nullable=False,
|
||||||
|
),
|
||||||
sa.Column("user_id", sa.String),
|
sa.Column("user_id", sa.String),
|
||||||
sa.Column("consent_given", sa.Boolean, nullable=False),
|
sa.Column("consent_given", sa.Boolean, nullable=False),
|
||||||
sa.Column("consent_timestamp", sa.DateTime(timezone=True), nullable=False),
|
sa.Column("consent_timestamp", sa.DateTime(timezone=True), nullable=False),
|
||||||
@@ -80,20 +99,18 @@ class Meeting(BaseModel):
|
|||||||
host_room_url: str
|
host_room_url: str
|
||||||
start_date: datetime
|
start_date: datetime
|
||||||
end_date: datetime
|
end_date: datetime
|
||||||
user_id: str | None = None
|
room_id: str | None
|
||||||
room_id: str | None = None
|
|
||||||
is_locked: bool = False
|
is_locked: bool = False
|
||||||
room_mode: Literal["normal", "group"] = "normal"
|
room_mode: Literal["normal", "group"] = "normal"
|
||||||
recording_type: Literal["none", "local", "cloud"] = "cloud"
|
recording_type: Literal["none", "local", "cloud"] = "cloud"
|
||||||
recording_trigger: Literal[
|
recording_trigger: Literal[ # whereby-specific
|
||||||
"none", "prompt", "automatic", "automatic-2nd-participant"
|
"none", "prompt", "automatic", "automatic-2nd-participant"
|
||||||
] = "automatic-2nd-participant"
|
] = "automatic-2nd-participant"
|
||||||
num_clients: int = 0
|
num_clients: int = 0
|
||||||
is_active: bool = True
|
is_active: bool = True
|
||||||
calendar_event_id: str | None = None
|
calendar_event_id: str | None = None
|
||||||
calendar_metadata: dict[str, Any] | None = None
|
calendar_metadata: dict[str, Any] | None = None
|
||||||
last_participant_left_at: datetime | None = None
|
platform: Platform = WHEREBY_PLATFORM
|
||||||
grace_period_minutes: int = 15
|
|
||||||
|
|
||||||
|
|
||||||
class MeetingController:
|
class MeetingController:
|
||||||
@@ -105,14 +122,10 @@ class MeetingController:
|
|||||||
host_room_url: str,
|
host_room_url: str,
|
||||||
start_date: datetime,
|
start_date: datetime,
|
||||||
end_date: datetime,
|
end_date: datetime,
|
||||||
user_id: str,
|
|
||||||
room: Room,
|
room: Room,
|
||||||
calendar_event_id: str | None = None,
|
calendar_event_id: str | None = None,
|
||||||
calendar_metadata: dict[str, Any] | None = None,
|
calendar_metadata: dict[str, Any] | None = None,
|
||||||
):
|
):
|
||||||
"""
|
|
||||||
Create a new meeting
|
|
||||||
"""
|
|
||||||
meeting = Meeting(
|
meeting = Meeting(
|
||||||
id=id,
|
id=id,
|
||||||
room_name=room_name,
|
room_name=room_name,
|
||||||
@@ -120,7 +133,6 @@ class MeetingController:
|
|||||||
host_room_url=host_room_url,
|
host_room_url=host_room_url,
|
||||||
start_date=start_date,
|
start_date=start_date,
|
||||||
end_date=end_date,
|
end_date=end_date,
|
||||||
user_id=user_id,
|
|
||||||
room_id=room.id,
|
room_id=room.id,
|
||||||
is_locked=room.is_locked,
|
is_locked=room.is_locked,
|
||||||
room_mode=room.room_mode,
|
room_mode=room.room_mode,
|
||||||
@@ -128,33 +140,36 @@ class MeetingController:
|
|||||||
recording_trigger=room.recording_trigger,
|
recording_trigger=room.recording_trigger,
|
||||||
calendar_event_id=calendar_event_id,
|
calendar_event_id=calendar_event_id,
|
||||||
calendar_metadata=calendar_metadata,
|
calendar_metadata=calendar_metadata,
|
||||||
|
platform=get_platform(room.platform),
|
||||||
)
|
)
|
||||||
query = meetings.insert().values(**meeting.model_dump())
|
query = meetings.insert().values(**meeting.model_dump())
|
||||||
await get_database().execute(query)
|
await get_database().execute(query)
|
||||||
return meeting
|
return meeting
|
||||||
|
|
||||||
async def get_all_active(self) -> list[Meeting]:
|
async def get_all_active(self) -> list[Meeting]:
|
||||||
"""
|
|
||||||
Get active meetings.
|
|
||||||
"""
|
|
||||||
query = meetings.select().where(meetings.c.is_active)
|
query = meetings.select().where(meetings.c.is_active)
|
||||||
return await get_database().fetch_all(query)
|
results = await get_database().fetch_all(query)
|
||||||
|
return [Meeting(**result) for result in results]
|
||||||
|
|
||||||
async def get_by_room_name(
|
async def get_by_room_name(
|
||||||
self,
|
self,
|
||||||
room_name: str,
|
room_name: str,
|
||||||
) -> Meeting:
|
) -> Meeting | None:
|
||||||
"""
|
"""
|
||||||
Get a meeting by room name.
|
Get a meeting by room name.
|
||||||
|
For backward compatibility, returns the most recent meeting.
|
||||||
"""
|
"""
|
||||||
query = meetings.select().where(meetings.c.room_name == room_name)
|
query = (
|
||||||
|
meetings.select()
|
||||||
|
.where(meetings.c.room_name == room_name)
|
||||||
|
.order_by(meetings.c.end_date.desc())
|
||||||
|
)
|
||||||
result = await get_database().fetch_one(query)
|
result = await get_database().fetch_one(query)
|
||||||
if not result:
|
if not result:
|
||||||
return None
|
return None
|
||||||
|
|
||||||
return Meeting(**result)
|
return Meeting(**result)
|
||||||
|
|
||||||
async def get_active(self, room: Room, current_time: datetime) -> Meeting:
|
async def get_active(self, room: Room, current_time: datetime) -> Meeting | None:
|
||||||
"""
|
"""
|
||||||
Get latest active meeting for a room.
|
Get latest active meeting for a room.
|
||||||
For backward compatibility, returns the most recent active meeting.
|
For backward compatibility, returns the most recent active meeting.
|
||||||
@@ -174,16 +189,11 @@ class MeetingController:
|
|||||||
result = await get_database().fetch_one(query)
|
result = await get_database().fetch_one(query)
|
||||||
if not result:
|
if not result:
|
||||||
return None
|
return None
|
||||||
|
|
||||||
return Meeting(**result)
|
return Meeting(**result)
|
||||||
|
|
||||||
async def get_all_active_for_room(
|
async def get_all_active_for_room(
|
||||||
self, room: Room, current_time: datetime
|
self, room: Room, current_time: datetime
|
||||||
) -> list[Meeting]:
|
) -> list[Meeting]:
|
||||||
"""
|
|
||||||
Get all active meetings for a room.
|
|
||||||
This supports multiple concurrent meetings per room.
|
|
||||||
"""
|
|
||||||
end_date = getattr(meetings.c, "end_date")
|
end_date = getattr(meetings.c, "end_date")
|
||||||
query = (
|
query = (
|
||||||
meetings.select()
|
meetings.select()
|
||||||
@@ -218,37 +228,27 @@ class MeetingController:
|
|||||||
return None
|
return None
|
||||||
return Meeting(**result)
|
return Meeting(**result)
|
||||||
|
|
||||||
async def get_by_id(self, meeting_id: str, **kwargs) -> Meeting | None:
|
async def get_by_id(
|
||||||
"""
|
self, meeting_id: str, room: Room | None = None
|
||||||
Get a meeting by id
|
) -> Meeting | None:
|
||||||
"""
|
|
||||||
query = meetings.select().where(meetings.c.id == meeting_id)
|
query = meetings.select().where(meetings.c.id == meeting_id)
|
||||||
|
|
||||||
|
if room:
|
||||||
|
query = query.where(meetings.c.room_id == room.id)
|
||||||
|
|
||||||
result = await get_database().fetch_one(query)
|
result = await get_database().fetch_one(query)
|
||||||
if not result:
|
if not result:
|
||||||
return None
|
return None
|
||||||
return Meeting(**result)
|
return Meeting(**result)
|
||||||
|
|
||||||
async def get_by_id_for_http(self, meeting_id: str, user_id: str | None) -> Meeting:
|
async def get_by_calendar_event(
|
||||||
"""
|
self, calendar_event_id: str, room: Room
|
||||||
Get a meeting by ID for HTTP request.
|
) -> Meeting | None:
|
||||||
|
|
||||||
If not found, it will raise a 404 error.
|
|
||||||
"""
|
|
||||||
query = meetings.select().where(meetings.c.id == meeting_id)
|
|
||||||
result = await get_database().fetch_one(query)
|
|
||||||
if not result:
|
|
||||||
raise HTTPException(status_code=404, detail="Meeting not found")
|
|
||||||
|
|
||||||
meeting = Meeting(**result)
|
|
||||||
if result["user_id"] != user_id:
|
|
||||||
meeting.host_room_url = ""
|
|
||||||
|
|
||||||
return meeting
|
|
||||||
|
|
||||||
async def get_by_calendar_event(self, calendar_event_id: str) -> Meeting | None:
|
|
||||||
query = meetings.select().where(
|
query = meetings.select().where(
|
||||||
meetings.c.calendar_event_id == calendar_event_id
|
meetings.c.calendar_event_id == calendar_event_id
|
||||||
)
|
)
|
||||||
|
if room:
|
||||||
|
query = query.where(meetings.c.room_id == room.id)
|
||||||
result = await get_database().fetch_one(query)
|
result = await get_database().fetch_one(query)
|
||||||
if not result:
|
if not result:
|
||||||
return None
|
return None
|
||||||
@@ -258,6 +258,28 @@ class MeetingController:
|
|||||||
query = meetings.update().where(meetings.c.id == meeting_id).values(**kwargs)
|
query = meetings.update().where(meetings.c.id == meeting_id).values(**kwargs)
|
||||||
await get_database().execute(query)
|
await get_database().execute(query)
|
||||||
|
|
||||||
|
async def increment_num_clients(self, meeting_id: str) -> None:
|
||||||
|
"""Atomically increment participant count."""
|
||||||
|
query = (
|
||||||
|
meetings.update()
|
||||||
|
.where(meetings.c.id == meeting_id)
|
||||||
|
.values(num_clients=meetings.c.num_clients + 1)
|
||||||
|
)
|
||||||
|
await get_database().execute(query)
|
||||||
|
|
||||||
|
async def decrement_num_clients(self, meeting_id: str) -> None:
|
||||||
|
"""Atomically decrement participant count (min 0)."""
|
||||||
|
query = (
|
||||||
|
meetings.update()
|
||||||
|
.where(meetings.c.id == meeting_id)
|
||||||
|
.values(
|
||||||
|
num_clients=sa.case(
|
||||||
|
(meetings.c.num_clients > 0, meetings.c.num_clients - 1), else_=0
|
||||||
|
)
|
||||||
|
)
|
||||||
|
)
|
||||||
|
await get_database().execute(query)
|
||||||
|
|
||||||
|
|
||||||
class MeetingConsentController:
|
class MeetingConsentController:
|
||||||
async def get_by_meeting_id(self, meeting_id: str) -> list[MeetingConsent]:
|
async def get_by_meeting_id(self, meeting_id: str) -> list[MeetingConsent]:
|
||||||
@@ -278,10 +300,9 @@ class MeetingConsentController:
|
|||||||
result = await get_database().fetch_one(query)
|
result = await get_database().fetch_one(query)
|
||||||
if result is None:
|
if result is None:
|
||||||
return None
|
return None
|
||||||
return MeetingConsent(**result) if result else None
|
return MeetingConsent(**result)
|
||||||
|
|
||||||
async def upsert(self, consent: MeetingConsent) -> MeetingConsent:
|
async def upsert(self, consent: MeetingConsent) -> MeetingConsent:
|
||||||
"""Create new consent or update existing one for authenticated users"""
|
|
||||||
if consent.user_id:
|
if consent.user_id:
|
||||||
# For authenticated users, check if consent already exists
|
# For authenticated users, check if consent already exists
|
||||||
# not transactional but we're ok with that; the consents ain't deleted anyways
|
# not transactional but we're ok with that; the consents ain't deleted anyways
|
||||||
|
|||||||
@@ -21,6 +21,7 @@ recordings = sa.Table(
|
|||||||
server_default="pending",
|
server_default="pending",
|
||||||
),
|
),
|
||||||
sa.Column("meeting_id", sa.String),
|
sa.Column("meeting_id", sa.String),
|
||||||
|
sa.Column("track_keys", sa.JSON, nullable=True),
|
||||||
sa.Index("idx_recording_meeting_id", "meeting_id"),
|
sa.Index("idx_recording_meeting_id", "meeting_id"),
|
||||||
)
|
)
|
||||||
|
|
||||||
@@ -28,10 +29,13 @@ recordings = sa.Table(
|
|||||||
class Recording(BaseModel):
|
class Recording(BaseModel):
|
||||||
id: str = Field(default_factory=generate_uuid4)
|
id: str = Field(default_factory=generate_uuid4)
|
||||||
bucket_name: str
|
bucket_name: str
|
||||||
|
# for single-track
|
||||||
object_key: str
|
object_key: str
|
||||||
recorded_at: datetime
|
recorded_at: datetime
|
||||||
status: Literal["pending", "processing", "completed", "failed"] = "pending"
|
status: Literal["pending", "processing", "completed", "failed"] = "pending"
|
||||||
meeting_id: str | None = None
|
meeting_id: str | None = None
|
||||||
|
# for multitrack reprocessing
|
||||||
|
track_keys: list[str] | None = None
|
||||||
|
|
||||||
|
|
||||||
class RecordingController:
|
class RecordingController:
|
||||||
|
|||||||
@@ -1,3 +1,4 @@
|
|||||||
|
import secrets
|
||||||
from datetime import datetime, timezone
|
from datetime import datetime, timezone
|
||||||
from sqlite3 import IntegrityError
|
from sqlite3 import IntegrityError
|
||||||
from typing import Literal
|
from typing import Literal
|
||||||
@@ -8,6 +9,7 @@ from pydantic import BaseModel, Field
|
|||||||
from sqlalchemy.sql import false, or_
|
from sqlalchemy.sql import false, or_
|
||||||
|
|
||||||
from reflector.db import get_database, metadata
|
from reflector.db import get_database, metadata
|
||||||
|
from reflector.schemas.platform import Platform
|
||||||
from reflector.utils import generate_uuid4
|
from reflector.utils import generate_uuid4
|
||||||
|
|
||||||
rooms = sqlalchemy.Table(
|
rooms = sqlalchemy.Table(
|
||||||
@@ -40,6 +42,8 @@ rooms = sqlalchemy.Table(
|
|||||||
sqlalchemy.Column(
|
sqlalchemy.Column(
|
||||||
"is_shared", sqlalchemy.Boolean, nullable=False, server_default=false()
|
"is_shared", sqlalchemy.Boolean, nullable=False, server_default=false()
|
||||||
),
|
),
|
||||||
|
sqlalchemy.Column("webhook_url", sqlalchemy.String, nullable=True),
|
||||||
|
sqlalchemy.Column("webhook_secret", sqlalchemy.String, nullable=True),
|
||||||
sqlalchemy.Column("ics_url", sqlalchemy.Text),
|
sqlalchemy.Column("ics_url", sqlalchemy.Text),
|
||||||
sqlalchemy.Column("ics_fetch_interval", sqlalchemy.Integer, server_default="300"),
|
sqlalchemy.Column("ics_fetch_interval", sqlalchemy.Integer, server_default="300"),
|
||||||
sqlalchemy.Column(
|
sqlalchemy.Column(
|
||||||
@@ -47,6 +51,12 @@ rooms = sqlalchemy.Table(
|
|||||||
),
|
),
|
||||||
sqlalchemy.Column("ics_last_sync", sqlalchemy.DateTime(timezone=True)),
|
sqlalchemy.Column("ics_last_sync", sqlalchemy.DateTime(timezone=True)),
|
||||||
sqlalchemy.Column("ics_last_etag", sqlalchemy.Text),
|
sqlalchemy.Column("ics_last_etag", sqlalchemy.Text),
|
||||||
|
sqlalchemy.Column(
|
||||||
|
"platform",
|
||||||
|
sqlalchemy.String,
|
||||||
|
nullable=True,
|
||||||
|
server_default=None,
|
||||||
|
),
|
||||||
sqlalchemy.Index("idx_room_is_shared", "is_shared"),
|
sqlalchemy.Index("idx_room_is_shared", "is_shared"),
|
||||||
sqlalchemy.Index("idx_room_ics_enabled", "ics_enabled"),
|
sqlalchemy.Index("idx_room_ics_enabled", "ics_enabled"),
|
||||||
)
|
)
|
||||||
@@ -63,15 +73,18 @@ class Room(BaseModel):
|
|||||||
is_locked: bool = False
|
is_locked: bool = False
|
||||||
room_mode: Literal["normal", "group"] = "normal"
|
room_mode: Literal["normal", "group"] = "normal"
|
||||||
recording_type: Literal["none", "local", "cloud"] = "cloud"
|
recording_type: Literal["none", "local", "cloud"] = "cloud"
|
||||||
recording_trigger: Literal[
|
recording_trigger: Literal[ # whereby-specific
|
||||||
"none", "prompt", "automatic", "automatic-2nd-participant"
|
"none", "prompt", "automatic", "automatic-2nd-participant"
|
||||||
] = "automatic-2nd-participant"
|
] = "automatic-2nd-participant"
|
||||||
is_shared: bool = False
|
is_shared: bool = False
|
||||||
|
webhook_url: str | None = None
|
||||||
|
webhook_secret: str | None = None
|
||||||
ics_url: str | None = None
|
ics_url: str | None = None
|
||||||
ics_fetch_interval: int = 300
|
ics_fetch_interval: int = 300
|
||||||
ics_enabled: bool = False
|
ics_enabled: bool = False
|
||||||
ics_last_sync: datetime | None = None
|
ics_last_sync: datetime | None = None
|
||||||
ics_last_etag: str | None = None
|
ics_last_etag: str | None = None
|
||||||
|
platform: Platform | None = None
|
||||||
|
|
||||||
|
|
||||||
class RoomController:
|
class RoomController:
|
||||||
@@ -120,13 +133,19 @@ class RoomController:
|
|||||||
recording_type: str,
|
recording_type: str,
|
||||||
recording_trigger: str,
|
recording_trigger: str,
|
||||||
is_shared: bool,
|
is_shared: bool,
|
||||||
|
webhook_url: str = "",
|
||||||
|
webhook_secret: str = "",
|
||||||
ics_url: str | None = None,
|
ics_url: str | None = None,
|
||||||
ics_fetch_interval: int = 300,
|
ics_fetch_interval: int = 300,
|
||||||
ics_enabled: bool = False,
|
ics_enabled: bool = False,
|
||||||
|
platform: Platform | None = None,
|
||||||
):
|
):
|
||||||
"""
|
"""
|
||||||
Add a new room
|
Add a new room
|
||||||
"""
|
"""
|
||||||
|
if webhook_url and not webhook_secret:
|
||||||
|
webhook_secret = secrets.token_urlsafe(32)
|
||||||
|
|
||||||
room = Room(
|
room = Room(
|
||||||
name=name,
|
name=name,
|
||||||
user_id=user_id,
|
user_id=user_id,
|
||||||
@@ -138,9 +157,12 @@ class RoomController:
|
|||||||
recording_type=recording_type,
|
recording_type=recording_type,
|
||||||
recording_trigger=recording_trigger,
|
recording_trigger=recording_trigger,
|
||||||
is_shared=is_shared,
|
is_shared=is_shared,
|
||||||
|
webhook_url=webhook_url,
|
||||||
|
webhook_secret=webhook_secret,
|
||||||
ics_url=ics_url,
|
ics_url=ics_url,
|
||||||
ics_fetch_interval=ics_fetch_interval,
|
ics_fetch_interval=ics_fetch_interval,
|
||||||
ics_enabled=ics_enabled,
|
ics_enabled=ics_enabled,
|
||||||
|
platform=platform,
|
||||||
)
|
)
|
||||||
query = rooms.insert().values(**room.model_dump())
|
query = rooms.insert().values(**room.model_dump())
|
||||||
try:
|
try:
|
||||||
@@ -153,6 +175,9 @@ class RoomController:
|
|||||||
"""
|
"""
|
||||||
Update a room fields with key/values in values
|
Update a room fields with key/values in values
|
||||||
"""
|
"""
|
||||||
|
if values.get("webhook_url") and not values.get("webhook_secret"):
|
||||||
|
values["webhook_secret"] = secrets.token_urlsafe(32)
|
||||||
|
|
||||||
query = rooms.update().where(rooms.c.id == room.id).values(**values)
|
query = rooms.update().where(rooms.c.id == room.id).values(**values)
|
||||||
try:
|
try:
|
||||||
await get_database().execute(query)
|
await get_database().execute(query)
|
||||||
@@ -202,6 +227,13 @@ class RoomController:
|
|||||||
|
|
||||||
return room
|
return room
|
||||||
|
|
||||||
|
async def get_ics_enabled(self) -> list[Room]:
|
||||||
|
query = rooms.select().where(
|
||||||
|
rooms.c.ics_enabled == True, rooms.c.ics_url != None
|
||||||
|
)
|
||||||
|
results = await get_database().fetch_all(query)
|
||||||
|
return [Room(**result) for result in results]
|
||||||
|
|
||||||
async def remove_by_id(
|
async def remove_by_id(
|
||||||
self,
|
self,
|
||||||
room_id: str,
|
room_id: str,
|
||||||
|
|||||||
@@ -1,22 +1,38 @@
|
|||||||
"""Search functionality for transcripts and other entities."""
|
"""Search functionality for transcripts and other entities."""
|
||||||
|
|
||||||
|
import itertools
|
||||||
|
from dataclasses import dataclass
|
||||||
from datetime import datetime
|
from datetime import datetime
|
||||||
from io import StringIO
|
from io import StringIO
|
||||||
from typing import Annotated, Any, Dict
|
from typing import Annotated, Any, Dict, Iterator
|
||||||
|
|
||||||
import sqlalchemy
|
import sqlalchemy
|
||||||
import webvtt
|
import webvtt
|
||||||
from pydantic import BaseModel, Field, constr, field_serializer
|
from databases.interfaces import Record as DbRecord
|
||||||
|
from fastapi import HTTPException
|
||||||
|
from pydantic import (
|
||||||
|
BaseModel,
|
||||||
|
Field,
|
||||||
|
NonNegativeFloat,
|
||||||
|
NonNegativeInt,
|
||||||
|
TypeAdapter,
|
||||||
|
ValidationError,
|
||||||
|
constr,
|
||||||
|
field_serializer,
|
||||||
|
)
|
||||||
|
|
||||||
from reflector.db import get_database
|
from reflector.db import get_database
|
||||||
from reflector.db.transcripts import SourceKind, transcripts
|
from reflector.db.rooms import rooms
|
||||||
|
from reflector.db.transcripts import SourceKind, TranscriptStatus, transcripts
|
||||||
from reflector.db.utils import is_postgresql
|
from reflector.db.utils import is_postgresql
|
||||||
from reflector.logger import logger
|
from reflector.logger import logger
|
||||||
|
from reflector.utils.string import NonEmptyString, try_parse_non_empty_string
|
||||||
|
|
||||||
DEFAULT_SEARCH_LIMIT = 20
|
DEFAULT_SEARCH_LIMIT = 20
|
||||||
SNIPPET_CONTEXT_LENGTH = 50 # Characters before/after match to include
|
SNIPPET_CONTEXT_LENGTH = 50 # Characters before/after match to include
|
||||||
DEFAULT_SNIPPET_MAX_LENGTH = 150
|
DEFAULT_SNIPPET_MAX_LENGTH = NonNegativeInt(150)
|
||||||
DEFAULT_MAX_SNIPPETS = 3
|
DEFAULT_MAX_SNIPPETS = NonNegativeInt(3)
|
||||||
|
LONG_SUMMARY_MAX_SNIPPETS = 2
|
||||||
|
|
||||||
SearchQueryBase = constr(min_length=1, strip_whitespace=True)
|
SearchQueryBase = constr(min_length=1, strip_whitespace=True)
|
||||||
SearchLimitBase = Annotated[int, Field(ge=1, le=100)]
|
SearchLimitBase = Annotated[int, Field(ge=1, le=100)]
|
||||||
@@ -24,6 +40,7 @@ SearchOffsetBase = Annotated[int, Field(ge=0)]
|
|||||||
SearchTotalBase = Annotated[int, Field(ge=0)]
|
SearchTotalBase = Annotated[int, Field(ge=0)]
|
||||||
|
|
||||||
SearchQuery = Annotated[SearchQueryBase, Field(description="Search query text")]
|
SearchQuery = Annotated[SearchQueryBase, Field(description="Search query text")]
|
||||||
|
search_query_adapter = TypeAdapter(SearchQuery)
|
||||||
SearchLimit = Annotated[SearchLimitBase, Field(description="Results per page")]
|
SearchLimit = Annotated[SearchLimitBase, Field(description="Results per page")]
|
||||||
SearchOffset = Annotated[
|
SearchOffset = Annotated[
|
||||||
SearchOffsetBase, Field(description="Number of results to skip")
|
SearchOffsetBase, Field(description="Number of results to skip")
|
||||||
@@ -32,15 +49,94 @@ SearchTotal = Annotated[
|
|||||||
SearchTotalBase, Field(description="Total number of search results")
|
SearchTotalBase, Field(description="Total number of search results")
|
||||||
]
|
]
|
||||||
|
|
||||||
|
WEBVTT_SPEC_HEADER = "WEBVTT"
|
||||||
|
|
||||||
|
WebVTTContent = Annotated[
|
||||||
|
str,
|
||||||
|
Field(min_length=len(WEBVTT_SPEC_HEADER), description="WebVTT content"),
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
|
class WebVTTProcessor:
|
||||||
|
"""Stateless processor for WebVTT content operations."""
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def parse(raw_content: str) -> WebVTTContent:
|
||||||
|
"""Parse WebVTT content and return it as a string."""
|
||||||
|
if not raw_content.startswith(WEBVTT_SPEC_HEADER):
|
||||||
|
raise ValueError(f"Invalid WebVTT content, no header {WEBVTT_SPEC_HEADER}")
|
||||||
|
return raw_content
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def extract_text(webvtt_content: WebVTTContent) -> str:
|
||||||
|
"""Extract plain text from WebVTT content using webvtt library."""
|
||||||
|
try:
|
||||||
|
buffer = StringIO(webvtt_content)
|
||||||
|
vtt = webvtt.read_buffer(buffer)
|
||||||
|
return " ".join(caption.text for caption in vtt if caption.text)
|
||||||
|
except webvtt.errors.MalformedFileError as e:
|
||||||
|
logger.warning(f"Malformed WebVTT content: {e}")
|
||||||
|
return ""
|
||||||
|
except (UnicodeDecodeError, ValueError) as e:
|
||||||
|
logger.warning(f"Failed to decode WebVTT content: {e}")
|
||||||
|
return ""
|
||||||
|
except AttributeError as e:
|
||||||
|
logger.error(
|
||||||
|
f"WebVTT parsing error - unexpected format: {e}", exc_info=True
|
||||||
|
)
|
||||||
|
return ""
|
||||||
|
except Exception as e:
|
||||||
|
logger.error(f"Unexpected error parsing WebVTT: {e}", exc_info=True)
|
||||||
|
return ""
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def generate_snippets(
|
||||||
|
webvtt_content: WebVTTContent,
|
||||||
|
query: SearchQuery,
|
||||||
|
max_snippets: NonNegativeInt = DEFAULT_MAX_SNIPPETS,
|
||||||
|
) -> list[str]:
|
||||||
|
"""Generate snippets from WebVTT content."""
|
||||||
|
return SnippetGenerator.generate(
|
||||||
|
WebVTTProcessor.extract_text(webvtt_content),
|
||||||
|
query,
|
||||||
|
max_snippets=max_snippets,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@dataclass(frozen=True)
|
||||||
|
class SnippetCandidate:
|
||||||
|
"""Represents a candidate snippet with its position."""
|
||||||
|
|
||||||
|
_text: str
|
||||||
|
start: NonNegativeInt
|
||||||
|
_original_text_length: int
|
||||||
|
|
||||||
|
@property
|
||||||
|
def end(self) -> NonNegativeInt:
|
||||||
|
"""Calculate end position from start and raw text length."""
|
||||||
|
return self.start + len(self._text)
|
||||||
|
|
||||||
|
def text(self) -> str:
|
||||||
|
"""Get display text with ellipses added if needed."""
|
||||||
|
result = self._text.strip()
|
||||||
|
if self.start > 0:
|
||||||
|
result = "..." + result
|
||||||
|
if self.end < self._original_text_length:
|
||||||
|
result = result + "..."
|
||||||
|
return result
|
||||||
|
|
||||||
|
|
||||||
class SearchParameters(BaseModel):
|
class SearchParameters(BaseModel):
|
||||||
"""Validated search parameters for full-text search."""
|
"""Validated search parameters for full-text search."""
|
||||||
|
|
||||||
query_text: SearchQuery
|
query_text: SearchQuery | None = None
|
||||||
limit: SearchLimit = DEFAULT_SEARCH_LIMIT
|
limit: SearchLimit = DEFAULT_SEARCH_LIMIT
|
||||||
offset: SearchOffset = 0
|
offset: SearchOffset = 0
|
||||||
user_id: str | None = None
|
user_id: str | None = None
|
||||||
room_id: str | None = None
|
room_id: str | None = None
|
||||||
|
source_kind: SourceKind | None = None
|
||||||
|
from_datetime: datetime | None = None
|
||||||
|
to_datetime: datetime | None = None
|
||||||
|
|
||||||
|
|
||||||
class SearchResultDB(BaseModel):
|
class SearchResultDB(BaseModel):
|
||||||
@@ -64,13 +160,18 @@ class SearchResult(BaseModel):
|
|||||||
title: str | None = None
|
title: str | None = None
|
||||||
user_id: str | None = None
|
user_id: str | None = None
|
||||||
room_id: str | None = None
|
room_id: str | None = None
|
||||||
|
room_name: str | None = None
|
||||||
|
source_kind: SourceKind
|
||||||
created_at: datetime
|
created_at: datetime
|
||||||
status: str = Field(..., min_length=1)
|
status: TranscriptStatus = Field(..., min_length=1)
|
||||||
rank: float = Field(..., ge=0, le=1)
|
rank: float = Field(..., ge=0, le=1)
|
||||||
duration: float | None = Field(..., ge=0, description="Duration in seconds")
|
duration: NonNegativeFloat | None = Field(..., description="Duration in seconds")
|
||||||
search_snippets: list[str] = Field(
|
search_snippets: list[str] = Field(
|
||||||
description="Text snippets around search matches"
|
description="Text snippets around search matches"
|
||||||
)
|
)
|
||||||
|
total_match_count: NonNegativeInt = Field(
|
||||||
|
default=0, description="Total number of matches found in the transcript"
|
||||||
|
)
|
||||||
|
|
||||||
@field_serializer("created_at", when_used="json")
|
@field_serializer("created_at", when_used="json")
|
||||||
def serialize_datetime(self, dt: datetime) -> str:
|
def serialize_datetime(self, dt: datetime) -> str:
|
||||||
@@ -79,84 +180,157 @@ class SearchResult(BaseModel):
|
|||||||
return dt.isoformat()
|
return dt.isoformat()
|
||||||
|
|
||||||
|
|
||||||
class SearchController:
|
class SnippetGenerator:
|
||||||
"""Controller for search operations across different entities."""
|
"""Stateless generator for text snippets and match operations."""
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def _extract_webvtt_text(webvtt_content: str) -> str:
|
def find_all_matches(text: str, query: str) -> Iterator[int]:
|
||||||
"""Extract plain text from WebVTT content using webvtt library."""
|
"""Generate all match positions for a query in text."""
|
||||||
if not webvtt_content:
|
if not text:
|
||||||
return ""
|
logger.warning("Empty text for search query in find_all_matches")
|
||||||
|
return
|
||||||
|
if not query:
|
||||||
|
logger.warning("Empty query for search text in find_all_matches")
|
||||||
|
return
|
||||||
|
|
||||||
try:
|
text_lower = text.lower()
|
||||||
buffer = StringIO(webvtt_content)
|
query_lower = query.lower()
|
||||||
vtt = webvtt.read_buffer(buffer)
|
start = 0
|
||||||
return " ".join(caption.text for caption in vtt if caption.text)
|
prev_start = start
|
||||||
except (webvtt.errors.MalformedFileError, UnicodeDecodeError, ValueError) as e:
|
while (pos := text_lower.find(query_lower, start)) != -1:
|
||||||
logger.warning(f"Failed to parse WebVTT content: {e}", exc_info=e)
|
yield pos
|
||||||
return ""
|
start = pos + len(query_lower)
|
||||||
except AttributeError as e:
|
if start <= prev_start:
|
||||||
logger.warning(f"WebVTT parsing error - unexpected format: {e}", exc_info=e)
|
raise ValueError("panic! find_all_matches is not incremental")
|
||||||
return ""
|
prev_start = start
|
||||||
|
|
||||||
@staticmethod
|
@staticmethod
|
||||||
def _generate_snippets(
|
def count_matches(text: str, query: SearchQuery) -> NonNegativeInt:
|
||||||
text: str,
|
"""Count total number of matches for a query in text."""
|
||||||
q: SearchQuery,
|
ZERO = NonNegativeInt(0)
|
||||||
max_length: int = DEFAULT_SNIPPET_MAX_LENGTH,
|
if not text:
|
||||||
max_snippets: int = DEFAULT_MAX_SNIPPETS,
|
logger.warning("Empty text for search query in count_matches")
|
||||||
) -> list[str]:
|
return ZERO
|
||||||
"""Generate multiple snippets around all occurrences of search term."""
|
assert query is not None
|
||||||
if not text or not q:
|
return NonNegativeInt(
|
||||||
return []
|
sum(1 for _ in SnippetGenerator.find_all_matches(text, query))
|
||||||
|
|
||||||
snippets = []
|
|
||||||
lower_text = text.lower()
|
|
||||||
search_lower = q.lower()
|
|
||||||
|
|
||||||
last_snippet_end = 0
|
|
||||||
start_pos = 0
|
|
||||||
|
|
||||||
while len(snippets) < max_snippets:
|
|
||||||
match_pos = lower_text.find(search_lower, start_pos)
|
|
||||||
|
|
||||||
if match_pos == -1:
|
|
||||||
if not snippets and search_lower.split():
|
|
||||||
first_word = search_lower.split()[0]
|
|
||||||
match_pos = lower_text.find(first_word, start_pos)
|
|
||||||
if match_pos == -1:
|
|
||||||
break
|
|
||||||
else:
|
|
||||||
break
|
|
||||||
|
|
||||||
snippet_start = max(0, match_pos - SNIPPET_CONTEXT_LENGTH)
|
|
||||||
snippet_end = min(
|
|
||||||
len(text), match_pos + max_length - SNIPPET_CONTEXT_LENGTH
|
|
||||||
)
|
)
|
||||||
|
|
||||||
if snippet_start < last_snippet_end:
|
@staticmethod
|
||||||
start_pos = match_pos + len(search_lower)
|
def create_snippet(
|
||||||
continue
|
text: str, match_pos: int, max_length: int = DEFAULT_SNIPPET_MAX_LENGTH
|
||||||
|
) -> SnippetCandidate:
|
||||||
|
"""Create a snippet from a match position."""
|
||||||
|
snippet_start = NonNegativeInt(max(0, match_pos - SNIPPET_CONTEXT_LENGTH))
|
||||||
|
snippet_end = min(len(text), match_pos + max_length - SNIPPET_CONTEXT_LENGTH)
|
||||||
|
|
||||||
snippet = text[snippet_start:snippet_end]
|
snippet_text = text[snippet_start:snippet_end]
|
||||||
|
|
||||||
if snippet_start > 0:
|
return SnippetCandidate(
|
||||||
snippet = "..." + snippet
|
_text=snippet_text, start=snippet_start, _original_text_length=len(text)
|
||||||
if snippet_end < len(text):
|
)
|
||||||
snippet = snippet + "..."
|
|
||||||
|
|
||||||
snippet = snippet.strip()
|
@staticmethod
|
||||||
|
def filter_non_overlapping(
|
||||||
|
candidates: Iterator[SnippetCandidate],
|
||||||
|
) -> Iterator[str]:
|
||||||
|
"""Filter out overlapping snippets and return only display text."""
|
||||||
|
last_end = 0
|
||||||
|
for candidate in candidates:
|
||||||
|
display_text = candidate.text()
|
||||||
|
# it means that next overlapping snippets simply don't get included
|
||||||
|
# it's fine as simplistic logic and users probably won't care much because they already have their search results just fin
|
||||||
|
if candidate.start >= last_end and display_text:
|
||||||
|
yield display_text
|
||||||
|
last_end = candidate.end
|
||||||
|
|
||||||
if snippet:
|
@staticmethod
|
||||||
snippets.append(snippet)
|
def generate(
|
||||||
last_snippet_end = snippet_end
|
text: str,
|
||||||
|
query: SearchQuery,
|
||||||
|
max_length: NonNegativeInt = DEFAULT_SNIPPET_MAX_LENGTH,
|
||||||
|
max_snippets: NonNegativeInt = DEFAULT_MAX_SNIPPETS,
|
||||||
|
) -> list[str]:
|
||||||
|
"""Generate snippets from text."""
|
||||||
|
assert query is not None
|
||||||
|
if not text:
|
||||||
|
logger.warning("Empty text for generate_snippets")
|
||||||
|
return []
|
||||||
|
|
||||||
start_pos = match_pos + len(search_lower)
|
candidates = (
|
||||||
if start_pos >= len(text):
|
SnippetGenerator.create_snippet(text, pos, max_length)
|
||||||
break
|
for pos in SnippetGenerator.find_all_matches(text, query)
|
||||||
|
)
|
||||||
|
filtered = SnippetGenerator.filter_non_overlapping(candidates)
|
||||||
|
snippets = list(itertools.islice(filtered, max_snippets))
|
||||||
|
|
||||||
|
# Fallback to first word search if no full matches
|
||||||
|
# it's another assumption: proper snippet logic generation is quite complicated and tied to db logic, so simplification is used here
|
||||||
|
if not snippets and " " in query:
|
||||||
|
first_word = query.split()[0]
|
||||||
|
return SnippetGenerator.generate(text, first_word, max_length, max_snippets)
|
||||||
|
|
||||||
return snippets
|
return snippets
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def from_summary(
|
||||||
|
summary: str,
|
||||||
|
query: SearchQuery,
|
||||||
|
max_snippets: NonNegativeInt = LONG_SUMMARY_MAX_SNIPPETS,
|
||||||
|
) -> list[str]:
|
||||||
|
"""Generate snippets from summary text."""
|
||||||
|
return SnippetGenerator.generate(summary, query, max_snippets=max_snippets)
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def combine_sources(
|
||||||
|
summary: NonEmptyString | None,
|
||||||
|
webvtt: WebVTTContent | None,
|
||||||
|
query: SearchQuery,
|
||||||
|
max_total: NonNegativeInt = DEFAULT_MAX_SNIPPETS,
|
||||||
|
) -> tuple[list[str], NonNegativeInt]:
|
||||||
|
"""Combine snippets from multiple sources and return total match count.
|
||||||
|
|
||||||
|
Returns (snippets, total_match_count) tuple.
|
||||||
|
|
||||||
|
snippets can be empty for real in case of e.g. title match
|
||||||
|
"""
|
||||||
|
|
||||||
|
assert (
|
||||||
|
summary is not None or webvtt is not None
|
||||||
|
), "At least one source must be present"
|
||||||
|
|
||||||
|
webvtt_matches = 0
|
||||||
|
summary_matches = 0
|
||||||
|
|
||||||
|
if webvtt:
|
||||||
|
webvtt_text = WebVTTProcessor.extract_text(webvtt)
|
||||||
|
webvtt_matches = SnippetGenerator.count_matches(webvtt_text, query)
|
||||||
|
|
||||||
|
if summary:
|
||||||
|
summary_matches = SnippetGenerator.count_matches(summary, query)
|
||||||
|
|
||||||
|
total_matches = NonNegativeInt(webvtt_matches + summary_matches)
|
||||||
|
|
||||||
|
summary_snippets = (
|
||||||
|
SnippetGenerator.from_summary(summary, query) if summary else []
|
||||||
|
)
|
||||||
|
|
||||||
|
if len(summary_snippets) >= max_total:
|
||||||
|
return summary_snippets[:max_total], total_matches
|
||||||
|
|
||||||
|
remaining = max_total - len(summary_snippets)
|
||||||
|
webvtt_snippets = (
|
||||||
|
WebVTTProcessor.generate_snippets(webvtt, query, remaining)
|
||||||
|
if webvtt
|
||||||
|
else []
|
||||||
|
)
|
||||||
|
|
||||||
|
return summary_snippets + webvtt_snippets, total_matches
|
||||||
|
|
||||||
|
|
||||||
|
class SearchController:
|
||||||
|
"""Controller for search operations across different entities."""
|
||||||
|
|
||||||
@classmethod
|
@classmethod
|
||||||
async def search_transcripts(
|
async def search_transcripts(
|
||||||
cls, params: SearchParameters
|
cls, params: SearchParameters
|
||||||
@@ -172,12 +346,7 @@ class SearchController:
|
|||||||
)
|
)
|
||||||
return [], 0
|
return [], 0
|
||||||
|
|
||||||
search_query = sqlalchemy.func.websearch_to_tsquery(
|
base_columns = [
|
||||||
"english", params.query_text
|
|
||||||
)
|
|
||||||
|
|
||||||
base_query = sqlalchemy.select(
|
|
||||||
[
|
|
||||||
transcripts.c.id,
|
transcripts.c.id,
|
||||||
transcripts.c.title,
|
transcripts.c.title,
|
||||||
transcripts.c.created_at,
|
transcripts.c.created_at,
|
||||||
@@ -187,24 +356,70 @@ class SearchController:
|
|||||||
transcripts.c.room_id,
|
transcripts.c.room_id,
|
||||||
transcripts.c.source_kind,
|
transcripts.c.source_kind,
|
||||||
transcripts.c.webvtt,
|
transcripts.c.webvtt,
|
||||||
sqlalchemy.func.ts_rank(
|
transcripts.c.long_summary,
|
||||||
|
sqlalchemy.case(
|
||||||
|
(
|
||||||
|
transcripts.c.room_id.isnot(None) & rooms.c.id.is_(None),
|
||||||
|
"Deleted Room",
|
||||||
|
),
|
||||||
|
else_=rooms.c.name,
|
||||||
|
).label("room_name"),
|
||||||
|
]
|
||||||
|
search_query = None
|
||||||
|
if params.query_text is not None:
|
||||||
|
search_query = sqlalchemy.func.websearch_to_tsquery(
|
||||||
|
"english", params.query_text
|
||||||
|
)
|
||||||
|
rank_column = sqlalchemy.func.ts_rank(
|
||||||
transcripts.c.search_vector_en,
|
transcripts.c.search_vector_en,
|
||||||
search_query,
|
search_query,
|
||||||
32, # normalization flag: rank/(rank+1) for 0-1 range
|
32, # normalization flag: rank/(rank+1) for 0-1 range
|
||||||
).label("rank"),
|
).label("rank")
|
||||||
]
|
else:
|
||||||
).where(transcripts.c.search_vector_en.op("@@")(search_query))
|
rank_column = sqlalchemy.cast(1.0, sqlalchemy.Float).label("rank")
|
||||||
|
|
||||||
|
columns = base_columns + [rank_column]
|
||||||
|
base_query = sqlalchemy.select(columns).select_from(
|
||||||
|
transcripts.join(rooms, transcripts.c.room_id == rooms.c.id, isouter=True)
|
||||||
|
)
|
||||||
|
|
||||||
|
if params.query_text is not None:
|
||||||
|
# because already initialized based on params.query_text presence above
|
||||||
|
assert search_query is not None
|
||||||
|
base_query = base_query.where(
|
||||||
|
transcripts.c.search_vector_en.op("@@")(search_query)
|
||||||
|
)
|
||||||
|
|
||||||
if params.user_id:
|
if params.user_id:
|
||||||
base_query = base_query.where(transcripts.c.user_id == params.user_id)
|
base_query = base_query.where(
|
||||||
|
sqlalchemy.or_(
|
||||||
|
transcripts.c.user_id == params.user_id, rooms.c.is_shared
|
||||||
|
)
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
base_query = base_query.where(rooms.c.is_shared)
|
||||||
if params.room_id:
|
if params.room_id:
|
||||||
base_query = base_query.where(transcripts.c.room_id == params.room_id)
|
base_query = base_query.where(transcripts.c.room_id == params.room_id)
|
||||||
|
if params.source_kind:
|
||||||
query = (
|
base_query = base_query.where(
|
||||||
base_query.order_by(sqlalchemy.desc(sqlalchemy.text("rank")))
|
transcripts.c.source_kind == params.source_kind
|
||||||
.limit(params.limit)
|
|
||||||
.offset(params.offset)
|
|
||||||
)
|
)
|
||||||
|
if params.from_datetime:
|
||||||
|
base_query = base_query.where(
|
||||||
|
transcripts.c.created_at >= params.from_datetime
|
||||||
|
)
|
||||||
|
if params.to_datetime:
|
||||||
|
base_query = base_query.where(
|
||||||
|
transcripts.c.created_at <= params.to_datetime
|
||||||
|
)
|
||||||
|
|
||||||
|
if params.query_text is not None:
|
||||||
|
order_by = sqlalchemy.desc(sqlalchemy.text("rank"))
|
||||||
|
else:
|
||||||
|
order_by = sqlalchemy.desc(transcripts.c.created_at)
|
||||||
|
|
||||||
|
query = base_query.order_by(order_by).limit(params.limit).offset(params.offset)
|
||||||
|
|
||||||
rs = await get_database().fetch_all(query)
|
rs = await get_database().fetch_all(query)
|
||||||
|
|
||||||
count_query = sqlalchemy.select([sqlalchemy.func.count()]).select_from(
|
count_query = sqlalchemy.select([sqlalchemy.func.count()]).select_from(
|
||||||
@@ -212,20 +427,52 @@ class SearchController:
|
|||||||
)
|
)
|
||||||
total = await get_database().fetch_val(count_query)
|
total = await get_database().fetch_val(count_query)
|
||||||
|
|
||||||
def _process_result(r) -> SearchResult:
|
def _process_result(r: DbRecord) -> SearchResult:
|
||||||
r_dict: Dict[str, Any] = dict(r)
|
r_dict: Dict[str, Any] = dict(r)
|
||||||
webvtt: str | None = r_dict.pop("webvtt", None)
|
|
||||||
|
webvtt_raw: str | None = r_dict.pop("webvtt", None)
|
||||||
|
webvtt: WebVTTContent | None
|
||||||
|
if webvtt_raw:
|
||||||
|
webvtt = WebVTTProcessor.parse(webvtt_raw)
|
||||||
|
else:
|
||||||
|
webvtt = None
|
||||||
|
|
||||||
|
long_summary_r: str | None = r_dict.pop("long_summary", None)
|
||||||
|
long_summary: NonEmptyString = try_parse_non_empty_string(long_summary_r)
|
||||||
|
room_name: str | None = r_dict.pop("room_name", None)
|
||||||
db_result = SearchResultDB.model_validate(r_dict)
|
db_result = SearchResultDB.model_validate(r_dict)
|
||||||
|
|
||||||
snippets = []
|
at_least_one_source = webvtt is not None or long_summary is not None
|
||||||
if webvtt:
|
has_query = params.query_text is not None
|
||||||
plain_text = cls._extract_webvtt_text(webvtt)
|
snippets, total_match_count = (
|
||||||
snippets = cls._generate_snippets(plain_text, params.query_text)
|
SnippetGenerator.combine_sources(
|
||||||
|
long_summary, webvtt, params.query_text, DEFAULT_MAX_SNIPPETS
|
||||||
|
)
|
||||||
|
if has_query and at_least_one_source
|
||||||
|
else ([], 0)
|
||||||
|
)
|
||||||
|
|
||||||
return SearchResult(**db_result.model_dump(), search_snippets=snippets)
|
return SearchResult(
|
||||||
|
**db_result.model_dump(),
|
||||||
|
room_name=room_name,
|
||||||
|
search_snippets=snippets,
|
||||||
|
total_match_count=total_match_count,
|
||||||
|
)
|
||||||
|
|
||||||
|
try:
|
||||||
results = [_process_result(r) for r in rs]
|
results = [_process_result(r) for r in rs]
|
||||||
|
except ValidationError as e:
|
||||||
|
logger.error(f"Invalid search result data: {e}", exc_info=True)
|
||||||
|
raise HTTPException(
|
||||||
|
status_code=500, detail="Internal search result data consistency error"
|
||||||
|
)
|
||||||
|
except Exception as e:
|
||||||
|
logger.error(f"Error processing search results: {e}", exc_info=True)
|
||||||
|
raise
|
||||||
|
|
||||||
return results, total
|
return results, total
|
||||||
|
|
||||||
|
|
||||||
search_controller = SearchController()
|
search_controller = SearchController()
|
||||||
|
webvtt_processor = WebVTTProcessor()
|
||||||
|
snippet_generator = SnippetGenerator()
|
||||||
|
|||||||
@@ -21,7 +21,7 @@ from reflector.db.utils import is_postgresql
|
|||||||
from reflector.logger import logger
|
from reflector.logger import logger
|
||||||
from reflector.processors.types import Word as ProcessorWord
|
from reflector.processors.types import Word as ProcessorWord
|
||||||
from reflector.settings import settings
|
from reflector.settings import settings
|
||||||
from reflector.storage import get_recordings_storage, get_transcripts_storage
|
from reflector.storage import get_transcripts_storage
|
||||||
from reflector.utils import generate_uuid4
|
from reflector.utils import generate_uuid4
|
||||||
from reflector.utils.webvtt import topics_to_webvtt
|
from reflector.utils.webvtt import topics_to_webvtt
|
||||||
|
|
||||||
@@ -88,6 +88,8 @@ transcripts = sqlalchemy.Table(
|
|||||||
sqlalchemy.Index("idx_transcript_created_at", "created_at"),
|
sqlalchemy.Index("idx_transcript_created_at", "created_at"),
|
||||||
sqlalchemy.Index("idx_transcript_user_id_recording_id", "user_id", "recording_id"),
|
sqlalchemy.Index("idx_transcript_user_id_recording_id", "user_id", "recording_id"),
|
||||||
sqlalchemy.Index("idx_transcript_room_id", "room_id"),
|
sqlalchemy.Index("idx_transcript_room_id", "room_id"),
|
||||||
|
sqlalchemy.Index("idx_transcript_source_kind", "source_kind"),
|
||||||
|
sqlalchemy.Index("idx_transcript_room_id_created_at", "room_id", "created_at"),
|
||||||
)
|
)
|
||||||
|
|
||||||
# Add PostgreSQL-specific full-text search column
|
# Add PostgreSQL-specific full-text search column
|
||||||
@@ -99,7 +101,8 @@ if is_postgresql():
|
|||||||
TSVECTOR,
|
TSVECTOR,
|
||||||
sqlalchemy.Computed(
|
sqlalchemy.Computed(
|
||||||
"setweight(to_tsvector('english', coalesce(title, '')), 'A') || "
|
"setweight(to_tsvector('english', coalesce(title, '')), 'A') || "
|
||||||
"setweight(to_tsvector('english', coalesce(webvtt, '')), 'B')",
|
"setweight(to_tsvector('english', coalesce(long_summary, '')), 'B') || "
|
||||||
|
"setweight(to_tsvector('english', coalesce(webvtt, '')), 'C')",
|
||||||
persisted=True,
|
persisted=True,
|
||||||
),
|
),
|
||||||
)
|
)
|
||||||
@@ -119,6 +122,15 @@ def generate_transcript_name() -> str:
|
|||||||
return f"Transcript {now.strftime('%Y-%m-%d %H:%M:%S')}"
|
return f"Transcript {now.strftime('%Y-%m-%d %H:%M:%S')}"
|
||||||
|
|
||||||
|
|
||||||
|
TranscriptStatus = Literal[
|
||||||
|
"idle", "uploaded", "recording", "processing", "error", "ended"
|
||||||
|
]
|
||||||
|
|
||||||
|
|
||||||
|
class StrValue(BaseModel):
|
||||||
|
value: str
|
||||||
|
|
||||||
|
|
||||||
class AudioWaveform(BaseModel):
|
class AudioWaveform(BaseModel):
|
||||||
data: list[float]
|
data: list[float]
|
||||||
|
|
||||||
@@ -174,6 +186,7 @@ class TranscriptParticipant(BaseModel):
|
|||||||
id: str = Field(default_factory=generate_uuid4)
|
id: str = Field(default_factory=generate_uuid4)
|
||||||
speaker: int | None
|
speaker: int | None
|
||||||
name: str
|
name: str
|
||||||
|
user_id: str | None = None
|
||||||
|
|
||||||
|
|
||||||
class Transcript(BaseModel):
|
class Transcript(BaseModel):
|
||||||
@@ -182,7 +195,7 @@ class Transcript(BaseModel):
|
|||||||
id: str = Field(default_factory=generate_uuid4)
|
id: str = Field(default_factory=generate_uuid4)
|
||||||
user_id: str | None = None
|
user_id: str | None = None
|
||||||
name: str = Field(default_factory=generate_transcript_name)
|
name: str = Field(default_factory=generate_transcript_name)
|
||||||
status: str = "idle"
|
status: TranscriptStatus = "idle"
|
||||||
duration: float = 0
|
duration: float = 0
|
||||||
created_at: datetime = Field(default_factory=lambda: datetime.now(timezone.utc))
|
created_at: datetime = Field(default_factory=lambda: datetime.now(timezone.utc))
|
||||||
title: str | None = None
|
title: str | None = None
|
||||||
@@ -611,7 +624,9 @@ class TranscriptController:
|
|||||||
)
|
)
|
||||||
if recording:
|
if recording:
|
||||||
try:
|
try:
|
||||||
await get_recordings_storage().delete_file(recording.object_key)
|
await get_transcripts_storage().delete_file(
|
||||||
|
recording.object_key, bucket=recording.bucket_name
|
||||||
|
)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.warning(
|
logger.warning(
|
||||||
"Failed to delete recording object from S3",
|
"Failed to delete recording object from S3",
|
||||||
@@ -635,6 +650,19 @@ class TranscriptController:
|
|||||||
query = transcripts.delete().where(transcripts.c.recording_id == recording_id)
|
query = transcripts.delete().where(transcripts.c.recording_id == recording_id)
|
||||||
await get_database().execute(query)
|
await get_database().execute(query)
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def user_can_mutate(transcript: Transcript, user_id: str | None) -> bool:
|
||||||
|
"""
|
||||||
|
Returns True if the given user is allowed to modify the transcript.
|
||||||
|
|
||||||
|
Policy:
|
||||||
|
- Anonymous transcripts (user_id is None) cannot be modified via API
|
||||||
|
- Only the owner (matching user_id) can modify their transcript
|
||||||
|
"""
|
||||||
|
if transcript.user_id is None:
|
||||||
|
return False
|
||||||
|
return user_id and transcript.user_id == user_id
|
||||||
|
|
||||||
@asynccontextmanager
|
@asynccontextmanager
|
||||||
async def transaction(self):
|
async def transaction(self):
|
||||||
"""
|
"""
|
||||||
@@ -700,11 +728,13 @@ class TranscriptController:
|
|||||||
"""
|
"""
|
||||||
Download audio from storage
|
Download audio from storage
|
||||||
"""
|
"""
|
||||||
transcript.audio_mp3_filename.write_bytes(
|
storage = get_transcripts_storage()
|
||||||
await get_transcripts_storage().get_file(
|
try:
|
||||||
transcript.storage_audio_path,
|
with open(transcript.audio_mp3_filename, "wb") as f:
|
||||||
)
|
await storage.stream_to_fileobj(transcript.storage_audio_path, f)
|
||||||
)
|
except Exception:
|
||||||
|
transcript.audio_mp3_filename.unlink(missing_ok=True)
|
||||||
|
raise
|
||||||
|
|
||||||
async def upsert_participant(
|
async def upsert_participant(
|
||||||
self,
|
self,
|
||||||
@@ -729,5 +759,27 @@ class TranscriptController:
|
|||||||
transcript.delete_participant(participant_id)
|
transcript.delete_participant(participant_id)
|
||||||
await self.update(transcript, {"participants": transcript.participants_dump()})
|
await self.update(transcript, {"participants": transcript.participants_dump()})
|
||||||
|
|
||||||
|
async def set_status(
|
||||||
|
self, transcript_id: str, status: TranscriptStatus
|
||||||
|
) -> TranscriptEvent | None:
|
||||||
|
"""
|
||||||
|
Update the status of a transcript
|
||||||
|
|
||||||
|
Will add an event STATUS + update the status field of transcript
|
||||||
|
"""
|
||||||
|
async with self.transaction():
|
||||||
|
transcript = await self.get_by_id(transcript_id)
|
||||||
|
if not transcript:
|
||||||
|
raise Exception(f"Transcript {transcript_id} not found")
|
||||||
|
if transcript.status == status:
|
||||||
|
return
|
||||||
|
resp = await self.append_event(
|
||||||
|
transcript=transcript,
|
||||||
|
event="STATUS",
|
||||||
|
data=StrValue(value=status),
|
||||||
|
)
|
||||||
|
await self.update(transcript, {"status": status})
|
||||||
|
return resp
|
||||||
|
|
||||||
|
|
||||||
transcripts_controller = TranscriptController()
|
transcripts_controller = TranscriptController()
|
||||||
|
|||||||
91
server/reflector/db/user_api_keys.py
Normal file
91
server/reflector/db/user_api_keys.py
Normal file
@@ -0,0 +1,91 @@
|
|||||||
|
import hmac
|
||||||
|
import secrets
|
||||||
|
from datetime import datetime, timezone
|
||||||
|
from hashlib import sha256
|
||||||
|
|
||||||
|
import sqlalchemy
|
||||||
|
from pydantic import BaseModel, Field
|
||||||
|
|
||||||
|
from reflector.db import get_database, metadata
|
||||||
|
from reflector.settings import settings
|
||||||
|
from reflector.utils import generate_uuid4
|
||||||
|
from reflector.utils.string import NonEmptyString
|
||||||
|
|
||||||
|
user_api_keys = sqlalchemy.Table(
|
||||||
|
"user_api_key",
|
||||||
|
metadata,
|
||||||
|
sqlalchemy.Column("id", sqlalchemy.String, primary_key=True),
|
||||||
|
sqlalchemy.Column("user_id", sqlalchemy.String, nullable=False),
|
||||||
|
sqlalchemy.Column("key_hash", sqlalchemy.String, nullable=False),
|
||||||
|
sqlalchemy.Column("name", sqlalchemy.String, nullable=True),
|
||||||
|
sqlalchemy.Column("created_at", sqlalchemy.DateTime(timezone=True), nullable=False),
|
||||||
|
sqlalchemy.Index("idx_user_api_key_hash", "key_hash", unique=True),
|
||||||
|
sqlalchemy.Index("idx_user_api_key_user_id", "user_id"),
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
class UserApiKey(BaseModel):
|
||||||
|
id: NonEmptyString = Field(default_factory=generate_uuid4)
|
||||||
|
user_id: NonEmptyString
|
||||||
|
key_hash: NonEmptyString
|
||||||
|
name: NonEmptyString | None = None
|
||||||
|
created_at: datetime = Field(default_factory=lambda: datetime.now(timezone.utc))
|
||||||
|
|
||||||
|
|
||||||
|
class UserApiKeyController:
|
||||||
|
@staticmethod
|
||||||
|
def generate_key() -> NonEmptyString:
|
||||||
|
return secrets.token_urlsafe(48)
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
def hash_key(key: NonEmptyString) -> str:
|
||||||
|
return hmac.new(
|
||||||
|
settings.SECRET_KEY.encode(), key.encode(), digestmod=sha256
|
||||||
|
).hexdigest()
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
async def create_key(
|
||||||
|
cls,
|
||||||
|
user_id: NonEmptyString,
|
||||||
|
name: NonEmptyString | None = None,
|
||||||
|
) -> tuple[UserApiKey, NonEmptyString]:
|
||||||
|
plaintext = cls.generate_key()
|
||||||
|
api_key = UserApiKey(
|
||||||
|
user_id=user_id,
|
||||||
|
key_hash=cls.hash_key(plaintext),
|
||||||
|
name=name,
|
||||||
|
)
|
||||||
|
query = user_api_keys.insert().values(**api_key.model_dump())
|
||||||
|
await get_database().execute(query)
|
||||||
|
return api_key, plaintext
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
async def verify_key(cls, plaintext_key: NonEmptyString) -> UserApiKey | None:
|
||||||
|
key_hash = cls.hash_key(plaintext_key)
|
||||||
|
query = user_api_keys.select().where(
|
||||||
|
user_api_keys.c.key_hash == key_hash,
|
||||||
|
)
|
||||||
|
result = await get_database().fetch_one(query)
|
||||||
|
return UserApiKey(**result) if result else None
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
async def list_by_user_id(user_id: NonEmptyString) -> list[UserApiKey]:
|
||||||
|
query = (
|
||||||
|
user_api_keys.select()
|
||||||
|
.where(user_api_keys.c.user_id == user_id)
|
||||||
|
.order_by(user_api_keys.c.created_at.desc())
|
||||||
|
)
|
||||||
|
results = await get_database().fetch_all(query)
|
||||||
|
return [UserApiKey(**r) for r in results]
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
|
async def delete_key(key_id: NonEmptyString, user_id: NonEmptyString) -> bool:
|
||||||
|
query = user_api_keys.delete().where(
|
||||||
|
(user_api_keys.c.id == key_id) & (user_api_keys.c.user_id == user_id)
|
||||||
|
)
|
||||||
|
result = await get_database().execute(query)
|
||||||
|
# asyncpg returns None for DELETE, consider it success if no exception
|
||||||
|
return result is None or result > 0
|
||||||
|
|
||||||
|
|
||||||
|
user_api_keys_controller = UserApiKeyController()
|
||||||
1
server/reflector/pipelines/__init__.py
Normal file
1
server/reflector/pipelines/__init__.py
Normal file
@@ -0,0 +1 @@
|
|||||||
|
"""Pipeline modules for audio processing."""
|
||||||
377
server/reflector/pipelines/main_file_pipeline.py
Normal file
377
server/reflector/pipelines/main_file_pipeline.py
Normal file
@@ -0,0 +1,377 @@
|
|||||||
|
"""
|
||||||
|
File-based processing pipeline
|
||||||
|
==============================
|
||||||
|
|
||||||
|
Optimized pipeline for processing complete audio/video files.
|
||||||
|
Uses parallel processing for transcription, diarization, and waveform generation.
|
||||||
|
"""
|
||||||
|
|
||||||
|
import asyncio
|
||||||
|
import uuid
|
||||||
|
from pathlib import Path
|
||||||
|
|
||||||
|
import av
|
||||||
|
import structlog
|
||||||
|
from celery import chain, shared_task
|
||||||
|
|
||||||
|
from reflector.asynctask import asynctask
|
||||||
|
from reflector.db.rooms import rooms_controller
|
||||||
|
from reflector.db.transcripts import (
|
||||||
|
SourceKind,
|
||||||
|
Transcript,
|
||||||
|
TranscriptStatus,
|
||||||
|
transcripts_controller,
|
||||||
|
)
|
||||||
|
from reflector.logger import logger
|
||||||
|
from reflector.pipelines import topic_processing
|
||||||
|
from reflector.pipelines.main_live_pipeline import (
|
||||||
|
PipelineMainBase,
|
||||||
|
broadcast_to_sockets,
|
||||||
|
task_cleanup_consent,
|
||||||
|
task_pipeline_post_to_zulip,
|
||||||
|
)
|
||||||
|
from reflector.pipelines.transcription_helpers import transcribe_file_with_processor
|
||||||
|
from reflector.processors import AudioFileWriterProcessor
|
||||||
|
from reflector.processors.audio_waveform_processor import AudioWaveformProcessor
|
||||||
|
from reflector.processors.file_diarization import FileDiarizationInput
|
||||||
|
from reflector.processors.file_diarization_auto import FileDiarizationAutoProcessor
|
||||||
|
from reflector.processors.transcript_diarization_assembler import (
|
||||||
|
TranscriptDiarizationAssemblerInput,
|
||||||
|
TranscriptDiarizationAssemblerProcessor,
|
||||||
|
)
|
||||||
|
from reflector.processors.types import (
|
||||||
|
DiarizationSegment,
|
||||||
|
TitleSummary,
|
||||||
|
)
|
||||||
|
from reflector.processors.types import (
|
||||||
|
Transcript as TranscriptType,
|
||||||
|
)
|
||||||
|
from reflector.settings import settings
|
||||||
|
from reflector.storage import get_transcripts_storage
|
||||||
|
from reflector.worker.webhook import send_transcript_webhook
|
||||||
|
|
||||||
|
|
||||||
|
class PipelineMainFile(PipelineMainBase):
|
||||||
|
"""
|
||||||
|
Optimized file processing pipeline.
|
||||||
|
Processes complete audio/video files with parallel execution.
|
||||||
|
"""
|
||||||
|
|
||||||
|
logger: structlog.BoundLogger = None
|
||||||
|
empty_pipeline = None
|
||||||
|
|
||||||
|
def __init__(self, transcript_id: str):
|
||||||
|
super().__init__(transcript_id=transcript_id)
|
||||||
|
self.logger = logger.bind(transcript_id=self.transcript_id)
|
||||||
|
self.empty_pipeline = topic_processing.EmptyPipeline(logger=self.logger)
|
||||||
|
|
||||||
|
def _handle_gather_exceptions(self, results: list, operation: str) -> None:
|
||||||
|
"""Handle exceptions from asyncio.gather with return_exceptions=True"""
|
||||||
|
for i, result in enumerate(results):
|
||||||
|
if not isinstance(result, Exception):
|
||||||
|
continue
|
||||||
|
self.logger.error(
|
||||||
|
f"Error in {operation} (task {i}): {result}",
|
||||||
|
transcript_id=self.transcript_id,
|
||||||
|
exc_info=result,
|
||||||
|
)
|
||||||
|
|
||||||
|
@broadcast_to_sockets
|
||||||
|
async def set_status(self, transcript_id: str, status: TranscriptStatus):
|
||||||
|
async with self.lock_transaction():
|
||||||
|
return await transcripts_controller.set_status(transcript_id, status)
|
||||||
|
|
||||||
|
async def process(self, file_path: Path):
|
||||||
|
"""Main entry point for file processing"""
|
||||||
|
self.logger.info(f"Starting file pipeline for {file_path}")
|
||||||
|
|
||||||
|
transcript = await self.get_transcript()
|
||||||
|
|
||||||
|
# Clear transcript as we're going to regenerate everything
|
||||||
|
async with self.transaction():
|
||||||
|
await transcripts_controller.update(
|
||||||
|
transcript,
|
||||||
|
{
|
||||||
|
"events": [],
|
||||||
|
"topics": [],
|
||||||
|
},
|
||||||
|
)
|
||||||
|
|
||||||
|
# Extract audio and write to transcript location
|
||||||
|
audio_path = await self.extract_and_write_audio(file_path, transcript)
|
||||||
|
|
||||||
|
# Upload for processing
|
||||||
|
audio_url = await self.upload_audio(audio_path, transcript)
|
||||||
|
|
||||||
|
# Run parallel processing
|
||||||
|
await self.run_parallel_processing(
|
||||||
|
audio_path,
|
||||||
|
audio_url,
|
||||||
|
transcript.source_language,
|
||||||
|
transcript.target_language,
|
||||||
|
)
|
||||||
|
|
||||||
|
self.logger.info("File pipeline complete")
|
||||||
|
|
||||||
|
await self.set_status(transcript.id, "ended")
|
||||||
|
|
||||||
|
async def extract_and_write_audio(
|
||||||
|
self, file_path: Path, transcript: Transcript
|
||||||
|
) -> Path:
|
||||||
|
"""Extract audio from video if needed and write to transcript location as MP3"""
|
||||||
|
self.logger.info(f"Processing audio file: {file_path}")
|
||||||
|
|
||||||
|
# Check if it's already audio-only
|
||||||
|
container = av.open(str(file_path))
|
||||||
|
has_video = len(container.streams.video) > 0
|
||||||
|
container.close()
|
||||||
|
|
||||||
|
# Use AudioFileWriterProcessor to write MP3 to transcript location
|
||||||
|
mp3_writer = AudioFileWriterProcessor(
|
||||||
|
path=transcript.audio_mp3_filename,
|
||||||
|
on_duration=self.on_duration,
|
||||||
|
)
|
||||||
|
|
||||||
|
# Process audio frames and write to transcript location
|
||||||
|
input_container = av.open(str(file_path))
|
||||||
|
for frame in input_container.decode(audio=0):
|
||||||
|
await mp3_writer.push(frame)
|
||||||
|
|
||||||
|
await mp3_writer.flush()
|
||||||
|
input_container.close()
|
||||||
|
|
||||||
|
if has_video:
|
||||||
|
self.logger.info(
|
||||||
|
f"Extracted audio from video and saved to {transcript.audio_mp3_filename}"
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
self.logger.info(
|
||||||
|
f"Converted audio file and saved to {transcript.audio_mp3_filename}"
|
||||||
|
)
|
||||||
|
|
||||||
|
return transcript.audio_mp3_filename
|
||||||
|
|
||||||
|
async def upload_audio(self, audio_path: Path, transcript: Transcript) -> str:
|
||||||
|
"""Upload audio to storage for processing"""
|
||||||
|
storage = get_transcripts_storage()
|
||||||
|
|
||||||
|
if not storage:
|
||||||
|
raise Exception(
|
||||||
|
"Storage backend required for file processing. Configure TRANSCRIPT_STORAGE_* settings."
|
||||||
|
)
|
||||||
|
|
||||||
|
self.logger.info("Uploading audio to storage")
|
||||||
|
|
||||||
|
with open(audio_path, "rb") as f:
|
||||||
|
audio_data = f.read()
|
||||||
|
|
||||||
|
storage_path = f"file_pipeline/{transcript.id}/audio.mp3"
|
||||||
|
await storage.put_file(storage_path, audio_data)
|
||||||
|
|
||||||
|
audio_url = await storage.get_file_url(storage_path)
|
||||||
|
|
||||||
|
self.logger.info(f"Audio uploaded to {audio_url}")
|
||||||
|
return audio_url
|
||||||
|
|
||||||
|
async def run_parallel_processing(
|
||||||
|
self,
|
||||||
|
audio_path: Path,
|
||||||
|
audio_url: str,
|
||||||
|
source_language: str,
|
||||||
|
target_language: str,
|
||||||
|
):
|
||||||
|
"""Coordinate parallel processing of transcription, diarization, and waveform"""
|
||||||
|
self.logger.info(
|
||||||
|
"Starting parallel processing", transcript_id=self.transcript_id
|
||||||
|
)
|
||||||
|
|
||||||
|
# Phase 1: Parallel processing of independent tasks
|
||||||
|
transcription_task = self.transcribe_file(audio_url, source_language)
|
||||||
|
diarization_task = self.diarize_file(audio_url)
|
||||||
|
waveform_task = self.generate_waveform(audio_path)
|
||||||
|
|
||||||
|
results = await asyncio.gather(
|
||||||
|
transcription_task, diarization_task, waveform_task, return_exceptions=True
|
||||||
|
)
|
||||||
|
|
||||||
|
transcript_result = results[0]
|
||||||
|
diarization_result = results[1]
|
||||||
|
|
||||||
|
# Handle errors - raise any exception that occurred
|
||||||
|
self._handle_gather_exceptions(results, "parallel processing")
|
||||||
|
for result in results:
|
||||||
|
if isinstance(result, Exception):
|
||||||
|
raise result
|
||||||
|
|
||||||
|
# Phase 2: Assemble transcript with diarization
|
||||||
|
self.logger.info(
|
||||||
|
"Assembling transcript with diarization", transcript_id=self.transcript_id
|
||||||
|
)
|
||||||
|
processor = TranscriptDiarizationAssemblerProcessor()
|
||||||
|
input_data = TranscriptDiarizationAssemblerInput(
|
||||||
|
transcript=transcript_result, diarization=diarization_result or []
|
||||||
|
)
|
||||||
|
|
||||||
|
# Store result for retrieval
|
||||||
|
diarized_transcript: Transcript | None = None
|
||||||
|
|
||||||
|
async def capture_result(transcript):
|
||||||
|
nonlocal diarized_transcript
|
||||||
|
diarized_transcript = transcript
|
||||||
|
|
||||||
|
processor.on(capture_result)
|
||||||
|
await processor.push(input_data)
|
||||||
|
await processor.flush()
|
||||||
|
|
||||||
|
if not diarized_transcript:
|
||||||
|
raise ValueError("No diarized transcript captured")
|
||||||
|
|
||||||
|
# Phase 3: Generate topics from diarized transcript
|
||||||
|
self.logger.info("Generating topics", transcript_id=self.transcript_id)
|
||||||
|
topics = await self.detect_topics(diarized_transcript, target_language)
|
||||||
|
|
||||||
|
# Phase 4: Generate title and summaries in parallel
|
||||||
|
self.logger.info(
|
||||||
|
"Generating title and summaries", transcript_id=self.transcript_id
|
||||||
|
)
|
||||||
|
results = await asyncio.gather(
|
||||||
|
self.generate_title(topics),
|
||||||
|
self.generate_summaries(topics),
|
||||||
|
return_exceptions=True,
|
||||||
|
)
|
||||||
|
|
||||||
|
self._handle_gather_exceptions(results, "title and summary generation")
|
||||||
|
|
||||||
|
async def transcribe_file(self, audio_url: str, language: str) -> TranscriptType:
|
||||||
|
"""Transcribe complete file"""
|
||||||
|
return await transcribe_file_with_processor(audio_url, language)
|
||||||
|
|
||||||
|
async def diarize_file(self, audio_url: str) -> list[DiarizationSegment] | None:
|
||||||
|
"""Get diarization for file"""
|
||||||
|
if not settings.DIARIZATION_BACKEND:
|
||||||
|
self.logger.info("Diarization disabled")
|
||||||
|
return None
|
||||||
|
|
||||||
|
processor = FileDiarizationAutoProcessor()
|
||||||
|
input_data = FileDiarizationInput(audio_url=audio_url)
|
||||||
|
|
||||||
|
# Store result for retrieval
|
||||||
|
result = None
|
||||||
|
|
||||||
|
async def capture_result(diarization_output):
|
||||||
|
nonlocal result
|
||||||
|
result = diarization_output.diarization
|
||||||
|
|
||||||
|
try:
|
||||||
|
processor.on(capture_result)
|
||||||
|
await processor.push(input_data)
|
||||||
|
await processor.flush()
|
||||||
|
return result
|
||||||
|
except Exception as e:
|
||||||
|
self.logger.error(f"Diarization failed: {e}")
|
||||||
|
return None
|
||||||
|
|
||||||
|
async def generate_waveform(self, audio_path: Path):
|
||||||
|
"""Generate and save waveform"""
|
||||||
|
transcript = await self.get_transcript()
|
||||||
|
|
||||||
|
processor = AudioWaveformProcessor(
|
||||||
|
audio_path=audio_path,
|
||||||
|
waveform_path=transcript.audio_waveform_filename,
|
||||||
|
on_waveform=self.on_waveform,
|
||||||
|
)
|
||||||
|
processor.set_pipeline(self.empty_pipeline)
|
||||||
|
|
||||||
|
await processor.flush()
|
||||||
|
|
||||||
|
async def detect_topics(
|
||||||
|
self, transcript: TranscriptType, target_language: str
|
||||||
|
) -> list[TitleSummary]:
|
||||||
|
return await topic_processing.detect_topics(
|
||||||
|
transcript,
|
||||||
|
target_language,
|
||||||
|
on_topic_callback=self.on_topic,
|
||||||
|
empty_pipeline=self.empty_pipeline,
|
||||||
|
)
|
||||||
|
|
||||||
|
async def generate_title(self, topics: list[TitleSummary]):
|
||||||
|
return await topic_processing.generate_title(
|
||||||
|
topics,
|
||||||
|
on_title_callback=self.on_title,
|
||||||
|
empty_pipeline=self.empty_pipeline,
|
||||||
|
logger=self.logger,
|
||||||
|
)
|
||||||
|
|
||||||
|
async def generate_summaries(self, topics: list[TitleSummary]):
|
||||||
|
transcript = await self.get_transcript()
|
||||||
|
return await topic_processing.generate_summaries(
|
||||||
|
topics,
|
||||||
|
transcript,
|
||||||
|
on_long_summary_callback=self.on_long_summary,
|
||||||
|
on_short_summary_callback=self.on_short_summary,
|
||||||
|
empty_pipeline=self.empty_pipeline,
|
||||||
|
logger=self.logger,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@shared_task
|
||||||
|
@asynctask
|
||||||
|
async def task_send_webhook_if_needed(*, transcript_id: str):
|
||||||
|
"""Send webhook if this is a room recording with webhook configured"""
|
||||||
|
transcript = await transcripts_controller.get_by_id(transcript_id)
|
||||||
|
if not transcript:
|
||||||
|
return
|
||||||
|
|
||||||
|
if transcript.source_kind == SourceKind.ROOM and transcript.room_id:
|
||||||
|
room = await rooms_controller.get_by_id(transcript.room_id)
|
||||||
|
if room and room.webhook_url:
|
||||||
|
logger.info(
|
||||||
|
"Dispatching webhook",
|
||||||
|
transcript_id=transcript_id,
|
||||||
|
room_id=room.id,
|
||||||
|
webhook_url=room.webhook_url,
|
||||||
|
)
|
||||||
|
send_transcript_webhook.delay(
|
||||||
|
transcript_id, room.id, event_id=uuid.uuid4().hex
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@shared_task
|
||||||
|
@asynctask
|
||||||
|
async def task_pipeline_file_process(*, transcript_id: str):
|
||||||
|
"""Celery task for file pipeline processing"""
|
||||||
|
|
||||||
|
transcript = await transcripts_controller.get_by_id(transcript_id)
|
||||||
|
if not transcript:
|
||||||
|
raise Exception(f"Transcript {transcript_id} not found")
|
||||||
|
|
||||||
|
pipeline = PipelineMainFile(transcript_id=transcript_id)
|
||||||
|
try:
|
||||||
|
await pipeline.set_status(transcript_id, "processing")
|
||||||
|
|
||||||
|
# Find the file to process
|
||||||
|
audio_file = next(transcript.data_path.glob("upload.*"), None)
|
||||||
|
if not audio_file:
|
||||||
|
audio_file = next(transcript.data_path.glob("audio.*"), None)
|
||||||
|
|
||||||
|
if not audio_file:
|
||||||
|
raise Exception("No audio file found to process")
|
||||||
|
|
||||||
|
await pipeline.process(audio_file)
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
logger.error(
|
||||||
|
f"File pipeline failed for transcript {transcript_id}: {type(e).__name__}: {str(e)}",
|
||||||
|
exc_info=True,
|
||||||
|
transcript_id=transcript_id,
|
||||||
|
)
|
||||||
|
await pipeline.set_status(transcript_id, "error")
|
||||||
|
raise
|
||||||
|
|
||||||
|
# Run post-processing chain: consent cleanup -> zulip -> webhook
|
||||||
|
post_chain = chain(
|
||||||
|
task_cleanup_consent.si(transcript_id=transcript_id),
|
||||||
|
task_pipeline_post_to_zulip.si(transcript_id=transcript_id),
|
||||||
|
task_send_webhook_if_needed.si(transcript_id=transcript_id),
|
||||||
|
)
|
||||||
|
post_chain.delay()
|
||||||
@@ -17,12 +17,11 @@ from contextlib import asynccontextmanager
|
|||||||
from typing import Generic
|
from typing import Generic
|
||||||
|
|
||||||
import av
|
import av
|
||||||
import boto3
|
|
||||||
from celery import chord, current_task, group, shared_task
|
from celery import chord, current_task, group, shared_task
|
||||||
from pydantic import BaseModel
|
from pydantic import BaseModel
|
||||||
from structlog import BoundLogger as Logger
|
from structlog import BoundLogger as Logger
|
||||||
|
|
||||||
from reflector.db import get_database
|
from reflector.asynctask import asynctask
|
||||||
from reflector.db.meetings import meeting_consent_controller, meetings_controller
|
from reflector.db.meetings import meeting_consent_controller, meetings_controller
|
||||||
from reflector.db.recordings import recordings_controller
|
from reflector.db.recordings import recordings_controller
|
||||||
from reflector.db.rooms import rooms_controller
|
from reflector.db.rooms import rooms_controller
|
||||||
@@ -32,6 +31,7 @@ from reflector.db.transcripts import (
|
|||||||
TranscriptFinalLongSummary,
|
TranscriptFinalLongSummary,
|
||||||
TranscriptFinalShortSummary,
|
TranscriptFinalShortSummary,
|
||||||
TranscriptFinalTitle,
|
TranscriptFinalTitle,
|
||||||
|
TranscriptStatus,
|
||||||
TranscriptText,
|
TranscriptText,
|
||||||
TranscriptTopic,
|
TranscriptTopic,
|
||||||
TranscriptWaveform,
|
TranscriptWaveform,
|
||||||
@@ -40,8 +40,9 @@ from reflector.db.transcripts import (
|
|||||||
from reflector.logger import logger
|
from reflector.logger import logger
|
||||||
from reflector.pipelines.runner import PipelineMessage, PipelineRunner
|
from reflector.pipelines.runner import PipelineMessage, PipelineRunner
|
||||||
from reflector.processors import (
|
from reflector.processors import (
|
||||||
AudioChunkerProcessor,
|
AudioChunkerAutoProcessor,
|
||||||
AudioDiarizationAutoProcessor,
|
AudioDiarizationAutoProcessor,
|
||||||
|
AudioDownscaleProcessor,
|
||||||
AudioFileWriterProcessor,
|
AudioFileWriterProcessor,
|
||||||
AudioMergeProcessor,
|
AudioMergeProcessor,
|
||||||
AudioTranscriptAutoProcessor,
|
AudioTranscriptAutoProcessor,
|
||||||
@@ -68,29 +69,6 @@ from reflector.zulip import (
|
|||||||
)
|
)
|
||||||
|
|
||||||
|
|
||||||
def asynctask(f):
|
|
||||||
@functools.wraps(f)
|
|
||||||
def wrapper(*args, **kwargs):
|
|
||||||
async def run_with_db():
|
|
||||||
database = get_database()
|
|
||||||
await database.connect()
|
|
||||||
try:
|
|
||||||
return await f(*args, **kwargs)
|
|
||||||
finally:
|
|
||||||
await database.disconnect()
|
|
||||||
|
|
||||||
coro = run_with_db()
|
|
||||||
try:
|
|
||||||
loop = asyncio.get_running_loop()
|
|
||||||
except RuntimeError:
|
|
||||||
loop = None
|
|
||||||
if loop and loop.is_running():
|
|
||||||
return loop.run_until_complete(coro)
|
|
||||||
return asyncio.run(coro)
|
|
||||||
|
|
||||||
return wrapper
|
|
||||||
|
|
||||||
|
|
||||||
def broadcast_to_sockets(func):
|
def broadcast_to_sockets(func):
|
||||||
"""
|
"""
|
||||||
Decorator to broadcast transcript event to websockets
|
Decorator to broadcast transcript event to websockets
|
||||||
@@ -106,6 +84,20 @@ def broadcast_to_sockets(func):
|
|||||||
message=resp.model_dump(mode="json"),
|
message=resp.model_dump(mode="json"),
|
||||||
)
|
)
|
||||||
|
|
||||||
|
transcript = await transcripts_controller.get_by_id(self.transcript_id)
|
||||||
|
if transcript and transcript.user_id:
|
||||||
|
# Emit only relevant events to the user room to avoid noisy updates.
|
||||||
|
# Allowed: STATUS, FINAL_TITLE, DURATION. All are prefixed with TRANSCRIPT_
|
||||||
|
allowed_user_events = {"STATUS", "FINAL_TITLE", "DURATION"}
|
||||||
|
if resp.event in allowed_user_events:
|
||||||
|
await self.ws_manager.send_json(
|
||||||
|
room_id=f"user:{transcript.user_id}",
|
||||||
|
message={
|
||||||
|
"event": f"TRANSCRIPT_{resp.event}",
|
||||||
|
"data": {"id": self.transcript_id, **resp.data},
|
||||||
|
},
|
||||||
|
)
|
||||||
|
|
||||||
return wrapper
|
return wrapper
|
||||||
|
|
||||||
|
|
||||||
@@ -147,15 +139,18 @@ class StrValue(BaseModel):
|
|||||||
|
|
||||||
|
|
||||||
class PipelineMainBase(PipelineRunner[PipelineMessage], Generic[PipelineMessage]):
|
class PipelineMainBase(PipelineRunner[PipelineMessage], Generic[PipelineMessage]):
|
||||||
transcript_id: str
|
def __init__(self, transcript_id: str):
|
||||||
ws_room_id: str | None = None
|
super().__init__()
|
||||||
ws_manager: WebsocketManager | None = None
|
|
||||||
|
|
||||||
def prepare(self):
|
|
||||||
# prepare websocket
|
|
||||||
self._lock = asyncio.Lock()
|
self._lock = asyncio.Lock()
|
||||||
|
self.transcript_id = transcript_id
|
||||||
self.ws_room_id = f"ts:{self.transcript_id}"
|
self.ws_room_id = f"ts:{self.transcript_id}"
|
||||||
self.ws_manager = get_ws_manager()
|
self._ws_manager = None
|
||||||
|
|
||||||
|
@property
|
||||||
|
def ws_manager(self) -> WebsocketManager:
|
||||||
|
if self._ws_manager is None:
|
||||||
|
self._ws_manager = get_ws_manager()
|
||||||
|
return self._ws_manager
|
||||||
|
|
||||||
async def get_transcript(self) -> Transcript:
|
async def get_transcript(self) -> Transcript:
|
||||||
# fetch the transcript
|
# fetch the transcript
|
||||||
@@ -184,8 +179,15 @@ class PipelineMainBase(PipelineRunner[PipelineMessage], Generic[PipelineMessage]
|
|||||||
]
|
]
|
||||||
|
|
||||||
@asynccontextmanager
|
@asynccontextmanager
|
||||||
async def transaction(self):
|
async def lock_transaction(self):
|
||||||
|
# This lock is to prevent multiple processor starting adding
|
||||||
|
# into event array at the same time
|
||||||
async with self._lock:
|
async with self._lock:
|
||||||
|
yield
|
||||||
|
|
||||||
|
@asynccontextmanager
|
||||||
|
async def transaction(self):
|
||||||
|
async with self.lock_transaction():
|
||||||
async with transcripts_controller.transaction():
|
async with transcripts_controller.transaction():
|
||||||
yield
|
yield
|
||||||
|
|
||||||
@@ -194,14 +196,14 @@ class PipelineMainBase(PipelineRunner[PipelineMessage], Generic[PipelineMessage]
|
|||||||
# if it's the first part, update the status of the transcript
|
# if it's the first part, update the status of the transcript
|
||||||
# but do not set the ended status yet.
|
# but do not set the ended status yet.
|
||||||
if isinstance(self, PipelineMainLive):
|
if isinstance(self, PipelineMainLive):
|
||||||
status_mapping = {
|
status_mapping: dict[str, TranscriptStatus] = {
|
||||||
"started": "recording",
|
"started": "recording",
|
||||||
"push": "recording",
|
"push": "recording",
|
||||||
"flush": "processing",
|
"flush": "processing",
|
||||||
"error": "error",
|
"error": "error",
|
||||||
}
|
}
|
||||||
elif isinstance(self, PipelineMainFinalSummaries):
|
elif isinstance(self, PipelineMainFinalSummaries):
|
||||||
status_mapping = {
|
status_mapping: dict[str, TranscriptStatus] = {
|
||||||
"push": "processing",
|
"push": "processing",
|
||||||
"flush": "processing",
|
"flush": "processing",
|
||||||
"error": "error",
|
"error": "error",
|
||||||
@@ -217,22 +219,8 @@ class PipelineMainBase(PipelineRunner[PipelineMessage], Generic[PipelineMessage]
|
|||||||
return
|
return
|
||||||
|
|
||||||
# when the status of the pipeline changes, update the transcript
|
# when the status of the pipeline changes, update the transcript
|
||||||
async with self.transaction():
|
async with self._lock:
|
||||||
transcript = await self.get_transcript()
|
return await transcripts_controller.set_status(self.transcript_id, status)
|
||||||
if status == transcript.status:
|
|
||||||
return
|
|
||||||
resp = await transcripts_controller.append_event(
|
|
||||||
transcript=transcript,
|
|
||||||
event="STATUS",
|
|
||||||
data=StrValue(value=status),
|
|
||||||
)
|
|
||||||
await transcripts_controller.update(
|
|
||||||
transcript,
|
|
||||||
{
|
|
||||||
"status": status,
|
|
||||||
},
|
|
||||||
)
|
|
||||||
return resp
|
|
||||||
|
|
||||||
@broadcast_to_sockets
|
@broadcast_to_sockets
|
||||||
async def on_transcript(self, data):
|
async def on_transcript(self, data):
|
||||||
@@ -355,7 +343,6 @@ class PipelineMainLive(PipelineMainBase):
|
|||||||
async def create(self) -> Pipeline:
|
async def create(self) -> Pipeline:
|
||||||
# create a context for the whole rtc transaction
|
# create a context for the whole rtc transaction
|
||||||
# add a customised logger to the context
|
# add a customised logger to the context
|
||||||
self.prepare()
|
|
||||||
transcript = await self.get_transcript()
|
transcript = await self.get_transcript()
|
||||||
|
|
||||||
processors = [
|
processors = [
|
||||||
@@ -363,7 +350,8 @@ class PipelineMainLive(PipelineMainBase):
|
|||||||
path=transcript.audio_wav_filename,
|
path=transcript.audio_wav_filename,
|
||||||
on_duration=self.on_duration,
|
on_duration=self.on_duration,
|
||||||
),
|
),
|
||||||
AudioChunkerProcessor(),
|
AudioDownscaleProcessor(),
|
||||||
|
AudioChunkerAutoProcessor(),
|
||||||
AudioMergeProcessor(),
|
AudioMergeProcessor(),
|
||||||
AudioTranscriptAutoProcessor.as_threaded(),
|
AudioTranscriptAutoProcessor.as_threaded(),
|
||||||
TranscriptLinerProcessor(),
|
TranscriptLinerProcessor(),
|
||||||
@@ -376,6 +364,7 @@ class PipelineMainLive(PipelineMainBase):
|
|||||||
pipeline.set_pref("audio:target_language", transcript.target_language)
|
pipeline.set_pref("audio:target_language", transcript.target_language)
|
||||||
pipeline.logger.bind(transcript_id=transcript.id)
|
pipeline.logger.bind(transcript_id=transcript.id)
|
||||||
pipeline.logger.info("Pipeline main live created")
|
pipeline.logger.info("Pipeline main live created")
|
||||||
|
pipeline.describe()
|
||||||
|
|
||||||
return pipeline
|
return pipeline
|
||||||
|
|
||||||
@@ -394,7 +383,6 @@ class PipelineMainDiarization(PipelineMainBase[AudioDiarizationInput]):
|
|||||||
async def create(self) -> Pipeline:
|
async def create(self) -> Pipeline:
|
||||||
# create a context for the whole rtc transaction
|
# create a context for the whole rtc transaction
|
||||||
# add a customised logger to the context
|
# add a customised logger to the context
|
||||||
self.prepare()
|
|
||||||
pipeline = Pipeline(
|
pipeline = Pipeline(
|
||||||
AudioDiarizationAutoProcessor(callback=self.on_topic),
|
AudioDiarizationAutoProcessor(callback=self.on_topic),
|
||||||
)
|
)
|
||||||
@@ -435,8 +423,6 @@ class PipelineMainFromTopics(PipelineMainBase[TitleSummaryWithIdProcessorType]):
|
|||||||
raise NotImplementedError
|
raise NotImplementedError
|
||||||
|
|
||||||
async def create(self) -> Pipeline:
|
async def create(self) -> Pipeline:
|
||||||
self.prepare()
|
|
||||||
|
|
||||||
# get transcript
|
# get transcript
|
||||||
self._transcript = transcript = await self.get_transcript()
|
self._transcript = transcript = await self.get_transcript()
|
||||||
|
|
||||||
@@ -597,6 +583,7 @@ async def cleanup_consent(transcript: Transcript, logger: Logger):
|
|||||||
|
|
||||||
consent_denied = False
|
consent_denied = False
|
||||||
recording = None
|
recording = None
|
||||||
|
meeting = None
|
||||||
try:
|
try:
|
||||||
if transcript.recording_id:
|
if transcript.recording_id:
|
||||||
recording = await recordings_controller.get_by_id(transcript.recording_id)
|
recording = await recordings_controller.get_by_id(transcript.recording_id)
|
||||||
@@ -607,8 +594,8 @@ async def cleanup_consent(transcript: Transcript, logger: Logger):
|
|||||||
meeting.id
|
meeting.id
|
||||||
)
|
)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.error(f"Failed to get fetch consent: {e}", exc_info=e)
|
logger.error(f"Failed to fetch consent: {e}", exc_info=e)
|
||||||
consent_denied = True
|
raise
|
||||||
|
|
||||||
if not consent_denied:
|
if not consent_denied:
|
||||||
logger.info("Consent approved, keeping all files")
|
logger.info("Consent approved, keeping all files")
|
||||||
@@ -616,25 +603,24 @@ async def cleanup_consent(transcript: Transcript, logger: Logger):
|
|||||||
|
|
||||||
logger.info("Consent denied, cleaning up all related audio files")
|
logger.info("Consent denied, cleaning up all related audio files")
|
||||||
|
|
||||||
if recording and recording.bucket_name and recording.object_key:
|
deletion_errors = []
|
||||||
s3_whereby = boto3.client(
|
if recording and recording.bucket_name:
|
||||||
"s3",
|
keys_to_delete = []
|
||||||
aws_access_key_id=settings.AWS_WHEREBY_ACCESS_KEY_ID,
|
if recording.track_keys:
|
||||||
aws_secret_access_key=settings.AWS_WHEREBY_ACCESS_KEY_SECRET,
|
keys_to_delete = recording.track_keys
|
||||||
)
|
elif recording.object_key:
|
||||||
try:
|
keys_to_delete = [recording.object_key]
|
||||||
s3_whereby.delete_object(
|
|
||||||
Bucket=recording.bucket_name, Key=recording.object_key
|
master_storage = get_transcripts_storage()
|
||||||
)
|
for key in keys_to_delete:
|
||||||
logger.info(
|
try:
|
||||||
f"Deleted original Whereby recording: {recording.bucket_name}/{recording.object_key}"
|
await master_storage.delete_file(key, bucket=recording.bucket_name)
|
||||||
)
|
logger.info(f"Deleted recording file: {recording.bucket_name}/{key}")
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.error(f"Failed to delete Whereby recording: {e}", exc_info=e)
|
error_msg = f"Failed to delete {key}: {e}"
|
||||||
|
logger.error(error_msg, exc_info=e)
|
||||||
|
deletion_errors.append(error_msg)
|
||||||
|
|
||||||
# non-transactional, files marked for deletion not actually deleted is possible
|
|
||||||
await transcripts_controller.update(transcript, {"audio_deleted": True})
|
|
||||||
# 2. Delete processed audio from transcript storage S3 bucket
|
|
||||||
if transcript.audio_location == "storage":
|
if transcript.audio_location == "storage":
|
||||||
storage = get_transcripts_storage()
|
storage = get_transcripts_storage()
|
||||||
try:
|
try:
|
||||||
@@ -643,18 +629,28 @@ async def cleanup_consent(transcript: Transcript, logger: Logger):
|
|||||||
f"Deleted processed audio from storage: {transcript.storage_audio_path}"
|
f"Deleted processed audio from storage: {transcript.storage_audio_path}"
|
||||||
)
|
)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.error(f"Failed to delete processed audio: {e}", exc_info=e)
|
error_msg = f"Failed to delete processed audio: {e}"
|
||||||
|
logger.error(error_msg, exc_info=e)
|
||||||
|
deletion_errors.append(error_msg)
|
||||||
|
|
||||||
# 3. Delete local audio files
|
|
||||||
try:
|
try:
|
||||||
if hasattr(transcript, "audio_mp3_filename") and transcript.audio_mp3_filename:
|
if hasattr(transcript, "audio_mp3_filename") and transcript.audio_mp3_filename:
|
||||||
transcript.audio_mp3_filename.unlink(missing_ok=True)
|
transcript.audio_mp3_filename.unlink(missing_ok=True)
|
||||||
if hasattr(transcript, "audio_wav_filename") and transcript.audio_wav_filename:
|
if hasattr(transcript, "audio_wav_filename") and transcript.audio_wav_filename:
|
||||||
transcript.audio_wav_filename.unlink(missing_ok=True)
|
transcript.audio_wav_filename.unlink(missing_ok=True)
|
||||||
except Exception as e:
|
except Exception as e:
|
||||||
logger.error(f"Failed to delete local audio files: {e}", exc_info=e)
|
error_msg = f"Failed to delete local audio files: {e}"
|
||||||
|
logger.error(error_msg, exc_info=e)
|
||||||
|
deletion_errors.append(error_msg)
|
||||||
|
|
||||||
logger.info("Consent cleanup done")
|
if deletion_errors:
|
||||||
|
logger.warning(
|
||||||
|
f"Consent cleanup completed with {len(deletion_errors)} errors",
|
||||||
|
errors=deletion_errors,
|
||||||
|
)
|
||||||
|
else:
|
||||||
|
await transcripts_controller.update(transcript, {"audio_deleted": True})
|
||||||
|
logger.info("Consent cleanup done - all audio deleted")
|
||||||
|
|
||||||
|
|
||||||
@get_transcript
|
@get_transcript
|
||||||
@@ -792,7 +788,7 @@ def pipeline_post(*, transcript_id: str):
|
|||||||
chain_final_summaries,
|
chain_final_summaries,
|
||||||
) | task_pipeline_post_to_zulip.si(transcript_id=transcript_id)
|
) | task_pipeline_post_to_zulip.si(transcript_id=transcript_id)
|
||||||
|
|
||||||
chain.delay()
|
return chain.delay()
|
||||||
|
|
||||||
|
|
||||||
@get_transcript
|
@get_transcript
|
||||||
|
|||||||
694
server/reflector/pipelines/main_multitrack_pipeline.py
Normal file
694
server/reflector/pipelines/main_multitrack_pipeline.py
Normal file
@@ -0,0 +1,694 @@
|
|||||||
|
import asyncio
|
||||||
|
import math
|
||||||
|
import tempfile
|
||||||
|
from fractions import Fraction
|
||||||
|
from pathlib import Path
|
||||||
|
|
||||||
|
import av
|
||||||
|
from av.audio.resampler import AudioResampler
|
||||||
|
from celery import chain, shared_task
|
||||||
|
|
||||||
|
from reflector.asynctask import asynctask
|
||||||
|
from reflector.db.transcripts import (
|
||||||
|
TranscriptStatus,
|
||||||
|
TranscriptWaveform,
|
||||||
|
transcripts_controller,
|
||||||
|
)
|
||||||
|
from reflector.logger import logger
|
||||||
|
from reflector.pipelines import topic_processing
|
||||||
|
from reflector.pipelines.main_file_pipeline import task_send_webhook_if_needed
|
||||||
|
from reflector.pipelines.main_live_pipeline import (
|
||||||
|
PipelineMainBase,
|
||||||
|
broadcast_to_sockets,
|
||||||
|
task_cleanup_consent,
|
||||||
|
task_pipeline_post_to_zulip,
|
||||||
|
)
|
||||||
|
from reflector.pipelines.transcription_helpers import transcribe_file_with_processor
|
||||||
|
from reflector.processors import AudioFileWriterProcessor
|
||||||
|
from reflector.processors.audio_waveform_processor import AudioWaveformProcessor
|
||||||
|
from reflector.processors.types import TitleSummary
|
||||||
|
from reflector.processors.types import Transcript as TranscriptType
|
||||||
|
from reflector.storage import Storage, get_transcripts_storage
|
||||||
|
from reflector.utils.string import NonEmptyString
|
||||||
|
|
||||||
|
# Audio encoding constants
|
||||||
|
OPUS_STANDARD_SAMPLE_RATE = 48000
|
||||||
|
OPUS_DEFAULT_BIT_RATE = 128000
|
||||||
|
|
||||||
|
# Storage operation constants
|
||||||
|
PRESIGNED_URL_EXPIRATION_SECONDS = 7200 # 2 hours
|
||||||
|
|
||||||
|
|
||||||
|
class PipelineMainMultitrack(PipelineMainBase):
|
||||||
|
def __init__(self, transcript_id: str):
|
||||||
|
super().__init__(transcript_id=transcript_id)
|
||||||
|
self.logger = logger.bind(transcript_id=self.transcript_id)
|
||||||
|
self.empty_pipeline = topic_processing.EmptyPipeline(logger=self.logger)
|
||||||
|
|
||||||
|
async def pad_track_for_transcription(
|
||||||
|
self,
|
||||||
|
track_url: NonEmptyString,
|
||||||
|
track_idx: int,
|
||||||
|
storage: Storage,
|
||||||
|
) -> NonEmptyString:
|
||||||
|
"""
|
||||||
|
Pad a single track with silence based on stream metadata start_time.
|
||||||
|
Downloads from S3 presigned URL, processes via PyAV using tempfile, uploads to S3.
|
||||||
|
Returns presigned URL of padded track (or original URL if no padding needed).
|
||||||
|
|
||||||
|
Memory usage:
|
||||||
|
- Pattern: fixed_overhead(2-5MB) for PyAV codec/filters
|
||||||
|
- PyAV streams input efficiently (no full download, verified)
|
||||||
|
- Output written to tempfile (disk-based, not memory)
|
||||||
|
- Upload streams from file handle (boto3 chunks, typically 5-10MB)
|
||||||
|
|
||||||
|
Daily.co raw-tracks timing - Two approaches:
|
||||||
|
|
||||||
|
CURRENT APPROACH (PyAV metadata):
|
||||||
|
The WebM stream.start_time field encodes MEETING-RELATIVE timing:
|
||||||
|
- t=0: When Daily.co recording started (first participant joined)
|
||||||
|
- start_time=8.13s: This participant's track began 8.13s after recording started
|
||||||
|
- Purpose: Enables track alignment without external manifest files
|
||||||
|
|
||||||
|
This is NOT:
|
||||||
|
- Stream-internal offset (first packet timestamp relative to stream start)
|
||||||
|
- Absolute/wall-clock time
|
||||||
|
- Recording duration
|
||||||
|
|
||||||
|
ALTERNATIVE APPROACH (filename parsing):
|
||||||
|
Daily.co filenames contain Unix timestamps (milliseconds):
|
||||||
|
Format: {recording_start_ts}-{participant_id}-cam-audio-{track_start_ts}.webm
|
||||||
|
Example: 1760988935484-52f7f48b-fbab-431f-9a50-87b9abfc8255-cam-audio-1760988935922.webm
|
||||||
|
|
||||||
|
Can calculate offset: (track_start_ts - recording_start_ts) / 1000
|
||||||
|
- Track 0: (1760988935922 - 1760988935484) / 1000 = 0.438s
|
||||||
|
- Track 1: (1760988943823 - 1760988935484) / 1000 = 8.339s
|
||||||
|
|
||||||
|
TIME DIFFERENCE: PyAV metadata vs filename timestamps differ by ~209ms:
|
||||||
|
- Track 0: filename=438ms, metadata=229ms (diff: 209ms)
|
||||||
|
- Track 1: filename=8339ms, metadata=8130ms (diff: 209ms)
|
||||||
|
|
||||||
|
Consistent delta suggests network/encoding delay. PyAV metadata is ground truth
|
||||||
|
(represents when audio stream actually started vs when file upload initiated).
|
||||||
|
|
||||||
|
Example with 2 participants:
|
||||||
|
Track A: start_time=0.2s → Joined 200ms after recording began
|
||||||
|
Track B: start_time=8.1s → Joined 8.1 seconds later
|
||||||
|
|
||||||
|
After padding:
|
||||||
|
Track A: [0.2s silence] + [speech...]
|
||||||
|
Track B: [8.1s silence] + [speech...]
|
||||||
|
|
||||||
|
Whisper transcription timestamps are now synchronized:
|
||||||
|
Track A word at 5.0s → happened at meeting t=5.0s
|
||||||
|
Track B word at 10.0s → happened at meeting t=10.0s
|
||||||
|
|
||||||
|
Merging just sorts by timestamp - no offset calculation needed.
|
||||||
|
|
||||||
|
Padding coincidentally involves re-encoding. It's important when we work with Daily.co + Whisper.
|
||||||
|
This is because Daily.co returns recordings with skipped frames e.g. when microphone muted.
|
||||||
|
Daily.co doesn't understand those frames and ignores them, causing timestamp issues in transcription.
|
||||||
|
Re-encoding restores those frames. We do padding and re-encoding together just because it's convenient and more performant:
|
||||||
|
we need padded values for mix mp3 anyways
|
||||||
|
"""
|
||||||
|
|
||||||
|
transcript = await self.get_transcript()
|
||||||
|
|
||||||
|
try:
|
||||||
|
# PyAV streams input from S3 URL efficiently (2-5MB fixed overhead for codec/filters)
|
||||||
|
with av.open(track_url) as in_container:
|
||||||
|
start_time_seconds = self._extract_stream_start_time_from_container(
|
||||||
|
in_container, track_idx
|
||||||
|
)
|
||||||
|
|
||||||
|
if start_time_seconds <= 0:
|
||||||
|
self.logger.info(
|
||||||
|
f"Track {track_idx} requires no padding (start_time={start_time_seconds}s)",
|
||||||
|
track_idx=track_idx,
|
||||||
|
)
|
||||||
|
return track_url
|
||||||
|
|
||||||
|
# Use tempfile instead of BytesIO for better memory efficiency
|
||||||
|
# Reduces peak memory usage during encoding/upload
|
||||||
|
with tempfile.NamedTemporaryFile(
|
||||||
|
suffix=".webm", delete=False
|
||||||
|
) as temp_file:
|
||||||
|
temp_path = temp_file.name
|
||||||
|
|
||||||
|
try:
|
||||||
|
self._apply_audio_padding_to_file(
|
||||||
|
in_container, temp_path, start_time_seconds, track_idx
|
||||||
|
)
|
||||||
|
|
||||||
|
storage_path = (
|
||||||
|
f"file_pipeline/{transcript.id}/tracks/padded_{track_idx}.webm"
|
||||||
|
)
|
||||||
|
|
||||||
|
# Upload using file handle for streaming
|
||||||
|
with open(temp_path, "rb") as padded_file:
|
||||||
|
await storage.put_file(storage_path, padded_file)
|
||||||
|
finally:
|
||||||
|
# Clean up temp file
|
||||||
|
Path(temp_path).unlink(missing_ok=True)
|
||||||
|
|
||||||
|
padded_url = await storage.get_file_url(
|
||||||
|
storage_path,
|
||||||
|
operation="get_object",
|
||||||
|
expires_in=PRESIGNED_URL_EXPIRATION_SECONDS,
|
||||||
|
)
|
||||||
|
|
||||||
|
self.logger.info(
|
||||||
|
f"Successfully padded track {track_idx}",
|
||||||
|
track_idx=track_idx,
|
||||||
|
start_time_seconds=start_time_seconds,
|
||||||
|
padded_url=padded_url,
|
||||||
|
)
|
||||||
|
|
||||||
|
return padded_url
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
self.logger.error(
|
||||||
|
f"Failed to process track {track_idx}",
|
||||||
|
track_idx=track_idx,
|
||||||
|
url=track_url,
|
||||||
|
error=str(e),
|
||||||
|
exc_info=True,
|
||||||
|
)
|
||||||
|
raise Exception(
|
||||||
|
f"Track {track_idx} padding failed - transcript would have incorrect timestamps"
|
||||||
|
) from e
|
||||||
|
|
||||||
|
def _extract_stream_start_time_from_container(
|
||||||
|
self, container, track_idx: int
|
||||||
|
) -> float:
|
||||||
|
"""
|
||||||
|
Extract meeting-relative start time from WebM stream metadata.
|
||||||
|
Uses PyAV to read stream.start_time from WebM container.
|
||||||
|
More accurate than filename timestamps by ~209ms due to network/encoding delays.
|
||||||
|
"""
|
||||||
|
start_time_seconds = 0.0
|
||||||
|
try:
|
||||||
|
audio_streams = [s for s in container.streams if s.type == "audio"]
|
||||||
|
stream = audio_streams[0] if audio_streams else container.streams[0]
|
||||||
|
|
||||||
|
# 1) Try stream-level start_time (most reliable for Daily.co tracks)
|
||||||
|
if stream.start_time is not None and stream.time_base is not None:
|
||||||
|
start_time_seconds = float(stream.start_time * stream.time_base)
|
||||||
|
|
||||||
|
# 2) Fallback to container-level start_time (in av.time_base units)
|
||||||
|
if (start_time_seconds <= 0) and (container.start_time is not None):
|
||||||
|
start_time_seconds = float(container.start_time * av.time_base)
|
||||||
|
|
||||||
|
# 3) Fallback to first packet DTS in stream.time_base
|
||||||
|
if start_time_seconds <= 0:
|
||||||
|
for packet in container.demux(stream):
|
||||||
|
if packet.dts is not None:
|
||||||
|
start_time_seconds = float(packet.dts * stream.time_base)
|
||||||
|
break
|
||||||
|
except Exception as e:
|
||||||
|
self.logger.warning(
|
||||||
|
"PyAV metadata read failed; assuming 0 start_time",
|
||||||
|
track_idx=track_idx,
|
||||||
|
error=str(e),
|
||||||
|
)
|
||||||
|
start_time_seconds = 0.0
|
||||||
|
|
||||||
|
self.logger.info(
|
||||||
|
f"Track {track_idx} stream metadata: start_time={start_time_seconds:.3f}s",
|
||||||
|
track_idx=track_idx,
|
||||||
|
)
|
||||||
|
return start_time_seconds
|
||||||
|
|
||||||
|
def _apply_audio_padding_to_file(
|
||||||
|
self,
|
||||||
|
in_container,
|
||||||
|
output_path: str,
|
||||||
|
start_time_seconds: float,
|
||||||
|
track_idx: int,
|
||||||
|
) -> None:
|
||||||
|
"""Apply silence padding to audio track using PyAV filter graph, writing to file"""
|
||||||
|
delay_ms = math.floor(start_time_seconds * 1000)
|
||||||
|
|
||||||
|
self.logger.info(
|
||||||
|
f"Padding track {track_idx} with {delay_ms}ms delay using PyAV",
|
||||||
|
track_idx=track_idx,
|
||||||
|
delay_ms=delay_ms,
|
||||||
|
)
|
||||||
|
|
||||||
|
try:
|
||||||
|
with av.open(output_path, "w", format="webm") as out_container:
|
||||||
|
in_stream = next(
|
||||||
|
(s for s in in_container.streams if s.type == "audio"), None
|
||||||
|
)
|
||||||
|
if in_stream is None:
|
||||||
|
raise Exception("No audio stream in input")
|
||||||
|
|
||||||
|
out_stream = out_container.add_stream(
|
||||||
|
"libopus", rate=OPUS_STANDARD_SAMPLE_RATE
|
||||||
|
)
|
||||||
|
out_stream.bit_rate = OPUS_DEFAULT_BIT_RATE
|
||||||
|
graph = av.filter.Graph()
|
||||||
|
|
||||||
|
abuf_args = (
|
||||||
|
f"time_base=1/{OPUS_STANDARD_SAMPLE_RATE}:"
|
||||||
|
f"sample_rate={OPUS_STANDARD_SAMPLE_RATE}:"
|
||||||
|
f"sample_fmt=s16:"
|
||||||
|
f"channel_layout=stereo"
|
||||||
|
)
|
||||||
|
src = graph.add("abuffer", args=abuf_args, name="src")
|
||||||
|
aresample_f = graph.add("aresample", args="async=1", name="ares")
|
||||||
|
# adelay requires one delay value per channel separated by '|'
|
||||||
|
delays_arg = f"{delay_ms}|{delay_ms}"
|
||||||
|
adelay_f = graph.add(
|
||||||
|
"adelay", args=f"delays={delays_arg}:all=1", name="delay"
|
||||||
|
)
|
||||||
|
sink = graph.add("abuffersink", name="sink")
|
||||||
|
|
||||||
|
src.link_to(aresample_f)
|
||||||
|
aresample_f.link_to(adelay_f)
|
||||||
|
adelay_f.link_to(sink)
|
||||||
|
graph.configure()
|
||||||
|
|
||||||
|
resampler = AudioResampler(
|
||||||
|
format="s16", layout="stereo", rate=OPUS_STANDARD_SAMPLE_RATE
|
||||||
|
)
|
||||||
|
# Decode -> resample -> push through graph -> encode Opus
|
||||||
|
for frame in in_container.decode(in_stream):
|
||||||
|
out_frames = resampler.resample(frame) or []
|
||||||
|
for rframe in out_frames:
|
||||||
|
rframe.sample_rate = OPUS_STANDARD_SAMPLE_RATE
|
||||||
|
rframe.time_base = Fraction(1, OPUS_STANDARD_SAMPLE_RATE)
|
||||||
|
src.push(rframe)
|
||||||
|
|
||||||
|
while True:
|
||||||
|
try:
|
||||||
|
f_out = sink.pull()
|
||||||
|
except Exception:
|
||||||
|
break
|
||||||
|
f_out.sample_rate = OPUS_STANDARD_SAMPLE_RATE
|
||||||
|
f_out.time_base = Fraction(1, OPUS_STANDARD_SAMPLE_RATE)
|
||||||
|
for packet in out_stream.encode(f_out):
|
||||||
|
out_container.mux(packet)
|
||||||
|
|
||||||
|
src.push(None)
|
||||||
|
while True:
|
||||||
|
try:
|
||||||
|
f_out = sink.pull()
|
||||||
|
except Exception:
|
||||||
|
break
|
||||||
|
f_out.sample_rate = OPUS_STANDARD_SAMPLE_RATE
|
||||||
|
f_out.time_base = Fraction(1, OPUS_STANDARD_SAMPLE_RATE)
|
||||||
|
for packet in out_stream.encode(f_out):
|
||||||
|
out_container.mux(packet)
|
||||||
|
|
||||||
|
for packet in out_stream.encode(None):
|
||||||
|
out_container.mux(packet)
|
||||||
|
except Exception as e:
|
||||||
|
self.logger.error(
|
||||||
|
"PyAV padding failed for track",
|
||||||
|
track_idx=track_idx,
|
||||||
|
delay_ms=delay_ms,
|
||||||
|
error=str(e),
|
||||||
|
exc_info=True,
|
||||||
|
)
|
||||||
|
raise
|
||||||
|
|
||||||
|
async def mixdown_tracks(
|
||||||
|
self,
|
||||||
|
track_urls: list[str],
|
||||||
|
writer: AudioFileWriterProcessor,
|
||||||
|
offsets_seconds: list[float] | None = None,
|
||||||
|
) -> None:
|
||||||
|
"""Multi-track mixdown using PyAV filter graph (amix), reading from S3 presigned URLs"""
|
||||||
|
|
||||||
|
target_sample_rate: int | None = None
|
||||||
|
for url in track_urls:
|
||||||
|
if not url:
|
||||||
|
continue
|
||||||
|
container = None
|
||||||
|
try:
|
||||||
|
container = av.open(url)
|
||||||
|
for frame in container.decode(audio=0):
|
||||||
|
target_sample_rate = frame.sample_rate
|
||||||
|
break
|
||||||
|
except Exception:
|
||||||
|
continue
|
||||||
|
finally:
|
||||||
|
if container is not None:
|
||||||
|
container.close()
|
||||||
|
if target_sample_rate:
|
||||||
|
break
|
||||||
|
|
||||||
|
if not target_sample_rate:
|
||||||
|
self.logger.error("Mixdown failed - no decodable audio frames found")
|
||||||
|
raise Exception("Mixdown failed: No decodable audio frames in any track")
|
||||||
|
# Build PyAV filter graph:
|
||||||
|
# N abuffer (s32/stereo)
|
||||||
|
# -> optional adelay per input (for alignment)
|
||||||
|
# -> amix (s32)
|
||||||
|
# -> aformat(s16)
|
||||||
|
# -> sink
|
||||||
|
graph = av.filter.Graph()
|
||||||
|
inputs = []
|
||||||
|
valid_track_urls = [url for url in track_urls if url]
|
||||||
|
input_offsets_seconds = None
|
||||||
|
if offsets_seconds is not None:
|
||||||
|
input_offsets_seconds = [
|
||||||
|
offsets_seconds[i] for i, url in enumerate(track_urls) if url
|
||||||
|
]
|
||||||
|
for idx, url in enumerate(valid_track_urls):
|
||||||
|
args = (
|
||||||
|
f"time_base=1/{target_sample_rate}:"
|
||||||
|
f"sample_rate={target_sample_rate}:"
|
||||||
|
f"sample_fmt=s32:"
|
||||||
|
f"channel_layout=stereo"
|
||||||
|
)
|
||||||
|
in_ctx = graph.add("abuffer", args=args, name=f"in{idx}")
|
||||||
|
inputs.append(in_ctx)
|
||||||
|
|
||||||
|
if not inputs:
|
||||||
|
self.logger.error("Mixdown failed - no valid inputs for graph")
|
||||||
|
raise Exception("Mixdown failed: No valid inputs for filter graph")
|
||||||
|
|
||||||
|
mixer = graph.add("amix", args=f"inputs={len(inputs)}:normalize=0", name="mix")
|
||||||
|
|
||||||
|
fmt = graph.add(
|
||||||
|
"aformat",
|
||||||
|
args=(
|
||||||
|
f"sample_fmts=s32:channel_layouts=stereo:sample_rates={target_sample_rate}"
|
||||||
|
),
|
||||||
|
name="fmt",
|
||||||
|
)
|
||||||
|
|
||||||
|
sink = graph.add("abuffersink", name="out")
|
||||||
|
|
||||||
|
# Optional per-input delay before mixing
|
||||||
|
delays_ms: list[int] = []
|
||||||
|
if input_offsets_seconds is not None:
|
||||||
|
base = min(input_offsets_seconds) if input_offsets_seconds else 0.0
|
||||||
|
delays_ms = [
|
||||||
|
max(0, int(round((o - base) * 1000))) for o in input_offsets_seconds
|
||||||
|
]
|
||||||
|
else:
|
||||||
|
delays_ms = [0 for _ in inputs]
|
||||||
|
|
||||||
|
for idx, in_ctx in enumerate(inputs):
|
||||||
|
delay_ms = delays_ms[idx] if idx < len(delays_ms) else 0
|
||||||
|
if delay_ms > 0:
|
||||||
|
# adelay requires one value per channel; use same for stereo
|
||||||
|
adelay = graph.add(
|
||||||
|
"adelay",
|
||||||
|
args=f"delays={delay_ms}|{delay_ms}:all=1",
|
||||||
|
name=f"delay{idx}",
|
||||||
|
)
|
||||||
|
in_ctx.link_to(adelay)
|
||||||
|
adelay.link_to(mixer, 0, idx)
|
||||||
|
else:
|
||||||
|
in_ctx.link_to(mixer, 0, idx)
|
||||||
|
mixer.link_to(fmt)
|
||||||
|
fmt.link_to(sink)
|
||||||
|
graph.configure()
|
||||||
|
|
||||||
|
containers = []
|
||||||
|
try:
|
||||||
|
# Open all containers with cleanup guaranteed
|
||||||
|
for i, url in enumerate(valid_track_urls):
|
||||||
|
try:
|
||||||
|
c = av.open(url)
|
||||||
|
containers.append(c)
|
||||||
|
except Exception as e:
|
||||||
|
self.logger.warning(
|
||||||
|
"Mixdown: failed to open container from URL",
|
||||||
|
input=i,
|
||||||
|
url=url,
|
||||||
|
error=str(e),
|
||||||
|
)
|
||||||
|
|
||||||
|
if not containers:
|
||||||
|
self.logger.error("Mixdown failed - no valid containers opened")
|
||||||
|
raise Exception("Mixdown failed: Could not open any track containers")
|
||||||
|
|
||||||
|
decoders = [c.decode(audio=0) for c in containers]
|
||||||
|
active = [True] * len(decoders)
|
||||||
|
resamplers = [
|
||||||
|
AudioResampler(format="s32", layout="stereo", rate=target_sample_rate)
|
||||||
|
for _ in decoders
|
||||||
|
]
|
||||||
|
|
||||||
|
while any(active):
|
||||||
|
for i, (dec, is_active) in enumerate(zip(decoders, active)):
|
||||||
|
if not is_active:
|
||||||
|
continue
|
||||||
|
try:
|
||||||
|
frame = next(dec)
|
||||||
|
except StopIteration:
|
||||||
|
active[i] = False
|
||||||
|
continue
|
||||||
|
|
||||||
|
if frame.sample_rate != target_sample_rate:
|
||||||
|
continue
|
||||||
|
out_frames = resamplers[i].resample(frame) or []
|
||||||
|
for rf in out_frames:
|
||||||
|
rf.sample_rate = target_sample_rate
|
||||||
|
rf.time_base = Fraction(1, target_sample_rate)
|
||||||
|
inputs[i].push(rf)
|
||||||
|
|
||||||
|
while True:
|
||||||
|
try:
|
||||||
|
mixed = sink.pull()
|
||||||
|
except Exception:
|
||||||
|
break
|
||||||
|
mixed.sample_rate = target_sample_rate
|
||||||
|
mixed.time_base = Fraction(1, target_sample_rate)
|
||||||
|
await writer.push(mixed)
|
||||||
|
|
||||||
|
for in_ctx in inputs:
|
||||||
|
in_ctx.push(None)
|
||||||
|
while True:
|
||||||
|
try:
|
||||||
|
mixed = sink.pull()
|
||||||
|
except Exception:
|
||||||
|
break
|
||||||
|
mixed.sample_rate = target_sample_rate
|
||||||
|
mixed.time_base = Fraction(1, target_sample_rate)
|
||||||
|
await writer.push(mixed)
|
||||||
|
finally:
|
||||||
|
# Cleanup all containers, even if processing failed
|
||||||
|
for c in containers:
|
||||||
|
if c is not None:
|
||||||
|
try:
|
||||||
|
c.close()
|
||||||
|
except Exception:
|
||||||
|
pass # Best effort cleanup
|
||||||
|
|
||||||
|
@broadcast_to_sockets
|
||||||
|
async def set_status(self, transcript_id: str, status: TranscriptStatus):
|
||||||
|
async with self.lock_transaction():
|
||||||
|
return await transcripts_controller.set_status(transcript_id, status)
|
||||||
|
|
||||||
|
async def on_waveform(self, data):
|
||||||
|
async with self.transaction():
|
||||||
|
waveform = TranscriptWaveform(waveform=data)
|
||||||
|
transcript = await self.get_transcript()
|
||||||
|
return await transcripts_controller.append_event(
|
||||||
|
transcript=transcript, event="WAVEFORM", data=waveform
|
||||||
|
)
|
||||||
|
|
||||||
|
async def process(self, bucket_name: str, track_keys: list[str]):
|
||||||
|
transcript = await self.get_transcript()
|
||||||
|
async with self.transaction():
|
||||||
|
await transcripts_controller.update(
|
||||||
|
transcript,
|
||||||
|
{
|
||||||
|
"events": [],
|
||||||
|
"topics": [],
|
||||||
|
},
|
||||||
|
)
|
||||||
|
|
||||||
|
source_storage = get_transcripts_storage()
|
||||||
|
transcript_storage = source_storage
|
||||||
|
|
||||||
|
track_urls: list[str] = []
|
||||||
|
for key in track_keys:
|
||||||
|
url = await source_storage.get_file_url(
|
||||||
|
key,
|
||||||
|
operation="get_object",
|
||||||
|
expires_in=PRESIGNED_URL_EXPIRATION_SECONDS,
|
||||||
|
bucket=bucket_name,
|
||||||
|
)
|
||||||
|
track_urls.append(url)
|
||||||
|
self.logger.info(
|
||||||
|
f"Generated presigned URL for track from {bucket_name}",
|
||||||
|
key=key,
|
||||||
|
)
|
||||||
|
|
||||||
|
created_padded_files = set()
|
||||||
|
padded_track_urls: list[str] = []
|
||||||
|
for idx, url in enumerate(track_urls):
|
||||||
|
padded_url = await self.pad_track_for_transcription(
|
||||||
|
url, idx, transcript_storage
|
||||||
|
)
|
||||||
|
padded_track_urls.append(padded_url)
|
||||||
|
if padded_url != url:
|
||||||
|
storage_path = f"file_pipeline/{transcript.id}/tracks/padded_{idx}.webm"
|
||||||
|
created_padded_files.add(storage_path)
|
||||||
|
self.logger.info(f"Track {idx} processed, padded URL: {padded_url}")
|
||||||
|
|
||||||
|
transcript.data_path.mkdir(parents=True, exist_ok=True)
|
||||||
|
|
||||||
|
mp3_writer = AudioFileWriterProcessor(
|
||||||
|
path=str(transcript.audio_mp3_filename),
|
||||||
|
on_duration=self.on_duration,
|
||||||
|
)
|
||||||
|
await self.mixdown_tracks(padded_track_urls, mp3_writer, offsets_seconds=None)
|
||||||
|
await mp3_writer.flush()
|
||||||
|
|
||||||
|
if not transcript.audio_mp3_filename.exists():
|
||||||
|
raise Exception(
|
||||||
|
"Mixdown failed - no MP3 file generated. Cannot proceed without playable audio."
|
||||||
|
)
|
||||||
|
|
||||||
|
storage_path = f"{transcript.id}/audio.mp3"
|
||||||
|
# Use file handle streaming to avoid loading entire MP3 into memory
|
||||||
|
mp3_size = transcript.audio_mp3_filename.stat().st_size
|
||||||
|
with open(transcript.audio_mp3_filename, "rb") as mp3_file:
|
||||||
|
await transcript_storage.put_file(storage_path, mp3_file)
|
||||||
|
mp3_url = await transcript_storage.get_file_url(storage_path)
|
||||||
|
|
||||||
|
await transcripts_controller.update(transcript, {"audio_location": "storage"})
|
||||||
|
|
||||||
|
self.logger.info(
|
||||||
|
f"Uploaded mixed audio to storage",
|
||||||
|
storage_path=storage_path,
|
||||||
|
size=mp3_size,
|
||||||
|
url=mp3_url,
|
||||||
|
)
|
||||||
|
|
||||||
|
self.logger.info("Generating waveform from mixed audio")
|
||||||
|
waveform_processor = AudioWaveformProcessor(
|
||||||
|
audio_path=transcript.audio_mp3_filename,
|
||||||
|
waveform_path=transcript.audio_waveform_filename,
|
||||||
|
on_waveform=self.on_waveform,
|
||||||
|
)
|
||||||
|
waveform_processor.set_pipeline(self.empty_pipeline)
|
||||||
|
await waveform_processor.flush()
|
||||||
|
self.logger.info("Waveform generated successfully")
|
||||||
|
|
||||||
|
speaker_transcripts: list[TranscriptType] = []
|
||||||
|
for idx, padded_url in enumerate(padded_track_urls):
|
||||||
|
if not padded_url:
|
||||||
|
continue
|
||||||
|
|
||||||
|
t = await self.transcribe_file(padded_url, transcript.source_language)
|
||||||
|
|
||||||
|
if not t.words:
|
||||||
|
continue
|
||||||
|
|
||||||
|
for w in t.words:
|
||||||
|
w.speaker = idx
|
||||||
|
|
||||||
|
speaker_transcripts.append(t)
|
||||||
|
self.logger.info(
|
||||||
|
f"Track {idx} transcribed successfully with {len(t.words)} words",
|
||||||
|
track_idx=idx,
|
||||||
|
)
|
||||||
|
|
||||||
|
valid_track_count = len([url for url in padded_track_urls if url])
|
||||||
|
if valid_track_count > 0 and len(speaker_transcripts) != valid_track_count:
|
||||||
|
raise Exception(
|
||||||
|
f"Only {len(speaker_transcripts)}/{valid_track_count} tracks transcribed successfully. "
|
||||||
|
f"All tracks must succeed to avoid incomplete transcripts."
|
||||||
|
)
|
||||||
|
|
||||||
|
if not speaker_transcripts:
|
||||||
|
raise Exception("No valid track transcriptions")
|
||||||
|
|
||||||
|
self.logger.info(f"Cleaning up {len(created_padded_files)} temporary S3 files")
|
||||||
|
cleanup_tasks = []
|
||||||
|
for storage_path in created_padded_files:
|
||||||
|
cleanup_tasks.append(transcript_storage.delete_file(storage_path))
|
||||||
|
|
||||||
|
if cleanup_tasks:
|
||||||
|
cleanup_results = await asyncio.gather(
|
||||||
|
*cleanup_tasks, return_exceptions=True
|
||||||
|
)
|
||||||
|
for storage_path, result in zip(created_padded_files, cleanup_results):
|
||||||
|
if isinstance(result, Exception):
|
||||||
|
self.logger.warning(
|
||||||
|
"Failed to cleanup temporary padded track",
|
||||||
|
storage_path=storage_path,
|
||||||
|
error=str(result),
|
||||||
|
)
|
||||||
|
|
||||||
|
merged_words = []
|
||||||
|
for t in speaker_transcripts:
|
||||||
|
merged_words.extend(t.words)
|
||||||
|
merged_words.sort(
|
||||||
|
key=lambda w: w.start if hasattr(w, "start") and w.start is not None else 0
|
||||||
|
)
|
||||||
|
|
||||||
|
merged_transcript = TranscriptType(words=merged_words, translation=None)
|
||||||
|
|
||||||
|
await self.on_transcript(merged_transcript)
|
||||||
|
|
||||||
|
topics = await self.detect_topics(merged_transcript, transcript.target_language)
|
||||||
|
await asyncio.gather(
|
||||||
|
self.generate_title(topics),
|
||||||
|
self.generate_summaries(topics),
|
||||||
|
return_exceptions=False,
|
||||||
|
)
|
||||||
|
|
||||||
|
await self.set_status(transcript.id, "ended")
|
||||||
|
|
||||||
|
async def transcribe_file(self, audio_url: str, language: str) -> TranscriptType:
|
||||||
|
return await transcribe_file_with_processor(audio_url, language)
|
||||||
|
|
||||||
|
async def detect_topics(
|
||||||
|
self, transcript: TranscriptType, target_language: str
|
||||||
|
) -> list[TitleSummary]:
|
||||||
|
return await topic_processing.detect_topics(
|
||||||
|
transcript,
|
||||||
|
target_language,
|
||||||
|
on_topic_callback=self.on_topic,
|
||||||
|
empty_pipeline=self.empty_pipeline,
|
||||||
|
)
|
||||||
|
|
||||||
|
async def generate_title(self, topics: list[TitleSummary]):
|
||||||
|
return await topic_processing.generate_title(
|
||||||
|
topics,
|
||||||
|
on_title_callback=self.on_title,
|
||||||
|
empty_pipeline=self.empty_pipeline,
|
||||||
|
logger=self.logger,
|
||||||
|
)
|
||||||
|
|
||||||
|
async def generate_summaries(self, topics: list[TitleSummary]):
|
||||||
|
transcript = await self.get_transcript()
|
||||||
|
return await topic_processing.generate_summaries(
|
||||||
|
topics,
|
||||||
|
transcript,
|
||||||
|
on_long_summary_callback=self.on_long_summary,
|
||||||
|
on_short_summary_callback=self.on_short_summary,
|
||||||
|
empty_pipeline=self.empty_pipeline,
|
||||||
|
logger=self.logger,
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
@shared_task
|
||||||
|
@asynctask
|
||||||
|
async def task_pipeline_multitrack_process(
|
||||||
|
*, transcript_id: str, bucket_name: str, track_keys: list[str]
|
||||||
|
):
|
||||||
|
pipeline = PipelineMainMultitrack(transcript_id=transcript_id)
|
||||||
|
try:
|
||||||
|
await pipeline.set_status(transcript_id, "processing")
|
||||||
|
await pipeline.process(bucket_name, track_keys)
|
||||||
|
except Exception:
|
||||||
|
await pipeline.set_status(transcript_id, "error")
|
||||||
|
raise
|
||||||
|
|
||||||
|
post_chain = chain(
|
||||||
|
task_cleanup_consent.si(transcript_id=transcript_id),
|
||||||
|
task_pipeline_post_to_zulip.si(transcript_id=transcript_id),
|
||||||
|
task_send_webhook_if_needed.si(transcript_id=transcript_id),
|
||||||
|
)
|
||||||
|
post_chain.delay()
|
||||||
@@ -18,22 +18,14 @@ During its lifecycle, it will emit the following status:
|
|||||||
import asyncio
|
import asyncio
|
||||||
from typing import Generic, TypeVar
|
from typing import Generic, TypeVar
|
||||||
|
|
||||||
from pydantic import BaseModel, ConfigDict
|
|
||||||
|
|
||||||
from reflector.logger import logger
|
from reflector.logger import logger
|
||||||
from reflector.processors import Pipeline
|
from reflector.processors import Pipeline
|
||||||
|
|
||||||
PipelineMessage = TypeVar("PipelineMessage")
|
PipelineMessage = TypeVar("PipelineMessage")
|
||||||
|
|
||||||
|
|
||||||
class PipelineRunner(BaseModel, Generic[PipelineMessage]):
|
class PipelineRunner(Generic[PipelineMessage]):
|
||||||
model_config = ConfigDict(arbitrary_types_allowed=True)
|
def __init__(self):
|
||||||
|
|
||||||
status: str = "idle"
|
|
||||||
pipeline: Pipeline | None = None
|
|
||||||
|
|
||||||
def __init__(self, **kwargs):
|
|
||||||
super().__init__(**kwargs)
|
|
||||||
self._task = None
|
self._task = None
|
||||||
self._q_cmd = asyncio.Queue(maxsize=4096)
|
self._q_cmd = asyncio.Queue(maxsize=4096)
|
||||||
self._ev_done = asyncio.Event()
|
self._ev_done = asyncio.Event()
|
||||||
@@ -42,6 +34,8 @@ class PipelineRunner(BaseModel, Generic[PipelineMessage]):
|
|||||||
runner=id(self),
|
runner=id(self),
|
||||||
runner_cls=self.__class__.__name__,
|
runner_cls=self.__class__.__name__,
|
||||||
)
|
)
|
||||||
|
self.status = "idle"
|
||||||
|
self.pipeline: Pipeline | None = None
|
||||||
|
|
||||||
async def create(self) -> Pipeline:
|
async def create(self) -> Pipeline:
|
||||||
"""
|
"""
|
||||||
|
|||||||
109
server/reflector/pipelines/topic_processing.py
Normal file
109
server/reflector/pipelines/topic_processing.py
Normal file
@@ -0,0 +1,109 @@
|
|||||||
|
"""
|
||||||
|
Topic processing utilities
|
||||||
|
==========================
|
||||||
|
|
||||||
|
Shared topic detection, title generation, and summarization logic
|
||||||
|
used across file and multitrack pipelines.
|
||||||
|
"""
|
||||||
|
|
||||||
|
from typing import Callable
|
||||||
|
|
||||||
|
import structlog
|
||||||
|
|
||||||
|
from reflector.db.transcripts import Transcript
|
||||||
|
from reflector.processors import (
|
||||||
|
TranscriptFinalSummaryProcessor,
|
||||||
|
TranscriptFinalTitleProcessor,
|
||||||
|
TranscriptTopicDetectorProcessor,
|
||||||
|
)
|
||||||
|
from reflector.processors.types import TitleSummary
|
||||||
|
from reflector.processors.types import Transcript as TranscriptType
|
||||||
|
|
||||||
|
|
||||||
|
class EmptyPipeline:
|
||||||
|
def __init__(self, logger: structlog.BoundLogger):
|
||||||
|
self.logger = logger
|
||||||
|
|
||||||
|
def get_pref(self, k, d=None):
|
||||||
|
return d
|
||||||
|
|
||||||
|
async def emit(self, event):
|
||||||
|
pass
|
||||||
|
|
||||||
|
|
||||||
|
async def detect_topics(
|
||||||
|
transcript: TranscriptType,
|
||||||
|
target_language: str,
|
||||||
|
*,
|
||||||
|
on_topic_callback: Callable,
|
||||||
|
empty_pipeline: EmptyPipeline,
|
||||||
|
) -> list[TitleSummary]:
|
||||||
|
chunk_size = 300
|
||||||
|
topics: list[TitleSummary] = []
|
||||||
|
|
||||||
|
async def on_topic(topic: TitleSummary):
|
||||||
|
topics.append(topic)
|
||||||
|
return await on_topic_callback(topic)
|
||||||
|
|
||||||
|
topic_detector = TranscriptTopicDetectorProcessor(callback=on_topic)
|
||||||
|
topic_detector.set_pipeline(empty_pipeline)
|
||||||
|
|
||||||
|
for i in range(0, len(transcript.words), chunk_size):
|
||||||
|
chunk_words = transcript.words[i : i + chunk_size]
|
||||||
|
if not chunk_words:
|
||||||
|
continue
|
||||||
|
|
||||||
|
chunk_transcript = TranscriptType(
|
||||||
|
words=chunk_words, translation=transcript.translation
|
||||||
|
)
|
||||||
|
|
||||||
|
await topic_detector.push(chunk_transcript)
|
||||||
|
|
||||||
|
await topic_detector.flush()
|
||||||
|
return topics
|
||||||
|
|
||||||
|
|
||||||
|
async def generate_title(
|
||||||
|
topics: list[TitleSummary],
|
||||||
|
*,
|
||||||
|
on_title_callback: Callable,
|
||||||
|
empty_pipeline: EmptyPipeline,
|
||||||
|
logger: structlog.BoundLogger,
|
||||||
|
):
|
||||||
|
if not topics:
|
||||||
|
logger.warning("No topics for title generation")
|
||||||
|
return
|
||||||
|
|
||||||
|
processor = TranscriptFinalTitleProcessor(callback=on_title_callback)
|
||||||
|
processor.set_pipeline(empty_pipeline)
|
||||||
|
|
||||||
|
for topic in topics:
|
||||||
|
await processor.push(topic)
|
||||||
|
|
||||||
|
await processor.flush()
|
||||||
|
|
||||||
|
|
||||||
|
async def generate_summaries(
|
||||||
|
topics: list[TitleSummary],
|
||||||
|
transcript: Transcript,
|
||||||
|
*,
|
||||||
|
on_long_summary_callback: Callable,
|
||||||
|
on_short_summary_callback: Callable,
|
||||||
|
empty_pipeline: EmptyPipeline,
|
||||||
|
logger: structlog.BoundLogger,
|
||||||
|
):
|
||||||
|
if not topics:
|
||||||
|
logger.warning("No topics for summary generation")
|
||||||
|
return
|
||||||
|
|
||||||
|
processor = TranscriptFinalSummaryProcessor(
|
||||||
|
transcript=transcript,
|
||||||
|
callback=on_long_summary_callback,
|
||||||
|
on_short_summary=on_short_summary_callback,
|
||||||
|
)
|
||||||
|
processor.set_pipeline(empty_pipeline)
|
||||||
|
|
||||||
|
for topic in topics:
|
||||||
|
await processor.push(topic)
|
||||||
|
|
||||||
|
await processor.flush()
|
||||||
34
server/reflector/pipelines/transcription_helpers.py
Normal file
34
server/reflector/pipelines/transcription_helpers.py
Normal file
@@ -0,0 +1,34 @@
|
|||||||
|
from reflector.processors.file_transcript import FileTranscriptInput
|
||||||
|
from reflector.processors.file_transcript_auto import FileTranscriptAutoProcessor
|
||||||
|
from reflector.processors.types import Transcript as TranscriptType
|
||||||
|
|
||||||
|
|
||||||
|
async def transcribe_file_with_processor(
|
||||||
|
audio_url: str,
|
||||||
|
language: str,
|
||||||
|
processor_name: str | None = None,
|
||||||
|
) -> TranscriptType:
|
||||||
|
processor = (
|
||||||
|
FileTranscriptAutoProcessor(name=processor_name)
|
||||||
|
if processor_name
|
||||||
|
else FileTranscriptAutoProcessor()
|
||||||
|
)
|
||||||
|
input_data = FileTranscriptInput(audio_url=audio_url, language=language)
|
||||||
|
|
||||||
|
result: TranscriptType | None = None
|
||||||
|
|
||||||
|
async def capture_result(transcript):
|
||||||
|
nonlocal result
|
||||||
|
result = transcript
|
||||||
|
|
||||||
|
processor.on(capture_result)
|
||||||
|
await processor.push(input_data)
|
||||||
|
await processor.flush()
|
||||||
|
|
||||||
|
if not result:
|
||||||
|
processor_label = processor_name or "default"
|
||||||
|
raise ValueError(
|
||||||
|
f"No transcript captured from {processor_label} processor for audio: {audio_url}"
|
||||||
|
)
|
||||||
|
|
||||||
|
return result
|
||||||
@@ -1,5 +1,7 @@
|
|||||||
from .audio_chunker import AudioChunkerProcessor # noqa: F401
|
from .audio_chunker import AudioChunkerProcessor # noqa: F401
|
||||||
|
from .audio_chunker_auto import AudioChunkerAutoProcessor # noqa: F401
|
||||||
from .audio_diarization_auto import AudioDiarizationAutoProcessor # noqa: F401
|
from .audio_diarization_auto import AudioDiarizationAutoProcessor # noqa: F401
|
||||||
|
from .audio_downscale import AudioDownscaleProcessor # noqa: F401
|
||||||
from .audio_file_writer import AudioFileWriterProcessor # noqa: F401
|
from .audio_file_writer import AudioFileWriterProcessor # noqa: F401
|
||||||
from .audio_merge import AudioMergeProcessor # noqa: F401
|
from .audio_merge import AudioMergeProcessor # noqa: F401
|
||||||
from .audio_transcript import AudioTranscriptProcessor # noqa: F401
|
from .audio_transcript import AudioTranscriptProcessor # noqa: F401
|
||||||
@@ -11,6 +13,13 @@ from .base import ( # noqa: F401
|
|||||||
Processor,
|
Processor,
|
||||||
ThreadedProcessor,
|
ThreadedProcessor,
|
||||||
)
|
)
|
||||||
|
from .file_diarization import FileDiarizationProcessor # noqa: F401
|
||||||
|
from .file_diarization_auto import FileDiarizationAutoProcessor # noqa: F401
|
||||||
|
from .file_transcript import FileTranscriptProcessor # noqa: F401
|
||||||
|
from .file_transcript_auto import FileTranscriptAutoProcessor # noqa: F401
|
||||||
|
from .transcript_diarization_assembler import (
|
||||||
|
TranscriptDiarizationAssemblerProcessor, # noqa: F401
|
||||||
|
)
|
||||||
from .transcript_final_summary import TranscriptFinalSummaryProcessor # noqa: F401
|
from .transcript_final_summary import TranscriptFinalSummaryProcessor # noqa: F401
|
||||||
from .transcript_final_title import TranscriptFinalTitleProcessor # noqa: F401
|
from .transcript_final_title import TranscriptFinalTitleProcessor # noqa: F401
|
||||||
from .transcript_liner import TranscriptLinerProcessor # noqa: F401
|
from .transcript_liner import TranscriptLinerProcessor # noqa: F401
|
||||||
|
|||||||
@@ -1,28 +1,78 @@
|
|||||||
|
from typing import Optional
|
||||||
|
|
||||||
import av
|
import av
|
||||||
|
from prometheus_client import Counter, Histogram
|
||||||
|
|
||||||
from reflector.processors.base import Processor
|
from reflector.processors.base import Processor
|
||||||
|
|
||||||
|
|
||||||
class AudioChunkerProcessor(Processor):
|
class AudioChunkerProcessor(Processor):
|
||||||
"""
|
"""
|
||||||
Assemble audio frames into chunks
|
Base class for assembling audio frames into chunks
|
||||||
"""
|
"""
|
||||||
|
|
||||||
INPUT_TYPE = av.AudioFrame
|
INPUT_TYPE = av.AudioFrame
|
||||||
OUTPUT_TYPE = list[av.AudioFrame]
|
OUTPUT_TYPE = list[av.AudioFrame]
|
||||||
|
|
||||||
def __init__(self, max_frames=256):
|
m_chunk = Histogram(
|
||||||
super().__init__()
|
"audio_chunker",
|
||||||
|
"Time spent in AudioChunker.chunk",
|
||||||
|
["backend"],
|
||||||
|
)
|
||||||
|
m_chunk_call = Counter(
|
||||||
|
"audio_chunker_call",
|
||||||
|
"Number of calls to AudioChunker.chunk",
|
||||||
|
["backend"],
|
||||||
|
)
|
||||||
|
m_chunk_success = Counter(
|
||||||
|
"audio_chunker_success",
|
||||||
|
"Number of successful calls to AudioChunker.chunk",
|
||||||
|
["backend"],
|
||||||
|
)
|
||||||
|
m_chunk_failure = Counter(
|
||||||
|
"audio_chunker_failure",
|
||||||
|
"Number of failed calls to AudioChunker.chunk",
|
||||||
|
["backend"],
|
||||||
|
)
|
||||||
|
|
||||||
|
def __init__(self, *args, **kwargs):
|
||||||
|
name = self.__class__.__name__
|
||||||
|
self.m_chunk = self.m_chunk.labels(name)
|
||||||
|
self.m_chunk_call = self.m_chunk_call.labels(name)
|
||||||
|
self.m_chunk_success = self.m_chunk_success.labels(name)
|
||||||
|
self.m_chunk_failure = self.m_chunk_failure.labels(name)
|
||||||
|
super().__init__(*args, **kwargs)
|
||||||
self.frames: list[av.AudioFrame] = []
|
self.frames: list[av.AudioFrame] = []
|
||||||
self.max_frames = max_frames
|
|
||||||
|
|
||||||
async def _push(self, data: av.AudioFrame):
|
async def _push(self, data: av.AudioFrame):
|
||||||
self.frames.append(data)
|
"""Process incoming audio frame"""
|
||||||
if len(self.frames) >= self.max_frames:
|
# Validate audio format on first frame
|
||||||
await self.flush()
|
if len(self.frames) == 0:
|
||||||
|
if data.sample_rate != 16000 or len(data.layout.channels) != 1:
|
||||||
|
raise ValueError(
|
||||||
|
f"AudioChunkerProcessor expects 16kHz mono audio, got {data.sample_rate}Hz "
|
||||||
|
f"with {len(data.layout.channels)} channel(s). "
|
||||||
|
f"Use AudioDownscaleProcessor before this processor."
|
||||||
|
)
|
||||||
|
|
||||||
|
try:
|
||||||
|
self.m_chunk_call.inc()
|
||||||
|
with self.m_chunk.time():
|
||||||
|
result = await self._chunk(data)
|
||||||
|
self.m_chunk_success.inc()
|
||||||
|
if result:
|
||||||
|
await self.emit(result)
|
||||||
|
except Exception:
|
||||||
|
self.m_chunk_failure.inc()
|
||||||
|
raise
|
||||||
|
|
||||||
|
async def _chunk(self, data: av.AudioFrame) -> Optional[list[av.AudioFrame]]:
|
||||||
|
"""
|
||||||
|
Process audio frame and return chunk when ready.
|
||||||
|
Subclasses should implement their chunking logic here.
|
||||||
|
"""
|
||||||
|
raise NotImplementedError
|
||||||
|
|
||||||
async def _flush(self):
|
async def _flush(self):
|
||||||
frames = self.frames[:]
|
"""Flush any remaining frames when processing ends"""
|
||||||
self.frames = []
|
raise NotImplementedError
|
||||||
if frames:
|
|
||||||
await self.emit(frames)
|
|
||||||
|
|||||||
32
server/reflector/processors/audio_chunker_auto.py
Normal file
32
server/reflector/processors/audio_chunker_auto.py
Normal file
@@ -0,0 +1,32 @@
|
|||||||
|
import importlib
|
||||||
|
|
||||||
|
from reflector.processors.audio_chunker import AudioChunkerProcessor
|
||||||
|
from reflector.settings import settings
|
||||||
|
|
||||||
|
|
||||||
|
class AudioChunkerAutoProcessor(AudioChunkerProcessor):
|
||||||
|
_registry = {}
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def register(cls, name, kclass):
|
||||||
|
cls._registry[name] = kclass
|
||||||
|
|
||||||
|
def __new__(cls, name: str | None = None, **kwargs):
|
||||||
|
if name is None:
|
||||||
|
name = settings.AUDIO_CHUNKER_BACKEND
|
||||||
|
if name not in cls._registry:
|
||||||
|
module_name = f"reflector.processors.audio_chunker_{name}"
|
||||||
|
importlib.import_module(module_name)
|
||||||
|
|
||||||
|
# gather specific configuration for the processor
|
||||||
|
# search `AUDIO_CHUNKER_BACKEND_XXX_YYY`, push to constructor as `backend_xxx_yyy`
|
||||||
|
config = {}
|
||||||
|
name_upper = name.upper()
|
||||||
|
settings_prefix = "AUDIO_CHUNKER_"
|
||||||
|
config_prefix = f"{settings_prefix}{name_upper}_"
|
||||||
|
for key, value in settings:
|
||||||
|
if key.startswith(config_prefix):
|
||||||
|
config_name = key[len(settings_prefix) :].lower()
|
||||||
|
config[config_name] = value
|
||||||
|
|
||||||
|
return cls._registry[name](**config | kwargs)
|
||||||
34
server/reflector/processors/audio_chunker_frames.py
Normal file
34
server/reflector/processors/audio_chunker_frames.py
Normal file
@@ -0,0 +1,34 @@
|
|||||||
|
from typing import Optional
|
||||||
|
|
||||||
|
import av
|
||||||
|
|
||||||
|
from reflector.processors.audio_chunker import AudioChunkerProcessor
|
||||||
|
from reflector.processors.audio_chunker_auto import AudioChunkerAutoProcessor
|
||||||
|
|
||||||
|
|
||||||
|
class AudioChunkerFramesProcessor(AudioChunkerProcessor):
|
||||||
|
"""
|
||||||
|
Simple frame-based audio chunker that emits chunks after a fixed number of frames
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(self, max_frames=256, **kwargs):
|
||||||
|
super().__init__(**kwargs)
|
||||||
|
self.max_frames = max_frames
|
||||||
|
|
||||||
|
async def _chunk(self, data: av.AudioFrame) -> Optional[list[av.AudioFrame]]:
|
||||||
|
self.frames.append(data)
|
||||||
|
if len(self.frames) >= self.max_frames:
|
||||||
|
frames_to_emit = self.frames[:]
|
||||||
|
self.frames = []
|
||||||
|
return frames_to_emit
|
||||||
|
|
||||||
|
return None
|
||||||
|
|
||||||
|
async def _flush(self):
|
||||||
|
frames = self.frames[:]
|
||||||
|
self.frames = []
|
||||||
|
if frames:
|
||||||
|
await self.emit(frames)
|
||||||
|
|
||||||
|
|
||||||
|
AudioChunkerAutoProcessor.register("frames", AudioChunkerFramesProcessor)
|
||||||
298
server/reflector/processors/audio_chunker_silero.py
Normal file
298
server/reflector/processors/audio_chunker_silero.py
Normal file
@@ -0,0 +1,298 @@
|
|||||||
|
from typing import Optional
|
||||||
|
|
||||||
|
import av
|
||||||
|
import numpy as np
|
||||||
|
import torch
|
||||||
|
from silero_vad import VADIterator, load_silero_vad
|
||||||
|
|
||||||
|
from reflector.processors.audio_chunker import AudioChunkerProcessor
|
||||||
|
from reflector.processors.audio_chunker_auto import AudioChunkerAutoProcessor
|
||||||
|
|
||||||
|
|
||||||
|
class AudioChunkerSileroProcessor(AudioChunkerProcessor):
|
||||||
|
"""
|
||||||
|
Assemble audio frames into chunks with VAD-based speech detection using Silero VAD
|
||||||
|
"""
|
||||||
|
|
||||||
|
def __init__(
|
||||||
|
self,
|
||||||
|
block_frames=256,
|
||||||
|
max_frames=1024,
|
||||||
|
use_onnx=True,
|
||||||
|
min_frames=2,
|
||||||
|
**kwargs,
|
||||||
|
):
|
||||||
|
super().__init__(**kwargs)
|
||||||
|
self.block_frames = block_frames
|
||||||
|
self.max_frames = max_frames
|
||||||
|
self.min_frames = min_frames
|
||||||
|
|
||||||
|
# Initialize Silero VAD
|
||||||
|
self._init_vad(use_onnx)
|
||||||
|
|
||||||
|
def _init_vad(self, use_onnx=False):
|
||||||
|
"""Initialize Silero VAD model"""
|
||||||
|
try:
|
||||||
|
torch.set_num_threads(1)
|
||||||
|
self.vad_model = load_silero_vad(onnx=use_onnx)
|
||||||
|
self.vad_iterator = VADIterator(self.vad_model, sampling_rate=16000)
|
||||||
|
self.logger.info("Silero VAD initialized successfully")
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
self.logger.error(f"Failed to initialize Silero VAD: {e}")
|
||||||
|
self.vad_model = None
|
||||||
|
self.vad_iterator = None
|
||||||
|
|
||||||
|
async def _chunk(self, data: av.AudioFrame) -> Optional[list[av.AudioFrame]]:
|
||||||
|
"""Process audio frame and return chunk when ready"""
|
||||||
|
self.frames.append(data)
|
||||||
|
|
||||||
|
# Check for speech segments every 32 frames (~1 second)
|
||||||
|
if len(self.frames) >= 32 and len(self.frames) % 32 == 0:
|
||||||
|
return await self._process_block()
|
||||||
|
|
||||||
|
# Safety fallback - emit if we hit max frames
|
||||||
|
elif len(self.frames) >= self.max_frames:
|
||||||
|
self.logger.warning(
|
||||||
|
f"AudioChunkerSileroProcessor: Reached max frames ({self.max_frames}), "
|
||||||
|
f"emitting first {self.max_frames // 2} frames"
|
||||||
|
)
|
||||||
|
frames_to_emit = self.frames[: self.max_frames // 2]
|
||||||
|
self.frames = self.frames[self.max_frames // 2 :]
|
||||||
|
if len(frames_to_emit) >= self.min_frames:
|
||||||
|
return frames_to_emit
|
||||||
|
else:
|
||||||
|
self.logger.debug(
|
||||||
|
f"Ignoring fallback segment with {len(frames_to_emit)} frames "
|
||||||
|
f"(< {self.min_frames} minimum)"
|
||||||
|
)
|
||||||
|
|
||||||
|
return None
|
||||||
|
|
||||||
|
async def _process_block(self) -> Optional[list[av.AudioFrame]]:
|
||||||
|
# Need at least 32 frames for VAD detection (~1 second)
|
||||||
|
if len(self.frames) < 32 or self.vad_iterator is None:
|
||||||
|
return None
|
||||||
|
|
||||||
|
# Processing block with current buffer size
|
||||||
|
print(f"Processing block: {len(self.frames)} frames in buffer")
|
||||||
|
|
||||||
|
try:
|
||||||
|
# Convert frames to numpy array for VAD
|
||||||
|
audio_array = self._frames_to_numpy(self.frames)
|
||||||
|
|
||||||
|
if audio_array is None:
|
||||||
|
# Fallback: emit all frames if conversion failed
|
||||||
|
frames_to_emit = self.frames[:]
|
||||||
|
self.frames = []
|
||||||
|
if len(frames_to_emit) >= self.min_frames:
|
||||||
|
return frames_to_emit
|
||||||
|
else:
|
||||||
|
self.logger.debug(
|
||||||
|
f"Ignoring conversion-failed segment with {len(frames_to_emit)} frames "
|
||||||
|
f"(< {self.min_frames} minimum)"
|
||||||
|
)
|
||||||
|
return None
|
||||||
|
|
||||||
|
# Find complete speech segments in the buffer
|
||||||
|
speech_end_frame = self._find_speech_segment_end(audio_array)
|
||||||
|
|
||||||
|
if speech_end_frame is None or speech_end_frame <= 0:
|
||||||
|
# No speech found but buffer is getting large
|
||||||
|
if len(self.frames) > 512:
|
||||||
|
# Check if it's all silence and can be discarded
|
||||||
|
# No speech segment found, buffer at {len(self.frames)} frames
|
||||||
|
|
||||||
|
# Could emit silence or discard old frames here
|
||||||
|
# For now, keep first 256 frames and discard older silence
|
||||||
|
if len(self.frames) > 768:
|
||||||
|
self.logger.debug(
|
||||||
|
f"Discarding {len(self.frames) - 256} old frames (likely silence)"
|
||||||
|
)
|
||||||
|
self.frames = self.frames[-256:]
|
||||||
|
return None
|
||||||
|
|
||||||
|
# Calculate segment timing information
|
||||||
|
frames_to_emit = self.frames[:speech_end_frame]
|
||||||
|
|
||||||
|
# Get timing from av.AudioFrame
|
||||||
|
if frames_to_emit:
|
||||||
|
first_frame = frames_to_emit[0]
|
||||||
|
last_frame = frames_to_emit[-1]
|
||||||
|
sample_rate = first_frame.sample_rate
|
||||||
|
|
||||||
|
# Calculate duration
|
||||||
|
total_samples = sum(f.samples for f in frames_to_emit)
|
||||||
|
duration_seconds = total_samples / sample_rate if sample_rate > 0 else 0
|
||||||
|
|
||||||
|
# Get timestamps if available
|
||||||
|
start_time = (
|
||||||
|
first_frame.pts * first_frame.time_base if first_frame.pts else 0
|
||||||
|
)
|
||||||
|
end_time = (
|
||||||
|
last_frame.pts * last_frame.time_base if last_frame.pts else 0
|
||||||
|
)
|
||||||
|
|
||||||
|
# Convert to HH:MM:SS format for logging
|
||||||
|
def format_time(seconds):
|
||||||
|
if not seconds:
|
||||||
|
return "00:00:00"
|
||||||
|
total_seconds = int(float(seconds))
|
||||||
|
hours = total_seconds // 3600
|
||||||
|
minutes = (total_seconds % 3600) // 60
|
||||||
|
secs = total_seconds % 60
|
||||||
|
return f"{hours:02d}:{minutes:02d}:{secs:02d}"
|
||||||
|
|
||||||
|
start_formatted = format_time(start_time)
|
||||||
|
end_formatted = format_time(end_time)
|
||||||
|
|
||||||
|
# Keep remaining frames for next processing
|
||||||
|
remaining_after = len(self.frames) - speech_end_frame
|
||||||
|
|
||||||
|
# Single structured log line
|
||||||
|
self.logger.info(
|
||||||
|
"Speech segment found",
|
||||||
|
start=start_formatted,
|
||||||
|
end=end_formatted,
|
||||||
|
frames=speech_end_frame,
|
||||||
|
duration=round(duration_seconds, 2),
|
||||||
|
buffer_before=len(self.frames),
|
||||||
|
remaining=remaining_after,
|
||||||
|
)
|
||||||
|
|
||||||
|
# Keep remaining frames for next processing
|
||||||
|
self.frames = self.frames[speech_end_frame:]
|
||||||
|
|
||||||
|
# Filter out segments with too few frames
|
||||||
|
if len(frames_to_emit) >= self.min_frames:
|
||||||
|
return frames_to_emit
|
||||||
|
else:
|
||||||
|
self.logger.debug(
|
||||||
|
f"Ignoring segment with {len(frames_to_emit)} frames "
|
||||||
|
f"(< {self.min_frames} minimum)"
|
||||||
|
)
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
self.logger.error(f"Error in VAD processing: {e}")
|
||||||
|
# Fallback to simple chunking
|
||||||
|
if len(self.frames) >= self.block_frames:
|
||||||
|
frames_to_emit = self.frames[: self.block_frames]
|
||||||
|
self.frames = self.frames[self.block_frames :]
|
||||||
|
if len(frames_to_emit) >= self.min_frames:
|
||||||
|
return frames_to_emit
|
||||||
|
else:
|
||||||
|
self.logger.debug(
|
||||||
|
f"Ignoring exception-fallback segment with {len(frames_to_emit)} frames "
|
||||||
|
f"(< {self.min_frames} minimum)"
|
||||||
|
)
|
||||||
|
|
||||||
|
return None
|
||||||
|
|
||||||
|
def _frames_to_numpy(self, frames: list[av.AudioFrame]) -> Optional[np.ndarray]:
|
||||||
|
"""Convert av.AudioFrame list to numpy array for VAD processing"""
|
||||||
|
if not frames:
|
||||||
|
return None
|
||||||
|
|
||||||
|
try:
|
||||||
|
audio_data = []
|
||||||
|
for frame in frames:
|
||||||
|
frame_array = frame.to_ndarray()
|
||||||
|
|
||||||
|
if len(frame_array.shape) == 2:
|
||||||
|
frame_array = frame_array.flatten()
|
||||||
|
|
||||||
|
audio_data.append(frame_array)
|
||||||
|
|
||||||
|
if not audio_data:
|
||||||
|
return None
|
||||||
|
|
||||||
|
combined_audio = np.concatenate(audio_data)
|
||||||
|
|
||||||
|
# Ensure float32 format
|
||||||
|
if combined_audio.dtype == np.int16:
|
||||||
|
# Normalize int16 audio to float32 in range [-1.0, 1.0]
|
||||||
|
combined_audio = combined_audio.astype(np.float32) / 32768.0
|
||||||
|
elif combined_audio.dtype != np.float32:
|
||||||
|
combined_audio = combined_audio.astype(np.float32)
|
||||||
|
|
||||||
|
return combined_audio
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
self.logger.error(f"Error converting frames to numpy: {e}")
|
||||||
|
|
||||||
|
return None
|
||||||
|
|
||||||
|
def _find_speech_segment_end(self, audio_array: np.ndarray) -> Optional[int]:
|
||||||
|
"""Find complete speech segments and return frame index at segment end"""
|
||||||
|
if self.vad_iterator is None or len(audio_array) == 0:
|
||||||
|
return None
|
||||||
|
|
||||||
|
try:
|
||||||
|
# Process audio in 512-sample windows for VAD
|
||||||
|
window_size = 512
|
||||||
|
min_silence_windows = 3 # Require 3 windows of silence after speech
|
||||||
|
|
||||||
|
# Track speech state
|
||||||
|
in_speech = False
|
||||||
|
speech_start = None
|
||||||
|
speech_end = None
|
||||||
|
silence_count = 0
|
||||||
|
|
||||||
|
for i in range(0, len(audio_array), window_size):
|
||||||
|
chunk = audio_array[i : i + window_size]
|
||||||
|
if len(chunk) < window_size:
|
||||||
|
chunk = np.pad(chunk, (0, window_size - len(chunk)))
|
||||||
|
|
||||||
|
# Detect if this window has speech
|
||||||
|
speech_dict = self.vad_iterator(chunk, return_seconds=True)
|
||||||
|
|
||||||
|
# VADIterator returns dict with 'start' and 'end' when speech segments are detected
|
||||||
|
if speech_dict:
|
||||||
|
if not in_speech:
|
||||||
|
# Speech started
|
||||||
|
speech_start = i
|
||||||
|
in_speech = True
|
||||||
|
# Debug: print(f"Speech START at sample {i}, VAD: {speech_dict}")
|
||||||
|
silence_count = 0 # Reset silence counter
|
||||||
|
continue
|
||||||
|
|
||||||
|
if not in_speech:
|
||||||
|
continue
|
||||||
|
|
||||||
|
# We're in speech but found silence
|
||||||
|
silence_count += 1
|
||||||
|
if silence_count < min_silence_windows:
|
||||||
|
continue
|
||||||
|
|
||||||
|
# Found end of speech segment
|
||||||
|
speech_end = i - (min_silence_windows - 1) * window_size
|
||||||
|
# Debug: print(f"Speech END at sample {speech_end}")
|
||||||
|
|
||||||
|
# Convert sample position to frame index
|
||||||
|
samples_per_frame = self.frames[0].samples if self.frames else 1024
|
||||||
|
frame_index = speech_end // samples_per_frame
|
||||||
|
|
||||||
|
# Ensure we don't exceed buffer
|
||||||
|
frame_index = min(frame_index, len(self.frames))
|
||||||
|
return frame_index
|
||||||
|
|
||||||
|
return None
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
self.logger.error(f"Error finding speech segment: {e}")
|
||||||
|
return None
|
||||||
|
|
||||||
|
async def _flush(self):
|
||||||
|
frames = self.frames[:]
|
||||||
|
self.frames = []
|
||||||
|
if frames:
|
||||||
|
if len(frames) >= self.min_frames:
|
||||||
|
await self.emit(frames)
|
||||||
|
else:
|
||||||
|
self.logger.debug(
|
||||||
|
f"Ignoring flush segment with {len(frames)} frames "
|
||||||
|
f"(< {self.min_frames} minimum)"
|
||||||
|
)
|
||||||
|
|
||||||
|
|
||||||
|
AudioChunkerAutoProcessor.register("silero", AudioChunkerSileroProcessor)
|
||||||
@@ -1,6 +1,7 @@
|
|||||||
from reflector.processors.base import Processor
|
from reflector.processors.base import Processor
|
||||||
from reflector.processors.types import (
|
from reflector.processors.types import (
|
||||||
AudioDiarizationInput,
|
AudioDiarizationInput,
|
||||||
|
DiarizationSegment,
|
||||||
TitleSummary,
|
TitleSummary,
|
||||||
Word,
|
Word,
|
||||||
)
|
)
|
||||||
@@ -37,18 +38,21 @@ class AudioDiarizationProcessor(Processor):
|
|||||||
async def _diarize(self, data: AudioDiarizationInput):
|
async def _diarize(self, data: AudioDiarizationInput):
|
||||||
raise NotImplementedError
|
raise NotImplementedError
|
||||||
|
|
||||||
def assign_speaker(self, words: list[Word], diarization: list[dict]):
|
@classmethod
|
||||||
self._diarization_remove_overlap(diarization)
|
def assign_speaker(cls, words: list[Word], diarization: list[DiarizationSegment]):
|
||||||
self._diarization_remove_segment_without_words(words, diarization)
|
cls._diarization_remove_overlap(diarization)
|
||||||
self._diarization_merge_same_speaker(words, diarization)
|
cls._diarization_remove_segment_without_words(words, diarization)
|
||||||
self._diarization_assign_speaker(words, diarization)
|
cls._diarization_merge_same_speaker(diarization)
|
||||||
|
cls._diarization_assign_speaker(words, diarization)
|
||||||
|
|
||||||
def iter_words_from_topics(self, topics: TitleSummary):
|
@staticmethod
|
||||||
|
def iter_words_from_topics(topics: list[TitleSummary]):
|
||||||
for topic in topics:
|
for topic in topics:
|
||||||
for word in topic.transcript.words:
|
for word in topic.transcript.words:
|
||||||
yield word
|
yield word
|
||||||
|
|
||||||
def is_word_continuation(self, word_prev, word):
|
@staticmethod
|
||||||
|
def is_word_continuation(word_prev, word):
|
||||||
"""
|
"""
|
||||||
Return True if the word is a continuation of the previous word
|
Return True if the word is a continuation of the previous word
|
||||||
by checking if the previous word is ending with a punctuation
|
by checking if the previous word is ending with a punctuation
|
||||||
@@ -61,7 +65,8 @@ class AudioDiarizationProcessor(Processor):
|
|||||||
return False
|
return False
|
||||||
return True
|
return True
|
||||||
|
|
||||||
def _diarization_remove_overlap(self, diarization: list[dict]):
|
@staticmethod
|
||||||
|
def _diarization_remove_overlap(diarization: list[DiarizationSegment]):
|
||||||
"""
|
"""
|
||||||
Remove overlap in diarization results
|
Remove overlap in diarization results
|
||||||
|
|
||||||
@@ -86,8 +91,9 @@ class AudioDiarizationProcessor(Processor):
|
|||||||
else:
|
else:
|
||||||
diarization_idx += 1
|
diarization_idx += 1
|
||||||
|
|
||||||
|
@staticmethod
|
||||||
def _diarization_remove_segment_without_words(
|
def _diarization_remove_segment_without_words(
|
||||||
self, words: list[Word], diarization: list[dict]
|
words: list[Word], diarization: list[DiarizationSegment]
|
||||||
):
|
):
|
||||||
"""
|
"""
|
||||||
Remove diarization segments without words
|
Remove diarization segments without words
|
||||||
@@ -116,9 +122,8 @@ class AudioDiarizationProcessor(Processor):
|
|||||||
else:
|
else:
|
||||||
diarization_idx += 1
|
diarization_idx += 1
|
||||||
|
|
||||||
def _diarization_merge_same_speaker(
|
@staticmethod
|
||||||
self, words: list[Word], diarization: list[dict]
|
def _diarization_merge_same_speaker(diarization: list[DiarizationSegment]):
|
||||||
):
|
|
||||||
"""
|
"""
|
||||||
Merge diarization contigous segments with the same speaker
|
Merge diarization contigous segments with the same speaker
|
||||||
|
|
||||||
@@ -135,7 +140,10 @@ class AudioDiarizationProcessor(Processor):
|
|||||||
else:
|
else:
|
||||||
diarization_idx += 1
|
diarization_idx += 1
|
||||||
|
|
||||||
def _diarization_assign_speaker(self, words: list[Word], diarization: list[dict]):
|
@classmethod
|
||||||
|
def _diarization_assign_speaker(
|
||||||
|
cls, words: list[Word], diarization: list[DiarizationSegment]
|
||||||
|
):
|
||||||
"""
|
"""
|
||||||
Assign speaker to words based on diarization
|
Assign speaker to words based on diarization
|
||||||
|
|
||||||
@@ -143,7 +151,7 @@ class AudioDiarizationProcessor(Processor):
|
|||||||
"""
|
"""
|
||||||
|
|
||||||
word_idx = 0
|
word_idx = 0
|
||||||
last_speaker = None
|
last_speaker = 0
|
||||||
for d in diarization:
|
for d in diarization:
|
||||||
start = d["start"]
|
start = d["start"]
|
||||||
end = d["end"]
|
end = d["end"]
|
||||||
@@ -158,7 +166,7 @@ class AudioDiarizationProcessor(Processor):
|
|||||||
# If it's a continuation, assign with the last speaker
|
# If it's a continuation, assign with the last speaker
|
||||||
is_continuation = False
|
is_continuation = False
|
||||||
if word_idx > 0 and word_idx < len(words) - 1:
|
if word_idx > 0 and word_idx < len(words) - 1:
|
||||||
is_continuation = self.is_word_continuation(
|
is_continuation = cls.is_word_continuation(
|
||||||
*words[word_idx - 1 : word_idx + 1]
|
*words[word_idx - 1 : word_idx + 1]
|
||||||
)
|
)
|
||||||
if is_continuation:
|
if is_continuation:
|
||||||
|
|||||||
74
server/reflector/processors/audio_diarization_pyannote.py
Normal file
74
server/reflector/processors/audio_diarization_pyannote.py
Normal file
@@ -0,0 +1,74 @@
|
|||||||
|
import os
|
||||||
|
|
||||||
|
import torch
|
||||||
|
import torchaudio
|
||||||
|
from pyannote.audio import Pipeline
|
||||||
|
|
||||||
|
from reflector.processors.audio_diarization import AudioDiarizationProcessor
|
||||||
|
from reflector.processors.audio_diarization_auto import AudioDiarizationAutoProcessor
|
||||||
|
from reflector.processors.types import AudioDiarizationInput, DiarizationSegment
|
||||||
|
|
||||||
|
|
||||||
|
class AudioDiarizationPyannoteProcessor(AudioDiarizationProcessor):
|
||||||
|
"""Local diarization processor using pyannote.audio library"""
|
||||||
|
|
||||||
|
def __init__(
|
||||||
|
self,
|
||||||
|
model_name: str = "pyannote/speaker-diarization-3.1",
|
||||||
|
pyannote_auth_token: str | None = None,
|
||||||
|
device: str | None = None,
|
||||||
|
**kwargs,
|
||||||
|
):
|
||||||
|
super().__init__(**kwargs)
|
||||||
|
self.model_name = model_name
|
||||||
|
self.auth_token = pyannote_auth_token or os.environ.get("HF_TOKEN")
|
||||||
|
self.device = device
|
||||||
|
|
||||||
|
if device is None:
|
||||||
|
self.device = "cuda" if torch.cuda.is_available() else "cpu"
|
||||||
|
|
||||||
|
self.logger.info(f"Loading pyannote diarization model: {self.model_name}")
|
||||||
|
self.diarization_pipeline = Pipeline.from_pretrained(
|
||||||
|
self.model_name, use_auth_token=self.auth_token
|
||||||
|
)
|
||||||
|
self.diarization_pipeline.to(torch.device(self.device))
|
||||||
|
self.logger.info(f"Diarization model loaded on device: {self.device}")
|
||||||
|
|
||||||
|
async def _diarize(self, data: AudioDiarizationInput) -> list[DiarizationSegment]:
|
||||||
|
try:
|
||||||
|
# Load audio file (audio_url is assumed to be a local file path)
|
||||||
|
self.logger.info(f"Loading local audio file: {data.audio_url}")
|
||||||
|
waveform, sample_rate = torchaudio.load(data.audio_url)
|
||||||
|
audio_input = {"waveform": waveform, "sample_rate": sample_rate}
|
||||||
|
self.logger.info("Running speaker diarization")
|
||||||
|
diarization = self.diarization_pipeline(audio_input)
|
||||||
|
|
||||||
|
# Convert pyannote diarization output to our format
|
||||||
|
segments = []
|
||||||
|
for segment, _, speaker in diarization.itertracks(yield_label=True):
|
||||||
|
# Extract speaker number from label (e.g., "SPEAKER_00" -> 0)
|
||||||
|
speaker_id = 0
|
||||||
|
if speaker.startswith("SPEAKER_"):
|
||||||
|
try:
|
||||||
|
speaker_id = int(speaker.split("_")[-1])
|
||||||
|
except (ValueError, IndexError):
|
||||||
|
# Fallback to hash-based ID if parsing fails
|
||||||
|
speaker_id = hash(speaker) % 1000
|
||||||
|
|
||||||
|
segments.append(
|
||||||
|
{
|
||||||
|
"start": round(segment.start, 3),
|
||||||
|
"end": round(segment.end, 3),
|
||||||
|
"speaker": speaker_id,
|
||||||
|
}
|
||||||
|
)
|
||||||
|
|
||||||
|
self.logger.info(f"Diarization completed with {len(segments)} segments")
|
||||||
|
return segments
|
||||||
|
|
||||||
|
except Exception as e:
|
||||||
|
self.logger.exception(f"Diarization failed: {e}")
|
||||||
|
raise
|
||||||
|
|
||||||
|
|
||||||
|
AudioDiarizationAutoProcessor.register("pyannote", AudioDiarizationPyannoteProcessor)
|
||||||
60
server/reflector/processors/audio_downscale.py
Normal file
60
server/reflector/processors/audio_downscale.py
Normal file
@@ -0,0 +1,60 @@
|
|||||||
|
from typing import Optional
|
||||||
|
|
||||||
|
import av
|
||||||
|
from av.audio.resampler import AudioResampler
|
||||||
|
|
||||||
|
from reflector.processors.base import Processor
|
||||||
|
|
||||||
|
|
||||||
|
def copy_frame(frame: av.AudioFrame) -> av.AudioFrame:
|
||||||
|
frame_copy = frame.from_ndarray(
|
||||||
|
frame.to_ndarray(),
|
||||||
|
format=frame.format.name,
|
||||||
|
layout=frame.layout.name,
|
||||||
|
)
|
||||||
|
frame_copy.sample_rate = frame.sample_rate
|
||||||
|
frame_copy.pts = frame.pts
|
||||||
|
frame_copy.time_base = frame.time_base
|
||||||
|
return frame_copy
|
||||||
|
|
||||||
|
|
||||||
|
class AudioDownscaleProcessor(Processor):
|
||||||
|
"""
|
||||||
|
Downscale audio frames to 16kHz mono format
|
||||||
|
"""
|
||||||
|
|
||||||
|
INPUT_TYPE = av.AudioFrame
|
||||||
|
OUTPUT_TYPE = av.AudioFrame
|
||||||
|
|
||||||
|
def __init__(self, target_rate: int = 16000, target_layout: str = "mono", **kwargs):
|
||||||
|
super().__init__(**kwargs)
|
||||||
|
self.target_rate = target_rate
|
||||||
|
self.target_layout = target_layout
|
||||||
|
self.resampler: Optional[AudioResampler] = None
|
||||||
|
self.needs_resampling: Optional[bool] = None
|
||||||
|
|
||||||
|
async def _push(self, data: av.AudioFrame):
|
||||||
|
if self.needs_resampling is None:
|
||||||
|
self.needs_resampling = (
|
||||||
|
data.sample_rate != self.target_rate
|
||||||
|
or data.layout.name != self.target_layout
|
||||||
|
)
|
||||||
|
|
||||||
|
if self.needs_resampling:
|
||||||
|
self.resampler = AudioResampler(
|
||||||
|
format="s16", layout=self.target_layout, rate=self.target_rate
|
||||||
|
)
|
||||||
|
|
||||||
|
if not self.needs_resampling or not self.resampler:
|
||||||
|
await self.emit(data)
|
||||||
|
return
|
||||||
|
|
||||||
|
resampled_frames = self.resampler.resample(copy_frame(data))
|
||||||
|
for resampled_frame in resampled_frames:
|
||||||
|
await self.emit(resampled_frame)
|
||||||
|
|
||||||
|
async def _flush(self):
|
||||||
|
if self.needs_resampling and self.resampler:
|
||||||
|
final_frames = self.resampler.resample(None)
|
||||||
|
for frame in final_frames:
|
||||||
|
await self.emit(frame)
|
||||||
@@ -16,37 +16,46 @@ class AudioMergeProcessor(Processor):
|
|||||||
INPUT_TYPE = list[av.AudioFrame]
|
INPUT_TYPE = list[av.AudioFrame]
|
||||||
OUTPUT_TYPE = AudioFile
|
OUTPUT_TYPE = AudioFile
|
||||||
|
|
||||||
|
def __init__(self, **kwargs):
|
||||||
|
super().__init__(**kwargs)
|
||||||
|
|
||||||
async def _push(self, data: list[av.AudioFrame]):
|
async def _push(self, data: list[av.AudioFrame]):
|
||||||
if not data:
|
if not data:
|
||||||
return
|
return
|
||||||
|
|
||||||
# get audio information from first frame
|
# get audio information from first frame
|
||||||
frame = data[0]
|
frame = data[0]
|
||||||
channels = len(frame.layout.channels)
|
output_channels = len(frame.layout.channels)
|
||||||
sample_rate = frame.sample_rate
|
output_sample_rate = frame.sample_rate
|
||||||
sample_width = frame.format.bytes
|
output_sample_width = frame.format.bytes
|
||||||
|
|
||||||
# create audio file
|
# create audio file
|
||||||
uu = uuid4().hex
|
uu = uuid4().hex
|
||||||
fd = io.BytesIO()
|
fd = io.BytesIO()
|
||||||
|
|
||||||
|
# Use PyAV to write frames
|
||||||
out_container = av.open(fd, "w", format="wav")
|
out_container = av.open(fd, "w", format="wav")
|
||||||
out_stream = out_container.add_stream("pcm_s16le", rate=sample_rate)
|
out_stream = out_container.add_stream("pcm_s16le", rate=output_sample_rate)
|
||||||
|
out_stream.layout = frame.layout.name
|
||||||
|
|
||||||
for frame in data:
|
for frame in data:
|
||||||
for packet in out_stream.encode(frame):
|
for packet in out_stream.encode(frame):
|
||||||
out_container.mux(packet)
|
out_container.mux(packet)
|
||||||
|
|
||||||
|
# Flush the encoder
|
||||||
for packet in out_stream.encode(None):
|
for packet in out_stream.encode(None):
|
||||||
out_container.mux(packet)
|
out_container.mux(packet)
|
||||||
out_container.close()
|
out_container.close()
|
||||||
|
|
||||||
fd.seek(0)
|
fd.seek(0)
|
||||||
|
|
||||||
# emit audio file
|
# emit audio file
|
||||||
audiofile = AudioFile(
|
audiofile = AudioFile(
|
||||||
name=f"{monotonic_ns()}-{uu}.wav",
|
name=f"{monotonic_ns()}-{uu}.wav",
|
||||||
fd=fd,
|
fd=fd,
|
||||||
sample_rate=sample_rate,
|
sample_rate=output_sample_rate,
|
||||||
channels=channels,
|
channels=output_channels,
|
||||||
sample_width=sample_width,
|
sample_width=output_sample_width,
|
||||||
timestamp=data[0].pts * data[0].time_base,
|
timestamp=data[0].pts * data[0].time_base,
|
||||||
)
|
)
|
||||||
|
|
||||||
|
|||||||
@@ -21,7 +21,11 @@ from reflector.settings import settings
|
|||||||
|
|
||||||
|
|
||||||
class AudioTranscriptModalProcessor(AudioTranscriptProcessor):
|
class AudioTranscriptModalProcessor(AudioTranscriptProcessor):
|
||||||
def __init__(self, modal_api_key: str | None = None, **kwargs):
|
def __init__(
|
||||||
|
self,
|
||||||
|
modal_api_key: str | None = None,
|
||||||
|
**kwargs,
|
||||||
|
):
|
||||||
super().__init__()
|
super().__init__()
|
||||||
if not settings.TRANSCRIPT_URL:
|
if not settings.TRANSCRIPT_URL:
|
||||||
raise Exception(
|
raise Exception(
|
||||||
|
|||||||
@@ -173,6 +173,7 @@ class Processor(Emitter):
|
|||||||
except Exception:
|
except Exception:
|
||||||
self.m_processor_failure.inc()
|
self.m_processor_failure.inc()
|
||||||
self.logger.exception("Error in push")
|
self.logger.exception("Error in push")
|
||||||
|
raise
|
||||||
|
|
||||||
async def flush(self):
|
async def flush(self):
|
||||||
"""
|
"""
|
||||||
@@ -240,14 +241,15 @@ class ThreadedProcessor(Processor):
|
|||||||
self.INPUT_TYPE = processor.INPUT_TYPE
|
self.INPUT_TYPE = processor.INPUT_TYPE
|
||||||
self.OUTPUT_TYPE = processor.OUTPUT_TYPE
|
self.OUTPUT_TYPE = processor.OUTPUT_TYPE
|
||||||
self.executor = ThreadPoolExecutor(max_workers=max_workers)
|
self.executor = ThreadPoolExecutor(max_workers=max_workers)
|
||||||
self.queue = asyncio.Queue()
|
self.queue = asyncio.Queue(maxsize=50)
|
||||||
self.task = asyncio.get_running_loop().create_task(self.loop())
|
self.task: asyncio.Task | None = None
|
||||||
|
|
||||||
def set_pipeline(self, pipeline: "Pipeline"):
|
def set_pipeline(self, pipeline: "Pipeline"):
|
||||||
super().set_pipeline(pipeline)
|
super().set_pipeline(pipeline)
|
||||||
self.processor.set_pipeline(pipeline)
|
self.processor.set_pipeline(pipeline)
|
||||||
|
|
||||||
async def loop(self):
|
async def loop(self):
|
||||||
|
try:
|
||||||
while True:
|
while True:
|
||||||
data = await self.queue.get()
|
data = await self.queue.get()
|
||||||
self.m_processor_queue.set(self.queue.qsize())
|
self.m_processor_queue.set(self.queue.qsize())
|
||||||
@@ -265,8 +267,19 @@ class ThreadedProcessor(Processor):
|
|||||||
)
|
)
|
||||||
finally:
|
finally:
|
||||||
self.queue.task_done()
|
self.queue.task_done()
|
||||||
|
except Exception as e:
|
||||||
|
logger.error(f"Crash in {self.__class__.__name__}: {e}", exc_info=e)
|
||||||
|
|
||||||
|
async def _ensure_task(self):
|
||||||
|
if self.task is None:
|
||||||
|
self.task = asyncio.get_running_loop().create_task(self.loop())
|
||||||
|
|
||||||
|
# XXX not doing a sleep here make the whole pipeline prior the thread
|
||||||
|
# to be running without having a chance to work on the task here.
|
||||||
|
await asyncio.sleep(0)
|
||||||
|
|
||||||
async def _push(self, data):
|
async def _push(self, data):
|
||||||
|
await self._ensure_task()
|
||||||
await self.queue.put(data)
|
await self.queue.put(data)
|
||||||
|
|
||||||
async def _flush(self):
|
async def _flush(self):
|
||||||
|
|||||||
33
server/reflector/processors/file_diarization.py
Normal file
33
server/reflector/processors/file_diarization.py
Normal file
@@ -0,0 +1,33 @@
|
|||||||
|
from pydantic import BaseModel
|
||||||
|
|
||||||
|
from reflector.processors.base import Processor
|
||||||
|
from reflector.processors.types import DiarizationSegment
|
||||||
|
|
||||||
|
|
||||||
|
class FileDiarizationInput(BaseModel):
|
||||||
|
"""Input for file diarization containing audio URL"""
|
||||||
|
|
||||||
|
audio_url: str
|
||||||
|
|
||||||
|
|
||||||
|
class FileDiarizationOutput(BaseModel):
|
||||||
|
"""Output for file diarization containing speaker segments"""
|
||||||
|
|
||||||
|
diarization: list[DiarizationSegment]
|
||||||
|
|
||||||
|
|
||||||
|
class FileDiarizationProcessor(Processor):
|
||||||
|
"""
|
||||||
|
Diarize complete audio files from URL
|
||||||
|
"""
|
||||||
|
|
||||||
|
INPUT_TYPE = FileDiarizationInput
|
||||||
|
OUTPUT_TYPE = FileDiarizationOutput
|
||||||
|
|
||||||
|
async def _push(self, data: FileDiarizationInput):
|
||||||
|
result = await self._diarize(data)
|
||||||
|
if result:
|
||||||
|
await self.emit(result)
|
||||||
|
|
||||||
|
async def _diarize(self, data: FileDiarizationInput):
|
||||||
|
raise NotImplementedError
|
||||||
33
server/reflector/processors/file_diarization_auto.py
Normal file
33
server/reflector/processors/file_diarization_auto.py
Normal file
@@ -0,0 +1,33 @@
|
|||||||
|
import importlib
|
||||||
|
|
||||||
|
from reflector.processors.file_diarization import FileDiarizationProcessor
|
||||||
|
from reflector.settings import settings
|
||||||
|
|
||||||
|
|
||||||
|
class FileDiarizationAutoProcessor(FileDiarizationProcessor):
|
||||||
|
_registry = {}
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def register(cls, name, kclass):
|
||||||
|
cls._registry[name] = kclass
|
||||||
|
|
||||||
|
def __new__(cls, name: str | None = None, **kwargs):
|
||||||
|
if name is None:
|
||||||
|
name = settings.DIARIZATION_BACKEND
|
||||||
|
|
||||||
|
if name not in cls._registry:
|
||||||
|
module_name = f"reflector.processors.file_diarization_{name}"
|
||||||
|
importlib.import_module(module_name)
|
||||||
|
|
||||||
|
# gather specific configuration for the processor
|
||||||
|
# search `DIARIZATION_BACKEND_XXX_YYY`, push to constructor as `backend_xxx_yyy`
|
||||||
|
config = {}
|
||||||
|
name_upper = name.upper()
|
||||||
|
settings_prefix = "DIARIZATION_"
|
||||||
|
config_prefix = f"{settings_prefix}{name_upper}_"
|
||||||
|
for key, value in settings:
|
||||||
|
if key.startswith(config_prefix):
|
||||||
|
config_name = key[len(settings_prefix) :].lower()
|
||||||
|
config[config_name] = value
|
||||||
|
|
||||||
|
return cls._registry[name](**config | kwargs)
|
||||||
58
server/reflector/processors/file_diarization_modal.py
Normal file
58
server/reflector/processors/file_diarization_modal.py
Normal file
@@ -0,0 +1,58 @@
|
|||||||
|
"""
|
||||||
|
File diarization implementation using the GPU service from modal.com
|
||||||
|
|
||||||
|
API will be a POST request to DIARIZATION_URL:
|
||||||
|
|
||||||
|
```
|
||||||
|
POST /diarize?audio_file_url=...×tamp=0
|
||||||
|
Authorization: Bearer <modal_api_key>
|
||||||
|
```
|
||||||
|
"""
|
||||||
|
|
||||||
|
import httpx
|
||||||
|
|
||||||
|
from reflector.processors.file_diarization import (
|
||||||
|
FileDiarizationInput,
|
||||||
|
FileDiarizationOutput,
|
||||||
|
FileDiarizationProcessor,
|
||||||
|
)
|
||||||
|
from reflector.processors.file_diarization_auto import FileDiarizationAutoProcessor
|
||||||
|
from reflector.settings import settings
|
||||||
|
|
||||||
|
|
||||||
|
class FileDiarizationModalProcessor(FileDiarizationProcessor):
|
||||||
|
def __init__(self, modal_api_key: str | None = None, **kwargs):
|
||||||
|
super().__init__(**kwargs)
|
||||||
|
if not settings.DIARIZATION_URL:
|
||||||
|
raise Exception(
|
||||||
|
"DIARIZATION_URL required to use FileDiarizationModalProcessor"
|
||||||
|
)
|
||||||
|
self.diarization_url = settings.DIARIZATION_URL + "/diarize"
|
||||||
|
self.file_timeout = settings.DIARIZATION_FILE_TIMEOUT
|
||||||
|
self.modal_api_key = modal_api_key
|
||||||
|
|
||||||
|
async def _diarize(self, data: FileDiarizationInput):
|
||||||
|
"""Get speaker diarization for file"""
|
||||||
|
self.logger.info(f"Starting diarization from {data.audio_url}")
|
||||||
|
|
||||||
|
headers = {}
|
||||||
|
if self.modal_api_key:
|
||||||
|
headers["Authorization"] = f"Bearer {self.modal_api_key}"
|
||||||
|
|
||||||
|
async with httpx.AsyncClient(timeout=self.file_timeout) as client:
|
||||||
|
response = await client.post(
|
||||||
|
self.diarization_url,
|
||||||
|
headers=headers,
|
||||||
|
params={
|
||||||
|
"audio_file_url": data.audio_url,
|
||||||
|
"timestamp": 0,
|
||||||
|
},
|
||||||
|
follow_redirects=True,
|
||||||
|
)
|
||||||
|
response.raise_for_status()
|
||||||
|
diarization_data = response.json()["diarization"]
|
||||||
|
|
||||||
|
return FileDiarizationOutput(diarization=diarization_data)
|
||||||
|
|
||||||
|
|
||||||
|
FileDiarizationAutoProcessor.register("modal", FileDiarizationModalProcessor)
|
||||||
65
server/reflector/processors/file_transcript.py
Normal file
65
server/reflector/processors/file_transcript.py
Normal file
@@ -0,0 +1,65 @@
|
|||||||
|
from prometheus_client import Counter, Histogram
|
||||||
|
|
||||||
|
from reflector.processors.base import Processor
|
||||||
|
from reflector.processors.types import Transcript
|
||||||
|
|
||||||
|
|
||||||
|
class FileTranscriptInput:
|
||||||
|
"""Input for file transcription containing audio URL and language settings"""
|
||||||
|
|
||||||
|
def __init__(self, audio_url: str, language: str = "en"):
|
||||||
|
self.audio_url = audio_url
|
||||||
|
self.language = language
|
||||||
|
|
||||||
|
|
||||||
|
class FileTranscriptProcessor(Processor):
|
||||||
|
"""
|
||||||
|
Transcript complete audio files from URL
|
||||||
|
"""
|
||||||
|
|
||||||
|
INPUT_TYPE = FileTranscriptInput
|
||||||
|
OUTPUT_TYPE = Transcript
|
||||||
|
|
||||||
|
m_transcript = Histogram(
|
||||||
|
"file_transcript",
|
||||||
|
"Time spent in FileTranscript.transcript",
|
||||||
|
["backend"],
|
||||||
|
)
|
||||||
|
m_transcript_call = Counter(
|
||||||
|
"file_transcript_call",
|
||||||
|
"Number of calls to FileTranscript.transcript",
|
||||||
|
["backend"],
|
||||||
|
)
|
||||||
|
m_transcript_success = Counter(
|
||||||
|
"file_transcript_success",
|
||||||
|
"Number of successful calls to FileTranscript.transcript",
|
||||||
|
["backend"],
|
||||||
|
)
|
||||||
|
m_transcript_failure = Counter(
|
||||||
|
"file_transcript_failure",
|
||||||
|
"Number of failed calls to FileTranscript.transcript",
|
||||||
|
["backend"],
|
||||||
|
)
|
||||||
|
|
||||||
|
def __init__(self, *args, **kwargs):
|
||||||
|
name = self.__class__.__name__
|
||||||
|
self.m_transcript = self.m_transcript.labels(name)
|
||||||
|
self.m_transcript_call = self.m_transcript_call.labels(name)
|
||||||
|
self.m_transcript_success = self.m_transcript_success.labels(name)
|
||||||
|
self.m_transcript_failure = self.m_transcript_failure.labels(name)
|
||||||
|
super().__init__(*args, **kwargs)
|
||||||
|
|
||||||
|
async def _push(self, data: FileTranscriptInput):
|
||||||
|
try:
|
||||||
|
self.m_transcript_call.inc()
|
||||||
|
with self.m_transcript.time():
|
||||||
|
result = await self._transcript(data)
|
||||||
|
self.m_transcript_success.inc()
|
||||||
|
if result:
|
||||||
|
await self.emit(result)
|
||||||
|
except Exception:
|
||||||
|
self.m_transcript_failure.inc()
|
||||||
|
raise
|
||||||
|
|
||||||
|
async def _transcript(self, data: FileTranscriptInput):
|
||||||
|
raise NotImplementedError
|
||||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user