mirror of
https://github.com/Monadical-SAS/reflector.git
synced 2025-12-20 20:29:06 +00:00
30cffeeb6b3f4e2323ac64277cffacac1298b1fb
Sample Chat generation with LLM
Reflector
Reflector server is responsible for audio transcription and summarization for now. The project is moving fast, documentation is currently unstable and outdated
Server
We currently use oogabooga as a LLM backend.
Using docker
Create a .env with
LLM_URL=http://IP:HOST/api/v1/generate
Then start with:
$ docker-compose up
Languages
Python
72.3%
TypeScript
26.9%
JavaScript
0.3%
Dockerfile
0.2%