mirror of
https://github.com/Monadical-SAS/reflector.git
synced 2026-02-04 09:56:47 +00:00
Task 3: LLM Streaming Integration - Import Settings, ChatMessage, MessageRole from llama-index - Configure LLM with temperature 0.7 on connection - Build system message with WebVTT transcript context (max 15k chars) - Initialize conversation history with system message - Handle 'message' type from client to trigger LLM streaming - Stream LLM response using Settings.llm.astream_chat() - Send tokens incrementally via 'token' messages - Send 'done' message when streaming completes - Maintain conversation history across multiple messages - Add error handling with 'error' message type - Add message protocol validation test Implements Tasks 3 & 4 from TASKS.md
5.8 KiB
5.8 KiB