mirror of
https://github.com/Monadical-SAS/reflector.git
synced 2025-12-21 04:39:06 +00:00
PR review comments
This commit is contained in:
@@ -2,7 +2,6 @@ import importlib
|
||||
import json
|
||||
import re
|
||||
from time import monotonic
|
||||
from typing import Union
|
||||
|
||||
from reflector.logger import logger as reflector_logger
|
||||
from reflector.settings import settings
|
||||
@@ -47,7 +46,7 @@ class LLM:
|
||||
pass
|
||||
|
||||
async def generate(
|
||||
self, prompt: str, logger: reflector_logger, schema: str = None, **kwargs
|
||||
self, prompt: str, logger: reflector_logger, schema: str | None = None, **kwargs
|
||||
) -> dict:
|
||||
logger.info("LLM generate", prompt=repr(prompt))
|
||||
try:
|
||||
@@ -63,7 +62,7 @@ class LLM:
|
||||
|
||||
return result
|
||||
|
||||
async def _generate(self, prompt: str, schema: Union[str | None], **kwargs) -> str:
|
||||
async def _generate(self, prompt: str, schema: str | None, **kwargs) -> str:
|
||||
raise NotImplementedError
|
||||
|
||||
def _parse_json(self, result: str) -> dict:
|
||||
|
||||
@@ -1,5 +1,4 @@
|
||||
import json
|
||||
from typing import Union
|
||||
|
||||
import httpx
|
||||
from reflector.llm.base import LLM
|
||||
@@ -16,7 +15,7 @@ class BananaLLM(LLM):
|
||||
"X-Banana-Model-Key": settings.LLM_BANANA_MODEL_KEY,
|
||||
}
|
||||
|
||||
async def _generate(self, prompt: str, schema: Union[str | None], **kwargs):
|
||||
async def _generate(self, prompt: str, schema: str | None, **kwargs):
|
||||
json_payload = {"prompt": prompt}
|
||||
if schema:
|
||||
json_payload["schema"] = json.dumps(schema)
|
||||
|
||||
@@ -1,5 +1,4 @@
|
||||
import json
|
||||
from typing import Union
|
||||
|
||||
import httpx
|
||||
from reflector.llm.base import LLM
|
||||
@@ -26,7 +25,7 @@ class ModalLLM(LLM):
|
||||
)
|
||||
response.raise_for_status()
|
||||
|
||||
async def _generate(self, prompt: str, schema: Union[str | None], **kwargs):
|
||||
async def _generate(self, prompt: str, schema: str | None, **kwargs):
|
||||
json_payload = {"prompt": prompt}
|
||||
if schema:
|
||||
json_payload["schema"] = json.dumps(schema)
|
||||
|
||||
@@ -1,5 +1,4 @@
|
||||
import json
|
||||
from typing import Union
|
||||
|
||||
import httpx
|
||||
from reflector.llm.base import LLM
|
||||
@@ -7,7 +6,7 @@ from reflector.settings import settings
|
||||
|
||||
|
||||
class OobaboogaLLM(LLM):
|
||||
async def _generate(self, prompt: str, schema: Union[str | None], **kwargs):
|
||||
async def _generate(self, prompt: str, schema: str | None, **kwargs):
|
||||
json_payload = {"prompt": prompt}
|
||||
if schema:
|
||||
json_payload["schema"] = json.dumps(schema)
|
||||
@@ -1,5 +1,3 @@
|
||||
from typing import Union
|
||||
|
||||
import httpx
|
||||
from reflector.llm.base import LLM
|
||||
from reflector.logger import logger
|
||||
@@ -17,7 +15,7 @@ class OpenAILLM(LLM):
|
||||
self.max_tokens = settings.LLM_MAX_TOKENS
|
||||
logger.info(f"LLM use openai backend at {self.openai_url}")
|
||||
|
||||
async def _generate(self, prompt: str, schema: Union[str | None], **kwargs) -> str:
|
||||
async def _generate(self, prompt: str, schema: str | None, **kwargs) -> str:
|
||||
headers = {
|
||||
"Content-Type": "application/json",
|
||||
"Authorization": f"Bearer {self.openai_key}",
|
||||
|
||||
Reference in New Issue
Block a user