From 763a6b8d6a40333f164a0e97a2054fbd41329211 Mon Sep 17 00:00:00 2001 From: "deepsource-autofix[bot]" <62050782+deepsource-autofix[bot]@users.noreply.github.com> Date: Thu, 19 Dec 2024 11:49:57 +0000 Subject: [PATCH] style: format code with Autopep8 This commit fixes the style issues introduced in 1ba3e08 according to the output from Autopep8. Details: None --- python/openwebui/pipe_mcts.py | 14 +++++++++++--- 1 file changed, 11 insertions(+), 3 deletions(-) diff --git a/python/openwebui/pipe_mcts.py b/python/openwebui/pipe_mcts.py index b0237e8..2b2b8c0 100644 --- a/python/openwebui/pipe_mcts.py +++ b/python/openwebui/pipe_mcts.py @@ -63,6 +63,7 @@ # ============================================================================= + class AsyncIteratorCallbackHandler(AsyncCallbackHandler): def __init__(self): self.queue = asyncio.Queue() @@ -86,6 +87,7 @@ async def __aiter__(self): break yield token + class LLMClient: def __init__(self, valves: "Pipe.Valves"): self.openai_api_key = valves.OPENAI_API_KEY @@ -182,12 +184,14 @@ def get_chunk_content(self, chunk): if "content" in delta: yield delta["content"] except json.JSONDecodeError: - logger.error(f'ChunkDecodeError: unable to parse "{chunk_str[:100]}"') + logger.error( + f'ChunkDecodeError: unable to parse "{chunk_str[:100]}"') # ============================================================================= # MCTS Classes + class Node: def __init__( self, @@ -240,6 +244,7 @@ def mermaid(self, offset=0, selected=None): logger.debug(f"Node Mermaid:\n{msg}") return msg + class MCTSAgent: def __init__( self, @@ -434,7 +439,8 @@ async def update_approach(self, answer: str, improvements: str): return await self.generate_completion(prompt) async def evaluate_answer(self, answer: str): - prompt = eval_answer_prompt.format(question=self.question, answer=answer) + prompt = eval_answer_prompt.format( + question=self.question, answer=answer) result = await self.generate_completion(prompt) try: score = int(re.search(r"\d+", result).group()) @@ -487,6 +493,7 @@ async def emit_replace(self, content: str): # Prompts + thoughts_prompt = """ Give a suggestion on how this answer can be improved. @@ -658,7 +665,8 @@ async def pipe( if match: backend, model_name = match.groups() else: - logger.error("Model ID should be in the format '*.mcts/backend/model_name'") + logger.error( + "Model ID should be in the format '*.mcts/backend/model_name'") logger.error(f"Invalid model ID: {model_id}") return ""