From 05f4506fcdf080cdef764638b17ea313aa0f9563 Mon Sep 17 00:00:00 2001 From: NeonKirill Date: Sat, 9 Mar 2024 22:45:07 +0100 Subject: [PATCH] removed response queue --- neon_llm_core/chatbot.py | 2 +- neon_llm_core/utils/personas/provider.py | 3 +-- 2 files changed, 2 insertions(+), 3 deletions(-) diff --git a/neon_llm_core/chatbot.py b/neon_llm_core/chatbot.py index 405da11..e8884d9 100644 --- a/neon_llm_core/chatbot.py +++ b/neon_llm_core/chatbot.py @@ -37,7 +37,7 @@ class LLMBot(ChatBot): def __init__(self, *args, **kwargs): ChatBot.__init__(self, *args, **kwargs) self.bot_type = "submind" - self.base_llm = kwargs.get("llm_name") # chat_gpt, fastchat, etc. + self.base_llm = kwargs.get("llm_name") # chatgpt, fastchat, etc. self.persona = kwargs.get("persona") self.mq_queue_config = self.get_llm_mq_config(self.base_llm) LOG.info(f'Initialised config for llm={self.base_llm}|' diff --git a/neon_llm_core/utils/personas/provider.py b/neon_llm_core/utils/personas/provider.py index 2a24d9a..657cd49 100644 --- a/neon_llm_core/utils/personas/provider.py +++ b/neon_llm_core/utils/personas/provider.py @@ -77,8 +77,7 @@ def _fetch_persona_config(self): queue = "get_configured_personas" response = send_mq_request(vhost=LLM_VHOST, request_data={"service_name": self.service_name}, - target_queue=queue, - response_queue=f'{queue}.{self.service_name}.response') + target_queue=queue) self.personas = response.get('items', []) for persona in self.personas: if persona: