@@ -770,7 +770,7 @@ def _chat_completion(self, messages, temperature=0.2, tools=None, stream=True, r
770770 display_error (f"Error in chat completion: { e } " )
771771 return None
772772
773- def chat (self , prompt , temperature = 0.2 , tools = None , output_json = None , output_pydantic = None , reasoning_steps = False ):
773+ def chat (self , prompt , temperature = 0.2 , tools = None , output_json = None , output_pydantic = None , reasoning_steps = False , stream = True ):
774774 # Log all parameter values when in debug mode
775775 if logging .getLogger ().getEffectiveLevel () == logging .DEBUG :
776776 param_info = {
@@ -912,7 +912,7 @@ def chat(self, prompt, temperature=0.2, tools=None, output_json=None, output_pyd
912912 agent_tools = agent_tools
913913 )
914914
915- response = self ._chat_completion (messages , temperature = temperature , tools = tools if tools else None , reasoning_steps = reasoning_steps )
915+ response = self ._chat_completion (messages , temperature = temperature , tools = tools if tools else None , reasoning_steps = reasoning_steps , stream = stream )
916916 if not response :
917917 return None
918918
@@ -949,7 +949,7 @@ def chat(self, prompt, temperature=0.2, tools=None, output_json=None, output_pyd
949949 "content" : "Function returned an empty output"
950950 })
951951
952- response = self ._chat_completion (messages , temperature = temperature )
952+ response = self ._chat_completion (messages , temperature = temperature , stream = stream )
953953 if not response :
954954 return None
955955 response_text = response .choices [0 ].message .content .strip ()
@@ -1019,7 +1019,7 @@ def chat(self, prompt, temperature=0.2, tools=None, output_json=None, output_pyd
10191019
10201020 logging .debug (f"{ self .name } reflection count { reflection_count + 1 } , continuing reflection process" )
10211021 messages .append ({"role" : "user" , "content" : "Now regenerate your response using the reflection you made" })
1022- response = self ._chat_completion (messages , temperature = temperature , tools = None , stream = True )
1022+ response = self ._chat_completion (messages , temperature = temperature , tools = None , stream = stream )
10231023 response_text = response .choices [0 ].message .content .strip ()
10241024 reflection_count += 1
10251025 continue # Continue the loop for more reflections
@@ -1199,7 +1199,7 @@ async def achat(self, prompt: str, temperature=0.2, tools=None, output_json=None
11991199 model = self .llm ,
12001200 messages = messages ,
12011201 temperature = temperature ,
1202- tools = formatted_tools
1202+ tools = formatted_tools ,
12031203 )
12041204 result = await self ._achat_completion (response , tools )
12051205 if logging .getLogger ().getEffectiveLevel () == logging .DEBUG :
0 commit comments