@@ -93,6 +93,7 @@ def event_content(response,
9393 reasoning_content_chunk = chunk .additional_kwargs .get ('reasoning_content' , '' )
9494 else :
9595 reasoning_content_chunk = reasoning_chunk .get ('reasoning_content' )
96+ content_chunk = reasoning ._normalize_content (content_chunk )
9697 all_text += content_chunk
9798 if reasoning_content_chunk is None :
9899 reasoning_content_chunk = ''
@@ -191,13 +192,15 @@ def execute(self, message_list: List[BaseMessage],
191192 manage , padding_problem_text , chat_user_id , chat_user_type ,
192193 no_references_setting ,
193194 model_setting ,
194- mcp_enable , mcp_tool_ids , mcp_servers , mcp_source , tool_enable , tool_ids , mcp_output_enable )
195+ mcp_enable , mcp_tool_ids , mcp_servers , mcp_source , tool_enable , tool_ids ,
196+ mcp_output_enable )
195197 else :
196198 return self .execute_block (message_list , chat_id , problem_text , post_response_handler , chat_model ,
197199 paragraph_list ,
198200 manage , padding_problem_text , chat_user_id , chat_user_type , no_references_setting ,
199201 model_setting ,
200- mcp_enable , mcp_tool_ids , mcp_servers , mcp_source , tool_enable , tool_ids , mcp_output_enable )
202+ mcp_enable , mcp_tool_ids , mcp_servers , mcp_source , tool_enable , tool_ids ,
203+ mcp_output_enable )
201204
202205 def get_details (self , manage , ** kwargs ):
203206 return {
@@ -264,7 +267,6 @@ def _handle_mcp_request(self, mcp_enable, tool_enable, mcp_source, mcp_servers,
264267
265268 return None
266269
267-
268270 def get_stream_result (self , message_list : List [BaseMessage ],
269271 chat_model : BaseChatModel = None ,
270272 paragraph_list = None ,
@@ -294,7 +296,8 @@ def get_stream_result(self, message_list: List[BaseMessage],
294296 else :
295297 # 处理 MCP 请求
296298 mcp_result = self ._handle_mcp_request (
297- mcp_enable , tool_enable , mcp_source , mcp_servers , mcp_tool_ids , tool_ids , mcp_output_enable , chat_model , message_list ,
299+ mcp_enable , tool_enable , mcp_source , mcp_servers , mcp_tool_ids , tool_ids , mcp_output_enable , chat_model ,
300+ message_list ,
298301 )
299302 if mcp_result :
300303 return mcp_result , True
@@ -319,7 +322,8 @@ def execute_stream(self, message_list: List[BaseMessage],
319322 tool_ids = None ,
320323 mcp_output_enable = True ):
321324 chat_result , is_ai_chat = self .get_stream_result (message_list , chat_model , paragraph_list ,
322- no_references_setting , problem_text , mcp_enable , mcp_tool_ids , mcp_servers , mcp_source , tool_enable , tool_ids ,
325+ no_references_setting , problem_text , mcp_enable , mcp_tool_ids ,
326+ mcp_servers , mcp_source , tool_enable , tool_ids ,
323327 mcp_output_enable )
324328 chat_record_id = uuid .uuid7 ()
325329 r = StreamingHttpResponse (
@@ -394,7 +398,9 @@ def execute_block(self, message_list: List[BaseMessage],
394398 # 调用模型
395399 try :
396400 chat_result , is_ai_chat = self .get_block_result (message_list , chat_model , paragraph_list ,
397- no_references_setting , problem_text , mcp_enable , mcp_tool_ids , mcp_servers , mcp_source , tool_enable , tool_ids , mcp_output_enable )
401+ no_references_setting , problem_text , mcp_enable ,
402+ mcp_tool_ids , mcp_servers , mcp_source , tool_enable ,
403+ tool_ids , mcp_output_enable )
398404 if is_ai_chat :
399405 request_token = chat_model .get_num_tokens_from_messages (message_list )
400406 response_token = chat_model .get_num_tokens (chat_result .content )
0 commit comments