@@ -885,27 +885,27 @@ async def execute(self, ctx: NodeContext) -> NodeResult:
885885 # next iteration. If we let the judge fire now it sees
886886 # "missing outputs" and injects RETRY feedback *before* the
887887 # user's answer, confusing the LLM.
888- _continue_count += 1
889- if ctx .runtime_logger :
890- iter_latency_ms = int ((time .time () - iter_start ) * 1000 )
891- verdict_fb = (
892- "Auto-blocked for user input (pre-interaction)"
893- if _cf_auto
894- else "Blocked for ask_user input (skip judge)"
895- )
896- ctx .runtime_logger .log_step (
897- node_id = node_id ,
898- node_type = "event_loop" ,
899- step_index = iteration ,
900- verdict = "CONTINUE" ,
901- verdict_feedback = verdict_fb ,
902- tool_calls = logged_tool_calls ,
903- llm_text = assistant_text ,
904- input_tokens = turn_tokens .get ("input" , 0 ),
905- output_tokens = turn_tokens .get ("output" , 0 ),
906- latency_ms = iter_latency_ms ,
907- )
908- continue
888+ # _continue_count += 1
889+ # if ctx.runtime_logger:
890+ # iter_latency_ms = int((time.time() - iter_start) * 1000)
891+ # verdict_fb = (
892+ # "Auto-blocked for user input (pre-interaction)"
893+ # if _cf_auto
894+ # else "Blocked for ask_user input (skip judge)"
895+ # )
896+ # ctx.runtime_logger.log_step(
897+ # node_id=node_id,
898+ # node_type="event_loop",
899+ # step_index=iteration,
900+ # verdict="CONTINUE",
901+ # verdict_feedback=verdict_fb,
902+ # tool_calls=logged_tool_calls,
903+ # llm_text=assistant_text,
904+ # input_tokens=turn_tokens.get("input", 0),
905+ # output_tokens=turn_tokens.get("output", 0),
906+ # latency_ms=iter_latency_ms,
907+ # )
908+ # continue
909909
910910 # 6i. Judge evaluation
911911 should_judge = (
0 commit comments