agent-runtime/agent/nodes/expert_base.py
Nico 1000411eb2 v0.15.0: Frame engine (v3), PA + Expert architecture (v4-eras), live test streaming
Frame Engine (v3-framed):
- Tick-based deterministic pipeline: frames advance on completion, not timers
- FrameRecord/FrameTrace dataclasses for structured per-message tracing
- /api/frames endpoint: queryable frame trace history (last 20 messages)
- frame_trace HUD event with full pipeline visibility
- Reflex=2F, Director=4F, Director+Interpreter=5F deterministic frame counts

Expert Architecture (v4-eras):
- PA node (pa_v1): routes to domain experts, holds user context
- ExpertNode base: stateless executor with plan+execute two-LLM-call pattern
- ErasExpertNode: eras2_production DB specialist with DESCRIBE-first discipline
- Schema caching: DESCRIBE results reused across queries within session
- Progress streaming: PA streams thinking message, expert streams per-tool progress
- PARouting type for structured routing decisions

UI Controls Split:
- Separate thinker_controls from machine controls (current_controls is now a property)
- Machine buttons persist across Thinker responses
- Machine state parser handles both dict and list formats from Director
- Normalized button format with go/payload field mapping

WebSocket Architecture:
- /ws/test: dedicated debug socket for test runner progress
- /ws/trace: dedicated debug socket for HUD/frame trace events
- /ws (chat): cleaned up, only deltas/controls/done/cleared
- WS survives graph switch (re-attaches to new runtime)
- Pipeline result reset on clear

Test Infrastructure:
- Live test streaming: on_result callback fires per check during execution
- Frontend polling fallback (500ms) for proxy-buffered WS
- frame_trace-first trace assertion (fixes stale perceived event bug)
- action_match supports "or" patterns and multi-pattern matching
- Trace window increased to 40 events
- Graph-agnostic assertions (has X or Y)

Test Suites:
- smoketest.md: 12 steps covering all categories (~2min)
- fast.md: 10 quick checks (~1min)
- fast_v4.md: 10 v4-eras specific checks
- expert_eras.md: eras domain tests (routing, DB, schema, errors)
- expert_progress.md: progress streaming tests

Other:
- Shared db.py extracted from thinker_v2 (reused by experts)
- InputNode prompt: few-shot examples, history as context summary
- Director prompt: full tool signatures for add_state/reset_machine/destroy_machine
- nginx no-cache headers for static files during development
- Cache-busted static file references

Scores: v3 smoketest 39/40, v4-eras fast 28/28, expert_eras 23/23

Co-Authored-By: Claude Opus 4.6 (1M context) <noreply@anthropic.com>
2026-03-29 17:10:31 +02:00

177 lines
6.3 KiB
Python

"""Expert Base Node: domain-specific stateless executor.
An expert receives a self-contained job from the PA, plans its own tool sequence,
executes tools, and returns a ThoughtResult. No history, no memory — pure function.
Subclasses override DOMAIN_SYSTEM, SCHEMA, and default_database.
"""
import asyncio
import json
import logging
from .base import Node
from ..llm import llm_call
from ..db import run_db_query
from ..types import ThoughtResult
log = logging.getLogger("runtime")
class ExpertNode(Node):
"""Base class for domain experts. Subclass and set DOMAIN_SYSTEM, SCHEMA, default_database."""
model = "google/gemini-2.0-flash-001"
max_context_tokens = 4000
# Override in subclasses
DOMAIN_SYSTEM = "You are a domain expert."
SCHEMA = ""
default_database = "eras2_production"
PLAN_SYSTEM = """You are a domain expert's planning module.
Given a job description, produce a JSON tool sequence to accomplish it.
{domain}
{schema}
Available tools:
- query_db(query, database) — SQL SELECT/DESCRIBE/SHOW only
- emit_actions(actions) — show buttons [{{label, action, payload?}}]
- set_state(key, value) — persistent key-value
- emit_display(items) — formatted data [{{type, label, value?, style?}}]
- create_machine(id, initial, states) — interactive UI with navigation
states: {{"state_name": {{"actions": [...], "display": [...]}}}}
- add_state(id, state, buttons, content) — add state to machine
- reset_machine(id) — reset to initial
- destroy_machine(id) — remove machine
Output ONLY valid JSON:
{{
"tool_sequence": [
{{"tool": "query_db", "args": {{"query": "SELECT ...", "database": "{database}"}}}},
{{"tool": "emit_actions", "args": {{"actions": [{{"label": "...", "action": "..."}}]}}}}
],
"response_hint": "How to phrase the result for the user"
}}
Rules:
- NEVER guess column names. If unsure, DESCRIBE first.
- Max 5 tools. Keep it focused.
- The job is self-contained — all context you need is in the job description."""
RESPONSE_SYSTEM = """You are a domain expert summarizing results for the user.
{domain}
Job: {job}
{results}
Write a concise, natural response. 1-3 sentences.
- Reference specific data from the results.
- Don't repeat raw output — summarize.
- Match the language: {language}."""
def __init__(self, send_hud, process_manager=None):
super().__init__(send_hud)
async def execute(self, job: str, language: str = "de") -> ThoughtResult:
"""Execute a self-contained job. Returns ThoughtResult."""
await self.hud("thinking", detail=f"planning: {job[:80]}")
# Step 1: Plan tool sequence
plan_messages = [
{"role": "system", "content": self.PLAN_SYSTEM.format(
domain=self.DOMAIN_SYSTEM, schema=self.SCHEMA,
database=self.default_database)},
{"role": "user", "content": f"Job: {job}"},
]
plan_raw = await llm_call(self.model, plan_messages)
tool_sequence, response_hint = self._parse_plan(plan_raw)
await self.hud("planned", tools=len(tool_sequence), hint=response_hint[:80])
# Step 2: Execute tools
actions = []
state_updates = {}
display_items = []
machine_ops = []
tool_used = ""
tool_output = ""
for step in tool_sequence:
tool = step.get("tool", "")
args = step.get("args", {})
await self.hud("tool_call", tool=tool, args=args)
if tool == "emit_actions":
actions.extend(args.get("actions", []))
elif tool == "set_state":
key = args.get("key", "")
if key:
state_updates[key] = args.get("value")
elif tool == "emit_display":
display_items.extend(args.get("items", []))
elif tool == "create_machine":
machine_ops.append({"op": "create", **args})
elif tool == "add_state":
machine_ops.append({"op": "add_state", **args})
elif tool == "reset_machine":
machine_ops.append({"op": "reset", **args})
elif tool == "destroy_machine":
machine_ops.append({"op": "destroy", **args})
elif tool == "query_db":
query = args.get("query", "")
database = args.get("database", self.default_database)
try:
result = await asyncio.to_thread(run_db_query, query, database)
tool_used = "query_db"
tool_output = result
await self.hud("tool_result", tool="query_db", output=result[:200])
except Exception as e:
tool_used = "query_db"
tool_output = f"Error: {e}"
await self.hud("tool_result", tool="query_db", output=str(e)[:200])
# Step 3: Generate response
results_text = ""
if tool_output:
results_text = f"Tool result:\n{tool_output[:500]}"
resp_messages = [
{"role": "system", "content": self.RESPONSE_SYSTEM.format(
domain=self.DOMAIN_SYSTEM, job=job, results=results_text, language=language)},
{"role": "user", "content": job},
]
response = await llm_call(self.model, resp_messages)
if not response:
response = "[no response]"
await self.hud("done", response=response[:100])
return ThoughtResult(
response=response,
tool_used=tool_used,
tool_output=tool_output,
actions=actions,
state_updates=state_updates,
display_items=display_items,
machine_ops=machine_ops,
)
def _parse_plan(self, raw: str) -> tuple[list, str]:
"""Parse tool sequence JSON from planning LLM call."""
text = raw.strip()
if text.startswith("```"):
text = text.split("\n", 1)[1] if "\n" in text else text[3:]
if text.endswith("```"):
text = text[:-3]
text = text.strip()
try:
data = json.loads(text)
return data.get("tool_sequence", []), data.get("response_hint", "")
except (json.JSONDecodeError, Exception) as e:
log.error(f"[expert] plan parse failed: {e}, raw: {text[:200]}")
return [], ""