HDP - cryptographic chain-of-custody for multi-agent delegation #7485
Replies: 4 comments 6 replies
-
|
The delegation-chain integrity problem is real and measurable. We run a test suite that specifically targets multi-agent delegation — injecting rogue instructions mid-chain, forging orchestrator messages, and testing whether downstream agents can distinguish legitimate delegations from spoofed ones. Short answer: without something like what you're describing, they can't. A few observations from testing that might be useful: What breaks without provenance: Our "confused deputy" tests inject adversarial instructions into a sub-agent's context that mimic orchestrator-style messages. In every framework we've tested (AutoGen, CrewAI, LangGraph), the sub-agent follows the injected instruction. There's no cryptographic or structural way to verify "this instruction actually came from my orchestrator." HDP's signed delegation chain would close this. The timestamp question: Your Verification cost: One thing to consider for AutoGen specifically — verification needs to be cheap enough that it doesn't become a bottleneck in high-throughput group chats. Have you benchmarked Ed25519 verification overhead per message in a typical AutoGen Our test findings are in |
Beta Was this translation helpful? Give feedback.
-
|
the Ed25519 chain approach makes sense for the core problem - current autogen setups have no way to distinguish a legitimate orchestrator message from a spoofed one once it's in the conversation history. everything looks like plaintext to the agents downstream. a few questions on autogen-specific dynamics: dynamic agent spawning - autogen can spawn agents at runtime based on task requirements. how does HDP handle the case where the spawning decision itself happens mid-chain? the human authorized "do X" but the orchestrator decides that X requires creating a new specialized agent. does that spawn operation get a derived token from the parent scope, or does it require a new human authorization? nested conversation termination - autogen's GroupChatManager can terminate and restart sub-conversations. if a sub-agent chain completes and a new one starts as part of the same task, does HDP treat that as a continuation (same session ID) or a new delegation that needs fresh authorization? scope declaration format - what does a scope string look like in practice? if an agent is authorized to "retrieve and summarize documents", does HDP validate that a file deletion call is out of scope, or is scope just informational metadata in the token? the IETF draft angle is interesting - standardized provenance tokens would make cross-framework delegation (autogen calling a crewai sub-agent) auditable in a way that nothing currently handles. that's probably the most underrated use case here. |
Beta Was this translation helpful? Give feedback.
-
|
This connects to the broader trust-boundary testing gap we've been working on. We just published a detailed breakdown of what fails across multi-agent frameworks when you test delegation chains adversarially: https://dev.to/mspro3210/agent-systems-are-failing-at-trust-boundaries-we-ran-332-tests-to-prove-it-5cod The short version: delegation handoffs are where trust assumptions break down. In default configurations across AutoGen, CrewAI, and LangGraph, context leaks across agent boundaries during delegation. The HDP approach of Ed25519-signed delegation tokens addresses the provenance side, but the behavioral side (does the downstream agent actually respect scope constraints?) still needs adversarial testing. The harness we discussed in #7432 now covers 332 tests across 24 modules, including delegation chain depth attacks and scope escalation. Happy to run the D004 module against an HDP-wrapped delegation flow if you want to see how the cryptographic chain holds up under adversarial conditions. |
Beta Was this translation helpful? Give feedback.
-
|
Cryptographic chain-of-custody for delegation is a great direction. I've been building something that complements this — proof-of-behavior for the actions within each delegation step. The chain-of-custody tells you who delegated to whom. Proof-of-behavior tells you what each agent did with that delegation. Together: who had authority + what they did with it + cryptographic evidence for both. The implementation: agents declare behavioral constraints, every action gets evaluated before execution, and decisions go into a SHA-256 hash chain. Each entry links to the previous — tamper with one and the chain breaks. For AutoGen multi-agent workflows, this could work per-agent: each agent in the pipeline gets its own behavioral covenant and its own hash-chained log. At the end of the workflow, you have the delegation chain (HDP) and the behavioral evidence for each step (proof-of-behavior). Open-source: github.com/arian-gogani/nobulex |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
Hi Autogen Community 👋,
been hitting a consistent gap when building multi-agent pipelines, when a task delegates across multiple agents, there's no standard way to verify that downstream actions were actually authorized by the originating human. a compromised tool or rogue sub-agent mid-chain can inject instructions that look identical to legitimate orchestrator messages. no forensic trail, no way to know where the chain broke.
built HDP (Human Delegation Provenance) to address this. every delegation hop is Ed25519 signed and encoded in a self-contained token. verification is fully offline, no registry, no network call, just a public key and a session ID. if the chain breaks, you know exactly where and at which hop.
what's shipped:
@helixar_ai/hdp, npm)configure(crew)calldraft-helixar-hdp-agentic-delegation-00(RATS WG)design decision worth flagging: HDP is a provenance layer, not enforcement. it records that a human authorized an action with a declared scope and traces the chain. what you do with a violation is up to the application.
AutoGen's architecture is exactly the kind of multi-agent system this was designed for and I'd genuinely value feedback from people building complex pipelines here on whether the model holds up, particularly around dynamic agent spawning and nested task delegation.
github: https://github.com/Helixar-AI/HDP
Beta Was this translation helpful? Give feedback.
All reactions