You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/index.md
+5-5Lines changed: 5 additions & 5 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -18,15 +18,15 @@ feature_text: |
18
18
19
19
{% include button.html text="Paper" link="" color="#4f2121" %} {% include button.html text="Code" link="https://github.com/hpai-bsc/pgeon" color="#113662" %} {% include button.html text="BlueSky" link="https://bsky.app/profile/hpai.bsky.social" color="rgb(32, 139, 254)" %} {% include button.html text="Linkedin" link="https://www.linkedin.com/company/hpai" color="#0a66c2" %}
20
20
21
-
###Abstract
21
+
# Abstract
22
22
23
23
> Agents are a special kind of AI-based software in that they interact in complex environments and have increased potential for emergent behaviour.
24
24
Explaining such behaviour is key to deploying trustworthy AI, but the increasing complexity and opaque nature of many agent implementations makes this hard.
25
25
In this work, we reuse the Policy Graphs method --which can be used to explain opaque agent behaviour-- and enhance it to query it with hypotheses of desirable situations. These hypotheses are used to compute a numerical value to examine agent intentions at any particular moment, as a function of how likely the agent is to bring about a hypothesised desirable situation. We emphasise the relevance of how this approach has full epistemic traceability, and each belief used by the algorithms providing answers is backed by specific facts from its construction process.
26
26
We show the numeric approach provides a robust and intuitive way
27
27
to provide telic explainability (explaining current actions from the perspective of bringing about situations), and allows to evaluate the interpretability of behaviour of the agent based on the explanations; and it opens the door to explainability that is useful not only to the human, but to an agent.
0 commit comments