Skip to content

Feature: Trust scoring extension for agent-to-agent delegation #1501

@imran-siddique

Description

@imran-siddique

Feature: Trust Scoring Extension for A2A Agent Delegation

Problem

The A2A protocol enables agents to discover and communicate with each other, but there''s no built-in mechanism for trust verification — an agent has no way to assess whether a remote agent is reliable before delegating tasks to it. In multi-agent systems, this creates risks:

  • A newly discovered agent could be unreliable or malicious
  • There''s no feedback loop — an agent that fails repeatedly keeps getting delegated to
  • No trust decay — stale trust from old interactions doesn''t degrade
  • No governance policies controlling which agents can delegate to which

Proposed Solution: Trust Scoring for A2A

An extension to the A2A protocol that adds trust metadata to agent interactions:

from a2a.trust import TrustScore, TrustRegistry

registry = TrustRegistry()

# Before delegation, check trust
target_trust = registry.get_trust("agent://data-analyst")
if target_trust.current() < 0.5:
    # Don''t delegate to untrusted agent, or require human approval
    raise TrustError(f"Agent trust too low: {target_trust.current():.2f}")

# After successful completion, reward
registry.record_success("agent://data-analyst", reward=0.05)

# After failure, penalize
registry.record_failure("agent://data-analyst", penalty=0.15)

# Trust decays over time — requires ongoing good behavior
# If an agent hasn''t been used in a while, trust score naturally decreases

Key Features

  1. Decay-based trust scores — Trust erodes without activity (exponential decay)
  2. Success/failure tracking — Asymmetric rewards (small reward for success, larger penalty for failure)
  3. Agent capability verification — Verify an agent''s AgentCard capabilities match what''s being delegated
  4. Trust thresholds — Configure minimum trust levels for different operation types (read-only vs write operations)
  5. Trust propagation — If Agent A trusts Agent B, and B trusts C, A can compute a transitive trust score for C

How it fits with A2A

The trust extension could be implemented as:

  • AgentCard metadata: Include trust-related fields in the agent card (trust_level, reputation_score)
  • Protocol extension: Add trust verification as an optional A2A message type
  • Client-side library: Trust registry that A2A clients maintain locally

Context

We''ve implemented trust scoring for Agent-OS, PydanticAI, CrewAI, and OpenAI Agents. A2A is the ideal place for trust scoring since it''s specifically about inter-agent communication.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions