The trust scoring layer is the interesting part here. The agent ecosystem has a discovery problem and a trust problem, and most tools today only tackle discovery. Being able to evaluate reliability before you connect changes the calculus.
Curious how the trust score works in practice. Is it purely automated test results, or do you plan to incorporate usage signals over time (uptime, response quality)?
Right now it's purely automated,50+ compliance checks against the A2A spec (agent card validation, endpoint testing, state machine, streaming, auth, error handling). Each check is weighted and rolled into the 0-100 score.
But you're right that automated spec compliance only tells part of the story. The roadmap includes usage signals, uptime monitoring, response latency tracking, and community ratings from developers who've actually integrated with an agent. The spec tells you if an agent CAN work. Usage data tells you if it DOES work.
The profile pages are designed with that in mind, test history over time already shows trends, and adding real world signals is the natural next layer.
Curious how the trust score works in practice. Is it purely automated test results, or do you plan to incorporate usage signals over time (uptime, response quality)?
But you're right that automated spec compliance only tells part of the story. The roadmap includes usage signals, uptime monitoring, response latency tracking, and community ratings from developers who've actually integrated with an agent. The spec tells you if an agent CAN work. Usage data tells you if it DOES work.
The profile pages are designed with that in mind, test history over time already shows trends, and adding real world signals is the natural next layer.