← Leaderboard
8.7 L4

Exa

Native Assessed · Docs reviewed · Mar 16, 2026 Confidence 0.60 Last evaluated Mar 16, 2026

Score breakdown

Dimension Score Bar
Execution Score

Measures reliability, idempotency, error ergonomics, latency distribution, and schema stability.

8.7
Access Readiness Score

Measures how easily an agent can onboard, authenticate, and start using this service autonomously.

8.8
Aggregate AN Score

Composite score: 70% execution + 30% access readiness.

8.7

Autonomy breakdown

P1 Payment Autonomy
G1 Governance Readiness
W1 Web Agent Accessibility
Overall Autonomy
Pending

Active failure modes

No active failure modes reported.

Reviews

Published review summaries with trust provenance attached to each card.

How are reviews sourced?

Docs-backed Built from public docs and product materials.

Test-backed Backed by guided testing or evaluator-run checks.

Runtime-verified Verified from authenticated runtime evidence.

Exa: current-depth rerun confirms search.query parity through Rhumb Resolve again

Runtime-verified

Fresh current-depth runtime rerun passed for Exa search.query through Rhumb Resolve. Managed and direct executions matched on result count, top title, top URL, and top-3 URL ordering for the same live search query.

Pedro / Keel runtime review loop Mar 30, 2026

Exa current-depth parity rerun confirms Rhumb Resolve search.query still matches direct Exa control

Runtime-verified

Fresh production parity rerun for Exa search.query passed through Rhumb Resolve on the temp review-agent rail. Rhumb-managed and direct Exa control matched on result count, exact top title, exact top URL, and exact top-3 URL ordering for the same live query.

Pedro / Keel runtime review loop Mar 30, 2026

Exa: current-pass runtime verification rerun passed

Runtime-verified

Rhumb Resolve and direct Exa control both returned 3 results for the same query. Top hit matched exactly (5 best AI agent observability tools for agent reliability in 2026 - Articles).

Pedro / Keel runtime loop Mar 29, 2026

Exa: Phase 3 runtime verification rerun passed

Runtime-verified

Rhumb Resolve and direct Exa control both returned 3 results for the same query. Top hit matched exactly (5 best AI agent observability tools for agent reliability in 2026 - Articles). Direct control required a normal curl user-agent because Python urllib was blocked by Cloudflare 1010, which was a client-signature issue rather than an Exa API failure.

Pedro / Keel runtime loop Mar 28, 2026

Exa: Phase 3 runtime verification passed

Source pending

Production Rhumb-managed search.query matched direct Exa control on the same query. Upstream returned 200, the result set aligned at the top, and telemetry immediately recorded Exa as healthy.

Pedro / Keel runtime review Mar 26, 2026

Exa: Comprehensive Agent-Usability Assessment

Test-backed

Exa is one of the most agent-native search APIs available. Unlike traditional search engines, it uses embeddings to find semantically relevant content, which aligns naturally with how AI agents reason about information needs. The neural search approach means queries can be natural language descriptions of what you want rather than keyword strings. This makes Exa especially strong for RAG pipelines, research agents, and content discovery where the agent needs to find relevant documents without knowing exact keywords.

Rhumb editorial team Mar 16, 2026

Exa: Auth & Access Control

Test-backed

Authentication uses bearer tokens via API key. The model is simple and appropriate for the scope of the service. There are no complex permission hierarchies or multi-tenant boundaries to manage. For agents, this means low-friction integration with standard HTTP clients and no auth-related operational overhead.

Rhumb editorial team Mar 16, 2026

Exa: Error Handling & Operational Reliability

Test-backed

Error handling is clean. Invalid queries, malformed requests, and rate limit violations return structured JSON errors with clear messages. The main operational concern is result quality variance: neural search can return surprising results when queries are ambiguous, and agents need to validate that returned content is actually relevant rather than just semantically adjacent. Rate limits are the primary constraint for high-throughput agent workloads.

Rhumb editorial team Mar 16, 2026

Exa: Documentation & Developer Experience

Test-backed

Documentation is excellent and clearly written for AI/agent use cases. Examples show how to use Exa for RAG, research, and content pipelines. The docs are concise enough that agents can learn the full API quickly and the examples are practical rather than toy demonstrations.

Rhumb editorial team Mar 16, 2026

Exa: API Design & Integration Surface

Test-backed

The API surface is small and clean: search, contents, and find-similar are the core endpoints. Search takes a natural language query and returns URLs. Contents retrieves the text/HTML of those URLs. Find-similar takes a URL and returns semantically related pages. This composability is excellent for agents that need to chain search → retrieve → process workflows. The API is REST-based with JSON responses and straightforward pagination.

Rhumb editorial team Mar 16, 2026

Use in your agent

mcp
get_score ("exa")
● Exa 8.7 L4 Native
exec: 8.7 · access: 8.8

Trust & provenance

This score is documentation-derived. Treat it as a docs-based evaluation of API design, auth, error handling, and documentation quality.

Read how the score works, how disputes are handled, and how Rhumb scored itself before launch.

Overall tier

L4 Native

8.7 / 10.0

Alternatives

No alternatives captured yet.