How the trust layer works

1) Any AI connects via open protocols

Business teams ask questions from whichever AI they already use — Claude via MCP, Gemini via A2A, Slack, or any AI assistant. Nodal's trust layer intercepts the request and checks it against your organization's verified query library.

Video not loading? Watch on YouTube

Business question entering Nodal

2) The verified query library returns a trusted answer

Nodal maps questions to verified query templates built from your existing dashboards, notebooks, and SQL. This is your organization's verified query library — the encoded institutional knowledge of how your business measures itself. The answer is consistent whether asked Monday or Friday, in Claude, Gemini, Codex, or Snowflake Cortex, because it comes from the same verified source.

Trusted query matching and verified answer path

3) Unknown questions route to experts — governance, not a limitation

When a question is ambiguous or genuinely new, the trust layer routes it to your data team with full context — not a blank email thread. This is the governance model: analysts review, verify, and approve. The verified answer then becomes part of the query library, available to every AI platform in your organization.

Video not loading? Watch on YouTube

Analyst escalation and review workflow

4) The verified query library compounds

Every question your team asks, every answer your analysts validate, every caveat they add — all of it grows the verified query library. Over time, more questions resolve instantly and fewer require analyst time.

A company using Nodal for a year has a verified query library that reflects a year of real business questions and real analyst decisions. That asset is yours — portable across any AI platform.

Continuous learning loop

5) Protocol-native. Every AI. Every warehouse.

Nodal is built on open protocols — MCP for Claude, A2A for Gemini, API for Codex, plus Snowflake Cortex and Databricks Genie, with more coming. Connect the trust layer once and every current and future AI agent your organization uses gets access to your verified query library. As the AI landscape shifts, your institutional knowledge stays portable, governed, and protocol-native.

Claude (MCP) Gemini (A2A) Any AI (open protocols) Snowflake BigQuery Postgres Redshift dbt Looker Tableau Metabase
Business-user tool integrations Analyst tool integrations

Same question. Very different results.

With Nodal

Trusted answer from validated logic

A business user asks "How does NRR compare between APAC and LATAM?" in Gemini. Nodal retrieves a verified query from a trusted dashboard — the same logic your analysts already use — and returns an answer in seconds.

Video not loading? Watch on YouTube

Faster Retrieves, not explores
Verified Query source
Consistent Same answer every time
Without Nodal

AI explores from scratch every time

The same question asked through Claude Code with only a Snowflake MCP connection. The AI spends dozens of tool calls discovering tables, guessing at schema, and generating SQL — with no guarantee the result matches how your organization defines the metric.

Video not loading? Watch on YouTube

Slower Explores schema from scratch
Generated Query source
Inconsistent Different answer each run

Analytics with the trust layer

With trust layer

Start from your priors, not from scratch

The trust layer gives Claude Code access to your verified queries. Instead of spending dozens of tool calls exploring your schema, it retrieves the right query in milliseconds and builds from there.

Coming soon Analytics with the trust layer in Claude Code
Without trust layer

Stateless exploration every time

Without the trust layer, Claude Code connects directly to Snowflake and starts from zero — discovering tables, inspecting columns, guessing at joins. Slower, more expensive, and the results may not match your team's definitions.

Coming soon Claude Code with only Snowflake MCP

See the trust layer in your stack. One integration. Every AI platform.

Request Demo