How dynamic consensus graphs use formal methods to give every fact a continuously updated confidence score — and why that changes everything about agent governance.
Most knowledge systems treat facts as binary — a thing is either in the database or it isn't. If two sources disagree, the most recent one wins. This is the "newest-wins" pattern, and it's how virtually every knowledge graph, vector store, and memory system works today.
The problem is obvious: newest doesn't mean correct. A single erroneous source uploaded last Tuesday can override ten authoritative sources from last month. In regulated industries — pharma, finance, legal — this isn't a UX annoyance. It's a compliance failure.
Brain takes a different approach. Every relationship in the knowledge graph carries a consensus score — a continuously recalculated confidence value that reflects the weight of all available evidence, not just the most recent write.
For a given claim — a specific entity-property-value triple in the knowledge graph — Brain computes a consensus score based on multiple dimensions:
The result: a single unverified source produces genuine uncertainty — the system doesn't pretend to know. Multiple high-authority sources with expert validation approach high confidence. But the score never reaches absolute certainty — there's always room for new contradictory evidence.
Critically, the scoring uses formal methods from truth discovery research — not ad hoc heuristics. The mathematical foundation ensures consistent, reproducible scores that hold up under audit.
Not all information ages the same way. A pharmaceutical efficacy study from 2024 is still relevant in 2026. A fintech company's ARR figure from 2024 is ancient history.
Brain uses domain-configurable decay rates. Pharmaceutical deployments use slow decay — clinical trial data remains weighted for years. Fintech deployments use fast decay — financial metrics lose relevance in weeks. But old evidence is never completely discarded. It just contributes less weight to the consensus score over time.
This means Brain can detect when a "fact" that was true last year has been contradicted by newer evidence — even if no one explicitly flagged it. The temporal dimension surfaces conflicts that static systems miss entirely.
Here's where consensus scoring becomes a compounding asset rather than a static system.
When a domain expert resolves a conflict — choosing Option A over Option B — Brain doesn't just update the single relationship. It recalibrates the entire model:
A team that resolves 50 conflicts per week isn't just cleaning up 50 facts — they're training the consensus model to increasingly favor reliable sources, trust experienced experts, and discount unreliable inputs. Over time, auto-resolution accuracy improves, noise decreases, and the system surfaces fewer but higher-signal conflicts.
This is the compounding dynamic that separates infrastructure from tooling. The more your team uses Brain, the smarter the truth layer gets. Every resolution is an investment that pays dividends across the entire knowledge graph.
One of the more powerful consequences of continuous consensus scoring is the ability to detect emergent conflicts — contradictions that only become visible when two high-confidence claims collide.
Example: Source A says a company's ARR is $10M with strong evidence backing. Source B, uploaded three months later, says ARR is $8M — and it has even stronger source authority. Neither claim was individually suspicious when ingested. But the consensus engine detects that the same entity now has two well-supported contradictory values.
This is a conflict that no LLM-based system would detect. An LLM asked "what is the company's ARR?" would just pick one value and present it confidently. Brain surfaces the conflict, scores both claims, and either auto-resolves (if one side clearly dominates) or escalates to the most qualified expert for that domain.
Not all conflicts require human attention. Brain uses a tiered autonomy model that's configurable per domain:
A pharmaceutical deployment might require human review for nearly everything — patient safety demands it. A general knowledge management deployment might auto-resolve most conflicts and only escalate genuinely contested claims. The thresholds are yours to set.
Adjust the inputs below to see how consensus changes. This is a conceptual model — the actual scoring engine uses additional dimensions — but it illustrates the key dynamics.
When you combine consensus scoring with deterministic conflict detection, you get something that doesn't exist anywhere else in the AI infrastructure stack: a continuously self-improving source of truth that agents can query before taking action.
The Query Gate API (POST /gate/query) returns one of three verdicts:
This is deterministic governance. Not a system prompt that can be overridden. Not an LLM checking another LLM's work. A formal query against a consensus-scored graph that returns the same verdict every time, given the same state.
And because the consensus model improves with every resolution, the system gets better at making these verdicts over time. More resolutions → better source authority scores → more accurate auto-resolution → fewer false positives → less noise → more trust.
That's the math behind the truth layer.
Watch Brain detect conflicts, score claims, and resolve contradictions in a live knowledge graph.
Get a Demo