Show & Tell — Phase 29.3 MetaCognitiveMonitor: Bias Detection & Confidence Calibration #636
web3guru888
started this conversation in
Show and tell
Replies: 0 comments
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Show & Tell — Phase 29.3 MetaCognitiveMonitor: Bias Detection & Confidence Calibration
Overview
The MetaCognitiveMonitor implements the monitoring level of Nelson & Narens' (1990) two-level metacognitive framework. While the IntrospectionEngine (29.2) observes what the system thinks, the MetaCognitiveMonitor evaluates how well it thinks — detecting cognitive biases, calibrating confidence, and generating thinking quality reports.
This component draws on decades of research in judgment and decision-making: Tversky & Kahneman's (1974) heuristics and biases program, Lichtenstein et al.'s (1982) calibration studies, Stanovich & West's (2000) individual differences in reasoning, and Gigerenzer et al.'s (1999) ecological rationality.
Architecture: Nelson & Narens Monitoring-Control Loop
Bias Detection Pipeline
The bias detector operates on thought chains from the IntrospectionEngine (29.2):
Eight Bias Types
CONFIRMATIONANCHORINGAVAILABILITYSUNK_COSTFRAMINGOVERCONFIDENCEGROUPTHINKRECENCYBrier Score Calibration
Confidence calibration uses the Brier score (Brier, 1950), the gold standard for evaluating probabilistic predictions:
Where
confidence_i ∈ [0, 1]is the system's stated confidence andoutcome_i ∈ {0, 1}is the actual binary outcome. A perfectly calibrated system has Brier score = reliability component ≈ 0.The calibrator maintains a rolling window of (confidence, outcome) pairs:
Why Brier Score Over Log-Loss?
Both are proper scoring rules, but Brier score has advantages for metacognitive monitoring:
ThinkingQualityReport
The
QualityReportergenerates periodic composite reports:Integration with ExecutiveController (Phase 28.4)
The MetaCognitiveMonitor feeds directly into the ExecutiveController's planning loop:
Prometheus Metrics
asi_metacog_brier_scoreasi_metacog_bias_alerts_totalbias_typeasi_metacog_quality_compositeasi_metacog_intervention_totaltypeasi_metacog_calibration_curvebinOpen Questions
Bias detection sensitivity — How do we tune the detection thresholds to minimize false positives without missing real biases? Adaptive thresholds (e.g., based on running baseline) may be better than fixed ones, but add complexity.
Intervention authority — Should the MetaCognitiveMonitor have the authority to directly intervene (e.g., force a re-search), or should it only recommend interventions to the ExecutiveController (28.4)? The current design recommends only, but strong biases might warrant automatic circuit-breaking.
Cross-session calibration — The Brier score window resets on restart. Should calibration history be persisted to the TemporalGraph (Phase 17.1) for cross-session continuity? This would enable long-term calibration trend analysis.
References: Nelson & Narens (1990), Tversky & Kahneman (1974), Brier (1950), Murphy (1973), Lichtenstein et al. (1982), Stanovich & West (2000), Gigerenzer et al. (1999)
Beta Was this translation helpful? Give feedback.
All reactions