The Deep-Tech Partner for AI, Simulation & Robotics

AI for Orebody Knowledge & Uncertainty Reduction [Part 3 of 3] - Uncertainty Quantification in AI-Generated Orebody Models: From Monte Carlo to Bayesian Neural Networks

Woody de Kafou | January 11, 2026

Mine-techAI Powered SimulationComputer VisionGenerative AIMachine Learning

Every mining executive has sat through a presentation showcasing an AI model predicting ore locations with "97% accuracy" – only to ask the question that kills the demo: "What about the 3%? And more importantly, where is that 3%?" Point predictions without uncertainty quantification are worse than useless in mining; they're dangerous. A mine plan based on confidently wrong AI predictions can destroy project economics before the first ore ton is processed. The industry learned this lesson decades ago with kriging – variability isn't a nuisance, it's fundamental geological reality that must be quantified, propagated through economic models, and managed via production strategy.

Now AI-generated orebody models promise step-change improvements in resource estimation, but they face a credibility gap: can machine learning provide the rigorous uncertainty bounds required for NI 43-101 or JORC compliance? The answer is yes – but it requires moving beyond naive neural networks to probabilistic architectures designed specifically for uncertainty quantification. Mining companies deploying these methods are already achieving uncertainty estimates that meet or exceed traditional geostatistical standards while handling geological complexity that breaks traditional approaches.

In this article, we explore the uncertainty quantification challenge, compare modern ML methods with traditional geostatistics, and present a practical framework for integrating AI uncertainty into resource estimation and mine planning.


The Uncertainty Problem: Why Mining Needs More Than Point Predictions

Traditional geostatistics has always recognized two uncertainty types:

Epistemic uncertainty (reducible ignorance): Uncertainty due to limited data. Add more drill holes and epistemic uncertainty decreases. This is what kriging variance estimates – how uncertain are we about the true grade at unsampled locations, given current data density?

Aleatoric uncertainty (irreducible randomness): Uncertainty due to inherent geological variability. Even with perfect information, ore grades vary meter-to-meter due to complex mineralization processes. No amount of additional sampling eliminates aleatoric uncertainty – you can only characterize it more precisely.

Mining companies need both quantified because they drive different decisions: epistemic uncertainty guides where to drill next (target high-uncertainty areas to maximize information gain), while aleatoric uncertainty determines production blending strategies and mill feed variability expectations.

Most machine learning models provide neither – they output a single grade prediction with no confidence intervals, no probabilistic distributions, and no distinction between data-limited uncertainty versus fundamental geological variability. This makes AI predictions unusable for mine planning, infeasible for compliance reporting, and unsuitable for risk analysis. The breakthrough: probabilistic ML architectures that match or exceed traditional geostatistical uncertainty quantification while handling geological complexity that traditional methods miss.


Traditional Geostatistical Approaches: Kriging and SGS

Kriging: Optimal Interpolation with Variance Estimates

Kriging remains the industry standard for a reason – it provides Best Linear Unbiased Predictions (BLUP) under stationarity assumptions, and critically, it quantifies prediction uncertainty via kriging variance. A kriging estimate isn't just "this block grades 2.4 g/t Au" – it's "this block grades 2.4 ± 0.6 g/t Au at 95% confidence, and variance is driven primarily by distance to nearest drill hole and grade continuity along this structural corridor."

Kriging variance has well-understood limitations: it only accounts for spatial configuration of data, not actual sample values. A block surrounded by 1.0 g/t samples has the same kriging variance as one surrounded by 10.0 g/t samples if the spatial geometry is identical. This means kriging variance underestimates uncertainty in high-grade zones and overestimates in low-grade areas – precisely where mine planning needs accurate uncertainty bounds.

Implementation is mature and well-understood – industry-standard software (Datamine, Leapfrog, Surpac), established workflows, and regulatory acceptance (QPs can sign off on kriging-based resources). For disseminated deposits in stationary geology with abundant drilling, kriging performs well and AI struggles to improve on it.

Sequential Gaussian Simulation: Capturing Spatial Uncertainty

Sequential Gaussian Simulation (SGS) generates multiple equally-probable realizations of an orebody, all honoring sample data and variogram-defined spatial continuity. Rather than a single estimate with variance, SGS provides 100-500 alternative models – each a possible "true" orebody that fits available data.

The power of SGS for mine planning: run each realization through your mine optimizer, production scheduler, and economic model. The distribution of NPVs quantifies project risk in a way single estimates cannot capture. You might discover: "70% of realizations are highly profitable, 20% are marginal, and 10% are uneconomic – and the uneconomic scenarios cluster in settings where high-grade shoots pinch out before reaching our planned mining horizons."

Limitations become apparent in complex geology – SGS assumes multi-Gaussian distributions (real ore often log-normal or heavily skewed), stationarity (geology changes across district-scale systems), and isotropic or simple anisotropic continuity (real structural controls are complex). When these assumptions fail, SGS generates geologically implausible realizations that may still honor data and variograms but violate geological understanding.


Modern ML Uncertainty Methods: Beyond Point Predictions

Ensemble Methods: Diverse Models, Quantified Disagreement

The conceptually simplest approach: train 50-100 different neural networks (varying architectures, hyperparameters, initialization), generate predictions from each, and treat their disagreement as uncertainty. High disagreement across ensemble members indicates uncertain predictions; low disagreement indicates confident predictions.

Ensemble uncertainty correlates with epistemic uncertainty – models disagree more in data-sparse regions where training signal is weak, and agree more in well-sampled regions where data constrains predictions. This makes ensembles useful for drilling optimization: target areas where model disagreement is high, indicating the next drill hole will maximally reduce model uncertainty.

Practical challenges include computational cost (training 100 models costs 100× a single model) and tendency toward overconfidence (if all ensemble members make the same systematic error, disagreement is low but predictions are wrong). Modern solutions use efficient ensemble methods – train a single network, then generate diverse predictions via dropout variation, weight space sampling, or hyperparameter sampling. This reduces cost 10-50× while maintaining diversity.

Monte Carlo Dropout: Uncertainty from Regularization

Dropout – randomly disabling neurons during training – was designed for regularization, but it also enables uncertainty estimation. The insight: dropout defines an implicit ensemble of sub-networks. Activating dropout during inference generates different predictions each time, and prediction variability estimates uncertainty.

MC Dropout implementation is trivial: add dropout layers to your network, train normally, then at inference time keep dropout active and run 50-100 forward passes per prediction. The distribution of predictions provides a principled uncertainty estimate – wider distributions indicate higher uncertainty, narrower distributions indicate confidence.

Calibration is critical but often overlooked – MC Dropout tends toward underconfident predictions (too wide uncertainty intervals) because dropout rate during inference affects uncertainty width independently of actual prediction quality. Calibration techniques (temperature scaling, quantile regression) ensure that "90% confidence intervals contain the true value 90% of the time" – matching predicted uncertainty to actual error rates.

Bayesian Neural Networks: Principled Probabilistic Predictions

Bayesian Neural Networks (BNNs) treat network weights as probability distributions rather than point values. Instead of learning "this weight equals 0.732," a BNN learns "this weight is normally distributed with mean 0.732 and standard deviation 0.043." Predictions then integrate over this weight distribution, naturally producing probabilistic outputs.

Theoretical advantages are compelling: BNNs provide calibrated uncertainty by construction, distinguish epistemic from aleatoric uncertainty, and avoid overconfidence through proper regularization. They're also sample-efficient – learning weight uncertainty means the model knows what it doesn't know, enabling active learning and targeted data collection.

Practical implementation remains challenging – exact Bayesian inference is intractable for neural networks, so approximations are required. Variational inference (optimizing a tractable distribution to approximate the true posterior), Monte Carlo methods (sampling weight configurations), and Laplace approximation (second-order Taylor expansion around the MAP estimate) all involve trade-offs between accuracy and computational cost.

Recent advances make BNNs practical for resource estimation: stochastic weight averaging (SWA), deep ensembles with explicit uncertainty prediction, and normalizing flow-based posteriors reduce computational costs to 2-5× standard training. For a mining project justifying $50M in drilling, spending $100K on better uncertainty quantification is trivially justified.

Conformal Prediction: Distribution-Free Uncertainty Bounds

Conformal prediction provides uncertainty intervals without assumptions about data distribution – no normality, no stationarity, no parametric models. The approach: hold out a calibration set, compute prediction errors on calibration data, then construct prediction intervals that include the true value for the desired fraction of calibration examples.

The guarantee is elegant: if you want 90% prediction intervals, conformal prediction delivers intervals that contain the true grade for 90% of samples – regardless of whether your model is neural network, random forest, or linear regression. This distribution-free property makes conformal prediction robust to model misspecification.

Practical advantages include trivial implementation (wrap any existing model, no retraining required), finite-sample guarantees (works even with limited calibration data), and adaptive uncertainty (intervals automatically widen in high-uncertainty regions, narrow in low-uncertainty regions). Limitations include conservative predictions (intervals may be wider than theoretically necessary) and challenges with covariate shift (if calibration and deployment geology differ, guarantees weaken).


Comparing Approaches: Traditional Geostatistics vs. Modern ML

The table below summarizes key differences:

Aspect

Kriging / SGS

ML Ensembles

Bayesian Neural Networks

Conformal Prediction

Uncertainty Type

Epistemic (kriging variance) + Aleatoric (SGS variability)

Primarily epistemic (model disagreement)

Both, distinguished

Both, mixed

Assumptions

Stationarity, Gaussian, variogram-defined continuity

Minimal (depends on architecture)

Minimal

None (distribution-free)

Computational Cost

Low (kriging) to Moderate (SGS)

Moderate (100 models)

Moderate to High (MCMC, VI)

Low (post-hoc calibration)

Geological Complexity

Struggles with non-stationary, complex structures

Handles complex nonlinear relationships

Handles complexity + principled uncertainty

Handles complexity, conservative

Calibration

Often poorly calibrated (kriging variance ≠ true uncertainty)

Requires manual calibration

Well-calibrated by design

Calibrated by construction

Regulatory Acceptance

Established (NI 43-101, JORC)

Emerging

Emerging

Emerging

Interpretability

High (variance maps, correlograms)

Moderate (why do models disagree?)

Low (posterior is complex)

Moderate (quantile-based)

Practical synthesis: Use traditional geostatistics for disseminated deposits in stationary geology with abundant drilling – it's proven, accepted, and sufficient. Deploy ML methods for structurally controlled deposits, sparse data settings, or properties with multi-modal data (geophysics, geochemistry, structure) that traditional methods can't integrate effectively. For compliance, generate both traditional and AI estimates, then submit the approach that best captures geological reality – increasingly, regulators accept ML if accompanied by rigorous validation and QP oversight.


The Regulatory Question: Can AI Meet NI 43-101 and JORC Standards?

The uncomfortable truth: current regulatory frameworks don't explicitly address AI-generated resource estimates. NI 43-101 requires "appropriate techniques" applied by "Qualified Persons"; JORC requires "techniques appropriate to the deposit style." Neither specifies kriging – they specify that methods must be fit for purpose, properly validated, and applied by competent practitioners.

Emerging regulatory guidance suggests AI models can achieve compliance if:

  1. Validation is rigorous: Cross-validation against hold-out drill data, comparison with traditional methods, geological plausibility checks, and reconciliation with production data (for brownfield assets).
  2. Uncertainty is quantified: Not just point predictions – probabilistic distributions, confidence intervals, or multiple realizations that support economic analysis and risk assessment.
  3. QP oversight is maintained: A Qualified Person must understand the AI method, validate its application, and take responsibility for results. QPs can't delegate judgment to black-box algorithms.
  4. Limitations are disclosed: AI models trained on certain deposit types may not generalize to others; epistemic uncertainty may be underestimated in extrapolated regions; specific assumptions (e.g., data normalization, geological domain boundaries) affect results.

Progressive jurisdictions (Australia, Canada) are accepting AI-enhanced estimates for early-stage resources (Inferred, Indicated) with traditional geostatistics required for Measured resources and reserves. This pragmatic approach lets companies leverage AI where uncertainty is already high (exploration) while requiring proven methods for production planning (reserves).

Forward-looking companies are positioning AI as complementary: "Our resource estimate uses kriging for compliance, but we also generated 500 AI realizations integrating geophysical constraints – these show 80% probability of resource expansion in the eastern fault block, guiding infill drilling prioritization." This frames AI as a strategic tool without triggering regulatory concerns about non-standard estimation.


Practical Framework: Integrating AI Uncertainty into Mine Planning

Step 1: Generate Probabilistic Realizations

Whether using ensembles, BNNs, or conformal prediction, generate 100-1000 distinct orebody models – each a geologically plausible configuration consistent with available data but exploring different scenarios (grade continuity, structural interpretation, alteration extent).

Quality checks are essential: validate that realizations honor drill data exactly (not approximately – exact), span geologically plausible configurations (no mineralization floating in air, no structures violating crosscutting relationships), and collectively reproduce input data statistics (mean, variance, spatial continuity).

Step 2: Propagate Through Economic Models

Run each realization through your mine planning workflow: pit optimization, production scheduling, metallurgical recovery modeling, cost estimation. Generate NPV distributions, production profiles, and equipment requirements for each scenario.

This reveals risks invisible in single-estimate plans: maybe 90% of realizations are profitable, but 10% lose money because high-grade shoots pinch out before reaching economic mining depth. Or maybe base-case profitability is $200M NPV, but upside scenarios reach $400M if grade extends into the fault block, while downside scenarios fall to $50M if sulfide content is lower than expected.

Step 3: Optimize for Robustness

Rather than optimizing plans for the single "best estimate" orebody, optimize for robust performance across the ensemble. This might mean: designing pits that capture value in 80% of scenarios (sacrificing some upside in high-grade scenarios for less downside in low-grade ones), scheduling production to blend variable ore (maintaining consistent mill feed across geological uncertainty), or staging development to preserve optionality (deferring capital until delineation drilling reduces uncertainty).

Risk-adjusted optimization trades expected value for reduced variance – a plan with $180M expected NPV but low variance might be preferable to one with $200M expected NPV but $150M standard deviation. This mirrors how mining companies actually think: avoid catastrophic scenarios even if it means giving up some upside.

Step 4: Active Learning and Iterative Refinement

Use uncertainty maps to prioritize infill drilling: drill where epistemic uncertainty is highest (maximum information gain) or where uncertainty most impacts project economics (maximum decision value). After each drill hole, update AI models, regenerate realizations, re-run economic analysis, and refine plans.

Value of information analysis quantifies this explicitly: estimate how much each potential drill hole reduces uncertainty (information value) and how that uncertainty reduction affects project NPV (decision value). Drill holes with highest decision value per dollar get prioritized – often not the holes targeting highest grade, but those resolving critical geological questions affecting mine design.


What Mining Companies Should Do Now

  1. Establish Uncertainty Requirements: Define what uncertainty metrics you need for different decision types. Exploration targeting needs epistemic uncertainty (where should we drill?). Mine planning needs full distributions (what's the NPV distribution?). Production scheduling needs aleatoric uncertainty (how variable is mill feed?). Match ML methods to specific needs.
  2. Benchmark Against Traditional Methods: Generate both AI and geostatistical estimates for a property with abundant drilling. Compare uncertainty width, coverage (do 90% confidence intervals actually contain 90% of truth?), and geological plausibility. This establishes baseline performance and builds internal confidence.
  3. Integrate into Workflows Incrementally: Don't replace your existing resource estimation overnight. Start with AI-enhanced targeting (using uncertainty maps for drill prioritization), graduate to dual estimation (both traditional and AI for comparison), eventually adopt AI as primary method where it demonstrably outperforms traditional approaches.
  4. Invest in QP Training: Your Qualified Persons need to understand AI fundamentals, uncertainty quantification principles, and validation strategies. Partner with universities offering courses in applied ML for geoscientists, or develop internal training programs. QPs who understand AI will advocate for it; QPs who see it as a black box will resist adoption regardless of performance.

Conclusion

Uncertainty quantification transforms AI-generated orebody models from interesting research to bankable resource estimates. The technology exists – Bayesian neural networks, conformal prediction, and well-calibrated ensembles provide uncertainty bounds matching or exceeding traditional geostatistics while handling geological complexity that breaks traditional methods. The regulatory path is clarifying – rigorous validation, QP oversight, and explicit uncertainty quantification enable compliance in progressive jurisdictions.

Mining companies investing in probabilistic AI now will be positioned to make better-informed decisions under uncertainty – optimal drilling programs, robust mine plans, and realistic risk assessments – while competitors continue relying on single-estimate thinking that ignores fundamental geological variability. The question isn't whether AI can quantify uncertainty well enough for mining. The question is whether your competitors will adopt it before you do.

30 minute technical discovery

Woody de Kafou

Woody de Kafou

Founder & CEO, AI Thought Leader

BOOK A MEETING