The Deep-Tech Partner for AI, Simulation & Robotics

AI for Orebody Knowledge & Uncertainty Reduction [Part 2 of 3] -Large World Models for Subsurface Intelligence: Teaching AI the Physics of Mineral Systems

Woody de Kafou | November 25, 2025

Mine-techGenerative AIAI Powered Simulation

Every decade or so, a breakthrough redefines mineral exploration. The 1960s brought portable XRF analyzers; the 1980s saw GPS revolutionize field mapping; the 2000s introduced 3D inversion of geophysical data. Now an even bigger shift looms: Large World Models (LWMs) trained specifically for subsurface intelligence. These AI systems don't just classify lithology from drill logs or predict grades from geochemistry – they understand 3D geological relationships, structural controls, and multi-stage alteration patterns the way an expert exploration geologist does. Think of moving from 2D pattern recognition to true spatial reasoning. Already, major mining companies are assembling "subsurface intelligence" AI teams, and junior explorers using early LWM prototypes report 60% reductions in time from discovery to preliminary economic assessment.

In this article, we explore what subsurface LWMs are, why traditional AI falls short for geological problems, and how mining companies can prepare for this paradigm shift.


What Are Large World Models for Subsurface Intelligence?

Traditional machine learning models for geology operate on flattened data – a drill hole becomes a 1D sequence of lithology codes and assays, a geological map becomes a 2D image, geochemical data becomes a table of element concentrations. These representations discard critical information: geometry, topology, and spatial relationships. A vein dipping 60° southeast behaves differently than one dipping 30° northwest, even if both have identical gold grades. Fault offsets create displacement that traditional AI can't reason about. Crosscutting relationships encode relative timing that flat data representations miss.

Large World Models, by contrast, aim to model the geological world itself – ingesting multi-modal inputs (drill core logs, geophysics, geochemistry, structural measurements, remote sensing) and generating an interactive 3D understanding of subsurface mineral systems. Instead of just predicting "high gold potential here," an LWM can reason: "This structure intersects favorable host rock at this depth, the alteration assemblage indicates 300°C fluids, the geochemical footprint shows a proximal source, and the geophysical signature suggests continuation 500m east – drill here."

The paradigm shift: from task-specific models (one for lithology prediction, another for grade estimation, another for targeting) to a unified foundation model that understands geological context across all data types and can tackle arbitrary exploration tasks through prompt engineering or fine-tuning.


Why Spatial Transformers and Geometric Deep Learning Matter

Standard neural networks struggle with 3D geological data because they lack built-in understanding of geometric relationships. A convolutional network trained to recognize alteration patterns in one orientation fails when the geology rotates – it hasn't learned that mineralization is invariant to coordinate system choice, only to geological reference frames.

Spatial Transformers: Handling Anisotropy

Narrow vein systems present a perfect example of why geometry matters. A quartz vein might extend 2km along strike, continue 800m down dip, but measure only 2m wide. Traditional 3D convolutions treat all directions equally – they blur the vein into surrounding wallrock, losing the sharp boundaries critical for resource estimation.

Spatial transformer networks (STNs) learn to warp input data into canonical orientations before processing. For a narrow vein system, the STN learns to: (1) identify vein orientation from geological data, (2) rotate coordinates so the vein aligns with a preferred axis, (3) process in this vein-aligned reference frame, (4) rotate outputs back to geographic coordinates. This approach lets the model use different receptive field shapes for along-strike (large) versus across-vein (small) processing – respecting the anisotropy inherent in geological systems.

Implementation for mineral exploration requires training STNs on diverse structural settings. A model trained only on flat-lying stratiform deposits fails when encountering steeply dipping veins. Training data must span orientations, structural complexities, and deposit types – or the model learns spurious correlations with coordinate systems rather than actual geological relationships.

Geometric Deep Learning: Topology-Aware Processing

Recent advances in geometric deep learning move beyond regular grids entirely, representing geology as graphs where nodes correspond to sample points (drill intersections, geophysical stations, geochemical samples) and edges encode spatial relationships. Graph neural networks (GNNs) then propagate information along these edges, respecting the actual topology of geological connectivity.

Consider a structurally controlled gold system: mineralization follows a shear zone, jumps across a fault offset, continues in the offset segment. Traditional approaches struggle because the high-grade zones are spatially separated in geographic coordinates. A GNN can learn that these zones are topologically connected via structural relationships – the graph edges follow structural geology rather than Euclidean distance, so information flows naturally from one ore shoot to its fault-offset continuation.

Message-passing architectures define how information propagates: each node aggregates information from neighbors (other drill holes, nearby geophysical stations), updates its internal representation based on this context, and passes new information onward. After multiple message-passing steps, each location has integrated information from its broader geological neighborhood – not just spatial proximity, but structural connectivity, stratigraphic position, and alteration context.


Multi-Modal Fusion: Integrating Diverse Data Types

Mineral exploration generates wildly different data types: categorical (lithology codes, alteration types), continuous (element concentrations, geophysical measurements), sequential (downhole logs, time-series geochemistry), spatial (structural measurements, oriented drill cores), and imagery (remote sensing, core photos, thin sections). Traditional workflows process these independently, then manually integrate results. LWMs learn to fuse multi-modal data into unified probabilistic representations.

Drill Core Logs: Sequential Geological Context

A drill log isn't just a list of intersections – it's a geological story with narrative structure. Passing through hanging wall, entering mineralized zone, hitting high-grade shoots, exiting to footwall – these sequences encode geological context that individual intersections miss.

Transformer architectures adapted for drill logs use self-attention to capture long-range dependencies: a high-grade intersection 200m downhole might relate to an alteration signature 150m above, with dozens of barren intervals between. Standard recurrent networks struggle with these long-range relationships; transformers explicitly attend to all previous intersections, learning which uphole features predict downhole mineralization.

Positional encoding matters critically – a 5m high-grade interval at 200m depth means something different than the same interval at 800m depth, even in identical host rock. Depth encoding, structural position (hanging wall versus footwall), and stratigraphic context all require explicit representation. Learned embeddings can capture "this is the typical depth for main-stage mineralization in this deposit type."

Geophysical Data: Dense Indirect Observations

Geophysics provides wall-to-wall coverage but indirect measurements. Gravity responds to bulk density, magnetics to magnetic susceptibility, IP to chargeability, EM to conductivity – none directly measure ore grades, but all correlate with alteration, sulfide content, or geological structure.

Modern LWM architectures treat geophysics as a auxiliary modality that conditions predictions. Cross-attention mechanisms let the model selectively attend to geophysical features when interpreting drill data: if IP shows high chargeability, pay more attention to sulfide mineralogy in drill logs; if gravity shows a density low, weight alteration types associated with density reduction. The model learns which geophysical responses are diagnostic for which geological features – and more importantly, learns when geophysics is ambiguous or misleading.

Joint embedding spaces provide an elegant solution: map drill logs into a learned embedding space, map geophysical responses into the same space, and train the model so geologically similar situations cluster together regardless of data modality. A magnetite alteration zone identified from magnetics embeds near drill-intersected magnetite alteration, even though the raw data (continuous magnetic susceptibility versus categorical alteration code) looks completely different.

Geochemistry: Multi-Element Signatures

Geochemical data presents unique challenges – hundreds of elements, spanning 8+ orders of magnitude, with complex inter-element correlations. Traditional approaches use simple ratios or PC analysis; LWMs can learn nonlinear element associations and contextualize them geologically.

Pathfinder elements exemplify the challenge: arsenic and antimony are pathfinders for Carlin-type gold deposits, but only in the right geological context. High As-Sb without the right structural setting, rock type, and alteration assemblage is meaningless. An LWM learns to condition element interpretation on geological context – the same geochemical signature means something different in carbonaceous shale versus limestone, in first-order structures versus third-order splays, in oxidized versus reduced alteration assemblages.

Multi-task learning helps overcome sparse labeling: most geochemical samples lack grade assays (too expensive for broad reconnaissance), but all have geological context (lithology, alteration, structure). Training the model to predict both grade (where available) and geological context (everywhere) lets it learn the relationships between element patterns and geological settings even when grade data is sparse.

Structural Measurements: Directional Data

Structural measurements – foliations, fault orientations, vein attitudes – are fundamentally directional. A fault striking 045° is identical to one striking 225° (opposite direction, same orientation), and coordinate rotations shouldn't affect predictions. Standard neural networks fail at directional data – they treat 045° and 046° as similar, but 045° and 225° as opposite, when geologically they're identical.

Von Mises-Fisher distributions on the sphere provide a principled approach: rather than predicting orientations as angles, predict probability distributions over the sphere of possible directions. Loss functions account for orientation ambiguity (strike versus back-strike), and rotation-equivariant layers ensure predictions don't depend on coordinate system choice.

Practical impact: structural controls are often the dominant factor in mineralization – everything from where Witwatersrand gold concentrates in paleochannel bends to where porphyry copper grades increase in stockwork density. An LWM that truly understands structural geometry can predict where favorable structural settings occur outside drilled areas – the geologist's classic "project the structure, drill the projection."


The Vision: AI That Thinks Like an Exploration Geologist

The end goal isn't replacing exploration geologists – it's building AI partners that match human geological reasoning while processing orders of magnitude more data. An experienced geologist might integrate information from 50-100 drill holes, mentally correlate geology across sections, reference similar deposits from career experience, and generate targeting hypotheses. An LWM can do this across 10,000 drill holes, 200 geophysical surveys, and 500 analogous deposits – not better judgment than an expert geologist, but more comprehensive data synthesis.

Reasoning with Long Context: The 1M Token Challenge

Most geological projects accumulate megabytes to gigabytes of data – thousands of drill holes, millions of geochemical analyses, terabytes of geophysical measurements. Current large language models handle 100K-200K tokens; geological foundation models need 1M+ token contexts to reason over entire project databases simultaneously.

Sparse attention mechanisms make this tractable: rather than every token attending to every other token (quadratic complexity, computationally infeasible), learn which tokens are likely relevant. Structural measurements 5km away probably don't affect local grade prediction, but structural measurements along the same fault corridor absolutely do – even if separated by distance. The attention pattern should follow geological connectivity, not spatial proximity.

Hierarchical processing offers another path: summarize detailed drill logs into deposit-scale representations, reason at deposit scale to identify promising areas, then zoom into fine-scale detail for those areas. This mirrors how geologists work – broad reconnaissance, targeted follow-up, detailed evaluation – but can operate at larger scale and faster iteration.

Interactive Planning: Exploration as Sequential Decision-Making

Mineral exploration is inherently sequential – each drill hole informs the next decision, geophysical surveys target areas refined by previous results, geochemical sampling fills gaps identified by other data. Optimal exploration is a reinforcement learning problem: take actions (drill holes, surveys) that maximize information gain given a budget constraint.

LWMs can enable this by predicting not just "where's the ore" but "what information will each potential drill hole provide." This means modeling epistemic uncertainty – how much does the model not know – separately from aleatoric uncertainty – how variable is the geology. A drill hole in a high epistemic uncertainty region teaches the model more than one in a well-understood area, even if both have similar grade predictions.

Practical implementation: the model outputs uncertainty maps showing where geological understanding is weakest, then optimization algorithms propose drill holes maximizing expected information gain per dollar spent. Early adopters report 30-40% reductions in total drilling to resource definition by avoiding redundant holes in well-characterized areas and focusing on information-sparse regions.


Implications for Exploration Strategy and Operations

Product Architecture: Deploying an LWM requires more than training a model – you need data integration pipelines (ingesting diverse formats from acQuire, Geosoft, Leapfrog, ArcGIS), 3D visualization interfaces (showing model predictions, uncertainty, and reasoning), and interactive targeting tools (letting geologists query "what if we drill here?" before committing resources).

Data Infrastructure: Training demands comprehensive multi-modal datasets. Start aggregating now: historical drill databases (including failures and abandoned properties), digitized paper logs, geophysical GEOSOFT grids, geochemical data from government surveys, structural measurements from field mapping. Expect storage requirements of 10-50 TB per trained model and pipeline throughput of 100+ GB/hour during training.

Go-to-Market for Service Providers: AI exploration consulting is emerging – companies without internal AI capability can partner with specialists who bring trained foundation models and fine-tune them on the client's data. For juniors, this offers access to technology requiring $5-10M to develop in-house. For consultants, it enables premium services differentiated from traditional geological consulting.


Navigating the Trade-offs

Adopting subsurface LWMs involves careful trade-offs:

Compute & Latency: Training a foundation model costs $500K-2M in compute (1,000+ GPU-hours), though inference is relatively cheap (~$50/property for comprehensive targeting). Fine-tuning on a specific property requires 10-20 GPU-hours ($200-400). You'll need cloud infrastructure partnerships (AWS, Azure, GCP) unless building internal GPU clusters.

Data Needs: LWMs hunger for diverse, comprehensive data – ideally 100+ deposits for pre-training, 10+ for domain adaptation, and complete multi-modal coverage (drill, geophysics, geochemistry, structure) for each property. Data is the moat – companies with superior databases will train superior models. Invest in data acquisition, digitization, and quality control now.

Reliability: These models can hallucinate geology – generating plausible-looking targets that violate physical or geological constraints. Rigorous validation (hold-out testing, geological plausibility checks, expert review) is essential. For compliance reporting (NI 43-101, JORC), plan on treating AI models as supplementary to QP-signed traditional estimation until regulatory guidance evolves.

Interpretability: A model predicting "drill here" isn't sufficient – geologists need to understand why. Attention visualization (which input data drove the prediction?), counterfactual analysis (what if this structural measurement were different?), and learned representation exploration (what did the model learn about this deposit type?) are critical for building trust and extracting geological insights beyond targeting.


What Exploration Companies Should Do Now

  1. Assess Data Readiness: Audit existing databases for completeness, consistency, and machine-readability. A company with 50 properties but inconsistent logging standards can't effectively train an LWM. Standardize lithology codes, alteration classifications, and assay databases before attempting AI integration.
  2. Identify High-Impact Use Cases: Focus on problems where traditional methods struggle – structurally complex deposits, sparse geophysics requiring better interpretation, properties with abundant legacy data but no recent success. These offer the largest opportunity for AI-driven improvements.
  3. Build or Partner: Large companies with 10+ projects should develop internal capabilities (hire ML engineers, train geologists on AI fundamentals). Juniors with 1-3 projects should partner with specialized consultants or platforms offering LWM-as-a-service. The break-even point for internal development is approximately $20M annual exploration budget.
  4. Pilot with Validation: Choose a property with abundant historical data and ground truth (subsequent drilling, mining, or updated interpretations). Generate LWM predictions from early-stage data, validate against later reality, quantify performance gains. Use this to build internal confidence and refine workflows before deploying on greenfield projects.

Conclusion

Large World Models push mineral exploration beyond statistical pattern-matching into genuine geological reasoning – understanding 3D relationships, structural controls, alteration sequences, and multi-modal data integration. They offer unprecedented capabilities but also significant implementation challenges. Exploration companies investing in subsurface LWM capabilities now will be positioned to make faster, more cost-effective exploration decisions while competitors continue struggling with data integration and sequential interpretation.

The core insight: geology is fundamentally spatial, multimodal, and hierarchical. Traditional AI treats it as flat, unimodal, and independent. LWMs bridge this gap – they think the way geological systems actually work. And that's the inflection point.

30 minute technical discovery

Woody de Kafou

Woody de Kafou

Founder & CEO, AI Thought Leader

BOOK A MEETING