Connected Patterns: Turning Measurements into Invariants
“An invariant is a promise the world keeps, even when your model changes.”
There is a reason conservation laws feel different from other scientific statements.
Flagship Router PickQuad-Band WiFi 7 Gaming RouterASUS ROG Rapture GT-BE98 PRO Quad-Band WiFi 7 Gaming Router
ASUS ROG Rapture GT-BE98 PRO Quad-Band WiFi 7 Gaming Router
A flagship gaming router angle for pages about latency, wired priority, and high-end home networking for gaming setups.
- Quad-band WiFi 7
- 320MHz channel support
- Dual 10G ports
- Quad 2.5G ports
- Game acceleration features
Why it stands out
- Very strong wired and wireless spec sheet
- Premium port selection
- Useful for enthusiast gaming networks
Things to know
- Expensive
- Overkill for simpler home networks
A curve fit can look good for a while and still be wrong. A classifier can score well and still be brittle. But when you find a true conservation law, you have found something that survives changes of scale, choice of coordinates, and even many changes of mechanism. It is the kind of claim that keeps paying rent, because it does not just describe what happened. It constrains what can happen.
That is why “discovering conservation laws from data” is one of the most exciting uses of AI in science, and also one of the easiest places to fool yourself. Data is noisy. Measurements are incomplete. Many systems only approximately conserve quantities under specific regimes. A naive workflow will gladly return a beautiful “law” that dissolves the moment you test it on new trajectories.
A practical workflow has a different goal:
- Treat candidate conservation laws as hypotheses, not conclusions
- Demand that invariants survive hold-out conditions, not just the training window
- Quantify how close the “conservation” is, and when it breaks
- Prefer simple, interpretable forms that can be stress-tested and communicated
What You Mean by “Conservation” in Real Data
In a textbook, a conserved quantity stays exactly constant over time.
In a lab or simulation pipeline, you usually see something messier:
- A quantity is conserved only after you correct for measurement bias
- Conservation holds only within a regime, like a range of temperatures or energies
- The “law” is approximate, but the residual has structure you can explain
- The invariant is not obvious in the raw variables, but appears after a transform
So the first discipline is to name the claim precisely.
A conservation-law claim should specify:
- The state variables you observe
- The time scale over which you assert conservation
- The conditions under which it holds
- The tolerance and error model you accept
- The tests that could falsify it
This sounds strict, but it is what turns “interesting pattern” into “defensible statement.”
The Core Workflow: Propose, Check, Stress-Test
Most approaches, whether symbolic or neural, reduce to a loop:
- Propose a candidate invariant I(x) from data
- Check whether I(x(t)) is constant along trajectories
- Stress-test that constancy under new conditions, new initial states, and new noise
The important part is the stress-test, because it is where fake invariants die.
Proposal engines that work
There are multiple ways to propose I(x). The best choice depends on how much structure you already believe exists.
Common proposal families:
- Symbolic candidates: polynomials, rational functions, sparse combinations of features
- Physics-informed candidates: energy-like sums, momentum-like terms, known dimensional forms
- Learned candidates: neural networks trained to output a scalar that stays constant along trajectories
- Hybrid candidates: a learned embedding followed by a sparse symbolic head for interpretability
The crucial requirement is that the proposal family is constrained enough that the result is testable and understandable.
If your candidate space is too flexible, the system will “memorize invariance” on the training traces and fail outside them.
Checking invariance without lying to yourself
The simplest check is to compute the variance of I(x(t)) over time.
That is necessary, but not sufficient.
You also need to check for the common ways apparent invariance arises:
- Drift cancellation: two errors with opposite sign hide the change
- Window bias: invariance holds only in a short segment you happened to sample
- Parameter leakage: the candidate indirectly encodes time or a hidden index
- Smoothing artifacts: preprocessing removes the very variations you are trying to explain
A better check includes:
- Multiple trajectories with different initial states
- Explicit hold-out trajectories not used in proposing the invariant
- Time-reversal or perturbation tests when applicable
- Simulated counterfactuals if you have a forward model
A Verification Ladder for Conservation Claims
Conservation law discovery should climb a ladder, not jump to the top.
| Verification rung | What you test | What could fool you | What makes it trustworthy |
|---|---|---|---|
| Stability on training traces | I(x(t)) stays near-constant | Overfitting to a narrow window | Multiple trajectories, no time leakage |
| Stability on hold-out traces | New initial conditions | Candidate memorizes training dynamics | Clear generalization without retuning |
| Regime robustness | Different parameter settings | Invariant is regime-specific | You map where it holds and where it fails |
| Noise robustness | Measurement noise, missingness | Smoothing creates fake constancy | Performance under realistic noise models |
| Mechanistic plausibility | Dimensional and structural sense | Coincidental cancellations | Interpretable form, aligns with constraints |
| Predictive constraint | Future states are restricted | “Invariant” does not constrain anything | You can rule out trajectories using the law |
The last rung is a powerful discriminator.
A good invariant is not just constant. It constrains behavior. It lets you say, “These futures are impossible unless something injects or removes the conserved quantity.”
Practical Methods That Show Up in Real Pipelines
Sparse regression for invariants
If you can build a library of candidate features, you can search for a combination that stays constant.
Typical pattern:
- Build features φ(x) such as monomials, trigonometric terms, or domain-specific quantities
- Search for coefficients c so that I(x) = c·φ(x) has minimal time-derivative along data
- Regularize for sparsity so the result is simple and robust
This can work extremely well when the real invariant is low-complexity.
Where it fails:
- When derivatives are noisy and amplify error
- When the invariant requires a transform you did not include
- When multiple near-invariants confuse the selection
Mitigation is not “use a bigger model.” It is “use better features and better checks.”
Neural invariants with structure
Neural networks can propose invariants without handcrafting features, but they need discipline.
Better patterns include:
- Learn an embedding z(x) and constrain I to be simple in z
- Penalize time-derivative of I along trajectories
- Add regularizers that enforce smoothness and avoid time leakage
- Force consistency across multiple trajectories and regimes
Then you take the neural candidate and try to distill it into a symbolic or simplified form.
The goal is not “a neural network that outputs a constant.” The goal is an invariant you can defend.
Distillation into interpretable laws
A practical approach:
- Use a flexible model to discover a candidate invariant
- Fit a simpler symbolic form to the candidate outputs
- Verify that the symbolic form still passes the stress-tests
Distillation is a truth test. If the “invariant” disappears when you ask for a simple expression, you likely had a fragile artifact.
Common Failure Modes and How to Catch Them
You can save months by assuming you will hit these.
Hidden time encoding
- Symptom: invariance is perfect, but only when using your exact data pipeline
- Fix: randomize time indexing, test with shuffled time stamps, remove any time features
Preprocessing-induced invariance
- Symptom: invariance improves when you smooth more
- Fix: evaluate on rawer data, vary smoothing, measure bias introduced by filters
Regime mismatch
- Symptom: invariant holds on one parameter set and breaks elsewhere
- Fix: treat the result as a regime-specific invariant and map its boundary
Multiple invariants competing
- Symptom: different runs return different laws with similar training scores
- Fix: compare under hold-out conditions and prefer the law that constrains prediction best
Confounded variables
- Symptom: invariant correlates with an unmeasured factor
- Fix: design experiments that vary suspected confounders independently
A good discipline is to keep an “invariant failure notebook” where you record every candidate that died and why. It becomes a map of your system’s true structure.
What a Strong Result Looks Like
A strong conservation-law discovery report can be summarized in a compact bundle:
- The invariant expression, in the simplest form you can justify
- A plot of I(x(t)) across many trajectories, including hold-outs
- A table mapping regimes where conservation holds or breaks
- An error model: expected variance under measurement noise
- A falsification plan: what new experiment could refute the law
- A mechanistic story: why this invariant makes sense
The mechanistic story matters. It is how you move from “pattern” to “understanding.”
When Conservation Is Approximate on Purpose
Sometimes the most valuable result is not a perfect invariant, but a controlled deviation.
If the system slowly leaks energy, or gradually loses mass, the residual tells you something.
Instead of forcing a fake conservation law, you can model:
- A conserved core plus a small drift term
- An invariant that holds under closed conditions, and breaks under open conditions
- A conservation law with an external forcing term you can estimate
This is still a discovery, because it tells you where the system is open to influence.
Keep Exploring AI Discovery Workflows
If you want to connect this topic to the rest of the discovery pipeline, these posts are the natural next steps.
• AI for Scientific Discovery: The Practical Playbook
https://ai-rng.com/ai-for-scientific-discovery-the-practical-playbook/
• Symbolic Regression for Discovering Equations
https://ai-rng.com/symbolic-regression-for-discovering-equations/
• AI for PDE Model Discovery
https://ai-rng.com/ai-for-pde-model-discovery/
• Inverse Problems with AI: Recover Hidden Causes
https://ai-rng.com/inverse-problems-with-ai-recover-hidden-causes/
• Uncertainty Quantification for AI Discovery
https://ai-rng.com/uncertainty-quantification-for-ai-discovery/
• From Data to Theory: A Verification Ladder
https://ai-rng.com/from-data-to-theory-a-verification-ladder/
