The UK Government’s report on the copyright and AI consultation was recently published. While the report confirms that balancing the interests of copyrights holders and AI developers is a complex exercise, it also provides an indication of likely scenarios to consider in this fast-evolving environment. The consultation focused on whether AI developers should be permitted

Researchers say deep learning can have a rigorous scientific theory and the math to prove it is already emerging – Startup Fortune

A new body of research is converging on a formal scientific theory of deep learning, using phase transitions from statistical physics and loss landscape geometry to explain why neural networks learn, not just that they do, with implications for training efficiency that could reshape how frontier models are built.
The criticism has followed deep learning since its commercial breakthrough. The models work, often spectacularly, but nobody fully understands why. Engineers and researchers have developed an extensive vocabulary of observed phenomena, grokking, double descent, emergent capabilities, loss spikes during training, but the underlying principles connecting these observations into a coherent explanatory framework have remained elusive. That is starting to change. Research published in recent months, building on years of work in singular learning theory and loss landscape analysis, is accumulating into something that looks, for the first time, like the foundations of a genuine scientific theory of how deep learning systems acquire knowledge.
The conceptual anchor is phase transitions. In physics, a phase transition is a sharp, qualitative change in a system’s behavior as a parameter crosses a threshold, water freezing at zero degrees Celsius being the simplest example. Researchers have been observing analogous transitions in neural network training for years: moments where the model’s internal structure reorganizes rapidly, where a capability that did not exist at step N appears suddenly at step N plus a small increment, where the geometry of the loss landscape shifts in ways that correspond to interpretable changes in what the model can do. A December 2025 paper from researchers working in the singular learning theory framework demonstrated analytically that these phase transitions are governed by saddle points in the loss landscape, providing a mathematical mechanism for what had previously been a qualitative observation.
Deep learning involves navigating a high-dimensional loss landscape, a surface defined by the model’s error across all possible parameter configurations. Training is the process of descending that surface toward a minimum. The landscape is not smooth: it contains saddle points, flat regions, clusters of nearly equivalent minima, and sharp valleys. Understanding its geometry is the key to understanding training dynamics. Work published in Nature Communications in April 2025, modeling loss landscapes as multifractal structures, unified a broad range of observed training phenomena including clustered degenerate minima, the edge of stability phenomenon, and anomalous optimization dynamics under a single theoretical framework. That kind of unification, multiple previously isolated observations falling out of a single model, is what distinguishes theoretical progress from empirical cataloguing.
The practical payoff is potentially significant. If the geometry of the loss landscape before training begins contains predictive information about how training will unfold, then the expensive, wasteful trial-and-error approach to training large models could be partially replaced by analytical prediction. A paper on feature learning phase transitions published on arXiv in January 2026 provided early evidence that specific architectural and initialization choices alter the phase transition structure in predictable ways, suggesting that practitioners may eventually be able to choose training configurations based on theoretical guarantees rather than empirical intuition backed by massive compute expenditure.
The Explainability Dividend
The theoretical case matters beyond training efficiency. Regulators in the European Union and in US federal agencies overseeing healthcare, finance, and autonomous systems have consistently identified the opacity of deep learning models as a barrier to deployment in high-stakes contexts. The EU AI Act’s requirements for transparency and explainability in high-risk AI systems do not require that every model decision be interpretable at the parameter level, but they do require that developers can characterize model behavior reliably and explain failures systematically. A mathematical framework that describes learning as a sequence of phase transitions in a geometric structure provides, for the first time, a vocabulary for that characterization that does not reduce to
Also read: A microwave clock and a Reddit thread just taught millions of people how AI image editing works • How to Build a Local Marketing Agency Without a Large Team – Using Agentic Agency • Google commits $40 billion to Anthropic and the AI funding arms race enters a new era
