Paper
Performance of universal machine learning potentials in global optimization
Authors
Edan T. Marcial, Laxman Chaudhary, Olesya Gorbunova, Aleksey N. Kolmogorov
Abstract
Rapid development of universal machine learning potentials (uMLPs) and expansion of training data sets are reshaping the state of the art in atomistic simulation, highlighting the need for concurrent systematic benchmarking of their capabilities. Global optimization is among the most demanding uMLP applications because unconstrained exploration includes probing motifs not present in reference sets. We examined the latest generation of uMLPs in unconstrained evolutionary searches to assess whether these models can consistently predict complex crystal structure ground states across diverse inorganic systems. Our findings demonstrate that the considered M3GNet, MACE, SevenNet, EquiformerV2, MatterSim, GRACE, eSEN, Orb-v3, and PET-MAD models span a wide performance range, from near ab initio to essentially non-predictive, in their ability to resolve competing phases within low-energy basins. Additional tests on hcp-Zn, MB$_4$ (M = Cr, Mn, and Fe), and LiB$_{y}$ ($y\approx 0.9$) ground states reveal that several uMLPs capture fine energy differences arising from subtle electronic structure features.
Metadata
Related papers
Fractal universe and quantum gravity made simple
Fabio Briscese, Gianluca Calcagni • 2026-03-25
POLY-SIM: Polyglot Speaker Identification with Missing Modality Grand Challenge 2026 Evaluation Plan
Marta Moscati, Muhammad Saad Saeed, Marina Zanoni, Mubashir Noman, Rohan Kuma... • 2026-03-25
LensWalk: Agentic Video Understanding by Planning How You See in Videos
Keliang Li, Yansong Li, Hongze Shen, Mengdi Liu, Hong Chang, Shiguang Shan • 2026-03-25
Orientation Reconstruction of Proteins using Coulomb Explosions
Tomas André, Alfredo Bellisario, Nicusor Timneanu, Carl Caleman • 2026-03-25
The role of spatial context and multitask learning in the detection of organic and conventional farming systems based on Sentinel-2 time series
Jan Hemmerling, Marcel Schwieder, Philippe Rufin, Leon-Friedrich Thomas, Mire... • 2026-03-25
Raw Data (Debug)
{
"raw_xml": "<entry>\n <id>http://arxiv.org/abs/2602.23515v1</id>\n <title>Performance of universal machine learning potentials in global optimization</title>\n <updated>2026-02-26T21:39:33Z</updated>\n <link href='https://arxiv.org/abs/2602.23515v1' rel='alternate' type='text/html'/>\n <link href='https://arxiv.org/pdf/2602.23515v1' rel='related' title='pdf' type='application/pdf'/>\n <summary>Rapid development of universal machine learning potentials (uMLPs) and expansion of training data sets are reshaping the state of the art in atomistic simulation, highlighting the need for concurrent systematic benchmarking of their capabilities. Global optimization is among the most demanding uMLP applications because unconstrained exploration includes probing motifs not present in reference sets. We examined the latest generation of uMLPs in unconstrained evolutionary searches to assess whether these models can consistently predict complex crystal structure ground states across diverse inorganic systems. Our findings demonstrate that the considered M3GNet, MACE, SevenNet, EquiformerV2, MatterSim, GRACE, eSEN, Orb-v3, and PET-MAD models span a wide performance range, from near ab initio to essentially non-predictive, in their ability to resolve competing phases within low-energy basins. Additional tests on hcp-Zn, MB$_4$ (M = Cr, Mn, and Fe), and LiB$_{y}$ ($y\\approx 0.9$) ground states reveal that several uMLPs capture fine energy differences arising from subtle electronic structure features.</summary>\n <category scheme='http://arxiv.org/schemas/atom' term='cond-mat.mtrl-sci'/>\n <published>2026-02-26T21:39:33Z</published>\n <arxiv:comment>14 pages, 3 tables, 6 figures</arxiv:comment>\n <arxiv:primary_category term='cond-mat.mtrl-sci'/>\n <author>\n <name>Edan T. Marcial</name>\n </author>\n <author>\n <name>Laxman Chaudhary</name>\n </author>\n <author>\n <name>Olesya Gorbunova</name>\n </author>\n <author>\n <name>Aleksey N. Kolmogorov</name>\n </author>\n </entry>"
}