Research

Paper

AI LLM March 13, 2026

MotionAnymesh: Physics-Grounded Articulation for Simulation-Ready Digital Twins

Authors

WenBo Xu, Liu Liu, Li Zhang, Dan Guo, RuoNan Liu

Abstract

Converting static 3D meshes into interactable articulated assets is crucial for embodied AI and robotic simulation. However, existing zero-shot pipelines struggle with complex assets due to a critical lack of physical grounding. Specifically, ungrounded Vision-Language Models (VLMs) frequently suffer from kinematic hallucinations, while unconstrained joint estimation inevitably leads to catastrophic mesh inter-penetration during physical simulation. To bridge this gap, we propose MotionAnymesh, an automated zero-shot framework that seamlessly transforms unstructured static meshes into simulation-ready digital twins. Our method features a kinematic-aware part segmentation module that grounds VLM reasoning with explicit SP4D physical priors, effectively eradicating kinematic hallucinations. Furthermore, we introduce a geometry-physics joint estimation pipeline that combines robust type-aware initialization with physics-constrained trajectory optimization to rigorously guarantee collision-free articulation. Extensive experiments demonstrate that MotionAnymesh significantly outperforms state-of-the-art baselines in both geometric precision and dynamic physical executability, providing highly reliable assets for downstream applications.

Metadata

arXiv ID: 2603.12936
Provider: ARXIV
Primary Category: cs.RO
Published: 2026-03-13
Fetched: 2026-03-16 06:01

Related papers

Raw Data (Debug)
{
  "raw_xml": "<entry>\n    <id>http://arxiv.org/abs/2603.12936v1</id>\n    <title>MotionAnymesh: Physics-Grounded Articulation for Simulation-Ready Digital Twins</title>\n    <updated>2026-03-13T12:30:42Z</updated>\n    <link href='https://arxiv.org/abs/2603.12936v1' rel='alternate' type='text/html'/>\n    <link href='https://arxiv.org/pdf/2603.12936v1' rel='related' title='pdf' type='application/pdf'/>\n    <summary>Converting static 3D meshes into interactable articulated assets is crucial for embodied AI and robotic simulation. However, existing zero-shot pipelines struggle with complex assets due to a critical lack of physical grounding. Specifically, ungrounded Vision-Language Models (VLMs) frequently suffer from kinematic hallucinations, while unconstrained joint estimation inevitably leads to catastrophic mesh inter-penetration during physical simulation. To bridge this gap, we propose MotionAnymesh, an automated zero-shot framework that seamlessly transforms unstructured static meshes into simulation-ready digital twins. Our method features a kinematic-aware part segmentation module that grounds VLM reasoning with explicit SP4D physical priors, effectively eradicating kinematic hallucinations. Furthermore, we introduce a geometry-physics joint estimation pipeline that combines robust type-aware initialization with physics-constrained trajectory optimization to rigorously guarantee collision-free articulation. Extensive experiments demonstrate that MotionAnymesh significantly outperforms state-of-the-art baselines in both geometric precision and dynamic physical executability, providing highly reliable assets for downstream applications.</summary>\n    <category scheme='http://arxiv.org/schemas/atom' term='cs.RO'/>\n    <category scheme='http://arxiv.org/schemas/atom' term='cs.CV'/>\n    <published>2026-03-13T12:30:42Z</published>\n    <arxiv:comment>5 figures</arxiv:comment>\n    <arxiv:primary_category term='cs.RO'/>\n    <author>\n      <name>WenBo Xu</name>\n    </author>\n    <author>\n      <name>Liu Liu</name>\n    </author>\n    <author>\n      <name>Li Zhang</name>\n    </author>\n    <author>\n      <name>Dan Guo</name>\n    </author>\n    <author>\n      <name>RuoNan Liu</name>\n    </author>\n  </entry>"
}