Human mesh reconstruction (HMR) provides direct insights into body-environment interaction,
which enables various immersive applications. While existing large-scale HMR datasets rely
heavily on line-of-sight RGB input, vision-based sensing is limited by occlusion, lighting
variation, and privacy concerns. To overcome these limitations, recent efforts have explored
radio-frequency (RF) mmWave radar for privacy-preserving indoor human sensing. However, current
radar datasets are constrained by sparse skeleton labels, limited scale, and simple in-place actions.
To advance the HMR research community, we introduce M4Human, the current largest-scale (661K-frame)
(9 times prior largest) multimodal benchmark,
featuring high-resolution mmWave radar, RGB, and depth data.
M4Human provides both raw radar tensors (RT) and processed radar point clouds (RPC)
to enable research across different levels of RF signal granularity. M4Human includes
high-quality motion capture (MoCap) annotations with 3D meshes and global trajectories,
and spans 20 subjects and 50 diverse actions, including in-place, sit-in-place, and free-space
sports or rehabilitation movements. We establish benchmarks on both RT and RPC modalities, as well as
multimodal fusion with RGB-D modalities. Extensive results highlight the significance of
M4Human for radar-based human modeling while revealing persistent challenges under fast,
unconstrained motion. The dataset and code will be released after the paper publication.