In-bed human mesh recovery can be crucial and enabling for
several healthcare applications, including sleep pattern
monitoring, rehabilitation support, and pressure ulcer
prevention. However, it is difficult to collect large real-world
visual datasets in this domain, in part due to privacy and
expense constraints, which in turn presents significant
challenges for training and deploying deep learning models.
Existing in-bed human mesh estimation methods often rely heavily
on real-world data, limiting their ability to generalize across
different in-bed scenarios, such as varying coverings and
environmental settings. To address this, we propose a
Sim-to-Real Transfer Framework for in-bed human mesh recovery
from overhead depth images, which leverages large-scale
synthetic data alongside limited or no real-world samples. We
introduce a diffusion model that bridges the gap between
synthetic data and real data to support generalization in
real-world in-bed pose and body inference scenarios. Extensive
experiments and ablation studies validate the effectiveness of
our framework, demonstrating significant improvements in
robustness and adaptability across diverse healthcare scenarios.
Our proposed framework addresses the challenge of developing reliable and generalizable in-bed human mesh recovery models in scenarios with limited or no real-world data. By leveraging a large volume of synthetic data generated through simulation, combined with a small amount of real-world data, our framework effectively reduces the reliance on costly and privacy-sensitive real-world data collection. The framework comprises three stages:
The symbol g in the diffusion model indicates the gender flag associated with the input. The Ref in the figure denotes the corresponding synthetic depth image during training and the corresponding RGB image for visualization purposes only.