SPATIUM: A Framework for Multimodal, LLM-Enhanced, and Spatiotemporal Machine Learning in Human-Centered Health Applications [paper]

SPATIUM redefines digital health by fusing physiological sensing, spatial intelligence, and context-aware machine learning into one cohesive framework.

It bridges the gap between body and environment to capture heart rate, HRV, respiration, and temperature from OmniBuds, while integrating Ultra-Wideband (UWB) indoor localization and environmental data such as light, sound, and temperature. The result is an immersive spatiotemporal health map rendered in 2D, 3D, and Apple Vision Pro environments, enabling intuitive visualization of how the body responds to its surroundings in real time.

Powered by large language models (LLMs), SPATIUM can interpret multimodal data, explain physiological changes within environmental context, and generate personalized insights and alerts in natural language.

SPATIUM Framework
Context-aware machine learning architecture for spatiotemporal health understanding.
Prototype System
Integration of OmniBuds sensing, UWB localization, and immersive visualization setup.
Immersive Visualization
Apple Vision Pro interface showing 3D spatiotemporal health data.
Spatiotemporal Data Mapping
Heart rate trajectories visualized across space and time in real-world environments.

A demo for the spatiotemporal visualization is shown in the video below.