IoT-Edge Hybrid Architecture with Cross-Modal Transformer and Federated Manifold Learning for Safety-Critical Gesture Control in Adaptive Mobility Platforms
Xinmin Jin, Jian Teng, Jiaji ChenThis research presents an IoT-empowered adaptive mobility framework that integrates high-dimensional gesture recognition with edge-cloud orchestration for safety-critical human–machine interaction. The system architecture establishes a three-tier IoT network: a perception layer with 60 GHz FMCW radar and TOF infrared arrays (12-node mesh topology, 15 cm baseline spacing) for real-time motion tracking; an edge intelligence layer deploying a time-aware neural network via NVIDIA Jetson Nano to achieve up to 99.1% recognition accuracy with latency as low as 48 ms under optimal conditions (typical performance: 97.8% ± 1.4% accuracy, 68.7 ms ± 15.3 ms latency); and a federated cloud layer enabling distributed model synchronization across 32 edge nodes via LoRaWAN-optimized protocols (κ = 0.912 consensus). A reconfigurable chassis with three operational modes (standing, seated, balance) employs IoT-driven kinematic optimization for enhanced adaptability and user safety. Using both radar and infrared sensors together reduces false detections to 0.08% even under high-vibration conditions (80 km/h), while distributed learning across multiple devices maintains consistent accuracy (variance < 5%) in different environments. Experimental results demonstrate 93% reliability improvement over HMM baselines and 3.8% accuracy gain over state-of-the-art LSTM models, while achieving 33% faster inference (48.3 ms vs. 72.1 ms). The system maintains industrial-grade safety certification with energy-efficient computation. Bridging adaptive mechanics with edge intelligence, this research pioneers a sustainable IoT-edge paradigm for smart mobility, harmonizing real-time responsiveness, ecological sustainability, and scalable deployment in complex urban ecosystems.