Patterns of human motion in outdoor and indoor environments are substantially different due to the scope of the scene and the typical intentions of people therein. While outdoor trajectory forecasting has received significant attention, indoor forecasting is still an underexplored research area. This paper proposes SITUATE, a novel approach to cope with indoor human trajectory prediction by leveraging equivariant and invariant geometric features and self-supervised vision representation. The geometric learning modules model the intrinsic symmetries and human movements inherent in indoor spaces. This concept becomes particularly important because self-loops at various scales and rapid direction changes often characterize indoor trajectories. On the other hand, the vision representation module is used to acquire spatial-semantic information about the environment to predict users' future locations more accurately. We evaluate our method through comprehensive experiments on the two most famous indoor trajectory forecasting datasets, i.e., THΓR and Supermarket, obtaining state-of-the-art performance. Furthermore, we also achieve competitive results in outdoor scenarios, showing that indoor-oriented forecasting models generalize better than outdoor-oriented ones.
@article{capogrosso2024situate,
title={SITUATE: Indoor Human Trajectory Prediction through Geometric Features and Self-Supervised Vision Representation},
author={Capogrosso, Luigi and Toaiari, Andrea and Avogaro, Andrea and Khan, Uzair and Jivoji, Aditya and Fummi, Franco and Cristani, Marco},
journal={arXiv preprint arXiv:2409.00774},
year={2024}
}