Login / Signup

The VISTA datasets, a combination of inertial sensors and depth cameras data for activity recognition.

Laura FioriniFederica Gabriella Cornacchia LoizzoAlessandra SorrentinoErika RoviniAlessandro Di NuovoFilippo Cavallo
Published in: Scientific data (2022)
This paper makes the VISTA database, composed of inertial and visual data, publicly available for gesture and activity recognition. The inertial data were acquired with the SensHand, which can capture the movement of wrist, thumb, index and middle fingers, while the RGB-D visual data were acquired simultaneously from two different points of view, front and side. The VISTA database was acquired in two experimental phases: in the former, the participants have been asked to perform 10 different actions; in the latter, they had to execute five scenes of daily living, which corresponded to a combination of the actions of the selected actions. In both phase, Pepper interacted with participants. The two camera point of views mimic the different point of view of pepper. Overall, the dataset includes 7682 action instances for the training phase and 3361 action instances for the testing phase. It can be seen as a framework for future studies on artificial intelligence techniques for activity recognition, including inertial-only data, visual-only data, or a sensor fusion approach.
Keyphrases
  • big data
  • electronic health record
  • artificial intelligence
  • machine learning
  • emergency department
  • deep learning
  • physical activity
  • adverse drug
  • convolutional neural network