Continuous Multimodal Multi-view Dataset of Human Fall (CMDFALL)

About the Project

The project is a collaboration between Computer Vision Department (COMVIS) of MICA International Research Institute (MICA) and Posts & Telecommunications Institute of Technology (PTIT). The objective of this project is to build a Continuous Multimodal Multi-view Dataset of human fall actions and to develop algorithms for human fall recognition. The dataset consists of both normal daily activities and simulated falls.

Description

Safety is a key concern for families with elderly people. Fall is the major risk for elderly people living at home alone, which might cause serious health consequence if medical treatment is not afforded in-time. Many techniques were proposed such to detect human fall to immediately alert other family members/medical staffs for in-time first aid. Two most potential types of sensing devices are cameras and accelerometers. Many computer vision/ accelerometer signal processing algorithms have been proposed for human fall recognition but there is a lack of a common, large-enough dataset of human fall actions for research community to evaluate their algorithms. This project aims at building a Continuous Multimodal Multi-view Dataset of human fall actions acquired from Microsoft Kinect cameras and wireless accelerometers simultaneously. The CMDFALL dataset contains 20 actions performed by 50 persons (30 males and 20 females in the range 21-40 years of age) with age and gender distribution as follows:

Figure 1: Age Distribution of 50 Subjects

Figure 2: Gender Distribution of 50 Subjects

Each person performed all 20 actions in 7.5 minutes, yielding about 375 minutes recording time in total. Falls were done in all different directions (front/back/left/right). A mattress was laid on the floor to protect the subjects during the simulated falls.

Figure 3: Data from seven Kinect cameras and two accelerometers

Hardware

In this section we describe the layout of the multimodal data collection system used for the collection of CMDFALL. Each action was simultaneously captured by two different sensing systems: seven Microsoft Kinect cameras and two wireless accelerometers.

The action list is divided into two scenes. Scene 1 is a room with no bed as shown in Figure 4. Scene 2 is a room with a bed as shown in Figure 5.

Figure 4: Scene 1 Room with no bed

Figure 5: Scene 2 Room with a bed

Kinect System

There are seven Kinect cameras (numbered from Kinect1 to Kinect7). The locations of devices are described as follows:

  1. Six Kinect cameras (from Kinect1 to Kinect6) are installed at height 1.8m.
  2. One Kinect camera (Kinect7) is installed on the ceiling at height 3.0m.

Each Kinect camera captured a color image with a resolution of 640x480 pixels and a 16-bit depth image, both with an acquisition rate of 20 Hz. The laptop connected to Kinect6 was set as a controller for time synchronization for all seven Kinect cameras.

Accelerometers

Two accelerometers are used for each subject during recording, one on the left wrist and one on the left hip of the subject to capture the physical activities. The sensor used in this project is 3-dimensional wireless accelerometer (WAX3) developed by researchers at Open Lab, Newcastle University. Originally, WAX3 was developed for embedding inside kitchen utensils and appliances. WAX3 is a wireless, small +/-16g 3-axis accelerometer with the dimension of 23 x 32.5 x 7.6 mm, and weighted just 7 grams. WAX is MEMS accelerometer that operates with the IEEE 802.15.4 low power radio protocol. It has the sampling rate of up to 2ks.sec-1 and supports Open Sound Control (OSC) message, binary, and American Standard Code for Information Interchange (ASCII) format. WAX consists of two components: WAX receiver (connecting to the computer via a USB port) and WAX transmitter (wireless communication to the receiver with the range of up to 25 m). WAX is also equipped with a re-chargeable Li-Polymer battery providing up to 8 hours of continuously signal transmission mode and up to 56 days in hibernate mode.

Figure 6: WAX3 Accelerometer

Action Checklist

For each subject, we collected the movements as shown in the following table.

Scene 1: Room with no bed

ID Action Recording Length (sec) Annotation Start Point Annotation End Point Sample Frame 1 Sample Frame 2 Sample Frame 3 Sample Frame 4 Sample Frame 5
1 Forward fall 30 Subject stands straight before forward fall Subject lies motionless on the floor

2 Back fall 30 Subject stands straight before back fall Subject lies motionless on the floor

3 Left fall 30 Subject stands straight before left fall Subject lies motionless on the floor

4 Right fall 30 Subject stands straight before right fall Subject lies motionless on the floor

7 Left fall while sitting on a chair 20 Subject sits on the chair before left fall Subject lies motionless on the floor

8 Right fall while sitting on a chair 20 Subject sits on the chair before right fall Subject lies motionless on the floor

9 Walk in all direction 20 Subject starts walking Subject stops walking

10 Run slowly 20 Subject starts running Subject stops running

11 Stagger 20 Subject starts staggering Subject stops staggering

12 Crawl 10 Subject starts crawling Subject stops crawling

13 Move the chair 20 Subject starts moving the chair Subject stops moving the chair

14 Move hands and knees 20 Subject starts moving hands and knees Subject stops moving hands and knees

15 Crouch down to pick up things by left hand 20 Subject stands straight before crouching down to pick up things by left hand Subject's head is at the highest position after picking up things

16 Crouch down to pick up things by right hand 20 Subject stands straight before crouching down to pick up things by right hand Subject's head is at the highest position after picking up things

17 Jump in place 20 Subject starts jumping Subject stops jumping

18 Sit down on a chair then stand up 20 Subject stands straight before sitting on the chair Subject stands straight after standing up

Scene 2: Room with a bed

ID Action Recording Length (sec) Annotation Start Point Annotation End Point Sample Frame 1 Sample Frame 2 Sample Frame 3 Sample Frame 4 Sample Frame 5
5 Left fall while lying on a bed 30 Subject lies on the bed before left fall Subject lies motionless on the floor

6 Right fall while lying on a bed 30 Subject lies on the bed before right fall Subject lies motionless on the floor

19 Sit down on a bed then stand up 20 Subject stands straight before sitting down on the bed Subject stands straight after standing up

20 Lie on a bed then sit up 20 Subject lies on the bed before sitting up After sitting up

Action Classification

Twenty actions are recorded among which eight actions are falls of different styles and twelve actions are daily and fall-like activities. For the task of fall detection, we categorize the actions into two groups: fall and non-fall. In the group of falls, we have three types of fall, each has different styles and directions. In the group of non-fall activities, we categorize in term of movement of the whole body or only component and direction of movement.

Group Activity Group ID
Fall While walking
While lying on the bed
While sitting on the chair
1,2,3,4
5,6
7,8
Non-fall Horizontal movement of the whole body
Hand and leg movement
Horizontal movement of the whole body
9,10,11,12,13
14,15,16
17,18,19,20

Collected Data and File Formats

In the table below, we provide information on the type and the size of data collected per subject by different sensors in the acquisition system.

System Data Format Collected Data File Name Approximate Size per Subject
Microsoft Kinect AVI
BIN
ASCII
RGB color data
16-bit depth map
UNIX time stamps
color.avi
depth.bin
timestamp.txt
7 GB
Accelerometers ASCII Acceleration in XYZ axes
UNIX time stamps
wrist-worn accelerometer.txt
hip-worn accelerometer.txt
2 MB

Demo Video

??

Download Instructions

Our CMDFALL dataset is available and free for research purpose. If you want to use our dataset, please contact thanh-hai.tran@mica.edu.vn.