Online Information about Wearable Fall Detection
1. Public Dataset - G-Sensor
1.1 UP-Fall Detection Dataset
A paper called “UP-Fall Detection Dataset: A Multimodal Approach (download PDF)”, the authors made the data available publicly for access at http://sites.google.com/up.edu.mx/har-up/.
Another paper Fall Detection: Threshold Analysis of Wrist-Worn Motion Sensor Signals, utilize a subset of the publicly available dataset - UP-Fall Detection dataset - to analyze acceleration and angular velocity signals measured on wrist-worn sensors to simulate smartwatch placement and give a 3-phased threshold analysis model approach (AVC to K-Means to SVM) whose accuracy is about 78 percent.
1.2 UR Fall Detection Dataset
This dataset contains 70 (30 falls + 40 activities of daily living) sequences. Fall events are recorded with 2 Microsoft Kinect cameras and corresponding accelerometric data. ADL events are recorded with only one device (camera 0) and accelerometer. Sensor data was collected using PS Move (60Hz) and x-IMU (256Hz) devices.
The related paper is:
Bogdan Kwolek, Michal Kepski, Human fall detection on embedded platform using depth maps and wireless accelerometer, Computer Methods and Programs in Biomedicine, Volume 117, Issue 3, December 2014, Pages 489-501, ISSN 0169-2607.
The above paper “Human fall detection on embedded platform using depth maps and wireless accelerometer” can be downloaded from HERE.
1.3 UMAFall: Fall Detection Dataset
The files contain the mobility traces generated by a group of 19 experimental subjects that emulated a set of predetermined ADL (Activities of Daily Life) and falls. The traces are aimed at evaluating fall detection algorithms.
Several video clips describing the performed movements are also included.
The source and authors of this publicly available dataset should be acknowledged in all publications in which it is utilized as by referencing any of the following papers as well as this web-site:
Santoyo-Ramón, José Antonio, Eduardo Casilari, and José Manuel Cano-García. “Analysis of a smartphone-based architecture with multiple mobility sensors for fall detection with supervised learning.” Sensors 18.4 (2018): 1155. Casilari, Eduardo, Jose A. Santoyo-Ramón, and Jose M. Cano-García. “UMAFall: A Multisensor Dataset for the Research on Automatic Fall Detection.” Procedia Computer Science 110 (2017): 32-39.
The paper could be downloaded from https://core.ac.uk/reader/132743211.
Below is the description of how the data were collected, and the dataset can also be downloaded from the link: http://webpersonal.uma.es/de/ECASILARI/Fall_ADL_Traces/UMA_FALL_ADL_dataset.html
1.4 machine-learning-databases
The dataset can be downloaded from below https://archive.ics.uci.edu/ml/machine-learning-databases/00455/
The related paper is named Simulated Falls and Daily Living Activities Data Set Data Set.
1.4 SmartFall Dataset
The related paper can be downloaded from SmartFall: A Smartwatch-Based Fall Detection System Using Deep Learning.
The description of the paper can be found from below link: https://github.com/IKKIM00/fall-detection-and-predction-using-GRU-and-LSTM-with-Transfer-Learning.
The dataset can be found from here: https://userweb.cs.txstate.edu/~hn12/data/SmartFallDataSet/, which includes three folders:
1.5 MobiAct Dataset
2. Public Dataset - Camera
2.1 Fall Detection Dataset
The datasets that are used for the simulation purpose are raw RGB and Depth images of size 320x240 recorded from a single uncalibrated Kinect sensor after resizing from 640x480. The Kinect sensor is fixed at roof height of approx 2.4m. The datasets contain a total of 21499 images. Out of total datasets of 22636 images, 16794 images can be used for training, 3299 images can be used for validation and 2543 images can be used for the test. The images in the dataset are recorded in 5 different rooms which consist of 8 different view angles. There are 5 different participants out of which there are two male participants of age 32 and 50 and three female participants of age 19, 28 and 40. All the activities of the participants represent 5 different categories of poses that are standing, sitting, lying, bending and crawling. There is only one participant in each image. Some images in the datasets are empty which are categorised as ‘other’. We have used images of 2 participants: the male of age 32 and the female of age 28 combining total of 16794 images for training, and 3299 images for validation which contains a male participant of age 32 from training set but is in a different room to that of training and testing set. Similarly, the test set contains images of 3 participants out of which 2 female participants are of age 19 and 40 and a male participant is of age 50. These images are recorded in a different room that is not seen in training or validation set. These total of 22636 images are in sequence but have not repeated anywhere in the sequence and all the sets have original and its horizontal flipped images added in sequence to increase the number of images in a set.
The related paper is HERE - PDF version.
K. Adhikari, H. Bouchachia and H. Nait-Charif, “Activity recognition for indoor fall detection using convolutional neural network,” 2017 Fifteenth IAPR International Conference on Machine Vision Applications (MVA), 2017, pp. 81-84, doi: 10.23919/MVA.2017.7986795.
2.2 Multiple cameras fall dataset
This dataset contain 24 scenarios recorded with 8 IP video cameras. The first 22 first scenarios contain a fall and confounding events, the last 2 ones contain only confounding events.
All technical information is given in this pdf file. All videos are gathered in one zipped file in this file.
E. Auvinet, C. Rougier, J.Meunier, A. St-Arnaud, J. Rousseau, “Multiple cameras fall dataset”, Technical report 1350, DIRO - Université de Montréal, July 2010.
2.3 Le2i Fall detection Dataset
Le2i Fall detection Dataset百度网盘链接:https://pan.baidu.com/s/188k4fKlQVKa0Jf7lr8qukg 提取码:qnu0
Note: the original data source is not available now. i.e., one cannot open this site: http://le2i.cnrs.fr/Fall-detection-Dataset?lang=fr
数据集大小:共有191个视频,视频中存在假摔部分,且存在视频中没有人的帧数
数据集内容:
- Video,视频部分
- Annotations,标签格式为txt格式
- 标签前两行为跌倒帧数的起始与终止点,也存在没有这两行的文件
- 之后一行就是六个字段,帧编号, XX,人物bbx
- DATASET (Home, Coffee room, Office, Lecture room)
- LE2I DIJON UMR6306
- FORMAT 320x240 25FPS
- For each subset GROUNDTRUTH in 'Annotation_files' folder
- For each video a ground truth file is given 'video (i).txt'
- each file contains:
- the frame number of the beginning of the fall
- the frame number of the end of the fall
- the height, the width and the coordinates of the center of the bounding box for each frame.
I. Charfi, J. Miteran, J. Dubois, M. Atri, and R. Tourki (2013), Optimized spatio-temporal descriptors for real-time fall detection: comparison of support vector machine and adaboost-based classification. Journal of Electronic Imaging 22 (4), pp. 041106.