Final published version
Research output: Contribution in Book/Report/Proceedings - With ISBN/ISSN › Conference contribution/Paper › peer-review
Publication date | 15/12/2016 |
---|---|
Host publication | Advanced Data Mining and Applications - 12th International Conference, ADMA 2016, Proceedings |
Editors | Jianxin Li, Xue Li, Shuliang Wang, Jinyan Li, Quan Z. Sheng |
Publisher | Springer Verlag |
Pages | 250-265 |
Number of pages | 16 |
ISBN (print) | 9783319495859 |
<mark>Original language</mark> | English |
Externally published | Yes |
Event | 12th International Conference on Advanced Data Mining and Applications, ADMA 2016 - Gold Coast, Australia Duration: 12/12/2016 → 15/12/2016 |
Conference | 12th International Conference on Advanced Data Mining and Applications, ADMA 2016 |
---|---|
Country/Territory | Australia |
City | Gold Coast |
Period | 12/12/16 → 15/12/16 |
Name | Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics) |
---|---|
Volume | 10086 LNAI |
ISSN (Print) | 0302-9743 |
ISSN (electronic) | 1611-3349 |
Conference | 12th International Conference on Advanced Data Mining and Applications, ADMA 2016 |
---|---|
Country/Territory | Australia |
City | Gold Coast |
Period | 12/12/16 → 15/12/16 |
Smartphones are widely available commercial devices and using them as a basis to creates the possibility of future widespread usage and potential applications. This paper utilizes the embedded sensors in a smartphone to recognise a number of common human actions and postures. We group the range of all possible human actions into five basic action classes, namely walking, standing, sitting, crouching and lying. We also consider the postures pertaining to three of the above actions, including standing postures (backward, straight, forward and bend), sitting postures (lean, upright, slouch and rest) and lying postures (back, side and stomach). Training data was collected through a number of people performing a sequence of these actions and postures with a smartphone in their shirt pockets. We analysed and compared three classification algorithms, namely k Nearest Neighbour (kNN), Decision Tree Learning (DTL) and Linear Discriminant Analysis (LDA) in terms of classification accuracy and efficiency (training time as well as classification time). kNN performed the best overall compared to the other two and is believed to be the most appropriate classification algorithm to use for this task. The developed system is in the form of an Android app. Our system can real-time accesses the motion data from the three sensors and on-line classifies a particular action or posture using the kNN algorithm. It successfully recognizes the specified actions and postures with very high precision and recall values of generally above 96%.