Deep Learning for Human Activity Recognition: A Resource Efficient Implementation on Low-Power Devices

Similar documents
Introduction to Machine Learning

Research on Hand Gesture Recognition Using Convolutional Neural Network

Radio Deep Learning Efforts Showcase Presentation

Convolutional neural networks

Smartphone Motion Mode Recognition

Author(s) Corr, Philip J.; Silvestre, Guenole C.; Bleakley, Christopher J. The Irish Pattern Recognition & Classification Society

Counterfeit Bill Detection Algorithm using Deep Learning

DeCAF: A Deep Convolutional Activation Feature for Generic Visual Recognition. ECE 289G: Paper Presentation #3 Philipp Gysel

Training Steps Files File Type File Count Total Size L3 embedding knowledge distillation (SONYC) Google audioset (environmental)

The Jigsaw Continuous Sensing Engine for Mobile Phone Applications!

Generating an appropriate sound for a video using WaveNet.

DERIVATION OF TRAPS IN AUDITORY DOMAIN

Lesson 08. Convolutional Neural Network. Ing. Marek Hrúz, Ph.D. Katedra Kybernetiky Fakulta aplikovaných věd Západočeská univerzita v Plzni.

LANDMARK recognition is an important feature for

PERFORMANCE ANALYSIS OF MLP AND SVM BASED CLASSIFIERS FOR HUMAN ACTIVITY RECOGNITION USING SMARTPHONE SENSORS DATA

Indirect structural health monitoring in bridges: scale experiments

arxiv: v2 [cs.sd] 22 May 2017

Mobile Sensing: Opportunities, Challenges, and Applications

Learning to Unlearn and Relearn Speech Signal Processing using Neural Networks: current and future perspectives

Biologically Inspired Computation

CSC321 Lecture 11: Convolutional Networks

AI Application Processing Requirements

Comparison of ridge- and intensity-based perspiration liveness detection methods in fingerprint scanners

Algorithms for processing accelerator sensor data Gabor Paller

Session 2: 10 Year Vision session (11:00-12:20) - Tuesday. Session 3: Poster Highlights A (14:00-15:00) - Tuesday 20 posters (3minutes per poster)

A Multiple Source Framework for the Identification of Activities of Daily Living Based on Mobile Device Data

PerSec. Pervasive Computing and Security Lab. Enabling Transportation Safety Services Using Mobile Devices

CROSS-LAYER FEATURES IN CONVOLUTIONAL NEURAL NETWORKS FOR GENERIC CLASSIFICATION TASKS. Kuan-Chuan Peng and Tsuhan Chen

A comparative study of different feature sets for recognition of handwritten Arabic numerals using a Multi Layer Perceptron

Tadeusz Stepinski and Bengt Vagnhammar, Uppsala University, Signals and Systems, Box 528, SE Uppsala, Sweden

Advanced Data Analysis Pattern Recognition & Neural Networks Software for Acoustic Emission Applications. Topic: Waveforms in Noesis

DYNAMIC CONVOLUTIONAL NEURAL NETWORK FOR IMAGE SUPER- RESOLUTION

Convolutional Neural Network-based Steganalysis on Spatial Domain

Augmenting Self-Learning In Chess Through Expert Imitation

arxiv: v1 [cs.sd] 12 Dec 2016

Human Activity Recognition using Single Accelerometer on Smartphone Put on User s Head with Head-Mounted Display

SIMULATION-BASED MODEL CONTROL USING STATIC HAND GESTURES IN MATLAB

Deep learning architectures for music audio classification: a personal (re)view

Classification of Road Images for Lane Detection

Deep Neural Network Architectures for Modulation Classification

23270: AUGMENTED REALITY FOR NAVIGATION AND INFORMATIONAL ADAS. Sergii Bykov Technical Lead Machine Learning 12 Oct 2017

On the Use of Convolutional Neural Networks for Specific Emitter Identification

Number Plate Detection with a Multi-Convolutional Neural Network Approach with Optical Character Recognition for Mobile Devices

RESEARCH AND DEVELOPMENT OF DSP-BASED FACE RECOGNITION SYSTEM FOR ROBOTIC REHABILITATION NURSING BEDS

DEEP LEARNING ON RF DATA. Adam Thompson Senior Solutions Architect March 29, 2018

A New Framework for Supervised Speech Enhancement in the Time Domain

Are there alternatives to Sigmoid Hidden Units? MLP Lecture 6 Hidden Units / Initialisation 1

Can you tell a face from a HEVC bitstream?

Rhythmic Similarity -- a quick paper review. Presented by: Shi Yong March 15, 2007 Music Technology, McGill University

Applications of Music Processing

Classifying the Brain's Motor Activity via Deep Learning

CLASSIFICATION OF CLOSED AND OPEN-SHELL (TURKISH) PISTACHIO NUTS USING DOUBLE TREE UN-DECIMATED WAVELET TRANSFORM

Environmental Sound Recognition using MP-based Features

Vehicle Color Recognition using Convolutional Neural Network

Design of Practical Color Filter Array Interpolation Algorithms for Cameras, Part 2

HeadScan: A Wearable System for Radio-based Sensing of Head and Mouth-related Activities

An Introduction to Convolutional Neural Networks. Alessandro Giusti Dalle Molle Institute for Artificial Intelligence Lugano, Switzerland

GESTURE RECOGNITION FOR ROBOTIC CONTROL USING DEEP LEARNING

Deep Neural Networks (2) Tanh & ReLU layers; Generalisation and Regularisation

Improving reverberant speech separation with binaural cues using temporal context and convolutional neural networks

Application of Classifier Integration Model to Disturbance Classification in Electric Signals

Design and Implementation of an Intuitive Gesture Recognition System Using a Hand-held Device

Constructing local discriminative features for signal classification

Integrated Navigation System

Wheel Health Monitoring Using Onboard Sensors

Coursework 2. MLP Lecture 7 Convolutional Networks 1

Comparison of Head Movement Recognition Algorithms in Immersive Virtual Reality Using Educative Mobile Application

Indoor navigation with smartphones

arxiv: v2 [eess.as] 11 Oct 2018

Singing Voice Detection. Applications of Music Processing. Singing Voice Detection. Singing Voice Detection. Singing Voice Detection

Applications and Challenges of Human Activity Recognition using Sensors in a Smart Environment

Satellite and Inertial Attitude. A presentation by Dan Monroe and Luke Pfister Advised by Drs. In Soo Ahn and Yufeng Lu

CP-JKU SUBMISSIONS FOR DCASE-2016: A HYBRID APPROACH USING BINAURAL I-VECTORS AND DEEP CONVOLUTIONAL NEURAL NETWORKS

Hardware-based Image Retrieval and Classifier System

Image Recognition of Tea Leaf Diseases Based on Convolutional Neural Network

Tiny ImageNet Challenge Investigating the Scaling of Inception Layers for Reduced Scale Classification Problems

Learning with Confidence: Theory and Practice of Information Geometric Learning from High-dim Sensory Data

A Novel Fuzzy Neural Network Based Distance Relaying Scheme

CSE 165: 3D User Interaction. Lecture #7: Input Devices Part 2

Driving Using End-to-End Deep Learning

Applications of Machine Learning Techniques in Human Activity Recognition

Neural network pruning for feature selection Application to a P300 Brain-Computer Interface

Learning the Speech Front-end With Raw Waveform CLDNNs

AUGMENTED CONVOLUTIONAL FEATURE MAPS FOR ROBUST CNN-BASED CAMERA MODEL IDENTIFICATION. Belhassen Bayar and Matthew C. Stamm

Study Impact of Architectural Style and Partial View on Landmark Recognition

GESTURE RECOGNITION WITH 3D CNNS

Next Generation Biometric Sensing in Wearable Devices

An Improved Voice Activity Detection Based on Deep Belief Networks

Global Contrast Enhancement Detection via Deep Multi-Path Network

EE 791 EEG-5 Measures of EEG Dynamic Properties

Exploiting the Unused Part of the Brain

Photo Selection for Family Album using Deep Neural Networks

Bio-Metric Authentication of an User using Hand Gesture Recognition

Recurrent neural networks Modelling sequential data. MLP Lecture 9 / 13 November 2018 Recurrent Neural Networks 1: Modelling sequential data 1

Music Genre Classification using Improved Artificial Neural Network with Fixed Size Momentum

Gait Recognition Using WiFi Signals

LIMITING NUMERICAL PRECISION OF NEURAL NETWORKS TO ACHIEVE REAL- TIME VOICE ACTIVITY DETECTION

Image Finder Mobile Application Based on Neural Networks

Assessing the likelihood of GNSS spoofing attacks on RPAS

Pervasive and mobile computing based human activity recognition system

Transcription:

Deep Learning for Human Activity Recognition: A Resource Efficient Implementation on Low-Power Devices Daniele Ravì, Charence Wong, Benny Lo and Guang-Zhong Yang To appear in the proceedings of the IEEE EMBS 13 th Annual International Body Sensor Networks (BSN) Conference, June 2016

Activity Recognition Main Goal Identify human activities (such as walking, jogging, cycling, etc.)

Applications Wellbeing Healthcare Sports

How is it recorded? Inertial sensors Accelerometer Gyroscope Magnetometer Camera GPS Audio Proximity Barometer

Can we use a phone?

What about wearables? Intel Edison Intel Atom dual-core CPU @ 500Hz 1GB DDR3 RAM 4GB emmc Flash Dimensions: 35mm 25mm

Deep Learning For many years, activity recognition approaches have been designed using shallow features Those methods are task dependent and limited With Deep Learning, features are extracted from the training data instead of being handcrafted for a specific application Shallow features

Our approach We propose an approach that combines a descriptive input domain with a deep learning method The method must be efficient It must also be robust against transformations and variations in sensor properties

Our approach

Our approach Spectrogram A spectrogram of an inertial signal is a new representation of the signal as a function of frequency and time The procedure for computing the spectrogram is to divide a longer time signal into shorter segments of equal length and then compute the Fourier transform separately on each shorter segment This results in a representation that describes the changing spectra as a function of time The representation captures the inertial input signal whilst providing a form of temporal and sample rate invariance

Our approach

Our approach Temporal convolution layer Each filter w i is applied to the spectrogram vertically and the weighted sum of the convolved signal is computed as follows: o[t][i] = Xst kwx j=1 k=1 w[i][j][k] input[dw (t 1) + k][j] These temporal convolutions produce an output layer of learned features with a small size for real-time processing This provides orientation invariance to the input signal The last two layers are used to finally classify the features

Training We use the error value in the backward propagation routine to update each weight of the network through the Stochastic Gradient Descent (SGD) approach To improve the training procedure of the weights, we have used 3 regularisations: Weight decay causes the weights to exponentially decay to zero if no other update is scheduled to avoid over-fitting Momentum accelerates gradient descent to move the global minimum of the function Dropout removes units randomly from the neural network to lower generalisation error

Datasets Four datasets are considered in our analysis We also release a new dataset called ActiveMiles It contains unconstrained real world data from 10 subjects It is one of the largest dataset (around 30 hours of labelled raw data) It is the first database that contains data captured using different devices

Experimental setup Classification accuracy changes when the spectrogram generation parameters are modified 98 X: 3 Y: 56 Z: 97.16 96 Accuracy (%) 94 92 90 88 2 # of time localised points 4 6 8 10 12 30 40 50 60 70 # of frequencies 80

Experimental setup The optimal size of the temporal convolution kernel is two or three, depending on the data being classified The proposed approach requires few levels in order to obtain good results ActiveMiles WISDM-v1 D-FoG Skoda Accuracy 99 98 97 96 95 94 93 92 91 90 89 88 1 2 3 4 5 Filter Size Accuracy 99 98 97 96 95 94 93 92 91 90 89 1 2 3 #of Levels

Experimental setup 15 filters and just 80 nodes in the fully convolution layer are sufficient for a good classification Accuracy 99 98 97 96 95 94 93 92 91 90 89 88 87 12 22 32 42 52 # of Filters 80 Nodes - WISDM-v1 50 Nodes - WISDM-v1 80 Nodes - ActiveMiles 50 Nodes - ActiveMiles 80 Nodes - D-FoG 50 Nodes - D-FoG 80 Nodes - Skoda 50 Nodes - Skoda

Results A comparison of HAR results using four baselines, existing methods, and the considered datasets are shown in Table III The accuracy of the proposed method is typically better in comparison to the other methods

Performance A comparison of the computation times required to classify activities on different devices All resulting times are within the requirements for real-time processing

Conclusion The proposed system generates discriminative features that are generally more powerful than handcrafted features The accuracy of the proposed approach is better or comparable against existing state-of-the-art methods The ability of the proposed method to generalise across different classification tasks is demonstrated using a variety of human activity datasets The computation times obtained from low-power devices are consistent with real-time processing

Thank you for your attention!