A Semi-Automatic Annotation Approach for Human Activity Recognition
Modern smartphones and wearables often contain multiple embedded sensors which generate significant amounts of data. This information can be used for body monitoring-based areas such as healthcare, indoor location, user-adaptive recommendations and transportation. The development of Human Activity R...
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2019-01-01
|
Series: | Sensors |
Subjects: | |
Online Access: | https://www.mdpi.com/1424-8220/19/3/501 |
id |
doaj-1fdf68fa7abe446481b37b196af64f65 |
---|---|
record_format |
Article |
spelling |
doaj-1fdf68fa7abe446481b37b196af64f652020-11-24T20:47:26ZengMDPI AGSensors1424-82202019-01-0119350110.3390/s19030501s19030501A Semi-Automatic Annotation Approach for Human Activity RecognitionPatrícia Bota0Joana Silva1Duarte Folgado2Hugo Gamboa3Associação Fraunhofer Portugal Research, Rua Alfredo Allen 455/461, 4200-135 Porto, PortugalAssociação Fraunhofer Portugal Research, Rua Alfredo Allen 455/461, 4200-135 Porto, PortugalAssociação Fraunhofer Portugal Research, Rua Alfredo Allen 455/461, 4200-135 Porto, PortugalAssociação Fraunhofer Portugal Research, Rua Alfredo Allen 455/461, 4200-135 Porto, PortugalModern smartphones and wearables often contain multiple embedded sensors which generate significant amounts of data. This information can be used for body monitoring-based areas such as healthcare, indoor location, user-adaptive recommendations and transportation. The development of Human Activity Recognition (HAR) algorithms involves the collection of a large amount of labelled data which should be annotated by an expert. However, the data annotation process on large datasets is expensive, time consuming and difficult to obtain. The development of a HAR approach which requires low annotation effort and still maintains adequate performance is a relevant challenge. We introduce a Semi-Supervised Active Learning (SSAL) based on Self-Training (ST) approach for Human Activity Recognition to partially automate the annotation process, reducing the annotation effort and the required volume of annotated data to obtain a high performance classifier. Our approach uses a criterion to select the most relevant samples for annotation by the expert and propagate their label to the most confident samples. We present a comprehensive study comparing supervised and unsupervised methods with our approach on two datasets composed of daily living activities. The results showed that it is possible to reduce the required annotated data by more than 89% while still maintaining an accurate model performance.https://www.mdpi.com/1424-8220/19/3/501human activity recognitionmachine learningactive learningsemi-supervised learningtime seriesself-training |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Patrícia Bota Joana Silva Duarte Folgado Hugo Gamboa |
spellingShingle |
Patrícia Bota Joana Silva Duarte Folgado Hugo Gamboa A Semi-Automatic Annotation Approach for Human Activity Recognition Sensors human activity recognition machine learning active learning semi-supervised learning time series self-training |
author_facet |
Patrícia Bota Joana Silva Duarte Folgado Hugo Gamboa |
author_sort |
Patrícia Bota |
title |
A Semi-Automatic Annotation Approach for Human Activity Recognition |
title_short |
A Semi-Automatic Annotation Approach for Human Activity Recognition |
title_full |
A Semi-Automatic Annotation Approach for Human Activity Recognition |
title_fullStr |
A Semi-Automatic Annotation Approach for Human Activity Recognition |
title_full_unstemmed |
A Semi-Automatic Annotation Approach for Human Activity Recognition |
title_sort |
semi-automatic annotation approach for human activity recognition |
publisher |
MDPI AG |
series |
Sensors |
issn |
1424-8220 |
publishDate |
2019-01-01 |
description |
Modern smartphones and wearables often contain multiple embedded sensors which generate significant amounts of data. This information can be used for body monitoring-based areas such as healthcare, indoor location, user-adaptive recommendations and transportation. The development of Human Activity Recognition (HAR) algorithms involves the collection of a large amount of labelled data which should be annotated by an expert. However, the data annotation process on large datasets is expensive, time consuming and difficult to obtain. The development of a HAR approach which requires low annotation effort and still maintains adequate performance is a relevant challenge. We introduce a Semi-Supervised Active Learning (SSAL) based on Self-Training (ST) approach for Human Activity Recognition to partially automate the annotation process, reducing the annotation effort and the required volume of annotated data to obtain a high performance classifier. Our approach uses a criterion to select the most relevant samples for annotation by the expert and propagate their label to the most confident samples. We present a comprehensive study comparing supervised and unsupervised methods with our approach on two datasets composed of daily living activities. The results showed that it is possible to reduce the required annotated data by more than 89% while still maintaining an accurate model performance. |
topic |
human activity recognition machine learning active learning semi-supervised learning time series self-training |
url |
https://www.mdpi.com/1424-8220/19/3/501 |
work_keys_str_mv |
AT patriciabota asemiautomaticannotationapproachforhumanactivityrecognition AT joanasilva asemiautomaticannotationapproachforhumanactivityrecognition AT duartefolgado asemiautomaticannotationapproachforhumanactivityrecognition AT hugogamboa asemiautomaticannotationapproachforhumanactivityrecognition AT patriciabota semiautomaticannotationapproachforhumanactivityrecognition AT joanasilva semiautomaticannotationapproachforhumanactivityrecognition AT duartefolgado semiautomaticannotationapproachforhumanactivityrecognition AT hugogamboa semiautomaticannotationapproachforhumanactivityrecognition |
_version_ |
1716810053000888320 |