Continuous Sign Language Recognition Through Cross-Modal Alignment of Video and Text Embeddings in a Joint-Latent Space
Continuous Sign Language Recognition (CSLR) refers to the challenging problem of recognizing sign language glosses and their temporal boundaries from weakly annotated video sequences. Previous methods focus mostly on visual feature extraction neglecting text information and failing to effectively mo...
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
IEEE
2020-01-01
|
Series: | IEEE Access |
Subjects: | |
Online Access: | https://ieeexplore.ieee.org/document/9090828/ |
id |
doaj-670962fc1cf947efa1a47fd7b92d156d |
---|---|
record_format |
Article |
spelling |
doaj-670962fc1cf947efa1a47fd7b92d156d2021-03-30T02:42:28ZengIEEEIEEE Access2169-35362020-01-018911709118010.1109/ACCESS.2020.29936509090828Continuous Sign Language Recognition Through Cross-Modal Alignment of Video and Text Embeddings in a Joint-Latent SpaceIlias Papastratis0https://orcid.org/0000-0003-4664-2626Kosmas Dimitropoulos1https://orcid.org/0000-0003-1584-7047Dimitrios Konstantinidis2Petros Daras3https://orcid.org/0000-0003-3814-6710Visual Computing Lab, Centre for Research and Technology Hellas-Information Technologies Institute, Thessaloniki, GreeceVisual Computing Lab, Centre for Research and Technology Hellas-Information Technologies Institute, Thessaloniki, GreeceVisual Computing Lab, Centre for Research and Technology Hellas-Information Technologies Institute, Thessaloniki, GreeceVisual Computing Lab, Centre for Research and Technology Hellas-Information Technologies Institute, Thessaloniki, GreeceContinuous Sign Language Recognition (CSLR) refers to the challenging problem of recognizing sign language glosses and their temporal boundaries from weakly annotated video sequences. Previous methods focus mostly on visual feature extraction neglecting text information and failing to effectively model the intra-gloss dependencies. In this work, a cross-modal learning approach that leverages text information to improve vision-based CSLR is proposed. To this end, two powerful encoding networks are initially used to produce video and text embeddings prior to their mapping and alignment into a joint latent representation. The purpose of the proposed cross-modal alignment is the modelling of intra-gloss dependencies and the creation of more descriptive video-based latent representations for CSLR. The proposed method is trained jointly with video and text latent representations. Finally, the aligned video latent representations are classified using a jointly trained decoder. Extensive experiments on three well-known sign language recognition datasets and comparison with state-of-the-art approaches demonstrate the great potential of the proposed approach.https://ieeexplore.ieee.org/document/9090828/Computer visioncontinuous sign language recognitioncross-modal learningdeep-learningjoint latent space |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Ilias Papastratis Kosmas Dimitropoulos Dimitrios Konstantinidis Petros Daras |
spellingShingle |
Ilias Papastratis Kosmas Dimitropoulos Dimitrios Konstantinidis Petros Daras Continuous Sign Language Recognition Through Cross-Modal Alignment of Video and Text Embeddings in a Joint-Latent Space IEEE Access Computer vision continuous sign language recognition cross-modal learning deep-learning joint latent space |
author_facet |
Ilias Papastratis Kosmas Dimitropoulos Dimitrios Konstantinidis Petros Daras |
author_sort |
Ilias Papastratis |
title |
Continuous Sign Language Recognition Through Cross-Modal Alignment of Video and Text Embeddings in a Joint-Latent Space |
title_short |
Continuous Sign Language Recognition Through Cross-Modal Alignment of Video and Text Embeddings in a Joint-Latent Space |
title_full |
Continuous Sign Language Recognition Through Cross-Modal Alignment of Video and Text Embeddings in a Joint-Latent Space |
title_fullStr |
Continuous Sign Language Recognition Through Cross-Modal Alignment of Video and Text Embeddings in a Joint-Latent Space |
title_full_unstemmed |
Continuous Sign Language Recognition Through Cross-Modal Alignment of Video and Text Embeddings in a Joint-Latent Space |
title_sort |
continuous sign language recognition through cross-modal alignment of video and text embeddings in a joint-latent space |
publisher |
IEEE |
series |
IEEE Access |
issn |
2169-3536 |
publishDate |
2020-01-01 |
description |
Continuous Sign Language Recognition (CSLR) refers to the challenging problem of recognizing sign language glosses and their temporal boundaries from weakly annotated video sequences. Previous methods focus mostly on visual feature extraction neglecting text information and failing to effectively model the intra-gloss dependencies. In this work, a cross-modal learning approach that leverages text information to improve vision-based CSLR is proposed. To this end, two powerful encoding networks are initially used to produce video and text embeddings prior to their mapping and alignment into a joint latent representation. The purpose of the proposed cross-modal alignment is the modelling of intra-gloss dependencies and the creation of more descriptive video-based latent representations for CSLR. The proposed method is trained jointly with video and text latent representations. Finally, the aligned video latent representations are classified using a jointly trained decoder. Extensive experiments on three well-known sign language recognition datasets and comparison with state-of-the-art approaches demonstrate the great potential of the proposed approach. |
topic |
Computer vision continuous sign language recognition cross-modal learning deep-learning joint latent space |
url |
https://ieeexplore.ieee.org/document/9090828/ |
work_keys_str_mv |
AT iliaspapastratis continuoussignlanguagerecognitionthroughcrossmodalalignmentofvideoandtextembeddingsinajointlatentspace AT kosmasdimitropoulos continuoussignlanguagerecognitionthroughcrossmodalalignmentofvideoandtextembeddingsinajointlatentspace AT dimitrioskonstantinidis continuoussignlanguagerecognitionthroughcrossmodalalignmentofvideoandtextembeddingsinajointlatentspace AT petrosdaras continuoussignlanguagerecognitionthroughcrossmodalalignmentofvideoandtextembeddingsinajointlatentspace |
_version_ |
1724184738082586624 |