Scaling Up Sign Spotting Through Sign Language Dictionaries
dc.contributor.author | Varol, G | |
dc.contributor.author | Momeni, L | |
dc.contributor.author | Albanie, S | |
dc.contributor.author | Afouras, T | |
dc.contributor.author | Zisserman, A | |
dc.date.accessioned | 2022-05-27T16:08:25Z | |
dc.date.available | 2022-05-27T16:08:25Z | |
dc.date.issued | 2022-06 | |
dc.date.submitted | 2021-05-01 | |
dc.identifier.issn | 0920-5691 | |
dc.identifier.other | s11263-022-01589-6 | |
dc.identifier.other | 1589 | |
dc.identifier.uri | https://www.repository.cam.ac.uk/handle/1810/337554 | |
dc.description.abstract | <jats:title>Abstract</jats:title><jats:p>The focus of this work is <jats:italic>sign spotting</jats:italic>–given a video of an isolated sign, our task is to identify <jats:italic>whether</jats:italic> and <jats:italic>where</jats:italic> it has been signed in a continuous, co-articulated sign language video. To achieve this sign spotting task, we train a model using multiple types of available supervision by: (1) <jats:italic>watching</jats:italic> existing footage which is sparsely labelled using mouthing cues; (2) <jats:italic>reading</jats:italic> associated subtitles (readily available translations of the signed content) which provide additional <jats:italic>weak-supervision</jats:italic>; (3) <jats:italic>looking up</jats:italic> words (for which no co-articulated labelled examples are available) in visual sign language dictionaries to enable novel sign spotting. These three tasks are integrated into a unified learning framework using the principles of Noise Contrastive Estimation and Multiple Instance Learning. We validate the effectiveness of our approach on low-shot sign spotting benchmarks. In addition, we contribute a machine-readable British Sign Language (BSL) dictionary dataset of isolated signs, <jats:sc>BslDict</jats:sc>, to facilitate study of this task. The dataset, models and code are available at our project page. </jats:p> | |
dc.language | en | |
dc.publisher | Springer Science and Business Media LLC | |
dc.subject | Article | |
dc.subject | Special issue on Advances in Computer Vision and Applications | |
dc.subject | Sign language recognition | |
dc.subject | Sign spotting | |
dc.subject | Few-shot learning | |
dc.title | Scaling Up Sign Spotting Through Sign Language Dictionaries | |
dc.type | Article | |
dc.date.updated | 2022-05-27T16:08:25Z | |
prism.endingPage | 1439 | |
prism.issueIdentifier | 6 | |
prism.publicationName | International Journal of Computer Vision | |
prism.startingPage | 1416 | |
prism.volume | 130 | |
dc.identifier.doi | 10.17863/CAM.84963 | |
dcterms.dateAccepted | 2022-01-21 | |
rioxxterms.versionofrecord | 10.1007/s11263-022-01589-6 | |
rioxxterms.version | VoR | |
rioxxterms.licenseref.uri | http://creativecommons.org/licenses/by/4.0/ | |
dc.contributor.orcid | Varol, G [0000-0002-8438-6152] | |
dc.identifier.eissn | 1573-1405 | |
cam.issuedOnline | 2022-04-05 |
Files in this item
This item appears in the following Collection(s)
-
Jisc Publications Router
This collection holds Cambridge publications received from the Jisc Publications Router