Abstract
Signed languages are the primary means of communication for many deaf and hard of hearing individuals. Since signed languages exhibit all the fundamental linguistic properties of natural language, we believe that tools and theories of Natural Language Processing (NLP) are crucial towards its modeling. However, existing research in Sign Language Processing (SLP) seldom attempt to explore and leverage the linguistic organization of signed languages. This position paper calls on the NLP community to include signed languages as a research area with high social and scientific impact. We first discuss the linguistic properties of signed languages to consider during their modeling. Then, we review the limitations of current SLP models and identify the open challenges to extend NLP to signed languages. Finally, we urge (1) the adoption of an efficient tokenization method; (2) the development of linguistically-informed models; (3) the collection of real-world signed language data; (4) the inclusion of local signed language communities as an active and leading voice in research.
Original language | English |
---|---|
Title of host publication | Proceedings of the 31st International Joint Conference on Artificial Intelligence, IJCAI 2022 |
Editors | Luc De Raedt, Luc De Raedt |
Publisher | International Joint Conferences on Artificial Intelligence |
Pages | 5369-5373 |
Number of pages | 5 |
ISBN (Electronic) | 9781956792003 |
DOIs | |
State | Published - 2022 |
Event | 31st International Joint Conference on Artificial Intelligence, IJCAI 2022 - Vienna, Austria Duration: 23 Jul 2022 → 29 Jul 2022 |
Publication series
Name | IJCAI International Joint Conference on Artificial Intelligence |
---|---|
ISSN (Print) | 1045-0823 |
Conference
Conference | 31st International Joint Conference on Artificial Intelligence, IJCAI 2022 |
---|---|
Country/Territory | Austria |
City | Vienna |
Period | 23/07/22 → 29/07/22 |
Bibliographical note
Publisher Copyright:© 2022 International Joint Conferences on Artificial Intelligence. All rights reserved.