Abstract
Sign language processing is often performed by processing each individual sign. Such an approach relies on an exhaustive description of the signs and does not take in account the spatial structure of the sentence. In this paper, we will present a general model of sign language sentences that uses the construction of the signing space as a representation of both the meaning and the realisation of the sentence. We will propose a computational model of this construction and explain how it can be attached to a sign language grammar model to help both analysis and generation of sign language utterances.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
Similar content being viewed by others
References
Starner, T., Pentland, A.: Real-time american sign language recognition from video using hidden markov models. Technical Report TR-375, M.I.T Media Laboratory Perceptual Computing Section (1995)
Braffort, A.: Argo: An architecture for sign language recognition and interpretation. In: Harling, P., et al. (eds.) Progress in Gestual Interaction, pp. 17–30. Springer, Heidelberg (1996)
Cuxac, C.: French sign language: proposition of a structural explanation by iconicity. In: Braffort, A., Gibet, S., Teil, D., Gherbi, R., Richardson, J. (eds.) GW 1999. LNCS (LNAI), vol. 1739, pp. 165–184. Springer, Heidelberg (2000)
Cuxac, C.: La langue des Signes française. Les voies de l’iconicité. ISBN 2-7080- 0952-4. Faits de langue, Ophrys, Paris (2000)
Stokoe, W., Casterline, D., Croneberg, C.: A dictionary of american sign language. Gallaudet College Press (1965)
Vogler, C., Metaxas, D.: Toward scalability in ASL recognition: Breaking down sign into phonemes. In: Braffort, A., Gibet, S., Teil, D., Gherbi, R., Richardson, J. (eds.) GW 1999. LNCS (LNAI), vol. 1739, p. 211. Springer, Heidelberg (2000)
Vogler, C., Metaxas, D.: Handshapes and movements: Multiple-channel asl recognition. In: Volpe, G., Camurri, A. (eds.) Proceedings of the Gesture Workshop 2003, Genova, Italy, April 15-17. LNCS (LNAI), vol. 2915, pp. 247–258. Springer, Heidelberg (2003)
Dalle, P., Lenseigne, B.: Vision-based sign language processing using a predictive approach ans linguistic knowledge. In: IAPR conference on Machine vision and application (MVA 2005), Tsukuba Science City, pp. 510–513 (May 2005)
Desclés, J.P.: La prédication opérée par les langues. In: Langages, pp. 83–97. Larousse, Paris (1991)
Bangham, J.A., et al.: An overview of ViSiCAST. In: IEEE Seminar on Speech and language processing for disabled and elderly people, London (April 2000)
Elliot, R., et al.: An overview of the SiGML and SiGMLsigning software system. In: Workshop on Representation and Processing of Sign Language. 4th Internationnal Conference on Language Resources and Evaluation (LREC 2004), Lisbone, Portugal, May 30, pp. 98–103 (2004)
Lenseigne, B., Gianni, F., Dalle, P.: A new gesture representation for sign language analysis. In: Workshop on Representation and Processing of Sign Language. 4th Internationnal Conference on Language Resources and Evaluation (LREC 2004), Lisbon, Portugal, May 30, pp. 85–90 (2004)
Huenerfauth, M.: Spatial representation of classifier predicates for machine translation into american sign language. In: Workshop on Representation and Processing of Sign Language, 4th Internationnal Conference on Language Ressources and Evaluation (LREC 2004), Lisbon, Portugal, May 30, pp. 24–31 (2004)
Langacker, R.: Foundations of cognitive grammar: Theoretical Prerequisites, vol. 1. Stanford University Press, Stanford (1987) (Hardcover, 1999) (Paperback)
Lenseigne, B.: Intégration de connaissances linguistiques dans un système de vision. Application à l’étude de la langue des Signes. PhD thesis, Université Paul Sabatier, Toulouse, décembre (2004)
Lenseigne, B., Dalle, P.: A model for sign language grammar. In: 2nd Language and technology Conference, Poznan, Poland, April 21-23 (2005)
Whyte, R.N., Somers, M.G.: Hand posture matching for irish sign language interpretation. In: ACM Proceedings of the 1st international symposium on Information and communication technologies, Dublin Ireland, September 24-26, pp. 439–444. Trinity College Dublin (2003)
Liang, R.H., Ouhyoung, M.: A real-time continuous gesture recognition system for sign language. In: 3rd International conference on automatic face and gesture recognition, Nara, Japan, pp. 558–565 (1998)
Sagawa, H., Takeuchi, M., Ohki, M.: Description and recognition methods for sign language based on gesture components. In: Proceedings of IUI, orlando, Florida (1997)
Prillwitz, S., et al.: HamNoSys. Version 2.0; Hamburg Notation System for Sign Languages. An introductory guide. Signum, Hamburg (1989)
Liddell, S., Johnson, R.E.: American sign language: The phonological base. Sign Language Studies 64, 195–227 (1989)
Gibet, S., Lebourque, T.: High level specification and control of communication gestures: the GESSYCA system. In: Computer Animation 1999, Geneva, May 26-28, Computer Graphics Society (CGS) and the IEEE Computer Society (1999)
Thom, R.: Stabilité structurelle et morphogénèse. Édisciences, 2nd edn., Paris, pp. 1972–1977
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2006 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Lenseigne, B., Dalle, P. (2006). Using Signing Space as a Representation for Sign Language Processing. In: Gibet, S., Courty, N., Kamp, JF. (eds) Gesture in Human-Computer Interaction and Simulation. GW 2005. Lecture Notes in Computer Science(), vol 3881. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11678816_3
Download citation
DOI: https://doi.org/10.1007/11678816_3
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-32624-3
Online ISBN: 978-3-540-32625-0
eBook Packages: Computer ScienceComputer Science (R0)