Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/1044588.1044623acmconferencesArticle/Chapter ViewAbstractPublication PagessiggraphConference Proceedingsconference-collections
Article

Language-support system using character recognition

Published: 16 June 2004 Publication History

Abstract

We have developed a prototype system that supports language translation for distance lecturing and local learning. The system automatically detects a word displayed in a video image by matching the identified character markers; it then translates the word into the other language and displays the translated word using text, image, and/or sound. The ARToolkit is used as an image-processing tool to search the video image for character markers, which are composed of square frames and registered characters. The use of image processing enables any language to be used as the source language; Japanese is used in the prototype. The use of video makes it easy to implement the system in existing telecommunication systems, because a user just prepares a PC for capturing video images. A Web browser is used as the presentation tool; it can handle any content format normally supported by a Web browser: images, sounds, 3D models, and characters. An automatic voice function automatically reads out words on user demand. The software consists of a character-recognition part (server) and a display part (client); the server controls the client through socket communications. This software architecture makes it possible for one lecturer to simultaneously teach students with various mother tongues, when the clients treat multi-language, because there is a different client for each student. Testing of this language-support system for distance lecturing between Thailand and Japan using a satellite communication system showed that it could be used for both fundamental language lecturing and language translation during international exchange.

References

[1]
Y. Ariki and J. Ogata, English CALL system with functions of speech segmentation and pronunciation evaluation using speech recognition technology, Proc. International Conference on Spoken Language Processing (ICSLP2002), vol. II, pp. 1229--1232, 2002.
[2]
ARToolkit, homepage: http://www.hitl.washington.edu/artoolkit/, as of 2003.
[3]
R. Azuma, A survey of augmented reality, Presence: Teleoperators and Virtual Environments, vol. 6, no. 4, pp. 355--385, 1997.
[4]
R. Azuma, Y. Baillot, R. Behringer, S. Feiner, S. Julier, and B. MacIntyre, Recent advances in augmented reality, IEEE Computer Graphics & Applications, pp. 34--47, 2001.
[5]
M. Billinghurst, Real world teleconferencing, IEEE Computer Graphics & Applications, vol. 22, no. 6, pp. 11--13, 2002.
[6]
M. Billinghurst, H. Kato, and I. Poupyrev, The MagicBook: Moving seamlessly between reality and virtuality, IEEE Computer Graphics and Applications, pp. 2--4, 2001.
[7]
K. Curran, A web-based collaboration teaching environment, IEEE Multimedia, pp. 72--76, 2002.
[8]
S. DiVerdi, D. Nurmi, and T. Hollerer, ARWin - A desktop augmented reality window manager, Proc. International Symposium on Mixed and Augmented Reality (ISMAR2003), pp. 298--299, 2003.
[9]
N. Hedley, L. Postner, R. May, M. Billinghurst, and H. Kato, Collaborative AR for geographic visualization, Proc. International Symposium on Mixed Reality (ISMR2001), pp. 11--18, 2001.
[10]
C. Ishii, N. Minematsu, K. Hirose, and R. Nishide, Identification of accent and intonation in sentences for CALL systems, Proc Eurospeech 2001, pp. 2455--2458, 2001.
[11]
H. Kato, M. Billinghurst, I. Poupyrev, K. Imamoto, and K. Tachibana, Virtual object manipulation on a table-top AR environment, Proc. International Symposium on Augmented Reality (ISAR 2000), pp. 111--119, 2000.
[12]
P. Milgram, H. Takemura, A. Utsumi, and F. Kishino, Augmented reality: a class of displays on the reality-virtuality continuum, SPIE proceedings: Telemanipulator and Telepresence Technologies, vol. 2351, pp. 282--292, 1994.
[13]
N. Nesselhauf and C. Tschichold, Collocations in CALL, An investigation of vocabulary-building software for EFL, International Journal of Computer Assisted Language Learning, vol. 15, no. 3, pp. 251--279, 2002.
[14]
H. Prendinger and M. Ishizuka, Let's talk! Socially intelligent agents for language conversation training, IEEE Trans. System, Man and Cybernetics, vol. 31, no. 5, pp. 465--471, 2001.
[15]
H. Regenbrecht, C. Ott, M. Wagner, T. Lum, P. Kohler, W. Wilke, and E. Mueller, An augmented virtuality approach to 3D videoconferencing, Proc. The Second IEEE and ACM International Symposium on Mixed and Augmented Reality (ISMAR2003), pp. 290--291, 2003.
[16]
S. G. Schar and H. Krueger, Using new learning technologies with multimedia, IEEE Multimedia, pp. 40--51, 2000.
[17]
D. Schmalstieg, A. Fuhrmann, G. Hesina, Z. Szalavari, L. M. Encarnacao, M. Gervautz, and W. Purgathofer, The studierstube augmented reality project, Presence: Teleoperators and Virtual Environments, vol. 11, pp. 33--54, 2002.
[18]
J. Vaupel and M. Sommer, Multimedia education, distance learning, and electronic commerce, Proc. International Conference on Virtual Systems and Multimedia (VSMM'97), pp. 174--175, 1997.
[19]
D. Wagner and I. Barakonyi, Augmented Reality Kanji Learning, Proc. The Second IEEE and ACM International Symposium on Mixed and Augmented Reality (ISMAR2003), pp. 335--336, 2003.

Cited By

View all
  • (2021)Emerging ExG-based NUI Inputs in Extended Realities: A Bottom-up SurveyACM Transactions on Interactive Intelligent Systems10.1145/345795011:2(1-49)Online publication date: 21-Jul-2021

Recommendations

Comments

Information & Contributors

Information

Published In

cover image ACM Conferences
VRCAI '04: Proceedings of the 2004 ACM SIGGRAPH international conference on Virtual Reality continuum and its applications in industry
June 2004
493 pages
ISBN:1581138849
DOI:10.1145/1044588
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected]

Sponsors

Publisher

Association for Computing Machinery

New York, NY, United States

Publication History

Published: 16 June 2004

Permissions

Request permissions for this article.

Check for updates

Author Tags

  1. ARToolkit
  2. character recognition
  3. distance lecture
  4. language-support
  5. videoconference

Qualifiers

  • Article

Acceptance Rates

Overall Acceptance Rate 51 of 107 submissions, 48%

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • Downloads (Last 12 months)7
  • Downloads (Last 6 weeks)1
Reflects downloads up to 03 Feb 2025

Other Metrics

Citations

Cited By

View all
  • (2021)Emerging ExG-based NUI Inputs in Extended Realities: A Bottom-up SurveyACM Transactions on Interactive Intelligent Systems10.1145/345795011:2(1-49)Online publication date: 21-Jul-2021

View Options

Login options

View options

PDF

View or Download as a PDF file.

PDF

eReader

View online with eReader.

eReader

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media