Abstract
The new ubiquitous interaction methods change people’s life and facilitate their tasks in everyday life and in the workplace, enabling people to access their personal data as well as public resources at any time and in any place. We found two solutions to enable ubiquitous interaction and put a stop to the limits imposed by the desktop mode: namely nomadism and mobility. Based on these two solutions, we have proposed three interfaces (Zhou et al. in HCI international 2011: human–computer interaction. Interaction techniques and environments, Springer, Berlin, pp 500–509, 2011): in-environment interface (IEI), environment dependent interface (EDI), and environment independent interface (EII). In this paper, we first discuss an overview of IEI, EDI, and EII, before excluding IEI and focusing on EDI and EII, their background, and distinct characteristics. We also propose a continuum from physical paper-based interface to digital projected interface in relation with EDI and EII. Then, to validate EDI and EII concepts, we design and implement a MobilePaperAccess system, which is a wearable camera-glasses system with paper-based interface and original input techniques allowing mobile interaction. Furthermore, we discuss the evaluation of the MobilePaperAccess system; we compare two interfaces (EDI and EII) and three input techniques (finger input, mask input, and page input) to test the feasibility and usability of this system. Both the quantitative and qualitative results are reported and discussed. Finally, we provide the prospects and our future work for improving the current approaches.
Similar content being viewed by others
References
Akaoka E, Ginn T, Vertegaal R (2010) DisplayObjects: prototyping functional physical interfaces on 3d styrofoam, paper or cardboard models. In: TEI 2010: proceedings of the 4th international conference on tangible and embedded interaction, pp 49–56
Asadzadeh P, Kulik L, Tanin E (2012) Gesture recognition using RFID technology. Pers Ubiquit Comput 16(3):225–234. doi:10.1007/s00779-011-0395-z
Ballagas R, Borchers J, Rohs M, Sheridan JG (2006) The smart phone: a ubiquitous input device. IEEE Pervasive Comput 5(1):70–77
Bradski G (2000) The opencv library. Dr Dobb’s J 25(11):120–126
Bradski GR (1998) Computer vision face tracking for use in a perceptual user interface. In: WACV 1998: proceedings IEEE workshop on application of computer vision, pp 214–219
Buxton W (1990) A three-state model of graphical input. In: INTERACT 1990: proceedings of 3rd IFIP international conference on human-computer interaction, pp 449–456
Choi J, Kim GJ (2013) Usability of one-handed interaction methods for handheld projection-based augmented reality. Pers Ubiquit Comput 17(2):399–409. doi:10.1007/s00779-011-0502-1
David BT, Chalon R (2007) IMERA: experimentation platform for computer augmented environment for mobile actors. In: WiMOB 2007: 3rd IEEE international conference on wireless and mobile computing, networking and communications, pp 51
David BT, Zhou Y, Xu T, Chalon R (2011) Mobile user interfaces and their utilization in a Smart City. In: ICOMP’11: the 2011 international conference on internet computing as part of WorldComp’2011 conference, CSREA Press, Boca Raton, FL
Dul J, Weerdmeester BA (2008) Ergonomics for beginners: a quick reference guide. CRC Press, Boca Raton, FL, pp 12
Fishkin KP (2004) A taxonomy for and analysis of tangible interfaces. Pers Ubiquit Comput 8(5):347–358. doi:10.1007/s00779-004-0297-4
Fitzmaurice GW, Ishii H, Buxton WAS (1995) Bricks: laying the foundations for graspable user interfaces. In: CHI 1995: proceedings of the SIGCHI conference on human fac tors in computing systems, pp 442–449
Grant DA (1948) The latin square principle in the design and analysis of psychological experiments. Psychol Bull 45(5):427
Grimes GJ (1983) Digital data entry glove interface device. Patent US4414537
Ha Y, Rolland J (2002) Optical assessment of head-mounted displays in visual space. Appl Opt 41(25):5282–5289
Harrison C, Benko H, Wilson AD (2011) OmniTouch: wearable multitouch interaction everywhere. In: UIST 2011: proceedings of the 24nd annual ACM symposium on user interface software and technology, pp 441–450
Harrison C, Hudson SE (2010) Minput: enabling interaction on small mobile devices with high-precision, low-cost, multipoint optical tracking. In: Proceedings of the 28th international conference on human factors in computing systems, pp 1661–1664
Harrison C, Tan D, Morris D (2010) Skinput: appropriating the body as an input surface. In: CHI 2010: proceedings of the 28th SIGCHI conference on human factors in computing systems, pp 453–462
Holman D, Vertegaal R, Altosaar M, Troje N, Johns D (2005) Paper windows: interaction techniques for digital paper. In: CHI 2005: proceedings of the SIGCHI conference on human factors in computing systems, pp 591–599
Hornecker E, Psik T (2005) Using ARToolKit markers to build tangible prototypes and simulate other technologies, Human-Computer Interaction-INTERACT 2005, vol 3585, pp 30–42. doi:10.1007/11555261_6
Ishii H (2008) Tangible bits: beyond pixels. In: TEI 2008: proceedings of the 2nd international conference on tangible and embedded interaction, pp xv–xxv
Ishii H, Ullmer B (1997) Tangible bits: towards seamless interfaces between people, bits and atoms. In: CHI 1997: proceedings of the SIGCHI conference on human factors in computing systems, pp 234–241
Kubicki S, Lepreux S, Kolski C (2012) RFID-driven situation awareness on TangiSense, a table interacting with tangible objects. Pers Ubiquit Comput 16(8):1079–1094. doi:10.1007/s00779-011-0442-9
Liao C, Tang H, Liu Q, Chiu P, Chen F (2010) FACT: fine-grained cross-media interaction with documents via a portable hybrid paper-laptop interface. In: MM 2010: proceedings of the international conference on multimedia, pp 361–370
Likert R (1932) A technique for the measurement of attitudes. Arch Psychol 22(140):1–55
Lyons K, Starner T, Gane B (2006) Experimental evaluations of the twiddler one-handed chording mobile keyboard. Hum Comput Interact 21(4):343–392
Lyons K, Starner T, Plaisted D, Fusia J, Lyons A, Drew A, Looney EW (2004) Twiddler typing: One-handed chording text entry for mobile phones. In: CHI 2004: proceedings of the SIGCHI conference on human factors in computing systems, pp 671–678
MacKenzie IS (1992) Fitts’ law as a research and design tool in human-computer interaction. Hum Comput Interact 7(1):91–139
Mistry P, Maes P (2008) Quickies: intelligent sticky notes. In: IET 2008: 4th international conference on intelligent environments, pp 1–4
Mistry P, Maes P, Chang L (2009) WUW-wear Ur world: a wearable gestural interface. In: CHI EA 2009: proceedings of the 27th international conference extended abstracts on human factors in computing systems, pp 4111–4116
Morris D (2010) Emerging input technologies for always-available mobile interaction. Found Trends® Hum Comput Interact 4(4):245–316
Ni T, Baudisch P (2009) Disappearing mobile devices. In: UIST 2009: proceedings of the 22th annual ACM symposium on user interface software and technology, pp 101–110
Rekimoto J, Ayatsuka Y (2000) CyberCode: designing augmented reality environments with visual tags. In: DARE 2000: proceedings of conference on designing augmented reality environments, pp 1–10
Rouillard J (2008) Contextual QR codes. In: ICCGI’08: the third international multi-conference on computing in the global information technology, pp 50–55
Rukzio E, Holleis P, Gellersen H (2012) Personal projectors for pervasive computing. IEEE Pervasive Comput 11(2):30–37. doi:10.1109/MPRV.2011.17
Shaer O, Hornecker E (2010) Tangible user interfaces: past, present, and future directions. Found Trends Hum Comput Interact 3(1–2):1–137
Spitzer MB, Rensing N, McClelland R, Aquilino P (1997) Eyeglass-based systems for wearable computing. In: Digest of papers. First international symposium on wearable computers, pp 48–51
Starner T, Mann S, Rhodes B, Healey J, Russell KB, Levine J, Pentland A (1995) Wearable computing and augmented reality. The Media Laboratory, Massachusetts Institute of Technology, Cambridge, MA, MIT Media Lab Vision and Modeling Group Technical Report, vol 355
Tamaki E, Miyaki T, Rekimoto J (2009) Brainy hand: an ear-worn hand gesture interaction device. In: CHI EA 2009: proceedings of the 27th of the international conference extended abstracts on human factors in computing systems, pp 4255–4260
Vandam A (1997) Post-WIMP user interfaces. Commun CM 40(2):63–67
Wang RY, Popović J (2009) Real-time hand-tracking with a color glove. In: SIGGRAPH 2009: ACM transactions on graphics (TOG), vol 28(3), Article No. 63
Weiser M (1991) The computer for the 21st century. Sci Am 265(3):94–104
Willis KD (2012) A pre-history of handheld projector-based interaction. Pers Ubiquit Comput 16(1):5–15. doi:10.1007/s00779-011-0373-5
Willis KDD, Poupyrev I, Shiratori T (2011) Motionbeam: a metaphor for character interaction with handheld projectors. In: CHI 2011: proceedings of the SIGCHI conference on human factors in computing systems, pp 1031–1040
Wilson ML, Craggs D, Robinson S, Jones M, Brimble K (2012) Pico-ing into the future of mobile projection and contexts. Pers Ubiquit Comput 16(1):39–52. doi:10.1007/s00779-011-0376-2
Zhou Y, David B, Chalon R (2011) Innovative user interfaces for wearable computers in real augmented environment. In: HCI international 2011: human–computer interaction. Interaction techniques and environments, Springer, Berlin, pp 500–509
Author information
Authors and Affiliations
Corresponding authors
Rights and permissions
About this article
Cite this article
Zhou, Y., Xu, T., David, B. et al. Innovative wearable interfaces: an exploratory analysis of paper-based interfaces with camera-glasses device unit. Pers Ubiquit Comput 18, 835–849 (2014). https://doi.org/10.1007/s00779-013-0697-4
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00779-013-0697-4