Abstract
The marginalization of people with disabilities, such as visually impaired individuals (VIIs), has driven scientists to take advantage of the fast growth of smart technologies and develop smart assistive systems (SASs) to bring VIIs back to social life, education and even to culture. Our research focuses on developing a human–computer interactive system that will guide VIIs in outdoor cultural environments by offering universal access to cultural information, social networking and safe navigation among other services. The VI users interact with computer-based SAS to control the system during its operation, while having access to remote connection with non-VIIs for external guidance and company. The development of such a system needs a user-centered design (UCD) that incorporates the elicitation of the necessary requirements for a satisfying operation for the VI users. In this paper, we present a novel SAS system for VIIs and its design considerations, which follow a UCD approach to determine a set of operational, functional, ergonomic, environmental and optional requirements of the system. Both VIIs and non-VIIs took part in a series of interviews and questionnaires, from which data were analyzed to form the requirements of the system for both the on-site and remote use. The final requirements are tested by trials and their evaluation and results are presented. The experimental investigations gave significant feedback for the development of the system, throughout the design process. The most important contribution of this study is the derivation of requirements applicable not only to the specific system under investigation, but also to other relevant SASs for VIIs.
Similar content being viewed by others
References
WHO: World Health Organization—blindness and visual impairement. http://www.who.int/news-room/fact-sheets/detail/blindness-and-visual-impairment (2018). Accessed Sept 2019
Vuletić, G., Šarlija, T., Benjak, T.: Quality of life in blind and partially sighted people. J. Appl. Health Sci. 2, 101–112 (2016)
Wahab, M.H.A., Talib, A.A., Kadir, H.A., Johari, A., Noraziah, A., Sidek, R.M., Mutalib, A.A.: Smart cane: assistive cane for visually-impaired people (2011). arXiv preprint arXiv:1110.5156
Ulrich, I., Borenstein, J.: The GuideCane-applying mobile robot technologies to assist the visually impaired. IEEE Trans. Syst. Man Cybern. Part A Syst. Hum. 31(2), 131–136 (2001)
Takizawa, H., Yamaguchi, S., Aoyagi, M., Ezaki, N., Mizuno, S.: Kinect cane: an assistive system for the visually impaired based on the concept of object recognition aid. Pers. Ubiquit. Comput. 19(5–6), 955–965 (2015)
Mandru, D., Lungu, I.B., Mociran, A.: Development of a mechatronic blind stick, fascicle of management and technological. Engineering 6(16), 797–802 (2007)
Gayathri, G., Vishnupriya, M., Nandhini, R., Banupriya, M.: Smart walking stick for visually impaired. IJECS 3(3), 4057–4061 (2014)
Nada, A.A., Fakhr, M.A., Seddik, A.F.: Assistive infrared sensor based smart stick for blind people. In: 2015 Science and Information Conference (SAI), pp. 1149–1154 (2015)
Sakhardande, J., Pattanayak, P., Bhowmick, M.: Smart cane assisted mobility for the visually impaired. World Academy of Science, Engineering and Technology, vol. 70 (2012)
“WeWALK”. https://wewalk.io/en (2019)
Alghamdi, S., Van Schyndel, R., Alahmadi, A.: Indoor navigational aid using active RFID and QR-code for sighted and blind people. In: 2013 IEEE Eighth International Conference on Intelligent Sensors, Sensor Networks and Information Processing, pp. 18–22 (2013)
Nassih, M., Cherradi, I., Maghous, Y., Ouriaghli, B., Salih-Alj, Y.: Obstacles recognition system for the blind people using RFID. In: 2012 Sixth International Conference on Next Generation Mobile Applications, Services and Technologies, pp. 60–63 (2012)
Ivanov, R.: RSNAVI: an RFID-based context-aware indoor navigation system for the blind. In: Proceedings of the 13th International Conference on Computer Systems and Technologies, pp. 313–320 (2012)
Qinghui, T., Malik, M.Y., Hong, Y., Park, J.: A real-time localization system using RFID for visually impaired (2011). arXiv preprint arXiv:1109.1879
Yelamarthi, K., Haas, D., Nielsen, D., Mothersell, S.: RFID and GPS integrated navigation system for the visually impaired. In: 2010 53rd IEEE International Midwest Symposium on Circuits and Systems, pp. 1149–1152 (2010)
Jafri, R., Ali, S.A., Arabnia, H.R.: Computer vision-based object recognition for the visually impaired using visual tags. In: Proceedings of the International Conference on Image Processing, Computer Vision, and Pattern Recognition (IPCV), p. 1 (2013)
Hakobyan, L., Lumsden, J., O’Sullivan, D., Bartlett, H.: Mobile assistive technologies for the visually impaired. Surv. Ophthalmol. 58(6), 513–528 (2013)
Liu, K.-C., Wu, C.-H., Tseng, S.-Y., Tsai, Y.-T.: Voice helper: a mobile assistive system for visually impaired persons. In: 2015 IEEE International Conference on Computer and Information Technology; Ubiquitous Computing and Communications; Dependable, Autonomic and Secure Computing; Pervasive Intelligence and Computing, pp. 1400–1405 (2015)
Mekhalfi, M.L., Melgani, F., Zeggada, A., De Natale, F.G., Salem, M.A.-M., Khamis, A.: Recovering the sight to blind people in indoor environments with smart technologies. Expert Syst. Appl. 46, 129–138 (2016)
Tian, Y., Yang, X., Arditi, A.: Computer vision-based door detection for accessibility of unfamiliar environments to blind persons. In: International Conference on Computers for Handicapped Persons, pp. 263–270 (2010)
Tian, Y., Yang, X., Yi, C., Arditi, A.: Toward a computer vision-based wayfinding aid for blind persons to access unfamiliar indoor environments. Mach. Vis. Appl. 24(3), 521–535 (2013)
Tee, Z., Ang, L., Seng, K.: Smart guide system to assist visually impaired people in an indoor environment. Iete Tech. Rev. 27(6), 455–464 (2010)
Jafri, R., Ali, S.A., Arabnia, H.R., Fatima, S.: Computer vision-based object recognition for the visually impaired in an indoors environment: a survey. Visual Comput. 30(11), 1197–1222 (2014)
Bai, J., Lian, S., Liu, Z., Wang, K., Liu, D.: Smart guiding glasses for visually impaired people in indoor environment. IEEE Trans. Consum. Electron. 63(3), 258–266 (2017)
He, H., Li, Y., Guan, Y., Tan, J.: Wearable ego-motion tracking for blind navigation in indoor environments. IEEE Trans. Autom. Sci. Eng. 12(4), 1181–1190 (2015)
Owayjan, M., Hayek, A., Nassrallah, H., Eldor, M.: Smart assistive navigation system for blind and visually impaired individuals. In: 2015 International Conference on Advances in Biomedical Engineering (ICABME), pp. 162–165 (2015)
Wang, H.-C., Katzschmann, R.K., Teng, S., Araki, B., Giarré, L., Rus, D.: Enabling independent navigation for visually impaired people through a wearable vision-based feedback system. In: 2017 IEEE International Conference on Robotics and Automation (ICRA), pp. 6533–6540 (2017)
Mascetti, S., Ahmetovic, D., Gerino, A., Bernareggi, C.: ZebraRecognizer: pedestrian crossing recognition for people with visual impairment or blindness. Pattern Recognit. 60, 405–419 (2016)
Wang, S., Pan, H., Zhang, C., Tian, Y.: RGB-D image-based detection of stairs, pedestrian crosswalks and traffic signs. J. Vis. Commun. Image Represent. 25(2), 263–272 (2014)
Wang, S., Tian, Y.: Detecting stairs and pedestrian crosswalks for the blind by RGBD camera. In: 2012 IEEE International Conference on Bioinformatics and Biomedicine Workshops, pp. 732–739 (2012)
Jafri, R., Ali, S.A.: Exploring the potential of eyewear-based wearable display devices for use by the visually impaired. In: 2014 3rd International Conference on User Science and Engineering (i-USEr), pp. 119–124 (2014)
Shinohara, K., Wobbrock, J.O.: In the shadow of misperception: assistive technology use and social interactions. In: Proceedings of the SIGCHI Conference on Human Factors in Computing Systems, pp. 705–714 (2011)
Law, C.M., Jaeger, P.T., McKay, E.: User-centered design in universal design resources? Univ. Access Inf. Soc. 9(4), 327–335 (2010)
Mirri, S., Roccetti, M., Salomoni, P.: Collaborative design of software applications: the role of users. Hum. Centric Comput. Inf. Sci. 8(1), 6 (2018)
Lopes, A., Valentim, N., Moraes, B., Zilse, R., Conte, T.: Applying user-centered techniques to analyze and design a mobile application. J. Softw. Eng. Res. Dev. 6(1), 5 (2018)
Vesin, B., Mangaroska, K., Giannakos, M.: Learning in smart environments: user-centered design and analytics of an adaptive learning system. Smart Learn. Environ. 5(1), 24 (2018)
Seyff, N., Todoran, I., Caluser, K., Singer, L., Glinz, M.: Using popular social network sites to support requirements elicitation, prioritization and negotiation. J. Internet Serv. Appl. 6(1), 7 (2015)
Lin, Y.-C., Yeh, C.-H., Wei, C.-C.: How will the use of graphics affect visual aesthetics? A user-centered approach for web page design. Int. J. Hum. Comput. Stud. 71(3), 217–227 (2013)
Costa, N.A., Holder, E., MacKinnon, S.: Implementing human centred design in the context of a graphical user interface redesign for ship manoeuvring. Int. J. Hum. Comput. Stud. 100, 55–65 (2017)
Augusto, J., Kramer, D., Alegre, U., Covaci, A., Santokhee, A.: The user-centred intelligent environments development process as a guide to co-create smart technology for people with special needs. Univ. Access Inf. Soc. 17(1), 115–130 (2018)
Lazar, J.: Web Usability: A User-Centered Design Approach. Pearson Addison Wesley, Boston (2006)
Juárez-Ramírez, R.: User-Centered Design and Adaptive Systems: Toward Improving Usability and accessibility. Springer, Berlin (2017)
Jafri, R., Khan, M.M.: User-centered design of a depth data based obstacle detection and avoidance system for the visually impaired. Hum. Centric Comput. Inf. Sci. 8(1), 14 (2018)
Mancas-Thillou, C., Ferreira, S., Demeyer, J., Minetti, C., Gosselin, B.: A multifunctional reading assistant for the visually impaired. J. Image Video Process. 2007(3), 5 (2007)
Bateman, A., Zhao, O.K., Bajcsy, A.V., Jennings, M.C., Toth, B.N., Cohen, A.J., Horton, E.L., Khattar, A., Kuo, R.S., Lee, F.A., et al.: A user-centered design and analysis of an electrostatic haptic touchscreen system for students with visual impairments. Int. J. Hum. Comput. Stud. 109, 102–111 (2018)
Meza-de-Luna, M.E., Terven, J.R., Raducanu, B., Salas, J.: A Social-Aware Assistant to support individuals with visual impairments during social interaction: a systematic requirements analysis. Int. J. Hum. Comput. Stud. 122, 50–60 (2019)
Koo, S.H., Fallon, K.: Explorations of wearable technology for tracking self and others. Fashion Text. 5(1), 8 (2018)
Robins, B., Ferrari, E., Dautenhahn, K., Kronreif, G., Prazak-Aram, B., Gelderblom, G., Tanja, B., Caprino, F., Laudanna, E., Marti, P.: Human-centred design methods: developing scenarios for robot assisted play informed by user panels and field trials. Int. J. Hum. Comput. Stud. 68(12), 873–898 (2010)
Iakovidis, D.K., Diamantis, D., Dimas, D., Ntakolia, C., Spyrou, E.: Digital enhancement of cultural experience and accessibility for the visually impaired. In: Technological Trends in Improved Mobility of the Visually Impaired, pp. 237–271. Springer (2020)
Loukopoulos, T., Koziri, M., Panagou, N., Papadopoulos, P.K., Iakovidis, D.K.: Cloud video guidance as ‘Deus ex Machina’ for the visually impaired. In: Technological Trends in Improved Mobility of the Visually Impaired. Springer, pp. 127–143 (2020)
Zowghi, D., Coulin, C.: Requirements elicitation: a survey of techniques, approaches, and tools. In: Engineering and Managing Software Requirements, pp. 19–46. Springer (2005)
Magnusson, C., Hedvall, P.-O., Caltenco, H.: Co-designing together with persons with visual impairments. In: Mobility of Visually Impaired People, pp. 411–434. Springer (2018)
I. O. for Standardization, “ISO 9241-210:2010” (2010)
Panchanathan, S., Black, J., Rush, M., Iyer, V.: iCare-a user centric approach to the development of assistive devices for the blind and visually impaired. In: 15th IEEE International Conference on Tools with Artificial Intelligence, 2003. Proceedings, pp. 641–648
Fryer, L.: Putting it into words: the impact of visual impairment on perception, experience and presence. Goldsmiths, University of London (2013)
Sosa-Garcia, J., Odone, F.: Hands on “Visual recognition for visually impaired users”. ACM Trans. Access. Comput. (TACCESS) 10(3), 8 (2017)
Panchanathan, S., Black, J., Rush, M., Iyer, V.: iCare-a user centric approach to the development of assistive devices for the blind and visually impaired. In: 15th IEEE International Conference on Tools with Artificial Intelligence, 2003. Proceedings, pp. 641–648 (2003)
Perakovi, D., Periša, M., Prci, A.B.: Possibilities of applying ICT to improve safe movement of blind and visually impaired persons. In: Cutting Edge Research in Technologies. InTech (2015)
Hersh, M.A., Johnson, M.A.: A robotic guide for blind people. Part 1. A multi-national survey of the attitudes, requirements and preferences of potential end-users. Appl. Bion. Biomech. 7(4), 277–288 (2010)
Zeng, L.: A survey: outdoor mobility experiences by the visually impaired. In: Mensch und Computer 2015—Workshopband (2015)
Conradie, P., de Goedelaan, G.K., Mioch, T., Saldien, J.: Blind user requirements to support tactile mobility. In: Tactile Haptic User Interfaces for Tabletops and Tablets (TacTT 2014), pp. 48–53 (2014)
Loomis, J.M., Marston, J.R., Golledge, R.G., Klatzky, R.L.: Personal guidance system for people with visual impairment: a comparison of spatial displays for route guidance. J. Visual Impair. Blindness 99(4), 219 (2005)
Petrie, H., Johnson, V., Strothotte, T., Raab, A., Fritz, S., Michel, R.: MOBIC: designing a travel aid for blind and elderly people. J. Navig. 49(01), 45–52 (1996)
Kulyukin, V., Gharpure, C., Nicholson, J., Pavithran, S.: RFID in robot-assisted indoor navigation for the visually impaired. In: 2004 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (IEEE Cat. No. 04CH37566) (2004)
Chebat, D.-R., Schneider, F., Kupers, R., Ptito, M.: Navigation with a sensory substitution device in congenitally blind individuals. NeuroReport 22, 342–347 (2011)
Johnson, L.A., Higgins, C.M.: A navigation aid for the blind using tactile-visual sensory substitution. In: 2006 International Conference of the IEEE Engineering in Medicine and Biology Society, pp. 6289–6292 (2006)
Benabid, A., AlZuhair, M.: User involvement in the development of indoor navigation system for the visually impaired: a needs-finding study. In: 2014 3rd International Conference on User Science and Engineering (i-USEr), pp. 97–102 (2014)
Alkhanifer, A., Ludi, S.: Towards a situation awareness design to improve visually impaired orientation in unfamiliar buildings: requirements elicitation study. In: 2014 IEEE 22nd International on Requirements Engineering Conference (RE), pp. 23–32 (2014)
Kalia, A.A., Legge, G.E., Roy, R., Ogale, A.: Assessment of indoor route-finding technology for people with visual impairment. J. Visual Impair. Blindness 104(3), 135 (2010)
Roentgen, U.R., Gelderblom, G.J., de Witte, L.P.: User evaluation of two electronic mobility aids for persons who are visually impaired: a quasi-experimental study using a standardized mobility course. Assist. Technol. 24(2), 110–120 (2012)
Hub, A., Diepstraten, J., Ertl, T.: Design and development of an indoor navigation and object identification system for the blind. In: Proceedings of the ACM SIGACCESS Conference on Computers and Accessibility—ASSETS textquotesingle04 (2004)
Guerrero, L.A., Vasquez, F., Ochoa, S.F.: An indoor navigation system for the visually impaired. Sensors 12(6), 8236–8258 (2012)
Strumillo, P.: Electronic interfaces aiding the visually impaired in environmental access, mobility and navigation. In: 3rd International Conference on Human System Interaction (2010)
Bohonos, S., Lee, A., Malik, A., Thai, C., Manduchi, R.: Universal real-time navigational assistance (URNA). In: Proceedings of the 1st ACM SIGMOBILE International Workshop on Systems and Networking Support for Healthcare and Assisted Living Environments—HealthNet textquotesingle07 (2007)
Sánchez, J., Maureira, E.: Subway mobility assistance tools for blind users. In: Universal Access in Ambient Intelligence, Lecture Notes on Computer Science, vol. 4397, pp. 386–404 (2007)
Kim, H.K., Han, S.H., Park, J., Park, J.: The interaction experiences of visually impaired people with assistive technology: a case study of smartphones. Int. J. Ind. Ergon. 55, 22–33 (2016)
Kane, S.K., Jayant, C., Wobbrock, J.O., Ladner, R.E.: Freedom to roam: a study of mobile device adoption and accessibility for people with visual and motor disabilities. In: Proceedings of the 11th International ACM SIGACCESS Conference on Computers and Accessibility, pp. 115–122 (2009)
Ungar, S., Blades, M., Spencer, C.: The construction of cognitive maps by children with visual impairments. In: Portugali, J. (ed.) The construction of cognitive maps, pp. 247–273. Springer Netherlands, Dordrecht (1996)
Periša, M., Cviti, I., Sente, R.E.: Social network customer requirements analysis for visually impaired people. In: 4th International Virtual Research Conference in Technical Disciplines, pp. 36–44 (2016)
Alkhafaji, A., Fallahkhair, S., Cocea, M., Crellin, J.: A survey study to gather requirements for designing a mobile service to enhance learning from cultural heritage. In: European Conference on Technology Enhanced Learning, pp. 547–550 (2016)
Asakawa, S., Guerreiro, J., Ahmetovic, D., Kitani, K.M., Asakawa, C.: The present and future of museum accessibility for people with visual impairments. In: Proceedings of the 20th International ACM SIGACCESS Conference on Computers and Accessibility, pp. 382–384 (2018)
Handa, K., Dairoku, H., Toriyama, Y.: Investigation of priority needs in terms of museum service accessibility for visually impaired visitors. Br. J. Visual Impair. 28(3), 221–234 (2010)
the I. C. O. ISO Central Secretariat and IEEE, “ISO/IEC/IEEE 29148: International Standard: Systems and Software Engineering—Life Cycle Processes—Requirements Engineering (2011)
Olphert, C., Damodaran, L.: Getting what you want, or wanting what you get?—Beyond user centred design. Des. Emot, pp. 138 (2004)
Klein, G., Kaempf, G.L., Wolf, S., Thorsden, M., Miller, T.: Applying decision requirements to user-centered design. Int. J. Hum. Comput. Stud. 46(1), 1–15 (1997)
Maguire, M., Bevan, N.: User requirements analysis. In: IFIP World Computer Congress, TC 13, pp. 133–148 (2002)
Lee, C.-H., Su, Y.-C., Chen, L.-G.: An intelligent depth-based obstacle detection system for visually-impaired aid applications. In: 2012 13th International Workshop on Image Analysis for Multimedia Interactive Services, pp. 1–4 (2012)
Wei, Z., Tang, K.L., Ngan, K.N.: Implementation of H. 264 on mobile device. IEEE Trans. Consum. Electron. 53(3), 1109–1116 (2007)
Dimas, G., Ntakolia, C., Iakovidis, D.K.: Obstacle detection based on generative adversarial networks and fuzzy sets for computer-assisted navigation. In: International Conference on Engineering Applications of Neural Networks, pp. 533–544 (2019)
Diamantis, D.E., Koutsiou, D.-C.C., Iakovidis, D.K.: Staircase detection using a lightweight look-behind fully convolutional neural network. In: International Conference on Engineering Applications of Neural Networks, pp. 522–532 (2019)
Acknowledgements
This research has been co-financed by the European Union and Greek national funds through the Operational Program Competitiveness, Entrepreneurship and Innovation, under the call RESEARCH—CREATE—INNOVATE (Project Code: T1EDK-02070). We would like to sincerely thank the NFB for the interview and the advice, as well as all the volunteers who participated in the study. We would also like to thank all the collaborators of the ENORASI project for their contributions.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
On behalf of all authors, the corresponding author states that there is no conflict of interest.
Ethical approval and consent for Publication
All procedures performed in this study involving human participants were in accordance with the ethical standards of the ethics committee of the Department of Computer Science and Biomedical Informatics of University of Thessaly. Informed consent was obtained from all individual participants included in this study.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix
Appendix
In Table 11, we present a summary of the requirements that have been identified in the literature review regarding smart assistive systems for visual impaired individuals.
Rights and permissions
About this article
Cite this article
Ntakolia, C., Dimas, G. & Iakovidis, D.K. User-centered system design for assisted navigation of visually impaired individuals in outdoor cultural environments. Univ Access Inf Soc 21, 249–274 (2022). https://doi.org/10.1007/s10209-020-00764-1
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10209-020-00764-1