Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3357236.3395433acmconferencesArticle/Chapter ViewAbstractPublication PagesdisConference Proceedingsconference-collections
research-article

Human-in-the-Loop Machine Learning to Increase Video Accessibility for Visually Impaired and Blind Users

Published: 03 July 2020 Publication History

Abstract

Video accessibility is crucial for blind and visually impaired individuals for education, employment, and entertainment purposes. However, professional video descriptions are costly and time-consuming. Volunteer-created video descriptions could be a promising alternative, however, they can vary in quality and can be intimidating for novice describers. We developed a Human-in-the-Loop Machine Learning (HILML) approach to video description by automating video text generation and scene segmentation and allowing humans to edit the output. The HILML approach facilitates human-machine collaboration to produce high quality video descriptions while keeping a low barrier to entry for volunteer describers. Our HILML system was significantly faster and easier to use for first-time video describers compared to a human-only control condition with no machine learning assistance. The quality of the video descriptions and understanding of the topic created by the HILML system compared to the human-only condition were rated as being significantly higher by blind and visually impaired users.

References

[1]
MAGpie 2.0. Accessed date 2019-07--17. (Accessed date 2019-07--17). https://main.wgbh.org/wgbh/pages/ncam_old/webaccess/magpie/
[2]
3play Media. Accessed Date 2019-09--15. (Accessed Date 2019-09--15). https://www.3playmedia.com/
[3]
Nayyer Aafaq, Syed Zulqarnain Gilani, Wei Liu, Ajmal Mian, and Mubarak Shah. 2018. Video Description: A Survey of Methods, Datasets and Evaluation Metrics. CoRR abs/1806.00186 (2018).
[4]
National Aeronautics and Space Administration (NASA). Accessed date 2019-07--17. NASA-TLX. (Accessed date 2019-07--17). https://humansystems.arc.nasa.gov/groups/TLX/tlxapp.php
[5]
Stanislaw Antol, Aishwarya Agrawal, Jiasen Lu, Margaret Mitchell, Dhruv Batra, C Lawrence Zitnick, and Devi Parikh. 2015. Vqa: Visual question answering. In Proceedings of the IEEE International Conference on Computer Vision, ICCV. 2425--2433.
[6]
Grigorios Antonellis, Andreas G. Gavras, Marios Panagiotou, Bruce L. Kutter, Gabriele Guerrini, Andrew C. Sander, and Patrick J. Fox. 2015. Shake Table Test of Large-Scale Bridge Columns Supported on Rocking Shallow Foundations. Journal of Geotechnical and Geoenvironmental Engineering 141, 5 (2015), 04015009.
[7]
Microsoft Azure. Accessed date 2019-07--17. (Accessed date 2019-07--17). https://azure.microsoft.com/en-us/services/media-services/video-indexer/
[8]
David Bar-El, Thomas Large, Lydia Davison, and Marcelo Worsley. 2018. Tangicraft: A Multimodal Interface for Minecraft. In Proceedings of the International ACM SIGACCESS Conference on Computers and Accessibility, ASSETS. 456--458.
[9]
Andrei Barbu, Alexander Bridge, Zachary Burchill, Dan Coroian, Sven Dickinson, Sanja Fidler, Aaron Michaux, Sam Mussman, Siddharth Narayanaswamy, Dhaval Salvi, Lara Schmidt, Jiangnan Shangguan, Jeffrey Mark Siskind, Jarrell Waggoner, Song Wang, Jinlian Wei, Yifan Yin, and Zhiqi Zhang. 2012. Video In Sentences Out. (2012).
[10]
United States Access Board. 2015. Proposed Information and Communication Technology (ICT) Standards and Guidelines. (2015). https://www.access-board.gov/guidelines-and-standards/communications-and-it/about-the-ict-refresh/proposed-rule
[11]
Rupert RA Bourne, Seth R Flaxman, Tasanee Braithwaite, Maria V Cicinelli, Aditi Das, Jost B Jonas, Jill Keeffe, John H Kempen, Janet Leasher, Hans Limburg, and others. 2017. Magnitude, temporal trends, and projections of the global prevalence of blindness and distance and near vision impairment: a systematic review and meta-analysis. The Lancet Global Health 5, 9 (2017), e888--e897.
[12]
Matthew Brand. 1997. The" Inverse hollywood problem": from video to scripts and storyboards via causal analysis. In AAAI/IAAI. Citeseer, 132--137.
[13]
Carmen Branje, Susan Marshall, Ashley Tyndall, and Deborah Fels. 2006. LiveDescribe. AMCIS 2006 Proceedings (2006), 368.
[14]
Claude Chapdelaine. 2012. Specialized DVD player to render audio description and its usability performance. In Proceedings of the International ACM SIGACCESS Conference on Computers and Accessibility, ASSETS. 203--204.
[15]
Claude Chapdelaine and Langis Gagnon. 2009. Accessible videodescription on-demand. In Proceedings of the International ACM SIGACCESS Conference on Computers and Accessibility, ASSETS. 221--222.
[16]
Kyunghyun Cho, Bart van Merrienboer, Dzmitry Bahdanau, and Yoshua Bengio. 2014. On the Properties of Neural Machine Translation: Encoder--Decoder Approaches. In Proceedings of the Eighth Workshop on Syntax, Semantics and Structure in Statistical Translation, SSST-8 (2014).
[17]
World Wide Web Consortium. 2018. Web Content Accessibility Guidelines 2.1. (2018). https://www.w3.org/TR/WCAG21/
[18]
Frederico da Rocha Tomé Filho, Pejman Mirza-Babaei, Bill Kapralos, and Glaudiney Moreira Mendonça Junior. 2019. Let's Play Together: Adaptation Guidelines of Board Games for Players with Visual Impairment. In Proceedings of the Conference on Human Factors in Computing Systems, CHI (CHI '19). 631:1--631:15.
[19]
Abhishek Das, Satwik Kottur, Khushi Gupta, Avi Singh, Deshraj Yadav, Jose M. F. Moura, Devi Parikh, and Dhruv Batra. 2017. Visual Dialog. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR.
[20]
Pradipto Das, Chenliang Xu, Richard F. Doell, and Jason J. Corso. 2013. A Thousand Frames in Just a Few Words: Lingual Description of Videos through Latent Topics and Sparse Object Stitching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR.
[21]
Benoît Encelle, Magali Ollagnier Beldame, and Yannick Prié. 2013. Towards the usage of pauses in audio-described videos. In Proceedings of the 10th International Cross-Disciplinary Conference on Web Accessibility. 31.
[22]
Benoît Encelle, Magali Ollagnier-Beldame, Stéphanie Pouchot, and Yannick Prié. 2011. Annotation-based video enrichment for blind people: A pilot study on the use of earcons and speech synthesis. In Proceedings of the International ACM SIGACCESS Conference on Computers and Accessibility, ASSETS. 123--130.
[23]
Be My Eyes. Accessed date 2019-07--17. (Accessed date 2019-07--17). https://www.bemyeyes.com/
[24]
Angela Fan, Mike Lewis, and Yann Dauphin. 2018. Hierarchical Neural Story Generation. In Proceedings of the 56th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). Association for Computational Linguistics.
[25]
Diamantino Freitas and Georgios Kouroupetroglou. 2008. Speech technologies for blind and low vision persons. Technology and Disability 20, 2 (2008), 135--156.
[26]
Giovanni Fusco, Ender Tekin, Richard E Ladner, and James M Coughlan. 2014. Using computer vision to access appliance displays. In Proceedings of the International ACM SIGACCESS Conference on Computers and Accessibility, ASSETS. 281--282.
[27]
Langis Gagnon, Claude Chapdelaine, David Byrns, Samuel Foucher, Maguelonne Heritier, and Vishwa Gupta. 2010. A Computer-Vision-Assisted System for Videodescription Scripting. In IEEE Computer Society Conference on Computer Vision and Pattern Recognition-Workshops. pp. 41--48.
[28]
Langis Gagnon, Samuel Foucher, Maguelonne Heritier, Marc Lalonde, David Byrns, Claude Chapdelaine, James Turner, Suzanne Mathieu, Denis Laurendeau, Nath Tan Nguyen, and others. 2009. Towards computer-vision software tools to increase production and accessibility of video description for people with vision loss. Universal Access in the Information Society 8, 3 (2009), 199--218.
[29]
Eitan Glinert and Lonce Wyse. 2007. AudiOdyssey: An Accessible Video Game for both Sighted and Non-sighted Gamers. In Proceedings of the Conference on Future Play. 251--252.
[30]
Deepak Gopinath, Siddarth Jain, and Brenna D Argall. 2016. Human-in-the-loop Optimization of Shared Autonomy in Assistive Robotics. IEEE Robotics and Automation Letters 2, 1 (2016), 247--254.
[31]
Alex Graves, Abdel-rahman Mohamed, and Geoffrey Hinton. 2013. Speech recognition with deep recurrent neural networks. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP.
[32]
Patrick Hanckmann, Klamer Schutte, and Gertjan J. Burghouts. 2012. Automated Textual Descriptions for a Wide Range of Video Events with 48 Human Actions. Lecture Notes in Computer Science (2012), 372--380.
[33]
SG Hart and others. 1988. Development of NASA-TLX: Results of empirical and theoretical research." inP. A. Hancock and N. Meshkati (eds.), Human Mental Workload. (1988).
[34]
Sepp Hochreiter and Jurgen Schmidhuber. 1997. Long Short-Term Memory. Neural Computation 9, 8 (1997), 1735--1780.
[35]
Andreas Holzinger, Markus Plass, Katharina Holzinger, Gloria Cerasela Cri¸ san, Camelia-M Pintea, and Vasile Palade. 2016. Towards interactive Machine Learning (iML): applying ant colony algorithms to solve the traveling salesman problem with the human-in-the-loop approach. In International Conference on Availability, Reliability, and Security. Springer, 81--95.
[36]
Ting-Yao Hsu, Chieh-Yang Huang, Yen-Chia Hsu, and Ting-Hao Huang. 2019a. Visual Story Post-Editing. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics. Association for Computational Linguistics, Florence, Italy, 6581--6586.
[37]
Ting-Yao Hsu, Chieh-Yang Huang, Yen-Chia Hsu, and Ting-Hao 'Kenneth' Huang. 2019b. Visual Story Post-Editing. (2019).
[38]
Ting-Hao Kenneth Huang, Francis Ferraro, Nasrin Mostafazadeh, Ishan Misra, Aishwarya Agrawal, Jacob Devlin, Ross Girshick, Xiaodong He, Pushmeet Kohli, Dhruv Batra, C. Lawrence Zitnick, Devi Parikh, Lucy Vanderwende, Michel Galley, and Margaret Mitchell. 2016. Visual Storytelling. In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies. Association for Computational Linguistics, San Diego, California, 1233--1239.
[39]
Common Objects in Context (COCO). Accessed date 2019-07--17. (Accessed date 2019-07--17). http://cocodataset.org/#home
[40]
Yu Jiang, Vivek Natarajan, Xinlei Chen, Marcus Rohrbach, Dhruv Batra, and Devi Parikh. 2018. Pythia v0.1: the Winning Entry to the VQA Challenge 2018. (2018).
[41]
Slava Kalyuga. 2003. The expertise reversal effect. Educational psychologist 38, 1 (2003), 23--31.
[42]
Slava Kalyuga. 2007. Expertise reversal effect and its implications for learner-tailored instruction. Educational Psychology Review 19, 4 (2007), 509--539.
[43]
Bongjun Kim and Bryan Pardo. 2017. I-SED: An Interactive Sound Event Detector. In Proceedings of the 22Nd International Conference on Intelligent User Interfaces (IUI '17). ACM, New York, NY, USA, 553--557.
[44]
Taehyeong Kim, Min-Oh Heo, Seonil Son, Kyoung-Wha Park, and Byoung-Tak Zhang. 2018. GLAC Net: GLocal Attention Cascading Networks for Multi-image Cued Story Generation. (2018).
[45]
Masatomo Kobayashi, Kentarou Fukuda, Hironobu Takagi, and Chieko Asakawa. 2009. Providing synthesized audio description for online videos. In Proceedings of the International ACM SIGACCESS Conference on Computers and Accessibility, ASSETS. 249--250.
[46]
Masatomo Kobayashi, Trisha O'Connell, Bryan Gould, Hironobu Takagi, and Chieko Asakawa. 2010. Are Synthesized Video Descriptions Acceptable?. In Proceedings of the International ACM SIGACCESS Conference on Computers and Accessibility, ASSETS. 163--170.
[47]
Dieter Koller, N Heinze, and Hans-Hellmut Nagel. 1991. Algorithmic characterization of vehicle trajectories from image sequences by motion verbs. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR. 90--95.
[48]
Richard E Ladner, Melody Y Ivory, Rajesh Rao, Sheryl Burgstahler, Dan Comden, Sangyun Hahn, Matthew Renzelmann, Satria Krisnandi, Mahalakshmi Ramasamy, Beverly Slabosky, and others. 2005. Automating Tactile Graphics Translation. In Proceedings of the International ACM SIGACCESS Conference on Computers and Accessibility, ASSETS. 150--157.
[49]
ListenByCode. Accessed date 2019-07--17. (Accessed date 2019-07--17). https://www.listenbycode.com/
[50]
LiveDescribe. Accessed date 2019-07--17. (Accessed date 2019-07--17).
[51]
Lara J Martin, Prithviraj Ammanabrolu, Xinyu Wang, William Hancock, Shruti Singh, Brent Harrison, and Mark O Riedl. 2018. Event representations for automated story generation with deep neural nets. In Thirty-Second AAAI Conference on Artifcial Intelligence.
[52]
Microsoft. Accessed date 2019-07--17. (Accessed date 2019-07--17). https://docs.microsoft.com/en-us/azure/media-services/video-indexer/video-indexer-use-apis
[53]
Joshua A Miele, Steven Landau, and Deborah Gilden. 2006. Talking TMAP: Automated Generation of Audio-Tactile Maps using Smith-Kettlewell's TMAP Software. British Journal of Visual Impairment 24, 2 (2006), 93--100.
[54]
Robert Munro. 2020. Human-in-the-Loop Machine Learning. MEAP.
[55]
United Nations. 2006. Convention on the Rights of Persons with Disabilities and Optional Protocol. (2006). https://www.un.org/disabilities/documents/convention/convoptprot-e.pdf
[56]
Devi Archana Paladugu, Zheshen Wang, and Baoxin Li. 2010. On Presenting Audio-tactile Maps to Visually Impaired Users for Getting Directions. In CHI '10 Extended Abstracts on Human Factors in Computing Systems (CHI EA '10). ACM, New York, NY, USA, 3955--3960.
[57]
Kishore Papineni, Salim Roukos, Todd Ward, and Wei-Jing Zhu. 2002. BLEU: a method for automatic evaluation of machine translation. In Proceedings of the 40th annual meeting on association for computational linguistics. Association for Computational Linguistics, 311--318.
[58]
Meg Pirrung, Nathan Hilliard, Artëm Yankov, Nancy O'Brien, Paul Weidert, Courtney D Corley, and Nathan O Hodas. 2018. Sharkzor: Interactive Deep Learning for Image Triage, Sort and Summary. (2018).
[59]
Venkatesh Potluri, Priyan Vaithilingam, Suresh Iyengar, Y Vidya, Manohar Swaminathan, and Gopal Srinivasa. 2018. CodeTalk: Improving Programming Environment Accessibility for Visually Impaired Developers. In Proceedings of the Conference on Human Factors in Computing Systems, CHI. 618.
[60]
Pythia. Accessed date 2019-07--17. (Accessed date 2019-07--17). https://learnpythia.readthedocs.io/en/latest/
[61]
Melissa Roemmele. 2016. Writing Stories with Help from Recurrent Neural Networks. In Proceedings of the AAAI Conference on Artifcial Intelligence, AAAI.
[62]
Jaime Sánchez and Matías Espinoza. 2011. Audio haptic videogaming for navigation skills in learners who are blind. In Proceedings of the International ACM SIGACCESS Conference on Computers and Accessibility, ASSETS. 227--228.
[63]
José Francisco Saray Villamizar, Benoît Encelle, Yannick Prié, and Pierre-Antoine Champin. 2011. An Adaptive videos enrichment system based on decision trees for people with sensory disabilities. In Proceedings of the International Cross-Disciplinary Conference on Web Accessibility.
[64]
Jessica Zeitz Self, Radha Krishnan Vinayagam, JT Fry, and Chris North. 2016. Bridging the gap between user intention and model parameters for human-in-the-loop data analytics. In Proceedings of the Workshop on Human-In-the-Loop Data Analytics. ACM, 3.
[65]
Karen Simonyan and Andrew Zisserman. 2014. Very Deep Convolutional Networks for Large-Scale Image Recognition. CoRR abs/1409.1556 (2014).
[66]
Ilya Sutskever, Oriol Vinyals, and Quoc V Le. 2014. Sequence to Sequence Learning with Neural Networks. In Advances in Neural Information Processing Systems 27, Z. Ghahramani, M. Welling, C. Cortes, N. D. Lawrence, and K. Q. Weinberger (Eds.). Curran Associates, Inc., 3104--3112. http://papers.nips.cc/paper/5346-sequence-to-sequence-learning-with-neural-networks. pdf
[67]
Christian Szegedy, Wei Liu, Yangqing Jia, Pierre Sermanet, Scott Reed, Dragomir Anguelov, Dumitru Erhan, Vincent Vanhoucke, and Andrew Rabinovich. 2015. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, CVPR.
[68]
Automatic Sync Technologies. Accessed Date 2019-09--15. (Accessed Date 2019-09--15). https://www.automaticsync.com/
[69]
Ashish Vaswani, Noam Shazeer, Niki Parmar, Jakob Uszkoreit, Llion Jones, Aidan N Gomez, Lukasz Kaiser, and Illia Polosukhin. 2017. Attention is all you need. In Advances in neural information processing systems. 5998--6008.
[70]
Steven Wall and Stephen Brewster. 2006. Feeling what you hear: tactile feedback for navigation of audio graphs. In Proceedings of the SIGCHI conference on Human Factors in Computing Systems. 1123--1132.
[71]
Yi Wang, Zhiming Luo, and Pierre-Marc Jodoin. 2017. Interactive deep learning method for segmenting moving objects. Pattern Recognition Letters 96 (2017), 66--75.
[72]
Doris Xin, Litian Ma, Jialin Liu, Stephen Macke, Shuchen Song, and Aditya Parameswaran. 2018. Accelerating human-in-the-loop machine learning: Challenges and opportunities. In Proceedings of the Second Workshop on Data Management for End-To-End Machine Learning.
[73]
YouDescribe. Accessed date 2019-07--17. (Accessed date 2019-07--17). https://youdescribe.org/

Cited By

View all
  • (2024)Automatic Video-to-Audiotactile Conversion of Golf Broadcasting on A Refreshable Pin ArrayProceedings of the 2024 ACM Symposium on Spatial User Interaction10.1145/3677386.3682092(1-8)Online publication date: 7-Oct-2024
  • (2024)Musical Performances in Virtual Reality with Spatial and View-Dependent Audio Descriptions for Blind and Low-Vision UsersProceedings of the 26th International ACM SIGACCESS Conference on Computers and Accessibility10.1145/3663548.3688492(1-5)Online publication date: 27-Oct-2024
  • (2024)Design considerations for photosensitivity warnings in visual mediaProceedings of the 26th International ACM SIGACCESS Conference on Computers and Accessibility10.1145/3663548.3675643(1-12)Online publication date: 27-Oct-2024
  • Show More Cited By

Index Terms

  1. Human-in-the-Loop Machine Learning to Increase Video Accessibility for Visually Impaired and Blind Users

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Conferences
      DIS '20: Proceedings of the 2020 ACM Designing Interactive Systems Conference
      July 2020
      2264 pages
      ISBN:9781450369749
      DOI:10.1145/3357236
      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 03 July 2020

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. blind users
      2. human-in-the-loop
      3. machine learning
      4. video accessibility
      5. video description
      6. visually impaired users

      Qualifiers

      • Research-article

      Funding Sources

      Conference

      DIS '20
      Sponsor:
      DIS '20: Designing Interactive Systems Conference 2020
      July 6 - 10, 2020
      Eindhoven, Netherlands

      Acceptance Rates

      Overall Acceptance Rate 1,158 of 4,684 submissions, 25%

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)128
      • Downloads (Last 6 weeks)15
      Reflects downloads up to 14 Jan 2025

      Other Metrics

      Citations

      Cited By

      View all
      • (2024)Automatic Video-to-Audiotactile Conversion of Golf Broadcasting on A Refreshable Pin ArrayProceedings of the 2024 ACM Symposium on Spatial User Interaction10.1145/3677386.3682092(1-8)Online publication date: 7-Oct-2024
      • (2024)Musical Performances in Virtual Reality with Spatial and View-Dependent Audio Descriptions for Blind and Low-Vision UsersProceedings of the 26th International ACM SIGACCESS Conference on Computers and Accessibility10.1145/3663548.3688492(1-5)Online publication date: 27-Oct-2024
      • (2024)Design considerations for photosensitivity warnings in visual mediaProceedings of the 26th International ACM SIGACCESS Conference on Computers and Accessibility10.1145/3663548.3675643(1-12)Online publication date: 27-Oct-2024
      • (2024)Towards Accessible Musical Performances in Virtual Reality: Designing a Conceptual Framework for Omnidirectional Audio DescriptionsProceedings of the 26th International ACM SIGACCESS Conference on Computers and Accessibility10.1145/3663548.3675618(1-17)Online publication date: 27-Oct-2024
      • (2024)Audio Description CustomizationProceedings of the 26th International ACM SIGACCESS Conference on Computers and Accessibility10.1145/3663548.3675617(1-19)Online publication date: 27-Oct-2024
      • (2024)Netizen A11y: Engaging Internet Users in Making Visual Media AccessibleCompanion Proceedings of the 29th International Conference on Intelligent User Interfaces10.1145/3640544.3645247(159-162)Online publication date: 18-Mar-2024
      • (2024)MIMOSA: Human-AI Co-Creation of Computational Spatial Audio Effects on VideosProceedings of the 16th Conference on Creativity & Cognition10.1145/3635636.3656189(156-169)Online publication date: 23-Jun-2024
      • (2024)SmartLearn: Visual-Temporal Accessibility for Slide-based e-learning VideosExtended Abstracts of the CHI Conference on Human Factors in Computing Systems10.1145/3613905.3650883(1-11)Online publication date: 11-May-2024
      • (2024)Making Short-Form Videos Accessible with Hierarchical Video SummariesProceedings of the 2024 CHI Conference on Human Factors in Computing Systems10.1145/3613904.3642839(1-17)Online publication date: 11-May-2024
      • (2024)SPICA: Interactive Video Content Exploration through Augmented Audio Descriptions for Blind or Low-Vision ViewersProceedings of the 2024 CHI Conference on Human Factors in Computing Systems10.1145/3613904.3642632(1-18)Online publication date: 11-May-2024
      • Show More Cited By

      View Options

      Login options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media