Collaborative Semantic Inference

Sebastian Gehrmann*, Hendrik Strobelt*, Robert Krueger, Hanspeter Pfister, Alexander M. Rush,

Automation of tasks can have critical consequences when humans lose agency over decision processes. Deep learning models are particularly susceptible since current black-box approaches lack explainable reasoning. We argue that both the visual interface and model structure of deep learning systems need to take into account interaction design. We propose a framework of collaborative semantic inference (CSI) for the co-design of interactions and models to enable visual collaboration between humans and algorithms. The approach exposes the intermediate reasoning process of models which allows semantic interactions with the visual metaphors of a problem, which means that a user can both understand and control parts of the model reasoning process. We demonstrate the feasibility of CSI with a co-designed case study of a document summarization system.

Paper (preprint)

Video

Citation

@ARTICLE {8805457,
author = {S. Gehrmann and H. Strobelt and R. Kruger and H. Pfister and A. M. Rush},
journal = {IEEE Transactions on Visualization & Computer Graphics},
title = {Visual Interaction with Deep Learning Models through Collaborative Semantic Inference},
year = {2020},
volume = {26},
number = {01},
issn = {1941-0506},
pages = {884-894},
abstract = {Automation of tasks can have critical consequences when humans lose agency over decision processes. Deep learning models are particularly susceptible since current black-box approaches lack explainable reasoning. We argue that both the visual interface and model structure of deep learning systems need to take into account interaction design. We propose a framework of collaborative semantic inference (CSI) for the co-design of interactions and models to enable visual collaboration between humans and algorithms. The approach exposes the intermediate reasoning process of models which allows semantic interactions with the visual metaphors of a problem, which means that a user can both understand and control parts of the model reasoning process. We demonstrate the feasibility of CSI with a co-designed case study of a document summarization system.},
keywords = {visualization;collaboration;semantics;tools;analytical models;cognition;predictive models},
doi = {10.1109/TVCG.2019.2934595},
publisher = {IEEE Computer Society},
address = {Los Alamitos, CA, USA},
month = {jan}
}


                                

Contact

Email: info (at) vizhub (dot) ai