We designed the AdHocProx system to serve as a spatial-sensing foundation for a range of possible cross-device interaction techniques. This foundation consists of not only the hardware attachment and sensor processing components described so far, but also a fundamental user interface layer to provide a cohesive experience across all AdHocProx-powered interaction techniques. This user interface layer realizes a visual representation and interactive behaviors that support the concept of a “portal” between devices.
5.1 AdHocProx Portals
An AdHocProx
portal serves as a semi-public, limited shared space between two devices, much in the way that the porch of a home acts as a constrained social space to greet a possibly unfamiliar guest [
33].
The sending user can only place a certain amount of content into a portal; the receiving user pulls the content out of the portal to complete the transaction. This design avoids a collaborator potentially interrupting and foisting content onto the main work-area of one’s device at an untimely moment, as well as preventing (possibly unwanted) content from appearing on a receiving device, unless accepted by the receiver (via dragging the content from the local portal to complete the gesture). Note that the latter half of this handshake,
i.e. explicitly dragging content out of the portal via touch, serves as a salient gesture that supports social awareness through its visibility to both sender and receiver (while also observable to other collaborators nearby, to support small-group awareness as well).
A portal automatically reveals itself along a screen edge when a remote device approaches from the corresponding direction. This design affords cross-device collaborative activity in accordance with observation
(F4) of our formative study, where participants formed different territories with devices placed closer together to reflect collaborator’s intent to work together. Moving an AdHocProx device within a threshold of approximately 0.8 meters triggers the portal. We selected this threshold based on Hall’s notion of proxemic zones, where an arm’s-length of interpersonal distance defines an important perceptual and social boundary for one’s personal space [
28].
Our sensor processing pipeline runs locally on each device, recognizing which side of the remote device lies closest – and in what cardinal direction. The AdHocProx user interface layer then opens a portal along the corresponding edge of the local screen; the portal dynamically updates its on-screen placement if the remote device moves closer to a different edge of the screen, or disappears when the remote device moves away.
Due to the completely peer-to-peer nature of our round-robin UWB protocol, these portals open based solely on information available to each device locally. No external server or cloud service is necessary for spatial awareness, nor for portals to appear (or disappear) in response to devices coming and going, respectively.
The sharing of content is likewise peer-to-peer. Namely, AdHocProx shares references to files (rather than the content itself) by sending a URL over the spare bytes of the UWB ranging frames. For example, this can be a permissioned (read/write or read-only) link to an online document. Further, during the interactive drag-and-drop phase of a handshake gesture, AdHocProx keeps the coordinates of file icons synced between devices using these bytes as well. Due to UWB’s limited bandwidth, we do not attempt to directly send media such as images and videos; rather, by sending a reference only, the other device can retrieve it when WiFi (or server access, if hosted in-cloud) becomes available at a later time.
5.2 Interaction Techniques
Based on the four phases and collaborative behaviors observed in our formative study (section 4.3), we designed four interaction techniques on top of the portals user interface layer, as detailed below (and in Figure
13.)
Note that we use the scenarios explored and behaviors revealed in our formative study as points of departure, rather than limiting the resulting interaction designs to a literal interpretation of passing pieces of paper around (for example). We also co-opt “digitally-authentic” behaviors such as copy or pan/zoom interactions that tablet and mobile device users have come to take for granted. Nonetheless, by having cross-device versions of such digital super-powers rooted in natural user behaviors, our hope is that the interface can embody these more abstract actions, making them more salient and memorable once discovered.
Move. Once a portal opens, a user can drag a file into it. This experience draws direct inspiration from phase 1 of the formative study, where the physical act of handing a sheet of paper to collaborators lets groups quickly triage and pass related data to one another.
Unlike email or using a cloud service, in which the sending of a file is an entirely disembodied digital experience, the physical metaphor behind this interaction remains consistent with the natural expectations of in-person collaboration. The interaction requires physical co-location and (unlike a digital sync progress bar) leverages the act of dragging the file as a salient physical indicator of the offer, exchange, and completion; a receiver can see the sender’s hand approaching the portal, and likewise, the sender can observe if the receiver accepts their offering. The portal appears automatically, but the exchange of content is designed as a dyadic gesture with one portion completed by each user. (However, since touchscreens cannot detect which user makes contact, it remains possible for a single person to complete the gesture by reaching into the other user’s personal device-space, but such a gesture is observable to both users, and would only be socially acceptable with tacit consent).
Copy. In addition to dragging a file into the portal, a tilt gesture (implemented as a tilt exceeding 20 degrees) triggers the duplication of a file that falls into the portal on its own. The tilt must correspond to the direction of an open portal; tilting in other directions has no effect. Hence, the gesture requires proximity as well as corresponding actions, while fluidly integrating three distinct elements – the verb (Copy), the selection source device, and the selection destination device – into a single, unified command phrase.
Since there is no physical metaphor for the duplication of a sheet of paper, we draw on a micro-mobility gesture [
50] to initiate this action – namely the embodied, mutually visible act of tilting one’s device towards a collaborator to afford sharing of content. This consideration also arose during the formative study, particularly during phase 2, where many participants asked the researcher conducting the study if some kind of copy operation was allowed in order for multiple people to look at the same item closely. Our observations with regards to lining-up devices
(F1) and tilting the device as an indication of sharing information with others
(F3) also informed the design of this technique.
Pan. Unlike the asynchronous Move and Copy operations, Pan enables users to zoom and pan around an image displayed across multiple devices. Respecting the limitations of the portal, to activate a Pan operation, the initiating user enlarges (i.e. zooms into) an image until it expands to reach the portal. In order to confirm the transaction, the receiving user can drag or zoom the image out of the portal on their own device. At this point, both users are able to perform zoom and pan operations on the image. The two screens functionally operate as one large screen, such that one user can touch their finger to one screen, and the other user drags their finger across the other screen, if desired, achieving a distributed zoom operation. Pan was inspired by multiple examples of participants arranging sheets of paper next to each other (mostly in phase 3 of the formative study, while sharing findings) in order to get a larger window into the data visualizations (F1). For example, in many cases, two participants would have two different line graphs that were sequential along the X-axis, so they would arrange the two line graphs to be next to each other to give the feeling that it was one large graph; Pan attempts to create a digital version of this experience. Our observation of the different roles devices played (F2), indicating that actions on one device can impact user expectations for another in a context-dependent way, also informed this interaction technique.
Note. A remote device held at a large tilt (greater than 70 degrees, as used by GroupTogether [
57]) – much steeper than the 20-degree tilt that distinguishes
Copy – creates a sticky-note icon in the portal that the user of the local device can conveniently drag out and write upon immediately. This interaction was inspired by instances of participants holding a sheet of paper up for another person to inspect or take notes on (for example, see Figure
8).
Since this facilitated one person more easily consuming content while writing on their own device, we assigned the role of capturing notes (writing) to the device flat on the table – with the devices thus serving different roles per (F2). As a result, for the Note interaction the vertical device serves as the “remote” device, while the horizontal device flat on the table acts as the “local” device.
In phases 2 and 4 of the formative study, participants took notes or produced poster content (i.e., headers for the posters) while in this pose, respectively.