Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
skip to main content
10.1145/3544548.3581354acmconferencesArticle/Chapter ViewFull TextPublication PageschiConference Proceedingsconference-collections
research-article
Open access

Stakeholder-Centered AI Design: Co-Designing Worker Tools with Gig Workers through Data Probes

Published: 19 April 2023 Publication History
  • Get Citation Alerts
  • Abstract

    AI technologies continue to advance from digital assistants to assisted decision-making. However, designing AI remains a challenge given its unknown outcomes and uses. One way to expand AI design is by centering stakeholders in the design process. We conduct co-design sessions with gig workers to explore the design of gig worker-centered tools as informed by their driving patterns, decisions, and personal contexts. Using workers’ own data as well as city-level data, we create probes—interactive data visuals—that participants explore to surface the well-being and positionalities that shape their work strategies. We describe participant insights and corresponding AI design considerations surfaced from data probes about: 1) workers’ well-being trade-offs and positionality constraints, 2) factors that impact well-being beyond those in the data probes, and 3) instances of unfair algorithmic management. We discuss the implications for designing data probes and using them to elevate worker-centered AI design as well as for worker advocacy.

    1 Introduction

    The advancement of AI has enabled a spectrum of innovative products and functions. From self-driving vehicles to mental health chatbots [8, 43], monitoring bank fraud to determining medical diagnoses [36]—when designed well, AI can be an assistive tool for humans in navigating their tasks.
    In the workplace, AI has increasingly automated repetitive, manual tasks and improved worker productivity and efficiency. Here, AI often manifests as algorithmic management—algorithms that take on managerial functions to oversee, assign tasks, and evaluate workers [34, 41]. For example, workplace tools such as Microsoft Suites or productivity apps include monitoring features so employers can track worker efficiency and throughput [3, 68]. Some workplaces may even continue to track and collect data on employees through these applications when they are off the clock [2]. These practices of algorithmic management are especially prevalent in gig work where data about workers is collected in droves and overbearing algorithms dictate worker tasks [66].
    Though vast amounts of data are collected through these applications on behalf of companies without consideration of the users, exploring one’s own data can be incredibly informative and empowering. Personal informatics (PI) research around how individuals desire to track, reflect on, and draw insights from their own data has shown how this can assist people in understanding their patterns [65] and even initiate behavior change [76]. Additionally, exploring and reflecting on their real data can support users in imagining future uses of their data, surfacing ideas that designers or practitioners might not otherwise come up with on their own [32].
    With the ubiquity of digital data, PI researchers have explored how individuals review and reflect on their own data collected by social media platforms [29, 47]. Researchers often use design probes, objects intended to help engage users in exploring open-ended questions [75]. Though often physical objects [28], design probes may also take form as digital artifacts and even incorporate data as a way to encourage users in investigating questions and design. For example, by presenting data to users through design probes, Gulotta et al. [29] demonstrate how probes can help individuals both surface insights about patterns in their data and inform designers on how to design digital systems. Similarly, Subramonyam et al. [73] demonstrated that these design probes using data, which they call "data probes", can also solidify AI system boundaries: they show how designers and engineers surfaced concrete use cases and limitations when considering user data while designing AI. These examples show promising direction toward the use of data for not only supporting individuals in understanding their own data but also advancing AI design. However, apart from [29], most existing work has explored how data probes can assist stakeholders in understanding their own data or practitioners in designing AI as opposed to how they might support stakeholders in co-designing AI.
    We explore this question—how to design AI with stakeholders to center their well-being and personal contexts—in the context of gig workers. Gig work platforms, where 16% of Americans were estimated to work in 2021 [5], is a fitting domain for us to explore our question due to the intense datafication of gig workers. With their work managed through mobile applications, a litany of data is collected about workers, from their hours worked and earnings, to more granular information such as their location even when they are not working. This presents a unique opportunity to explore how to co-design AI tools with workers by helping them harness their own work data through the exploration of data probes. We define data probes as digital objects which use historical data to help participants recall past or current patterns and events as they complete co-design activities.
    We design five data probes using workers’ own data—four visualizations and one AI work planning prototype—that workers interact with to surface their work patterns, well-being concerns, and positionality. During individual co-design sessions with 12 rideshare drivers, drivers reflect on well-being and positionality before interacting with the data probes created from their own work data as well as city-level aggregate data. For each probe, we introduce drivers to the metrics presented in the visualization and ask questions to encourage their exploration. We discover drivers use data probes as boundary objects to describe their work patterns and contexts, suggesting implications for AI design rooted in these contexts. Using probes, drivers 1) shared the well-being trade-offs and positionality factors they face in their work, 2) identified factors beyond those represented in their data probes which impact their work and well-being, and 3) identified instances of algorithmic management in their data. For each finding, we also share corresponding AI design considerations such as the creation of assistive AI to track and reflect on a driver’s well-being goals. Finally, we discuss the implications of our study on the design of data probes and the use of data probes for elevating worker expertise in AI design, including revealing individual contexts to inform AI design, as well as the potential for worker advocacy.

    2 Related Work

    We first describe the challenges researchers have faced when designing AI or ML and the approaches that have been taken thus far to expand our understanding of designing AI with the use of data. We then describe approaches advancing PI and stakeholder-centered AI in the context of gig work to motivate our design and methodology.

    2.1 Designing AI: Challenges and Recent Approaches

    Despite the increasing prevalence of AI technologies, the challenge of designing AI persists [24, 31, 80, 83]. Because of its dynamic nature, AI often resists the standard prototyping methods designers are used to, requiring necessary adaptations [81], and often leading to costly prototyping [24, 80, 83]. For example, Yang et al. [81] explored how designers used sketching to design NLP systems, revealing how sketching approaches had to be modified to support NLP designers. Additionally, designers often face uncertainty around AI capabilities [24, 83], potentially due to a gap in AI knowledge, which contributes to their struggles in envisioning AI use cases [24, 80]. Though this gap can often be bridged by designers working with developers to quickly understand the feasibility of their ideas [24, 58, 80, 85], designers may not always have access to these individuals while working with AI.
    One domain of research has investigated how the use of data may be able to assist with designing AI or digital systems. Researchers have employed data-based prototyping methods. For example, Holstein et al. [32] employed historical data through "Replay Enactments" to conduct feature prototyping, simulating experiences for teachers on technical systems. Zhang et al. [88] also used historical data with focus groups to elicit stakeholder feedback about future organizational decision-making practices. Subramonyam et al. [73] used data probes where designer-engineer teams considered end-user data to surface use cases and outcomes of AI. Others have studied how to use social media data with users to explore the design of digital experiences [29, 47]. Though they focused on designing a physical product rather than AI, Bogers et al. [11] used data as a design material to improve physical prototypes. Through interviews and a diary study, they describe how the collection and analysis of qualitative data allowed them to deepen their understanding of quantitative data (i.e., sensor-collected data) which was largely meaningless on its own. This method of data-enabled design allows for greater immersion of data probes, resulting in more engaging designs. These methods have primarily employed data for simulating experiences, technologies, and behaviors. We identify an opportunity to expand on the idea of data probes put forth by Subramonyam et al. [73] to explore how stakeholders’ own data may be explored and interpreted by them in order to design AI.

    2.2 Personal Informatics and the Gig Economy

    As our study seeks to create data probes that support stakeholders in reflecting on their data patterns, we turn to personal informatics (PI) literature, a well-explored field in a variety of domains, particularly for its support of individual user reflection. In their influential model, Li et al. [44] proposed using 5 stages to design PI systems. Significant research has sought to assist individuals in the latter two stages—reflection and action—when bridging the gap between users’ goals and data [17, 19, 20, 54]: these papers emphasize the importance of care in messaging taken between ’positive’ and ’negative’ nudges, and ongoing reflective support to create the ability for equally nuanced and personal reflections for users. You et al. [86] explored these two stages with drivers and their partners, finding that the combination of technology-sensing probes (wearables) with social sense-making (having drivers and their partners share reflections about the drivers’ behaviors) was promising for encouraging work-life balance for drivers. They facilitated drivers in reflection through summary tables of tracked data and diary entries, suggesting future work for increasing the granularity of data analysis for drivers.
    However, attention must also be paid to creating understandable and actionable measures for participants who are inexperienced with data, otherwise, researchers run the risk of participants abandoning their learned practices as soon as the study concludes [12, 26]. Boulard-Masson et al. [12] specifically suggested contextual and understandable quantifications for users, perhaps even at the cost of simplicity. For that reason, we chose to not only create data probes based on individual participants’ own data, but also present them side-by-side with local city-level data, so that they would be able to place their behaviors in the context of their peers. We predict that this will be an effective method of not only contextualization, but also as a self-reflective exercise, comparing self to others.
    Algorithmic management is a particularly difficult, yet important, domain to examine PI data. Compared to traditional PI apps, the goals of rideshare apps for data collection center around maximizing profit rather than helping users, such as drivers, understand their data patterns; thus platform data practices are intentionally opaque. This leaves drivers to collectively sensemake gamified metrics [22, 74]. Although there are some PI tools for rideshare drivers—including apps such as Gridwise, Hurdlr, or Stride [27]—these apps have several challenges. These include paywalls or ’premium’ features for personalized insights, different features between apps, mixed functionality across apps and platforms, complexity of using separate applications simultaneously, importing data, and drivers interpreting features and statistics on their own. As a result, drivers often have to calculate statistics themselves mentally or use simpler data analysis tools, limiting their ability to learn more advanced insights [87].
    Drawing on the goals of these tools and recent data privacy policy changes, we identify the potential to democratize personalized data tools for gig workers and for researchers to study the effect of access to these tools on well-being and performance. For example, the EU’s General Data Protection Regulation (GDPR) ensures the user’s right to their own data including on gig work platforms. This opens up the possibility for gig workers to investigate and harness their own data for exploration and insights. While the United States may not have policies on the same scale, a number of companies have shifted towards open data practices by allowing users to download their own data (e.g., Facebook, Google, Uber). This presents a prime opportunity for researchers to investigate the design of AI in the domain of digital work.

    2.3 Stakeholder Involvement in AI Design

    Past research on human-centered design methods for AI has primarily focused on assisting UX or ML practitioners, for example, Yang et al. [82] interviewing UX practitioners for their challenges and practices when using AI as design material, Subramonyam et al. [73] conducting co-design sessions to see how data probes can support designer-engineer teams developing AI, and Yildirim et al. [85] exploring the practices of enterprise designers working on cross-functional teams. In recent years though, calls have been made to use human-centered AI design methods to address the ethical issues that arise when AI is created void of stakeholder contexts to inform potential outcomes and harms of a technology [46, 77]. Resources and guidelines have been published to support designers in the pursuit of human-centered AI that is more inclusive of impacted stakeholders [6, 21, 23, 69, 78].
    Researchers have conducted interviews or workshops [14, 72] and co-design or participatory design sessions [60, 87] with impacted stakeholders and shown how this can generate AI design ideas based on stakeholders’ own experiences and situations. Researchers have also worked specifically with gig workers. Zhang et al. [87] held co-design sessions with rideshare drivers to surface drivers’ ideas for how to address or redesign algorithmic management; Bates et al. [9] used co-design with courier drivers to examine their complex relationship with data, finding that reliance on PI tracking tools like Strava repurposed for gig work, can help fill the pieces left missing by the platform; and Alvarez de la Vega et al. [4] utilized design fiction as a means of co-designing systems to support freelancers. Co-design has also been used as a method for creating PI systems [38, 63, 84], however, most such studies focus on users as consumers. Worker PI data, encompassing their job and livelihood, is a meaningful direction for our work.
    Integrating these methods to design AI for gig work is particularly relevant given the lack of user context in algorithmic management or worker-related models. Uber and Lyft currently use an advanced dynamic matching and pricing algorithm to connect drivers with riders, often in just seconds [79]. While highly efficient, these algorithms also require advanced decision-making from drivers, usually before they’ve even completed their previous task. Named dynamic waiting, this has negative impacts on driver well-being by including very limited driver input to indicate work preferences [87]. Separate from platforms, researchers such as [15] and [59] have developed algorithms using worker data to understand their earnings or how to improve them. However, these models do not currently incorporate worker context, which can lead to tools workers cannot use or miss the opportunity to discover unique worker-centered questions.
    These instances and others centering stakeholders [4, 7, 33, 40, 45, 89] illustrate the potential in working with stakeholders to inform, through personal contexts and experiences, how to design AI.

    3 Background: Partnership with Independent Drivers Guild

    In the Summer of 2021, we began working with the co-founder of Independent Drivers Guild of Illinois in the U.S. to better understand the well-being challenges and algorithmic management faced by drivers for our then-ongoing study design. Shortly after, we brought him on in a compensated position as a community advisor to periodically seek his feedback about whether our research direction aligned with what would be most useful for drivers. In Fall 2021, following our first study, we presented the results to our community advisor to get his opinion about our ideas for follow-up research directions as well as discuss his ideas given his direct driver advocacy work. He was particularly interested in the data-driven ideas of drivers from our study, suggesting the potential for using data to assist drivers in comparing their individual metrics with city-level drivers. For example, he conceptualized an AI tool that could provide a weekly plan to the driver, customized to an individual’s driving habits for where and when they should work given their goals and constraints. He introduced us to Chicago’s public dataset of rideshare data and encouraged us to explore ways this could be used to assist drivers.
    This was a key factor in guiding our research direction toward how to design AI with gig workers through data probes created from individual data and city-level data. We began by using public city data to create data visuals on Tableau about rideshare earnings. We then reached out to our community advisor in Winter 2021 to gain feedback around these visuals and his thoughts on which ones drivers may find useful to explore, and ultimately moved forward with the visual that he felt was the most beneficial—a preliminary work planning tool to predict weekly driving earnings. We also began exploring drivers’ individual data in order to see how this data could be used to help drivers surface ideas from their work to design AI. We learned the Uber platform allows drivers to request their data, and we reached out to two previous study participants in Spring 2022 who graciously agreed to provide their driving data in a compensated pilot exploring individual data.1
    From these pilots, we gained a better sense of how the drivers might benefit from exploring their data using data probes—most prominently, we noted that both drivers naturally used the visualizations they saw (see Appendix for Figures 3 and 4) to reflect and narrate specific events they recalled (e.g., a trip that went to Milwaukee), temporal patterns they had internalized (e.g., bar closing times), and well-being or personal situations that affected their driving patterns (e.g., a family member passing away from COVID). Additionally, one of the drivers used the data to explain specific ways he wanted to review his data to investigate whether certain patterns he followed led to successful earnings or not. For example, he explained he anticipated a drop in summer ridership volume and devised a tactic to try to increase his tips and subsequently his earnings per trip. He explained that reviewing his data for this time period could help him determine whether that idea had a tangible effect: "It’d be nice to know if, if I’m—if that extra effort is having an impact."
    These interactions inspired us to revise the data visualizations to their current version in our study, where we focus on using data probes to support drivers’ reflection and storytelling and surface specific contexts of their driving well-being and strategy formulation to inform AI design. In section 4, we describe our design goals, principles, and challenges that we address in formulating our session activities and data probes in order to reveal sensitizing concepts for practitioners designing AI.

    4 Designing Data Probes

    Table 1:
    No.Data Probe NameData TypeDescription
    1Driving Animation
    (Animation)
    PersonalVisceral animated gif showing a specific day
    of a worker’s movement patterns on a map of their city
    2Neighborhood Map
    (Map)
    Personal
    &
    City-level
    Interactive map showing the neighborhoods of where
    pick-ups have occurred. Drivers can hover or click on
    neighborhoods to see statistics such as when and how
    many trips occurred, average fare, and miles per trip.
    3CalendarPersonal
    &
    City-level
    Intuitive monthly and weekly calendars that display
    earning trends and breakdown of trips for a
    specific day or day of the week.
    4HourlyPersonal
    &
    City-level
    Interactive bar charts that use personal or city-level
    data to display trends around earnings and trips
    depending on hour of the day.
    5Work Planner
    (Planner)
    City-levelInteractive prototype that allows a user to input work
    parameters (e.g., hours, days, and neighborhoods
    worked) and uses city-level data to provide an
    estimation of base earnings, tips, and mileage for the
    driver in a week
    Table 1: Summary of the 5 Data Probes. Further details about the data used to create them can be found in Sections 4 and 5.2.
    We were drawn to how detailed and reflective our pilot participants’ insights were about their data, from the specific rides they recalled to how their personal contexts shaped their patterns. We wondered how the contexts drivers surface from exploring their data might inform AI design or AI-based tools for drivers. Designing AI can refer to different design phases—such as design space exploration which Zhang et al. [87] pursued through scenario-based co-design sessions with drivers to explore algorithmic management interventions—or it may refer to surfacing specific components for AI tools such as data types, feature engineering, prediction and optimization goals, and constraints. One set of ideas drivers surfaced in [87] centered around using their own data to discover insights and support their well-being. Building off of this, our pilot studies, and past work around AI design with data probes (Section 2.1), our goal became to co-design AI—explore components of AI tools such as feature engineering and constraints—with drivers through data probes—visualizations created using their own data. To design effective data probes that drivers feel empowered to use, regardless of experience in data analysis, we identified design principles to adhere to.

    4.1 Supporting Reflection and Action

    For our first principle, we wanted to ensure the data probes supported drivers’ ease of understanding when reflecting on their data and identifying subsequent actions they can take [44]. Cognizant of how data analysis can be daunting to participants without analysis experience [52], we wanted to design data probes to reflect drivers’ data in forms that are familiar to them, ideally, formats that can remind them of everyday representations.
    This led us to design one data probe that displays driver data on a calendar. We chose this based on how drivers in our previous study often discussed reviewing earnings at the end of the day or week. Additionally, given the ubiquity of calendars, we hoped a calendar data probe could support drivers in reflecting on their personal calendar of activities and discuss how life events or contexts influenced their driving.
    Based on discussions with our community advisor and the fact that maps are the most common artifact drivers interact with, we created two more data probes: 1) an interactive map of Chicago with earnings trends by neighborhood, and 2) an animation displaying a day of the drivers’ trips. Drivers we spoke to often referenced the maps within rideshare apps when describing the information they use for decision-making. However, these maps only display immediate demand of certain regions rather than insights about a driver’s personal patterns. Thus, we created an interactive map to support drivers in generating location-specific insights. Similarly, we presented drivers with personalized animations that trace their point-by-point movements to make their data more tangible for reflections.
    We included the work planning tool, Planner, as a data probe as our community advisor believed it could be beneficial for drivers and because it presents a use case that can help drivers anchor their ideas for AI components such as predictions or constraints to consider. Drivers can input driving parameters such as times and places they work, and weekly expenses to obtain weekly predictions (e.g., total earnings, trips, and miles in a week). This is displayed as a table and text summary to support individual preferences of reading and reflecting on data. See Figure 2 for the Planner’s full input and output options and Section 5.2.2 for how the Planner works.
    For our final data probe representing drivers’ earnings based on trip pick-up hour, we cycled through many iterations, finally landing on a bar chart due to its simplicity as a visual compared to the other options.
    To further support drivers in reflection, we incorporated heat maps into the calendar and map data probes. The colors allow participants to easily identify profitable days or neighborhoods. We also designed questions to begin task-based and then become open-ended, as opposed to asking open-ended questions like "what are your thoughts?" which can be intimidating given the data they are presented with. These data probes are summarized in Table 1 as well as Figure 1 which situates each probe in the co-design session. We describe the data populating these probes in Section 5.2.

    4.2 Centering Context

    Our second principle centers on individual contexts that drivers hold which influence their work preferences, patterns, and limitations. Having the drivers discuss their strategies and patterns within the contexts that bind their work may make AI capabilities more certain and outputs more attuned to what drivers need. Our approach is largely informed by suggestions from D’ignazio and Klein [23]. We use visualizations for our data probes to make patterns visible, followed by probing questions about the data to uncover the context behind the presented patterns.
    In order to elevate driver context, we incorporate well-being and positionality considerations prior to exploring data probes and then throughout questions as they explore data probes. To encourage drivers to consider the personal contexts that influence their work, we introduced participants to the concepts of well-being and positionality at the beginning of each session. For this study, we focus on three types of well-being: physical, psychological, and financial, building from Hickok et al. [30]’s definitions of the terms [42]. We also have drivers complete a positionality activity—"the personal values, views, and location in time and space that influence how one engages with and understands the world" [35]—by having them consider what factors advantage and disadvantage them as a driver. Some of these factors are derived from existing positionality wheels (e.g., "Level of Education, Age, Race & Ethnicity") [56] while others we added specific to gig work ("Own vs. Rent My Car", "Single vs. Multi-Platform Worker"). Later, as drivers interact with data probes, we ask how their work patterns affect their well-being and remind them of the positionality factors they identified to see if/how any of their factors impact their choice to work or not at a certain time or place.

    5 Methods

    5.1 Study Participants

    Table 2:
    PAgeRaceGenderEducationGig-work
    Tenure
    Hours
    per Week
    Household
    Size
    Other
    Job?
    Platforms Worked
    (primary first)
    133AsianMalePrefer not to answer5 years+30 - 45Prefer not
    to answer
    NoUber, Lyft Amazon
    Flex, DoorDash,
    Grubhub
    232HispanicMaleSome graduate school2-5 years45 - 602NoUber, Lyft, DoorDash
    335AsianMalePostgraduate/prof. degree6-12 months45 - 603NoUber
    456WhiteMaleSome college, no degree5 years+0 - 152YesUber
    550AsianMaleSome college, no degree5 years+30 - 454NoUber
    655WhiteMaleSome college, no degree5 years+30 - 451NoUber, Lyft
    743WhiteMalePostgraduate/prof. degree2-5 years30 - 451NoUber, Lyft
    849WhiteFemalePostgraduate/prof. degree2-5 years15 - 301NoLyft, Uber
    935AsianMaleHigh school5 Years+45 - 603NoUber
    1040WhiteMalePostgraduate/prof. degree5 Years+0 - 153YesUber, Lyft
    1154WhiteFemaleTwo-year associate degree5 Years+30 - 453YesLyft, Instacart
    1245WhiteMaleSome college, no degree2-5 years45 - 607NoUber, Lyft, Rodeo
    Table 2: Participant Demographics and Work History

    5.1.1 Participant Criteria.

    We recruited active Chicago Uber drivers. We required participants to be Uber drivers primarily because Uber is the only rideshare platform we identified as allowing drivers to easily request and download their data. The participants had to be actively driving because one of the data probes we showed was based on their geodata which Uber only provides for the past 30 days. Finally, the participants had to be working in Chicago because of available city-level rideshare data with earnings.
    We used multiple recruitment methods: Facebook Ads, posts in Facebook Groups, posts on subreddits, Reddit ads, word-of-mouth, and emails through our partner channels. We distributed a pre-screening survey. In addition to basic demographic data, we asked respondents to self-classify the permanence of their gig work careers [25], whether they had another job, and how much they relied upon their rideshare income ("nice to have, but not essential to my budget" or "essential for meeting basic needs" [71]). Our research team reached out to respondents, detailing instructions on how to download their Uber data and share with the research team.

    5.1.2 Participants.

    We had n=12 total participants. Their average age was 43.9 years (SD = 2.57). 7 of our 12 (58.3%) participants identified as White, 4 (33.3%) as Asian, and 1 as Hispanic (8.3%). 10 of our participants (83.3%) identified as male, with the remaining 2 identifying as female (16.7%). 11 of our participants (91.7%) have been working for rideshare platforms for 2 years or more. 10 participants (83.3%) also reported working at least 30 hours a week on gig platforms, with the same 10 reporting that their gig work income was "essential for meeting basic needs"; 2 participants reported that their income was "nice to have, but not essential to my budget".

    5.2 Data Probe Creation

    We first describe the data sources we used and the four visualization data probes. We then explain the fifth data probe, the Planner—its mechanics and computations. Most data probes were created primarily using Tableau, an interactive data visualization platform.

    5.2.1 Data Sources to Create Data Probes.

    Individual Data. To create individual data probes (see Figure 5) we utilized the driver’s own Uber data2.
    Personal Trip Animation Data Probe. We created a personalized trip animation for each driver, combining trips with location data from the past 30 days. With this file, we generated a time-lapse animation of participants’ movements overlaid on a map of Chicago using Unfolded.ai3. For the session, we isolated the animation to one recent day of trips. The purpose of the animation was to give participants a visceral representation of their movement data as they described their general work strategy.
    Individual Hourly, Calendar, and Map Data Probes. Next, participants saw their hourly data probe, a bar chart displaying the average fare per minute at each hour of the day for the month of June. Then, they reviewed their calendar displaying similar metrics, summarized for each day since June 2022. For the map data probe, participants saw two personalized maps, colored in varying shades based on the average fares per minute in each neighborhood. The first showed their common pick-up locations (as well as total pick-ups per area upon hover). The second showed their common drop-off locations. These maps were interactive: selecting a pick-up neighborhood altered the drop-off map to display only drop-off locations for trips beginning in the selected neighborhood.
    City-Level Data. The data for city-level data probes came directly from the City of Chicago4. Available back to 2018, we limited our dataset to the most recently available month (June 2022) because of size and processing power limitations. To prepare the dataset for analysis, we used Pandas—a Python-based data manipulation tool—to classify all the pick-up and drop-off coordinates into Chicago’s official neighborhoods, as defined by the City of Chicago5. Next, we used Pandas to add time-appropriate weather data acquired from VisualCrossing, a live weather API6.
    City-Level Hourly, Calendar, and Map Data Probes. City-level data probes were all presented side-by-side with their personal counterparts. These were virtually identical to their personal counterparts, only differing in that these versions utilized city-level driver data (see Figure 6). For the calendar, weekly calendars were used to compare individual and city-level data because the Chicago dataset does not identify unique drivers. Drivers saw a matrix of 4 variables per day of the week: total number of trips (for the entire city’s dataset), average fare, average trip duration (with passengers in the vehicle), and average fare per minute. The map data probe displayed an interactive map of Chicago, color-coded to indicate neighborhoods with the highest average fares per minute.

    5.2.2 Work Planner.

    We created the Work Planner (Planner), a comprehensive simulation of predicted earnings and expenses using the aforementioned Chicago data. Figure 2 displays the full tool details including input and output variables.
    To use it, drivers input driving preferences: the number of hours driven in a hypothetical week, days of the week driven, hours of the day driving, neighborhoods they choose to pick up in, outdoor temperature, and precipitation. Weather variables were included based on past drivers sharing they believe weather may impact their earnings. Drivers also enter expenses including the price of gas, their vehicle’s mileage, their car insurance cost, and other miscellaneous expenses. Expense calculation, something rideshare platforms currently lack, was included to give drivers a more realistic sense of their net profit. They also adjusted other variables important to the simulation including the portion of their fare that the platform withdraws, and the percent of their time spent with passengers in their car. The first was included to respect drivers’ personal beliefs of how much platforms take which often differs from what platforms formally claim. The second was included as drivers previously warned their hourly earnings rate includes unpaid time spent looking for passengers or driving to them. For these two variables, we set the default parameters to 25% and 55%, standard values for Chicago [49].
    Output statistics are generated by using the participant inputs to filter the city-level data to the subset of trips which match these. For example, to calculate weekly earnings, the Planner first calculates the average fare (AF, $) and average trip duration (ATD, minutes) directly from the data subset. Next, it determines the number of projected trips (PT) through the equation:
    \(\begin{equation} PT = \frac{60}{ATD}*TPC*HPW \end{equation} \)
    (1)
    where TPC refers to the percentage of time a driver spends with passengers in the car and HPW represents the hours per week the driver inputs as working. Finally, the weekly earnings (WE, $) is obtained:
    \(\begin{equation} WE = AF * PT \end{equation} \)
    (2)
    Similar calculations are done to obtain the other statistics.
    We explained to participants that the calculations were based on June 2022 data from only the Chicago area. We also explained that based on their selections for time and location of work, the tool’s output would adjust accordingly to take into consideration trips started on those days, hours, or neighborhoods.
    Figure 1:
    Figure 1: Co-Design Procedure: 1. Participants interact with the Planner to input parameters of their typical work schedule and expense-related information. They review their predicted earnings for whether it aligns with their real experiences. 2. Participants discuss well-being and positionality to identify well-being goals they have and the advantages or disadvantages they face as drivers. 3. Participants view an animation of the trips they drove on a map of Chicago, using their own driving data. 4. Participants review individual data probes to identify trends and/or reflect on what affects their work habits. 5. They compare individual data probes with ones made from Chicago city-level driver data to reflect on similarities and differences with the city driving population. 6. Participants create another work schedule based on the trends they identified from the probes and compare the results with the schedule from step (1) to review which satisfies their preferences.

    5.3 Co-Design Procedure

    Participants began with the Planner to create their typical weekly schedule. Next, we introduced them to the three types of well-being and asked whether they wanted to improve any of them in their work. Participants completed a positionality activity, where they considered 1-4 factors that advantage them and 1-4 factors that disadvantage them in their work from a set of factors often used in positionality exercises [56] in addition to ones related to gig work. The section concluded with participants answering, "Is there anything you’d like to know or want to do to improve your work strategy?"
    Participants then explored data probes created using their personal Uber data. The first probe was an animation of their Uber trips that they watched as they described their typical driving strategy. The second probe was their hourly bar chart of the earnings they made during June 20227. The third probe was the calendar, from June to their most recent available data, where days were shaded based on their total earnings. Finally, we showed participants their map data probe reflecting their pick-up and drop-off locations and fare patterns for each neighborhood based on the past 30 days.
    With each probe, we guided participants through task-based questions (e.g., "Which days does the calendar show you made the most money?"), followed by questions to understand their patterns, the factors that influence them, and the impact on their well-being. Participants could also hover over each probe to review metrics about their work (see supplementary information to view full metrics per probe in the hover).
    Next, participants compared their individual data probes with corresponding ones created from Chicago drivers: hourly, calendar, and map data probes. Participants viewed their individual and city-level hourly data probes side-by-side and were asked to identify differences and similarities, and reasons for why they believed those exist. This continued for the calendar and map data probes.
    Finally, participants created a new schedule using the Planner based on the insights they drew from the data probes about optimal hours, days, and locations to work. They compared the results of their original and new schedule and shared their reflections and preferences between the two. Figure 1 shows the co-design session flow. The session concluded with participants reflecting on their experience with the five data probes, thinking specifically about how the probes helped them and which visualizations they liked the most.

    5.4 Analysis

    All co-design sessions were conducted and recorded using Zoom, and transcribed using Otter.ai. Each session lasted two hours on average, and participants were compensated with a $90 gift card. Notes and transcriptions were reviewed by the interviewers and notetakers to produce summaries for each participant. These summaries (containing notes, memos, and reflections) were clustered based on protocol areas and data probes. The data was then analyzed following Patton [62]’s qualitative data analysis method. Two researchers coded the data and grouped them for emerging themes. The entire research team then discussed these themes to derive the final themes reported in this paper.

    5.5 Background on Uber Features

    We provide a brief explanation of Uber platform features referenced by our participants to provide clarity for our findings. Quest promotions are Uber’s incentive-based method of encouraging drivers to increase their short-term driving8. Quests offer drivers a limited time incentive for completing a set number of trips; e.g., ’Complete 60 trips between Friday and Sunday for a $150 bonus’. Quests appear unevenly, with varying thresholds and rewards. Many drivers suspect that Quests unfairly favor newer or inactive drivers. However, because of the opacity of the promotions, they lack the means to definitively prove this suspicion [87]. Surge promotions are Uber’s method of matching the supply of drivers with the demand from riders9. Uber will temporarily create "surges" in high demand areas to incentivize drivers to relocate to those areas. Uber Pro is Uber’s driver rewards programs that automatically enrolls all Uber drivers10. There are four tiers of Uber Pro: blue, gold, platinum, and diamond. Driver tier is determined by a variety of factors including driver rating (the average of their last 500 ratings from passengers), cancellation rate, trip acceptance rate, and the number of ’points’ earned. Points are re-calculated every 3 months, and earned by completing trips with certain high-demand times earning more points per trip.

    6 Findings

    In co-design sessions, we observed that data probes acted as boundary objects for participants to effectively communicate their work patterns and contexts with our research team and suggest AI design implications rooted in these contexts. In this section, we describe three main insights surfaced by participants through data probes and corresponding (AI) design considerations. First, using data probes, participants shared how well-being trade-offs and immovable positionality factors pose constraints for how they could use the Planner—the Tableau schedule and earnings simulator we created—practically, but suggested alternative ways for an algorithmic work planning tool to center well-being within their constraints. Second, participants explored data probes to identify factors beyond those represented in their data probes that impact their work and well-being, offering new prediction ideas for AI tools that can support them. Third, participants used data probes to identify specific instances of unfair algorithmic management and suggested ways to repurpose data probes and design algorithmic tools to audit platforms.
    Figure 2:
    Figure 2: Work Planner data probe that participants interacted with to view predictions of their schedules and surface design considerations.

    6.1 Well-Being Trade-Offs and Positionality Constraints That Impact Worker’s Strategies

    Participants used their data probes to explain how these visualizations reflect the ways they currently balance well-being in their strategy as well as how personal obligations constrain their work flexibility. It is important for worker-centered AI design to be considerate of these limitations in order to provide well-being support that is actually feasible for workers to take.

    6.1.1 Surfacing Well-Being and Positionality Through Hourly, Calendar, and Map Data Probes.

    Initial reflections around well-being and positionality. When we first introduced well-being and positionality, a handful of participants raised well-being concerns such as safety or exhaustion (P3, P8, P10), stress over precarity of earnings and passenger attitudes (P3, P12), and frustration over traffic or unfair treatment by platforms (P8, P11). For positionality, participants shared advantages that directly impact their earnings such as “Being a Multi-Platform Worker” or reduce their expenses such as “Owning My Car”, and disadvantages that limit their earning ability such as living far away from centrally located, high earning neighborhoods.
    Articulating well-being trade-offs and positionality constraints using hourly and map data probes. Through data probes, participants surfaced or expanded on initially identified well-beings and positionalities. The hourly and map data probes were especially effective as tools for participants to discuss specific well-being trade-offs they make—such as balancing physical and mental health with financial gain—and positionality constraints they face—particularly familial obligations. These probes were likely effective given participants held safety concerns about Chicago’s high rate of carjackings and shootings (P1-3, P5, P8-12), which are highly associated with late night/early morning hours and specific neighborhoods [1, 57]. Participants explicitly identified high earning hours and neighborhoods on hourly and map data probes of city-level data, saying the probes confirm what they already “know” are profitable strategies. Then, they explained their individual data probes reflect their conscious decisions to not drive at these times or areas because of well-being concerns and positionality limitations. P2 said, "maybe it [working late night hours] could have a positive impact on my finances, rates, more earnings, but I just wouldn’t feel safe. And I would feel, you know, anxiety”, and P10 referenced how his fear of his daughter losing her father prevents him from driving in certain areas. In contrast, P4 did not hold the same concerns over his well-being, instead reflecting a high level of empathy towards the positionalities of residents of these neighborhoods. He accepted rides regardless of pick-up or drop-off locations because he did not want to see people lose access to rides simply because they lived in more "dangerous" neighborhoods.
    Surfacing obligations imposed by workers’ positionalities using the calendar data probe. The calendar data probe surfaced more unique perspectives around the kinds of limitations worker positionalities impose on their work patterns. As people often use calendars to record and view their daily tasks, this probe was intended to help people intuitively recall their data in the context of their responsibilities and routines. Accordingly, while exploring his calendar, P2 explained his data probe showed few working Saturdays and Sundays because he aligns his days off to be with his girlfriend. Comparing the city-level and his individual weekly calendars, P3 lamented being unable to capitalize on high-earning weekends due to caring for his infant child. P1 explored his individual monthly calendar for memorable low- or high-earning days, and expressed regret when he remembered that he had been unable to work on July 4th because of family obligations. P5, P7, P9-10, and P12 also expressed similar constraints in their weekly schedules due to commitments with family and friends.

    6.1.2 Design Consideration for Centering Well-Being in AI Design.

    Developing assistive AI for tracking and reflecting on well-being goals. While exploring their data, several participants commented how they try to prioritize their physical and psychological well-being with days off but are not always successful. For example, P5 told us Sundays are family days, yet his calendar showed him working half of all Sundays. P6 tries to prioritize time with family and friends as he gets older, but rather than treating it as time off, he feels he must make up for it later. P2 expressed surprise when he saw his calendar, explaining he thought he only worked 15 days in June when in reality, he worked 25 days in the month. Data probes illuminated the difference between what drivers believe they do for their well-being versus what occurs in practice. AI tools can focus on supporting drivers in the same regard. This could be through an AI work planning tool which uses a driver’s data to present statistics about recent days worked so they can see whether/how they are meeting well-being goals and how to integrate that knowledge into the next week’s plan.
    AI tools could also assist drivers with well-being centered schedules by predicting when to take breaks or what past factors led to profitable days. Unsurprisingly, long-tenured drivers felt confident in their driving strategy and did not seem receptive to change. But a few mentioned an interest in using their data to influence their strategy. P6 wanted to use his data to predict profitability so he could schedule appointments or breaks on days with lower predicted earnings. P12 shared how he would like a predictive model to help him decipher what factors led to him having a good or bad day (in earnings), "As a driver, you can analyze it and be like, ’Man, I’ve really had a bad day. Why did I have such a bad day?’ Or ’I had a great day. What is it that made it such a great day?’" Similar to P6, P8 was interested in using her data to plan her day off, adding the benefit of taking time for herself: “I tend to do better with tips back after a couple days off, which means I’m just getting burned out of people.” Addressing her comment, AI tools could also use workers’ own data to enable them to track and reflect on their time off and see when they may benefit from a break.
    Respecting driver contexts and knowledge. For most participants, viewing data probes did not lead to changes in their work strategies due to immutable constraints—not driving at night due to safety concerns despite the potential profit or being a caretaker limiting the time they may work. However, they gave suggestions for how to design AI tools that can assist them in decision-making for well-being. This is a reminder that a tool designed without considering worker contexts may very well violate their personal constraints or expertise and will unlikely be adopted. Instead, AI tools should take on assistive roles, leaving final decision-making to workers themselves. For example, by their own admission, drivers get burned out, thus a tool that can track and help them reflect on the effect of taking time off could be more effective in advancing worker well-being than one which prescribes to them to take a break.
    Additionally, we observed that for a few participants, the session did lead to recognizing potential changes to improve their well-being. P2, though an experienced driver, described himself as being in a “transition period”: he was moving and his girlfriend recently changed her days off, so he was developing a new schedule. Though confident in the merits of his old schedule, he showed interest in using the Planner to simulate how to tweak his schedule. P1 and P11 both considered testing new schedules which estimated fewer earnings but also meant fewer or different hours in order to improve their psychological and physical well-being. Though they were not interested in change, P8 and P12 shared that reviewing the data probes and using the Planner could help beginner drivers instead of established drivers. Once again, through understanding their context, we can more appropriately determine the use and audience of tools like algorithmic work planning. Rather than catering to general users, a work planning tool could be designed to center around onboarding new drivers or experienced drivers with an existing interest in tweaking their strategy.

    6.2 Identifying New Factors Not Included in Data Probe Design

    Through data probe exploration, participants identified additional factors which impact well-being as well as ideas of predictions they wanted to see from AI tools. Participants also used the data probes to recognize and explain gaps in their Uber data which resulted from other platform work.

    6.2.1 Identifying Additional Factors that Impact Well-Being Through All Data Probes.

    Our data probes centered around participants reflecting on schedule-related factors of work (e.g., time, day, location). Encouragingly, exploring their data in these ways also allowed participants to remember factors we did not include in the data probes but that they felt were influential on well-being, including passenger characteristics, general work precarity, perceived success of driving patterns, and multi-platform work.
    Passenger characteristics. Reviewing their hourly data probes, P7 and P9 felt they currently work optimal hours which do not lead to exhaustion or stress. Instead, P7 reflected that far more impactful on his well-being are interactions with passengers—negative interactions can pull his mood down: "No one wants to interact with someone who...is mistreating them or being obnoxious." Additionally, positive interactions with passengers can boost drivers’ moods. By inspecting her calendar data probe for high earnings days, P8 recalled that Saturday afternoons are often profitable due to tourists, adding that a good day is when she can connect with her passengers: “I feel at times like I’m an advocate for my city”.
    Work precarity. Because of how it allowed participants to view trends in their earnings across days, the calendar data probe led to several participants raising work precarity as another feature that negatively impacts their well-being (P3, P5-6). For them, work precarity led to continuous stress trying to meet financial goals and having to make up the hours elsewhere. They described that this worry often manifests as feeling unmotivated at the beginning of the week and then anxiously trying to make up for it at the end with nonstop work and late hours. Participants illustrated this situation by pointing to the color gradients on the calendars where Mondays and Tuesdays were light in color (lower earning) and Fridays and Saturdays were darker (higher earning). In contrast, part-time drivers P4 and P10 described being unaffected by gig work precarity because the income is not essential so they have freedom to stop and start at their convenience.
    Perceived success of driving patterns. Some associated well-being with how successfully they felt they followed driving preferences or plans. They expressed how not being able to follow these patterns impacted their day. Using the map data probe, P1 mentioned the stress he feels driving downtown versus areas that he likes, such as the airport. P3 explained if he experienced a trip he considered a "mistake", such as driving in the suburbs during rush hour, it could wreck his day. As long as she was able to maximize her efficiency through higher earnings and shorter trip durations, P8 felt comfortable both financially and physically.
    Missing data from other work platforms. Across all data probes, we observed the challenges multi-platform workers (P1, P4, P6-12) face in truly contextualizing their work patterns and understanding their well-being. When reviewing his calendar, P7 explained that he held little stock in the patterns displayed as they were partial information. But if he had data from all his platforms, he could make “a better determination of what I’m doing, and what I’m making and why that particular time and particular day.” Still, we found that even in the presence of missing data, reviewing their probes to identify hours or days they had very little to no earnings helped many recall switching to another platform due to better bonuses or tasks, and expand on their typical patterns for multi-app work (P4, P6-P12).

    6.2.2 Design Consideration For Algorithmic Work Planners or Other AI Tools.

    New predictions to make to address well-being factors. Some participants suggested new things they wanted to see predicted in the Planner. P11 suggested predicting passenger tips which directly impacts their total income, echoing the comments by other participants around how the uncertainty of passenger behaviors greatly affected their moods (P7, P9). P11 also suggested predicting optimal starting times and places to support drivers in structuring their work. This could also help address the challenges drivers face with work precarity by anchoring them with a starting point to stay on track each day. P5 also made a suggestion for how to help workers stay on track with schedules, explaining how he was interested in having upcoming purchases or expenses integrated with a calendar data probe. Based on his comment, an AI tool could assist drivers in simulating more control around their work plans by taking their historical data and future spending to provide feedback for how long it may take to meet purchase goals.
    Using data probes to understand gaps imposed by missing platform data. Ideally, tools to support workers will integrate data from all platforms they work to give a comprehensive understanding and provide useful support for future decision-making. However, this is not always possible because of platforms’ different policies for providing data to workers. In lieu of this, we observed how data probes can play a role to at least fill some gaps of understanding left behind by missing platform data. We recall how P8 used her hourly data probe to review strange patterns and recall specific multi-platform behavior: she had no Uber data for June 2022 on specific hours (e.g., 1pm, 3pm) and remembered taking advantage of a bonus Lyft ran on odd afternoon hours that month. One common problem for data practitioners is determining how to handle missing data when building models—working with workers can surface specific reasons for why missing data exists (e.g., promotions on another platform), and how to handle it as informed by the workers’ own experiences.

    6.3 Instances of Unfair Algorithmic Management

    Algorithmic management came up frequently and unprompted during sessions as participants discussed how they lack control over their strategy in the face of unfair platform algorithms. Using data probes, participants identified instances of algorithmic management and imagined how probes or other algorithmic tools could help them resist or investigate algorithmic management.

    6.3.1 Demonstrating Unfair Algorithmic Management Through Animation, Hourly, and Map Data Probes.

    Although some participants used the Planner and found their new schedule promising to explore further, others (P3, P5, P11-12) lamented that the changes they made for locations to drive in are impossible because Uber does not allow them to entirely control the locations of trip drop-offs they are offered. This lack of control is related to heavy-handed algorithmic management identified by drivers in prior work [51, 87].
    Identifying unfair algorithmic management with the animation and map data probes. The animation and map data probes were integral to participants identifying trips that deviate from their usual preferences. Though P3, who started less than a year ago, characterized these deviations as made "probably by accident", others discussed these trips as a result of manipulative algorithmic management (P2, P5-P6, P10-11). For them, their probes demonstrated how platforms reduce driver autonomy by assigning them rides that go against their preferences and well-being. Watching his animation, P10 expressed frustration over how platforms seemingly intentionally send him rides in the opposite direction of where he wants to go at inopportune times: "Yeah, I remember that ride…I wanted to end the night and get home, and then of course, it’s like that always seems to happen where it’s like Uber or Lyft just happen to know that...So then that was a struggle to get back home to find rides, you know, at 2, 3 in the morning to get back. It’s not easy."
    Some participants mentioned that Uber does allow drivers to set two destination filters a day to provide them with a limited level of control. However, quite a few shared that these destination filters “don’t work” or have stopped working in the past month (P3, P8, P10-11). P9 mentioned the possibility of Uber treating drivers differently depending on status, saying he believes the trip assignment algorithm favors him for his high Uber status (i.e., little downtime between ride requests) but that a driver of lower status probably experiences high downtime between requests. If true, this hypothesis could also apply to the disparate experiences of the destination filter amongst our participants.
    Identifying platform manipulation with the hourly and map data probes. A couple participants mentioned how the data probes reflect Uber’s manipulation tactic to keep people driving on the platform as long as possible. When asked what influenced his driving patterns as shown by his hourly data probe, P3 explained them as the product of Uber’s Quests. Whereas he would otherwise sign out during slow hours to take a break, Uber pressures him to keep driving in the attempt to complete Quests. Reviewing the map data probe, P6 talked about how he tries to stay close to places where surges often occur. He explained that from past mistakes, he has learned not to chase surges after realizing it is a bait and switch tactic Uber uses to lure drivers to a certain region and then retract the surge bonus.

    6.3.2 Design Consideration for Algorithmic Worker Tools To Combat Algorithmic Management.

    Validating hunches of algorithmic unfairness through quantitative data. Viewing the data probes, participants gave suggestions that point to how tools or data probes can support drivers against algorithmic management. Participants liked generally being able to see their strategy and movement patterns on their animation in a visceral manner. P1-2 and P6 added how it helps them quantify their hunches about their movement patterns and where trips take them. P1 and P6 specifically shared that the animations confirmed their beliefs that platforms regularly send them rides to places they do not want to go: "So that’s kind of a nice way to just see, oh, I do end up—this isn’t my imagination. I am getting sent on these trips that take me here all the time." (P6). These suggestions for using data probes to provide evidence points to an excellent direction for tools to support drivers: How can we use participant data so drivers can investigate their hunches of platform manipulation in a measurable way?
    Reverse engineering platform algorithms. P6 also shared an idea around how to help drivers predict surges through reverse engineering Uber’s surge algorithm. He observed that the passenger Uber app seems to incorporate surge pricing a few minutes before the drivers see it, thus he said a tool to help drivers effectively predict surge would need to connect to the “rider side of the app that’s showing where...the demands are typically highest”. He explained that some drivers already track this information through notebooks and screenshots, although he does not personally track it.
    These ideas of reverse engineering and platform auditing through data echo work done by [15] where workers may be able to collectively share and analyze data to identify widely-encompassing platform manipulation or unfairness.

    7 Discussion

    Our findings suggest that participants were able to use the different data probes to help them identify and communicate work patterns as well as personal or situational contexts that impact their strategies. We discuss the implications on the design of data probes as well as using data probes to design AI with stakeholders and for worker advocacy.

    7.1 How to Use and Improve Data Probes

    7.1.1 Reflections From Co-Design Sessions with Data Probes.

    One of the greatest benefits of using data probes is the ability to surface situational contexts or limitations workers face which may not be obvious from their data alone, similar to [11]’s use of diary entries and interviews to understand raw sensor data. Used without context, the data in an AI tool could lead to inappropriate recommendations. For example, the data alone could identify what drivers told us is a trend: the slow hours for work are typically 10AM-2PM. However, for P1, these hours overlap with when he stops working to begin caretaker duties for his father. If a tool were created to suggest the best hours for P1 to work, it might assume he chooses not to work these hours because of low demand and fail to treat it as a hard constraint for him.
    Using multiple data probes was also crucial, not just to prompt different insights, but because it often took repeated mentions of well-being or positionality before participants fully revealed their circumstances. This may be due to a need to establish trust during the session. For example, P1 did not select caregiver as a positionality trait, nor did he mention it during the hourly probe though it restricts his hours. Yet he disclosed this limitation during the calendar probe. Using different probes may also be necessary to dislodge worker notions of what valid well-being concerns are. P9 initially described his well-being as “physically, I’m really good, financially, I’m making good money, and psychologically, I’m perfect”. But discussing how he feels driving late night hours, it appeared that his view on well-being is influenced by his concern of platform punishment: he explained he strategically masks physical exhaustion from passengers out of fear of platform deactivation.

    7.1.2 Improvements for Data Probes.

    In order to better compare themselves with Chicago drivers, participants suggested different metrics to display and subsets of driver data to view. Though we displayed metrics such as average fare or miles per trip or per minute, participants associate different metrics with different types of strategies: some people try to maximize earnings per trip while others try to maximize total number of trips. P2 follows the latter and wanted to compare number of drop-offs with all Chicago drivers, explaining that “you can have really high average trip per min average, but maybe not as high as in terms of the number of trips.” Participants also suggested comparing themselves to similar subsets of drivers rather than all Chicago drivers including experienced full-time drivers only (P6, P8), type of car (P8), and Uber Pro status (P9). P4, based in a suburb, and P8, based in the city, suggested comparisons filtered by trips originating from the suburbs or city would be more reliable because of inherent differences: city trips are lower-earning but fast, while suburban trips are higher-earning but longer. P4 remarked, "While I’m lumped in Chicago data, I’m not Chicago. It’s a completely different animal".
    Participants also gave suggestions to improve how features are defined in data probes. P2 and P9 reminded us the Planner should consider car expenses that occur at non-weekly intervals such as car maintenance. Nearly all participants said our current day and hour inputs for the Planner are too limiting as these features vary depending on the day and week. Finally, exploring the calendar probes led participants to point out the need to align how data probes define begin/end of days and weekdays versus weekends with how Uber, and consequently drivers, define them (P1, P4, P9-10)—e.g., drivers define weekends based on weekend Quests which usually run Friday 4AM-Monday 4AM.
    We propose that these improvements for data probes also translate to AI design considerations due to similarities in characteristics they present—e.g., suggestions for redefining feature boundaries align with feature engineering. Also, while some features like expenses may be knowable without data probes, others may not necessarily surface naturally without tools like data probes. For instance, in an interview, it may not seem obvious to ask drivers (or anyone), “How do you define the hours of a day/weekend?” Misunderstandings might be perpetuated if data analysis tools are built using flawed logic (which is not normally displayed to users). Instead, in our sessions, participants explored data probes to view hourly breakdown on the days in the calendar, allowing them to correct us on how to define a deceptively simple feature.

    7.2 Using Data Probes to Elevate Stakeholder Expertise in AI Co-Design

    First, we share the realization we made in our findings around our own assumptions about drivers in order to demonstrate the importance of building tools with and not for them. Specifically, we discuss how positionality and well-being shape work preferences and constraints participants hold, summarizing our findings around the ideas and contexts they surfaced while exploring their data probes and how these can inform AI design.
    Participants’ personal contexts and situational factors we learned can inform directions and challenges for AI practitioners and UX designers to consider. One of the implications for AI design from our co-design sessions is the importance of surfacing hard constraints for AI products that support workers. As we learned in our sessions, participants’ positionalities did not always reveal themselves during our well-being and positionality activities, but became more explicit through the exploration of data probes. These hard constraints are also informative for AI designers in the consideration of tools—for example, in the design of nudges, to show recommendations that respect what a stakeholder is able to do within their confines, designers must surface and integrate these nuanced hard constraints.
    Prior work has explored different approaches for modelling and forecasting passenger travel behavior to predict where passenger demand will occur within certain time intervals [18, 55], even using external features such as weather, demographic data, and crime rates [18]. However, these studies focus on demand as the principal element of passenger behavior. Future work can be done to predict other passenger characteristics such as tipping habits, travel patterns, and propensity for conversation depending on the availability of passenger data and issues of data privacy and confidentiality. Of course, when devising and integrating these insights, care must also be taken to ensure that they do not become a means of discrimination against certain groups of passengers and exclude them from access to rides as a result.
    Treating workers as experts can also help AI practitioners and designers recognize challenges to consider that they may not have thought of before. For example, participants’ inconsistent work patterns and concerns around Uber’s constant app changes would impact the patterns exhibited in their data used for AI tools. We recall how P8 and P11 have to alter their work strategy often due to Uber’s frequent changes. We also recall how multi-platform drivers switched working between apps in inconsistent ways and had gaps in their Uber data due to working on multiple platforms. These problems that participants identified are reminiscent of the data drift problem, where there is a mismatch between the data that a model was trained on and future data. To help augment those explorations in this domain, workers can provide additional explanations to contextualize how or why their patterns are changing, based on the platform app changes or their own personal situations.
    Participants’ suggestions of additional factors for the Planner or new predictions the Planner could provide as discussed before further support the value of AI co-design with stakeholders. Based on their everyday driving, participants were able to share what has influenced their driving patterns and earnings in the past, including a driver’s neighborhood and commuting tendencies, large events that lead to high surges, and weather or other seasonal data. Participants also shared ideas for what would be more useful as predictions for them than the earnings we displayed, such as optimal start location and time, and tipping probability of passengers in different neighborhoods. These participants’ ideas reiterate the value of incorporating them as experts in AI design to promote the co-design of worker-centered tools.

    7.3 Rideshare Worker Advocacy

    Based on our findings, we propose there is potential for the use of data probes as boundary objects to work with researchers, advocates, and policymakers advancing worker rights. Collective driver goals typically consist of increased algorithmic and data transparency [37, 70], greater and guaranteed pay11[39], and increased access to benefits such as healthcare, wellness support, and education reimbursement12.
    We observed how the data probes were very effective tools for workers to communicate complex ideas and trends with us, and for us to understand their work strategies and how they are shaped. Several participants (P1, P6, P10) noted during their session how their data probes allowed them to quantify or prove beliefs regarding algorithmic management. The animation allowed them to quantify when Uber was deliberately not honoring the in-app destination filter feature that Uber claims helps drivers target locations they want to go to. The combination of how the data probes worked as boundary objects for us with participants and how the data probes helped participants identify instances of manipulative algorithmic management leads us to propose the data probes we have developed as methods that other researchers, advocates, and policymakers working with gig workers can adopt to advance worker rights.

    7.3.1 Boundary Objects for Advocacy.

    The tools designed for this study could serve as a framework for future tools designed for non-driver audiences. Policymakers have a limited understanding of platform practices, and because of Uber and Lyft’s opaque processes and limited data availability, it can be difficult for governments and regulatory agencies to get an accurate grasp of the real-world driver context. However, local governments and courts are increasingly issuing legislation and decisions requiring rideshare companies to publicize their data, including New York in 2017 [53], Seattle in 2018 [13], and San Francisco in 2019 [67]. Because of the increased availability of data, it is becoming more feasible to build similar visualizations for different cities and at a larger scale. This type of interactive tool allows a depth of understanding beyond the typical statistics seen in a white paper or fact sheet, acting as a boundary object between drivers and policymakers. This can advance drivers’ collective interests by adding more accessible public-facing data-driven tools to raise public awareness of driver issues.
    While some cities and states have established minimum wages for drivers such as New York City in 2018 [16], California in 2020 [48], and Washington in 2022 [10], the lack of legislation in other locations—such as Chicago—means that drivers are often being paid less than local minimum wages after subtracting expenses [49, 64]. Countless studies have been performed examining rideshare datasets [49, 61], however, without publicly available data, this type of independent analysis is impossible. However, there are still problems with data-driven studies as a means of activism. Two prior studies attempted to calculate average hourly earnings in Seattle and came up with two very different values ($23.25/hour vs. $9.73) due to wildly differing methodologies in calculating expenses [50]. Intense care must be taken regarding understanding the methodology and meanings of ridesharing terms before issuing declarative statements or legislation on the basis of data. Through our partnership with the Independent Drivers Guild, we hope to develop further public-facing tools to build awareness of the realities of rideshare work and advance worker rights.

    8 Limitations

    We acknowledge that our study has limitations. We faced challenges in recruiting a diverse range of drivers, possibly due to the requirement in our study for participant data; our participants came primarily from Reddit, Facebook driver forums, or a driver advocacy group, thus they may be more data literate and aware of platform issues than a standard driver. We conducted this study with Chicago drivers where we had access to publicly available data. Additionally, due to the limitations of Uber’s data, we were unable to calculate certain metrics for drivers. The duration of our sessions limited how much participants could share with us. Our sessions lasted for two hours, yet some participants wanted to discuss more after it ended. Finally, as a qualitative study, the findings we offer around worker-centered AI design should be further investigated through other research methods with a larger number of participants.

    9 Conclusion

    Using workers’ own data, we designed data probes and conducted co-design sessions with workers to surface their work patterns and how those are affected by their well-being and positionality. We discovered these data probes functioned as boundary objects to help participants surface three main insights: 1) participants described well-being trade-offs and positionality factors they face in their work, 2) they identified factors not in their data probes which impact their work and well-being such as unpredictability of passengers, and 3) they used data probes to identify instances of algorithmic management. For each finding, we also share the design considerations raised by participants using their data probes. We discuss the implications of our study on the design of data probes and the use of data probes to elevate worker expertise in AI design, including revealing individual contexts to inform the constraints of AI products, as well as the potential for data probes to be used in the future to support worker advocacy efforts.

    Acknowledgments

    This research was partially supported by the following: the National Science Foundation CNS-1952085, IIS-1939606, DGE-2125858 grants; Good Systems, a UT Austin Grand Challenge for developing responsible AI technologies13; and UT Austin’s School of Information. We are grateful to Mark Smithivas for his indispensable counsel that guided our research direction, Bianca Talabis for creating the illustrations in this paper, Charles Miele and Yihan Xi for their assistance in creating the animation data probes, the anonymous reviewers who provided invaluable feedback, and our participants for their trust in our work and thoughtful insights during sessions.

    10 Appendix

    Figure 3:
    Figure 3: One of the figures created using driver’s individual data for the pilots. This uses their trip history to depict average earnings for the time and distance they drove in trips.
    Figure 4:
    Figure 4: One of the figures created using driver’s individual data for the pilots. This uses their trip history to plot each trip for its duration and distance so drivers can identify patterns or outliers of trips they drive.
    Figure 5:
    Figure 5: Individual Data Probes
    Figure 6:
    Figure 6: City-level Data Probes

    Footnotes

    1
    These pilot participants are excluded from Table 2
    3
    https://www.unfolded.ai/ is a geospatial analytics platform
    7
    For P4 and P10, this was modified to their entire 2021-2022 data because of the sparsity of their trips as part-time drivers

    Supplementary Material

    Supplemental Materials (3544548.3581354-supplemental-materials.zip)
    MP4 File (3544548.3581354-talk-video.mp4)
    Pre-recorded Video Presentation

    References

    [1]
    2022. Tracking Chicago’s carjackings. CBS News (Apr 2022). https://www.cbsnews.com/chicago/news/carjacking/
    [2]
    Ifeoma Ajunwa. 2018. Algorithms at work: productivity monitoring applications and wearable technology as the new data-centric research agenda for employment and labor law. . Louis ULJ 63(2018), 21.
    [3]
    Ifeoma Ajunwa, Kate Crawford, and Jason Schultz. 2017. Limitless worker surveillance. Calif. L. Rev. 105(2017), 735.
    [4]
    Juan Carlos Alvarez de la Vega, Marta E Cecchinato, and John Rooksby. 2022. Design Opportunities for Freelancing Platforms: Online Freelancers’ Views on a Worker-Centred Design Fiction. In 2022 Symposium on Human-Computer Interaction for Work. 1–19.
    [5]
    Monica Anderson, Colleen McClain, Michelle Faverio, and Risa Gelles-Watnick. 2021. The state of gig work in 2021. (2021).
    [6]
    Cecilia Aragon, Shion Guha, Marina Kogan, Michael Muller, and Gina Neff. 2022. Human-Centered Data Science: An Introduction. MIT Press.
    [7]
    Amid Ayobi, Katarzyna Stawarz, Dmitri Katz, Paul Marshall, Taku Yamagata, Raúl Santos-Rodríguez, Peter Flach, and Aisling Ann O’Kane. 2021. Machine Learning Explanations as Boundary Objects: How AI Researchers Explain and Non-Experts Perceive Machine Learning. (2021).
    [8]
    Petter Bae Bae Brandtzæg, Marita Skjuve, Kim Kristoffer Kristoffer Dysthe, and Asbjørn Følstad. 2021. When the Social Becomes Non-Human: Young People’s Perception of Social Support in Chatbots. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. 1–13.
    [9]
    Oliver Bates, Carolynne Lord, Hayley Alter, Adrian Friday, and Ben Kirman. 2021. Lessons from one future of work: opportunities to flip the gig economy. IEEE Pervasive Computing 20, 4 (2021), 26–34.
    [10]
    Tina Bellon. 2022. Washington governor signs uber, lyft driver pay guarantee into law. https://www.reuters.com/business/autos-transportation/washington-gov-signs-state-wide-uber-lyft-driver-pay-guarantee-into-law-2022-04-01/
    [11]
    Sander Bogers, Joep Frens, Janne Van Kollenburg, Eva Deckers, and Caroline Hummels. 2016. Connected baby bottle: A design case study towards a framework for data-enabled design. In Proceedings of the 2016 ACM conference on designing interactive systems. 301–311.
    [12]
    Cécile Boulard-Masson, Tommaso Colombino, and Antonietta Grasso. 2018. Analysis of ‘Quantified-Self Technologies’: An Explanation of Failure. In Congress of the International Ergonomics Association. Springer, 579–583.
    [13]
    James Brasuell and Brian Minkoff. 2018. Uber and lyft reveal Seattle-area trip data. https://www.planetizen.com/news/2018/11/101400-uber-and-lyft-reveal-seattle-area-trip-data
    [14]
    Anna Brown, Alexandra Chouldechova, Emily Putnam-Hornstein, Andrew Tobin, and Rhema Vaithianathan. 2019. Toward algorithmic accountability in public services: A qualitative study of affected community perspectives on algorithmic decision-making in child welfare services. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. 1–12.
    [15]
    Dan Calacci and Alex Pentland. 2022. Bargaining with the Black-Box: Designing and Deploying Worker-Centric Tools to Audit Algorithmic Management. Proceedings of the ACM on Human-Computer Interaction 6, CSCW2(2022), 1–24.
    [16]
    Alexia Fernández Campbell. 2018. New York City passes nation’s first minimum pay rate for Uber and Lyft drivers. Vox. December 5(2018), 2018.
    [17]
    Ana Caraban, Evangelos Karapanos, Daniel Gonçalves, and Pedro Campos. 2019. 23 ways to nudge: A review of technology-mediated nudging in human-computer interaction. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. 1–15.
    [18]
    Long Chen, Piyushimita Vonu Thakuriah, and Konstantinos Ampountolas. 2021. Short-term prediction of demand for ride-hailing services: A deep learning approach. Journal of Big Data Analytics in Transportation 3, 2(2021), 175–195.
    [19]
    Janghee Cho, Tian Xu, Abigail Zimmermann-Niefield, and Stephen Voida. 2022. Reflection in Theory and Reflection in Practice: An Exploration of the Gaps in Reflection Support among Personal Informatics Apps. In CHI Conference on Human Factors in Computing Systems. 1–23.
    [20]
    Felicia Cordeiro, Daniel A Epstein, Edison Thomaz, Elizabeth Bales, Arvind K Jagannathan, Gregory D Abowd, and James Fogarty. 2015. Barriers and negative nudges: Exploring challenges in food journaling. In Proceedings of the 33rd annual ACM conference on human factors in computing systems. 1159–1162.
    [21]
    Sasha Costanza-Chock. 2020. Design justice: Community-led practices to build the worlds we need. The MIT Press.
    [22]
    Jackie Davalos and Drake Bennett. 2022. Gamification in the Gig Economy: Who Really Wins?https://www.bloomberg.com/news/features/2022-05-27/how-uber-and-lyft-gamify-the-gig-economy?leadSource=uverify%20wall
    [23]
    Catherine D’ignazio and Lauren F Klein. 2020. Data feminism. MIT press.
    [24]
    Graham Dove, Kim Halskov, Jodi Forlizzi, and John Zimmerman. 2017. UX design innovation: Challenges for working with machine learning as a design material. In Proceedings of the 2017 chi conference on human factors in computing systems. 278–288.
    [25]
    Michael Dunn. 2020. Making gigs work: digital platforms, job quality and worker motivations. New Technology, Work and Employment 35, 2 (2020), 232–249.
    [26]
    Daniel A Epstein, Monica Caraway, Chuck Johnston, An Ping, James Fogarty, and Sean A Munson. 2016. Beyond abandonment to next steps: understanding and designing for life after personal informatics tool use. In Proceedings of the 2016 CHI conference on human factors in computing systems. 1109–1113.
    [27]
    Gabe Ets-Hokin. 2022. Mileage tracker review - the best free apps to Track Your Miles [updated for 2022]. https://therideshareguy.com/day-5-what-are-the-best-apps-to-track-your-mileage/
    [28]
    Bill Gaver, Tony Dunne, and Elena Pacenti. 1999. Design: cultural probes. interactions 6, 1 (1999), 21–29.
    [29]
    Rebecca Gulotta, Aisling Kelliher, and Jodi Forlizzi. 2017. Digital systems and the experience of legacy. In Proceedings of the 2017 Conference on Designing Interactive Systems. 663–674.
    [30]
    Eloanni Hickok, Ann Skeet, Christina Colclough, Jonnie Penn, and Richard Zuroff. 2022. Framework for Promoting Workforce Well-being in the AI-Integrated Workplace. (2022).
    [31]
    Lars Erik Holmquist. 2017. Intelligence on tap: artificial intelligence as a new design material. interactions 24, 4 (2017), 28–33.
    [32]
    Kenneth Holstein, Erik Harpstead, Rebecca Gulotta, and Jodi Forlizzi. 2020. Replay enactments: Exploring possible futures through historical data. In Proceedings of the 2020 ACM Designing Interactive Systems Conference. 1607–1618.
    [33]
    Kenneth Holstein, Bruce M McLaren, and Vincent Aleven. 2019. Co-designing a real-time classroom orchestration tool to support teacher–AI complementarity. Journal of Learning Analytics 6, 2 (2019).
    [34]
    Mohammad Hossein Jarrahi, Gemma Newlands, Min Kyung Lee, Christine T Wolf, Eliscia Kinder, and Will Sutherland. 2021. Algorithmic management in a work context. Big Data & Society 8, 2 (2021), 20539517211020332.
    [35]
    Darla Benton Kearney. 2022. 4.2: Positionality and Intersectionality. https://ecampusontario.pressbooks.pub/universaldesign/chapter/positionality-intersectionality/#: :text=Positionality%20refers%20to%20the%20personal,dynamics%20of%20power%20and%20privilege.
    [36]
    Daniel S Kermany, Michael Goldbaum, Wenjia Cai, Carolina CS Valentim, Huiying Liang, Sally L Baxter, Alex McKeown, Ge Yang, Xiaokang Wu, Fangbing Yan, 2018. Identifying medical diagnoses and treatable diseases by image-based deep learning. Cell 172, 5 (2018), 1122–1131.
    [37]
    Vera Khovanskaya, Lynn Dombrowski, Jeffrey Rzeszotarski, and Phoebe Sengers. 2019. The Tools of Management: Adapting Historical Union Tactics to Platform-Mediated Labor. Proceedings of the ACM on Human-Computer Interaction 3, CSCW(2019), 1–22.
    [38]
    John Kildea, John Battista, Briana Cabral, Laurie Hendren, David Herrera, Tarek Hijal, Ackeem Joseph, 2019. Design and development of a person-centered patient portal using participatory stakeholder co-design. Journal of medical Internet research 21, 2 (2019), e11371.
    [39]
    Sage Lazzaro. 2016. Uber Drivers Plan Boycott After Fare Cuts Slash Their Earnings to Below Minimum Wage. Observer 19(2016), 2016.
    [40]
    Min Kyung Lee, Daniel Kusbit, Anson Kahng, Ji Tae Kim, Xinran Yuan, Allissa Chan, Daniel See, Ritesh Noothigattu, Siheon Lee, Alexandros Psomas, 2019. WeBuildAI: Participatory framework for algorithmic governance. Proceedings of the ACM on Human-Computer Interaction 3, CSCW(2019), 1–35.
    [41]
    Min Kyung Lee, Daniel Kusbit, Evan Metsky, and Laura Dabbish. 2015. Working with machines: The impact of algorithmic and data-driven management on human workers. In Proceedings of the 33rd annual ACM conference on human factors in computing systems. 1603–1612.
    [42]
    Min Kyung Lee, Ishan Nigam, Angie Zhang, Joel Afriyie, Zhizhen Qin, and Sicun Gao. 2021. Participatory algorithmic management: Elicitation methods for worker well-being models. In Proceedings of the 2021 AAAI/ACM Conference on AI, Ethics, and Society. 715–726.
    [43]
    Yi-Chieh Lee, Naomi Yamashita, and Yun Huang. 2020. Designing a chatbot as a mediator for promoting deep self-disclosure to a real mental health professional. Proceedings of the ACM on Human-Computer Interaction 4, CSCW1(2020), 1–27.
    [44]
    Ian Li, Anind Dey, and Jodi Forlizzi. 2010. A stage-based model of personal informatics systems. In Proceedings of the SIGCHI conference on human factors in computing systems. 557–566.
    [45]
    Phoebe Lin and Jessica Van Brummelen. 2021. Engaging teachers to Co-design integrated AI curriculum for K-12 classrooms. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. 1–12.
    [46]
    Daria Loi, Christine T Wolf, Jeanette L Blomberg, Raphael Arar, and Margot Brereton. 2019. Co-designing AI futures: Integrating AI ethics, social computing, and design. In Companion publication of the 2019 on designing interactive systems conference 2019 companion. 381–384.
    [47]
    Caitlin Lustig, Artie Konrad, and Jed R Brubaker. 2022. Designing for the Bittersweet: Improving Sensitive Experiences with Recommender Systems. In CHI Conference on Human Factors in Computing Systems. 1–18.
    [48]
    Kim Lyons. 2020. Uber and lyft roll out new benefits for California drivers under Prop 22. https://www.theverge.com/2020/12/14/22174600/uber-lyft-new-benefits-california-drivers-prop-22-gig-economy
    [49]
    Frank Manzo and Robert Bruno. 2021. https://illinoisepi.files.wordpress.com/2021/01/ilepi-pmcr-on-demand-workers-sub-minimum-wages-final.pdf
    [50]
    Aarian Marshall. 2020. Seattle’s Uber and lyft drivers make $23.25 an hour-or $9.73. https://www.wired.com/story/seattles-uber-lyft-drivers-make-2325-hour-973/
    [51]
    Marieke Möhlmann and Lior Zalmanson. 2017. Hands on the wheel: Navigating algorithmic management and Uber drivers’. In Autonomy’, in proceedings of the international conference on information systems (ICIS), Seoul South Korea. 10–13.
    [52]
    Jimmy Moore, Pascal Goffin, Jason Wiese, and Miriah Meyer. 2021. Exploring the Personal Informatics Analysis Gap:“There’s a Lot of Bacon”. IEEE Transactions on Visualization and Computer Graphics 28, 1(2021), 96–106.
    [53]
    DZ Morris. 2017. New York City Says Uber Must Share Ride Data. Fortune.
    [54]
    Sean Munson. 2012. Mindfulness, reflection, and persuasion in personal informatics. (2012).
    [55]
    Kun Niu, Cheng Cheng, Jielin Chang, Huiyang Zhang, and Tong Zhou. 2018. Real-time taxi-passenger prediction with L-CNN. IEEE Transactions on Vehicular Technology 68, 5 (2018), 4122–4129.
    [56]
    Lesley-Ann Noel and Marcelo Paiva. 2021. Learning to recognize exclusion. Journal of Usability Studies 16, 2 (2021), 63–72.
    [57]
    City of Chicago. 2022. Violence and victimization trends. https://www.chicago.gov/city/en/sites/vrd/home/violence-victimization.html
    [58]
    Fatih Kursat Ozenc, Miso Kim, John Zimmerman, Stephen Oney, and Brad Myers. 2010. How to support designers in getting hold of the immaterial material of software. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. 2513–2522.
    [59]
    Menghai Pan, Yanhua Li, Xun Zhou, Zhenming Liu, Rui Song, Hui Lu, and Jun Luo. 2019. Dissecting the learning curve of taxi drivers: A data-driven approach. In Proceedings of the 2019 SIAM International Conference on Data Mining. SIAM, 783–791.
    [60]
    Hyanghee Park, Daehwan Ahn, Kartik Hosanagar, and Joonhwan Lee. 2022. Designing Fair AI in Human Resource Management: Understanding Tensions Surrounding Algorithmic Evaluation and Envisioning Stakeholder-Centered Solutions. In CHI Conference on Human Factors in Computing Systems. 1–22.
    [61]
    James A Parrott and Michael Reich. 2018. An earnings standard for new york city’s app-based drivers. New York: The New School: Center for New York City Affairs (2018).
    [62]
    Michael Quinn Patton. 2014. Qualitative research & evaluation methods: Integrating theory and practice. Sage publications.
    [63]
    Kyrill Potapov and Paul Marshall. 2020. LifeMosaic: Co-design of a personal informatics tool for youth. In Proceedings of the interaction design and children conference. 519–531.
    [64]
    David Roeder. 2022. Contractor status keeps Rideshare Drivers’ pay below minimum wage, report says. https://chicago.suntimes.com/2022/3/14/22973357/contractor-employee-rideshare-drivers-transportation-lyft-uber-pay
    [65]
    John Rooksby, Mattias Rost, Alistair Morrison, and Matthew Chalmers. 2014. Personal tracking as lived informatics. In Proceedings of the SIGCHI conference on human factors in computing systems. 1163–1172.
    [66]
    Alex Rosenblat and Luke Stark. 2016. Algorithmic labor and information asymmetries: A case study of Uber’s drivers. International journal of communication 10 (2016), 27.
    [67]
    Carolyn Said. 2019. Uber Must Share Data with San Francisco, Appeals Court Rules. San Francisco Chronicle(2019).
    [68]
    Rachel Sandler. 2020. Microsoft’s New ‘Productivity Score’ Lets Your Boss Monitor How Often You Use Email and Attend Video Meetings.
    [69]
    Ben Shneiderman. 2022. Human-Centered AI. Oxford University Press.
    [70]
    Faiz Siddiqui. 2019. Uber and Lyft drivers strike for pay transparency—after algorithms made it harder to understand. Washington Post (2019).
    [71]
    Aaron Smith. 2016. Gig work, online selling and home sharing. (2016).
    [72]
    Logan Stapleton, Min Hun Lee, Diana Qing, Marya Wright, Alexandra Chouldechova, Ken Holstein, Zhiwei Steven Wu, and Haiyi Zhu. 2022. Imagining new futures beyond predictive systems in child welfare: A qualitative study with impacted stakeholders. In 2022 ACM Conference on Fairness, Accountability, and Transparency. 1162–1177.
    [73]
    Hariharan Subramonyam, Colleen Seifert, and Eytan Adar. 2021. Towards a process model for co-creating AI experiences. In Designing Interactive Systems Conference 2021. 1529–1543.
    [74]
    Krishnan Vasudevan and Ngai Keung Chan. 2022. Gamification and work games: Examining consent and resistance among Uber drivers. new media & society 24, 4 (2022), 866–886.
    [75]
    Jayne Wallace, John McCarthy, Peter C Wright, and Patrick Olivier. 2013. Making design probes work. In Proceedings of the SIGCHI Conference on Human Factors in Computing Systems. 3441–3450.
    [76]
    Greg Walsh and Jennifer Golbeck. 2014. StepCity: a preliminary investigation of a personal informatics-based social game on behavior change. In CHI’14 Extended Abstracts on Human Factors in Computing Systems. 2371–2376.
    [77]
    Christine T Wolf, Haiyi Zhu, Julia Bullard, Min Kyung Lee, and Jed R Brubaker. 2018. The changing contours of" participation" in data-driven, algorithmic ecosystems: Challenges, tactics, and an agenda. In Companion of the 2018 ACM Conference on Computer Supported Cooperative Work and Social Computing. 377–384.
    [78]
    Wei Xu. 2019. Toward human-centered AI: a perspective from human-computer interaction. interactions 26, 4 (2019), 42–46.
    [79]
    Chiwei Yan, Helin Zhu, Nikita Korolko, and Dawn Woodard. 2020. Dynamic pricing and matching in ride-hailing platforms. Naval Research Logistics (NRL) 67, 8 (2020), 705–724.
    [80]
    Qian Yang. 2018. Machine learning as a UX design material: how can we imagine beyond automation, recommenders, and reminders?. In AAAI Spring Symposia.
    [81]
    Qian Yang, Justin Cranshaw, Saleema Amershi, Shamsi T Iqbal, and Jaime Teevan. 2019. Sketching nlp: A case study of exploring the right things to design with language intelligence. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. 1–12.
    [82]
    Qian Yang, Alex Scuito, John Zimmerman, Jodi Forlizzi, and Aaron Steinfeld. 2018. Investigating how experienced UX designers effectively work with machine learning. In Proceedings of the 2018 designing interactive systems conference. 585–596.
    [83]
    Qian Yang, Aaron Steinfeld, Carolyn Rosé, and John Zimmerman. 2020. Re-examining whether, why, and how human-AI interaction is uniquely difficult to design. In Proceedings of the 2020 chi conference on human factors in computing systems. 1–13.
    [84]
    Yaxing Yao, Justin Reed Basdeo, Smirity Kaushik, and Yang Wang. 2019. Defending my castle: A co-design study of privacy mechanisms for smart homes. In Proceedings of the 2019 chi conference on human factors in computing systems. 1–12.
    [85]
    Nur Yildirim, Alex Kass, Teresa Tung, Connor Upton, Donnacha Costello, Robert Giusti, Sinem Lacin, Sara Lovic, James M O’Neill, Rudi O’Reilly Meehan, 2022. How Experienced Designers of Enterprise Applications Engage AI as a Design Material. In CHI Conference on Human Factors in Computing Systems. 1–13.
    [86]
    Chuang-Wen You, Chien Wen Yuan, Nanyi Bi, Min-Wei Hung, Po-Chun Huang, and Hao-Chuan Wang. 2021. Go Gig or Go Home: Enabling Social Sensing to Share Personal Data with Intimate Partner for the Health and Wellbeing of Long-Hour workers. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. 1–16.
    [87]
    Angie Zhang, Alexander Boltz, Chun Wei Wang, and Min Kyung Lee. 2022. Algorithmic management reimagined for workers and by workers: Centering worker well-being in gig work. In CHI Conference on Human Factors in Computing Systems. 1–20.
    [88]
    Angie Zhang, Olympia Walker, Kaci Nguyen, Jiajun Dai, Anqing Chen, and Min Kyung Lee. 2023. Deliberating with AI: Improving Decision-Making for the Future through Participatory AI Design and Stakeholder Deliberation. Proceedings of the ACM on Human-Computer Interaction 7, CSCW1(2023), 1–33.
    [89]
    Haiyi Zhu, Bowen Yu, Aaron Halfaker, and Loren Terveen. 2018. Value-sensitive algorithm design: Method, case study, and lessons. Proceedings of the ACM on human-computer interaction 2, CSCW(2018), 1–23.

    Cited By

    View all
    • (2024)Advancing HCI and Design Methods to Empower Gig WorkersCompanion Publication of the 2024 ACM Designing Interactive Systems Conference10.1145/3656156.3665123(37-40)Online publication date: 1-Jul-2024
    • (2024)The AI-DEC: A Card-based Design Method for User-centered AI ExplanationsProceedings of the 2024 ACM Designing Interactive Systems Conference10.1145/3643834.3661576(1010-1028)Online publication date: 1-Jul-2024
    • (2024)"My Sense of Morality Leads to My Suffering, Battling, and Arguing": The Role of Platform Designers in (Un)Deciding Gig Worker IssuesProceedings of the 2024 ACM Designing Interactive Systems Conference10.1145/3643834.3660713(3501-3514)Online publication date: 1-Jul-2024
    • Show More Cited By

    Index Terms

    1. Stakeholder-Centered AI Design: Co-Designing Worker Tools with Gig Workers through Data Probes

      Recommendations

      Comments

      Information & Contributors

      Information

      Published In

      cover image ACM Conferences
      CHI '23: Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems
      April 2023
      14911 pages
      ISBN:9781450394215
      DOI:10.1145/3544548
      Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components of this work owned by others than the author(s) must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to redistribute to lists, requires prior specific permission and/or a fee. Request permissions from [email protected].

      Sponsors

      Publisher

      Association for Computing Machinery

      New York, NY, United States

      Publication History

      Published: 19 April 2023

      Permissions

      Request permissions for this article.

      Check for updates

      Author Tags

      1. AI Design
      2. Co-Design
      3. Data Probes
      4. Gig Work
      5. Worker Well-Being
      6. Worker-Centered Worker Tools

      Qualifiers

      • Research-article
      • Research
      • Refereed limited

      Funding Sources

      Conference

      CHI '23
      Sponsor:

      Acceptance Rates

      Overall Acceptance Rate 6,199 of 26,314 submissions, 24%

      Upcoming Conference

      CHI PLAY '24
      The Annual Symposium on Computer-Human Interaction in Play
      October 14 - 17, 2024
      Tampere , Finland

      Contributors

      Other Metrics

      Bibliometrics & Citations

      Bibliometrics

      Article Metrics

      • Downloads (Last 12 months)1,654
      • Downloads (Last 6 weeks)230

      Other Metrics

      Citations

      Cited By

      View all
      • (2024)Advancing HCI and Design Methods to Empower Gig WorkersCompanion Publication of the 2024 ACM Designing Interactive Systems Conference10.1145/3656156.3665123(37-40)Online publication date: 1-Jul-2024
      • (2024)The AI-DEC: A Card-based Design Method for User-centered AI ExplanationsProceedings of the 2024 ACM Designing Interactive Systems Conference10.1145/3643834.3661576(1010-1028)Online publication date: 1-Jul-2024
      • (2024)"My Sense of Morality Leads to My Suffering, Battling, and Arguing": The Role of Platform Designers in (Un)Deciding Gig Worker IssuesProceedings of the 2024 ACM Designing Interactive Systems Conference10.1145/3643834.3660713(3501-3514)Online publication date: 1-Jul-2024
      • (2024)Design, Development, and Deployment of Context-Adaptive AI Systems for Enhanced User AdoptionExtended Abstracts of the 2024 CHI Conference on Human Factors in Computing Systems10.1145/3613905.3638195(1-5)Online publication date: 11-May-2024
      • (2024)Sensible and Sensitive AI for Worker Wellbeing: Factors that Inform Adoption and Resistance for Information WorkersProceedings of the CHI Conference on Human Factors in Computing Systems10.1145/3613904.3642716(1-30)Online publication date: 11-May-2024
      • (2024)Designing Gig Worker Sousveillance ToolsProceedings of the CHI Conference on Human Factors in Computing Systems10.1145/3613904.3642614(1-19)Online publication date: 11-May-2024
      • (2024)"At the end of the day, I am accountable": Gig Workers' Self-Tracking for Multi-Dimensional Accountability ManagementProceedings of the CHI Conference on Human Factors in Computing Systems10.1145/3613904.3642151(1-20)Online publication date: 11-May-2024
      • (2024)Data Probes as Boundary Objects for Technology Policy Design: Demystifying Technology for Policymakers and Aligning Stakeholder Objectives in Rideshare Gig WorkProceedings of the CHI Conference on Human Factors in Computing Systems10.1145/3613904.3642000(1-21)Online publication date: 11-May-2024
      • (2024)Not Just A Dot on The Map: Food Delivery Workers as InfrastructureProceedings of the CHI Conference on Human Factors in Computing Systems10.1145/3613904.3641918(1-15)Online publication date: 11-May-2024
      • (2023)Designing Individualized Policy and Technology Interventions to Improve Gig Work ConditionsProceedings of the 2nd Annual Meeting of the Symposium on Human-Computer Interaction for Work10.1145/3596671.3598576(1-9)Online publication date: 13-Jun-2023
      • Show More Cited By

      View Options

      View options

      PDF

      View or Download as a PDF file.

      PDF

      eReader

      View online with eReader.

      eReader

      HTML Format

      View this article in HTML Format.

      HTML Format

      Get Access

      Login options

      Media

      Figures

      Other

      Tables

      Share

      Share

      Share this Publication link

      Share on social media