1 Introduction
The advancement of AI has enabled a spectrum of innovative products and functions. From self-driving vehicles to mental health chatbots [
8,
43], monitoring bank fraud to determining medical diagnoses [
36]—when designed well, AI can be an assistive tool for humans in navigating their tasks.
In the workplace, AI has increasingly automated repetitive, manual tasks and improved worker productivity and efficiency. Here, AI often manifests as algorithmic management—algorithms that take on managerial functions to oversee, assign tasks, and evaluate workers [
34,
41]. For example, workplace tools such as Microsoft Suites or productivity apps include monitoring features so employers can track worker efficiency and throughput [
3,
68]. Some workplaces may even continue to track and collect data on employees through these applications when they are off the clock [
2]. These practices of algorithmic management are especially prevalent in gig work where data about workers is collected in droves and overbearing algorithms dictate worker tasks [
66].
Though vast amounts of data are collected through these applications on behalf of companies without consideration of the users, exploring one’s own data can be incredibly informative and empowering. Personal informatics (PI) research around how individuals desire to track, reflect on, and draw insights from their own data has shown how this can assist people in understanding their patterns [
65] and even initiate behavior change [
76]. Additionally, exploring
and reflecting on their real data can support users in imagining future uses of their data, surfacing ideas that designers or practitioners might not otherwise come up with on their own [
32].
With the ubiquity of digital data, PI researchers have explored how individuals review and reflect on their own data collected by social media platforms [
29,
47].
Researchers often use design probes, objects intended to help engage users in exploring open-ended questions [75]. Though often physical objects [28], design probes may also take form as digital artifacts and even incorporate data as a way to encourage users in investigating questions and design. For example, by presenting data to users through design probes, Gulotta et al. [
29] demonstrate how probes can help individuals
both surface insights about patterns in their data and inform designers on how to design digital systems.
Similarly, Subramonyam et al. [73] demonstrated that these design probes using data, which they call "data probes", can also solidify AI system boundaries:
they show how designers and engineers surfaced
concrete use cases and limitations
when considering user data
while designing AI. These examples show promising direction toward the use of data for not only supporting individuals in understanding their own data but also advancing AI design.
However, apart from [29], most existing work has explored how data probes can assist stakeholders in understanding their own data or practitioners in designing AI as opposed to how they might support stakeholders in co-designing AI.We explore this question—how to design AI with stakeholders to center their well-being and personal contexts—in the context of gig workers. Gig work platforms, where 16% of Americans were estimated to work in 2021 [
5],
is a fitting domain for us to explore our question due to the intense datafication of gig workers. With their work managed through mobile applications, a litany of data is collected about workers, from their hours worked and earnings, to more granular information such as their location even when they are not working. This presents a unique opportunity to explore how to co-design AI tools with workers by helping them harness their own work data through
the exploration of data probes.
We define data probes as digital objects which use historical data to help participants recall past or current patterns and events as they complete co-design activities.We design five data probes using workers’ own data—four visualizations and one AI work planning prototype—that workers interact with to surface their work patterns, well-being concerns, and positionality. During individual co-design sessions with 12 rideshare drivers, drivers reflect on well-being and positionality before interacting with the data probes created from their own work data as well as city-level aggregate data. For each probe, we introduce drivers to the metrics presented in the visualization and ask questions to encourage their exploration. We discover drivers use data probes as boundary objects to describe their work patterns and contexts, suggesting implications for AI design rooted in these contexts. Using probes, drivers 1) shared the well-being trade-offs and positionality factors they face in their work, 2) identified factors beyond those represented in their data probes which impact their work and well-being, and 3) identified instances of algorithmic management in their data. For each finding, we also share corresponding AI design considerations such as the creation of assistive AI to track and reflect on a driver’s well-being goals. Finally, we discuss the implications of our study on the design of data probes and the use of data probes for elevating worker expertise in AI design, including revealing individual contexts to inform AI design, as well as the potential for worker advocacy.
3 Background: Partnership with Independent Drivers Guild
In the Summer of 2021, we began working with the co-founder of Independent Drivers Guild of Illinois in the U.S. to better understand the well-being challenges and algorithmic management faced by drivers for our then-ongoing study design. Shortly after, we brought him on in a compensated position as a community advisor to periodically seek his feedback about whether our research direction aligned with what would be most useful for drivers. In Fall 2021, following our first study, we presented the results to our community advisor to get his opinion about our ideas for follow-up research directions as well as discuss his ideas given his direct driver advocacy work. He was particularly interested in the data-driven ideas of drivers from our study, suggesting the potential for using data to assist drivers in comparing their individual metrics with city-level drivers. For example, he conceptualized an AI tool that could provide a weekly plan to the driver, customized to an individual’s driving habits for where and when they should work given their goals and constraints. He introduced us to Chicago’s public dataset of rideshare data and encouraged us to explore ways this could be used to assist drivers.
This was a key factor in guiding our research direction toward how to design AI with gig workers through
data probes created from individual data and city-level data. We began by using public city data to create data visuals on Tableau about rideshare earnings. We then reached out to our community advisor in Winter 2021 to gain feedback around these visuals and his thoughts on which ones drivers may find useful to explore, and ultimately moved forward with the visual that he felt was the most beneficial—a preliminary
work planning tool to predict weekly driving earnings. We also began exploring
drivers’ individual data in order to see how this data could be used to help drivers surface ideas from their work to design AI. We learned the Uber platform allows drivers to request their data, and we reached out to two previous study participants in Spring 2022 who graciously agreed to provide their driving data in a compensated pilot exploring individual data.
1From these pilots, we gained a better sense of how the drivers might benefit from exploring their data using data probes—most prominently, we noted that both drivers naturally used the visualizations they saw (see Appendix for Figures
3 and
4) to reflect and narrate specific events they recalled (e.g., a trip that went to Milwaukee), temporal patterns they had internalized (e.g., bar closing times), and well-being or personal situations that affected their driving patterns (e.g., a family member passing away from COVID). Additionally, one of the drivers used the data to explain specific ways he wanted to review his data to investigate whether certain patterns he followed led to successful earnings or not. For example, he explained he anticipated a drop in summer ridership
volume and devised a tactic to try to increase his tips and subsequently his earnings per trip. He explained that reviewing his data for this time period could help him determine whether that idea had a tangible effect: "It’d be nice to know if, if I’m—if that extra effort is having an impact."
These interactions inspired us to revise the data visualizations to their current version in our study, where we focus on
using data probes to support drivers’ reflection and storytelling and surface specific contexts of their driving well-being and strategy formulation to inform AI design. In section
4, we describe our design goals, principles, and challenges that we address in formulating our session activities and data probes
in order to reveal sensitizing concepts for practitioners designing AI.
4 Designing Data Probes
We were drawn to how detailed and reflective our pilot participants’ insights were about their data, from the specific rides they recalled to how their personal contexts shaped their patterns. We wondered how the contexts drivers surface from exploring their data might inform AI design or AI-based tools for drivers.
Designing AI can refer to different design phases—such as design space exploration which Zhang et al. [87] pursued through scenario-based co-design sessions with drivers to explore algorithmic management interventions—or it may refer to surfacing specific components for AI tools such as data types, feature engineering, prediction and optimization goals, and constraints. One set of ideas drivers surfaced in [87] centered around using their own data to discover insights and support their well-being. Building off of this, our pilot studies, and past work around AI design with data probes (Section
2.1), our goal became to co-design AI—explore components of AI tools such as feature engineering and constraints—with drivers through data probes—visualizations created using their own data. To design effective data probes that drivers feel empowered to use, regardless of experience in data analysis, we identified design principles to adhere to.
4.1 Supporting Reflection and Action
For our first principle, we wanted to ensure the data probes supported drivers’ ease of understanding when reflecting on their data and identifying subsequent actions they can take [
44]. Cognizant of how data analysis can be daunting to participants without analysis experience [
52], we wanted to design data probes to reflect drivers’ data in forms that are familiar to them, ideally, formats that can remind them of everyday representations.
This led us to design one data probe that displays driver data on a calendar. We chose this based on how drivers in our previous study often discussed reviewing earnings at the end of the day or week. Additionally, given the ubiquity of calendars, we hoped a calendar data probe could support drivers in reflecting on their personal calendar of activities and discuss how life events or contexts influenced their driving.
Based on discussions with our community advisor and the fact that maps are the most common artifact drivers interact with, we created two more data probes: 1) an interactive map of Chicago with earnings trends by neighborhood, and 2) an animation displaying a day of the drivers’ trips. Drivers we spoke to often referenced the maps within rideshare apps when describing the information they use for decision-making. However, these maps only display immediate demand of certain regions rather than insights about a driver’s personal patterns. Thus, we created an interactive map to support drivers in generating location-specific insights. Similarly, we presented drivers with personalized animations that trace their point-by-point movements to make their data more tangible for reflections.
We included the work planning tool, Planner,
as a data probe as our community advisor believed it could be beneficial for drivers
and because it presents a use case that can help drivers anchor their ideas for AI components such as predictions or constraints to consider.
Drivers can input driving parameters such as times and places they work, and weekly expenses to obtain weekly predictions (e.g., total earnings, trips, and miles in a week).
This is displayed as a table and text summary to support individual preferences of reading and reflecting on data.
See Figure 2 for the Planner’s full input and output options and Section 5.2.2 for how the Planner works.For our final data probe representing drivers’ earnings based on trip pick-up hour, we cycled through many iterations, finally landing on a bar chart due to its simplicity as a visual compared to the other options.
To further support drivers in reflection, we incorporated heat maps into the calendar and map
data probes. The colors allow participants to easily identify profitable days or neighborhoods. We also designed questions to begin task-based and then become open-ended, as opposed to asking open-ended questions like "what are your thoughts?" which can be intimidating given the data they are presented with.
These data probes are summarized in Table 1 as well as Figure 1 which situates each probe in the co-design session. We describe the data populating these probes in Section 5.2.4.2 Centering Context
Our second principle centers on individual contexts that drivers hold which influence their work preferences, patterns, and limitations. Having the drivers discuss their strategies and patterns within the contexts that bind their work may make AI capabilities more certain and outputs more attuned to what drivers need. Our approach is largely informed by suggestions from D’ignazio and Klein [
23]. We use visualizations
for our data probes to make patterns visible, followed by probing questions about the data to uncover the context behind the presented patterns.
In order to elevate driver context, we incorporate well-being and positionality considerations prior to exploring data probes and then throughout questions as they explore data probes. To encourage drivers to consider the personal contexts that influence their work, we introduced participants to the concepts of well-being and positionality at the beginning of each session. For this study, we focus on three types of well-being: physical, psychological, and financial, building from Hickok et al. [
30]’s definitions of the terms [
42]. We also have drivers complete a positionality activity—"the personal values, views, and location in time and space that influence how one engages with and understands the world" [
35]—by having them consider what factors advantage and disadvantage them as a driver. Some of these factors are derived from existing positionality wheels (e.g., "Level of Education, Age, Race & Ethnicity") [
56] while others we added specific to gig work ("Own vs. Rent My Car", "Single vs. Multi-Platform Worker"). Later, as drivers interact with data probes, we ask how their work patterns affect their well-being and remind them of the positionality factors they identified to see if/how any of their factors
impact their choice to work or not at a certain time or place.
5 Methods
5.1 Study Participants
5.1.1 Participant Criteria.
We recruited active Chicago Uber drivers. We required participants to be Uber drivers primarily because Uber is the only rideshare platform we identified as allowing drivers to easily request and download their data. The participants had to be actively driving because one of the data probes we showed was based on their geodata which Uber only provides for the past 30 days. Finally, the participants had to be working in Chicago because of available city-level rideshare data with earnings.
We used multiple recruitment methods: Facebook Ads, posts in Facebook Groups, posts on subreddits, Reddit ads, word-of-mouth, and emails through our partner channels. We distributed a pre-screening survey. In addition to basic demographic data, we asked respondents to self-classify the permanence of their gig work careers [
25], whether they had another job, and how much they relied upon their rideshare income ("nice to have, but not essential to my budget" or "essential for meeting basic needs" [
71]). Our research team reached out to respondents, detailing instructions on how to download their Uber data and share with the research team.
5.1.2 Participants.
We had n=12 total participants. Their average age was 43.9 years (SD = 2.57). 7 of our 12 (58.3%) participants identified as White, 4 (33.3%) as Asian, and 1 as Hispanic (8.3%). 10 of our participants (83.3%) identified as male, with the remaining 2 identifying as female (16.7%). 11 of our participants (91.7%) have been working for rideshare platforms for 2 years or more. 10 participants (83.3%) also reported working at least 30 hours a week on gig platforms, with the same 10 reporting that their gig work income was "essential for meeting basic needs"; 2 participants reported that their income was "nice to have, but not essential to my budget".
5.2 Data Probe Creation
We first describe the data sources we used and the four visualization data probes. We then explain the fifth data probe, the Planner—its mechanics and computations. Most data probes were created primarily using Tableau, an interactive data visualization platform.
5.2.1 Data Sources to Create Data Probes.
Individual Data. To create individual
data probes (see Figure
5) we utilized the driver’s own Uber data
2.
Personal Trip Animation Data Probe. We created a personalized trip animation for each driver, combining trips with location data from the past 30 days. With this file, we generated a time-lapse animation of participants’ movements overlaid on a map of Chicago using Unfolded.ai
3. For the session, we isolated the animation to one recent day of trips. The purpose of the animation was to give participants a visceral representation of their movement data as they described their general work strategy.
Individual Hourly, Calendar, and Map Data Probes. Next, participants saw their hourly data probe, a bar chart displaying the average fare per minute at each hour of the day for the month of June. Then, they reviewed their calendar displaying similar metrics, summarized for each day since June 2022. For the map data probe, participants saw two personalized maps, colored in varying shades based on the average fares per minute in each neighborhood. The first showed their common pick-up locations (as well as total pick-ups per area upon hover). The second showed their common drop-off locations. These maps were interactive: selecting a pick-up neighborhood altered the drop-off map to display only drop-off locations for trips beginning in the selected neighborhood.
City-Level Data. The data for city-level
data probes came directly from the City of Chicago
4. Available back to 2018, we limited our dataset to the most recently available month (June 2022) because of size and processing power limitations. To prepare the dataset for analysis, we used Pandas—a Python-based data manipulation tool—to classify all the pick-up and drop-off coordinates into Chicago’s official neighborhoods, as defined by the City of Chicago
5. Next, we used Pandas to add time-appropriate weather data acquired from VisualCrossing, a live weather API
6.
City-Level Hourly, Calendar, and Map Data Probes. City-level data probes were all presented side-by-side with their personal counterparts. These were virtually identical to their personal counterparts, only differing in that these versions utilized city-level driver data (see Figure
6).
For the calendar, weekly calendars were used to compare individual and city-level data because the Chicago dataset does not identify unique drivers. Drivers saw a matrix of 4 variables per day of the week: total number of trips (for the entire city’s dataset), average fare, average trip duration (with passengers in the vehicle), and average fare per minute. The map data probe displayed an interactive map of Chicago, color-coded to indicate neighborhoods with the highest average fares per minute.
5.2.2 Work Planner.
We created the Work Planner (Planner), a comprehensive simulation of predicted earnings and expenses using the aforementioned Chicago data.
Figure 2 displays the full tool details including input and output variables.To use it, drivers input driving preferences: the number of hours driven in a hypothetical week, days of the week driven, hours of the day driving, neighborhoods they choose to pick up in, outdoor temperature, and precipitation.
Weather variables were included based on past drivers sharing they believe weather may impact their earnings. Drivers also enter expenses including the price of gas, their vehicle’s mileage, their car insurance cost, and other miscellaneous expenses.
Expense calculation, something rideshare platforms currently lack, was included to give drivers a more realistic sense of their net profit. They also adjusted other variables important to the simulation including
the portion of their fare that the platform withdraws, and the percent of their time spent with passengers in their car.
The first was included to respect drivers’ personal beliefs of how much platforms take which often differs from what platforms formally claim. The second was included as drivers previously warned their hourly earnings rate includes unpaid time spent looking for passengers or driving to them. For these two variables, we set the default parameters to 25% and 55%, standard values for Chicago [
49].
Output statistics are generated by using the participant inputs to filter the city-level data to the subset of trips which match these. For example, to calculate weekly earnings, the Planner first calculates the average fare (AF, $) and average trip duration (ATD, minutes) directly from the data subset. Next, it determines the number of projected trips (PT) through the equation: where TPC refers to the percentage of time a driver spends with passengers in the car and HPW represents the hours per week the driver inputs as working. Finally, the weekly earnings (WE, $) is obtained: Similar calculations are done to obtain the other statistics. We explained to participants that the calculations were based on June 2022 data from only the Chicago area. We also explained that based on their selections for time and location of work, the tool’s output would adjust accordingly to take into consideration trips started on those days, hours, or neighborhoods. 5.3 Co-Design Procedure
Participants began with the Planner to create their typical weekly schedule. Next, we introduced them to the three types of well-being and asked whether they wanted to improve any of them in their work. Participants completed a positionality activity, where they considered 1-4 factors that advantage them and 1-4 factors that disadvantage them in their work from a set of factors often used in positionality exercises [
56] in addition to ones related to gig work. The section concluded with participants answering, "Is there anything you’d like to know or want to do to improve your work strategy?"
Participants then explored data probes created using their personal Uber data. The first probe was an animation of their Uber trips
that they watched as they described their typical driving strategy. The second probe was their hourly bar chart of the earnings they made during June 2022
7. The third probe was the calendar, from June to their most recent available data, where days were shaded based on their total earnings. Finally, we showed participants
their map data probe reflecting their pick-up and drop-off locations and fare patterns
for each neighborhood based on the past 30 days.
With each probe, we guided participants through task-based questions (e.g., "Which days does the calendar show you made the most money?"), followed by questions to understand their patterns, the factors that influence them, and the impact on their well-being. Participants could also hover over each probe to review metrics about their work (see supplementary information to view full metrics per probe in the hover).
Next, participants compared their individual data probes with corresponding ones created from Chicago drivers: hourly, calendar, and map data probes. Participants viewed their individual and city-level hourly data probes side-by-side and were asked to identify differences and similarities, and reasons for why they believed those exist. This continued for the calendar and map data probes.
Finally, participants created a new schedule using the Planner based on the insights they drew from the data probes about optimal hours, days, and locations to work. They compared the results of their original and new schedule and shared their reflections and preferences between the two. Figure
1 shows the co-design session flow. The session concluded with participants reflecting on their experience with the five data probes, thinking specifically about how the probes helped them and which visualizations they liked the most.
5.4 Analysis
All co-design sessions were conducted and recorded using Zoom, and transcribed using Otter.ai. Each session lasted two hours on average, and participants were compensated with a $90 gift card. Notes and transcriptions were reviewed by the interviewers and notetakers to produce summaries for each participant. These summaries (containing notes, memos, and reflections) were clustered based on protocol areas and data probes. The data was then analyzed following Patton [
62]’s qualitative data analysis method. Two researchers coded the data and grouped them for emerging themes. The entire research team then discussed these themes to derive the final themes reported in this paper.
5.5 Background on Uber Features
We provide a brief explanation of Uber platform features referenced by our participants to provide clarity for our findings.
Quest promotions are Uber’s incentive-based method of encouraging drivers to increase their short-term driving
8. Quests offer drivers a
limited time incentive for completing a set number of trips; e.g.,
’Complete 60 trips between Friday and Sunday for a $150 bonus’. Quests appear unevenly, with varying thresholds and rewards. Many drivers suspect that Quests unfairly favor newer or inactive drivers. However, because of the opacity of the promotions, they lack the means to definitively prove this suspicion [
87].
Surge promotions are Uber’s method of matching the supply of drivers with the demand from riders
9. Uber will temporarily create "surges" in high demand areas to incentivize drivers to relocate to those areas.
Uber Pro is Uber’s driver rewards programs that automatically enrolls all Uber drivers
10. There are four tiers of Uber Pro: blue, gold, platinum, and diamond. Driver tier is determined by a variety of factors including driver rating (the average of their last 500 ratings from passengers), cancellation rate, trip acceptance rate, and the number of ’points’ earned. Points are re-calculated every 3 months, and earned by completing trips with certain high-demand times earning more points per trip.
6 Findings
In co-design sessions, we observed that data probes acted as boundary objects for participants to effectively communicate their work patterns and contexts with our research team and suggest AI design implications rooted in these contexts. In this section, we describe three main insights surfaced by participants through data probes and corresponding (AI) design considerations. First, using data probes, participants shared how well-being trade-offs and immovable positionality factors pose constraints for how they could use the Planner—the Tableau schedule and earnings simulator we created—practically, but suggested alternative ways for an algorithmic work planning tool to center well-being within their constraints. Second, participants explored data probes to identify factors beyond those represented in their data probes that impact their work and well-being, offering new prediction ideas for AI tools that can support them. Third, participants used data probes to identify specific instances of unfair algorithmic management and suggested ways to repurpose data probes and design algorithmic tools to audit platforms.
6.1 Well-Being Trade-Offs and Positionality Constraints That Impact Worker’s Strategies
Participants used their data probes to explain how these visualizations reflect the ways they currently balance well-being in their strategy as well as how personal obligations constrain their work flexibility. It is important for worker-centered AI design to be considerate of these limitations in order to provide well-being support that is actually feasible for workers to take.
6.1.1 Surfacing Well-Being and Positionality Through Hourly, Calendar, and Map Data Probes.
Initial reflections around well-being and positionality. When we first introduced well-being and positionality, a handful of participants raised well-being concerns such as safety or exhaustion (P3, P8, P10), stress over precarity of earnings and passenger attitudes (P3, P12), and frustration over traffic or unfair treatment by platforms (P8, P11). For positionality, participants shared advantages that directly impact their earnings such as “Being a Multi-Platform Worker” or reduce their expenses such as “Owning My Car”, and disadvantages that limit their earning ability such as living far away from centrally located, high earning neighborhoods.
Articulating well-being trade-offs and positionality constraints using hourly and map data probes. Through data probes, participants surfaced or expanded on initially identified well-beings and positionalities. The hourly and map data probes were especially effective as tools for participants to discuss specific well-being trade-offs they make—such as balancing physical and mental health with financial gain—and positionality constraints they face—particularly familial obligations. These probes were likely effective given participants held safety concerns about Chicago’s high rate of carjackings and shootings (P1-3, P5, P8-12), which are highly associated with late night/early morning hours and specific neighborhoods [1, 57]. Participants explicitly identified high earning hours and neighborhoods on hourly and map data probes of city-level data, saying the probes confirm what they already “know” are profitable strategies. Then, they explained their individual data probes reflect their conscious decisions to not drive at these times or areas because of well-being concerns and positionality limitations. P2 said, "maybe it [working late night hours] could have a positive impact on my finances, rates, more earnings, but I just wouldn’t feel safe. And I would feel, you know, anxiety”, and P10 referenced how his fear of his daughter losing her father prevents him from driving in certain areas. In contrast, P4 did not hold the same concerns over his well-being, instead reflecting a high level of empathy towards the positionalities of residents of these neighborhoods. He accepted rides regardless of pick-up or drop-off locations because he did not want to see people lose access to rides simply because they lived in more "dangerous" neighborhoods. Surfacing obligations imposed by workers’ positionalities using the calendar data probe. The calendar data probe surfaced more unique perspectives around the kinds of limitations worker positionalities impose on their work patterns. As people often use calendars to record and view their daily tasks, this probe was intended to help people intuitively recall their data in the context of their responsibilities and routines. Accordingly, while exploring his calendar, P2 explained his data probe showed few working Saturdays and Sundays because he aligns his days off to be with his girlfriend. Comparing the city-level and his individual weekly calendars, P3 lamented being unable to capitalize on high-earning weekends due to caring for his infant child. P1 explored his individual monthly calendar for memorable low- or high-earning days, and expressed regret when he remembered that he had been unable to work on July 4th because of family obligations. P5, P7, P9-10, and P12 also expressed similar constraints in their weekly schedules due to commitments with family and friends.
6.1.2 Design Consideration for Centering Well-Being in AI Design.
Developing assistive AI for tracking and reflecting on well-being goals. While exploring their data, several participants commented how they try to prioritize their physical and psychological well-being with days off but are not always successful. For example, P5 told us Sundays are family days, yet his calendar showed him working half of all Sundays. P6 tries to prioritize time with family and friends as he gets older, but rather than treating it as time off, he feels he must make up for it later. P2 expressed surprise when he saw his calendar, explaining he thought he only worked 15 days in June when in reality, he worked 25 days in the month. Data probes illuminated the difference between what drivers believe they do for their well-being versus what occurs in practice. AI tools can focus on supporting drivers in the same regard. This could be through an AI work planning tool which uses a driver’s data to present statistics about recent days worked so they can see whether/how they are meeting well-being goals and how to integrate that knowledge into the next week’s plan.
AI tools could also assist drivers with well-being centered schedules by predicting when to take breaks or what past factors led to profitable days. Unsurprisingly, long-tenured drivers felt confident in their driving strategy and did not seem receptive to change. But a few mentioned an interest in using their data to influence their strategy. P6 wanted to use his data to predict profitability so he could schedule appointments or breaks on days with lower predicted earnings. P12 shared how he would like a predictive model to help him decipher what factors led to him having a good or bad day (in earnings), "As a driver, you can analyze it and be like, ’Man, I’ve really had a bad day. Why did I have such a bad day?’ Or ’I had a great day. What is it that made it such a great day?’" Similar to P6, P8 was interested in using her data to plan her day off, adding the benefit of taking time for herself: “I tend to do better with tips back after a couple days off, which means I’m just getting burned out of people.” Addressing her comment, AI tools could also use workers’ own data to enable them to track and reflect on their time off and see when they may benefit from a break.
Respecting driver contexts and knowledge. For most participants, viewing data probes did not lead to changes in their work strategies due to immutable constraints—not driving at night due to safety concerns despite the potential profit or being a caretaker limiting the time they may work. However, they gave suggestions for how to design AI tools that can assist them in decision-making for well-being. This is a reminder that a tool designed without considering worker contexts may very well violate their personal constraints or expertise and will unlikely be adopted. Instead, AI tools should take on assistive roles, leaving final decision-making to workers themselves. For example, by their own admission, drivers get burned out, thus a tool that can track and help them reflect on the effect of taking time off could be more effective in advancing worker well-being than one which prescribes to them to take a break.
Additionally, we observed that for a few participants, the session did lead to recognizing potential changes to improve their well-being. P2, though an experienced driver, described himself as being in a “transition period”: he was moving and his girlfriend recently changed her days off, so he was developing a new schedule. Though confident in the merits of his old schedule, he showed interest in using the Planner to simulate how to tweak his schedule. P1 and P11 both considered testing new schedules which estimated fewer earnings but also meant fewer or different hours in order to improve their psychological and physical well-being. Though they were not interested in change, P8 and P12 shared that reviewing the data probes and using the Planner could help beginner drivers instead of established drivers. Once again, through understanding their context, we can more appropriately determine the use and audience of tools like algorithmic work planning. Rather than catering to general users, a work planning tool could be designed to center around onboarding new drivers or experienced drivers with an existing interest in tweaking their strategy.
6.2 Identifying New Factors Not Included in Data Probe Design
Through data probe exploration, participants identified additional factors which impact well-being as well as ideas of predictions they wanted to see from AI tools. Participants also used the data probes to recognize and explain gaps in their Uber data which resulted from other platform work.
6.2.1 Identifying Additional Factors that Impact Well-Being Through All Data Probes.
Our data probes centered around participants reflecting on schedule-related factors of work (e.g., time, day, location). Encouragingly, exploring their data in these ways also allowed participants to remember factors we did not include in the data probes but that they felt were influential on well-being, including passenger characteristics, general work precarity, perceived success of driving patterns, and multi-platform work.
Passenger characteristics. Reviewing their hourly data probes, P7 and P9 felt they currently work optimal hours which do not lead to exhaustion or stress. Instead, P7 reflected that far more impactful on his well-being are interactions with passengers—negative interactions can pull his mood down: "No one wants to interact with someone who...is mistreating them or being obnoxious." Additionally, positive interactions with passengers can boost drivers’ moods. By inspecting her calendar data probe for high earnings days, P8 recalled that Saturday afternoons are often profitable due to tourists, adding that a good day is when she can connect with her passengers: “I feel at times like I’m an advocate for my city”.
Work precarity. Because of how it allowed participants to view trends in their earnings across days, the calendar data probe led to several participants raising work precarity as another feature that negatively impacts their well-being (P3, P5-6). For them, work precarity led to continuous stress trying to meet financial goals and having to make up the hours elsewhere. They described that this worry often manifests as feeling unmotivated at the beginning of the week and then anxiously trying to make up for it at the end with nonstop work and late hours. Participants illustrated this situation by pointing to the color gradients on the calendars where Mondays and Tuesdays were light in color (lower earning) and Fridays and Saturdays were darker (higher earning). In contrast, part-time drivers P4 and P10 described being unaffected by gig work precarity because the income is not essential so they have freedom to stop and start at their convenience.
Perceived success of driving patterns. Some associated well-being with how successfully they felt they followed driving preferences or plans. They expressed how not being able to follow these patterns impacted their day. Using the map data probe, P1 mentioned the stress he feels driving downtown versus areas that he likes, such as the airport. P3 explained if he experienced a trip he considered a "mistake", such as driving in the suburbs during rush hour, it could wreck his day. As long as she was able to maximize her efficiency through higher earnings and shorter trip durations, P8 felt comfortable both financially and physically.
Missing data from other work platforms. Across all data probes, we observed the challenges multi-platform workers (P1, P4, P6-12) face in truly contextualizing their work patterns and understanding their well-being. When reviewing his calendar, P7 explained that he held little stock in the patterns displayed as they were partial information. But if he had data from all his platforms, he could make “a better determination of what I’m doing, and what I’m making and why that particular time and particular day.” Still, we found that even in the presence of missing data, reviewing their probes to identify hours or days they had very little to no earnings helped many recall switching to another platform due to better bonuses or tasks, and expand on their typical patterns for multi-app work (P4, P6-P12).
6.2.2 Design Consideration For Algorithmic Work Planners or Other AI Tools.
New predictions to make to address well-being factors. Some participants suggested new things they wanted to see predicted in the Planner. P11 suggested predicting passenger tips which directly impacts their total income, echoing the comments by other participants around how the uncertainty of passenger behaviors greatly affected their moods (P7, P9). P11 also suggested predicting optimal starting times and places to support drivers in structuring their work. This could also help address the challenges drivers face with work precarity by anchoring them with a starting point to stay on track each day. P5 also made a suggestion for how to help workers stay on track with schedules, explaining how he was interested in having upcoming purchases or expenses integrated with a calendar data probe. Based on his comment, an AI tool could assist drivers in simulating more control around their work plans by taking their historical data and future spending to provide feedback for how long it may take to meet purchase goals.
Using data probes to understand gaps imposed by missing platform data. Ideally, tools to support workers will integrate data from all platforms they work to give a comprehensive understanding and provide useful support for future decision-making. However, this is not always possible because of platforms’ different policies for providing data to workers. In lieu of this, we observed how data probes can play a role to at least fill some gaps of understanding left behind by missing platform data. We recall how P8 used her hourly data probe to review strange patterns and recall specific multi-platform behavior: she had no Uber data for June 2022 on specific hours (e.g., 1pm, 3pm) and remembered taking advantage of a bonus Lyft ran on odd afternoon hours that month. One common problem for data practitioners is determining how to handle missing data when building models—working with workers can surface specific reasons for why missing data exists (e.g., promotions on another platform), and how to handle it as informed by the workers’ own experiences.
6.3 Instances of Unfair Algorithmic Management
Algorithmic management came up frequently and unprompted during sessions as participants discussed how they lack control over their strategy in the face of unfair platform algorithms. Using data probes, participants identified instances of algorithmic management and imagined how probes or other algorithmic tools could help them resist or investigate algorithmic management.
6.3.1 Demonstrating Unfair Algorithmic Management Through Animation, Hourly, and Map Data Probes.
Although some participants used the Planner and found their new schedule promising to explore further, others (P3, P5, P11-12) lamented that the changes they made for locations to drive in are impossible because Uber does not allow them to entirely control the locations of trip drop-offs they are offered. This lack of control is related to heavy-handed algorithmic management identified by drivers in prior work [51, 87]. Identifying unfair algorithmic management with the animation and map data probes. The animation and map data probes were integral to participants identifying trips that deviate from their usual preferences. Though P3, who started less than a year ago, characterized these deviations as made "probably by accident", others discussed these trips as a result of manipulative algorithmic management (P2, P5-P6, P10-11). For them, their probes demonstrated how platforms reduce driver autonomy by assigning them rides that go against their preferences and well-being. Watching his animation, P10 expressed frustration over how platforms seemingly intentionally send him rides in the opposite direction of where he wants to go at inopportune times: "Yeah, I remember that ride…I wanted to end the night and get home, and then of course, it’s like that always seems to happen where it’s like Uber or Lyft just happen to know that...So then that was a struggle to get back home to find rides, you know, at 2, 3 in the morning to get back. It’s not easy."
Some participants mentioned that Uber does allow drivers to set two destination filters a day to provide them with a limited level of control. However, quite a few shared that these destination filters “don’t work” or have stopped working in the past month (P3, P8, P10-11). P9 mentioned the possibility of Uber treating drivers differently depending on status, saying he believes the trip assignment algorithm favors him for his high Uber status (i.e., little downtime between ride requests) but that a driver of lower status probably experiences high downtime between requests. If true, this hypothesis could also apply to the disparate experiences of the destination filter amongst our participants.
Identifying platform manipulation with the hourly and map data probes. A couple participants mentioned how the data probes reflect Uber’s manipulation tactic to keep people driving on the platform as long as possible. When asked what influenced his driving patterns as shown by his hourly data probe, P3 explained them as the product of Uber’s Quests. Whereas he would otherwise sign out during slow hours to take a break, Uber pressures him to keep driving in the attempt to complete Quests. Reviewing the map data probe, P6 talked about how he tries to stay close to places where surges often occur. He explained that from past mistakes, he has learned not to chase surges after realizing it is a bait and switch tactic Uber uses to lure drivers to a certain region and then retract the surge bonus.
6.3.2 Design Consideration for Algorithmic Worker Tools To Combat Algorithmic Management.
Validating hunches of algorithmic unfairness through quantitative data. Viewing the data probes, participants gave suggestions that point to how tools or data probes can support drivers against algorithmic management. Participants liked generally being able to see their strategy and movement patterns on their animation in a visceral manner. P1-2 and P6 added how it helps them quantify their hunches about their movement patterns and where trips take them. P1 and P6 specifically shared that the animations confirmed their beliefs that platforms regularly send them rides to places they do not want to go: "So that’s kind of a nice way to just see, oh, I do end up—this isn’t my imagination. I am getting sent on these trips that take me here all the time." (P6). These suggestions for using data probes to provide evidence points to an excellent direction for tools to support drivers: How can we use participant data so drivers can investigate their hunches of platform manipulation in a measurable way?
Reverse engineering platform algorithms. P6 also shared an idea around how to help drivers predict surges through reverse engineering Uber’s surge algorithm. He observed that the passenger Uber app seems to incorporate surge pricing a few minutes before the drivers see it, thus he said a tool to help drivers effectively predict surge would need to connect to the “rider side of the app that’s showing where...the demands are typically highest”. He explained that some drivers already track this information through notebooks and screenshots, although he does not personally track it.
These ideas of reverse engineering and platform auditing through data echo work done by [15] where workers may be able to collectively share and analyze data to identify widely-encompassing platform manipulation or unfairness. 7 Discussion
Our findings suggest that participants were able to use the different data probes to help them identify and communicate work patterns as well as personal or situational contexts that impact their strategies. We discuss the implications on the design of data probes as well as using data probes to design AI with stakeholders and for worker advocacy.
7.1 How to Use and Improve Data Probes
7.1.1 Reflections From Co-Design Sessions with Data Probes.
One of the greatest benefits of using data probes is the ability to surface situational contexts or limitations workers face which may not be obvious from their data alone, similar to [11]’s use of diary entries and interviews to understand raw sensor data. Used without context, the data in an AI tool could lead to inappropriate recommendations. For example, the data alone could identify what drivers told us is a trend: the slow hours for work are typically 10AM-2PM. However, for P1, these hours overlap with when he stops working to begin caretaker duties for his father. If a tool were created to suggest the best hours for P1 to work, it might assume he chooses not to work these hours because of low demand and fail to treat it as a hard constraint for him. Using multiple data probes was also crucial, not just to prompt different insights, but because it often took repeated mentions of well-being or positionality before participants fully revealed their circumstances. This may be due to a need to establish trust during the session. For example, P1 did not select caregiver as a positionality trait, nor did he mention it during the hourly probe though it restricts his hours. Yet he disclosed this limitation during the calendar probe. Using different probes may also be necessary to dislodge worker notions of what valid well-being concerns are. P9 initially described his well-being as “physically, I’m really good, financially, I’m making good money, and psychologically, I’m perfect”. But discussing how he feels driving late night hours, it appeared that his view on well-being is influenced by his concern of platform punishment: he explained he strategically masks physical exhaustion from passengers out of fear of platform deactivation.
7.1.2 Improvements for Data Probes.
In order to better compare themselves with Chicago drivers, participants suggested different metrics to display and subsets of driver data to view. Though we displayed metrics such as average fare or miles per trip or per minute, participants associate different metrics with different types of strategies: some people try to maximize earnings per trip while others try to maximize total number of trips. P2 follows the latter and wanted to compare number of drop-offs with all Chicago drivers, explaining that “you can have really high average trip per min average, but maybe not as high as in terms of the number of trips.” Participants also suggested comparing themselves to similar subsets of drivers rather than all Chicago drivers including experienced full-time drivers only (P6, P8), type of car (P8), and Uber Pro status (P9). P4, based in a suburb, and P8, based in the city, suggested comparisons filtered by trips originating from the suburbs or city would be more reliable because of inherent differences: city trips are lower-earning but fast, while suburban trips are higher-earning but longer. P4 remarked, "While I’m lumped in Chicago data, I’m not Chicago. It’s a completely different animal".
Participants also gave suggestions to improve how features are defined in data probes. P2 and P9 reminded us the Planner should consider car expenses that occur at non-weekly intervals such as car maintenance. Nearly all participants said our current day and hour inputs for the Planner are too limiting as these features vary depending on the day and week. Finally, exploring the calendar probes led participants to point out the need to align how data probes define begin/end of days and weekdays versus weekends with how Uber, and consequently drivers, define them (P1, P4, P9-10)—e.g., drivers define weekends based on weekend Quests which usually run Friday 4AM-Monday 4AM.
We propose that these improvements for data probes also translate to AI design considerations due to similarities in characteristics they present—e.g., suggestions for redefining feature boundaries align with feature engineering. Also, while some features like expenses may be knowable without data probes, others may not necessarily surface naturally without tools like data probes. For instance, in an interview, it may not seem obvious to ask drivers (or anyone), “How do you define the hours of a day/weekend?” Misunderstandings might be perpetuated if data analysis tools are built using flawed logic (which is not normally displayed to users). Instead, in our sessions, participants explored data probes to view hourly breakdown on the days in the calendar, allowing them to correct us on how to define a deceptively simple feature.
7.2 Using Data Probes to Elevate Stakeholder Expertise in AI Co-Design
First, we share the realization we made in our findings around our own assumptions about drivers in order to demonstrate the importance of building tools with and not for them. Specifically, we discuss how positionality and well-being shape work preferences and constraints participants hold, summarizing our findings around the ideas and contexts they surfaced while exploring their data probes and how these can inform AI design.
Participants’ personal contexts and situational factors we learned can inform directions and challenges for AI practitioners and UX designers to consider. One of the implications for AI design from our co-design sessions is the importance of surfacing hard constraints for AI products that support workers. As we learned in our sessions, participants’ positionalities did not always reveal themselves during our well-being and positionality activities, but became more explicit through the exploration of data probes. These hard constraints are also informative for AI designers in the consideration of tools—for example, in the design of nudges, to show recommendations that respect what a stakeholder is able to do within their confines, designers must surface and integrate these nuanced hard constraints.
Prior work has explored different approaches for modelling and forecasting passenger travel behavior to predict where passenger demand will occur within certain time intervals [
18,
55], even using external features such as weather, demographic data, and crime rates [
18]. However, these studies focus on demand as the principal element of passenger behavior. Future work can be done to predict other passenger characteristics such as tipping habits, travel patterns, and propensity for conversation depending on the availability of passenger data and issues of data privacy and confidentiality. Of course, when devising and integrating these insights, care must also be taken to ensure that they do not become a means of discrimination against certain groups of passengers and exclude them from access to rides as a result.
Treating workers as experts can also help AI practitioners and designers recognize challenges to consider that they may not have thought of before. For example, participants’ inconsistent work patterns and concerns around Uber’s constant app changes would impact the patterns exhibited in their data used for AI tools. We recall how P8 and P11 have to alter their work strategy often due to Uber’s frequent changes. We also recall how multi-platform drivers switched working between apps in inconsistent ways and had gaps in their Uber data due to working on multiple platforms. These problems that participants identified are reminiscent of the data drift problem, where there is a mismatch between the data that a model was trained on and future data. To help augment those explorations in this domain, workers can provide additional explanations to contextualize how or why their patterns are changing, based on the platform app changes or their own personal situations.
Participants’ suggestions of additional factors for the Planner or new predictions the Planner could provide as discussed before further support the value of AI co-design with stakeholders. Based on their everyday driving, participants were able to share what has influenced their driving patterns and earnings in the past, including a driver’s neighborhood and commuting tendencies, large events that lead to high surges, and weather or other seasonal data. Participants also shared ideas for what would be more useful as predictions for them than the earnings we displayed, such as optimal start location and time, and tipping probability of passengers in different neighborhoods. These participants’ ideas reiterate the value of incorporating them as experts in AI design to promote the co-design of worker-centered tools.
7.3 Rideshare Worker Advocacy
Based on our findings, we propose there is potential
for the use of data probes as boundary objects to work with researchers, advocates, and policymakers advancing worker rights. Collective driver goals typically consist of increased algorithmic and data transparency [
37,
70], greater and guaranteed pay
11[
39], and increased access to benefits such as healthcare, wellness support, and education reimbursement
12.
We observed how the data probes were very effective tools for workers to communicate complex ideas and trends with us, and for us to understand their work strategies and how they are shaped. Several participants (P1, P6, P10) noted during their session how their data probes allowed them to quantify or prove beliefs regarding algorithmic management. The animation allowed them to quantify when Uber was deliberately not honoring the in-app destination filter feature that Uber claims helps drivers target locations they want to go to. The combination of how the data probes worked as boundary objects for us with participants and how the data probes helped participants identify instances of manipulative algorithmic management leads us to propose the data probes we have developed as methods that other researchers, advocates, and policymakers working with gig workers can adopt to advance worker rights.
7.3.1 Boundary Objects for Advocacy.
The tools designed for this study could serve as a framework for future tools designed for non-driver audiences. Policymakers have a limited understanding of platform practices, and because of Uber and Lyft’s opaque processes and limited data availability, it can be difficult for governments and regulatory agencies to get an accurate grasp of the real-world driver context. However, local governments and courts are increasingly issuing legislation and decisions requiring rideshare companies to publicize their data, including New York in 2017 [
53], Seattle in 2018 [
13], and San Francisco in 2019 [
67]. Because of the increased availability of data, it is becoming more feasible to build similar visualizations for different cities and at a larger scale. This type of interactive tool allows a depth of understanding beyond the typical statistics seen in a white paper or fact sheet, acting as a boundary object between drivers and policymakers. This can advance drivers’ collective interests by adding more accessible public-facing data-driven tools to raise public awareness of driver issues.
While some cities and states have established minimum wages for drivers such as New York City in 2018 [
16], California in 2020 [
48], and Washington in 2022 [
10], the lack of legislation in other locations—such as Chicago—means that drivers are often being paid less than local minimum wages after subtracting expenses [
49,
64]. Countless studies have been performed examining rideshare datasets [
49,
61], however, without publicly available data, this type of independent analysis is impossible. However, there are still problems with data-driven studies as a means of activism. Two prior studies attempted to calculate average hourly earnings in Seattle and came up with two very different values ($23.25/hour vs. $9.73) due to wildly differing methodologies in calculating expenses [
50]. Intense care must be taken regarding understanding the methodology and meanings of ridesharing terms before issuing declarative statements or legislation on the basis of data. Through our partnership with the Independent Drivers Guild, we hope to develop further public-facing tools to build awareness of the realities of rideshare work and advance worker rights.
8 Limitations
We acknowledge that our study has limitations. We faced challenges in recruiting a diverse range of drivers, possibly due to the requirement in our study for participant data; our participants came primarily from Reddit, Facebook driver forums, or a driver advocacy group, thus they may be more data literate and aware of platform issues than a standard driver. We conducted this study with Chicago drivers where we had access to publicly available data. Additionally, due to the limitations of Uber’s data, we were unable to calculate certain metrics for drivers. The duration of our sessions limited how much participants could share with us. Our sessions lasted for two hours, yet some participants wanted to discuss more after it ended. Finally, as a qualitative study, the findings we offer around worker-centered AI design should be further investigated through other research methods with a larger number of participants.