Scalable crisis relief: Crowdsourced SMS translation and
categorization with Mission 4636
Vaughn Hester
Aaron Shaw
Lukas Biewald
CrowdFlower
455 Valencia Street
San Francisco, CA 94103
vaughn@crowdflower.com
UC Berkeley
Department of Sociology
410 Barrows Hall
Berkeley, CA 94720-1980
adshaw@berkeley.edu
CrowdFlower
455 Valencia Street
San Francisco, CA 94103
lukas@crowdflower.com
ABSTRACT
Crowdsourced crisis response harnesses distributed human networks
in combination with information and communication technology
(ICT) to create scalable, flexible and rapid communication systems
that promote well-being, survival, and recovery during the acute
phase of an emergency. In this paper, we analyze a recent experience in which CrowdFlower conducted crowdsourced translation,
categorization and geo-tagging for SMS-based reporting as part of
Mission 4636 after a 7.0 magnitude earthquake struck Haiti on January 12, 2010. We discuss CrowdFlower’s approach to this task,
lessons learned from the experience, and opportunities to generalize the techniques and technologies involved for other ICT for development (ICTD) applications. We find that CrowdFlower’s most
significant contribution to Mission 4636 and to the broader field of
crowdsourced crisis relief lies in the flexible, scalable nature of the
pool of earthquake survivors, volunteers, workers, and machines
that the organization engaged during the emergency response efforts.
Categories and Subject Descriptors
H.4.3 [Information Systems Applications]: Communications Applications; J.m [Computer Applications]: Miscellaneous; K.4.2
[Computers and Society]: Social Issues
General Terms
Human Factors, Design, Management, Languages
Keywords
Human Computation, Crowdsourcing, Crisis Response, Haiti, Translation, Geolocation, ICTD, Development
1. INTRODUCTION
Crowdsourced crisis response harnesses distributed human networks in combination with information and communication technology (ICT) to create scalable, rapid communication systems that
Permission to make digital or hard copies of all or part of this work for
personal or classroom use is granted without fee provided that copies are
not made or distributed for profit or commercial advantage and that copies
bear this notice and the full citation on the first page. To copy otherwise, to
republish, to post on servers or to redistribute to lists, requires prior specific
permission and/or a fee.
ACM DEV’10, December 17–18, 2010, London, United Kingdom.
Copyright 2010 ACM 978-1-4503-0473-3 ...$10.00.
promote well-being, survival, and recovery during the acute phase
of an emergency. In this paper, we analyze a recent experience
in which CrowdFlower conducted crowdsourced translation and
geo-tagging for SMS-based emergency reporting as part of Mission 4636 in response to the 7.0 magnitude earthquake that struck
Haiti on January 12, 2010.
Distributed, SMS-based systems for information sharing can facilitate effective responses to humanitarian crises worldwide. During the violence that followed the disputed 2007 presidential elections in Kenya, Ushahidi pioneered the use of FrontlineSMS for
crisis mapping and response. Since then, Ushahidi and others have
deployed FrontlineSMS and similar reporting systems in a variety
of settings. One constraint of such systems emerges when linguistic and other factors prevent a scalable means of filtering, verifying, translating, and geo-tagging all the incoming SMS reports.
Crowdsourcing, or the act of engaging distributed groups of people to complete microtasks or generate information, offers unique
advantages in such situations when combined with (a) willing communities of volunteers and (b) a highly elastic “on-demand” labor
platform.
Mission 4636 was a collaboration among technology companies,
international non-governmental organizations and emergency relief
agencies in the aftermath of the 2010 Haitian earthquakes. Mission 4636 combined an SMS-based reporting system with crowdsourced translation and geolocation to collect, categorize and map
reports of emergency data from a distributed human network in
Haiti. Many organizations and steps were involved in this process.
CrowdFlower, a technology company based in San Francisco, provides a crowdsourcing plaform with rigorous quality control tools.
The platform enables access to labor on demand, or to pools of
distributed human judges around the world. Users of the platform
create online projects, or microtasks, which are completed by multiple workers working in parallel. CrowdFlower facilitates qualitycontrolled completion of microtasks such as text or image moderation, sentiment analysis, and business listing verification. CrowdFlower’s participation in Mission 4636 was one of the first times
this crowdsourcing platform had been applied to emergency relief
efforts.
This paper focuses on CrowdFlower’s approach to the crowdsourcing translation and geolocation components of the broader
Mission 4636 effort. In particular, we describe the system design
and performance as well as lessons learned from the experience.
The contributions of other organizations and the details of other
parts of this group effort are, for the most part, beyond the scope of
this article.
While our ability to evaluate the project was severely limited for
many reasons, we find that Mission 4636’s unusual capacity to engage survivors both at the stage of information collection and at the
stage of information processing (the latter in exchange for money)
has the potential to produce a triple impact by (a) alleviating emergency conditions, (b) expanding employment opportunities in the
midst of a crisis, and (c) contributing to individuals’ overall sense
of participation in the recovery process.
Finally, we explore several ways in which the crowdsourced translation and geolocation might be improved, as well as opportunities
to generalize these techniques and technologies for application in
other emergency responses.
2. CROWDSOURCING AND ICTD
2.1
Crowdsourcing: Definition, Origins and
Applications
We define crowdsourcing as “the act of engaging distributed groups
of people to complete microtasks or generate information.”1 The
new term conceals a rich history of “human computing,” which
David Alan Grier has traced back to early applications of calculus, including 18th century astronomy and World War I era national statistics offices whose responsibilities included the creation
of ballistics tables for combat use [11]. In light of this history,
we view the recent introduction of digital, networked ICTs into
circuits of human computation as a catalyst for a sort of computing renaissance. With the global spread of the Internet and cellular telecommunications, the opportunities to harness the potential
of distributed human intelligence and cognition have never been
greater.
The increase in the number of applications of crowdsourcing follows the broader expansion of digital ICTs. The proliferation of
wireless, cellular and satellite networks as well as the diffusion
of devices connected through these networks facilitates the transmission of many different kinds of digital information and creates
opportunities to collect, process, and analyze data on a scale previously unavailable.
In this regard, the spread of networked, digital ICTs has facilitated a rapid expansion in the tools, techniques and applications for
crowdsourcing and human computation. Several online labor markets and commercial crowdsourcing service providers, including
(among the U.S. based firms) LiveOps, Amazon.com’s Mechanical Turk, Casting Words, and CrowdFlower (a) contribute to the
growth of the tools, techniques, and applications for crowdsourcing and (b) constitute an emergent sector within the ICT industry.
The analysis and application of crowdsourcing has grown in a
similar manner. Among many noteworthy examples in the computational and social sciences, Von Ahn and collaborators have developed widely adopted image labeling, text transcription, and music
tagging tools that harness distributed human intelligence [27, 19,
26]. Also, a range of research applications involving the use of
crowdsourcing to conduct behavioral and labor experiments [16,
15, 6, 5], filter inexpert judgments [24], conduct user studies [18]
and analyze large bodies of textual content [13] have introduced
crowdsourcing into diverse disciplinary contexts.
The rise of crowdsourcing applications has elicited divided opinions among researchers about the ethical, economic and political
implications of these phenomena. Some scholars have emphasized
the potential for crowdsourcing and related practices to bring about
social welfare gains by facilitating (a) innovation, (b) distributed
problem solving and (c) access to information resources[2]. Others
have made the case that crowdsourcing (as an example of human
1
This definition builds on the original usage by Jeff Howe, who
coined the term[17].
computing) also makes it easier to exploit workers, conduct malicious tasks, or generally devalue human intelligence[28].
We contend that crowdsourcing can and will produce positive social welfare benefits in a variety of contexts, but that nothing in the
technologies or techniques makes this inherently true. We support
this position through our analysis of Mission 4636, which entailed
the confluence of several different actors, technological platforms,
networks, and circumstances following the 2010 earthquakes in
Haiti.
2.2
Relevance to ICTD
Mission 4636 exemplifies the growing number of projects that
incorporate crowdsourcing and SMS networks to advance the social welfare aims at the core of ICT for development (ICTD) research.2 In this sense, Mission 4636 extends many existing tools
and approaches while also contributing several unique innovations
into the fields of ICTD and emergency response. We review a few
salient examples of ICTD and emergency response projects that
use crowdsourcing before turning to our analysis of Mission 4636
below.
Efforts to incorporate crowdsourcing technologies and practices
into ICTD projects build on existing work with networks of various
kinds. For example, the use of sensor networks to monitor water
quality or irrigation levels also harnesses distributed, large-scale
information gathering to comparable effects, but it involves a less
continuous flow of human participation than most crowdsourcing
work[22, 23].
Crowdsourcing has begun to play a more prominent role in development initiatives and emergency response efforts. An ideal example of this trend, the San Francisco nonprofit Samasource facilitates digital work opportunities for underserved groups such as
women, children, and refugees around the world. As part of a
well-publicized 2007 search and rescue effort, emergency workers
collaborated with users of Google Earth and Amazon Mechanical
Turk to look for Steve Fossett, a pilot whose plane had disappeared
over the Nevada desert.3 While the crowdsourced search ultimately
failed, we find no reason to believe that the failure will prevent similar attempts in the future and find the effort to achieve a large-scale
search and rescue operation using a combination of paid and unpaid
volunteers especially relevant to the case of Mission 4636.
Similarly, SMS-based crowdsourcing projects represent a relatively new addition to the existing repertoire of ICTD applications for mobile phones. SMS offers numerous practical advantages in regions with high cell phone diffusion rates and GSM network coverage such as many parts of the Caribbean[7, 14]. The
txteagle project represents the most innovative effort to use SMSbased crowdsourcing to enable cell phones subscribers in Kenya
and Rwanda to earn income in exchange for completing microtasks[8].
Rapid emergency response adds an additional layer of complexity and difficulty to ICTD projects and research[20]. Ken Banks
and the creators of FrontlineSMS first applied an SMS-based reporting system to crowdsource election monitoring in Nigeria in
2007[1]. Subsequently, the FrontlineSMS system was extended
2
Brewer et al. provide an early, synthetic overview of ICTD[3].
Following recent work by Burrell and Toyama, we refrain from
attempting to narrowly define ICTD or development, but use the
terms here to refer to the scholarly community that (a) identifies
itself with similar terms and (b) focuses its research on the social
impact and implications of ICTs in geographic regions historically
considered to be “developing” or “underdeveloped”[4].
3
More information about the search and links to news stories can
be found at: http://www.stevefossett.com/
by Ushahidi to help map post-election violence in Kenya later that
year. FrontlineSMS itself incorporated this system into the FrontlineSMS:Medic project to promote health care through effective data
collection and patient-physician interactions.4
Building on previous ICTD, crowdsourcing, emergency response,
and SMS-reporting projects, the key innovations of Mission 4636
consist of several distinct components. First, the project expands
on the efforts of FrontlineSMS, Ushahidi and related groups to integrate both the survivors of emergencies as well as the promoters of ongoing community development into the recovery process.
Second, Mission 4636 also takes advantage of the distributed effort
of volunteers and paid workers to achieve a response of sufficient
scale and speed in the wake of an unexpected event. Finally, and
perhaps most uniquely, Mission 4636 represents an attempt to bypass the dual barriers of scale and language through a complex,
multi-step workflow that incorporated a number of organizational
actors around the world. Such an adoption of multiple tools as well
as collaborative strategies in response to a sudden crisis are becoming a viable and valuable resource for emergency responders[10].
Each worker was shown an SMS message from Haiti. The worker
was asked whether he or she could understand and translate the
message. If not, the worker had the option of moving onto another
message. A second question allowed the worker to flag a particular message if it was indecipherable or if it did not contain any
information.
The worker then completed as many fields as possible in response to questions about the SMS message.5
3. CROWDSOURCING CRISIS RESPONSE
IN HAITI
Once the messages had been processed and categorized using
this framework, CrowdFlower posted feeds of translated and tagged
messages.
3.1
January 2010 Earthquakes
On January 12, 2010, a magnitude 7.0 earthquake struck Haiti 25
miles west of Port-au-Prince. In the 12 days following this quake,
52 aftershocks of magnitude 4.5 or higher were recorded.
Despite massive devastation of basic infrastructure in Port-auPrince as well as complete or periodic disruption of many essential
services, most of Haiti’s cell tower infrastructure remained intact.
Over 20 countries sent military personnel to contribute to relief
efforts on the ground. Dozens of international NGOs and agencies deployed staff and resources to Haiti as well. Communication,
coordination and oversight were chaotic in the first few days after
January 12.
3.2
Mission 4636
Within 48 hours of the first earthquake, Josh Nesbit of FrontlineSMS:Medic and Katie Stanton of the U.S. State Department convinced DigiCel, the largest telco in Haiti, to set up a short code –
“4636” – that any individual could text for free to request help.
Robert Munro of Energy for Opportunity and Brian Herbert of
Ushahidi set up a workflow where Kreyol-speaking volunteers could
translate the messages into English and classify them. The translated messages were then routed to aid agencies and workers who
could send relief. Once the system was working, InSTEDD (in collaboration with Thompson Reuters) worked on the ground to broadcast the existence of the “4636” short code to as many Haitians as
possible using radio and other accessible forms of communication.
3.3
Distributed Tagging and Translation with
Crowdflower
People throughout Haiti used this short code system to send very
urgent messages and pleas for help. As the volume of messages
grew, so did the need for a more robust workflow platform. CrowdFlower began receiving feeds of SMS messages via Ushahidi (and
its Haitian telco partners) and facilitating their review. Volunteers
on the CrowdFlower platform provided translation, categorization
and geo-tagging.
4
A case study by Goldstein and Rotich situates Ushahidi’s early
work within the broader context of the Kenyan elections[9].
• Sender’s first and last name (free text entry)
• Sender’s gender (radio button)
• Sender’s status: missing, survived, wounded, deceased or
other
• Category of the aid request. The seven main categories were
Emergency, Threats, Vital Logistics, Response, Other, Persons News and Child Alone.
• Zone or neighborhood of the request/report
4. RESULTS
4.1
Worker Population
Kreyol speakers around the world were asked to contribute to the
translation and tagging efforts. Through word of mouth and a small
amount of viral online marketing about Mission 4636, the pool of
volunteer translators grew throughout the Haitian diaspora. Mission 4636 also created a short viral video to publicize its efforts.6
The Kreyol SMS translation and classification work was completed with contributions from a large number of volunteers around
the world. In Haiti, residents also played an integral role in the
SMS-based relief efforts. Before the first earthquake, Samasource
launched a new work center in Haiti, which now made it possible
to hire earthquake survivors to contribute to crowdsourced relief efforts. This Samasource service partner assumed a large amount of
the earthquake relief responsibilities, not only providing labor for
the emergency message routing, but also creating jobs for the local
economy.
We report geocoded IP information on the locations of the workers who translated and tagged SMS messages via CrowdFlower in
Table 1.
The distribution of workers’ IP addresses provides an approximate sense of the geographic distribution of the workers themselves. The vast proportion of U.S. addresses illustrates that these
crowdsourcing efforts were dominated by volunteers and workers
living outside the earthquake affected region.
4.2
System Throughput
The combined effort of the workers, volunteers and various partner organizations involved produced low latency and accessible results. CrowdFlower processed approximately 100,000 SMS messages in total and collected over 16,000 judgments (translations)
on the messages that were decipherable and legible. At peak volume, CrowdFlower workers and volunteers filtered over 5,000 SMS
5
A
preview
of
the
task
is
available
through
the
Crowdflower
internal
judgment
interface:
http://www.crowdflower.com/judgments/mob/4980.
6
Available at http://www.youtube.com/watch?v=QVvYauvydac
Figure 1: Traffic per day (workers)
Table 1: Worker Location by Country (geocoded IP address)
Count
Percentage†
USA
Canada
Haiti
Switzerland
Colombia
Germany
France
Guadeloupe
Japan
Great Britain
Kenya
11616
853
232
155
52
39
16
10
4
2
1
89.49
6.57
1.79
1.19
0.40
0.30
0.12
0.08
0.03
0.02
0.01
Total
12980
100.00
†
No IP information was
available for 9289 workers or
41.7% of the aggregate workforce.
messages in one hour. The average response time to filter, translate, map/geocode and categorize a message did not exceed two
minutes. Parts of the emergency SMS message feeds – and the
maps generated by Ushahidi – are now used by a growing number
of organizations, including the Red Cross, Plan International, charity:water, the U.S. State Department, International Medical Corps,
AIDG, USAID, FEMA, the U.S. Coast Guard Task Force, World
Food Program, SOUTHCOM, OFDA and UNDP.
Figure 1 illustrates the overall throughput of the system as the
number of individual workers participating in Mission 4636 through
the Crowdflower platform per day during the months immediately
following the earthquakes. The graph begins January 28, 2010 and
ends May 12, 2010.
Figure 1 also conveys the uneven spikes typical of many crowdsourcing jobs. We are not sure what caused the unlikely spike
around mid April. To reiterate, the ability to accomodate unpredictable contractions and expansions in work volume and the related volume of workers needed to complete this work is a tremendous advantage of dedicated, crowdsourcing systems.
A key component of the task consisted in categorizing the messages that came through. Table 2 provides summary information
about the intelligible aid request messages that were categorized
by workers and volunteers through the CrowdFlower platform. The
categories are sorted by prevalence.
Note the prevalence of the “Other” and “Blank” categories, which
made up over 80% of the total. Also note that despite the high number of blanks, “the crowd” applied a total of 12,364 “non-blank”
categories, or 55.52% of the total.
4.3
Quality Control
The urgent nature of Mission 4636 did not allow for the application of redundancy that is a traditional step in CrowdFlower quality
control. In other crowdsourcing projects, multiple workers respond
to the same question, and an aggregation method (such as majority voting) then produces a robust estimate of the single accurate
response. In the Mission 4636 efforts, the majority of the SMS
messages were reviewed by a single judge only. Accordingly, it
was not possible to calculate precision or accuracy by measuring
agreement across multiple workers on a single SMS message.
Despite the absence of multiple judgments on each incoming
SMS, CrowdFlower employed several strategies for quality control.
A tactic for maximizing the skills of Kreyol speakers was to filter
out illegible, indecipherable or nonsense messages. By allowing
workers to flag and remove these SMS data units, overall latency
was streamlined and translators remained focused on substantive
messages to translate and categorize.
Mission 4636 also established a chat room for the online workers to facilitate collaboration and clarification. This parallel stream
of communication allowed workers to confirm and share additional
details on exact locations and addresses in Haiti. We believe this
communication contributed to heightened accuracy and more efficient throughput.7
Once an SMS message was processed, translated and categorized
by CrowdFlower volunteers, it was immediately pushed to an outbound feed to Ushahidi. During the project, a CrowdFlower effort
was made to improve the quality of the processed SMS messages
by improving the speed and accessibility of the outbound feed of
the translated and categorized messages. CrowdFlower and Mission 4636 published the feed with the aim of making the messages
available to as many relief agencies and actors as possible.
5. DISCUSSION
5.1
Impact
One challenge of operating in an acutely affected environment
after a major emergency is the difficulty of accurately assessing the
impact of many emergency response efforts. This was complicated
by the fact that the CrowdFlower SMS translation and categorization effort employed workers from all over the world and served
primarily to distribute information across dozens of separate organizations and agencies operating in Haiti. We were unable to collect
7
A brief report by Robert Munro posted on the Ushahidi blog supports this claim[21].
Table 2: Summary of SMS Categorization Results
systematic data to evaluate CrowdFlower’s role in Mission 4636 in
a rigorous way. In the absence of such data, we discuss the impact of the project in this section with the caveat that our findings
are not conclusive and may diverge from the observations of other
individuals and organizations involved in the relief effort.
As described in the previous section, overall system performance
scaled well to match the flow of incoming messages. The massive
number of volunteers worldwide in combination with the Haitibased workforce brought online by Samasource meant that the system processed most emergency messages with extremely low latency. For messages like “Non mwen se lucaint lucoit madanm
mwen ansent li rive le poul akouche nou nan delma 31 ri maryen
n 21 nan lakou legliz apostolik anfas site jeremi, mpa” (“condition bloody about. undergoing children delivery corner of delmas
31 and rue marine”) it was crucial not just to be fast, but to have
local knowledge to get the exact longitude and latitude from an ambiguous 140 character message as well as an accurate classification
so that the right aid agency can be deployed. In the case of this
particular SMS, USGS responded “just got emergency SMS, child
delivery, USCG are acting, and, the GPS coordinates of the location
we got from someone of your team were 100% accurate”[12].
The organizations receiving and using Mission 4636 information
to direct relief services provided strong, positive feedback. Craig
Clark of the U.S. Marine Corps commented on the text message
project: “I wish I had time to document to you every example, but
there are too many and our operation is moving too fast ... I say
with confidence that there are hundreds of these kinds of [success]
stories. The Marine Corps is using your project every second of the
day to get aid and assistance to the people that need it most.” Many
other relief actors, including the State Department, shared similar
comments on the utility of Mission 4636 and the unprecedented
efficiency it added to relief efforts[12].
Another consideration with respect to the overall project impact
was the speed with which it was deployed. The SMS translation
project was developed and launched extremely quickly, and there
were numerous challenges throughout this process. Publishing an
RSS feed of the translated SMS that was accessible to many different groups was key in allowing aid agencies to start downloading
the particular categories of messages that were important to their
work. Ushahidi’s maps of the reported incidents and the requests
for help made an important delineation of the overall trends separate from the individual needs.
After evaluating the lessons learned from CrowdFlower’s experience with Mission 4636, we found a few key limitations and recommendations.
The biggest technical constraint had to do with the various SMS
providers changing the formats of their feeds during the aftermath.
Such changes imposed major obstacles for SMS message processing, as CrowdFlower engineers had to alter the underlying structure
of the message-processing system in a hurry after each change. We
recommend that telecommunications providers do everything in
their power to maintain consistent data formats during emergency
situations.
As described above, CrowdFlower was constrained in its ability
to implement a more robust set of quality control tools in this case.
Previous research has demonstrated the significant gains in judgment accuracy and precision that can result from even relatively
simple quality measurement techniques[25, 24]. CrowdFlower deploys many of these same techniques with its clients through its
online platform; however, the Mission 4636 tasks did not incorporate many of these tools as a result of the urgency and speed with
which the response system was deployed.
This experience reflects the fact that in crowdsourcing, as in
many computational endeavors, we often encounter the tension between precision and recall. In the case of Mission 4636, the crowdsourcing system was optimized for recall, perhaps at the expense
of precision.
In future iterations of crowdsourced microtasks to support disaster relief, it would also be advantageous to add flexible functionality to the CrowdFlower platform. Specifically, certain categories or
types of SMS messages may warrant review by more than one human judge. This dynamic or variable redundancy on highly urgent
messages could reveal more details about the actual location of the
message and prevent any mistakes that a single translator may have
made.
Another lesson learned concerns the design and presentation of
the SMS data itself through the CrowdFlower worker interface. The
task display included information about the number of SMS messages awaiting judgment. If this number was too big, volunteers
reported that they felt hopeless, yet others reported that the transparency surrounding the pending workload provided a great motivator for the people churning through the tasks. We draw two
conclusions from these reports. First, we might make the display
of the queue an optional feature, allowing workers and volunteers
the latitude to hide it if they wish. Another possible improvement
would be to pre-populate the task with a Google translation of the
SMS text to ease the workload of the translators, thereby enhancing
the speed of the system and reducing frustration simultaneously.
Workers could simply correct or verify the machine translations,
pointing out any specific or missing details that need enhancement.
A final concern regards the overall quality of the information
flowing through the system. Volunteered information and crowdsourced analysis may produce results of questionable precision and
accuracy. At the same time, the effective integration of such “imperfect signals” can complement existing emergency response systems. In other words, emergency information that may be imperfect is better than an absence of any information in the context of
the acute phase of an emergency cycle. An evaluation of this claim
should be left to future work. As in all crowdsourcing endeavors,
the incentives of the participants will affect the end result of the
project. Disaster victims in dire situations are, in our opinion, less
likely to attempt to scam a lifesaving system than unethical online
workers in an enterprise crowdsourcing task may be.
5.2
6. FUTURE WORK
Category
†
Blank
Other
Response request
Vital logistics
Emergency
Persons news
Threats
Child alone
Total
Count
Percentage
9905
9204
1435
734
691
235
52
13
44.48
41.33
6.44
3.30
3.10
1.06
0.23
0.06
22269
100
†
These messages were analyzed, but
category information was left blank.
Limitations and Recommendations
In future crowdsourced emergency response efforts, we propose
adding layers of quality control along the lines laid out in previous research[25, 24]. Given the success of this project in KreyolEnglish translation, we also can engage Kreyol speakers to verify
the quality of translations. This creates a multi-step workflow to
validate results, which would increase completion time and reduce
throughput. Our hypothesis is that this would generate substantial
improvements in quality.
In future applications of crowdsourcing in disaster relief, more
nuanced results could be available through more use of the CrowdFlower platform. Specifically, having more people review or verify certain SMS messages (i.e., the more urgent ones) could allow
workers to extract more details about the actual location of the message. Multiple judgments per message would prevent any mistakes
that a particular single translator may have made. 8
It is clear that new technology and cross-platform collaboration
stand to make significant contributions to future disaster preparedness and relief efforts. Crowdsourcing has a role within these efforts and should be considered a highly flexible, highly responsive
tool for any and all actors. As mobile phone penetration grows
around the world, we can anticipate growth in the coverage of
SMS-based relief efforts in the event of an emergency. As Internet penetration grows, we anticipate an increase in the available
worker pools to assist with crowdsourced translation, categorization and tagging efforts to support on-the-ground relief efforts. Finally, as governments, nonprofits and humanitarian relief agencies
learn more about the possibilities that lie in this new tool for emergency relief, we are confident that greater numbers of lives will not
only be changed but may very well be saved.
Ongoing efforts by CrowdFlower and others are focused on establishing a more permanent version of this short code system and
SMS-based reporting for hurricane season in Haiti. International
relief agencies and researchers should examine the feasibility of
integrating crowdsourcing into crisis response assessment methods
and toolkits. Crowdsourcing can complement existing emergency
response systems and networks, but greater information and education on crowdsourced options and approaches will need to be circulated among the international relief community. Local authorities
around the world should also look to their version of 911 systems
to see whether parts of the response cycle could be crowdsourced
efficiently. Situations that are less urgent than natural disasters can
create spikes in the volume of digital information that enters these
automated systems. Crowdsourcing can facilitate processing such
unexpected spikes.
A major danger of expanding crowdsourced emergency response
stems from the potential for poor integration with existing systems.
In the aforementioned example of the search and rescue efforts to
find Steve Fossett, Mechanical Turk workers produced a vast number of false-positives, straining emergency agencies’ resources and
willingness to collaborate.
Future projects should evaluate the effectiveness of these kinds
of solutions as well as a more comprehensive evaluation of their integration and utility in acute emergency repsonse. Future research
should also assess SMS-reporting system performance as well as
short and long-term impacts on survivor populations and future
8
As of the time of publication of this paper, CrowdFlower was
involved with the Pakreport project, a similar SMS-based relief collaboration with Ushahidi and other agencies. In this
instance, there was an opportunity to draw upon the lessons
learned with Mission 4636 and to structure the SMS processing such that each message was moderated by three volunteers.
More details on the broader Pakreport effort can be found at
http://www.pakreport.org/ushahidi/
emergency preparedness and response efforts.
7. CONCLUSIONS
We have presented an overview of CrowdFlower’s efforts to maximize the quality of the crowdsourced information geolocation and
categorization as part of Mission 4636. Our focus in this analysis on evaluating tools and techniques rather than project participants’ interpretations therefore positions these findings closer to the
“problem solving” pole of the pragmatist-interpretivist spectrum of
ICTD research laid out in a recent essay by Burrell and Toyama[4].
We find that CrowdFlower’s most significant contribution to Mission 4636 and to the broader field of crowdsourced crisis relief
stems from the scalability of the pool of earthquake survivors, volunteers, workers, and machines that the organization engaged at
an acute stage of the emergency response effort. This collaboration generated a high volume of filtered, categorized, and geocoded
messages that were used by relief workers and agencies to direct
assistance where it was needed most.
While we are unable to conduct a more thorough review of the
project impact in Haiti among earthquake survivors and their families, our analysis offers several significant take-away points:
• The importance of task design.
• The importance of effective collaboration with other organizations and actors involved.
• The opportunities made possible by efficient, large -scale information processing.
• The opportunities to improve system throughput and quality
control.
Mission 4636 came to life because multiple organizations were
able to work together quickly and effectively. It illustrated the advantages of a flexible, crowdsourced workflow for disaster relief
management. Businesses use crowdsourcing because it eliminates
the need to forecast or anticipate work capacity in advance. Disaster relief is simply an extreme extension of this same situation.
Traditionally, harnessing thousands of volunteers would represent
a logistical quagmire, but crowdsourcing platforms are specifically
suited for these types of amorphous, virtual labor forces. Through
viral channels and a microtask framework, it was possible (and will
be possible in the future) to have thousands of people around the
world doing mission-critical work within hours or days.
8. ACKNOWLEDGMENTS
The authors are grateful for the support and collaboration of everyone at CrowdFlower and the participating organizations in Mission 4636. We especially thank Anisha Sekar and Diyang Tang
(CrowdFlower interns) for their invaluable assistance with the preparation of this manuscript.
9. REFERENCES
[1] K. Banks. And then came the nigerian elections: The story of
frontlinesms. SAUTI: The Stanford Journal of African
Studies, (Spring/Fall):1–4, 2007.
[2] Y. Benkler. The Wealth of Networks: How Social Production
Transforms Markets and Freedom. Yale University Press,
2006.
[3] E. Brewer, M. Demmer, B. Du, M. Ho, M. Kam,
S. Nedevschi, J. Pal, R. Patra, S. Surana, and K. Fall. The
case for technology in developing regions. IEEE Computer,
38(6):25–38, 2005.
[4] J. Burrell and K. Toyama. What constitutes good ictd
research? Information Technologies International
Development, 5(3), Fall 2009.
[5] D. Chandler and A. Kapelner. Breaking monotony with
meaning: Motivation in crowdsourcing markets. University
of Chicago mimeo, 2010.
[6] D. L. Chen and J. J. Horton. The wages of paycuts: Evidence
from a field experiment. Working Paper, 2009.
[7] H. Dunn. Teleworking the mobile caribbean: Emerging
patterns of broadband-assisted remote work among the
marginalized in jamaica and trinidad and tobago. Information
Technologies International Development, 5(2), 2009.
[8] N. Eagle. txteagle: Mobile crowdsourcing. In
Internationalization, Design and Global Development, pages
447–456. Springer, Berlin, 2009.
[9] J. Goldstein and J. Rotich. Digitally networked technology in
kenya’s 2007-2008 post-election crisis. Berkman Center for
Internet and Society: Internet & Democracy Case Study
Series, 2008.
[10] M. F. Goodchild and J. A. Glennon. Crowdsourcing
geographic information for disaster response: a research
frontier. International Journal of Digital Earth, 2010.
[11] D. A. Grier. When Computers Were Human. Princeton
University Press, 2005.
[12] V. Hester. Crowdflower impacts the future of disaster relief.
http://haitirewired.wired.com/profiles/blogs/crowdflowerimpacts-the-future.
[13] D. Hopkins and G. King. A method of automated
nonparametric content analysis for social science. American
Journal of Political Science, 54(1):229–247, 2010.
[14] H. Horst and D. Miller. The Cell Phone: An Anthropology of
Communication. Berg Publishers, 2006.
[15] J. Horton and L. Chilton. The labor economics of paid
crowdsourcing. Proceedings of the ACM Conference on
Electronic Commerce 2010 (forthcoming), 2010.
[16] J. J. Horton, D. Rand, and R. J. Zeckhauser. The online
laboratory: Conducting experiments in a real labor market.
SSRN eLibrary, 2010.
[17] J. Howe. Crowdsourcing: Why the Power of the Crowd Is
Driving the Future of Business. Crown Business, 2008.
[18] A. Kittur, E. H. Chi, and B. Suh. Crowdsourcing user studies
with mechanical turk. Proceedings of Computer Human
Interaction (CHI-2008), 2008.
[19] E. Law, L. Von Ahn, R. Dannenberg, and M. Crawford.
Tagatune: A game for music and sound annotation. In
International Conference on Music Information Retrieval,
pages 361–364. Citeseer, 2003.
[20] C. F. Maitland, N. Pogrebnyakov, and A. F. van Gorp. A
fragile link: Disaster relief, icts and development. In
Information and Communication Technologies and
Development, 2006, 2006.
[21] R. Munro. The nuts and bolts behind 4636 in haiti.
http://blog.ushahidi.com/index.php/2010/01/22/the-nuts-andbolts-behind-4636-in-haiti/.
[22] J. Panchard, S. Rao, T. Prabhakar, H. Jamadagni, and
J. Hubaux. Common-sense net: Improved water management
for resource-poor farmers via sensor networks. In
Information and Communication Technologies and
Development, 2006., 2006.
[23] N. Ramanathan, L. Balzano, D. Estrin, M. Hansen,
T. Harmon, J. Jay, W. Kaiser, and G. Sukhatme. Designing
[24]
[25]
[26]
[27]
[28]
wireless sensor networks as a shared resource for sustainable
development. In Information and Communication
Technologies and Development, 2006., 2006.
V. S. Sheng, F. Provost, and P. G. Ipeirotis. Get another
label? improving data quality and data mining using
multiple, noisy labelers. Knowledge Discovery and Data
Minding 2008 (KDD-2008), 2008.
R. Snow, B. O’Connor, D. Jurafsky, and A. Y. Ng. Cheap and
fast—but is it good? evaluating non-expert annotations for
natural language tasks. Proceedings of the Conference on
Empirical Methods in Natural Language Processing
(EMNLP 2008), 2008.
L. Von Ahn. Games with a purpose. IEEE Computer
Magazine, 39(6):96–98, 2006.
L. Von Ahn and L. Dabbish. Labeling images with a
computer game. In Proceedings of the SIGCHI conference
on Human factors in computing systems, pages 319–326.
ACM, 2004.
J. Zittrain. Ubiquitous human computing. Philosophical
Transactions of the Royal Society A, 366:3813–3821, 2008.