Web Science Challenges 1
Web Science Challenges 1
The act of searching for security flaws (vulnerabilities) in a piece of software was previously considered to be the preserve of malicious
actors, or at least actors who wished to cause chaos. Increasingly, however, companies are recognising the value of running a bug
bounty program, where they will pay ‘white hat’ hackers to locate and disclose security flaws in their applications in order that they
can fix it. This is known as a ‘bug bounty’ or a ‘vulnerability reward program’, and at present has seen comparatively little research.
This paper introduces two existing research on bug bounties in two areas: as a means of regulating the sale of vulnerabilities; and as a
form of crowdsourcing. We argue that the nature of bug bounties makes Web science particularly suitable to drive forward research.
We identify gaps in the current literature, and propose areas which we consider to be particularly promising for future research.
CCS Concepts: • Information systems → Crowdsourcing; • Security and privacy → Social aspects of security and privacy;
Additional Key Words and Phrases: Crowdsourcing, Vulnerability research, Bug bounties
1 INTRODUCTION
A ‘bug bounty’ or ‘vulnerability reward program’ (VRP) is the process for rewarding the discovery of a flaw or
vulnerability in a piece of software. The concept has been around for a long time, notably Donald Knuth offering
rewards for omissions in his The Art of Programming books, or flaws in his LATEXsoftware, and in the 1990s Netscape
offered a reward for flaws in its browser. Despite this history, examples of its application have been sparse up until the
last few years where its popularity has increased, as this decade high profile programs from companies such as Mozilla
and Google[26], and even the US Department of Defense in 2016 have started. There now exist services which act as
middlemen in connecting companies with people who are prepared to search their systems for weaknesses.
This paper will provide a brief review of some of the key research into bug bounties. For the most part, this has
been tangential, merely acknowledging their existence, as a part of overall Web or application security. Nevertheless,
there has been more recent research which has considered bug bounties in their own right, analysing the behaviour of
the participants, or the means in which a company operating a bug bounty might seek to optimise the quality of the
results of their bug bounty. We will identify the gaps in the literature, and identify areas where crowdsourcing and Web
Science research can assist in driving research forward in this area.
In Section 2, we introduce some terminology, and background research, before describing our literature review
in Section 3. We then present opportunities for future research in Section 4 based on crowdsourcing literature, and
discuss the suitability of this research area for Web Science. We conclude with a summary of our paper, and our
recommendations.
Permission to make digital or hard copies of all or part of this work for personal or classroom use is granted without fee provided that copies are not
made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. Copyrights for components
of this work owned by others than ACM must be honored. Abstracting with credit is permitted. To copy otherwise, or republish, to post on servers or to
redistribute to lists, requires prior specific permission and/or a fee. Request permissions from permissions@acm.org.
© 2017 ACM. Manuscript submitted to ACM
3 BUG BOUNTIES
3.1 Methodology
Despite the increasing popularity of bug bounties [6], and their seeming relationship with crowdsourcing, we were
unaware of any work which considered bug bounties within the context of crowdsourcing. The one exception to this
was Su & Pan, who proposed a system to introduce microtasking to the process, where additional actors would test and
verify the vulnerability submitted by another researcher [36].
As a result, we conducted a literature review, based on the methodology of Mao et al’s review of the related area
of crowdsourced software engineering [24]. The search was for the phrases “bug bount(y|ies)", “vulnerability reward
program", “vulnerability disclosure", in any available field in seven online search engines: ACM Digital Library, IEEE
Digital Library, Springerlink Online Library, Wiley Online Library, Elsevier ScienceDirect, ProQuest, and Google Scholar.
As a fallback, we additionally used snowballing of references where further titles were identified. To identify relevant
literature, the title, abstract and introduction sections of each paper were read, which was usually enough to identify it
as being outside the criteria for inclusion. Where this was not the case, the whole paper was read.
As an exploratory study, our research question was: what are the gaps in the existing literature related to bug bounties,
which can be addressed by crowdsourcing? As a result, the inclusion criteria for the literature review was that the paper
in question was about bug bounties specifically, or contained analysis of a bug bounty program, platform, or behaviour
of the workers in a program. Literature was excluded where it merely mentioned the existence of bug bounties, or
it focused on vulnerability management more generally. In future work, it is intended that this inclusion criteria be
widened, because these are all relevant with regards to policy implications, as well as assessing cost-effectiveness for
starting a bug bounty.
In conducting this search, a total of 11 papers were discovered which were primarily about bug bounties. This
includes a paper which is unpublished[40], and also non-academic work by Bugcrowd[6].
Having established the literature, we make use of established crowdsourcing literature as a means of solving the
overall research question. We do not attempt to conduct a survey of crowdsourcing literature, since it is beyond
the scope of this research and existing works have conducted surveys of crowdsourcing generally[9], and within
software engineering in particular[24]. We contend that according to Estellés-Arolas criteria[12], considering a bug
bounty program as crowdsourcing is valid. There is a defined crowd (security researchers) with a clear goal (locate
vulnerabilities) and a defined benefit for both the worker and (clearly defined) requester. It is an online process to solve
a problem, which uses the Internet, and has some degree of open call. Many calls will have restrictions imposed by the
platform based on the reputation of the researcher, particularly in regards to signal to noise ratio yet many will have a
completely open call subject only to self-selection by those considering themselves to have enough skill.
Despite this, as a format it is quite different to most crowdsourced programs. The nature of bug bounties is that it
accepts all valid unique submissions, provided they are within the scope of the call. Consensus is generally not required,
making it differ from many HIT tasks such as image tagging. Competition or innovation crowdsourcing calls will
generally accept only one, or a handful of submissions which best solve the problem[8].
1 See http://blog.trendmicro.com/pwn2own-returns-for-2017-to-celebrate-10-years-of-exploits/
Manuscript submitted to ACM
3.2 Literature
Whatever reservations there may be about ethics, efficiency or cost effectiveness, and with initial scepticism from major
players[13], bug bounties have been embraced by many of the major technology companies as well as gaining support
in other industries[6]. Some research about the participants in the programs has started, as well as means of getting
around some of the difficulties with running a bug bounty program.
As indicated by [27], prior to the introduction of a formal mechanism for buying and selling bugs through bug
bounties, obtaining a seller was challenging. Two reports by Ring illustrated this in further detail, discussing the
competing opinion of whether companies should offer bounties to vulnerability research - and additionally of some
companies prosecuting those discovering vulnerabilities[32, 33]. Kuehn & Mueller[16, 17] consider the changing
dynamics in information security towards bug bounties being considered a norm. After case studies on Microsoft
& Facebook’s bug bounty they conclude that bug bounty programs exist as a way of reducing uncertainty when
exchanging an information good as a reason for their development.
There are currently two major operators who have had mention made in the literature who facilitate bug bounties:
Bugcrowd2 , and Hackerone3 , although other websites offer a list of other Web applications offering a bounty. Bugcrowd
now publish an annual report on current trends in the bug bounty area, the most recent being in 2016[6]. Previously,
Wooyun offered a forum for researchers to disclose bugs, and had a more coercive model - the Web applications in
question were given a certain period of time to fix the flaws, before the flaw was made public. However, the website has
been out of action since July 2016 when the founder Fang Xiaodun was reportedly arrested[22]. As of March 2017, the
website still displays a message indicating that it is not operational4 .
Two of the older bounty programs, those of Mozilla and Google for their Web browsers Firefox and Chrome
respectively were studied in 2013 [26]. Both were found to be better value for the company than hiring a security
researcher on a permanent basis when considering the severe security flaws they discovered relative to the cost. They
found Google’s bug bounty program gleaned more vulnerabilities for a comparable amount of money, which they
suggested was due to the tiered reward system they operated compared to Mozilla’s flat fee.
In two separate papers, Zhao analysed the behaviour of white hats on the Wooyun[38] and Hackerone[39] platforms.
In both platforms they observed the behaviour of white hats in the different systems. In [38], it was observed that the
distribution of effort followed a power law, similar to that observed Lotka about academic publication frequency and
supporting observations by[26], with a maximum of 291 submissions and an overall average of 4.8. Analysis of both
revealed that when divided into categories of productivity each group reported a comparable amount of vulnerabilities,
in addition to the severity of the vulnerability and the ranking of the website.
Maillart et al. focus more on the misaligned incentives involved between the companies running a bug bounty
program and the researchers themselves[23]. The interest of the company is to exhaust the amount of flaws to a residual
level, whereas the interest of the researcher is the cumulative payoff they will gain from discovering bugs. This is best
served for the researchers by diversifying their efforts across different programs, since there will be bugs to discover
which are easier to locate, and there should be less competition. Analysing 35 programs on Hackerone, they follow
[38, 39] and observe a windfall effect within a few weeks of the start of the program, after which the amount of reports
reduce significantly in quantity. They attribute this to timing effects, where researchers switch to a new program, or
possibly stockpile vulnerabilities in advance of the program opening.
2 https://bugcrowd.com/
3 https://www.hackerone.com/
4 http://wooyun.org/ Last accessed 15 March 2017
Manuscript submitted to ACM
These papers regard the interest of a company running a bug bounty that they should seek to encourage as many
researchers as possible, in the hope of finding flaws. The diversification meaning that, the higher proportion of white
hats, the higher the probability that any flaw discovered is by a white hat rather than an attacker. This is supported by is
supported by [10], however they also discovered a positive correlation (r = 0.3591, p = 0.0307) between the amount of
accurate reports, and the amount of false positives. This indicates some tension here in regards to the costs of obtaining
too many results of low quality.
Laszka et al use economic modeling to analyse various models employed by Hackerone to mitigate this and hopefully
ensure a higher quality of submission[18]. They found that policies such as restricting access to those of high reputation,
or rate-limiting submissions could be effective, although care needed to be taken in implementing them otherwise
the overall utility would go down. Zhao et al expanded on this, although this work does not yet appear to have been
published[40].
REFERENCES
[1] Christian Medeiros Adriano and Andre van der Hoek. 2016. Exploring Microtask Crowdsourcing as a Means of Fault Localization. arXiv preprint
arXiv:1612.03015 (2016). https://arxiv.org/abs/1612.03015
[2] Abdullah M. Algarni and Yashwant K. Malaiya. 2013. Most Successful Vulnerability Discoverers: Motivation and Methods. In Proceedings of the
International Conference on Security and Management (SAM). 1.
[3] Ross Anderson, Chris Barton, Rainer Böhme, Richard Clayton, Michel JG Van Eeten, Michael Levi, Tyler Moore, and Stefan Savage. 2013. Measuring
the cost of cybercrime. In The economics of information security and privacy. Springer, 265–300.
[4] Michael S. Bernstein, Greg Little, Robert C. Miller, BjÃűrn Hartmann, Mark S. Ackerman, David R. Karger, David Crowell, and Katrina Panovich.
2015. Soylent: A Word Processor with a Crowd Inside. Commun. ACM 58, 8 (July 2015), 85–94.
[5] Rainer Böhme. 2006. A comparison of market approaches to software vulnerability disclosure. In Emerging trends in information and communication
security. Springer, 298–311. http://link.springer.com/chapter/10.1007/11766155_21
[6] Bugcrowd. 2016. The State of Bug Bounty. (June 2016).
[7] Bugcrowd. 2017. Defensive Vulnerability Pricing Model. (2017). https://pages.bugcrowd.com/whats-a-bug-worth
Manuscript submitted to ACM
[8] Thierry Burger-Helmchen and Julien Pénin. 2010. The limits of crowdsourcing inventive activities: What do transaction cost theory and the
evolutionary theories of the firm teach us. In Workshop on Open Source Innovation, Strasbourg, France. 1–26.
[9] A. I. Chittilappilly, L. Chen, and S. Amer-Yahia. 2016. A Survey of General-Purpose Crowdsourcing Techniques. IEEE Transactions on Knowledge
and Data Engineering 28, 9 (Sept. 2016), 2246–2266.
[10] Anne Edmundson, Brian Holtkamp, Emanuel Rivera, Matthew Finifter, Adrian Mettler, and David Wagner. 2013. An Empirical Study on the
Effectiveness of Security Code Review. In Engineering Secure Software and Systems. Springer, Berlin, Heidelberg, 197–212.
[11] Carsten Eickhoff and Arjen de Vries. 2011. How crowdsourcable is your task. In Proceedings of the workshop on crowdsourcing for search and data
mining (CSDM). 11–14.
[12] Enrique Estellés-Arolas and Fernando González-Ladrón-De-Guevara. 2012. Towards an integrated crowdsourcing definition. Journal of Information
science 38, 2 (2012), 189–200.
[13] Dennis Fisher. 2010. Microsoft Says No to Paying Bug Bounties. (July 2010). https://threatpost.com/microsoft-says-no-paying-bug-bounties-072210/
74249/
[14] Chris Grier, Lucas Ballard, Juan Caballero, Neha Chachra, Christian J Dietrich, Kirill Levchenko, Panayiotis Mavrommatis, Damon McCoy, Antonio
Nappa, Andreas Pitsillidis, and others. 2012. Manufacturing compromise: the emergence of exploit-as-a-service. In Proceedings of the 2012 ACM
conference on Computer and communications security. ACM, 821–832.
[15] Susan Halford, Catherine Pope, and Leslie Carr. 2010. A manifesto for Web Science. Journal of Web Science (2010).
[16] Andreas Kuehn and Milton Mueller. 2014. Analyzing bug bounty programs: An institutional perspective on the economics of software vulnerabilities.
In TPRC Research Conference on Communication, Information and Internet Policy.
[17] Andreas Kuehn and Milton Mueller. 2014. Shifts in the Cybersecurity Paradigm: Zero-Day Exploits, Discourse, and Emerging Institutions. In
Proceedings of the 2014 New Security Paradigms Workshop (NSPW ’14). ACM, New York, NY, USA, 63–68. DOI:http://dx.doi.org/10.1145/2683467.
2683473
[18] Aron Laszka, Mingyi Zhao, and Jens Grossklags. 2016. Banishing misaligned incentives for validating reports in bug-bounty platforms. In European
Symposium on Research in Computer Security. Springer, 161–178. http://link.springer.com/chapter/10.1007/978-3-319-45741-3_9
[19] Thomas D. LaToza, W. Ben Towne, Christian M. Adriano, and AndrÃľ van der Hoek. 2014. Microtask Programming: Building Software with a
Crowd. In Proceedings of the 27th Annual ACM Symposium on User Interface Software and Technology (UIST ’14). ACM, New York, NY, USA, 43–54.
DOI:http://dx.doi.org/10.1145/2642918.2647349
[20] Thomas D LaToza, W Ben Towne, André Van Der Hoek, and James D Herbsleb. 2013. Crowd development. In Cooperative and Human Aspects of
Software Engineering (CHASE), 2013 6th International Workshop on. IEEE, 85–88.
[21] Thomas D LaToza and André Van Der Hoek. 2015. A vision of crowd development. In Software Engineering (ICSE), 2015 IEEE/ACM 37th IEEE
International Conference on, Vol. 2. IEEE, 563–566.
[22] Gene Lin. 2016. Founder of China’s largest ’ethical hacking’ community arrested. (July 2016). https://www.hongkongfp.com/2016/07/30/
founder-chinas-largest-ethical-hacking-community-arrested/
[23] T Maillart, M Zhao, J Grossklags, and J Chuang. 2016. Given enough eyeballs, all bugs are shallow? Revisiting Eric Raymond with bug bounty
markets. (2016).
[24] Ke Mao, Licia Capra, Mark Harman, and Yue Jia. 2016. A survey of the use of crowdsourcing in software engineering. Journal of Systems and
Software (Sept. 2016).
[25] Winter Mason and Duncan J Watts. 2010. Financial incentives and the performance of crowds. ACM SigKDD Explorations Newsletter 11, 2 (2010),
100–108.
[26] Matthew Finifter, Devdatta Akhawe, and David Wagner. 2013. An Empirical Study of Vulnerability Rewards Programs. In Proceedings of the 22nd
USENIX Security Symposium. USENIX Association, Washington DC, 273–288.
[27] Charlie Miller. 2007. The legitimate vulnerability market: Inside the secretive world of 0-day exploit sales. In In Sixth Workshop on the Economics of
Information Security.
[28] Andy Ozment. 2004. Bug auctions: Vulnerability markets reconsidered. In Third Workshop on the Economics of Information Security. 19–26.
[29] Andy Ozment. 2005. The Likelihood of Vulnerability Rediscovery and the Social Utility of Vulnerability Hunting.. In WEIS. Citeseer. http:
//citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.479.7888&rep=rep1&type=pdf
[30] Sam Ransbotham, Sabyasachi Mitra, and Jon Ramsey. 2008. Are markets for vulnerabilities effective? ICIS 2008 Proceedings (2008), 24.
[31] E. Rescorla. 2005. Is finding security holes a good idea? IEEE Security Privacy 3, 1 (Jan. 2005), 14–19. DOI:http://dx.doi.org/10.1109/MSP.2005.17
[32] Tim Ring. 2014. Why bug hunters are coming in from the wild. Computer Fraud & Security 2014, 2 (Feb. 2014), 16–20.
[33] Tim Ring. 2015. White hats versus vendors: the fight goes on. Computer Fraud & Security 2015, 10 (Oct. 2015), 12–17.
[34] Ari Schwartz and Rob Knake. 2016. Government’s Role in Vulnerability Disclosure: Creating a Permanent and Accountable Vulnerability Equities
Process. Technical Report. Discussion Paper 2016-04, Cyber Security Project, Belfer Center for Science and International Affairs, Harvard Kennedy
School.
[35] Elena Simperl. 2015. How to Use Crowdsourcing Effectively: Guidelines and Examples. LIBER Quarterly 25, 1 (Aug. 2015).
[36] H. J. Su and J. Y. Pan. 2016. Crowdsourcing platform for collaboration management in vulnerability verification. In 2016 18th Asia-Pacific Network
Operations and Management Symposium (APNOMS). 1–4.