Algorithmic Transparency Via Quantitative Input Influence
Algorithmic Transparency Via Quantitative Input Influence
Input Influence
Abstract Algorithmic systems that employ machine learning are often opaque—
it is difficult to explain why a certain decision was made. We present a formal
foundation to improve the transparency of such decision-making systems. Specif-
ically, we introduce a family of Quantitative Input Influence (QII) measures that
capture the degree of input influence on system outputs. These measures provide a
foundation for the design of transparency reports that accompany system decisions
(e.g., explaining a specific credit decision) and for testing tools useful for internal
and external oversight (e.g., to detect algorithmic discrimination). Distinctively,
our causal QII measures carefully account for correlated inputs while measuring
influence. They support a general class of transparency queries and can, in
particular, explain decisions about individuals and groups. Finally, since single
inputs may not always have high influence, the QII measures also quantify the
joint influence of a set of inputs (e.g., age and income) on outcomes (e.g. loan
decisions) and the average marginal influence of individual inputs within such a
set (e.g., income) using principled aggregation measures, such as the Shapley value,
previously applied to measure influence in voting.
Abbreviation
Symbols
Influence
A Algorithm
QA Quantity of Interest
1 Introduction
Instead of releasing the entire source code of the decision making algorithm,
our proposed transparency report provides answers to personalized transparency
queries about the influence of various inputs (or features), such as race or recent
criminal history, on the system’s decision. These transparency reports can in turn be
studied by oversight agencies or the general public in order to gain much needed
insights into the use of sensitive user information (such as race or gender) by the
decision making algorithm; as we later discuss, transparency reports can also point
out potential systematic differences between groups of individuals (e.g. race-based
discrimination). Our transparency reports can also be used to identify harms and
errors in input data, and guide the design of better decision making algorithms.
Our Model We focus on a setting where a transparency report is generated with
black-box access to the decision-making system1 and knowledge of the input dataset
on which it operates. This setting models the kind of access available to a private or
public sector entity that proactively publishes transparency reports. It also models a
useful level of access required for internal or external oversight of such systems to
identify harms introduced by them. For the former use case, our approach provides
a basis for design of transparency mechanisms; for the latter, it provides a formal
basis for testing. Returning to our predictive policing example, the law enforcement
agency that employs such a system could proactively publish transparency reports,
and test the system for early detection of harms like race-based discrimination. An
oversight agency could also use transparency reports for post hoc identification of
harms.
Our Approach We formalize transparency reports by introducing a family of
Quantitative Input Influence (QII) measures that capture the degree of influence
of inputs on outputs of the system. Three desiderata drive the definitions of these
measures.
First, we seek a formalization of a general class of transparency reports that
allows us to answer many useful transparency queries related to input influence,
including but not limited to the example forms described above about the system’s
decisions about individuals and groups.
Second, we seek input influence measures that appropriately account for corre-
lated inputs—a common case for our target applications. For example, consider a
system that assists in hiring decisions for a moving company. Gender and the ability
to lift heavy weights are inputs to the system. They are positively correlated with
one another and with the hiring decisions. Yet transparency into whether the system
uses weight lifting ability or gender in making its decisions (and to what degree)
has substantive implications for determining if it is engaging in discrimination (the
business necessity defense could apply in the former case [16]). This observation
makes us look beyond correlation coefficients and other associative measures.
1
By “black-box access to the decision-making system” we mean a typical setting of software
testing with complete control of inputs to the system and full observability of the outputs.
74 A. Datta et al.
2 Unary QII
For QII to quantify the use of an input for individual outcomes, we define the
quantity of interest to be the classification outcome for a particular individual. Given
a particular individual x, we define Qxind ./ to be E.c./ D 1 j X D x/. The influence
measure is therefore:
When the quantity of interest is not the probability of positive classification but
the classification that x actually received, a slight modification of the above QII
measure is more appropriate:
Y
grp .i/ D E.c.X/ D 1 j X 2 Y / E.c.Xi Ui / D 1 j X 2 Y / (4)
E.c.X/ D 1 j X 2 Y /
E.c.X/ D 1 j X 62 Y /
QY
disp ./ D jE.c./ D 1 j X 2 Y / E.c./ D 1 j X 62 Y /j (5)
Y Y Y
disp .i/ D Qdisp .X/ Qdisp .Xi Ui /: (6)
3 Transparency Schema
The forms of transparency reports discussed so far are specific instances of a general
schema for algorithmic transparency via input influence that we describe now.
Recall that the influence of an input is the difference in a quantity of interest before
and after an intervention. This definition features the following three elements of
our transparency schema.
• A quantity of interest, which captures the aspect of the system we wish to gain
transparency into.
• An intervention distribution, which defines how the alternate distribution is
constructed from the true distribution.
• A difference measure, which quantifies the difference between two quantities of
interest.
The particular instantiation of elements in this schema is determined by the
question the analyst wishes to answer about the system. Table 1 contains some
examples of these instantiations and we describe the elements in detail below.
The first element, the quantity of interest, is determined by the subject of the
question. In the previous section, the systems considered were deterministic, and
the quantities of interest considered were represented by a number such as an
expectation or a probability. However, for more complex systems, the analyst may
be interested in richer quantities about the system. For instance, if the system is
randomized, a particular quantity of interest is the distribution over outcomes for a
given individual.
Algorithmic Transparency via Quantitative Input Influence 79
The second element is the causal intervention used to compute influence. The
particular randomized intervention considered in this paper, Xi Ui , where Ui is
drawn from the prior distribution of feature i, is suitable when no alternatives are
suggested in the analyst’s question. For example, the question “What is the influence
of age on Bob’s classification?” does not suggest any particular alternative age. On
the other hand, when a particular alternative is suggested, a constant intervention is
more suited. For example, for the question “Was age influential in Bob’s rejection
and Carl’s acceptance?” Carl’s age is the natural alternate to consider, and Ui is
drawn from the constant distribution corresponding to Carl’s age.
Finally, we use a difference measure to compare the quantity of interest before
and after the intervention. Apart from subtraction, considered above, examples of
other suitable difference measures are absolute difference, and distance measures
over distributions. The absolute difference measure can be employed when the
direction of change of outcomes before and after an intervention is unimportant.
When the quantity of interest is the distribution over outcomes, subtraction does
not make sense, and a suitable measure of difference should be employed, such
as the earthmover distance or the mutual information between the original and the
intervened distributions.
2
The adult dataset contains approximately 31k datapoints of users’ personal attributes, and
whether their income is more than $50k per annum; see Sect. 5 for more details.
80 A. Datta et al.
For most individuals, all features have zero influence: changing the state of one
feature alone is not likely to change the outcome of a classifier. Of the 19,537
datapoints we evaluate, more than half have x .i/ D 0 for all i 2 N; indeed, changes
to outcome are more likely to occur if we intervene on sets of features. In order
to better understand the influence of a feature i 2 N, we should measure its effect
when coupled with interventions on other features. We define the influence of a set
of inputs as a straightforward extension of individual input influence. Essentially,
we wish the influence of a set of inputs S N to be the same as when the set of
inputs is considered to be a single input; when intervening on S, we draw the states
of i 2 S based on the joint distribution of the states of features in S.
We define the random variable XS US representing an intervention on a set of
features instead of a single feature. XS US has the states of features in N n S fixed
to their original values in x, but features in S take on new values according the
joint marginal distribution over the features in S. This is identical to the intervention
considered in unary QII, when the features in S are viewed as a single composite
feature. Using this generalization of an intervention to sets of features, we can define
set QII that measures the influence of a set of features.
Definition 2 (Set QII) For a quantity of interest Q, and an input i, the Quantitative
Input Influence of set S N on Q is defined to be
Notice that marginal QII can also be viewed as a difference in set QIIs: Q .S [
fig/ Q .S/. Informally, the difference between Q .S [ fig/ and Q .S/ measures the
“added value” obtained by intervening on S [ fig, versus intervening on S alone.
The marginal contribution of i may vary significantly based on S. Thus, we are
interested in the aggregate marginal contribution of i to S, where S is sampled from
some natural distribution over subsets of N n fig. In what follows, we describe
a few measures for aggregating the marginal contribution of a feature i to sets,
based on different methods for sampling sets. The primary method of aggregating
the marginal contribution is the Shapley value [39]. The less theoretically inclined
reader can choose to proceed to Sect. 5 without a loss in continuity.
In this section, we discuss how measures from the theory of cooperative games
define measures for aggregating marginal influence. In particular, we observe that
the Shapley value [39] is uniquely characterized by axioms that are natural in our
setting. Definition 2 measures the influence that an intervention on a set of features
S N has on the outcome. One can naturally think of Set QII as a function v W
2N ! R, where v.S/ is the influence of S on the outcome. With this intuition in
mind, one can naturally study influence measures using cooperative game theory,
and in particular, using the Shapley value. The Shapley value can be thought of as
an influence aggregation method, which, given an influence measure v W 2N ! R,
outputs a vector 2 Rn , whose i-th coordinate corresponds in some natural way to
the aggregate influence, or aggregate causal effect, of feature i.
The original motivation for game-theoretic measures is revenue division [29,
Chapter 18]: the function v describes the amount of money that each subset of
players S N can generate; assuming that the set N generates a total revenue
of v.N/, how should v.N/ be divided amongst the players? A special case of
revenue division that has received significant attention is the measurement of voting
power [40]. In voting systems with multiple agents (think of political parties in a
parliament), whose weights differ, voting power often does not directly correspond
to agent weights. For example, the US presidential election can roughly be modeled
as a cooperative game where each state is an agent. The weight of a state is the
number of electors in that state (i.e., the number of votes it brings to the presidential
candidate who wins that state). Although states like California and Texas have
higher weight, swing states like Pennsylvania and Ohio tend to have higher power
in determining the outcome of elections.
82 A. Datta et al.
A voting system can be modeled as a cooperative game: players are voters, and
the value of a coalition S N is 1 if S can make a decision (e.g. pass a bill,
form a government, or perform a task), and is 0 otherwise. Note the similarity to
classification, with players being replaced by features (in classification, functions
of the form v W 2N ! f0; 1g are referred to as boolean functions [33]). The
game-theoretic measures of revenue division are a measure of voting power: how
much influence does player i have in the decision making process? Thus the notions
of voting power and revenue division fit naturally with our goals when defining
aggregate QII influence measures: in both settings, one is interested in measuring
the aggregate effect that a single element has, given the actions of subsets.
Several canonical influence measures rely on the fundamental notion of marginal
contribution. Given a player i and a set S N n fig, the marginal contribution of
i to S is denoted mi .S; v/ D v.S [ fig/ v.S/ (we simply write mi .S/ when v
is clear from the context). Marginal QII, as defined above, can be viewed as an
instance of a measure of marginal contribution. Given a permutation 2 ˘.N/ of
the elements in N, we define Pi . / D fj 2 N j .j/ < .i/g; this is the set of i’s
predecessors in . We can now similarly define the marginal contribution of i to
a permutation 2 ˘.N/ as mi . / D mi .Pi . //. Intuitively, one can think of the
players sequentially entering a room, according to some ordering ; the value mi . /
is the marginal contribution that i has to whoever is in the room when she enters it.
Generally speaking, game theoretic influence measures specify some reasonable
way of aggregating the marginal contributions of i to sets S N. In our setting,
we argue for the use of the Shapley value. Introduced by the late Lloyd Shapley,
the Shapley value is one of the most canonical methods of dividing revenue in
cooperative games. It is defined as follows:
1 X
'i .N; v/ D E Œmi . / D mi . /
nŠ
2˘.N/
Intuitively, the Shapley value describes the following process: players are sequen-
tially selected according to some randomly chosen order ; each player receives a
payment of mi . /. The Shapley value is the expected payment to the players under
this regime.
In this work, the Shapley value is our function of choice for aggregating marginal
feature influence. To justify our choice, we provide a brief exposition of axiomatic
game-theoretic value theory. We present a set of axioms that uniquely define the
Shapley value, and discuss why they are desirable in the QII setting.
Algorithmic Transparency via Quantitative Input Influence 83
5 Experimental Evaluation
We illustrate the utility of the QII framework by developing two simple machine
learning applications on real datasets. In Sect. 5.1, we illustrate the distinction
between different quantities of interest on which Unary QII can be computed. We
also illustrate the effect of discrimination on the QII measure. In Sect. 5.2, we
analyze transparency reports of three individuals to demonstrate how Marginal QII
can provide insights into individuals’ classification outcomes.
We use the following datasets in our experiments:
• adult[27]: This standard machine learning benchmark dataset is a subset of US
census data classifying individual income; it contains factors such as age, race,
gender, marital status and other socio-economic parameters. We use this dataset
to train a classifier that predicts the income of individuals from other parameters.
Such a classifier could potentially be used to assist in credit decisions.
84 A. Datta et al.
In Fig. 2, we illustrate the use of different Unary QII measures. Figure 2a, b show the
Average QII measure (Eq. (3)) computed for features of a decision forest classifier.
For the income classifier trained on the adult dataset, the feature with highest
influence is Marital Status, followed by Occupation, Relationship and Capital Gain.
Sensitive features such as Gender and Race have relatively lower influence. For the
predictive policing classifier trained on the arrests dataset, the most influential
input is Drug History, followed by Gender, and Smoking History. We observe that
influence on outcomes may be different from influence on group disparity.
QII on Group Disparity Figure 2c, d show influences of features on group disparity
(Eq. (6)) for two different settings. Figure 2c shows feature influence on group
disparity by Gender in the adult dataset; Fig. 2d shows the influence on group
disparity by Race in the arrests dataset. For the income classifier trained on
the adult dataset, we observe that most inputs have negative influence on group
disparity; randomly intervening on most inputs would lead to a reduction in group
disparity. In other words, a classifier that did not use these inputs would be fairer.
Interestingly, in this classifier, marital status—rather than gender—has the highest
influence on group disparity by gender.
For the arrests dataset, most inputs have the effect of increasing group
disparity if randomly intervened on. In particular, Drug history has the highest
positive influence on disparity in arrests. Although Drug history is correlated
with race, using it reduces disparate impact by race, i.e. makes fairer decisions.
Algorithmic Transparency via Quantitative Input Influence 85
(a) QII of inputs on Outcomes for the adult (b) QII of inputs on Outcomes for the arrests
dataset dataset
(c) QII of Inputs on Group Disparity by Gender (d) Influence on Group Disparity by Race in the
in the adult dataset arrests dataset
Fig. 2 QII measures for the adult and arrests datasets. (a) QII of inputs on outcomes for
the adult dataset, (b) QII of inputs on Outcomes for the arrests dataset, (c) QII of Inputs on
group disparity by gender in the adult dataset, (d) Influence on group disparity by race in the
arrests dataset
In both examples, features correlated with the sensitive attribute are the most
influential for group disparity according to the sensitive attribute rather than the
sensitive attribute itself. It is in this sense that QII measures can identify proxy
variables that cause associations between outcomes and sensitive attributes.
QII with Artificial Discrimination We simulate discrimination using an artificial
experiment. We first randomly assign “ZIP codes” to individuals in our dataset
(these are simply arbitrary numerical identifiers). Then, to simulate systematic bias,
we make an f fraction of the ZIP codes discriminatory in the following sense: all
individuals in the protected set are automatically assigned a negative classification
outcome. We then study the change in the influence of features as we increase f .
Figure 3a, shows that the influence of Gender increases almost linearly with f .
Recall that Marital Status was the most influential feature for this classifier without
any added discrimination. As f increases, the importance of Marital Status decreases
as expected, since the number of individuals for whom Marital Status is pivotal
decreases.
86 A. Datta et al.
(a) Change in QII of inputs as discrimination by (b) Change in QII of inputs as discrimination by
Zip Code increases in the adult dataset Zip Code increases in the arrests dataset
Fig. 3 The effect of discrimination on QII. (a) Change in QII of inputs as discrimination by Zip
code increases in the adult dataset, (b) Change in QII of inputs as discrimination by Zip code
increases in the arrests dataset
Age 23
Workclass Private
Education 11th
Education-Num 7
Marital Status Never-married
Occupation Craft-repair
Relationship Own-child
Race Asian-Pac-Islander
Gender Male
Capital Gain 14344
Capital Loss 0
Hours per week 40
Country Vietnam
Fig. 4 Mr. X’s profile and transparency report for negative classification
Age 27
Workclass Private
Education Preschool
Education-Num 1
Marital Status Married-civ-spouse
Occupation Farming-fishing
Relationship Other-relative
Race White
Gender Male
Capital Gain 41310
Capital Loss 0
Hours per week 24
Country Mexico
Fig. 5 Mr. Y’s profile and transparency report for negative classification
Mr. Z: The third example, who we refer to as Mr. Z (Fig. 6) is from the arrests
dataset. History of drug use and smoking are both strong indicators of arrests.
However, Mr. X received positive classification by this classifier even without
any history of drug use or smoking. Upon examining his classifier, it appears that
race, age and gender were most influential in determining his outcome. In other
words, the classifier that we train for this dataset (a decision forest) has picked
up on the correlations between race (Black), and age (born in 1984) to infer that
this individual is likely to engage in criminal activity. Indeed, our interventional
approach indicates that this is not a mere correlation effect: race is actively being
used by this classifier to determine outcomes. Of course, in this instance, we
have explicitly offered the race parameter to our classifier as a viable feature.
However, our influence measure is able to pick up on this fact, and alert us of the
problematic behavior of the underlying classifier. More generally, this example
illustrates a concern with the black box use of machine learning which can lead
to unfavorable outcomes for individuals.
88 A. Datta et al.
Fig. 6 Mr. Z’s profile and transparency report for positive classification
Due to the widespread and black box use of machine learning in aiding decision
making, there is a legitimate concern of algorithms introducing and perpetuating
social harms such as racial discrimination [4, 35]. As a result, the algorithmic
foundations of fairness in personal information processing systems have received
significant attention recently [10, 12, 15, 22, 48]. While many of the algorithmic
approaches [10, 17, 22, 48] have focused on group parity as a metric for achieving
fairness in classification, Dwork et al. [15] argue that group parity is insufficient as a
basis for fairness, and propose an approach which prescribes that similar individuals
should receive similar classification outcomes. However, this approach requires a
similarity metric for individuals which is often subjective and difficult to construct.
QII does not suggest any normative definition of fairness. Instead, we view QII as
a diagnostic tool to aid fine-grained fairness determinations. In fact, QII can be used
in the spirit of the similarity based definition of [15] by comparing the personalized
privacy reports of individuals who are perceived to be similar but received different
classification outcomes, and identifying the inputs which were used by the classifier
to provide different outcomes.
When group parity is used as a criteria for fairness, QII can identify the features
that lead to group disparity, thereby identifying features being used by a classifier
as a proxy for sensitive attributes. The determination of whether using certain
proxies for sensitive attributes is discriminatory is often a task-specific normative
judgment. For example, using standardized test scores (e.g., SAT scores) for college
admissions decisions is by and large accepted. However, SAT scores may act as
a proxy for several protected attributes, leading to concerns of unfairness [43, 45].
Our goal is not to provide such normative judgments. Rather we seek to provide fine-
grained transparency into input usage (e.g., what’s the extent to which SAT scores
influence decisions), which is useful to make determinations of discrimination from
a specific normative position.
Algorithmic Transparency via Quantitative Input Influence 89
7 Related Work
between inputs that may lead to leakage. The dual problem of transparency, on the
other hand, requires us to destroy correlations while analyzing the outcomes of a
system to identify the causal paths for information leakage. Measures of association
have also been widely used for detecting discrimination (see Sect. 6).
Proxy Influence An emerging direction in this space is the identification of proxy
or indirect use of sensitive features as opposed to direct causal use, as captured
by QII. Adler et al. [1] quantify the indirect influence of an attribute by obscuring
the attribute (along with associations) from a dataset and comparing the prediction
accuracy of a model before and after obscuring. In a different approach, Datta
et al. (Use privacy in data-driven systems: theory and experiments with machine
learnt programs, Unpublished Manuscript) identify intermediate computations in a
program that are associated with an attribute and use QII to measure the causal
influence of the intermediate computations on the outcome.
Interpretable Machine Learning An orthogonal approach to adding interpretability
to machine learning is to constrain the choice of models to those that are inter-
pretable by design. This can either proceed through regularization techniques such
as Lasso [44] that attempt to pick a small subset of the most important features, or
by using models that structurally match human reasoning such as Bayesian Rule
Lists [26], Supersparse Linear Integer Models [46], or Probabilistic Scaling [37].
Since the choice of models in this approach is restricted, a loss in predictive accuracy
is a concern, and therefore, the central focus in this line of work is the minimization
of the loss in accuracy while maintaining interpretability. On the other hand, our
approach to interpretability is forensic. We add interpretability to machine learning
models after they have been learnt. As a result, our approach does not constrain the
choice of models that can be used.
Experimentation on Web Services Another emerging body of work focuses on
systematic experimentation to enhance transparency into Web services such as
targeted advertising [3, 12, 18, 24, 25]. The setting in this line of work is different
since they have restricted access to the analytics systems through publicly available
interfaces. As a result they only have partial control of inputs, partial observability
of outputs, and little or no knowledge of input distributions. The intended use of
these experiments is to enable external oversight into Web services without any
cooperation. Our framework is more appropriate for a transparency mechanism
where an entity proactively publishes transparency reports for individuals and
groups. Our framework is also appropriate for use as an internal or external oversight
tool with access to mechanisms with control and knowledge of input distributions,
thereby forming a basis for testing.
Game-Theoretic Influence Measures Recent years have seen game-theoretic influ-
ence measures used in various settings. Datta et al. [13] also define a measure for
quantifying feature influence in classification tasks. Their measure does not account
for the prior on the data, nor does it use interventions that break correlations between
sets of features. In the terminology of this paper, the quantity of interest used by
Datta et al. [13] is the ability of changing the outcome by changing the state of
Algorithmic Transparency via Quantitative Input Influence 91
a feature. Strumbelj and Kononenko [42] also use the Shapley value to compute
influences for individual classification. This work greatly extends and generalizes
the concepts presented in [13] and [42], by both accounting for interventions
on sets, and by generalizing the notion of influence to include a wide range of
system behaviors, such as group disparity, group outcomes and individual outcomes.
Essentially, the instantiations of our transparency schema define a wide range of
transparency reports.
Game theoretic measures have been used by various research disciplines to
measure influence. Indeed, such measures are relevant whenever one is interested
in measuring the marginal contribution of variables, and when sets of variables are
able to cause some measurable effect. Prominent domains where game theoretic
measures have recently been used are terrorist networks [28, 30], protein inter-
actions [8], and neurophysical models [23]. The novelty in our use of the game
theoretic power indices lies in the conception of a cooperative game via a valuation
function .S/, defined by a randomized intervention on input S. Such an intervention
breaks correlations and allows us to compute marginal causal influences on a wide
range of system behaviors.
In this chapter, we present QII, a general family of metrics for quantifying the
influence of inputs in systems that process personal information. In particular,
QII lends insights into the behavior of opaque machine learning algorithms by
allowing us to answer a wide class of transparency queries ranging from influence
on individual causal outcomes to influence on disparate impact. To achieve this,
QII breaks correlations between inputs to allow causal reasoning, and computes
the marginal influence of inputs in situations where inputs cannot affect outcomes
alone. Also, we demonstrate that QII can be efficiently approximated, and can be
made differentially private with negligible noise addition in many cases.
We do not consider situations where inputs do not have well understood seman-
tics. Such situations arise often in settings such as image or speech recognition,
and automated video surveillance. With the proliferation of immense processing
power, complex machine learning models such as deep neural networks have
become ubiquitous in these domains. Similarly, we do not consider the problem of
measuring the influence of inputs that are not explicitly provided but inferred by the
model. Defining transparency and developing analysis techniques in such settings is
important future work.
92 A. Datta et al.
References
1. Adler, P., Falk, C., Friedler, S., Rybeck, G., Schedegger, C., Smith, B., Venkatasubramanian,
S.: Auditing black-box models for indirect influence. In: Proceedings of the 2016 IEEE
International Conference on Data Mining (ICDM), ICDM’16, pp. 339–348. IEEE Computer
Society, Washington (2016)
2. Alloway, T.: Big data: Credit where credit’s due (2015). http://www.ft.com/cms/s/0/7933792e-
a2e6-11e4-9c06-00144feab7de.html
3. Barford, P., Canadi, I., Krushevskaja, D., Ma, Q., Muthukrishnan, S.: Adscape: harvesting and
analyzing online display ads. In: Proceedings of the 23rd International Conference on World
Wide Web, WWW’14, pp. 597–608. ACM, New York (2014)
4. Barocas, S., Nissenbaum, H.: Big data’s end run around procedural privacy protections.
Commun. ACM 57(11), 31–33 (2014)
5. Big data in education (2015). https://www.edx.org/course/big-data-education-
teacherscollegex-bde1x
6. Big data in government, defense and homeland security 2015–2020 (2015). http://www.
prnewswire.com/news-releases/big-data-in-government-defense-and-homeland-security-
2015---2020.html
7. Bishop, C.M.: Pattern Recognition and Machine Learning (Information Science and Statistics).
Springer, New York (2006)
8. Bork, P., Jensen, L., von Mering, C., Ramani, A., Lee, I., Marcott, E.: Protein interaction
networks from yeast to human. Curr. Opin. Struct. Biol. 14(3), 292–299 (2004)
9. Breiman, L.: Random forests. Mach. Learn. 45(1), 5–32 (2001)
10. Calders, T., Verwer, S.: Three naive Bayes approaches for discrimination-free classification.
Data Min. Knowl. Disc. 21(2), 277–292 (2010)
11. Cover, T.M., Thomas, J.A.: Elements of Information Theory. Wiley, New York (2012)
12. Datta, A., Tschantz, M., Datta, A.: Automated experiments on ad privacy settings: a tale
of opacity, choice, and discrimination. In: Proceedings on Privacy Enhancing Technologies
(PoPETs 2015), pp. 92–112 (2015)
13. Datta, A., Datta, A., Procaccia, A., Zick, Y.: Influence in classification via cooperative game
theory. In: Proceedings of the 24th International Joint Conference on Artificial Intelligence
(IJCAI 2015), pp. 511–517 (2015)
14. Datta, A., Sen, S., Zick, Y.: Algorithmic transparency via quantitative input influence: theory
and experiments with learning systems. In: Proceedings of 37th Symposium on Security and
Privacy (Oakland 2016), pp. 598–617 (2016)
15. Dwork, C., Hardt, M., Pitassi, T., Reingold, O., Zemel, R.: Fairness through awareness. In:
Proceedings of the 3rd Innovations in Theoretical Computer Science Conference (ITCS 2012),
pp. 214–226 (2012)
16. E.G. Griggs v. Duke Power Co., 401 U.S. 424, 91 S. Ct. 849, 28 L. Ed. 2d 158 (1977)
17. Feldman, M., Friedler, S.A., Moeller, J., Scheidegger, C., Venkatasubramanian, S.: Certifying
and removing disparate impact. In: Proceedings of the 21th ACM SIGKDD International
Conference on Knowledge Discovery and Data Mining, KDD’15, pp. 259–268. ACM, New
York (2015)
18. Guha, S., Cheng, B., Francis, P.: Challenges in measuring online advertising systems. In:
Proceedings of the 10th ACM SIGCOMM Conference on Internet Measurement, IMC’10,
pp. 81–87. ACM, New York (2010)
19. Guyon, I., Elisseeff, A.: An introduction to variable and feature selection. J. Mach. Learn. Res.
3, 1157–1182 (2003)
20. Janzing, D., Balduzzi, D., Grosse-Wentrup, M., Schölkopf, B.: Quantifying causal influences.
Ann. Statist. 41(5), 2324–2358 (2013)
21. Jelveh, Z., Luca, M.: Towards diagnosing accuracy loss in discrimination-aware classification:
an application to predictive policing. In: Fairness, Accountability and Transparency in Machine
Learning, pp. 137–141 (2014)
Algorithmic Transparency via Quantitative Input Influence 93
22. Kamishima, T., Akaho, S., Sakuma, J.: Fairness-aware learning through regularization
approach. In: Proceedings of the 2011 IEEE 11th International Conference on Data Mining
Workshops (ICDMW 2011), pp. 643–650 (2011)
23. Keinan, A., Sandbank, B., Hilgetag, C., Meilijson, I., Ruppin, E.: Fair attribution of functional
contribution in artificial and biological networks. Neural Comput. 16(9), 1887–1915 (2004)
24. Lécuyer, M., Ducoffe, G., Lan, F., Papancea, A., Petsios, T., Spahn, R., Chaintreau, A.,
Geambasu, R.: Xray: enhancing the web’s transparency with differential correlation. In:
Proceedings of the 23rd USENIX Conference on Security Symposium, SEC’14, pp. 49–64.
USENIX Association, Berkeley (2014)
25. Lecuyer, M., Spahn, R., Spiliopolous, Y., Chaintreau, A., Geambasu, R., Hsu, D.: Sunlight:
Fine-grained targeting detection at scale with statistical confidence. In: Proceedings of the 22nd
ACM SIGSAC Conference on Computer and Communications Security, CCS’15, pp. 554–566.
ACM, New York (2015)
26. Letham, B., Rudin, C., McCormick, T.H., Madigan, D.: Interpretable classifiers using rules and
Bayesian analysis: building a better stroke prediction model. Ann. Appl. Stat. 9(3), 1350–1371
(2015)
27. Lichman, M.: UCI machine learning repository (2013). http://archive.ics.uci.edu/ml
28. Lindelauf, R., Hamers, H., Husslage, B.: Cooperative game theoretic centrality analysis of
terrorist networks: the cases of Jemaah Islamiyah and Al Qaeda. Eur. J. Oper. Res. 229(1),
230–238 (2013)
29. Maschler, M., Solan, E., Zamir, S.: Game Theory. Cambridge University Press, Cambridge
(2013)
30. Michalak, T., Rahwan, T., Szczepanski, P., Skibski, O., Narayanam, R., Wooldridge, M., Jen-
nings, N.: Computational analysis of connectivity games with applications to the investigation
of terrorist networks. In: Proceedings of the 23rd International Joint Conference on Artificial
Intelligence (IJCAI 2013), pp. 293–301 (2013)
31. Murdoch, T.B., Detsky, A.S.: The inevitable application of big data to health care. http://jama.
jamanetwork.com/article.aspx?articleid=1674245
32. National longitudinal surveys (2017). http://www.bls.gov/nls/
33. O’Donnell, R.: Analysis of Boolean Functions. Cambridge University Press, New York (2014)
34. Perry, W.L., McInnis, B., Price, C.C., Smith, S.C., Hollywood, J.S.: Predictive policing: the
role of crime forecasting in law enforcement operations. RAND Corporation, Santa Monica
(2013)
35. Podesta, J., Pritzker, P., Moniz, E., Holdern, J., Zients, J.: Big data: seizing opportunities,
preserving values. Technical Report, Executive Office of the President - the White House
(2014)
36. Rényi, A.: On measures of entropy and information. In: Proceedings of the Fourth Berkeley
Symposium on Mathematical Statistics and Probability, Volume 1: Contributions to the Theory
of Statistics, pp. 547–561. University of California Press, Berkeley (1961)
37. Rüping, S.: Learning interpretable models. Ph.D. Thesis, Dortmund University of Technology
(2006). Http://d-nb.info/997491736
38. Shannon, C.E.: A mathematical theory of communication. Bell Syst. Tech. J. 27(3), 379–423
(1948)
39. Shapley, L.: A value for n-person games. In: Contributions to the Theory of Games, vol. 2,
Annals of Mathematics Studies, No. 28, pp. 307–317. Princeton University Press, Princeton
(1953)
40. Shapley, L.S., Shubik, M.: A method for evaluating the distribution of power in a committee
system. Am. Polit. Sci. Rev. 48(3), 787–792 (1954)
41. Smith, G.: Quantifying information flow using min-entropy. In: Proceedings of the 8th
International Conference on Quantitative Evaluation of Systems (QEST 2011), pp. 159–167
(2011)
42. Strumbelj, E., Kononenko, I.: An efficient explanation of individual classifications using game
theory. J. Mach. Learn. Res. 11, 1–18 (2010)
94 A. Datta et al.
43. The National Center for Fair and Open Testing: 850+ colleges and universities that do not
use SAT/ACT scores to admit substantial numbers of students into bachelor degree programs
(2015). http://www.fairtest.org/university/optional
44. Tibshirani, R.: Regression shrinkage and selection via the lasso: a retrospective. J. R. Stat. Soc.
Ser. B 73(3), 273–282 (2011)
45. University, G.W.: Standardized test scores will be optional for GW applicants (2015). https://
gwtoday.gwu.edu/standardized-test-scores-will-be-optional-gw-applicants
46. Ustun, B., Tracà, S., Rudin, C.: Supersparse linear integer models for interpretable classifica-
tion. ArXiv e-prints (2013). http://arxiv.org/pdf/1306.5860v1
47. Young, H.: Monotonic solutions of cooperative games. Int. J. Game Theory 14(2), 65–72
(1985)
48. Zemel, R., Wu, Y., Swersky, K., Pitassi, T., Dwork, C.: Learning fair representations.
In: Proceedings of the 30th International Conference on Machine Learning (ICML 2013),
pp. 325–333 (2013)