Location via proxy:   [ UP ]  
[Report a bug]   [Manage cookies]                
HACKING BACK: OPTIMAL USE OF SELF-DEFENSE IN CYBERSPACE Jay P. Kesan* and Ruperto P. Majuca** *College of Law **Department of Economics University of Illinois at Urbana-Champaign 1. INTRODUCTION Internet security is a big problem. Several approaches have been suggested to deal with the problem, ranging from technological, to law-based, to economics-based solutions. 1 One approach emerging is the notion of self-help – using reasonable force in self-defense against hackers. 2 Hitherto, the law has not taken a clear position on whether or not counterstrike should be allowed in cyberspace. Here, we try to understand the optimality of hackback and articulate what the law on self-defense in cyberspace ought to be. In particular, we seek to answer the following questions: Should society permit hackback? How should the law on self-defense in cyberspace be designed? Which among the tools of combating cybercrimes – law enforcement, court litigation, hacking back the hacker – should be used to most effectively address cybercrimes? What optimal mix of these alternatives should be used to combat cyber-attacks? What role does technology play? One major argument for hackback is that traditional law enforcement schemes simply do not work in cyberspace because of the speed by which attacks cause great damage to 1 See generally Kesan and Majuca (2005) for a survey of these various approaches to Internet security. In real space, various instances of self-help have been recognized by the law, ranging from the use of reasonable force in self-defense or in defense of property in criminal law (see American Law Institute [1985], secs. 3.04 and 3.06), to recovery of property and summary abatement of nuisance in tort law, to repossession and commercial arbitration in commercial law, to the right of restraint and self-help eviction remedies in landlord-tenant relations (see Brandon et al. 1984), and even to such areas as the first amendment, trade secret law, copyright law, and patent law (see Lichtman 2005). 2 1 e-commerce sites and also because hackers can stage attacks from multiple jurisdictions with varying cybercrime laws and procedures for prosecuting Internet crimes. As Smith (2005), for example, point outs, while forensic investigation takes time, a virus or worm spreads quickly, underscoring the need to act right away in order to mitigate the grave damage that security incidents can cause. Epstein (2005) believes that even when legal remedies are available, self-help still plays a role because of the numerous instances when the judicial remedy is inadequate or too slow. Thus, Lichtman (2005) points out the importance of bringing together, and capitalizing on the interchangeability between, public and private means, particularly when legal remedies respond slowly to technological risks. However, some commentators also point out the potential dangers associated with hackback. Kerr (2005), for example, is concerned about counter-strikers hitting innocent third parties rather than the hacker since, in his view, it is easy to conceal the real source of the attack in the Internet. 3 Katyal (2005), on the other hand, argues that private selfhelp methods not only raise distributional issues (since the rich would be more able to afford themselves of the private measures than the poor), but also fragment the community spirit by weakening the connectivity between people. 4 He thus proposes methods toward community action against cybercrimes. 3 See also Himma (2004) (“For example, a set of zombies network could … have the direct effect of impairing the performance of the life-support system and hence could result in death of any number of innocent bystanders”). If such argument is correct, would driving a car be morally wrong because of a remote probability that the driver can fatally hit somebody? 4 In Katyal’s (2005) view, individual self-help can cripple interconnectivity and destroy reciprocity. In our view, however, the community spirit is in many circumstances already fragmented by other factors (such as, in the case of the Internet, the anonymity of the actors), and although individual self-help may contribute to the fragmentation, it may be not be entirely fair to withhold such option to individuals since it may be the rational response to opt for individual self-help remedies rather than to wait for the community to address the problem, especially when the community action is not forthcoming. 2 Those aforementioned papers articulate well the arguments for and against hackback. We believe, however, that there is room for the middle position that articulates in what situations hackback would be a good remedy and when it would not be. Thus, we differ from the more polar views of Kerr (2005) and Katyal (2005) and Smith (2005). Our position is closest to Epstein (2005) in that with him we think that neither blanket permission nor total prohibition of hackback is the right solution, and Lichtman (2005) in that we deem it important to use public means and private methods like self-help to combat cybercrimes. Unlike Epstein (2005) and Lichtman (2005) who do not lay down the criteria for the valid exercise of hackback however, here we actually formulate what are the requirements for a valid exercise of self-help in cyberspace, in the same manner that Posner (1971) proposed the conditions for the use of deadly force in real space. However, unlike Posner (1971), we employ formal modeling to generate our criteria. Hence, in this paper, we employ formal game theory to model the strategic interaction between the firm and the hacker. This allows us to study the behavior of the hacker given the effectiveness of law enforcement and the potential counter-actions of the firm, and vice versa, and also capture the interaction between law enforcement, court remedies, and self-help remedies. From the Nash equilibria that flow from the model, we observe that the firm will find police enforcement works best in certain instances, while in some cases, resort to the courts based on civil liability litigation will be the better approach, and in The building of community spirit is potentially a fruitful endeavor and thus needs to be pursued more, with the nuts and bolts of the proposal further tightened. (For instance, how does one deal with free riders and shirkers? How are groups formed? How are the responsibilities and costs allocated among the group members? How does one implement and enforce the obligation of each individual in the community?) We agree with Professor Katyal that community-based solutions are probably fruitful pursuits, but this need not entail that individual self-help should be banned outright. In fact, the two may very well go hand-in-hand, particularly in cyberspace where the quickness of the attacks, for instance, may entail that individuals should defend themselves until the community is able to act. 3 still other situations, self-defense and self-help will best address the cybercrimes problem. Furthermore, from the social planner’s perspective, we show that social welfare is higher when hackback is permitted in society versus when it is not. Also, by identifying the divergence between the private and the socially-optimal solutions, we are able to formulate regulations that are needed in order to bring the private solution closer to the socially-optimal outcome. Thus, explicit modeling enables us to develop litmus tests and criteria that determine if hackback is the proper remedy in certain cases, as well as formulate regulations governing proper conduct during hackback. The model results generate the following criteria for the valid exercise of self-defense in cyberspace: (1) accounting for trace back costs, the damage to the attacked firm’s (that is, the entity that is hacking back) systems that can be potentially mitigated outweigh the potential damage to third parties; (2) there is a relatively high chance of hitting the hacker, instead of innocent third parties; and (3) recourse to police enforcement or civilaction based litigation is either ineffective or impractical. The results also underscore the importance of using good technology (that is, intrusion detection systems (IDS), and trace back technology) in order for hackback to be effective as a deterrent against cyberattacks. When such criteria are satisfied, resort to hackback would be justified, and the rules governing proper conduct during counterstrike would come into play: (i) counter-strikers should not cause undue damage to the hacker’s computer systems and use only reasonable and proportionate means to defend themselves; and (ii) counter-strikers would be held liable for whatever damages may be suffered by innocent third parties caught in 4 the crossfire. These added regulations are necessary in order to move the firm’s Nash equilibrium outcome towards the socially optimal result. For example, making firms liable for third-party damages will cause them to internalize in their decision-making the potential damage to others and behave closer to the socially-optimal outcome. As it turns out, these conditions resemble the traditional formulation of the “just war” doctrine, 5 which requires the following necessary elements for a valid counterstrike: (1) there is grave damage (greater than the damage that might result from the action) that will be inflicted to the defender unless it counter-strikes, (2) there is a serious prospect of success, and (3) other means for stopping the evil are either impractical or ineffective (see United States Catholic Conference 1997, ¶ 2309). Interestingly, our requirement that counterstrikers should not wantonly damage the hacker’s system and use only necessary force echoes the classical authors’ position that war must not be waged for “revengeful cruelty” (Augustine 400, ¶ 74) and that only necessary and proportionate force ought to be used (Grotius 1625). Since our reasonableness conditions were generated from the social planner’s optimization of social welfare, they are consistent with the economic approach to tort law which balances the rights of firms seeking to mitigate damages to their systems and of third parties not being forced to suffer economic harm. 5 Aurelius Agustinus (354-430), generally acknowledged as the first to have articulated the “just war” doctrine, points out that war must be exercised by the sovereign (¶ 75), and must be waged in order to achieve peace and not for “love of violence, revengeful cruelty, fierce and implacable enmity, wild resistance, and the lust of power, and such like” (¶ 74) (Augustine 400, XXII, ¶¶ 73-79; see also Augustine 423, XIX, chap. 7). Aquinas (c.1271, II, II, Q.40, Art. 1) contributed to the discussion by identifying the three necessary elements for a war to be just: authority of the sovereign waging the war, just cause, and rightful intention. Hugo Grotius, generally known as the father of modern international law, articulated that a just war must contain these basic elements: immediate danger to the nation, necessity of the force employed used is necessary to adequately defend the nation's interests, and proportionality of the force employed to the threatened danger (DeForrest 1997, citing Grotius 1625). 5 Section 2 presents the basic model. In Section 3, we introduce intrusion detection system (IDS) into the model and examine the role of technology in deterring the hacker and the effectiveness of hackback. Section 4 considers the social planner’s perspective and analyzes the divergence between the private and social motive to engage in hackback. Section 5 discusses the proper liability rule for damages to innocent third parties. Based on the model results, Section 6 summarizes what the law of self-help in cyberspace should be. Section 7 concludes our discussions together with some final comments. 2. THE BASIC MODEL 2.1. The Model Set-up In this paper, we model the interaction between several measures – IDS and traceback (technology), criminal law enforcement and liability-based court litigation (legal remedies), and costs/benefits associated with hackback (economic incentives) – in order to determine how to optimally mix these methods to best address the cybercrimes problem. We start with the basic model of hacker and firm interaction when IDS is not available, and in the next section, we consider the role of IDS technology. The timing of the game is as follows: (A ) Not Hack, S i mpl yR eco ver Da mages (SR D) (A ) Not Hack, S i mpl yR eco ver Da mages (SR D) (A ) Not Hack, S i mpl yR eco ver Da mages (SR D) 0 1 2 Nature determines if an individual is a hacker or not Hacker decides whether to hack or not Firm decides whether to litigate, hack back or simply recover damages 6 The solution to this game was calculated by specifying the pay-off functions of the parties, solving the first-order conditions, and then calculating the Nash equilibria (see the Appendix for the game tree, the pay-offs and the Nash equilibrium calculations.) 2.2. Equilibrium When Police Enforcement is Effective Lemma 1. If the probability of catching the hacker times the magnitude of the penalty is bigger than what the hacker gains from hacking, the hacker will not hack (and there is no need for the firm to hack back or to litigate). 6 (This corresponds to area A of Figures 1 and 2 below.) Lemma 1 states that when the expected punishment exceeds the expected benefit to the hacker, cyberintrusions will be completely eliminated, and there is no need for the firm to resort to hackback or litigation. Effective cybercrime laws and police enforcement therefore act as a broad deterrent against cybercrimes. However, in the Internet where hackers can situate themselves in different and several jurisdictions with varying computer crime laws of several jurisdictions, the costs associated with the discovery and prosecution of hackers can be prohibitive, and hence, traditional law enforcement measures cannot be entirely relied upon to address cybercrimes. Thus, firms can provide additional deterrence by resorting to litigation and/or counterstrike. In Propositions 1 to 5, we tackle the optimal behavior of the firm when law enforcement is inadequate. 6 Proofs are presented in the Appendix. 7 2.3. Equilibrium When Litigation is Not Beneficial Proposition 1. When litigation is not beneficial, the following Nash equilibria obtain: Probability times Magnitude of Fine (A) Not Hack, Simply Recover Damages (SRD) [Hacker’s strategy, Firm’s strategy] (D) Mixed Hack, Mixed Hackback/SRD [Hacker’s strategy, Firm’s strategy] (B) Hack [Hacker’s strategy], SRD [Firm’s strategy] (C) Hack, Hackback [Hacker’s strategy, Firm’s strategy] Hackback’s Net Benefit-to-Cost Ratio Figure 1. Nash equilibria when litigation is not beneficial (No IDS available) When the net payoff from going to court (times the proportion of hackers) is lower than the trace costs, going to court would not be beneficial for the firm. In such a situation, when both police enforcement and litigation are ineffective or impractical, self-help can be useful in mitigating the damage to the firm’s e-assets. The firm’s first-order (optimality) condition (see Appendix) shows that the firm’s decision to hack back will crucially depend on the damage that can be mitigated to firm’s systems relative to those that can potentially be caused to third parties. The firm will find it advantageous to counter the attack only when it calculates that the damage that it can 8 mitigate by counter-striking would be considerably greater than the potential liability. Moreover, because of the discipline induced by the liability rule, the firm’s objective in hacking back will be limited to mitigating damages to its systems and due care will be exercised in order to lessen the damages inflicted to others. That is, the liability rule induces the firm to internalize the potential damages that it may cause others. Also, the net pay-off from hackback depends crucially on the probability of hitting the right person instead of innocent ones. Thus, from the firm’s perspective, the optimality of hackback depends on the available traceback technology: hackback will make sense only if the probability of successfully tracing and hitting the hacker exceed a certain threshold level. 7 If the probability is below this level, then active defense is not optimal. The firm will thus hack back only if there is “serious prospect of success”. In effect, our model shows that the concern of other authors about innocent parties getting hit (see, e.g., Himma 2004, Kerr 2004) is alleviated by liability rules. Liability for damages to innocent third parties causes the firm to strike back only if the probability of hitting the hacker (instead of innocent third parties) is large enough relative to the amounts of damages involved. The fact that the firm would be liable makes it cautious in calculating its chances of (and benefits from) success as compared to the potential liability. Figure 1 also shows that the propensity to hack back decreases with the effectiveness of law enforcement. Thus, increasing either the probability or the magnitude of the fine has two important effects. First, it reduces the hacker’s intrusion rate (compare, for example, the hacker’s equilibrium strategy for regions (C), (D) and (A)). Hence, an 7 This threshold level of probability depends on ratio of third-party damages to the sum of the mitigated damages and third-party damages. 9 investment in more police resources and international coordination of enforcement efforts, for example, can reduce hacking activity. Second, better police enforcement also reduces the firm’s propensity to hack back, as the firm perceives a higher level of protection that law enforcement affords (compare, for example, the firm’s strategy for regions (C), (D) and (A)). This emphasizes the substitutability between self-help and law enforcement. This also reduces the force of the argument that since hackers are anonymous in the Internet, hackback should be prohibited (see, for example, Kerr 2005). True, the fact that hackers can attack anonymously in the Internet could mean that innocent third parties may be caught in the crossfire. 8 But that fact, as well as the fact that hackers can situate themselves in different jurisdictions, also mean that cybercriminals are harder for the police to pin down thus lowering the efficacy of traditional police enforcement measures, and increasing the need for self-help measures to substitute in for the slack by providing additional remedy and deterrence against intrusions. Thus, equilibrium D illustrates the deterrent effect of hackback when law enforcement per se is not sufficient to completely eliminate cybercrimes. When effective police enforcement lies in region D, the hacker adopts a mixed hack strategy in response to the firm’s adopting a mixed hackback strategy. In contrast, in the same region of police enforcement, if the firm does not hack back, the hacker will definitely hack (region B). Thus, when the probability times the magnitude of the fine falls below the gain from hacking, self-help can supplement law enforcement because the hacker incorporates in his decision-making the potential damage a counterstrike can inflict on its system. 8 This concern though, as mentioned, could be handled by liability rules holding firms responsible for the consequences of their actions. 10 2.4. Equilibrium When Litigation is Beneficial Proposition 2. When litigation is beneficial, the following Nash equilibria obtain: Probability times Magnitude of Fine (A) Not Hack, Simply Recover Damages (SRD) [Hacker’s strategy, Firm’s strategy] (D) Mixed Hack, Mixed Hackback/SRD (B1) Mixed Hack, Mixed Litigate/SRD (C) Hack, Hackback (B2) Hack, Litigate Hackback’s Net Benefit-to-Cost Ratio Figure 2. Nash equilibria when litigation is beneficial (No IDS available) When the proportion of hackers times the net pay-off from litigation exceeds the trace costs, litigation is a beneficial option for the firm. In this case, the firm will choose between litigation or hackback, depending on whether the net pay-off from litigation exceeds that of hackback’s (equilibria B1 and B2 on the left side of Figure 2) or vice versa (equilibria C and D on the right). The firm will prefer to rely on active defense if its net pay-off exceeds that of litigation. Thus, if active defense will allow the firm to save further damage, or if there 11 are high transaction costs of going court, self-help will be the more effective remedy. When it is more beneficial for the firm to engage in self-help remedies (equilibrium C and D), the law should perhaps not compel the firm to go to court. However, self-help is not the better remedy in all situations. Equilibria B1 and B2 illustrate cases where the benefits from self-help are small compared to the more effective relief afforded by the courts. In these cases, self-help measures are not costeffective, and the courts should be provided as an alternative that the firm can resort to. This illustrates why there would still be a need for the legal system to provide formal legal protections since in certain instances, self-help remedies do not provide complete assistance (see Epstein 2005; Lichtman 2005). The law should thus permit hackback as an option, but not force it as a requirement (see Brandon et al. 1984, p. 870 et seq. for examples of judicially required self-help). 3. THE ROLE OF IDS TECHNOLOGY In this section, we study the effect of introducing IDS into the model and analyze the role that technology such as IDS can play against cyber-attacks. If the firm installs an IDS in its security architecture, the timing of the game is modified (see Figure A4 of the Appendix for the modified game tree). With this new IDS set-up, the firm has to consider strategy for two cases: (a) when the IDS signals an intrusion; and (b) when the IDS does not signal an intrusion. Proposition 3 below 12 presents the case where litigation is not beneficial regardless of whether the IDS signals an intrusion. 9 Proposition 3. When litigation is not beneficial irrespective of whether or not the IDS signals an intrusion, the following Bayesian Nash equilibria obtain: Probability times Magnitude of Fine (A) Not Hack, SRD, SRD [Hacker’s strategy, Firm’s strategy when IDS signals an intrusion, Firm’s strategy when IDS does not signal an intrusion] (D1) Mixed Hack, Mixed Hackback/SRD, SRD (B) Hack, SRD, SRD (C1) Mixed Hack, Hackback SRD (D2) Mixed Hack, Hackback, Mixed HB/SRD (C2) Hack, Hackback, Hackback Hackback’s Net Benefit-to-Cost Ratio Figure 3. Nash equilibria when litigation is not beneficial (IDS available) In general, the firm will counter-strike with a higher probability when the IDS signals an intrusion than when it does not. Thus, in region C1 of Figure 3, the firm hacks back when the IDS sends a signal and does not hack back when the IDS does not signal an 9 Proposition 4 (which covers situations where litigation is beneficial in both the signal and non-signal states) and Proposition 5 (which contemplates cases where litigation is beneficial when the IDS signals an intrusion, but not otherwise) are presented in the Appendix. 13 intrusion. Also, as Region D1, for example, illustrates, the probability of hacking back is greater when the IDS signals an intrusion than in the previous case when IDS was not available. So too, the firm hacks back with less frequency when the IDS does not signal an intrusion compared to the previous no-IDS case. The intuition behind these results is that with the IDS, the firm’s information as to the probability of intrusion is updated using Bayes’ rule (see the Appendix). The IDS thus enables the firm to have better information as to whether or not its systems are under attack. Better information in turn enables an organization to better identify an imminent danger so as to determine if the proper self-help response is a defensive one or a more pro-active one. The IDS thus allows the firm to fine-tune its strategy and be more efficient with its hackback/litigation response (see also Cavusoglu, Mishra, and Raghunathan 2005). The better the IDS configuration (that is, the lower the false positives and false negatives), the more efficient the firm’s hackback/litigation strategy will be, and hackback’s effectiveness as a cybercrime countermeasure increases. 4. SOCIALLY-OPTIMAL SOLUTION 4.1 The Social Planner’s Problem In Sections 3 and 4, we considered the (firm’s) private solution to the hackback game. Since the private and socially-optimal solutions can diverge, we need to consider the perspective of the social planner in order to analyze whether permitting hackback would be optimal for society or not. In contrast to the firm, the social planner takes into account the total pay-off to all members of society (the hacker, the firm, and the third party) (see equation A10 of the 14 Appendix for the social planner’s pay-off). From the social planner’s first-order conditions, we know that the social planner will never litigate because from the social planner’s perspective, litigation will result in court costs without any corresponding net social gain (since the amount awarded by the court constitute a mere transfer from the hacker to the firm). Consequently, for the social planner, litigation is never beneficial. Hence, its Nash equilibria would be: Probability times Magnitude of Fine (A) Not Hack, Simply Recover Damages (SRD) [Hacker’s strategy, Social planner’s strategy] (D) (B) Hack, SRD Mixed Hack, Mixed Hackback/SRD (C) Hack, Hackback Divergence of private and sociallyoptimal solutions Hackback’s Net Benefit-to-Cost Ratio Figure 4. Nash equilibria of the social planner’s problem By comparing the Nash equilibria of the social planner’s problem with that of the firm’s solution, we can see the divergence between the private incentive to hack back and the socially-optimal level of hackback. Thus, from the shaded region above, we can see that the firm has an incentive to engage in excessive hackback. This is because the firm does 15 not take into consideration the damage to the hacker’s systems in its decision to hack back while the social planner views such damages as losses to society. We thus think that hackback must be regulated in order to steer the firm’s behavior closer towards the socially-optimal solution. The law, for one, should require that in conducting hackback, firms must exert efforts not to wantonly destroy the digital assets of the hacker. In Section 6, we discuss how the law on self-defense in cyberspace should be written. 4.2. The Optimality of Hackback If hackback were to be for the good of society, it must be the case that the overall social welfare is higher when hackback is allowed, compared to the case when it is not. Hence, here we compare the societal welfare under both regimes by calculating the total societal pay-off across the different regions of the Nash equilibria under both regimes. Figure 5 below summarizes the social welfare comparisons between the two cases (see the Appendix for the details of the calculations). In region A, the two cases – (a) hackback available, and (b) hackback not available – have similar pay-offs. In region B, cases (a) and (b) again have similar pay-offs. In region C, societal pay-off is higher if hackback is available. Hence, hackback is “good” for society in region C. Finally, in region D, the net society’s pay-off is higher when hackback is available provided that the expected net social waste from hacking is positive. We thus conclude that, in general, the availability of hackback is beneficial to society. 16 Probability times Magnitude of Fine (A) Hackback has no effect on the overall social welfare (B) Hackback has no effect on the overall social welfare (D) Hackback increases social welfare if θ (1 − r ) D f > U − pe F (C) Hackback increases overall social welfare Hackback’s Net Benefit-to-Cost Ratio Figure 5. Hackback vs. no hackback social welfare comparisons 5. PROPER LIABILITY RULE FOR DAMAGES TO INNOCENT THIRD PARTIES We assumed in our model in Section 2 that counter-strikers are liable for third-party damages. If, instead, they are not, the Nash equilibria would be different, as shown below: 17 Probability times Magnitude of Fine (A) Not Hack, Simply Recover Damages (SRD) [Hacker’s strategy, Firm’s strategy] (D) (B) Hack SRD Mixed Hack, Mixed Hackback/SRD (C) Hack, Hackback No liability: Excessive Hackback Hackback’s Net Benefit-to-Cost Ratio Figure 6. Nash equilibria of the firm’s problem (No liability rule) Figure 6 shows that, not holding counterstrikers liable for third party damages would cause a distortion that compounds the wedge already caused by the firm not incorporating the damage to the hacker’s systems into its cost-benefit calculations (Figure 4). With both these distortions, the overall divergence of the private and socially-optimal solutions would be larger, as depicted in Figure 7. 18 Probability times Magnitude of Fine (A) Not Hack, Simply Recover Damages (SRD) [Hacker’s strategy, Social planner’s strategy] (D) (B) Hack, SRD Mixed Hack, Mixed Hackback/SRD (C) Hack, Hackback Total divergence of private and sociallyoptimal solutions Hackback’s Net Benefit-to-Cost Ratio Figure 7. Divergence of private and socially-optimal equilibria Thus, not holding firms liable for third-party damages would exacerbate the inefficiency of the laissez-faire hackback regime. Hence, there will now be two distortions that cause the private solution to diverge away from the socially-optimal solution: (a) the fact that the damage to the hacker’s system are social losses not considered by the firm (striped lines), and (b) the fact that the damage to third parties are social losses not internalized by active defenders (dotted lines). Thus, our model shows that liability rules for third-party damages function like an “invisible hand” guiding the private solution closer to the socially-optimal solution. On top of liability rules, however, regulations are still needed to handle the other distortion. Besides, even if third-party liability rules are present, several “frictions” could cause a wedge between the efficient amount of hackback and the actual amount hackback 19 (see Figure A10 in the Appendix for the Nash equilibria when these frictions are present). For example, several of those caught in the line of crossfire are likely to not have enough computer sophistication to even detect that their systems have been hit. 10 Also, others may decide not to sue given the transaction costs involved with going to court. 11 Because of these concerns, liability rules for damage caused to third parties, in and of themselves, may not be enough to generate the optimal outcome. Consequently, on top of having liability rules, regulations establishing criteria and guidelines for valid hackback should be set in order to constrain the parties closer to the socially-optimal outcome. 6. WHAT THE LAW ON SELF-DEFENSE IN CYBERSPACE SHOULD BE Given the results of the model, we are now ready to formulate what the law on selfdefense in cyberspace ought to be. First and foremost, we do not see any overriding reason why the law should outright deny firms the right to exercise self-defense in cyberspace (see, for example, equilibria C and D of Figure 5 where the availability of hackback increases the social welfare). With the time and expense associated with court-administered remedies, the availability of self-help could provide an equitable solution (Brandon et al. 1984, pp. 869-70). Moreover, by deterring criminals, active defense can supplement law enforcement: once precedents of hackers experiencing damage from counterstrike are established, a large number of script kiddies would vanish (see discussions in Section 2.3 on the deterrent 10 We thank George Deltas for bringing up this point. Also, the damage may be de minimis, in which case resort to courts may not be available (see Epstein 2005). 11 20 effect of hackback). In sum, we think that absent the showing of widespread misuse, active defense should not be outlawed at the outset (see also Epstein 2005). Secondly, counterstrikes can, however, function as a wrong against innocent third parties and what passes as self-defense may in reality be another wrong. Similarly, selfhelp though originally justified can bring about several harmful results.12 For these reasons, reasonableness standards must be instituted and resort to legal remedies prescribed when the planned counter-actions fall outside their boundaries (Epstein 2005). Thus, in our view, the law should in some instances allow (though not require) resort to self-help remedies, yet at the same time regulate the exercise of the privilege so as to check against its potential abuse. This is where we differ from Kerr (2005) and Katyal (2005) in that while we recognize that excesses and abuses can potentially occur, for us, this does not necessarily mean that the privilege of self-defense should be denied outright. Given the potential benefits self-help can generate when used responsibly, we think that regulating the exercise of the privilege is the best way to deal with these potential excesses. Based on our model results, the governing regulation on self-defense in cyberspace should have the following features: (1) Attacked firms and individuals can hack back if, and only if, the following requirements are satisfied: (a) Hackback does not result in greater harm to innocent parties compared to the damage to the defender’s systems that is sought to be mitigated. Furthermore, due care 12 The standard litany of criticisms of hackback include misidentification problems, use of automated program by counter-strikers, shooting matches between trigger-happy defenders and intruders, selfproclaimed “white hats” releasing worm patches with good intention but with terrible results, etc.. (see, e.g., Katyal 2005). 21 should be exercised to avoid or minimize damage to third parties and the purpose of the hackback should be limited to the prevention of damage to the firm’s information technology infrastructure. (b) Recourse to other alternatives is either ineffective or impractical. In particular, this occurs when: (i) police enforcement is ineffective. Lemmas 1 and 3 show that effective criminal law enforcement provides wide-ranging deterrence against cybercrimes and does away with the need for counterstrikes or civil litigation. (ii) litigation is impractical (see discussions in Sections 2.3 and 2.4). (iii) a more defensive strategy, such as simply recovering damages or simply dropping incoming packets, would not deter the hacker. In short, active defense is an extraordinary remedy, available only when other alternatives are ineffective or impractical. (c) There is a serious prospect of success. There must be a relatively high chance of hitting the hacker, instead of hitting innocent persons. Thus, reasonable effort must be exerted to employ state-of-the-art traceback technology. (d) Reasonable effort must be exerted to employ good IDS technology. This helps the firm to more carefully ascertain the existence or the imminence of the attack/danger; it also decreases the error of hitting innocent persons; and enhances the deterrent effect of hackback (see Section 3). If a firm hacks back without these conditions being present, it oversteps the bounds for reasonable exercise of self-defense in cyberspace. The law can hold those who exercise self-help not in a legally permissible manner, liable for penalties. 22 (2) Even if those preconditions are present – and thus the exercise of the privilege is justified – the conduct during hackback must also be regulated by the law: (a) In order to internalize the damage to third parties, active defenders should be held liable to third parties caught in the crossfire. Not holding active defenders responsible for the consequences of their action will result in externalities and excessive amount of hackback activity (see Figure 6). (b) Counter-strikers must also use only “proportionate force”, that is, they must not wantonly damage the hackers’ digital systems out of retaliation, but rather, only use force that is necessary to avoid damage to their own systems (see Figure 4). In sum, the law needs to layer liability rules on top of the reasonableness conditions. 7. CONCLUSION We believe that self-defense springs from the natural instinct for self-preservation. Hence, hackback should not be banned outright – it is generally accepted that one has the right to defend one’s self and one’s property and, toward this end, use reasonable force (see Aquinas c.1271, II, II, Q.64, Art. 7). The fact that the exercise of this right can be abused does not necessarily mean that the right should be denied at the outset; it does, however, mean that the exercise of the privilege should be regulated. In this paper, we formulated criteria and guidelines that articulate under what circumstances self-defense is proper in cyberspace, and in what situations should we instead rely on the police or resort to the courts. Using a game-theoretic model of the interaction between the defender and the hacker, we were able to capture the interplay 23 between legal remedies (police enforcement and court litigation), technology (IDS and traceback), and economic incentives (cost and benefits of self-help remedies), and thus develop specific rules or tests for resolving whether resort to hackback is justified vel non. Based on the results from the model, the criteria for valid resort to hackback are: (1) other alternatives, such as police enforcement and resort to courts, are either ineffective or ineffectual; (2) there is a genuine prospect of hitting the hacker instead of innocent third parties; and (3) the damage that can be mitigated to the defender’s systems outweigh the potential damage to third parties. Additionally, when hackback is justified, the following rules govern conduct during hackback: (4) defenders must not use excessive force, that is, they must only use force necessary to defend their property and not needlessly destroy the hacker’s digital assets; and (5) counter-strikers would be held liable for damage to other third parties. Thus, liability rules should be set in place so that firms will internalize the damage to third parties, thereby bringing the private incentive to hackback closer to the socially-optimal outcome. In sum, the law should layer the thirdparty liability rules on top of the reasonableness conditions mentioned above. 24 APPENDIX Summary of Notations U = hacker’s utility from hacking Df = firm’s damage due to hacking Dh= damage to the hacker if the firm hacks back F = fine the hacker pays if caught by law enforcers W = amount awarded the firm if it wins the litigation K = cost of going to court r = percentage of damage recovered if the firm decides to “simply recover damage” h = percentage of damage mitigated by resorting to hackback d = damage incurred by third parties (as percentage of the counter-striker’s damage) ph = probability of hitting the hacker if the firm hacks back pw = probability of winning the case if the firm litigates θ = proportion of hackers in the population π = fraction of affected third parties who forgo suit against the counterstriker q1= probability of getting an IDS signal given that there is an intrusion q2= probability of not getting an IDS signal given that there is no intrusion η1 = probability of intrusion given the IDS signals an intrusion η2 = probability of intrusion given the IDS does not signal an intrusion δ = probability that the hacker hacks σ1 = probability that the firm hacks back in the no IDS case σ2 = probability that the firm litigates in the no IDS case α1= probability that the firm hacks back if the IDS does not signal an intrusion α2= probability that the firm litigates if the IDS does not signal an intrusion β1= probability that the firm hacks back if the IDS signals an intrusion β2= probability that the firm litigates if the IDS signals an intrusion 25 The Model without Intrusion Detection System hack back (σ1) litigate (σ2) hack (δ) simply recover damages (1-σ1-σ2) hacker (θ) do not (1-δ) hack back (σ1) not (1-θ) litigate (σ2) simply recover damages (1-σ1-σ2) Figure A1. Game tree, no IDS case The total expected pay-off for the firm and the hacker, are respectively: ⎧σ 1 [ p h (1 − r − h) D f + (1 − p h )(1 − r + d ) D f ]⎫ ⎪ ⎪ F (σ 1 , σ 2 , δ ) ≡ −C t (σ 1 + σ 2 ) − θδ ⎨+ σ 2 [(1 − r ) D f + K − p wW ] ⎬ ⎪ ⎪ ⎩+ (1 − σ 1 − σ 2 )(1 − r ) D f ⎭ (A1) H (σ 1 , σ 2 , δ ) ≡ δ (U − pe F − σ 1 p h Dh − σ 2 p wW ) (A2) The first-order conditions are: ∂F (δ ) = −C t + θδ [ p h h − (1 − p h )d ]D f ∂σ 1 ∂F (δ ) = −C t + θδ [ p wW − K ] ∂σ 2 ∂H (σ 1 , σ 2 ) = U − p e F − σ 1 p h Dh − σ 2 p wW . ∂δ 26 (A3) (A4) (A5) Proof of Lemma 1. (A5) ⇒ ∂H < 0 ⇒ δ = 0 . Substituting, ∂δ ∂F ∂F (δ = 0) = −C t < 0 and (δ = 0) = −C t < 0 ⇒ σ 1 = σ 2 = 0 . ∂σ 1 ∂σ 2 Proof of Proposition 1. Probability times Magnitude of Fine, peF (A) δ = 0, σ 1 = 0, σ 2 = 0 U U − p h Dh (D) δ = (B) δ = 1, σ1 = 0, σ2 = 0 Ct U − pe F , σ1 = ,σ2 = 0 p h Dh θ [ p h h − (1 − p h )d ]D f (C) δ = 1, σ1 = 1, σ2 = 0 Hackback’s Net , [ p h h − (1 − p h )d ]D f Ct Benefit-to-Cost Ratio 1 θ Figure A2. Nash equilibria when litigation is not beneficial (No IDS available) C ∂F = − t + θ [ p wW − K ] < 0 ⇒ σ 2 = 0 in all equilibria here. δ ∂σ 2 − Ct ∂F Equilibrium B: (A3) ⇒ = + θ [ p h h − (1 − p h )d ]D f < 0 ⇒ σ 1 = 0 . Given δ ∂σ 1 ∂F σ 1 = σ 2 = 0 , H (σ 1 = σ 2 = 0) = δ (U − p e F ) . > 0 ⇒ δ = 1 . Equilibrium C: ∂σ 1 ∂F ∂H (A3) ⇒ > 0 ⇒ σ 1 = 1 . Given σ 1 = 1 and σ 2 = 0 , > 0 ⇒ δ = 1 . Equilibrium ∂σ 1 ∂δ θ [ p wW − K ] < C t => 27 D: We know by Nash (1950) that if U ≥ p e F > U − p h Dh and θ [ p h h − (1 − p h )d ]D f > C t , an equilibrium in mixed strategies exists. Setting Ct ∂F =0⇒ δ = . Given σ 2 = 0 , ∂σ 1 θ [ p h h − (1 − p h )]D f U − pe F ∂H = 0 ⇒ σ1 = . ∂δ p h Dh H (σ 2 = 0) = δ (U − p e F − σ 1 p h Dh ) . Hence, Proposition 2. Probability times Magnitude of Fine, peF (A) δ = 0, σ1 = 0, σ2 = 0 U (B1) Ct δ = , θ [ p wW − K ] U − p h Dh U − p wW σ 1 = 0, σ 2 = (D) δ = U − pe F p wW Ct U − pe F , σ1 = ,σ2 = 0 θ [ p h h − (1 − p h )d ]D f p h Dh (C) δ = 1, σ1 = 1, σ2 = 0 δ = 1, (B2) σ1 = 0, σ2 = 1 1 θ p wW − K Ct Hackback’s Net , [ p h h − (1 − p h )d ]D f Benefit-to-Cost Ratio Ct Figure A3. Nash equilibria when litigation is beneficial (No IDS available) The proof, which is omitted due to space constraints, is available from the authors upon request. 28 Model with Intrusion Detection System hack back (β1) litigate (β2) signal (q1) hack (δ) hacker (θ) simply recover damages (1-β1-β2) hack back (β1) no signal (1-q1) litigate (β2) simply recover damages (1-β1-β2) hack back (α1) do not (1-δ) signal (1-q2) litigate (α2) simply recover damages (1-α1-α2) not (1-θ) hack back (α1) no signal (q2) litigate (α2) simply recover damages (1-α1-α2) Figure A4. Game tree, IDS case Firm’s expected pay-off in the signal states: Fs ( β 1 , β 2 , δ ) ≡ −( β 1 + β 2 )C t −η 1 β 1 [ p h (1 − r − h) D f + (1 − p h )(1 − r + d ) D f ] − η1 β 2 [(1 − r ) D f + K − p wW ] − η1 (1 − β 1 − β 2 )(1 − r ) D f where η1 = Pr(intrusion|signal) = (A6) q1θδ . q1θδ + (1 − q 2 )(1 − θδ ) Firm’s expected pay-off in the non-signal states: Fn (α 1 , α 2 , δ ) ≡ −(α 1 + α 2 )C t −η 2 α 1 [ p h (1 − r − h) D f + (1 − p h )(1 − r + d ) D f ] − η 2α 2 [(1 − r ) D f + K − p wW ] − η 2 (1 − α 1 − α 2 )(1 − r ) D f where η2 = Pr(intrusion| no signal) = (1 − q1 )θδ . (1 − q1 )θδ + q 2 (1 − θδ ) Firm’s overall expected pay-off: 29 (A7) F (α 1 , α 2 , β 1 , β 2 , δ ) ≡ [q1θδ + (1 − q 2 )(1 − θδ )]Fs + [(1 − q1 )θδ + q 2 (1 − θδ )]Fn . (A8) Hacker’s expected pay-off: H (α 1 , α 2 , β1 , β 2 , δ ) = δ [U − pe F − (1 − q1 )(α 1 p h Dh + α 2 p wW ) − q1 ( β 1 p h Dh + β 2 p wW )]. (A9) [Note: Subsequent first-order conditions and proofs are omitted due to space limitations.] Lemma 2. (a) β 2 ≥ α 2 and (b) β 1 ≥ α 1 . Lemma 3. U < p e F ⇒ δ = 0, β 1 = β 2 = 0 . Proposition 3. When p wW − K q θ + (1 − q 2 )(1 − θ ) < a1 ≡ 1 , the following Bayesian q1θ Ct Nash equilibria obtain: Probability times Magnitude of Fine, peF (A) δ = 0, β1 = 0, α1 = 0, β2 = 0, α2 = 0 U (D1) δ = mixed , β 1 = mixed , α 1 = β 2 = α 2 = 0 U − q1 p h Dh U − p h Dh (B) δ = 1, β1 = 0, β2 = 0, α1 = 0, α2 = 0 (D2) δ = mixed , β 1 = 1, α 1 = mixed , β 2 = α 2 = 0 (C1) δ=mixed β1=1 α1 =α2= β2 = 0 a1 1 θ (C2) δ = α1 = β1 = 1, α2 =β2 = 0 Hackback’s Net , [ p h h − (1 − p h )d ]D f Ct Benefit-to-Cost Ratio a2 Figure A5. Nash equilibria when litigation is not beneficial (IDS available) 30 Proposition 4. When a1 < a 2 ≡ (1 − q1 )θ + q 2 (1 − θ ) p wW − K < , the following (1 − q1 )θ Ct Bayesian Nash equilibria obtain: Probability times Magnitude of Fine, peF (A) δ = 0, α1 = α2 = β1 = β2 = 0 U U − q1 p h Dh (B1) δ, β2 = mixed, α1= α2 =β1 = 0 (D1) δ, β1 = mixed, α1= α2 = β2 =0 (D2) δ, α1 = mixed, β1 = 1, α2 =β2= 0 U − q1 p wW U − p h Dh U − p wW (B2) δ, α2 = mixed, β2 = 1, α1 =β1 = 0 (C2) δ = α1 = β1 =1, α2 = β2 = 0 (C1) δ = α2 = β2 =1, α1 = β 1 = 0 a1 1 θ a2 p wW − K Ct Hackback’s Net , [ p h h − (1 − p h )d ]D f Ct Benefit-to-Cost Ratio Figure A6. Nash equilibria when litigation is beneficial (IDS available) Lemma 4. η1 > δ (given q1 , q 2 > 0.5 ) Proposition 5. When a1 < p wW − K < a 2 , the following Bayesian Nash equilibria Ct obtain: 31 Probability times Magnitude of Fine, peF (A) δ = 0, β1 = β2 = α1 = α2 = 0 U U − q1 p h Dh U − q1 p wW U − p h Dh (B1) δ, β2 = mixed, α1, α2, β1 = 0 (B2) δ = mixed β2=1, α1, α2, β1 =0 a1 (D1) δ, β1 = mixed, β2 = α1= α2 =0 (D2) δ, α1 = mixed, β1 = 1, β2= α2 =0 (C1) δ= mixed β1=1, α1, α2, β2 = 0 p wW − K Ct (C2) δ = 1 α1 =β1 =1, α2 =β2 = 0 a2 Hackback’s Net , [ p h h − (1 − p h )d ]D f Benefit-to-Cost Ratio C t Figure A7. Nash equilibria when litigation is beneficial when the IDS signals an intrusion, but not otherwise Socially-Optimal Solution If Hackback is Available Social planner’s problem: ⎧σ 1 [ p h (1 − r − h) D f + (1 − p h )(1 − r + d ) D f ]⎫ ⎪ ⎪ Max S = −C t (σ 1 + σ 2 ) − θδ ⎨+ σ 2 [(1 − r ) D f + K − p wW ] ⎬ ⎪ ⎪ ⎩(1 − σ 1 − σ 2 )(1 − r ) D f ⎭ (A10) + δ [U − p e F − σ 1 p h Dh − σ 2 p wW ] Hacker’s problem: Max δ [U − p e F − σ 1 p h Dh − σ 2 p wW ] (A11) 32 Proposition 6: Nash equilibria: Probability times Magnitude of Fine, peF (A) δ = 0, σ 1 = 0, σ 2 = 0 U (D) δ = (B) δ = 1, σ1 = 0, σ2 = 0 U − p h Dh U − p wW 1 θ Ct U − pe F ,σ1 = ,σ 2 = 0 θ [ p h h − (1 − p h )d ]D f − p h Dh p h Dh (C) δ = 1, σ1 = 1, σ2 = 0 1 ⎡ p h Dh ⎤ 1+ ⎥ Ct ⎦ θ ⎢⎣ Hackback’s Net , [ p h h − (1 − p h )d ]D f Benefit-to-Cost Ratio Ct Figure A8. Nash equilibria of the social planner’s problem, (No IDS available) If Hackback is Not Available Social planner’s problem: ⎧⎪σ 2 [(1 − r ) D f + K − p wW ]⎫⎪ Max S = −C t σ 2 − θδ ⎨ ⎬ + δ [U − pe F − σ 2 p wW ] ⎪⎩+ (1 − σ 2 )(1 − r ) D f ⎪⎭ (A12) Hacker’s problem: Max δ [U − p e F − σ 2 p wW ] (A13) Nash equilibria has simply 2 regions. 33 Social Welfare Comparisons (omitted due to space limitations) If the Firm is Not Held Liable for Third Party Damages Firm’s pay-off: ⎧σ 1 [ p h (1 − r ) D f − p h hD f ] ⎫ ⎪ ⎪ F = −C t (σ 1 + σ 2 ) − θδ ⎨+ σ 2 [(1 − r ) D f + K − p wW ]⎬ . ⎪ ⎪ ⎭ ⎩+ (1 − σ 1 − σ 2 )(1 − r ) D f (A14) Nash equilibria: Probability times Magnitude of Fine, peF (A) Not Hack, Simply Recover Damages (SRD) U (D) U − p h Dh (B) Hack SRD Mixed Hack, Mixed Hackback/SRD (C) Hack, Hackback No liability: Excessive Hackback 1 θ − (1 − p h )dD f 1 Ct θ Hackback’s Net Benefit-to-Cost Ratio Figure A9. Nash equilibria of the firm’s problem (No liability rule) 34 Frictions in the Application of the Third Party Liability Rule Due to frictions, only a fraction 1-π goes to court. Nash equilibria: Probability times Magnitude of Fine, peF (A) Not Hack, Simply Recover Damages (SRD) U Mixed Hack, Mixed Hackback/SRD (D) U − p h Dh (B) Hack SRD (C) Hack, Hackback Excessive Hackback Due to Frictions in Enforcing Liability Rules 1 θ − (1 − ph )πdD f 1 Ct θ Hackback’s Net Benefit-to-Cost Ratio Figure A10. Nash equilibria of the firm’s problem (Frictions in enforcing liability rules) 35 REFERENCES Aquinas, Thomas. c.1271. Summa Theologica. http://www.newadvent.org/summa/ American Law Institute. 1965. Restatement (Second) of Torts. Philadelphia: American Law Institute. American Law Institute. 1985. Model Penal Code and Commentaries (Official Draft and Revised Comments). Philadelphia: American Law Institute. Augustinus, Aurelius. 400. Contra Faustum Manichaeum. http://www.newadvent.org/fathers/1406.htm. Augustinus, Aurelius. c.423. De Civitate Dei. http://www.newadvent.org/fathers/120119.htm. Brandon, Douglas Ivor, Melinda Lee Cooper, Jeremy H. Greshin, Alvin Louis Harris, James M. Head Jr., Keith R. Jacques, and Lea Wiggins. 1984. Special Project, Selfhelp: Extrajudicial Rights, Privileges and Remedies in Contemporary American Society. Vanderbilt Law Review 37:845-1040. Cavusoglu, Huseyin, Birenda Mishra, and Srinivasan Raghunathan. 2005. The Value of Intrusion Detection Systems (IDSs) in Information Technology (IT) Security. Information Systems Research, forthcoming, March. DeForrest, Mark Edward. 1997. Just War Theory and the Recent U.S. Air Strikes Against Iraq. Gonzaga University, School of Law. http://law.gonzaga.edu/borders/documents/deforres.htm Epstein, Richard A. 2005. Self-help: From Stone Age to the Internet Age. Journal of Law, Economics, and Policy 1. Grotius, Hugo. 1625. Jure Belli ac Pacis ("On the Law of War and Peace"). http://www.geocities.com/Athens/Thebes/8098/ Himma, Kenneth Einar. 2004. Targeting the Innocent: Active Defense and the Moral Immunity of Innocent Persons from Aggression. Journal of Information, Communication, and Ethics in Society 2, no. 1. Katyal, Neal. 2005. Community Self-help. Journal of Law, Economics, and Policy 1 Kesan, Jay P., and Ruperto P. Majuca. 2005. Cybercrimes and Cyber-Attacks: Technological, Economic, and Law-Based Solutions. In Encyclopedia of Cybercrimes, edited by Pauline C. Reich. Forthcoming: Oceania Press. Lichtman, Douglas. 2005. How the Law Responds to Self-help. Journal of Law, Economics, and Policy 1. Nash, John. 1950. Equilibrium Points in n-Person Games. Proceedings of the National Academy of Sciences 36:48-49. Posner, Richard A. 1971. Killing or Wounding to Protect a Property Interest,” Journal of Law and Economics 14:201-32. Smith, Bruce P. 2005. Hacking, Poaching, and Counterattacking. Journal of Law, Economics, and Policy 1. United States Catholic Conference. 1997. Catechism of the Catholic Church. 2d ed. Washington, DC: USCC Publishing Services. 36