Breaking The Cyber-Security Dilemma: Aligning Security Needs and Removing Vulnerabilities
Breaking The Cyber-Security Dilemma: Aligning Security Needs and Removing Vulnerabilities
DOI 10.1007/s11948-014-9551-y
ORIGINAL PAPER
Received: 21 November 2013 / Accepted: 13 April 2014 / Published online: 30 April 2014
Springer Science+Business Media Dordrecht 2014
123
702 M. Dunn Cavelty
Introduction
Cyber-threats and the measures necessary to counter them are the security issue of
the hour. In recent years, a number of sophisticated cyber-attacks and intensifying
media attention have combined to give the impression that cyber-incidents are
becoming more frequent, more organised, more costly, and altogether more
dangerous. As a result, cyber-fears have percolated upwards, from the expert level
to executive decision-makers and politicians; and diffused horizontally, advancing
from mainly being an issue of relevance to the US to one that is at the top of the
threat list of more and more countries, resulting in a flurry of government-led and
private-led cyber-security initiatives.1
However, despite concerted efforts and increasing sums of money spent on
various aspects of cyber-security over the years, cyberspace does not seem to
become more secure—rather the opposite, considering the plethora of technical and
governmental reports that use the language of urgency and general doom.
Furthermore, the actions of some states convey an additional level of unease:
Though consolidated numbers are hard to come by, the amount of money spent on
defence-related aspects of cyber security is rising (Brito and Watkins 2011;
Boulanin 2013). Furthermore, an increasing number of states go (semi) public about
opening up ‘cyber-commands’, which are military units for (potentially offensive)
cyber war activities.
If we assume that more—rather than less—security in and through cyberspace is
one, if not the key goal of cyber-security policies, then the current approach to
cyber-security is not working. Worse, as I will show in this article, actions geared
towards gaining more security are (directly and indirectly) to blame for making both
the virtual but also, by implication, the real world less and not more secure. What
we seem to be facing is a ‘‘security dilemma’’, where efforts by one actor
(traditionally, states) to enhance its security decrease the security of others (Jervis
1978). Because cyber-capabilities cannot easily be divulged by normal intelligence
gathering activities, uncertainty and mistrust are on the rise. Although most states
still predominantly focus on cyber-defence issues, measures taken by some nations
are seen by others as covert signs of aggression by others and will likely fuel more
efforts to master ‘‘cyber-weapons’’ worldwide (Dunn Cavelty 2012; Rueter 2011).
That said, the cyber-security dilemma, like other security dilemmas before it,
extends to much more than just the security of and between states. In its basic form,
cyber-security signifies a multifaceted set of technologies, processes and practices
designed to protect networks, computers, programs and data from attack, damage or
unauthorized access. The related security discourse is about a diverse set of threat
forms, ranging from basic computer viruses to cyber-crime and cyber-espionage
activities, as well as cyber-terror and cyber-war. Each sub-issue is represented and
treated in a distinct way in the political process: Multiple actors employ different
political, private, societal, and corporate notions of security to mobilise (or de-
mobilise) different audiences (Stevens and Betz 2013; Dunn Cavelty 2008). In a less
1
Several governments have released or updated cyber-security or cyber-defense strategies in the last
several years. See http://www.ccdcoe.org/328.html for a good overview.
123
Breaking the Cyber-Security Dilemma 703
123
704 M. Dunn Cavelty
administrator, but is hardly ever a specific and direct referent object of security. The
threat to (national) security is presented as possible disruption to a specific way of
life—one building on information technologies, economic performance and
‘‘critical’’ functions of infrastructures—but the direct threat to human security,
especially a threat that undermines acquired values such as anonymity, privacy,
freedom of speech, free access to information, etc. does not figure prominently in
the policy discourse.
123
Breaking the Cyber-Security Dilemma 705
The selection of a ‘‘referent object’’ of security is closely interrelated with how the
threat is represented. As mentioned, some objects—commonly called infrastruc-
tures—and the functions they perform are regarded as ‘critical’ by the authorities (in
the sense of ‘vital’, ‘crucial’, ‘essential’) because their prolonged unavailability
harbours the potential for major crisis, both political and social (Burgess 2007). In
the mid-1990s, the issue of cyber-security was persuasively interlinked with this
topic of ‘‘critical infrastructures’’ and their necessary protection and in the process
made into a salient national security issue (PCCIP 1997). Because critical
infrastructures combine symbolic and instrumental values, attacking them becomes
integral to a modern logic of destruction (Coward 2009: 408f.) that seeks maximum
impact.
One classical goal of (national) security is to throw a ‘‘protective or preservative
measure […] around a valued subject or object’’ (Dillon and Lobo-Guerrero 2008:
276). Before this security can unfold, the valued subject/object needs to be identified
and also localized in space. In cyber-security linked to critical infrastructure
protection, the identification and designation of the protection-worthy is performed
following the well-established steps of (technical) risk analysis techniques, which
contains both an act of ‘‘naming’’ and an act of prioritizing. At the beginning of such
an analysis stands the identification of the assets (including services) that are
critical: Criticality is seen as a measure of the consequences associated with the loss
or degradation of a particular asset or object. Therefore, criticality needs a reference
point: it can only exist in relation to something (pre-)defined as important and
normal/desirable (Brunner et al. 2010).
Cyber-security linked to critical infrastructures creates and is implemented in a
special type of security environment. Whereas the traditional logic of national
security suggests unilateral government action and policy, the policies of cyber-
security are inevitably blurred by liberalization, domestic considerations and other
policy imperatives (Coaffee and Murakami Wood 2006). The management of
infrastructure is in general not (or no longer) the prerogative of government; instead
it is based on the logic of the market. While it remains the essential task of a
123
706 M. Dunn Cavelty
123
Breaking the Cyber-Security Dilemma 707
field also seems to be an effect of the issue that human security scholarship has
already tackled decades ago: that too much focus on the state and national security
tends to crowd out consideration for the individual citizen, with often detrimental
effects for security overall (cf. Burgess and Owen 2004). I look at both aspects and
their consequences for security below and then turn to the clash between this type of
security and human security.
A focus on technical objects is not a bad thing per-se. In fact, the type of security
that emerges directly from the wish to ensure cyber-security is one that seemingly
dodges problematic issues normally associated with security, at least in the first
instance. Ultimately, we are looking at the practice of protecting inanimate things;
the regulation of machines and their performance. Computers, servers, and the
computer-powered infrastructures are non-human objects, which are someone’s
legitimate property and have a certain (usually undisputed) value for societies.
Cyber-security measures thus imagined have little to no bearing on citizens’ lives
directly. Most importantly, there are no concerns about freedom/security trade-offs,
and no civil liberty issues (Buzan et al. 1998). This security does not depend upon
the invocation of a state of emergency, but is ‘clean’ and ultimately, ‘good’, since
everybody seems to benefit from an interruption-free performance of vital systems.
However, this view is inevitably problematized, because these machines cannot
be isolated from human life. The image of modern complex critical infrastructures is
one in which it becomes futile to try and separate the human from the technological.
Technology is not simply a tool that makes life livable: rather, technologies become
constitutive of novel forms of ‘a complex subjectivity’, which is characterized by an
inseparable ensemble of material and human elements (Coward 2009: 414).
Therefore, even if technologies may appear to regulate objectively and apolitically,
there is always a connection to a place, to a space, to a space of protection, to values,
to life. An even closer look at the seemingly apolitical management of a technical
issue with technical means reveals a deeply political nature, because the selection of
referent objects as described above always entails a larger argument about
protection: Endangered entities are judged to have legitimate claims to protection
(while others do not). In other words, this type of security will only provide relief to
a valued referent object—not necessarily ‘‘the citizen’’ or humans more generally.
In cyber-security, as argued above, economic imperatives like profit maximiza-
tion are decisive. It is not a given, then, that cyber-security is a truly public good,
understood as security for all. Quite the opposite: the type of security that emerges
mainly benefits a few and already powerful entities and has no or even negative
effects for the rest. The type of referent object to be protected and by implication,
the type of life to be saved, is represented by the uninterrupted flow of information
linked to the accumulation of capital and economic growth (Swyngedouw 2007),
which in turn is linked to national security. This is at the heart of the cyber-security
dilemma, in which the dominant form of security is making large parts of the
population arguably less secure. Various security needs are not aligned; and while
they do not always have to be, more awareness of the clash between them is needed.
123
708 M. Dunn Cavelty
Referent objects also reveal a lot about (hidden) power structures. Contrary to the
beautiful utopia of cyber-libertarians like Barlow (1996), who saw cyberspace as a
serious challenge to traditional state power, the dystopian reality is more like a
‘‘feudal power structure’’ that consolidates power in the hands of the few (Schneier
2012a). Even though the cyber-realm has challenged us to think about power
differently, the most power rests with a few IT companies that act with little
restraint in their own self-interest, often changing social norms by accident or
deliberately, at all times using ‘‘the users’’ to increase their profits. At the same time,
states are asserting their power positions rather forcefully (Schneier 2013), mostly in
the name of security.
Assertion of state power is linked to the possibility (and desirability) to create
borders in cyberspace, which results in a changing topology of cyberspace as we
know it (Mueller et al. 2013). Prominent concepts like ‘‘Cyber-Westphalia’’ tap into
the founding myths of a stable political world order based on state power and invoke
images of a delimited and thus defendable and securable place, newly reordered by
the state as the real guarantor of security (Demchak and Dombrowski 2011). In this
view, held by many government actors, the process of re-establishing state control
in cyberspace is inevitable, because security is the most basic need of human beings
and seeking security will triumph over other, lesser, inferior needs (such as privacy).
Furthermore, the more the issue is presented like a traditional national security
issue, the more natural it seems that the keeper of the peace in cyberspace should be
the military, and the most relevant concepts are cyber-defense, cyber-deterrence,
etc. However, actions by military actors with relation to cyberspace directly fuel the
cyber-security dilemma as we have seen.
Of course, there is a certain appeal to a vision in which the unruly, anarchical and
dangerous side of cyberspace is kept ‘‘outside’’, and relative security can be
established among states. However, this image simplifies complex matters in an
unbeneficial way: Not only does inside-outside generally not apply easily to
cyberspace, state control also often means control over information flows: Indeed,
an increasing number of governments are already controlling what their citizens can
and cannot do on the Internet. Totalitarian governments are embracing a growing
‘‘cyber-sovereignty’’ movement to further consolidate their power. But democratic
states are doing very similar things: There is more government surveillance, more
government censorship, and more government propaganda than ever before (Deibert
2013; Wagner 2014).
State controlled borders in cyberspace would in most cases amount to (at least
partial) governmental control over information flows. Certainly, this does not mean
that all states would start misusing this power, but trust in their benign intent with
regards to civil liberties, most notably privacy, has taken a serious hit last year with
Edward Snowden’s NSA revelations. Most notably, the NSA scandal has focused
123
Breaking the Cyber-Security Dilemma 709
attention on the fact that there are direct human security implications arising from
mass surveillance in the name of national security.
In this day and age, more and more user or system specific data is up for grabs—
for anybody who is interested in it, ranging from business, to criminals, and the
intelligence services. While just the extensive data collection by companies and
intelligence agencies is already cause for concern, the consequences of this for
human security becomes fully apparent when the possibilities of its analysis are
taken into account. With a relatively simple network analysis, detailed insight into
the private lives and relationships of each individual can be gained. More
sophisticated methods of calculation are less interested in the present but are geared
towards the prediction of future behaviour (and motivations) of people (cf. McCue
2007). Such techniques are already used for targeted advertising, whereby an
algorithm defines that if Person X buys this or that product, it is very likely that X is
also interested in this or that product. In predictive policing, similar techniques are
used to calculate crime hot spots (Perry et al. 2013). A goal of intelligence services
is to be able to have advance warning of i.e. radicalization or terrorist behaviour,
based on data combination that could look like this: If Person X visit this website
and that website, is in contact with this and that person and has this specific motion
profile, then it is likely that Person X will commit a terrorist attack in the next
2 years.
From a data protection perspective, these developments are daunting, particularly
because the so-called commercialization of data is not done against the wishes of
the user, but rather because it seems to make our lives so much more efficient and
convenient. Sure, targeted advertising is at best intrusive and does not yet constitute
a human security threat. However, much more unpleasant implications of individual
risk profiles are already apparent today, with people being excluded from certain
services, because aspects of their (private) life does not meet the requirements of a
company (Amoore and de Goede 2005). In the future, it is not unlikely that even
more unpleasant and more directly political relevant implications arise when
democratic rights, such as political resistance or dissidence, are seen as an
opportunity for government intervention in the sense of ‘‘proactive security’’ (i.e. at
airports).
Add to these developments a fantasy about a version of cyberspace in which
crime or even attacks by state actors become impossible or at least very hard. Given
that the prime issue for traditional law enforcement methods like punishment or
well-proven military tools like deterrence is the ‘‘attribution problem’’ (the difficulty
of clearly identifying those initially responsible for a cyber-attack), and given that
the attribution problem arises from technological protocols that guarantee a great
deal of anonymity for its users, taking away said anonymity, in parts or fully, is
sometimes seen as one of the best solutions for a secure internet of the future (cf.
CSIS 2008: 61ff.). Here, the clash of different types of security becomes directly
visible. From a human and political rights perspective, anonymity is not a threat to
security, it is a crucial part of it. An Internet without the attribution problem, which
would most likely have a negligible effect on security overall, would introduce a
new issue: citizens could be readily identified and punished for their political
activities (Zittrain 2011).
123
710 M. Dunn Cavelty
123
Breaking the Cyber-Security Dilemma 711
flood of official documents lamenting the security-deficit. Second, the type of cyber-
security that is produced is based on economic maxims, often without consideration
for the particular security-needs of the population. Third, extending a notion of
national security based on border control to cyberspace will almost inevitably have
an impact on civil liberties, especially on the right to privacy and the freedom of
speech. Fourth, cyber-exploitation by intelligence agencies linked to the manipu-
lation of vulnerabilities is directly making cyber-space more insecure. What
becomes exceedingly clear from the developments and lessons of the last decade is
that we cannot have both: a strategically exploitable cyberspace full of vulnera-
bilities—and a secure and resilient cyberspace that all the cyber-security policies
call for.
At the heart of this challenge is, as so often when human security is implicated,
the state (cf. Kerr 2007). On the one hand, state practices are emerging as a major
part of the problem, constantly creating more insecurity and in fact also hindering
the removal of known insecurities. At the same time, a secure, safe, and open
cyberspace is not possible without involvement of the state. How, then, can this
dilemma be overcome? Because it is a dilemma extending to more than the state,
solutions are not to be found solely in the cooperation between states (cf. Booth and
Wheeler 2008). Rather, a focus on a common issue of interest for all the
stakeholders that are interested in more security is needed. Such a common ground
is held by vulnerabilities.
If we want a secure and resilient cyberspace, then a strategically exploitable
cyberspace full of vulnerabilities has to be actively worked against. This is a
compromise that some state actors need to make if they want a type of national
security that extends to cyberspace. If such a compromise is not made, then the
quest for more national security will always mean less cyber-security, which will
always mean less national security because of vulnerabilities in critical infrastruc-
tures. The reason why vulnerabilities persist and even proliferate has already been
identified above: the current incentive structures in the market are skewed (Dynes
et al. 2008). This is where states are needed to help improve cyber-security through
additional regulation (and through further encouragement of voluntary arrangement
for the increase of cyber-security in the corporate sector). Furthermore, there is no
doubt from a human security perspective that the zero-day exploit ‘‘market’’ needs
to be regulated internationally for security reasons (Kuehn 2013). In addition, prime
human security concerns like the freedom of speech and the right to privacy should
no longer be seen as anti-security, but as pro-security if linked to vulnerabilities:
reducing the amount of data that is unencrypted will substantially reduce cyber-
crime and cyber-espionage, with benefits for both human-centred and state-centred
security.
In turn, the ethics that should guide our future engagement with cyber-security
have to take into account the special and all-embracing characteristics of
cyberspace. So far, ethical considerations with bearing on cyber-security have
mainly been made from a military perspective, following the tradition to address
new forms of warfare and weapons systems under ethical viewpoints (cf. Rowe
2010; Dipert 2010; Barrett 2013). Cyber-security, as argued in the very beginning, is
far more than this, however: From both a state and a human security perspective,
123
712 M. Dunn Cavelty
References
Amoore, L., & De Goede, M. (2005). Governance, risk and dataveillance in the war on terror. Crime, Law
and Social Change, 43(2–3), 149–173.
Anderson, B. (2010). Preemption, precaution, preparedness: Anticipatory action and future geographies.
Progress in Human Geography, 34(6), 777–798.
Anderson, R., & Moore, T. (2006). The economics of information security. Science, 314, 610–623.
Axworthy, L. (2001). Human security and global governance: Putting people first. Global Governance,
7(1), 19–24.
Barlow, J. P. (1996). A declaration of the independence of cyberspace, electronic frontier foundation
website. http://homes.eff.org/*barlow/Declaration-Final.html.
Barnard-Wills, D., & Ashenden, D. (2012). Securing virtual space: Cyber war, cyber terror, and risk.
Space and Culture, 15(2), 110–123.
Barrett, E. T. (2013). Warfare in a new domain: The ethics of military cyber-operations. Journal of
Military Ethics, 12(1), 4–17.
Böhme, R. (2005). Vulnerability markets—What is the economic value of a zero-day exploit? Paper held
at the 2005 Chaos Communication Congress Berlin, Germany. http://events.ccc.de/congress/2005/
fahrplan/attachments/542-Boehme2005_22C3_VulnerabilityMarkets.pdf.
Booth, K., & Wheeler, N. (2008). The security dilemma: Fear, cooperation and trust in world politics.
New York: Palgrave.
Boulanin, V. (2013). Cybersecurity and the arms industry. SIPRI Yearbook 2013: Armaments,
disarmament and international security (pp. 218–226). Oxford: Oxford University Press.
Brito, J., & Watkins, T. (2011). Loving the cyber bomb? The dangers of threat inflation in cybersecurity
policy. Mercatus Center George Mason University, Working Paper No. 11-24, April 2011.
123
Breaking the Cyber-Security Dilemma 713
Brunner, E., Dunn Cavelty, M., Giroux, J., & Suter, M. (2010). Protection goals. Focal report on Critical
Infrastructure Protection for the Federal Office for Civil Protection, No. 4. Zurich: Center for
Security Studies.
Burgess, J. P. (2007). Social values and material threat: The European Programme for Critical
Infrastructure Protection. International Journal of Critical Infrastructures, 3(3–4), 471–487.
Burgess, J. P. & Owen, T. (Eds.) (2004). Special section: What is ‘human security’?, Security Dialogue,
35(3), 345–346.
Buzan, B., Wæver, O., & de Wilde, J. (1998). Security: A new framework for analysis. Boulder: Lynne
Rienner.
Capurro, R. (2006). Towards an ontological foundation of information ethics. Ethics and Information
Technology, 8(4), 175–186.
Clarke, R. A., Morell, M. J., Stone, G. R., Sunstein, C. R., & Swire, P. (2013). Liberty and security in a
changing world: Report and Recommendations of The President’s Review Group on Intelligence
and Communications Technologies. Washington, DC. http://www.whitehouse.gov/sites/default/
files/docs/2013-12-12_rg_final_report.pdf.
Coaffee, J., & Murakami Wood, D. (2006). Security is coming home: Rethinking scale and constructing
resilience in the global urban response to terrorist risk. International Relations, 20(4), 503–517.
Collier, S. J. & Lakoff, A. (2008). The vulnerability of vital systems: How critical infrastructure became a
security problem. In M. Dunn Cavelty & K. S. Kristensen (Eds.), The politics of securing the
homeland: Critical infrastructure, risk and securitization (pp. 17–39). New York: Routledge.
Conway, M. (2008). The media and cyberterrorism: A study in the construction of ‘reality. In M. Dunn
Cavelty & K.S. Kristensen (Eds.), The politics of securing the homeland: Critical infrastructure, risk
and securitisation (pp. 109–129). London: Routledge.
Coward, M. (2009). Network-centric violence, critical infrastructure and the urbanization of security.
Security Dialogue, 40(4–5), 399–418.
CSIS Center for Strategic and International Studies (2008). Securing Cyberspace for the 44th Presidency
A Report of the CSIS Commission on Cybersecurity for the 44th Presidency. Washington, DC.
http://csis.org/files/media/csis/pubs/081208_securingcyberspace_44.pdf.
Deibert, R. J. (2013). Black code: Inside the battle for cyberspace. Toronto: McClelland & Stewart.
Deibert, R. J., Palfrey, J. G., Rohozinski, R., & Zittrain, J. (2008). The practice and policy of global
internet filtering. Cambridge: MIT Press.
Demchak, C. & Dombrowski, P. (2011). Rise of a cybered westphalian age. Strategic Studies Quarterly,
Spring, pp. 32–61.
Der Derian, J. & Finkelstein, J. (2008). Critical infrastructures and network pathologies: The semiotics
and biopolitics of heteropolarity. In M. Dunn Cavelty & K. S. Kristensen (Eds.), The politics of
securing the homeland: critical infrastructure, risk and securitisation (pp. 84–105). London:
Routledge.
Dillon, M., & Lobo-Guerrero, L. (2008). Biopolitics of security in the 21st century: An introduction.
Review of International Studies, 34(2), 265–292.
Dipert, R. R. (2010). The ethics of cyberwarfare. Journal of Military Ethics, 9(4), 384–410.
Dunn Cavelty, M. (2008). Cyber-security and threat politics: US efforts to secure the information age.
London: Routledge.
Dunn Cavelty, M. (2012). Militarizing cyberspace: Why less may be better. In C. Czosseck, R. Ottis, &
K. Ziolkowski (Eds.), Proceedings of the 4th International Conference on cyber conflict (pp.
141–153). Tallinn: CCD COE Publications.
Dunn Cavelty, M. & Kristensen, K.S. (2008). Introduction: Securing the homeland: Critical infrastruc-
ture, risk, and (in)security. In M. Dunn Cavelty & K. S. Kristensen (Eds.), The politics of securing
the homeland: Critical infrastructure, risk and securitization (pp. 1–14). New York: Routledge.
Dunn Cavelty, M. & Suter, M. (2012). The art of CIIP strategy: Taking stock of content and processes. In
J. Lopez, R. Setola, S. D. Wolthusen (Eds.). Critical infrastructure protection: Information
infrastructure models, analysis, and defense (pp. 15–38). Springer: Berlin.
Dynes, S., Goetz, E., & Freeman, M. (2008). Cyber Security: Are economic incentives adequate? In E.
Goetz & S. Shenoi (Eds.), Critical infrastructure protection, IFIP International Federation for
Information Processing (Vol. 253, pp. 15–27). Boston: Springer.
Floridi, L. (1998). Does information have a moral worth in itself? Paper presented at Computer Ethics:
Philosophical Enquiry in Association with the ACM SIG on Computers and Society, London School
of Economics and Political Science, London, December 14–15, 1998. http://papers.ssrn.com/sol3/
papers.cfm?abstract_id=144548.
123
714 M. Dunn Cavelty
Floridi, L. (1999). Information ethics: On the theoretical foundations of computer ethics. Ethics and
Information Technology, 1(1), 37–56.
Floridi, L. (2001). Ethics in the Infosphere. The Philosophers’ Magazine, 6, 18–19.
Floridi, L. & Sanders, J. W. (1999). Entropy as evil in information ethics. Etica & Politica, special issue
on Computer Ethics, 1(2).
Floridi, L., & Sanders, J. W. (2001). Artificial evil and the foundation of computer ethics. Ethics and
Information Technology, 3(1), 55–66.
Fung, B. (2013). The NSA hacks other countries by buying millions of dollars’ worth of computer
vulnerabilities. Washington Post. http://www.washingtonpost.com/blogs/the-switch/wp/2013/08/31/
the-nsa-hacks-other-countries-bybuying-millions-of-dollars-worth-of-computer-vulnerabilities/.
Greenwald, G. & MacAskill, E. (2013). Obama orders US to draw up overseas target list for cyber-
attacks, The Guardian. http://www.theguardian.com/world/2013/jun/07/obama-china-targets-cyber-
overseas.
Hagmann, J., & Dunn Cavelty, M. (2012). National risk registers: Security scientism and the propagation
of permanent insecurity. Security Dialogue, 43(1), 80–97.
Hoogensen, G., & Stuvøy, K. (2006). Gender, resistance and human security. Security Dialogue, 37(2),
207–228.
Jervis, R. (1978). Cooperation under the security dilemma. World Politics, 30(2), 167–214.
Kerr, P. (2007). Human security. In A. Collins (Ed.), Contemporary security studies (pp. 122–134).
Oxford: Oxford University Press.
Kristensen, K.S. (2008). The absolute protection of our citizens: Critical infrastructure protection and the
practice of security. In M. Dunn Cavelty & K. S. Kristensen (Eds.), The politics of securing the
homeland: Critical infrastructure, risk and securitisation (pp. 63–83). London: Routledge.
Kuehn, A. (2013). Extending cybersecurity, securing private internet infrastructure: The U.S. Einstein
Program and its Implications for Internet Governance. In R. Radu, J.-M. Chenou & R.H. Weber
(Eds.) The evolution of global internet governance (pp. 157–167). Schulthess: Zürich.
McCue, C. (2007). Data mining and predictive analysis: Intelligence gathering and crime analysis.
Oxford: Butterworth Heinemann.
Miller, C. (2007). The legitimate vulnerability market: The secretive world of 0-day exploit sales. In 6th
Workshop on the Economics of Information Security (WEIS 2007). http://weis2007.econinfosec.
org/papers/29.pdf.
Morozov, E. (2013). To save everything, click here: Technology, solutionism, and the urge to fix problems
that don’t exist. London: Allen Lane.
Mueller, M., Schmidt, A., & Kuerbis, B. (2013). Internet security and networked governance in
international relations. International Studies Review, 15(19), 86–104.
NIST (2002). NIST Special Publication 800-30, Risk Management Guide for Information Security.
PCCIP President’s Commission on Critical Infrastructure Protection. (1997). Critical foundations:
Protecting America’s infrastructures. Washington: US Government Printing Office.
Perlroth, N., & Sanger, D. E. (2013). Nations buying as hackers sell knowledge of software flaws. The
New York Times, 14, A1.
Perry, W. L., McInnis, B., Price, C. C., Smith, S. C., & Hollywood, J. S. (2013). Predictive policing: The
role of crime forecasting in law enforcement operations. Santa Monica: RAND.
Rowe, N. C. (2010). The ethics of cyberweapons in warfare. International Journal of Techoethics, 1(1),
20–31.
Rueter, N. (2011). The Cybersecurity Dilemma. MA thesis. Duke University.
Schneier, B. (2012a). The vulnerabilities market and the future of security. Forbes, May 30. http://www.
forbes.com/sites/bruceschneier/2012/05/30/the-vulnerabilities-market-and-the-future-ofsecurity/.
Schneier, B. (2012b). When it comes to security, we’re back to Feudalism. Wired, http://www.wired.com/
opinion/2012/11/feudal-security/.
Schneier, B. (2013). The battle for power on the internet. The Atlantic, http://www.theatlantic.com/
technology/archive/2013/10/the-battle-for-power-on-the-internet/280824.
Simonite, T. (2013). NSA’s own hardware backdoors may still be a ‘‘problem from hell’’, http://www.
technologyreview.com/news/519661/nsas-own-hardware-backdoors-may-still-be-a-problem-from-
hell/.
Stevens, T., & Betz, D. J. (2013). Analogical reasoning and cyber security. Security Dialogue, 44(2),
147–164.
123
Breaking the Cyber-Security Dilemma 715
123
Reproduced with permission of the copyright owner. Further reproduction prohibited without
permission.