Balahmadi Thesis
Balahmadi Thesis
Balahmadi Thesis
Operation Centres
List of Figures ix
List of Tables xi
1 Introduction 1
1.1 Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.1.1 Monitoring for Malware in SOCs . . . . . . . . . . . . . . . 3
1.1.2 Malware Detection Tools . . . . . . . . . . . . . . . . . . . . 4
1.2 Research Gaps . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.3 Problem Statement and Research Aims . . . . . . . . . . . . . . . . 7
1.4 Thesis Organisation . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
1.5 Publications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 10
2 Background 13
2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
2.2 Security Operations Center (SOC) . . . . . . . . . . . . . . . . . . . 14
2.2.1 People . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
2.2.2 Process . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
2.2.3 Technology . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
2.3 Malware Network Communications . . . . . . . . . . . . . . . . . . 18
2.3.1 Propagation . . . . . . . . . . . . . . . . . . . . . . . . . . . 19
2.3.2 Rallying . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
2.3.3 Operation . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
2.4 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
3 Literature Review 22
3.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
3.2 Network Intrusion Detection Systems (NIDS) . . . . . . . . . . . . 23
3.2.1 Machine Learning-Based NIDS . . . . . . . . . . . . . . . . 25
3.3 Malware Analysis, Detection, and Classification . . . . . . . . . . . 26
3.4 Cyber Data Fusion . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
3.5 Security Operation Centres . . . . . . . . . . . . . . . . . . . . . . . 33
3.6 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35
v
Contents vi
4 Methodology 37
4.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37
4.2 Participant Recruitment and Demographics . . . . . . . . . . . . . 38
4.3 Quantitative Method: Online Questionnaire . . . . . . . . . . . . . 40
4.3.1 Data Collection: Online Questionnaire . . . . . . . . . . . . 40
4.4 Qualitative Method: Semi-structured Interviews . . . . . . . . . . . 41
4.4.1 Data Collection: Semi-Structured Interviews . . . . . . . . . 41
4.4.2 Data Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . 43
4.4.3 Qualitative Research Limitations . . . . . . . . . . . . . . . 45
4.5 Experimental Design . . . . . . . . . . . . . . . . . . . . . . . . . . 45
4.5.1 Dataset . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 45
4.5.2 Classifiers Design . . . . . . . . . . . . . . . . . . . . . . . . 46
4.5.3 Evaluation Metrics . . . . . . . . . . . . . . . . . . . . . . . 47
4.6 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 47
References 186
Appendices
ix
List of Figures x
xi
List of Tables xii
xiii
List of Abbreviations xiv
1.1 Motivation
Malware has evolved from viruses attacking single victims to more sophisticated
malware such as botnets, ransomware, or Advanced Persistent Threats (APTs)
used by attackers for monetising or much more disruptive purposes. Malware is
the source of many Internet threats such as information or credit card theft (e.g.
Aurora [2]), network service disruption through DDoS (e.g. DDoS on Estonia [3]),
email spam (e.g. Pushdo [4]), ClickFraud (e.g. ClickBot.A [5]), and spreading
malware (e.g. Zeus [6]).
Such sophisticated attacks are challenging to detect as they employ multiple
attack vectors to achieve their objectives by targeting an organisation for data
exfiltration, impeding critical aspects, or positioning itself to carry out these
objectives in the future. Large multinational organisations hit with destructive
1
1. Introduction 2
malware can experience an average cost of $239 million and lose an average of
12,316 devices1 .
Building out an organisation’s cybersecurity capabilities can be a daunting
task. The number of advanced and emerging threats will continue to outpace the
capabilities of security personnel within organisations. Security Operations Centres
(SOCs), a centralised group of security personnel 2
responsible for the 24/7 moni-
toring, detection, and response to security attacks, are the first line of defence in an
organisation. Previous research suggests that security practitioners may benefit from
better tools for their job [7–9] and a need for a better understanding of the human
side of security operation centres [10, 11]. For example, recent work highlighted
the significant technical and human-centric issues SOCs face [12]. Oftentimes SOC
analysts need to manually go through alarms performing monotonous tasks, wasting
time on lower priority alerts while the more critical ones slip by.
To enable SOC practitioners to work more efficiently and focus their efforts
on detecting more challenging threats, additional levels of intelligence need to be
incorporated into SOC operations. As the level of automation in the SOC scales
upward, analysts will be free to embrace a more effective “hunter role” detecting
the more dangerous threats.
In 2013, Target was hit by the most prominent retail hack in U.S. history
that managed to infiltrate Target’s network installing malware designed to steal
customer’s credit cards. Target has installed a new malware detection technology
by FireEye a few months prior to the breach that cost $1.6 million. This technology
creates a virtual honeypot environment, executing each file received over the network
to determine if it is malicious. The tool was able to detect the malware used in
the breach, generating an alarm that was picked up by Target’s SOC in Bangalore.
However, when the alarm was escalated to the Minneapolis SOC, it was ignored [13].
The deployed technology could have been able to automatically respond by
deleting the malicious file, a capability that had been disabled by Target’s security
personnel. This configuration is not unusual, as most SOCs would want to avoid
any automated decision that could cause business disruption - a lack of trust caused
by the prevalence of false positive alarms created by detection tools. This breach
resulted in more than 90 lawsuits filed against Target by customers and banks
for negligence, compensatory damages and significant financial loss of $61 million
for the breach response alone [13].
1
https://www.ibm.com/downloads/cas/XZGZLRVD
2
We will refer to personnel that work in a SOC as security practitioners or SOC practitioners
interchangeably.
1. Introduction 3
The malware used in Target’s breach was far from sophistication. Nevertheless,
the malware was able to bypass a large and resourceful organisation’s security
controls and procedures, hinting that the problem goes beyond just a SOC’s
technological capabilities. What this case study highlights is that the challenges in
detecting malware through a SOC are complex and cut across multiple processes
and technologies. To understand both the importance and the challenges of the
SOC it is worth exploring two areas in more detail: (1) difficulties related to the
monitoring for malware in SOCs and the role of humans (analysts) in configuring
and using technology, and (2) weaknesses in the malware detection technologies.
determining the threat and the best response. Analysts use sandboxes and online
engines (e.g. VirusTotal3 ) to get information about a malicious binary. However, in
Target’s case, although the FireEye technology detecting the malware gave a high
risk score, the security team may have been skeptical about the information [13],
hence not prioritizing the alarm.
the Internet edge. However, this assumption is no longer valid in modern networks.
Well defined security perimeters no longer exist due to the heterogeneous nature
of Internet connectivity in organisations and the adoption of technologies such as
cloud computing, VPN, the borderless architecture of the IoT, and the increase
of mobile devices and bring-your-own devices (BYOD).
Attackers have also grown in sophistication and regularly employ new methods
to hide their activities and how they bypass these perimeter intrusion detection
devices. They utilise stealthy malware that is very discrete, traversing in the
network very slowly, taking days, weeks, or months to accomplish their objectives
to avoid detection. Hence, as networks grow in size and speed, detecting such
malware will be a difficult task.
work. Analysts (humans) are essential in a SOC, interacting and working together,
using the information provided by the SIEM, and following specific workflows
based on the incident at hand.
Today, more companies are outsourcing the security management and network
monitoring to companies that offer SOC as a service, to escape significant investments
and mitigate their lack of cybersecurity expertise [24]. Although these companies
monitor the client’s systems and networks for signs of malicious activity, they may
not have direct access to the client’s hosts for privacy or policy reasons. When a
client’s host is suspected of being infected by malware, gaining access to that host
and retrieving the malware binary for analysis is a challenge. However, existing
dynamic analysis tools for malware classification require SOC analysts to have
access to these malicious binary executables. Even if access to hosts is possible,
the time required to negotiate this access and retrieve the executable, before its
analysis, is not efficient where speedy detection and response are critical. For
example, in situations where the implications are severe, such as in ransomware
attacks, classifying the attack as ransomware is critical, ensuring rapid response
to avoid catastrophic implications.
Behavioural malware detection systems continue to dominate the most effective
solutions for malware detection. One of the main limitations for many previously
proposed network-based behavioural monitoring systems is their reliance on content-
based features. Such technological solutions are privacy-invasive and not favourable
in a real-world deployment, especially when monitoring is outsourced to a third
party and with clients that require security clearance. Gartner estimated that by
2019 80% of organisations’ web traffic would be encrypted (i.e. using HTTPS) [25].
Regulations such as General Data Protection Regulation (GDPR) require the
enforcement of appropriate security measures for protecting personal data; hence,
traffic encryption is favoured to ensure compliance. However, similarly attackers are
increasingly employing HTTPS to conceal malware and evade detection, therefore,
developing encryption resilient inspection measures is a requisite.
More than 317 million new variants of malware were observed in 2014 (i.e.
malware automatically generated by modification of a previous malware binary that
results in a new hash), an estimate of 1 million unique malware variants each day,
increasing the total number of malware to approximately 1.7 billion [26]. Proposed
detection solutions need to adapt to this growing threat landscape, developing
systems that adapt to malware use of new attack tactics, and are resilient to
malware evasion and obfuscation.
1. Introduction 7
RQ2: What are the advantages and disadvantages that security practi-
tioners perceive of network-monitoring tools?
they are also able to determine what malicious activities the malware is launching
and choose the most effective response measure. Chapter 10 concludes the thesis
and proposes future research.
1.5 Publications
The work described in this thesis has resulted in peer-reviewed publications, with
additional papers still under review. The publications are derived from chapters, as
illustrated in Figure 1.1, which also illustrates the relationship between research
questions and the thesis structure.
In all publications, I was the first author and was responsible for the research
reported and for writing the paper, with some writing contributions from the other
authors. The survey and interviews reported in Chapters 3, 7, and 6 were part of
a wider set of interviews carried out jointly with another author (Louise Axon); the
questions relating to this thesis were devised, asked, and analysed by myself.
1. Alahmadi, B.A., Axon, L. , Webb H. and Martinovic, I., Down the Rabbit
Hole: Understanding Security-Monitoring Processes in Security Operations
Centres. (Under Review S&P’21 )
2. Alahmadi, B.A., Axon, L. and Martinovic, I., 99% False Positives: On SOC
Analysts’ Love-Hate Relationship with Security Monitoring Tools. (Under
Review Usenix Security’21 )
4. Alahmadi, B.A., Mariconti, E., Spolaor, R., Stringhini, G. and Martinovic, I.,
BOTection: Bot Detection by Building Markov Chain Models of Bots Network
Behavior. (AsiaCCS’20 )
I was also lead author and a co-author in other publications that are outside
of the focus of this thesis.
3. Alahmadi, B.A., Legg P.A, Nurse J.RC , Using Internet Activity Profiling for
Insider-threat Detection. (12th Special Session on Security in Information
Systems - WOSIS 2015 )
1. Introduction 12
Introduction
Background
Literature
Review
Methodology
Contribution
Chapter 5
RQ2: Tools Contribution
Weaknesses and Paper 2
Strengths Chapter 7
RQ3: On-the-wire Contribution
Malware Paper 3
Classification Chapter 8
RQ4: Advanced Contribution
Malware Paper 4
Detection Chapter 9
Conclusion
Figure 1.1: Relationship between research questions, thesis structure and publications.
Background
2
Contents
2.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . 13
2.2 Security Operations Center (SOC) . . . . . . . . . . . . 14
2.2.1 People . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
2.2.2 Process . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
2.2.3 Technology . . . . . . . . . . . . . . . . . . . . . . . . . 16
2.3 Malware Network Communications . . . . . . . . . . . 18
2.3.1 Propagation . . . . . . . . . . . . . . . . . . . . . . . . . 19
2.3.2 Rallying . . . . . . . . . . . . . . . . . . . . . . . . . . . 20
2.3.3 Operation . . . . . . . . . . . . . . . . . . . . . . . . . . 20
2.4 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
2.1 Introduction
In this chapter, we present a background overview to introduce the reader to the
concepts and terminology used in subsequent chapters.
We start by describing the security operation centres (SOCs) landscape and the
integration of people, process and technology. We focus on describing the primary
technology used in SOCs —SIEMs. We present the SIEM architecture, and how
existing monitoring technologies and security logs are integrated within. We then
go on describing the botnet architecture and the various network communications
it launches, which serve as a baseline of what malware detectors should detect.
13
2. Background 14
Process
Intel/ Threat
Technology Network and
People System Owners
Firewall Escalation
5
Level 1 Level 2 6
2 Case
1 SIEM
Web Server Closed Incident
4 Handler
Network
Engineer
3
7
Proxy
SIEM Storage
Business
Monitoring
Tier 1 Open Tickets, Close False Positives
Basic Investigation and Mitigation
Deep investigation/CSIRT
Tier 2 Mitigation/recommends changes
Advanced Investigation/CSIRT
Prevention
Threat Hunting
Tier 3 Forensics
Counter-intelligence
Malware reverser
Figure 2.2: Example of a three-tier SOC and personnel responsibilities, adapted from [30]
in Figure 2.4 an example for a SOC structure and its three dimensions (people,
process, technology). In such a structure, when a threat enters the network, security
devices and logs detect this threat. These detection are collected by the SIEM that
produces an alarm. The alarm is verified by level 1 analysts and is escalated to
incident handlers to handle the incident by liaising with the system owners. Reports
of organisations threats are communicated to the business management.
When discussing the challenges in threat monitoring in SOCS, it’s essential
to identify first the various Tools, Processes and People involved in the opera-
tion of the SOC.
2.2.1 People
The SOC team’s goal is to detect, analyse, and respond to cybersecurity incidents
using a combination of technology solutions and a strong set of processes. To do
that, analysts have to maintain Situational Awareness (SA) of events from the
systems and networks they monitor. Situational Awareness is defined as: “Within
a volume of time and space, the perception of an enterprise’s security posture and
its threat environment; the comprehension/meaning of both taken together (risk);
and the projection of their status into the near future.” [29].
The SOC team is typically staffed with security analysts, engineers, incident
responders, hunters, contractors, as well as managers who oversee security operations.
SOC engineers are responsible for providing and supporting the SOC with the
required software (e.g. SIEM scripts or configurations). Incident responders handle
2. Background 16
events that are escalated by the analysts that need in-depth investigation and foren-
sics. SOC practitioners’ team structure and responsibilities varies [23]. For example,
analysts can be categorised by their tasks to multiple levels, described in Figure 2.2.
2.2.2 Process
The SOC process involves the various workflows SOC security practitioners follow
in their everyday tasks. For example, this includes process followed for SIEM
monitoring and alarming, event management process, security incident ticket
management, incident handling, reporting and escalation process. All these processes
are documented in a Wiki Portal, share point or share drive for team reference.
There are standard guidelines that direct analysts in SOC operations. For
example, The European Network and Information Security Agency (ENISA) [31],
and the National Institute of Standards and Technology (NIST) [32] have published
guidelines for incident response teams. To provide a rapid automated response for
incident identification, detection, response to SOC team communications, procedures
are documented in a playbook [33], a document prepared by experienced SOC
analysts that details steps an analyst follow to deal with a security alarm.
2.2.3 Technology
SIEMs
One of the most frequently chosen tools in a SOC is Security Information and
Event Management (SIEM)— e.g, ArcSight 1 , AlienVault 2 . SIEMs are systems
that combine SIM (security information management), and SEM (security event
management) functions into one security management system. SEM deals with real-
time monitoring, correlation of events, notifications and console views. SIMs provide
long-term storage as well as analysis, manipulation and reporting of log data and
security records of the type collated by SEM software. SIEMs replace the need for
analysts to access traditional security tools directly. Instead, the SIEM aggregates
the logs from the multiple data sources, and processes them to detect threats.
We show the SIEM architecture for monitoring and detection of alerts in
Figure 2.3. The first step is to parse the data collected from the data sources
and normalise it to a standard format produced as a security event. Multiple
security events may be correlated to create a correlation rule or alarm. When the
rule is triggered, an alarm is fired and prioritised. Then, it is up to the analyst
to determine if the alarm is a false positive (FP) or it needs to be escalated. We
discuss each step in further detail in the following.
Data Sources— The SIEM has data source plug-ins called collectors where
data sources are fed into it. These data sources could be either raw logs or security
events generated by security devices. Such data sources could be, for example,
network-based security tools (e.g. firewalls, IDS), host-based security tools (e.g.
Anti-Virus), logs (e.g. operating systems logs, web server logs). Contextual data
provided by threat intelligence platforms and other processes (e.g. vulnerability
scans) could also feed into the SIEM.
Ideally, these data sources are received from Security Device Event Exchange
(SDEE) enabled devices/hosts. SDEE is a standard proposed by the International
Computer Security Association (ICSA)3 that specifies the format of messages and
protocols used to communicate events generated by security devices. SDEE enables
devices to collect logs in the device itself, and the SIEM retrieves these logs. If
the raw logs match a specific criterion, then part of the message is inserted into
the SIEM database as a security event. Other devices send the logs directly
to the SIEM to store.
Normalisation— Logs/messages received from SDEE enabled devices are
intrinsically suited to the SIEM. They do not require manipulation because they
1
https://www.microfocus.com/en-us/products/siem-security-information-event-management/
overview
2
https://cybersecurity.att.com/solutions/siem-log-management
3
https://www.icsalabs.com/
2. Background 18
are in the right format. However, some applications/devices were never designed
to generate logs. Therefore, they have to be heavily edited by scripts to produce
a log that will fit the SIEM’s requirements.
Correlation— Most alarms in a SIEM are directive alarms, also known as
correlation rules. Using correlation rules, SOCs can identify potential security
threats by detecting behaviour patterns in disparate yet related events. Directive
alarms link these different events to generate an alarm that is more useful than
any event seen in isolation. If the organisation has a particular threat use-case,
then they can create their own directives.
Alarming and Prioritisation— Its worth noting that there is a distinction
in the definition of alarm, alert, and event in a SOC operation. Security tools
and networking devices produce alerts when they detect a threat. Similarly, these
threats might be written in logs as an event. The alerts and events are then
aggregated through the SIEM to produce an alarm.
When multiple alarms are flagged, a frequent scenario in a SOC, what alarm is
looked at depends on what is monitored. For example, if an organisation is only
concerned about where data is going, then network traffic and IDS alerts alone
might be sufficient. However, receiving multiple alerts from different sources related
to the same asset provides assurance of the validity of the alarm. How they choose
the alarms they investigate depends on several factors, which could be defined
using the prioritisation module in the SIEM engine.
Reporting and Visualisations— Alarms produced by the SIEM are presented
to the analyst through visualisation (e.g. dashboards). In addition, the SIEM
provides reporting capabilities, meaning analysts can auto-generate reports.
Bot Spam
C&C DDoS
Bot
Crypto Mining
Botmaster
ClickFraud
Bot
Ransom
2.3.1 Propagation
There is strength in numbers, and the power of the botnet increases when the
number of bots grows. Therefore, bots aim to increase the number of bots by
recruiting new bots through propagation. The propagation process may initially
start with a synchronisation stage. Bots need to synchronise/coordinate with each
other through obtaining a seed input that is used by the bot Domain Generation
Algorithm (DGA) [35]. This input could be a synchronized system time, for example,
Conficker-A sends an empty HTTP request to a popular website such as google.com
to get the current date/time, or Torpig [35] uses trending topics on Twitter or
the Stock exchange. This synchronised token is then used as input to the Domain
Generation Algorithm (DGA) to communicate with C&C server domain name.
The bot then attempts to infect another machine either through active means (e.g.
scanning) or passive propagation (e.g. phishing emails). Some bots employ multiple
stages of infection, and once the victim is infected with the bot binary it downloads
additional files from the C&C server. When the infection process is complete, the
new bot registers with the C&C server. Propagation could be categorised as passive
or active based on the recruitment mechanism used.
Active— Active propagation means that the bots exhibit worm-like behaviour,
actively attempting to infect new victims, using existing vulnerabilities (e.g. Duqu
2.0). Bots leverage vulnerabilities that enable them to gain administrative privileges
on the infected machine [34]. Some bots perform a network scan before the actual
infection (e.g. Conficker), either by sending ICMP echo requests to detect reachable
hosts or port scanning to detect vulnerable services (e.g. Neris) [36]. For example,
Sality performed a scan of the entire IPv4 address space for 12 days [34]. Phatbot
could either have hard-coded configuration or be instructed to scan a certain network
2. Background 20
range [36]. Botnets may rely on scanning to propagate to reduce resources cost or as
a way of obfuscation if the attack operation’s communication protocol is used seldom
on the victim network [36]. However, scanning is noisy and results in additional
network traffic; thus stealthy bots may resort to a passive propagation approach.
Passive— Bots can propagate through other means such as phishing emails,
drive-by-download or infected removable media, known as Passive propagation. For
example, Zeus is distributed over phishing emails and drive-by-download. Although
the underlying network communications resulting from passive approaches (e.g.
receiving a phishing email, visiting a website) can be detected, correlating these
activities with the actual bot infection is difficult using network-based detectors [36].
This is as a result of the passive propagation and the actual binary infection
possibly occurring in different times [36].
2.3.2 Rallying
The first step a new bot does is registering with the C&C server, a process also
known as rallying [34]. There are several techniques that bots use to contact the
C&C server. The basic approach is to contact a domain hard-coded in the bot
binary. However, this approach is vulnerable to domain takedowns, and when the
domain becomes unreachable then the bot can no longer receive commands. Some
bots use a Domain Generation Algorithm (DGA) (e.g. Zeus, Conficker), generating
a different domain name from a changing seed input obtained in the coordination
stage. This makes taking down the botnet domain a more complicated process;
once a domain is taken down by the authorities, the bots can simply connect to
the next domain generated by the DGA.
To add to the layer of resiliency, the bot can apply fast-fluxing where multiple
bots register and de-register their IP addresses to the same DNS host record,
mapping a single domain name to possibly thousands of bots. Thus instead of bots
connecting to the C&C directly, they connect to one these active proxy bots that
rally the messages to the C&C server. Storm uses Double-flux, a sophisticated
fast-flux approach where these bots register their IP addresses as part of DNS
Name Server record list for the DNS zone.
2.3.3 Operation
As part of the bot’s operation, the bot can attempt to infect new victims through
port scanning to detect vulnerable services, or ICMP scans to detect reachable
hosts. Botnets could be used to carry out various types of attacks, each having
different network flow communication structures.
2. Background 21
For example, bots could be used to launch Distributed Denial of Service Attacks
(DDoS), sending huge amounts of requests (usually TCP/UDP HTTP requests) to
overload systems and seeking to make a machine or network resource unavailable.
Bots could also be used to send unsolicited emails or SPAM, usually using Simple
Mail Transport Protocol (SMTP). ClickFraud is another attack that uses bots to
access advertising URLs (pay-per-click ads), forcing advertisers to pay for irrelevant
clicks, leveraging HTTP and HTTPS protocols.
Botnets also utilise network communications for other operation tasks, such
as scanning the network to collect network information (e.g. Duqu 2.0, Phatbot),
login information harvesting (e.g. Storm), self-updating (e.g. Zeus, Phatbot),
bitcoin mining (e.g. Miner), receiving C&C instructions (e.g. Stuxnet), or network
traffic modification (e.g. Miner). Other protocols bots might leverage include
Server Message Block (SMB) (e.g. Duqu 2.0) and Internet Relay Chat (IRC) (e.g.
Rbot), all visible in a bots network traffic. More details on the bots communication
patters could be found in [36].
2.4 Summary
We provided in this chapter an overview description of SOCs, focusing on describing
its leading technological platform (i.e. SIEM) and the monitoring and detection
of malware. SOCs have various other responsibilities and apply other tools that
we do not address and are outside the scope of this thesis.
Malware requires network communication to interact with its C&C, execute its
attack operations and propagate to recruit further bots. Network-based malware
detection systems take advantage of malware’s need for network connections,
spotting malware by monitoring for malicious traffic. Substantial previous work has
been proposed for malware network detection. We review the literature on network-
based malware detection and its deployment in SOCs in the following chapter.
Literature Review
3
Contents
3.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . 22
3.2 Network Intrusion Detection Systems (NIDS) . . . . . 23
3.2.1 Machine Learning-Based NIDS . . . . . . . . . . . . . . 25
3.3 Malware Analysis, Detection, and Classification . . . . 26
3.4 Cyber Data Fusion . . . . . . . . . . . . . . . . . . . . . 31
3.5 Security Operation Centres . . . . . . . . . . . . . . . . 33
3.6 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . 35
3.1 Introduction
In this chapter, we provide an overview of the existing literature on network-based
malware detection systems used in SOCs. Such technologies can be divided into
Network-based Intrusion Detection Systems (NIDS) and Malware-specific detection
systems (i.e. techniques tailored to malware detection). NIDS can be signature-based
or anomaly-based NIDS. Machine Learning (ML) can be applied to either type (i.e.
ML-based NIDS). For malware-specific technologies, in addition to providing a review
of network-based solutions, we also discuss host-based approaches for completeness.
These technologies can be deployed in a SOC to generate an alarm on its
own or be fed into a SIEM to be combined with other alert sources as part of
a correlation rule. Such an aggregation or fusion of multiple sources is a known
functionality of a SIEM. Hence, in this chapter, we also discuss related work on
multi-log and multi-source data fusion. We then conclude with a review of related
22
3. Literature Review 23
SIEM
Anomaly-based NIDS [43–53]
[54–57]
Machine Learning-based NIDS
work on Security Operations Centres (SOCs) and the challenges therein. We give
an illustration of this structure in Figure 3.1.
NIDS monitors the network traffic, analysing the network, transport, and
application protocols to identify suspicious activities. In NIDS monitored network
packets are analysed for rule violations by a pattern recognition algorithm [139].
Depending on the model of detection, network-based IDs can be categorized as
either signature-based or anomaly-based. A signature detection system identifies
patterns of network traffic presumed to be malicious while anomaly detection
systems compare network activities against a “normal” baseline. Garcia et al. [140]
produced a survey that analyzes and compares the most important efforts carried
out in a network-based detection area.
Signature-based IDS detect attacks by monitoring the network for pre-defined
signatures of known attacks [37–42]. This approach is useful in detecting known
attacks with a low false-positive rate. However, it is inadequate in detecting
unknown attacks such as new or polymorphic malware. It requires a signature to
be defined for all of the possible attacks that an attacker may launch against a
network, thus requiring frequent rule-base updates and signature updates.
There have been several research papers on network-based anomaly detection
(i.e. [43–52]). Anomaly-based IDS initially learns the normal network behaviour,
then monitors for any deviations in this behaviour [138]. A survey on commercial
anomaly-based IDS is provided in [53]. Compared to signature-based IDS, anomaly-
based IDS could detect unknown intrusion attacks known as “zero-day” attacks. In
addition, the network profiles of normal activity are unique to every network, making
it very difficult for an attacker to know with certainty what actions it can carry out
without being detected [139]. However, Anomaly-based IDS has its drawbacks. For
example, they produce a high percentage of false positives, fail to scale to gigabit
speeds, and have difficulty determining which specific event triggered an alarm [139].
The high rate of false positives in these alarms result in attacks or malicious activities
on the network going undetected for periods of time. Since security analysts are
looking for abnormal activity rather than a specific attack, they are prone to be
overwhelmed by time-consuming false alarms, thus increasing the time until they
detect the attack. We refer the reader to [141], a survey on anomaly detection.
Developing intelligent models for network-based intrusion detection systems
is not trivial. NIDS deal with huge network volumes, highly imbalanced data
distribution, and the challenge of recognising decision boundaries of normal and
anomaly behaviour. In addition, NIDS needs to be adaptable to the constant change
in networks without a lot of manual tuning [142]. A robust and effective NIDS
needs to solve these challenges and be designed to adapt to different patterns of
network traffic, to successfully detect unknown malicious attacks [21].
3. Literature Review 25
their network data with researchers. In addition, the anonymisation efforts of these
data are not always effective to motivate the safe sharing of network data [151].
The mass volume of network traffic also presents obstacles in data storage and
computation for researchers. Moreover, data of such volume are often unlabeled
and lack ground truth needed for system validation. As a result of these challenges,
researchers may settle by taking the proposed system to where the data is held, for
example, through collaborations with industry partners which introduces challenges
in Intellectual Property (IP) ownership. Due to the aforementioned challenges,
researchers usually resolve to lab simulation, which may not produce realistic results.
We reviewed the literature related to NIDS that monitor the network for a
specific attack signature (signature-based) or any deviation of the network’s normal
behaviour (anomaly-based). Although IDS aim to detect malicious activities in a
network, it is difficult for IDS to detect a malware-infected host, as they do not
consume a lot of bandwidth and so fly under the radar. In the following section,
we review research contributions that focus on detecting specifically malware-
oriented attacks.
monitoring networks at the perimeter and did not include local host modification
and internal network propagation.
BotHunter is only effective on unencrypted network traffic and requires multiple
bot infections on the network to detect bot activity [36]. Similarly, BotMiner [81]
requires multiple bot infections for detection and does not consider active bot
propagation through scanning [36]. However, it does improve on BotHunter
by not requiring unencrypted traffic. Ashfaq et al. [83] extended the infection
dialog proposed by BotHunter [82] to include passive propagation mechanisms
such as spam.
Gu et al. proposed BotSniffer [78], a network anomaly-based botnet detection
system, to detect the malware’s C&C communications with protocols such as
IRC and HTTP. The approach utilizes spatial-temporal correlation and statistical
algorithms to detect the synchronisation of hosts’ network traffic. The authors
hypothesised that botnet C&C communication has a certain pattern that is a
result of pre-programmed activities. However, this approach requires multiple
infected bots on the same network.
Gu et al. also proposed BotMiner [81], a network anomaly detection system
that exploits the similarity of botnets C&C communications and malicious network
patterns. Tegeler et al. proposed BotFinder [84], a bot detection system that
monitors network traffic for C&C communication to detect bot-infected hosts. It
builds a model of C&C network patterns of the different malware families; thus,
traffic that matches the model is flagged as malicious. Compared to BotSniffer [78]
and BotMiner [81], the approach detects individually infected hosts and therefore
does not need to correlate network activities of multiple hosts. Moreover, it does
not rely on payload information, but high-level network data such as NetFlow in the
detection, making it applicable to encrypted botnet traffic. Abaid et al. [85] analyzed
botnet network behaviours and identified those that are synchronised across multiple
bots. However, they only focused on the detection of spam and port scanning.
C&C Detection— Since malware is normally distributed, managed, and
coordinated through remote servers, research efforts have focused on monitoring,
detecting, and blocking these malicious domains. Such malware communicates to
C&C servers to receive commands using protocols such as IRC, or more commonly
HTTP since web traffic is not usually blocked in most organisations.
For example, Botzilla [77] monitored malware network traffic, specifically a bot’s
communication to the C&C server, in a sandboxed environment to find patterns
and generate network signatures. Bilge et al. proposed Disclosure [79] that extracts
flow size-based features, client access patterns, and temporal behaviour features
3. Literature Review 29
malicious domains through analysing DNS Queries. The approach repeatedly issues
DNS queries to a set of possible malicious domains and collects information about
the resolved IP addresses, to classify each domain name into either fast-flux or non-
fast-flux [96–98]. However, active probing may be detected by the adversary, which
then avoids responding to probing queries to prevent unveiling further information.
Manadhata et al. [99] modelled the malicious domain detection problem as an
inference problem on large host-domain graphs. A host-domain graph is constructed
from the proxy log and a seed of minimal ground truth; they then apply belief
propagation to estimate the marginal probability that the domain is malicious.
Similarly, this approach could be applied to other logs, such as DNS queries.
Malware Family Classification— Rossow et al.[76] proposed a dynamic
malware analysis environment by observing the network behaviour of 100,000
malware samples over long periods of time. The authors aimed to tackle the
limitations of previously proposed malicious network traffic tools [73, 74], such as
the short analysis period of a few minutes and the lack of detailed network behaviour
analysis. They also provide insights on DNS and HTTP malware traffic trends.
Rafique et al. [17] investigated the application of several machine learning algorithms
for malware classification. They propose a framework to classify malware based
on different network protocols (e.g. HTTP).
Malware behavioural clustering approaches aim to group malware behaviour
to reveal similarities between malware samples that may not be captured using
system-level malware analysis. Perdisci et al. in [103], proposed a network-level
clustering system to derive similarities in HTTP-based malware behaviour for
detection. FIRMA [104] used a similar approach to cluster and generate network
signatures for malware network traffic. Previous research also extracted malware
behavioural features as a sequence of values. For example, Santos et al. [105]
used n-grams, which are sub-strings of a larger string with length n, to generate
file signatures for malware detection.
Wressnegger et al. [102] studied the applicability of n-grams for anomaly
detection and classification. Mohaisen et al. [100] observed the high-level network
features of malware and applied n-gram document analysis for family classification.
Similarly, Mekky et al. [101] used n-grams to encode the order of sub-sequences of
malware network communication events to build malware classification models. The
authors applied a similar approach to [100] by first isolating malware traffic from
the benign traffic using Independent Component Analysis (ICA). However, their
approach differs from [100] as they use coarse-grained groups of network events so
that inbound, DNS, and port number are considered one network event. Similar
to [100], they use a count-based approach for network flow identification.
3. Literature Review 31
was “Big Data”, where the organisation devices generate billions of log alerts per
average a day stored in a SIEM, and “efficient data-reduction algorithms and
techniques” are required. Another challenge was “big velocity” as the organisation
deploys various heterogeneous devices with a wide verity of data structures, making
correlation of these events difficult. Moreover, they described some data as dirty, as
devices may generate logs that are either incomplete or inconsistent (different time-
stamps), making analysis an even more significant challenge. However, Beehive was
quite successful in detecting 784 security incidents compared to the eight detected
by the organisation’s existing security system, reducing the alerts inspected by 74%
by removing the white-listed target hosts. Hence, data log aggregation is found
to be a promising approach for effective intrusion detection.
Bocchin et al. [114] used network connectivity graphs to extract host events
relevant to a seed event to detect malicious network activities. The approach
passively monitors network traffic at the edge and extracts and logs events, such as
HTTP requests, DNS queries, and response or use of unknown protocols. When
an intrusion detection system flags an event, this event is the seed that triggers
the proposed system to correlate the logs correlated with the seed event. It then
models network activities of hosts over time (different network traffic samples
from the same host), space (connecting similar patterns across different hosts),
and network layers (HTTP, DNS, etc.). The approach provides analysts with
more visibility to what network traffic is associated with the IDS alert, thus more
understanding of the underlying activities.
APT Detection— Oprea et al. [115] propose a graph-based approach based
on belief propagation to detect malicious domains utilized by malware in the
early stages of an Advanced Persistent Threats (APT). The method was evaluated
using a large dataset of web proxy logs and utilizes supervised and unsupervised
machine learning algorithms. Given a seed of a known malicious host or domain,
the approach infers other possible domains or compromised hosts that belong
to the same criminal network.
Milajerdi et al. proposed Holmes [164], a real-time APT detection system
that correlates tactics, techniques, and procedures that might be used to carry
out each APT stage. HOLMES generates a high-level graph that summarizes
the attacker’s steps in real-time. Pei et al. proposed Hercule [165], a multi-stage
log-based APT attack detection. Inspired by graph analysis approaches, it builds a
graph by correlating log entries across multiple logs, then discovers attacks through
community detection algorithms. Similarly, Peng et al. [166] used a graph-based
approach for malicious domain detection.
3. Literature Review 33
Active Theory Model, they identified and list contradictions in SOCs that manifest
as conflicts. They present a Pentagon model for improving the SOC operations
by identifying tasks that can be automated and resolve conflicts [128]. Kokulu
et al. [12] applied a qualitative approach to identify technological, human, and
operational issues in SOCs across different sectors. The authors discussed limited
issues arising from technology (e.g. malfunctioning of SOC tools). One of the
most interesting findings was that analysts don’t perceive false positives to be a
major issue, despite academia’s belief.
Akinrolabu et al. [174] found that current IDS are inadequate in detecting
multi-stage attacks stealthy attacks. Although existing ML detection approaches
exists, the authors discussed how they rely on statistical features derived from
training sets limiting the models ability to detect new malware variants. The
authors conducted interviews with analysts, eliciting 17 behavioural-based network
features, capable of detecting novel attacks.
Paul et al. [175] studied how cybersecurity analysts establish and maintain
Situation Awareness (SA) by conducting interviews, observations, and a card
sorting activity. They identified questions analysts ask themselves when faced with
a networking event, developing a taxonomy of the questions for event detection
and orientation. D’Amico et al. [130] applied Cognitive Task Analysis (CTA) to
understand the work processes, cognitive skills, and tools that analysts rely on to
achieve SA. They also identified the cognitive challenges and obstacles that impede
SA. Haney et al. [176] investigated the methods, tools, and challenges of Red and
Blue teams, identifying examples of successful integration and opportunities to
enhance SA. They also discuss design implications for tools that can facilitate SA
among multiple cyber-defense teams by supporting data fusion, change detection,
network mapping, and access tracking.
M’manga et al. [129] used Distributed Cognition and Grounded Theory, a
qualitative approach, to understand how security analysts make decisions about
risk. D’Amico et al. [131] studied the analysts’ analytic goals, tasks, and the
processes they follow, the decisions they make, and the data sources and tools used
to make those decisions and challenges they face.
3.6 Summary
In this chapter, we reviewed the existing related literature identifying their limi-
tations and the research gaps. Although previous work that considered malware
network communication patterns exists [81–85], as identified by [36] these either
3. Literature Review 36
require unencrypted network traffic (e.g. [82]), require multiple malware (i.e. bot)
infections on network (e.g. [78, 81, 82]), require active propagation through scanning
(e.g. [81]), or don’t consider local malware (i.e. bots) attacking local targets (e.g. [81]).
Importantly, most previous work on malware detection using machine learning
failed to evaluate the system in detecting unseen or new malware. This is crucial in
determining the systems’ performance over time in detecting new malware families.
Our review of existing literature studying SOCs revealed a shortage of systematic
and comprehensive research in describing the SOC landscape and the technological
and operational challenges therein. For example, research by Sundaramurthy et
al. [10, 11, 22, 128] applied ethnography research methods to study educational
SOCs, focusing only on their operational and personnel difficulties. Hence, the
strengths and weaknesses of the deployed technological solutions were not discussed.
To describe how these research gaps were addressed, we continue in the following
chapter by outlining our methodology.
Methodology
4
Contents
4.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . 37
4.2 Participant Recruitment and Demographics . . . . . . 38
4.3 Quantitative Method: Online Questionnaire . . . . . . 40
4.3.1 Data Collection: Online Questionnaire . . . . . . . . . . 40
4.4 Qualitative Method: Semi-structured Interviews . . . 41
4.4.1 Data Collection: Semi-Structured Interviews . . . . . . 41
4.4.2 Data Analysis . . . . . . . . . . . . . . . . . . . . . . . . 43
4.4.3 Qualitative Research Limitations . . . . . . . . . . . . . 45
4.5 Experimental Design . . . . . . . . . . . . . . . . . . . . 45
4.5.1 Dataset . . . . . . . . . . . . . . . . . . . . . . . . . . . 45
4.5.2 Classifiers Design . . . . . . . . . . . . . . . . . . . . . . 46
4.5.3 Evaluation Metrics . . . . . . . . . . . . . . . . . . . . . 47
4.6 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . 47
4.1 Introduction
We discuss in this chapter the research methods applied to investigate the research
aims identified in Chapter 1: (1) the role of SOC dimensions (people, process,
technology) on malware monitoring, detection and response; (2) designing novel
malware classification and detection systems that improves over existing state-
of-the-art tools.
For the first aim, we applied quantitative and qualitative research methods
reporting the findings in Chapters 5, 6, and 7. Ethical approval for this study
37
4. Methodology 38
Our participants work across seven different SOCs from organisations of var-
ious sizes serving government and non-government organisations. We show the
demographics of our survey participants in Table 4.1 and interview participants
in Table 4.2. In survey responses, participants provided their demographics at
the beginning of the questionnaire. Hence, demographics such as Expertise Level
indicated in Table 4.1 were based on the participants self-evaluation. On the other
hand, the demographics of interview participants and their organisations were
extracted from the participant’s answer to the interview question: “Please describe
your position/job role/level”. Accordingly, the participants’ expertise reported in
Table 4.2 is based on the analyst level and number of years in the position. Hence,
level-3 analysts or lead analysts are scored as high on expertise compared to level-1
analysts or those who are new to the job, who are scored low.
starting the survey, participants were presented with information about the research
goals, handling their data and the researchers’ contact information. Participants
then gave their consent for their participation. We offered participants the option to
comment on the questions and sections in the survey by providing free-text spaces.
In designing the survey questions, we incorporated feedback from subject-matter
experts (a security analyst, senior security professional and a SOC manager).
In our survey, we presented the analysts with a set of assertions on the human-
in-the-loop level in SOCs and the importance of maintaining awareness of the
network. Participants were asked to indicate their level of agreement with each
assertion using a Likert-Scale (1: “Strongly Disagree”, 2: “Disagree”, 3: “Neutral”,
4: “Agree”, 5: “Strongly Agree”). There is a discrepancy in the community as to
how to handle Likert scale data [178]. Therefore, we calculated both the mode and
median to analyse responses to the assertions. Mode or median values higher than 3
constituted overall agreement with an assertion as 3 was the neutral value. We also
calculated comparison of non-neutral scores (CNNS), which represents the ratio
of scores less than (1,2) and greater than (4,5) the neutral value (3). We present
the full list of assertions in Appendix (Tables B.1 and B.2).
Theme Description
SOC tasks that rely on analysts (humans) intelligence
1 Human-centered Tasks
and cognitive abilities for monitoring, detection and response.
• Define the Apriori Themes: We started with defining ten themes we were
interested in after reviewing the literature and the results of our survey to
guide our analysis, shown in Table 4.3.
• Produce Initial Template: We use the codes that arise from the subset of
the data to produce the initial template. The initial template may include
further sub-codes, thus producing a template with a hierarchy of codes within
each theme. We show a snapshot of the initial template in Appendix C.
• Interpretation: Using the final template produced by coding all the interview
data, we interpreted the data and wrote our findings.
4.5.1 Dataset
We used three datasets in evaluating our proposed systems.
ISCX Botnet Dataset— This dataset contains malicious and benign traffic
and is split into training and testing sets with 5 and 13 bots respectively. The
dataset contains traffic from 5283 and 9896 benign hosts for training and testing
respectively with 27 bot-infected hosts. The benign traffic in the dataset contains
activities such as SSH, HTTP, Web browsing, World of Warcraft gaming, bit-
torrent clients such as Azureus, web and email traffic and backup and streaming
media, and SMTP mimicking users’ behaviour. We refer the reader to [182] for
a detailed description of the dataset.
4. Methodology 46
accuracy, especially when there is noise in the data, whilst larger values slows the
classifier performance and can lead to overfitting. Overfitting is a modeling error
that occurs when the algorithm function is too closely or exact fit to a particular
set of data points, therefore, failing to classify additional data reliably.
Random Forest— Random Forest is an ensemble classifier that leverages
multiple decision trees to produce a better prediction accuracy. It aggregates the
decision trees individual predictions and combine into a final prediction, based on a
majority voting of the individual predictions. The trees are trained using different
subsets of the training set, thus overcoming over-fitting issues of an individual
decision tree. However, random forests is an ensemble model. Hence, its predictions
are inherently less interpretable than individual decision trees.
4.6 Summary
In this chapter, we discussed the methodology followed to carryout the research
goals identified in Chapter 1. We applied quantitative methods in form of an
online survey to gain an understanding of the cyber threats our participants face
and the monitoring and detection tools they deploy. Qualitative method is then
applied through the use of semi-structured interviews that allow us to gain a
4. Methodology 48
rich understanding of the inner workings of the SOC, their tool use and the
challenges therein. We also discussed the experimental design used to develop
malware detection and classification systems, in designing the ML classifiers, and
the datasets and metrics used for evaluation.
As these research methods are used consistently throughout the thesis, this chap-
ter provides a reference to these methods, complementing additional methodology
details provided in each chapter. In the next chapter, we present the findings of our
online survey. These findings highlighted interesting research questions about the
participating SOCs that we followup on using semi-structured interviews.
SOC Analysts’ Perspective on Network
5
Monitoring Tools
Contents
5.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . 49
5.2 Security Tools Detection Capability . . . . . . . . . . . 50
5.3 Security Monitoring Tools Use . . . . . . . . . . . . . . 53
5.4 Network Features & Indicators of Compromise (IOC) 56
5.5 Network Monitoring Efforts . . . . . . . . . . . . . . . . 57
5.6 Assertions . . . . . . . . . . . . . . . . . . . . . . . . . . . 58
5.7 Summary of Findings . . . . . . . . . . . . . . . . . . . . 61
5.1 Introduction
We commence our study on understanding malware monitoring and detection in
SOCs by conducting quantitative research through an online questionnaire. The goal
of the survey was to understanding the operational landscape that SOC analysts
are working in well enough to develop research questions. SOCs are diverse with
distinctive setups and goals; thus, the people, technology, and processes will be
unique as well. Hence, we followed an inductive approach and used a quantitative
study as a starting point for our qualitative research. The results of the survey
helped us to design our qualitative study, which we discuss in Chapters 6 and 7.
In the following, we explore the results provided by our survey: how security
practitioners view tool detection capabilities, which tools they use, network features
49
5. SOC Analysts’ Perspective on Network Monitoring Tools 50
and indicators of compromise valuable in the threat detection; and the practitioners
agreement to assertions.
10
20
30
40
50
60
70
80
90
10
20
30
40
50
60
70
80
90
10
20
30
40
50
60
70
80
90
100
100
100
Figure 5.1: Threats organisations face, what tools detect them and what needs improvement based on participating organisation size.
51
Threats Tools Detect Well Tools Need Improvment
Web defacement 33 29 33 12
Viruses, Worms, Trojans 67 76 100 94 18
Unauthorised access attempts 100 59 100 88 18
Scanning (reconnaissance e.g. port scans) 67 53 100 94 12
Ransomware 33 65 76 67 24
Policy violation (e.g. gaming, streaming) 67 47 100 88 18
Phishing emails (including spear-phishing) 33 82 100 82 12
Next-generation malware (e.g. Bots) 33 35 33 53 100 41
Insider threats 67 65 33 47 67 41
Denial of service attacks 33 47 100 82 12
Brute-force attacks 67 41 100 88 6
Advanced Persistent Threats (APTs) 100 76 67 65 67 65
Abnormal user activity 100 82 100 88 41
Abnormal network activity 100 53 67 76 33 47
10
20
30
40
50
60
70
80
90
10
20
30
40
50
60
70
80
90
10
20
30
40
50
60
70
80
90
100
100
100
Government Other
5. SOC Analysts’ Perspective on Network Monitoring Tools
Figure 5.2: Threats organisations face, what tools detect them and what needs improvement based on of SOC client (government and
non-government).
52
5. SOC Analysts’ Perspective on Network Monitoring Tools 53
had the tools. We asked:Which of these security incidents do your systems detect?,
and Which of the following network security incidents do you believe you could
do better at addressing if you had the tools?
Most participants (90%) reported that their tools detect abnormal user activity,
which is the main threat they face daily. Similarly, most analysts find that the
tools they have are capable of detecting threats such as phishing emails (85%),
unauthorized access (90%), scanning (95%), policy violations (90%), DoS (85%),
brute force attacks (90%), and viruses, trojans, and worms (95%).
Although 80% of the participants reported that APT is a threat their organisation
faces, only 65% of the participants believe that the tools they have can detect them.
Still, 65% of the participants believe that they can do better if they had the tools.
Similarly, 60% reported that ransomware is a threat, with 65% believing that current
tools detect them. Participants that reported that their tools detect ransomware
are from non-government organisations. Interestingly only 30% of the participants
think they would be better in addressing ransomware if they had the tools.
Insider threats are also a challenge as 45% believe that current tools can detect
insiders, while 45% think they can do better if they had the tools. Although 50% of
participants reported that next-generation malware (more advanced malware such
as bots) is a top threat they face, the participants were divided as 50% reported
that their tools are capable of detecting it, whilst 50% think they can do better
if they had the tools. Analysts from government SOCs all agreed that they can
do better in detecting malware if they had the tools.
Figure 5.3: Which of the following network monitoring tools do you use?
Figure 5.4: How important are the following features in choosing an ideal network
monitoring system?
We show the results in Figure 5.5. Most data sources are monitored through a
SIEM. However, 30% of the analysts reported that they monitor network packet
captures individually, with 10% indicating they do not monitor packet captures at
all. Similarly, 20% of the analysts stated they do not monitor Netflow data, and
15% monitor it individually. 25% indicated they do not monitor website logs. This
is not surprising, as only one analyst reported that web defacement is one of the
threats their organisation faces, as shown in Figure 5.1.
We want to determine which logs are considered important in detecting malicious
activities on a network. We ask: Q24: How important are the following data sources
in detecting malicious activity on the network? We show the results in Figure 5.6.
Overall, analysts find these data sources important. Other data sources that analysts
reported in the comments are threat intelligence and user reporting. One analyst
commented, “User reporting is very important and also mail server logs are essential
unless you are running PCAP device across your mail traffic.”
5. SOC Analysts’ Perspective on Network Monitoring Tools 56
Website logs 55 5 25 15
Web proxy logs 75 5 5 15
Server Logs 85 5 5 5
Network packet Captures 55 30 10 5
Netflow 55 15 20 10
IPS Logs 75 15 10
IDS Logs 80 20
Host Logs 80 5 10 5
Firewal logs 85 10 5
DNS logs 70 5 10 15
Authentication logs 80 5 15
Antivirus logs 75 10 10 5
Anomaly Detection 65 10 10 15
Participants (%)
Monitor using a SIEM Monitor individually Do not monitor I don't know
Figure 5.5: Which network security data sources do you monitor in your work? For
each source you monitor, please indicate whether you monitor the source using a Security
Incident and Event Management (SIEM) tool or individually?
Figure 5.7 shows the analysts opinion on the importance of network traffic features.
Analysts found that source and destination IP addresses are an important feature.
Equally as important is Packet Content, although, it is an issue when traffic is
encrypted. However, as pointed out by one participant, the importance of these
features depends on the nature of the attack in hand.
When investigating a possible incident, analysts look for Indicator of Compromise
(IOC). Figure 5.8 shows the analysts’ opinions of the importance of IOC. 70% found
that suspicious registry or system file changes and anomalies in the account activity
of a privileged user are very important features, both which are host-based activities.
However, one participant noted that they look at a combination of IOCs rather
than one specific IOC, as threat feeds generally group IOCs.
5. SOC Analysts’ Perspective on Network Monitoring Tools 57
Figure 5.6: How important are the following data sources in detecting malicious activity
on the network?
Web refferer 5 25 30 35 5
User-agent string 5 50 40 5
Source Port 10 35 55
Source IP 30 70
Sent/received bytes 20 50 25 5
Protocol 5 50 45
Packet Size 10 25 35 30
Packet Content 10 20 70
HTTP Status code 10 50 40
Domain reputation 10 40 50
Domain Category 10 15 35 40
Destination port 45 55
Destination IP 25 75
Destination Domain URL 30 70
DNS Response 10 65 25
Byte Size 10 15 45 30
Participants (%)
Figure 5.7: How important are the following network traffic features in detecting
malicious activity on the network?
threat). Another participant stated a 200:50 ratio. Both participants are not from
a large organisation. Participants from small/medium organisations have reported
that they investigate more alarms than participants from large organisations, thus
encountering more false positives. The sophisticated capabilities of the tools deployed
by large organisations, due to having more resources, may lead to producing more
reliable alarms. Hence, the overwhelming number of FPs produced in small/medium
participating organisation requires them to focus their resources on detection.
5.6 Assertions
In order to obtain the analysts opinion on various areas of the SOC operation, we
asked analysts to indicate their level of agreement with each statement. We used
Likert-Scale (1: Strongly Disagree, 2: Disagree, 3: Neutral, 4: Agree, 5: Strongly
Agree). We present the full results of the assertions in Tables B.1 and B.2.
Human in the Loop— We wanted to measure how much analysts rely on
their experience in analysing and detecting malicious activity as opposed to relying
on security monitoring system alerts. We show the results in Figure 5.11. 55%
analysts indicated that they rely on their experience over 51% of the time. Overall,
5. SOC Analysts’ Perspective on Network Monitoring Tools 59
Figure 5.8: How important are the following Indicators of Compromise in detecting
malicious activity on the network?
Monitor IDS
Incident handling
Monitoring network and systems logs
Produce technical documents
Provide training and security awareness
Track and trace intruders
Provide and answer a hotline
Perform security policy development
Security configuration administration
Vulnerability handling
Perform virus handling
Vulnerability assessments
Publish advisories or alerts
Perform artifact analysis
Vulnerability scanning
Perform forensic evidence collection
Security product development
Other
Pursue legal investigations
Penetration testing
0 20 40 60 80
Participants (%)
Figure 5.9: Which tasks are you required to carry out in your role?
as the experience of the analyst gets higher the more reliant they are on their
experience and intuition. That is expected, as experienced analysts are assigned
more complex tasks such as hunting. In addition, in Table B.2 — Human in the
loop analysts agree on the importance of the human role in detection, filtering
false positives, and preliminary analysis of events.
Data Fusion— We wanted to determine whether the aggregation of multiple
5. SOC Analysts’ Perspective on Network Monitoring Tools 60
Less Than 5K
I don't know
Over 150K
100K-150K
5K-10K
Other
50K-100K
10K-50K
0 10 20 30 40
Participants (%)
Figure 5.10: How many network security alarms does your organisation receive on
average daily?
data sources are helpful for the analyst to perform their job. The analysts agree
that they are required to simultaneously monitor information from multiple data
sources (M edian = 4, M ode = 4), as well as monitor the state of multiple network
devices(M edian = 4, M ode = 4). Overall, the analysts believe that the tools
they have are adequate technologies for aggregation and correlation from multiple
sources (M edian = 4, M ode = 4). However, the analysts reported that the
heterogeneous data structures or incomplete/inconsistent logs are still a challenge
(M edian = 4, M ode = 4).
Network Monitoring Tools— We asked analysts how satisfied they are
with the network monitoring tools. Specifically, we asked them about the FPs,
FNs, and their use of scripts for the detection of threats and aggregation. Most
analysts agreed that current network monitoring tools frequently produce false
positive alarms (M edian = 4, M ode = 4). However, the analysts’ response on
whether the tools deployed in their organisations produce false-negatives was
inconclusive (M edian = 3, M ode = 3).
Intrusion Detection Systems— As we discussed in Section 5.2, 90% of
analysts reported they used IDS. We wanted to determine analysts’ opinions of on
IDS capabilities. Overall, the analysts’ responses were quite split between agree and
disagree and thus inconclusive. However, analysts do agree that they find that the
number of alerts generated by most IDS is overwhelming (M edian = 4, M ode = 4).
Knowing your Network— We investigate how important analysts believe
knowing the network is in analyzing incoming threats. Overall, the analysts
agree that it is important to keep up with changes in the configuration of the
network and knowing the network environment (M edian = 4, M ode = 4). Most
analysts agreed that they keep track of the network configurations using a database
5. SOC Analysts’ Perspective on Network Monitoring Tools 61
intuition
100 50 0 50 100
Percentage
1% - 10% of the time 31% -50% of the time 71% - 80% of the time I don't know
Response
11% - 30% of the time 51% - 70% of the time 81% - 100% of the time
Figure 5.11: How much do you rely on your experience in analysing and aggregating
data sources to detect malicious activity as opposed to relying on security monitoring
system alerts?
Figure 5.12: How do you choose the security alerts you process?
(M edian = 4, M ode = 4), whilst when asked if they track these configurations using
their memory, results were inconclusive (M edian = 3, M ode = 3). Knowledge of
the network is also important in detecting anomalies on the network. Analysts were
asked: How do you choose the security alerts you process? As shown in Figure 5.12,
57.9% of the analysts indicated that they chose the alerts based on awareness of
normal network activity, and 47.4% said based on the alert severity rating.
Influencing Factors— Our survey results imply that customers from public
sector handle threats and tools differently than other sectors. For example, tools
detecting ransomware were seen mostly used in non-government organisations.
Moreover, a survey participant noted (in one of the free-text comments boxes
included in the survey interface) that answers to questions on SOC operations may
vary based on whether the SOC is in-house or an MSSP: “You should consider the
model of an MSSP in SOC monitoring practices.[..] The model an MSSP will follow
is everything is put in one queue rather than an analyst being devoted to a particular
sensor or customer”. We therefore aim to investigate further the way that factor (in-
house or MSSP) influences SOC processes and identify further factors in Chapter 6.
Threat Investigation— In our survey, we asked the analysts to rate the
importance of a list of network traffic features and indicators of compromise (IOC)
in detecting malicious activity on the network. However, as raised by one of
the participants—“Basically depends on the nature of the attack; the approach is
different”. We explore this further in Chapter 6 to study how such indicators are
thought of by analysts when investigating a possible incident.
Importance of Knowing your Network— We found that practitioners
believe that their knowledge of the network is essential in detecting anomalous
behaviour on the network. Furthermore, analysts reported that they track the
network configurations as it is necessary to provide context to diagnose an alert.
We explore this further in Chapter 7, identifying how such context is vital to
the analysts’ job.
Network Traffic Collection— Only 30% of the analysts reported that they
monitor network packets captures individually. We explore this finding further
in Chapter 7. In particular, when in the incident monitoring, detection and
response process are packet captures collected and examined. In addition, we
will also explore whether SOCs collect and store network traffic and at what
granularity (e.g. Netflow, PCAPs).
SOC Tools— We gained initial insight into the tools analysts use. In Chapter 7,
we investigate such technologies further. In particular, how these tools are used
and analysts’ opinions on their weaknesses and strengths. In addition, we plan
to explore their use of machine learning-based tools.
Malware Detection Processes in Security
6
Operations Centres
Contents
6.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . 64
6.2 SOC Operations Overview . . . . . . . . . . . . . . . . . 64
6.2.1 Preparation . . . . . . . . . . . . . . . . . . . . . . . . . 66
6.2.2 Detection . . . . . . . . . . . . . . . . . . . . . . . . . . 70
6.2.3 Investigation . . . . . . . . . . . . . . . . . . . . . . . . 72
6.2.4 Response . . . . . . . . . . . . . . . . . . . . . . . . . . 78
6.3 Factors Influencing SOC Operations . . . . . . . . . . . 81
6.3.1 Security Tools Budget . . . . . . . . . . . . . . . . . . . 81
6.3.2 Security Clearance . . . . . . . . . . . . . . . . . . . . . 82
6.3.3 Size of Organisation . . . . . . . . . . . . . . . . . . . . 83
6.3.4 Change Management Process . . . . . . . . . . . . . . . 83
6.3.5 Customer’s Risk Appetite . . . . . . . . . . . . . . . . . 84
6.3.6 Maturity of the Customer . . . . . . . . . . . . . . . . . 85
6.3.7 Type of SOC . . . . . . . . . . . . . . . . . . . . . . . . 85
6.3.8 Communication with the Customer . . . . . . . . . . . . 88
6.3.9 Third parties . . . . . . . . . . . . . . . . . . . . . . . . 89
6.3.10 SOC Maturity . . . . . . . . . . . . . . . . . . . . . . . 90
6.4 Discussion . . . . . . . . . . . . . . . . . . . . . . . . . . 90
6.4.1 Preparation . . . . . . . . . . . . . . . . . . . . . . . . . 91
6.4.2 Detection . . . . . . . . . . . . . . . . . . . . . . . . . . 93
6.4.3 Investigation . . . . . . . . . . . . . . . . . . . . . . . . 94
6.4.4 Response . . . . . . . . . . . . . . . . . . . . . . . . . . 95
6.5 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . 96
63
6. Malware Detection Processes in Security Operations Centres 64
6.1 Introduction
In this chapter, we describe the SOC operations, focusing on the previously
understudied role that humans play. We explore the workflows security practitioners
follow to prepare, detect, and investigate a threat — including the kinds of activities
security practitioners engage in daily, the tools they use, and the skills required.
As we identified Chapter 5, security practitioners believe that their experience
plays a significant role in their everyday tasks. We investigate this further and de-
termine the current state of balance between the human and technology, understand
why some malware attacks slip through and identify opportunities for automation
of operational bottlenecks. SOC operations are impacted by various factors that
we elicit during our interviews, such as the type of customer and SOC.
Drawing on the findings from the quantitative study (discussed in Chapter 5),
we designed interview questions to address the following research questions:
Knowledge of Cusomer
Threat Modeling Playbook
Pull-the-plug
Tacit Knowledge
Knowledge of Network
Risk Assessment Hunter Analyst
Asset Criticality Alarm Block
Define Usecases Determine
Technology Escalation Path
Correlated
Baselining SIEM Alarm
Usecases Yes Call Customer
FP?
Escalation Paths
Data Escalation Path
Sources No
Investigate
Check SIEM Further
....
Prioritize Alarms
Escalate
DNS Logs
Anti-virus Logs
Web Proxy Logs
Security Event
Security Tools Check Raw
Customer Logs
Update Customer Profile
6. Malware Detection Processes in Security Operations Centres
Profile
6.2.1 Preparation
Participants across four SOCs (P17, P2, P9, P5), have referred to the “Threat
Modelling” or “Customer On-boarding”. During this process, the SOC aims to
answer these main questions: What threats does the organisation care about? What
does a threat look like? and How should the SOC respond to the threat? The
answers to these questions guide the operations and decision making in a SOC. An
in-house SOC goes through this process for their organisation only. In contrast,
MSSP SOCs need to do this process for each customer. Hence, in this section
we will focus on MSSP SOCs participants reflection on how they liaise with their
customers during the Preparation phase.
What threats does the organisation care about? How analysts monitor,
detect, and respond to incidents depends on the customer’s risk appetite. Risk
Appetite can be defined as “The amount and type of risk that an organisation is
prepared to pursue, retain or take” [188]. The risk appetite for a customer depends
on their Risk Assessment and influences SOC operations. For example, it impacts
the prioritization of alarms generated by SIEMs and how analysts respond to
incidents, ensuring practitioners focus their efforts on detecting threats that target
important business operations. Asset Criticality (an output of the Risk Assessment)
also is considered in the prioritization. SOC practitioners engage with customers to
understand business risks and identify the most important assets. The SOC works
together with the customer to identify such information and document it.
What does a threat look like? Even if technology provides full automation,
humans are still required to for configuration. When asked about how they were able
to detect a particular incident, P5 explained that the analysts were able while the
technology failed to detect. He attributed that to technology not being configured
by the human to detect such a threat. Hence, human are needed not only to detect
things that technology missed but to design scenarios or use cases of what should
be detected (what a threat looks like) and configure the technology accordingly.
When P20 described designing the use-case, he referred to imagination, a human abil-
ity.
P20: All of these things are easily doable, you just have to sit down and
write the use-case, and the limit of your imagination is the limit to what
you can more-than-likely do with most SIEM tools.
The SOC will request the basic data sources needed for threat monitoring
(e.g. Anti-Virus (AV) logs). Then, more explicit use cases are discussed with
the customer. As P17 remarks:
P17: Whenever we take a customer on board, we basically aim for a
baseline. We want your DNS logs, we want your DHCP logs, we want
your AV logs, we want firewalls, we want...you know, all the core stuff.
It is important to identify the data sources required for each use-case, collecting
only the bare minimum needed. As P20 explained.
P20: You have to choose those logs appropriately to meet your use-cases
rather than just say "I’ll collect everything" because then your network
becomes more about sending logs than it does about doing anything else.
So, it’s choosing the minimum number of logs to meet the use-case,
rather than - so yes it can, a SIEM can do absolutely anything you want
it to, as long as you can collect the logs.
6. Malware Detection Processes in Security Operations Centres 68
P9: We’ll then dig into those and look for, is there noise in there we
will look at the tickets that were generated for the customers, and how
many of those will come back from the customers as “well we sort of
want to know about this but we sort of don’t".
P9: When taking in a new customer, the SOC provider in the first
months attempts to understand what’s normal within the customer’s
network.
Through understanding the customer’s normal environment use, the SOC can
adjust the security tools’ parameters accordingly. For example, when deploying an
anomaly detection device, you need to configure the device to alert you when the
network traffic goes above a certain threshold. If the threshold is configured too low
in an environment where the traffic could be high, then this will generate multiple
false positives. So it is down to the analyst to tune the parameters.
P6: A human will have to adjust those parameters [...] or else you’d be
getting high, high, high and it would turn out to be nothing.
The process of base-lining the new asset might take up to a month, depending
on the type of the asset. For example, a webserver will be generating many types
of alarms that need to be filtered out to the point both the customer and SOC
are satisfied. As P9 noted “So then we will track that through on a ticket until
ourselves, and the customer are happy that the output that they’re getting from that
alarming for that box are now appropriate to the new asset type."
How does the SOC respond to a threat? What happens when an alarm
is triggered needs to be determined in advance with the customer. In particular,
how the SOC contacts the customer (Escalation Paths) and how the SOC handles
the incident. The latter is determined when defining the use-cases and documented
in the Customer Profile.
6. Malware Detection Processes in Security Operations Centres 69
Escalation Paths— One of the main objectives of the preparation stage is defining
the escalation paths for each customer. Specifically, when and whom the SOC can
contact from the customer or upper management to escalate an incident. In addition,
the customer needs to know their point of contact in the SOC. Escalation paths
are defined for both in-house and MSSP alike. These pre-defined communications
are defined in advance to help avoid disorientation when an incident occurs. As
P9 remarks on escalation paths:
P17: Then that’s [Threat Modelling] collected on the day, and over a
period of time, it takes quite a while to get host information and network
information from customers because usually they’re all still writing it
down..".
P2: So we are doing that, I believe once a quarter with each customer to
actually continuously improve the log sources and the visibility we have
got.
6.2.2 Detection
The previous step prepares the SOC to monitor the customer’s network and host
environments for threats. In particular, (1) identifying the assets and their criticality
for alarm prioritization, (2) Configure SIEM alarms from use-cases and thresholds
for security tools, (3) Obtain the required logs and deploy proprietary sensors, and
(4) document escalation path and customer’s business and technical details.
The SOC is now ready to start the monitoring and detection of potential threats.
Detecting a threat may occur through reports from end-users or other teams in
the organisation, through hunting performed by analysts on an ad-hoc basis, or
it may be accomplished by using technology.
In this section, we will discuss how the SOC analysts detect potential threats
through the use of detection tools (Technology) and through Hunting.
Technology— SOCs rely on various technologies for monitoring and detection
of forensics. For example, as we discussed in Section 6.2.1, SOC analysts need to
get a baseline of the customer’s normal network behaviour for anomaly detection
tools. As we found in Chapter 5, one of the most frequently used tools was SIEMs.
SOC practitioners (humans) configure the SIEM to monitor and generate alarms
based on the use-cases defined in Section 6.2.1. These use-cases are usually thought
of during the Threat modelling process, from what the scenarios are to what
data sources (logs) are needed. The SIEM then monitors for these use-cases and
generates an alarm when a use-case was flagged.
6. Malware Detection Processes in Security Operations Centres 71
For example, P20 believes that a SIEM can be configured to do what the
analysts needs, as long as you can boil that down to basic logic, and where you’re
going to get that information from. He provided an example of incorporating
physical access security system in the SIEM tool so that if somebody logged on
and they were supposed to be offsite, the nearest camera to their desk can be
activated, and the feed can be observed.
Once the SIEM raises an alarm, it is prioritized depending on the use-case
triggered, the asset criticality (obtained from customer profile), and the predefined
agreement logged in the customer profile.
P17: If we could see that there’s a lot of traffic going out to a host,
say, in a suspicious geographic region, for example, and it looks like
they’re transmitting a lot of data outbound and, say, it was a critical
server internally, domain controller or file server or something like that
then that would shoot right up there in our high priorities to get the
communications blocked, especially if it’s something that we haven’t seen
before.
However, in certain cases, an alarm might not be generated by the SIEM at all.
In other cases, the attack is stealthy (e.g. APTs), and their activities are spread
out across months, hence, not captured by technology.
P15: An alert may never be produced, but just because you have a hunting
capability that can make a more complex human-based estimation of why
something looks abnormal over a 3-4 month period, which the big data
now is trying to address as well, so you are able to query much faster,
much larger datasets.
P15: It’s the hunting human capability, I haven’t at least seen a big data,
autonomous solution that actually is able to give you security insights
on its own.
Hunting is the process of actively going through data sources to find any
abnormal or “weird” activity. As P15 and P19 noted:
P15: The work of a hunter never starts with there is an alert that has
been correlated.
P19: The words I don’t want to hear out of his mouth is “I’m bored”.
Because “I’m bored” means he’s going to go and look for something,
which means I’m going to spend the next ten days running around.
Similarly, P2 and P17 explained how they conduct a hunting session monthly.
P17: Once a month we’ll have a threat hunting session where we’d get
in more analysts for the same day, so that would be a bit more active
team building and sifting through the logs trying to find anomalies.
Analysts seek things that they do not expect to produce security events, yet
they believe are uncommon or not normal. For example, an analyst mentioned
an anomalous activity could be a discrepancy between certain applications and
protocols and usage, or data and usage. In addition, hunting involves looking
for larger or “big” picture discrepancies that do not necessarily produce an alert
for technical reasons. P15 noted:
P15: So the assumption is, you seek for something that won’t correlate,
that appears very normal, very casual, and won’t trigger anything,
won’t trigger a network-monitoring event, won’t trigger a SIEM event
afterwards, won’t do any of that.
6.2.3 Investigation
Once a SIEM reports an alarm or a hunter detects unusual activity, the analyst
starts investigating the event. Although technology might detect and generate
the alarm, it is still the analyst who needs to evaluate that alarm to determine if
its a false positive, benign trigger, and how to act and respond to the alarm. In
doing so, analysts rely on their human cognitive abilities (e.g. pattern matching,
association, reasoning, and computation).
6. Malware Detection Processes in Security Operations Centres 73
P6: And I think that that’s where the human element still remains,
because even when you get an alert, the alert will have to be sent to a
human to make that intelligent decision.
P2: I know everyone says ideally there are processes in place we should
follow. You should do this that and the other. I am not great at that.
I tend to go dig for the interesting stuff first and fill in the blanks
afterwards.
Although the procedures for handling an alarm are defined, analysts fall back
to their human cognitive abilities to investigate an incident. When our participants
described their thought process and actions, we found that although tools such
as SIEMs may assist them in correlating alarms, in most cases, analysts are more
capable in connecting these patterns to detect threats than SIEMs.
P2: Looking for multiple things like chaining together and stuff like that.
Whether it will manage it, I think it is just a bit too much intelligence
to put into an automated tool, and I think you always need that little bit
of human interaction just to say, "That’s not normal".
Similarly, as P17 described, the detection process usually involves looking at multiple
observations and deriving patterns and correlations.
P17: If we could see that there’s a lot of traffic going out to a host,
say, in a suspicious geographic region, for example, and it looks like
they’re transmitting a lot of data outbound and, say, it was a critical
server internally, domain controller or file server or something like that
then that would shoot right up there in our high priorities to get the
communications blocked, especially if it’s something that we haven’t seen
before.
as a file name, mutex, loading or side-loading of a file. It also includes log entries
such as failed log-ins for a local account, privilege escalation, or creation of a new
service on the host that are found in the Operating System (OS) security events.
The unusual network traffic detected by the SIEM may be a result of a malware-
infected host. So when analysts cannot determine the source of the traffic, they
may resolve to remove the host from the network and perform on-demand AV scans
using the latest AV. If it is a definite hit and its definitely a malware, then the
analysts need first to determine what the malware does.
What is the malware? The analysts need to determine the type of malware.
Is it going to attempt to spread, or is it going to try and stay local and hide
itself? Some analysts noted that they would try to determine what the malware
is before isolating the network.
P7: Different malware work in different ways, if it is spreadable it could
affect other machines we have to know that. Does it download further
malware, we have to know the characteristics of what we are actually
dealing with at the time. That is possibly one of the first things we look
at, what can this thing do.
If it is a known malware detected, for example, by a security device, then you
have its name and so have immediate access to the description (IOC). Based on
that information, the analysts may decide to notify the customer to pull the plug.
As P9 noted, “I would see the indicator of compromise, do a little bit of research
as quick as possible in terms of the capabilities of said malware, and then based
upon that is how I would make my decision."
Unfortunately, establishing the type of malware directly from the security devices
may not always be straightforward. The analysts themselves would then need to
research the malware name by looking for the IOC.
P2: If you are looking at abnormal activity, you can say ’ There is
something on that machine, I am not sure what but it is doing really
crazy stuff.’ So you are looking yourself, looking for indicators, You are
looking for changes. So you don’t always know straight away.
Analysts often refer to public sites such as VirusTotal or AV sites, to get
information on a detected executable. Getting information directly from such sites
allows analysts to start incident handling quicker.
P2: However, it’s more important to limit the spread first (So as a
Forensic step). So if I can grab the name of it (from Snort), jump on
one of the many blogs which has already profiled it This is how it spread.
You are looking for port 4898 traffic’ then I can start monitoring for it
a lot quicker. Sandbox, 10-15 minutes to get results half the time people
have done it for you already.
6. Malware Detection Processes in Security Operations Centres 76
However, such public sites, although useful, may provide different classification to
the same malware, and it is up to the analysts to determine which one to rely on.
P7: Its bit difficult sometimes, a lot of AV sites and information, each
have a different intake on what a piece of malware does, so yes it is
difficult to pick which one to use. We do go through forums and such
and if they all say the same thing, yes this thing does do this, we act
accordingly.
Analysts are also faced with a dilemma when using public websites to get answers
as quickly as possible as this risks the malware author knowing their detection,
specifically in more targeted or APT malware. However, P19 remarks: “We also
use the malware site, if it’s publicly available information, obviously we can’t use
internet tools for most of the stuff that we do onsite here.” Similarly, P9 remarks:
“So if you have seen something that could be very targeted, you wouldn’t go to any
of that stuff, because that could indicate to your attacker that you’re aware of it,
and then they can change their methods.”
As a last resort, the analyst may run the malware executable (if they can get
access to it) in a sand-boxed environment to determine what it does. However, this
is usually done in the forensics phase as it takes time to get the knowledge you want.
P2: We do have sandboxes set up. [..] But yeah if you do explode it you
have got a massive 50 odd page report of everything that is going on. It
is going to take you a while to get what you need out of it.
P2: If you logged under the workstation, then set a tail for the username,
IP, hostname, basically monitor everything your host is doing. Stick
that on one screen. Have a summary of events running down, basically
saying ’This is what that machine is doing.
6. Malware Detection Processes in Security Operations Centres 77
Analysts reported that they usually interact with the SIEM platform to determine
these interactions. Therefore, it is vital during the Threat Modelling Process
that they capture all of the log types, that would answer questions about such
communications without requiring to go to the customer onsite.
In addition to looking at the SIEM, the analysts may look at other security
devices that can see any other known indicators of compromise, e.g. firewall. One
participant mentioned that they would start looking at technologies that they have
the “highest respect for or has proved to be effective for a customer”. Usually, for
malware, one analyst argues that the “first and optimal choice” is to have APT-based
available technology to determine if it can say something about the detected malware.
P15: So, the technology that is most contemporary and that has proven
in real-life battles is likely, that we know is available [..]. So, that is the
way we would start looking at it.
Other sources of information analysts may look at for malware is the built-in
technologies, like the Windows advanced security log, or the system log of Windows.
One useful host-based indicator found by analysts for investigating APT is Windows
PowerShell. Although the host user may use PowerShell for automation, obfuscating
the usage of PowerShell gives an indication of something more suspicious. A P15
remarks: “Power shell usage, very sort of contemporary and favoured by the threat
actors, APT-based threat actors.”
Investigating malware depends on what type of malware was detected. Certainly,
in more advanced malware (e.g. APT), analysts have to rely on offline tools to
determine what that malware is capable of. For example, when looking at worms,
analysts need to limit the infection spread by identifying what other hosts it
attempted to connect to. As P9 noted on investigating a worm:
P9: So we would work from the first point of compromise, and look
at where that had talked to next, look at those for any indication of
compromise, isolate and remove those from the network if there’s a
relative transaction, and then work out where they have spoken out to.
Participant P15 noted that interacting directly with the host through remote
desktop tools for hands-on live response is possible but used as as a last resort
and not recommended. It is only possible when the SOC provides live response
service to the customer.
6. Malware Detection Processes in Security Operations Centres 78
6.2.4 Response
At some point during the investigation, the SOC practitioners need to make a
decision on how to respond to a threat. For example, the analysts may need to Pull-
the-Plug, and segregate the infected hosts from the network. Determining whether
a SOC should respond by pulling the plug depends on what the encountered threat.
However, pulling the plug is not always the ideal solution to handle an incident.
Specifically for APT, where disconnecting the host means revealing the detection
to the adversary.
P15: So, if it’s APT, you don’t want to tip of the threat actor, right?
You wouldn’t want to disconnect, unless this disconnect - you don’t want
to move about it really quick, in a way that you may reveal yourself to
the malicious threat actor right."
When analysts determine that what they are seeing is, in fact, an APT, they
remediate as soon as possible.
Pulling the plug is always a business decision that the customer needs to make
themselves. The response decision depends on its impact on the business, and
the pre-agreed escalation paths defined with the customer during the customer
threat modelling. The MSSP SOC is responsible for monitoring and informing
the customer of detected threats and the potential impact it has on the business.
Hence, this decision is not determined by the SOC practitioner alone but through
communications between SOC practitioner and customer decision-makers. It is
then the customer’s responsibility to make the decision.
P17: From our perspective, we’re really monitoring, alerting and notify-
ing customers. We don’t have the authority to shut down communications
or do anything interaction on the network.
6. Malware Detection Processes in Security Operations Centres 79
P19: We’re the security team. We don’t have the business decision
to say, "stop". That’s the business’s choice. I can make a very strong
representation to the business owner, to the senior risk owner and say,
"if you do not do this, this is what’s going to happen, choice is yours".
It would be an interesting discussion.
Pulling the plug sometimes has a major impact on the business. Therefore, this
decision differs based on the customer and its tolerance to risk. Some device actions
are more performance-costly than others. For example, as analyst P15 noted,
blocking a communication costs more than permitting it. So costly decisions such
as enabling packet capture are likely to be crucial decisions that one needs to think
about again upfront. Determining the best action (response) depends on various
variables, requiring human intelligence to make a decision.
Incident Response & Forensics— Finally, once the infected machines are
pulled from the network, depending on the SOC setup, the investigation will be
passed to the incident response or Forensics team, which is a team that might
not be necessarily part of the SOC.
The incident team will advise the customer on how to handle the incident
and possible course of action.
Then based on the agreement with the customer, the forensics team may need
to go onsite to capture evidence, isolate hosts, or take snapshots of machines for
after-event analysis to determine the level of compromise that customers may need
for insurance or evidence for prosecution.
It is the incident response team’s job to work with the customer on the possible
actions they should take to remediate the threat. As P9 noted: “so what we will tend
to do when we take a customer on is build an incident plan that’s specific for their
6. Malware Detection Processes in Security Operations Centres 80
industry or environment, so that we can give them the quickest course to isolating a
problem with the minimum disruption to other aspects of the business.” For example,
if the detected threat was a worm, they inform the customer how it spreads, whether
it is a known worm or a variant of a known worm, and whether it has spread in
the customer’s network. One possible action the incident team may recommend is
isolating sections of the network that are likely to be infected. However, there are
other possible actions that the customer can take that do not disrupt the business.
P9: So if we know it spreads by shares then we can make all the shares
read-only for instance, so that people can still work, but they can’t
compromise the shares if they get infected.
If an APT was detected, the incident response team needs to determine how far
it is in the cyber kill chain, a term used to describe the series of stages of a cyber
attack from the early reconnaissance stages to the exfiltration of data.
P15: And then we go ahead and say, well this is what we suggest needs
to be done to evaluate if this is a single case, if this is just starting, or
has much stronger foothold into your environment, and this is just being
called now while doing a site investigation. This is very common that
you investigate something and you see something else in the meantime
that you need to follow up, or later.
Table 6.1: Factors Influencing SOC Operations that participants mentioned during their
interview.
logs directly rather than have a SIEM to aggregate the logs, slowing down their
detection and hunting abilities.
P19: The biggest problems we’ve got, is a lack of people who are suitable
qualified to use it. We just can’t find enough. I have posts that have been
vacant for 9 months, and I can’t find people. We’re constrained because
everybody has to have a top-secret clearance. And there aren’t that many
people who have a top-secret clearance and are good at Arcsight.
To solve the recruitment challenge, SOCs may train internal employees in cyberse-
curity, As P19 continues: “but we’re training people internally to do stuff, because
I just can’t find people. I think that’s a universal problem. So, we’re starting to
train our own, which is the only way we can do things.” . Alternatively, they
may hire contractors with security clearance as P19 mentioned: “75% of my
SOC are actually contractors, because I can’t get permanents, so they’re temporary
staff, including my SOC manager.”
Tools used— Customers with vetting and security requirements affect which
tools can be used on the network, such as machine learning tools. For example, P18
explained how the tools need to be accredited to be used on the customer’s network.
P18: So, one of the problems we face is getting things accredited for
use on our networks because of the security levels involved. So, I mean
especially things like machine learning which are really cutting edge still
at the moment but they’re really the realms of academic stuff, so getting
something like that approved for use on the network is probably several
years away.
Security clearance also affects how security practitioners learn from the reported
incidents. Once they report an incident, they sometimes do not know the outcome
as it is considered “classified”.
Another challenge introduced as a result is storage. The bigger the size of the
customer’s network, the more logs it generates that need to be looked at.
P9: But it’s quite a mammoth task, especially for bigger networks, it’s a
lot of data that you’re going to have to store, and you might not review
it for a week, it’s a long time to store a lot of data.
On the other hand, some customers are not usually aware of the change in their
network or systems until the MSSP SOC asks them, as P9 remarks:
6. Malware Detection Processes in Security Operations Centres 84
The maturity of the customer’s network topology and network (e.g. how
the customer segregates the network) also has an impact on SOC operations.
As P2 remarks:
P2: We have some customers who are absolutely brilliant, who segregate
their users pretty well and everything goes over a proper router, so we
will actually have logs. Some people have thousands of users sat in one
subnet and you are not going to get any logs.
In addition, whether the internal network is monitored or not affects the detection
of threats that emerge internally, such as detecting worm propagation or insider
threats. As P2 remarks:
P2: We have limited network taps on the internal, between the important
segments of the network. It depends on the maturity of the business.
P17: From our perspective we’re really monitoring, alerting and notifying
customers. We don’t have the authority to shut down communications
or do anything interaction on the network.
It is then up to the customer to make the decision on how they should respond.
This is a challenge when faced with threats such as a worm that propagates
in the network and requires a prompt response. Analysts need to contact the
customer to pull a host from the network. In the event they are unable to get
in touch with the customer, then they can only hope the worm does not spread.
As participant P18 mentioned:
Monitoring for a customer adds pressure on the analysts that they need to raise
every suspicious alarm to the customer. However, this introduces a dilemma about
whether to raise every suspicion to the customer or only raise those that they are
sure about, so that the provider does not look incompetent. As P4 remarks:
P4: Being as analysts, most of them are afraid to not raise them to the
customer. They do tend to raise quite a lot of alarms to the customers,
obviously to be on the safe side.
Making a decision to inform the customer is also magnified when not providing
a certain level of service results in fines. As P18 remarks:
P18: The security teams there are there to maintain availability above
all else because it’s when it’s not available to the [customer] that you
start seeing fines and things like that put in place. So, from a commercial
side, availability is the most important".
Using Playbook— Most SOCs have a manual called a “playbook”, that details
steps an analyst follow to deal with a security alarm. In addition to having
the playbook, the MSSP SOC refers to the customer profile to determine how
the customer wants them to respond to the threat. When monitoring multiple
customers, MSSP SOCs rely on such documentation as the method of communication
of each customer’s requirements..
P9: So we have a playbook for internal events just like we have a playbook
for external events, and the people that we’ll interact with are primed
in the same way, so they know what we will be telling them during an
event, and the same for escalation paths as well.
6. Malware Detection Processes in Security Operations Centres 87
However, in in-house SOCs, such documentation may not be as critical, as the SOC
analysts are all monitoring one network/customer. Hence, when the analysts detect
an attack, they communicate with each other directly and their senior members
on how to handle an incident. As noted by P19, this is much faster than looking
it up in a playbook. However, P19 believes that a playbook is more needed in an
MSSP SOC where they monitor for multiple customers.
P19: But it’s much better, quicker, because our SOC is only the size
of that bit of the room. [showed us - approx 8x6m] Tom [analyst] turns
around and John[ SOC manager] sat 3 foot away, and says "John, what
do I do about this?"
Access to Logs— The SIEM aggregates the logs from the multiple data sources,
and processes them to detect threats. Although obtaining the data sources from
an organisation for an in-house SOC might be feasible, an MSSP is limited by the
data that the customer is willing to share. As a P2 noted:
P17: So, for the most part it’s about understanding the network topology,
but we do try to get involved in their change processes so we can see
when a new server has been added or if a server has been taken offline,
as well as understanding what’s configured on security enforcing devices
such as firewalls or switches. We need to get a bit of information around
that. It depends on how much the customer wants to share but the more
they share the better a job we can do.
Likewise, for patching, knowing which assets are vulnerable to which vulnerability
assists in SOC operations.
For MSSP SOCs, some SOCs assign each customer an analyst (i.e. Technical
Account Manager) as a point of contact. That analyst will be responsible for
understanding the customers’ network and assets. On the other hand, some SOCs
find that having the customer exposed to all analysts provides ease of continuity
and consistency for the customers. As P9 remarks:
P9: For a lot of our customers, it’s quite a stressful experience when
they’re told that there’s something you need to investigate that looks
suspicious. By having exposure to all of the engineers, there’s no sudden
shock of dealing with somebody you don’t know.
6.4 Discussion
The aim of this study was to to provide the research community a better un-
derstanding of the inner workings of SOCs and the challenges that impact its
operations. We summarise in Figure 6.1 the overall SOC workflow, and summarise
the human-centric tasks in Figure 6.2.
Our findings highlight several tensions and contradictions that influence how SOC
analysts prepare, detect, investigate, and respond to threats using technology. We
summarise these findings below, per SOC process module, as broader lessons that re-
searchers and security tool vendors can consider for future contributions in this field.
6. Malware Detection Processes in Security Operations Centres 91
P2, P6, P9, P13, P16, P17, P18 Filtering False Positives/Benign Triggers
Investigation
P2, P5, P17 Correlation and Brainstorming
P9 Forensics
6.4.1 Preparation
Data Sources— As we discussed in Section 6.2.1, the outcomes of the Preparation
process influences how the SOC monitors for and detects threats. Our analysis
revealed that the outcome of this stage has an impact on the technology as well.
For example, the data sources identified (e.g. logs) are fed into the technology,
specifically the SIEM. Based on that data, SIEM correlation rules are written,
and alarms are defined.
Identifying these data sources is reliant on understanding the customer’s network
and systems. Therefore, access to documentation of the customer’s network topology,
systems design, and usage is crucial. However, as reported by the analysts, customers
are still documenting this information. In the absence of such data, analysts need
to find other ways to obtain information required for analysis. P2 explained how
patching reports could, for example, detail the operating system of a host. In
6. Malware Detection Processes in Security Operations Centres 92
the absence of such reports, the analysts need to inspect the network traffic to
get such information manually.
MSSP SOC practitioners hinted that they would want their customers to share
more data. However, SOCs charge customers based on the amount of data they
have; hence, customers may be reluctant to share much. However, this adds burden
to SOC analysts to get that information through other means. As with the patching
logs example, analysts may obtain such data through indirect methods (e.g. network
traffic). Such practice delays the investigation and exhausts resources. This causes
tensions as analysts may be evaluated by the number of tickets closed daily, and
SOC bill customers by the number of triaged alarms.
Configuring Technology— Knowledge analysts gather during the Preparation
stage is the first step of configuring security monitoring tools (e.g. SIEM/IDS).
For example, the knowledge of the customer’s network and “normal” traffic helps
the analyst in tuning security monitoring tools parameters. Similarly, defining
the SIEM use cases requires analysts to communicate with the customer to think
of the possible scenarios. Analysts (humans) determine what threats to look for,
through defining use-cases. Although technology is more effective than humans in
detection scale, still, it requires humans to configure it. However, this dependency
makes configuration prone to human error [14].
Customer Profile— During the preparation stage, SOC analysts document
business information about the customer in the customer profile. Such information
has proved very valuable for analysts, especially in prioritisation of alarms and
determining the best remediation to deploy. On the other hand, other information
that may be equally useful may not be included, such as change management process
reports or patching logs. Such customer information, although important during
the SOC process, is looked at adhoc and on request based on the incident in hand.
Escalation paths are documented in the customer profile, an essential practice
to determine how the SOC and customer communicate when faced with an incident,
preplanned to establish trust. In the case of outsourced SOCs, some SOCs assign a
lead analyst to each customer. On the other hand, this may introduce a problem
when the lead analyst is not present (e.g. on vacation). Hence, some MSSP SOCs
instead choose to have all SOC analysts in communication with the customer to
ensure all analysts are knowledgeable of the customer and its environment. Although
such an arrangement is useful to avoid customer disturbance, it requires MSSP
analysts to be knowledgeable of all customers’ environments, causing pressure to
keep such knowledge up to date and stay informed.
6. Malware Detection Processes in Security Operations Centres 93
6.4.2 Detection
Technology vs. Human— One of the valuable characteristics of the deployed
technology in SOCs is its ability to monitor and detect “known attacks” or “‘known
anomalies” at scale. However, although the technology may identify a possible
threat, analysts need to review the alarms to determine if they are valid or a
false positive. In general, analysts give priority to alarms produced by the SIEM
that are either (1) built using a use case (designed by the analysts based on a
specific scenario) or (2) is an alarm that is a correlation of multiple alerts from
multiple security devices. As described by P15, analysts prioritize technologies
that they respect and which have proven effective for a customer from previous
experience. Nevertheless, reviewing such alarms to filter out false positives is still
one of the analysts’ most troublesome tasks. In doing so, analysts rely on their
tacit knowledge and their knowledge of the monitored environment, built over
experience and documentation of the customer profile.
The aforementioned detection processes are only dependable for known attacks.
Still, the detection of more sophisticated or new attacks relies heavily on humans’
capabilities. Indeed, our analysis revealed that technology is currently deployed in
the SOC as a method to collect the billions of logs, identifying points of interest to
narrow the threat search space. Possible threats are then conveyed to the analysts
as alarms or visualizations. Then it depends on the humans (analysts) and their
ability to correlate and link patterns observed over long periods — “chaining things
together”(P9, P2). This is essential in detecting more stealthy attacks (e.g. APTs),
where activities may be observed in a month’s time.
For example, for hunting, technology is there as a way to organize logs and data
in a structured format (e.g. Elastic Search) so that analysts using their analytical
thinking to query such tools to make connections. Therefore, technology here is
a means for humans to sort through data quickly to detect zero-days or APTs
that may not trigger a network monitoring alarm.
Human-Technology Balance— Overall, the human-technology balance de-
pends on the SOC maturity. The more established the SOC is, the more reliant
they are on technology for triaging and filtering out false positives, having analysts’
expertise in responding to incidents, and hunting for more advanced threats (e.g.
APTs). Therefore, the first objective of automation in SOCs is to automate level-1
SOC analysts’ tasks. On the other hand, in more advanced incidents, there is
still a reliance on human intelligence that some analysts feel is difficult to replace
(P9, P2, P15). As P15 described, “there is still no technology that can give
security insights on its own”.
6. Malware Detection Processes in Security Operations Centres 94
Lack of Skills— There are many mentions of the lack of cybersecurity ex-
pertise [190]. Cybersecurity job descriptions require candidates of a technical
background. Indeed, our SOC participants showed that a technical education
(e.g. networking) is needed in their jobs. However, our analysis has revealed a
contradiction in such assumptions. As described in Section 9.6, there is a reliance
on humans to spot things that “don’t look right” or “weird”. However, such abilities
are not necessarily present in “technical” analysts.
Interestingly, one participant who was a new hire and was able to detect an
incident during their first weeks used to be an accountant. The event was detected
by spotting deviating patterns in visualization. Another alarm was attributed
to a misconfigured IP address. Both were missed by technical analysts and was
spotted by the new hire. The new hire’s “eye for detail”, as described by their
manager, was a valued ability in the SOC, especially level-one analysts. In fact,
the manager was reluctant to provide technical training to the new hire, as such
technical information might distract the analyst, effecting their valued ability.
Nevertheless, technical knowledge is valuable in investigating triaged incidents (e.g.
discrepancies in protocol usage). However, other skills, such as the ability to spot
irregularities, are also valued for level-1 analysts.
6.4.3 Investigation
When describing the triaged incident investigation process, analysts revealed how
they use technology to respond to queries emerging from their analytical thinking.
Again, technology is a method to support humans in an investigation and making
decisions rather than automation. Our analysis revealed a contradiction in whether
analysts should follow their tacit knowledge or follow a set of guidelines defined
in the playbook when investigating the incident at hand.
The investigation is a cognitive process, requiring humans to observe multiple
data sources and make connections using their expertise and knowledge of the
environment. Each person has their unique cognitive abilities of imagination,
association, and pattern matching, all abilities needed in such investigative tasks.
Therefore, enforcing analysts to follow procedures defined in a playbook may
do more harm than good by limiting their intuitive ability. As reported by
P2, such procedures represent a burden for them, and therefore, they rely on
their own methods.
Malware Classification— In certain scenarios where the incident involves
malware, determining the type of malware impacts the decision (P7, P9). For
example, ransomware and worms spread in the network requiring immediate
6. Malware Detection Processes in Security Operations Centres 95
6.4.4 Response
Reporting an Incident— Once an incident is verified, SOC analysts need to
report this detection to the customer. Our analysis revealed a tension in MSSP
SOC analysts reporting. To avoid appearing incompetent, analysts are under
pressure to ensure that the incident reported is indeed an incident and not a FP.
For example, one analyst reported that they spend time investigating a possible
malware infection before notifying the customer. However, delays in dealing with
malware might have huge repercussions, especially if its a ransomware or a worm.
On the other hand, MSSP SOCs may be susceptible to fines if an incident is not
handled on time under the SLA.
Response Decision— Once an incident is triaged, the SOC needs to decide on
the most suitable response. In such decisions, the type of SOC profoundly impacts
who makes the judgment. For MSSP SOCs, the response is not determined by the
analysts themselves but is based on the pre-agreed response documented in the
customer’s profile. As the analysts describe it, they are only there to monitor and
detect threats, and the response is a business decision. The SOC will communicate
the threat and its risks to the customer, and they make the decision. However,
not all customers are at a maturity level to make such a decision. Therefore, some
MSSP SOCs might advise their customers to get them to a business maturity to
make such decisions. In addition, specific response actions (e.g. Pull-the-Plug)
might have a higher cost than others, and therefore, decisions need to be scrutinized
based upon how it will impact the business. Similarly, in internal SOCs, decisions
are made by people up the hierarchy.
Response Automation— Once a decision is made (e.g. Pull-the-Plug), the
type of SOC impacts how the response action is administered. Analysts in internal
SOC inform whoever is in charge of the impacted asset to disconnect the asset
from the network. On the other hand, MSSP analysts advise the customer of
the best course of action, which is then executed by the customer’s security team.
This is mainly a manual process and involves communication between the SOC
6. Malware Detection Processes in Security Operations Centres 96
SOC Manager
SIEM Engineers
SOC Analysts
People
ce
an
ten
Fe
ain
ed
Co
l)
ba
ba
nt
M
er
ex
ck
&
, v
tu
(e
ion
ail
al
.g.
t
ta
Al
FP
(e
en
ar
Communication
s)
c
m
ms
ho
cu
Ad
Do
Patching SIEM
Network Operations Security monitoring tools
Process Knowledge (e.g. Customer profile, Technology (e.g. firewall, AV, IDS/IPS)
Change Management
change management, patching)
Risk Assessment Ticketing System
members and other teams (internal) or SOC and the customer’s team (MSSP).
One participant noted that automation of such a task was seen performed by one
customer (described as scripts) to respond to the incident, but that is still deemed
rare. However, an automated decision may not be as effective as human involvement.
For example, when deciding on the response action for a malware that spreads, its
implication on business disruption needs to be considered. Hence, as described by
P9, an action might be to make file shares read-only to ensure that data necessary
for business continuity is available but still prevent malware from spreading. Such
an orchestrated response currently needs to be thought out by humans.
6.5 Summary
Research on SOCs and their operation are still untapped due to the sensitive nature
of the job. This is a problem for the security research community, as most security
technologies are used mainly by analysts in the SOC. In this chapter, we aimed to
address this gap by describing the SOC workflow, obtained by interviewing a wide
variety of SOCs. From our study, we found that the SOC is still heavily human-
oriented, and they would benefit from automation tools that assist the analysts in
their decision making. We hope that by understanding the inner working of the
SOC, researchers can propose solutions that automate aspects of the monitoring,
detection, and response processes.
Technology has proven to be more useful in detecting threats at scale with the
human direction of what is a threat (e.g. use-cases, signatures). Human involvement
is then centered more on filtering out the FPs. FP filtering requires the presence of
6. Malware Detection Processes in Security Operations Centres 97
knowledge accumulated in the customer profile and existing processes (e.g. change
management) that are currently not incorporated into technology. This transparency
in knowledge is challenged in environments that require security clearance and
acquire confidentiality levels. Our findings revealed a precondition that needs to
be dealt with before attempting automation. We found that communication and
transparency are critical in SOCs operations.
Previous work highlighted the need for communication between analysts and
managers as well between analysts and other organisation teams [12]. However,
communication and transparency not only need to be between people but also
embedded between all SOC elements: People, Process, and Technology, as we
summarise in Figure 6.3, and discuss in the following.
Process-Technology— For example, the process of change management and
patching needed to communicate its outcome to technology (e.g. SIEM) as input to
be considered when designing SIEM use-cases to reduce FPs. Similarly, embedding
patching reports to technology will provide transparency on which asset is vulnerable,
to identify possible immediate threats. Incorporation of the output of the various
customer’s processes (e.g. risk assessments, change management, patching, HR
records etc.) into the technology can result in generating more context-aware alarms.
People-Process— Of course, such process outcomes are not useful unless
people keep it up to date, communicating changes into the process (e.g. change
management process). As we showed in Figure 6.2, customer profiling is a very
human-intensive task; removing such a load from analysts using automation can
release them to work on threat hunting. Currently, such communication between
process and people (e.g. the result of vulnerability scans) is adhoc, done verbally
or via email [129].
People-Technology— One of the most human-centric tasks in the SOC is FP
filtering. Analysts providing feedback to the technology on the reported alarms and
whether they are false positives can facilitate future FP filtering automation.
As we summarised in Figure 6.2, analysts are overwhelmed with mundane tasks.
Automation is needed to automate such tasks to provide analysts with time to work
on more interesting and career rewarding tasks (e.g. hunting). Moreover, we found
that contextual knowledge is important and needs to be incorporated in the produced
alarms. In the following chapter, we continue our qualitative study by focusing on
tool usage by analysts, identifying advantages and disadvantages analysts perceive
from these tools, discussing how contextual knowledge can be incorporated in these
tools and provide recommendations for future tool development.
Malware Detection Tools in Security
7
Operations Centres
Contents
7.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . 99
7.2 The Perception of False-Positives . . . . . . . . . . . . . 100
7.3 Intrusion Detection Systems . . . . . . . . . . . . . . . . 101
7.3.1 Strengths . . . . . . . . . . . . . . . . . . . . . . . . . . 101
7.3.2 Weaknesses . . . . . . . . . . . . . . . . . . . . . . . . . 102
7.4 SIEMs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 106
7.4.1 Strengths . . . . . . . . . . . . . . . . . . . . . . . . . . 107
7.4.2 Weaknesses . . . . . . . . . . . . . . . . . . . . . . . . . 110
7.5 Machine Learning-based Tools . . . . . . . . . . . . . . 112
7.6 Challenges in Network-based Monitoring Tools . . . . 114
7.6.1 Encrypted Network Communications . . . . . . . . . . . 114
7.6.2 Internal Network Traffic Monitoring . . . . . . . . . . . 115
7.7 Towards Contextual Alarms . . . . . . . . . . . . . . . 118
7.8 Discussion: Lessons Learned . . . . . . . . . . . . . . . . 121
7.8.1 Human intelligence vs. Automation . . . . . . . . . . . 122
7.8.2 Contextual Alarms . . . . . . . . . . . . . . . . . . . . . 123
7.8.3 Internal Network Monitoring . . . . . . . . . . . . . . . 124
7.8.4 Customised Machine Learning Models . . . . . . . . . . 124
7.8.5 Technology Evaluation Metrics . . . . . . . . . . . . . . 125
7.8.6 Compatibility with Existing Technologies . . . . . . . . 125
7.8.7 Malware Detection Systems . . . . . . . . . . . . . . . . 126
7.9 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . 126
98
7. Malware Detection Tools in Security Operations Centres 99
7.1 Introduction
As we discussed in the previous chapter, SOC analysts use several tools in their
day-to-day tasks. For example, one of the leading tools used in SOCs are SIEMs.
Alarms generated by SIEMs rely on humans to review them, as most of these alarms
are FPs. For example, NIDS have been found to produce 99% FPs [20]).
Previous research found that false positives can overload security practition-
ers [10]. Consequently, academia has been focusing on improving network-monitoring
systems, specifically addressing the challenge of false-positives (e.g. [192–194]).
However, Kokulu et al. [12] have found that SOC security practitioners do not
consider false-positives in automatic malicious activity detection as a significant
issue in SOC operations. Such findings contradict academia’s current understanding.
Similarly, in Chapter 5, we found that although analysts find the number of
alerts generated by most IDS to be overwhelming, the results were neutral on the
IDS’s inadequacy in detecting threats. We aim to explore this further to determine
analysts’ perspectives on the strengths and weaknesses of these tools and the reasons
why such tools that produce high false positives are not seen as flawed by analysts.
Our survey results show that knowledge of the network is vital in detecting
anomalous behaviour on the network. In fact, analysts choose the alerts they process
based on that awareness. We explore this further to investigate how such contextual
knowledge can be incorporated into tools to reduce the number of false positives.
To summarise, in this chapter, we address the following research questions.
• What are the advantages and disadvantages that security practitioners perceive
of network-monitoring tools?
P2: The SIEM in itself is too noisy. We know 99% of the alarms we
generate false positives, but we still have to look at them.
Interestingly, when we asked our participants about false positives during the
interviews, P9, P4, P15 asked us to clarify if we mean a false alarm or a benign
trigger/noise. In fact, one analyst emphasized the importance of distinguishing
between benign trigger and false positives when evaluating a security monitoring
tool performance and its accuracy. False alarms are used to describe an alarm being
generated without a true security-related event (the boy who cried wolf ). In contrast,
benign triggers are true alarms; meaning they match an existing signature (e.g.
vulnerable java version) but the organisation chooses to ignore it (e.g. due to legacy
systems). P9 provided an example of a benign trigger due to outdated Java versions.
P4: So they are very accurate I believe, but whether they’re accurate and
a false positive, now, it depends on how you perceive a false positive. So
if we raise an alarm to the customer, and they’re aware of it, they will
just call out, we are aware of it, but it’s a false positive in that sense
Hence, an IDS detecting benign triggers does not mean it is not accurate.
It is accurate because as it is the same trigger a malicious actor can utilise.
Hence, labelling such benign triggers as false positives gives the impression that the
technology itself is flawed when it is detecting what it was designed to detect.
7. Malware Detection Tools in Security Operations Centres 101
P15: So, you have a very high number of events right, and it doesn’t mean
they’re a false positive, but it means many enough are again triggered by
benign triggers. Which means the condition is perfectly matched, and
the filter as such works, but the circumstances are completely legitimate.
This is benign, because the purpose is business-justified, and it’s not
malicious, it’s just manifests in the same way as particular malware
would, right.
7.3.1 Strengths
Good first indicator— Some analysts found IDS Alerts to be a good first indicator
for an intrusion. As P4 explained “Most of the anomalies are basically network traffic
because that’s the first point of call for any sort of intrusion”. Similarly, P2 reported:
P2: So IDS analysis and network traffic for us performs very well. It is
a very good early indicator that something is going on.
Deal with high volumes— One of the reasons why IDS is very popular is
its capability to handle and cope with high data rates (high volume of traffic),
as we discussed in Chapter 5.
P15: If we look at the volume itself, the top talkers are without any
doubt the IPS/IDS appliances. And that is expected, given the relative
detection uncertainty that they apply to implement detection as such,
like regular expressions and anomaly-based detection right.
7. Malware Detection Tools in Security Operations Centres 102
Well Maintained— Having the tool be well maintained is also a sought after
tool property. For example, updating the tool with signatures of the latest threats.
P4: In terms of the good parts of the tool, I believe Suricata or Snort
[signature-based NIDS], because it’s pretty well maintained and there
are new rules every day added to it, I think it’s probably one of the best
tools for network intrusion detection systems.
Easy to Write Signatures— There are three types of signatures deployed in
a NIDS: (1) signatures released by the product vendor, (2) signatures shared by
the community or other organisations, (3) user-defined or use-case-based signatures
(e.g. admin logins, TOR). Some analysts think that one of the good features
of intrusion detection systems is the capability of defining custom signatures —
User Defined (UD) signatures. Every organisation is different in its networks,
systems, users, and its use. Defining custom configurations that work for that
environment is a sought after trait.
P9: In terms of the good features they have capabilities for you to deploy
your own customised signatures, you know, UD, you know, user-defined
ones.
Creating custom signatures also assists analysts in providing a quick fix to a
new threat. For example, in case of new malware, analysts can write signatures
to detect it based on the malware known Indicators of compromise (IOC). As
P9 pointed out in the WannaCry case:
P9: So like WannaCry, instantly, we wrote as many signatures for that
as possible, for everything we could, and then put it in place. There
were already signatures there, but we just wanted to be as thorough as
possible.
Work Across Protocols— P18 and P6 also commented that the ability for
the signature-based IDS to work across protocols is an advantage.
P18: I suppose the good thing about the IPS is that it can work across
protocols.
7.3.2 Weaknesses
Mostly False Alarms– A high number of alarms generated in the SOC are
attributed to NIDS [20]. As P4 explained: “We do tend to get a lot of alarms based
on the network traffic, and 90% of the time they are based on the Suricata rules.”
Similarly P6 explained: “I think that most of the tools will generate a higher volume
of false positives, especially when it comes to a signature-based event.”
However, some analysts find that these alerts are mostly benign triggers rather
than attributing to a malicious threat.
7. Malware Detection Tools in Security Operations Centres 103
P15: We had many detections that are accurate, as much as the filter
mechanics is concerned, but I can’t recall to be honest, a genuine
malicious detection, attributable to TippingPoint or any other IPS as
a detecting device. And this is concerning, given that at the same time
they represent most of the volume of the SOC, when you come to think
of it.
Likewise, P18 explained that a poorly written signature will result in many
alarms, that analysts can not review, rendering the signature not “useful”.
P18: Yeah, it’s kind of an internal battle of having a signature for the
latest threats and having a useful signature because if it’s just constantly
firing, nobody’s got the time to review all of them, so it has to be well
written and they’re not always.
Black Box— One of the disadvantages of IDS/IPS is that they are closed
systems. Meaning that the analyst receives an alert but does not know the reason
for the detection. The analysts themselves need to lookup more information about
the alarm using, for example, the filter name to understand what the alarm is for.
P15: You know the filter name and you have the filter description,
and this may or may not be tied to a CVE number, for a particular
vulnerability, which you can then read up, see what operating system,
software, etc., it is related to.
Likewise, P4 explained that alarms produced are not clear and require analysts
to search further for the source of the alarm.
P4: I think the fact, I think your last question about the anomaly-based,
although it gives us some sort of indication, it’s not very clear and It
does require us to dig in deeper and find out where the source was.
7. Malware Detection Tools in Security Operations Centres 104
However, such ambiguity not only requires more effort for the analyst to
investigate, as P15 explained, they also lose their trust in the validity of the alarm.
P15: So that’s one issue of traditional IPS/IDS, that we also don’t really
have high respect for, because if you don’t tell me the reason you fire,
you’re not ready to open up the reason you produce an event for me,
and I have only a very high level of description that may not be usable,
it’s not usually usable, let alone actionable, then I can’t have really high
respect on that.
Loosely Written— Some analysts reported that signatures are “loosely writ-
ten”, meaning that it is written to be general to capture a variety of activities,
thus resulting in a high volume of false positives.
P6: I think that most of the tools will generate a higher volume of false
positives [..], especially when it comes to a signature-based event. We
have IPSs and we get signatures from the vendor. Some of the signatures
are written very loosely in order to allow it to capture a wide range of
activities and those are some of the downfalls of using it
This is also true for signatures that are designed to detect malware. As the
malware evolves and generates various attack behaviours, signatures can be written
to be too broad to capture all these possible behaviours.
P18: Malware can change so quickly now that, yeah, a lot of their
signatures can be very loose and very broad to try and capture every
possibility.
Similarly, when creating malware signatures, they are written to capture part of the
malware activities. Although, the alarm is triggered, the analysts are left confused
about which kind of malware triggered it. P4 explained:
P4: But what happens is, sometimes when they write the rules, they
only write to capture a part of the malware and not all, not other pieces
of it. So what happens when the rule triggers and we have an alarm, it
may give a false indication that that alarm is for this particular malware,
when in reality, it’s for a different one. And that could be, that is an issue
because as malware perform different, sort of, actions on a system, we
don’t really know what sort of remediation advice to give to a customer.
Whether it’s for that malware or whether it’s for this.
One analyst provided an example of how such loose signatures results in the
tool generating a false alarm. Such an alarm was determined as a false alarm
on examining the network traffic.
7. Malware Detection Tools in Security Operations Centres 105
P9: But some of the signatures are very poor, so they will look for
the words "select" and "from", in clear text, in a packet, but it could
be "from" and "select" in a packet, there’s no context applied, or "drop
tables" a classic one, so we have a customer who sells a [Drop Table],
I’m assuming it’s a fold-down side table, every time they sell one of
them it fires an alert, but it fires it on the SQL injection alarm, so you
can’t turn that signature off, as poor as it is, because that’s one of the
SQL injection alarms that’s part of that product set for.
However, such a manual inspection of the packet might not be possible with
encrypted traffic, as P9 continues.
P9: which is quite time-wasting really, but then if the limitation is, so
if that traffic is then encrypted, you might get the alarm but you can’t
see the raw text.
Lack Context— Some analysts remarked that one of the reasons that IDS
alerts are not reliable is that it does not perform contextual analysis. Meaning, that
the alert lacks context that might help the analysts determine if its a true alarm.
P15: It’s a bit of a paradox there, you have quite a bit of an output, at
the same time, when the detection is accurate, and the regular expression
is correct and everything is fine, it doesn’t really mean it’s malicious,
and the reason is, it doesn’t do usually contextual analysis, it doesn’t
add more event sources into it, which is what then the SIEM does, and
the analysis of the human element does, of the human actor.
For example, some network intrusion detection systems do not capture the
packets that triggered the filter. This may be due to storage limitations.
P9: Some of the devices will be configured to give you the trigger packet,
so the one packet that contains the string that triggered the alarm. In
order to kind of contextualise some of these trigger packets, it would
be useful to have the packets before it, the packets after it, so you can
understand what was going on in the bigger picture. Where we do have
that, analysis is just made loads easier. Where we don’t have that, you’ve
got to start making assumptions, which is where a weakness could lie.
7. Malware Detection Tools in Security Operations Centres 106
However, some analysts find that more products are now adding more context to
their alarms, hence are getting better as P9 continues to explain.
P9: So, things are getting better with that, Cisco have sort of got rid of a
lot of that technology and moved on with the SourceFire and FirePower
stuff, so that’s more contextualised but yeah it can be a real challenge.
Detecting New Threats— In general, most IDS systems used by SOCs are
signature-based, as P16 remarks:
P6: From the network monitoring its all still signature-based, so natu-
rally if there is no signature it will not cause an alarm.
P2: Yeah, so with a lot of the APT malware you don’t notice it on the
way in. Snort signatures aren’t written for it, just because it is not very
well known. So a lot of the time it will get through.
P18: So, with something like WannaCry when it first hit on the Friday
afternoon, we didn’t have a signature for it. So, that was something we
needed to develop as quickly as possible and then, of course, again on a
huge network it takes time to deploy all this stuff.
7.4 SIEMs
As we discussed in Chapter 5, 80% of our participants use SIEMs. SIEMs replace
the need for analysts to access the traditional security tools, sometimes called
eye-on-glass, directly.
7.4.1 Strengths
Visibility — SIEMs are capable of handling large numbers of security events
from multiple sources [195].
P2: So any one source on its own, I don’t find that useful. That is why
most SOCS now, the core bit of technology is a SIEM, because you can
tie all the different technologies in together.
Hence, the SIEM can see all logs and alerts generated by security devices you
connect to it, as P2 explained: “Yeah, so we have a SIEM platform that ingests
logs from all sorts of stuff.” Everything goes into this one central repository, so all
the logs and alerts are in one place. In addition, the data sources are monitored
continuously to make sure they are working. Thus, if the log data source becomes
silent for some time, the SOC receives a warning saying, for example, “You haven’t
received an IDS alert for 20 minutes.” This functionality allows analysts to have
visibility of the monitored network and systems.
Analysts having access to host and network logs is attractive, as the more
information, they can view the better their situational awareness [195].
P12: The good thing is we have recordings of all the networks and all
the emails and everything, so all the files and directories that any system
user will have so if there’s any malware detected or any malicious mail
received by them we all have the same tool, all logs are collected via
SIEM, a so in that way it’s good that we have visibility to all the systems
via SIEM.
P19: So they will tell arcsight, “I want to see how many people have
logged onto so-and-so in the last 24 hours”, and it goes “dzzzz” back.
For example, defining a use-case and the required logs to monitor that use-
case will reduce the size of the acquired logs to the bare minimum. This makes
storing them more feasible as P2 noted: “You can store it that bit longer because
you don’t get too many of them.”
Moreover, defining use-cases and the bare-minimum logs needed ensures that
analysts are not swamped with far too much information. P20 noted:
7. Malware Detection Tools in Security Operations Centres 108
P20: I mean a SIEM tool can do pretty much anything you want it to,
as long as you can see how you can boil that down to basic logic, and
where you’re going to get that information from. As long as you can do
that, then the SIEM tool can do absolutely anything.
P20 explained how an analyst could write a use-case that integrates the physical
security system into the SIEM. In such a use-case, the SOC can detect log-ins
originating from users who are currently offsite. P20: “a log-on by someone who
has left the organisation is a particularly good example of how it can work well.”
However, such detection is only possible if the SIEM data (e.g. employee leavers)
is up-to-date and maintained.
Cross-event, Cross-Platform Correlation — One of the main advantages
of SIEMs over traditional IDS is its capability of correlating multiple alerts and
log events generated by heterogeneous platforms.
P16: With the network threat it all goes Scericata (or snort of whatever),
where I think if the, it cant do like comparing 2 events at the same time,
is just an IDS really, it can’t put the 2 together. If it was going to the
SIEM platform that would be easy, because that is what it does.
For example, as described by P2, a correlation rule could be “If you have seen
an IP doing any one of these activities, followed by authentication success or a
large amount of data transfer to it, generate an alarm.”
Correlation Rules/Directive alarms could be alarms that monitor the network
for “anomalies”. For example, one SOC had correlation rules for noticing spikes
in traffic. These rules learn normal traffic patterns over two weeks by looking at
host traffic, creating upper thresholds. Thus when traffic for a host goes beyond
that threshold (threshold determined during the baselining phase in preparation),
an alarm is generated.
P2: It’s a bit of both. You signature it, you write the rule and then you
baseline it from there, so it’s a learning rule you would call it. So you
are basically saying learn what activity is there for several days and then
monitor what changes for the next however many hours.
SOCs usually have a fairly extensive set of correlated rules, which the analyst
is expected to react to immediately. The analyst find these rules effective: P2
explained “So yeah we do correlation rules like that. It proved quite effective, just
because you are generating that event yourself.”
7. Malware Detection Tools in Security Operations Centres 109
P9: So, bit of traffic bouncing off your firewall is not super-interesting,
a host sending a bit more traffic than it normally does, you know peaks
and troughs in traffic do change, but join those two together within your
SIEM, now that’s suddenly become more interesting because that’s been
persistently hit externally and now suddenly more things are going out,
so that’s a type of anomaly that that kind of kit would spot.
There are several reasons why a SOC may define directive alarms. Defining
correlation rules or directive alarms allows the SOC to store these alarms for an
extended period so that correlating these alarms over time might allow capturing
more stealthy attacks such as Advanced Persistent Threats (APTs).
P6: The analysts that you’ve got run in 24-hour shifts, session, so if
something is missed during the day, they still have to run historical,
maybe do trend analysis, you know, where you analyse, maybe, a 24-hour
period log against the previous 24 hours. So, chances are, you’ll be able
to pick something that was missed by whoever was on shift before you
came on and rectified and those logs will still be there.
P9: I think how the SIEM platform is configured to receive that in-
formation and then deal with it accordingly helps, so it’s a lot easier
for manual analysis, when things are structured properly or categorised
properly..
P2: So the SIEM platform itself will assign a risk based priority score
to the alarm that it generates. These scores are based on the criticality
of the host, where it sits in the network, the criticality of the alert. So if
it is a strong signature which is only seen as information it assigns a
lower score. Basically the computation in the background assigns a score
to it, and obviously the higher ones we jump on as quick as possible.
Fast MTTR— One of the metrics used to measure a SOC’s maturity is the
time needed to take action and neutralize the threat is Mean Time to Respond
(MTTR). SOCs aim to lower the MTTR, and the SIEM’s ability to aggregate
datasource helps to achieve that. However, some customers may not have a SIEM,
and analysts have reported that they would need to go onsite and look at the
host. Specifically, they look for historical evidence of how the infected machine
interacted with the network. To determine that, they would examine the file shares
that can access it, local logs stored on the infected host, switch logs in switches,
and firewalls that it may have passed through.
Nevertheless, the best option is always to have such information feed into a
SIEM. As a participant noted, going onsite takes time and delays the investigation.
P9: Even with the best will in the world, you might take an hour, two
hours to get somebody to somebody’s site, if those logs are available
for inspection within the SIEM, you’re going to be investigating within
minutes.
7.4.2 Weaknesses
Overwhelm Analysts— The SIEM’s ability to plug-in data sources, although
attractive for comprehensive monitoring, might result in a huge collection of data
and alarms that overwhelm the analysts. Hence, when an analyst wants to find
particular information, it might not be a straightforward process.
Therefore, it is best practice when deploying SIEMs is to properly design the use-
cases, and collect data needed only for these cases, as noted by P20, P11, and P3.
P20: It’s just a case of getting the bare minimum that you need to meet
your security use-case.
7. Malware Detection Tools in Security Operations Centres 111
Use of structure databases— The volume of data becomes even more prob-
lematic when the deployed SIEM uses structured databases that take time to
retrieve queries.
P13: When handling large amounts of data in terms of alarms, it’s
not got quite as much resource to handle it, cause at the moment the
platform manager uses the Microsoft SQL, so when handling big data,
structure database doesn’t really work for it. In terms of just taking in
data, it can handle it cause it uses a ElasticSearch and backhand, but
the actual alarms are still structured.
P2: The SIEM in itself is too noisy. We know 99 percent of the alarms
we generate false positives, but we still have to look at them.
Cost— One of the main challenges of SIEMs is cost. A SIEM platform with
hundreds of connectors and archival databases may require $3 to $5 million of
upfront hardware cost as well as a yearly maintenance cost [195]. Moreover, adding
additional functionalities as “plug-ins” will accumulate an additional cost.
P12: Pattern discovery, yeah, so you have to buy this stuff separately
and it will do all your IOCs together, its so smart but in other tools
like Splunk we don’t have that pattern discovery. But we have to do it
manually.
P20: They [SIEM] are good at detecting anomalies, but if you’re clever
enough to make sure that you don’t look like an anomaly then the SIEM
is not going to do you any good, then you’re reliant on an analyst saying
– that’s fairly normal but it just looks a little bit odd. And then you’re
reliant on the human being making a connection.
P15: So that’s the sort of behaviour-based approach that you don’t have
a-priori built-in conception of what to look for, but you see what it does,
machine learning,
Furthermore, P1 explained that they are planning to deploy a tool that applies
ML to network analysis in the next year.
P1: When it comes to the network side, we are in the process to sort
of look for machine learning activities in there. But right now it’s just
based on signatures.
P1: I really like to, you know, it’s, in a perfect world it will be, all the
processes would be known by my machine learning tool or the AI tool.
P18 explained how machine learning is being used to automate FP filtering, one
of the most human-centric tasks as we found in Chapter 6. Such automation
functionality is often referred to as “orchestration” [196].
7. Malware Detection Tools in Security Operations Centres 113
P18: Well, I was at InfoSec yesterday and it seems that a lot of the big
vendors think it can be. They were very heavily pushing their tools that
do away with all of the false positives for you, or they’re supposed to, to
the point where it was even kind of raising a ticket for you and things
like this. I can’t remember what they called it. Orchestration.
P17 explained that they attend conferences and vendor demos to keep up to date
on the latest technologies. However, they have not seen commercial products that
stood out, and they were skeptical of the tools’ application of machine learning.
P17: When I’ve looked at machine learning in the past it seems like it’s
not entirely machine learning and they say they’ve got sensors deployed
which have got algorithms that get updated by the vendor, but it really
sounds like they’re updating some form of signatures on there.
P9: I think it would be good to get some of the noise away that might be
generated within there, although one of the big tasks that comes through
the SOC is the cost of baselining, taking the noise out of the equation.
P18: So, one of the problems we face is getting things accredited for
use on our networks because of the security levels involved. So, I mean
especially things like machine learning which are really cutting edge still
at the moment but they’re really the realms of academic stuff, so getting
something like that approved for use on the network is probably several
years away.
7. Malware Detection Tools in Security Operations Centres 114
When the deployed tools perform content inspection, encryption renders them
obsolete, as P9 explained.
P9: If the traffic is encrypted, we won’t be able to do that [content
inspection], unfortunately, but provided it’s not, yes it will take full
packet captures and then it will run the Snort signature base through all
the packet captures, and see what it can match on. So yes, I would call
it like content inspection, yes.
In fact, analysts find that host-based tools provide more useful information and
are more accurate. As P17 explained, host-based solutions provide insight into
all of the processes and activities occurring on the end host, while maintaining
the host user’s privacy. Similarly, P2 explained:
P2: The most accurate ones are normally the host agent ones, so we are
looking for a sequence of things. So a change to a registry key followed by
an application starting with a parent process which is normal, followed
by network traffic out to a suspicious network block and stuff like that.
7. Malware Detection Tools in Security Operations Centres 115
P9: I know from what you hear in the industry, a lot of people put a
very very large focus on protecting themselves around the perimeter, but
then, say you do manage to get through that, a lot of networks they don’t
spend half as much time securing the innards of it all, so once you’re
actually in, a lot of the time it’s very easy to just propagate throughout
these networks and gain a foothold
P9: I think that’s where people start wising up, because things like
the WannaCry, it got into people’s networks, but it also entered at the
perimeter, but what people are struggling to find at the moment is they
can’t find whether they got it via an email, and somebody clicked on
something they shouldn’t, or whether they had a public-facing share and
that’s how it got into their network. And there’s quite a lot of companies
that just don’t know yet [...] But because people aren’t monitoring at
that level on the inside, that bit is a little bit of a grey area for a lot of
companies at the moment.
P17: So, you’d either need full packet archive or NetFlow and knowledge
of what the internal subnets look like. So, we collect the info on what
internal subnets are used for what. So, a workstation would be LAN
specific department network and where the servers live and we could use
NetFlow to detect how much traffic is being transmitted between two.
7. Malware Detection Tools in Security Operations Centres 116
There are several reasons why they do not monitor internal traffic. For example,
P4 mentioned that it is hard to monitor internal traffic because of geographically
how the network topology has been configured. So if the SOC monitors the
internal part of the customer’s firewall, then when host A tries to connect to host
B, and they are on the same network, that traffic will not be captured because
it will never hit the firewall.
Another main challenge is storage. Capturing and storing full packet captures
of the ingress and egress points to the network alone is costly. Thus, to additionally
collect internal network traffic in even greater volume is a challenge.
P9: And again, that’s where you get into vast amounts of storage, if
you’re going to capture all the packets of all of that internal interaction,
at the perimeter you’ve got a few devices, and then it’s whoever lands
at the perimeter, talking outside, but there’s lots and lots of traffic
happening all the time on the inside, that if you were trying to capture
that’s quite a vast amount of data to sort of look at.
Similarly, an MSSP SOC, collecting PCAPs depends on how much the customer
is willing to pay for storage. Although they cover the customer perimeter network
connections, monitoring internal traffic is still costly for the customer.
P2: We cover all ingress and egress into networks, but it is how much
the customer wants to pay for, for us to cover within the network. Most
of it at the end of the day comes down to storage and how many logs
we can process a second for them. There is always a cost and they are
always trying to save money, and it’s normal, "Well as long as we’ve
got the internal ... in and out covered then internal shouldn’t matter."
It is a very alternate way of thinking. We don’t like it but ... yeah.
Internal network traffic may still be visible in the SIEM, but that depends on
whether use-cases that the customer defined rely on part of the internal network
communications as a data source. For example, the customer might be interested in
monitoring the internal network traffic of specific servers. As such, the SOC defines
use-cases that monitor that part of the network. As P2 mentioned:
Some SOCs may attempt to monitor the internal network behaviour through
the host logs. However, when defining such use cases, baselining what is considered
a normal activity for an user asset is a challenge.
7. Malware Detection Tools in Security Operations Centres 117
P17: At the moment we don’t have any alerts on that kind of traffic
because it’s hard to really determine for that side what’s normal really
because sometimes people log on to servers, SharePoint or something,
download a lot of files because of working on some big documents, so
that would be quite challenging to baseline, but we definitely use that
during an investigation. So, then whilst we’re looking through the traffic
for a host we’ll see the logs related to that host so if they’ve connect to
a bunch of servers which they don’t usually connect to then we can flag
that as well.
PCAPs are useful in understanding “what they are dealing with”, but are usually
only examined after a SIEM alarm detected malicious activity. Some SOCs collect
the PCAPs but only look at it during the investigation process.
P9: So you can go through, review the traffic manually, it’s an immense
job, you wouldn’t want to review all of the traffic manually, without
having some sort of intelligence to point you to a particular thing. But
that’s one way, and I suppose it works the same in Splunk or any kind
of big data platform - you can manually go and review it, but with the
quantities of data and stuff in there, it’s going to be very difficult unless
you have got something to point you in the right direction
Some SOCs monitor the internal network through IPS. However, the size and
scale of the network make it difficult to push signatures to all IPS. Such as in the
WannaCry attack, as P18 discussed in Section 7.3.2.
To overcome storage challenges, some SOCs are attempting to collect internal
traffic communications in alternative formats such as network flow (e.g. Net-
Flow). When asked to explain the reasons they do not collect internal network
traffic, P17 replied:
One of the main questions we were interested in is at what stage of the process
is network traffic analysis useful. Network traffic can be captured as a packet
capture (PCAP) or in a NetFlow format. As we have shown in Chapter 5, only
55% or our participants monitor network packet captures and NetFlow through
the SIEM. However, some analysts find NetFlow to be only helpful in forensics
7. Malware Detection Tools in Security Operations Centres 118
and so these are not actively monitored (20%). In addition, 80% use Wireshark
to analyse network traffic.
P15: NetFlow is not so much popular, although it is to be found for
some forensics purposes, so for ad-hoc specific tasks, but this doesn’t
usually feed into security-event management systems.
In contrast, some analysts have configured the SIEM to collect PCAPs of the
network when it receives an alarm from a network-based IDS. Although the intrusion
was picked up by the NID, investigating the alarm and determining why that alarm
was fired was done though the PCAP.
P4: If it’s a NID-based alarm so through Suricata or Snort, then
AlienVault automatically performs a capture, a PCAP capture, so we
can download the PCAP and look at why that alarm was triggered, and
then looking at the PCAP we can understand what is because it matches
certain criteria which that that rule was looking for ..
Although some SOCs might attempt to collect the customers’ network traffic,
they may keep them for a day and a half to three days. After that, they will store
the metadata, which will go back a couple of months enabling them to perform
retrospective searching if required. Similar to what we discussed in Chapter 6, some
SOCs only start collecting the PCAP after they start investigating an incident.
So they will take a PCAP of the traffic for 120 sec, then wait for a couple of
minutes before collecting another one.
For example, one of the main weaknesses reported by the analysts in the tools
they use, discussed in Section 7.3.2, is the lack of context in the signature description,
leaving them to determine the cause of the alarm, adding more effort to a very
tiresome task. We discuss in the following how adding context to alarms can help
analysts quickly determine its genuineness.
To be able to determine if a network activity is “normal”, the analyst needs
to know the monitored network. This knowledge spans, for example, the network
topology, network devices, what these devices are used for, location, and the
owner of these devices.
For example, the analyst might need to know the connection patterns of each asset—
i.e. what services are running on this asset and what are their frequent connections.
Before analysts start looking at logs, they need first to have an awareness of
what they are looking at. So they go back to the customer profile to determine,
for example, what is the server for the customer? Is it a domain control or a
web server? Is there a new server being tested? The abnormal activity can be
an indication that the server changed or a different IP address was reassigned
to the server. Such changes, as well as benign triggers, asset usage, and other
information about the network, are recorded in the customer profile (i.e. wiki). The
analyst may also contact the server administrator to validate abnormal activities
or check historically in the logs to see if there was any similar connection seen
previously and at what times.
Knowledge of the asset and network helps practitioners eliminate false positives
and determine the path of investigation. For example, knowing an asset is a windows
machine will assist when receiving an alarm for a Linux signature for that machine.
7. Malware Detection Tools in Security Operations Centres 120
Not only do analysts need to know the network they are defending, they also
need to be aware of the customer’s business. Knowledge of the customer is critical
in many aspects in the SOC operations, from ruling out false positives to making
a decision on how to respond to a threat.
P4: So it’s not just about what you see in the security events, or in
the tools, it’s about what you know of the customer and the customer’s
nature, you know, business nature.
Knowledge of the customer and the nature of its business helps analysts in
investigating an incident. One participant provided an example of how valuable it
could be for determining a false positive. On detecting large outgoing connection
from the customer to another company, after spending time investigating, they
researched if there was any business between these two companies. They found out
that their customer acquired that company recently. Similarly, P16 remarked on
how knowing the customers’ working hours helped in filtering a false positive.
7. Malware Detection Tools in Security Operations Centres 121
Analysts might acquire knowledge from third parties, such as ISP, security
vendors, and the security community to investigate threats. For example, the ISP
might inform the customer of any changes in the network they need to be aware of.
However, that information might not reach the SOC, especially an outsourced SOC.
P6: You’ve got [ISP] and sometimes you’ve got to talk to each other,
but it becomes difficult because, for instance, [ISP] would be doing some
configuration change on setting a limit for the network. They might tell
somebody within [customer], but that information may not necessarily
come to you.
Security vendors are responsible for maintaining their products and providing
signatures for new threats. P6 discussed how they rely on security vendors to
share that knowledge, pushing out any new signatures and any new indicators
of compromise (IOC).
P6: Something like WannaCry, for example, the vendor also has the
responsibility because they have to maintain activity of their system so
when it happened, straightaway McAfee, for instance, was on the case
developing signatures ready for it to be deployed.
The security community can also assist the SOC process through intelligence
sharing, writing blogs about how a threat behave, and spreads. As we found in
our survey, 35% of our participants triage alarms according to newly announced
vulnerabilities in security blogs. The SOC practitioners search online for information
about a threat they encounter to determine the best countermeasure to deploy.
P2: So WannaCry was great, because straight away people were publishing
vlogs, this is how it spreads. This is what you need to look for, and you
could go and look for it before there were any signatures for it. We were
finding things quite quickly.
P4 Well Maintained
Strengths
P6, P9,P18 Easy to Write Signatures
P9, P6
Fast Signature Creation
Traditional P6, P18
Tools Work Across Protocols
(IDS)
P6, P4, P2, P9, P15 Mostly False Alarms
P6, P18, P9
Loosely Written
Figure 7.1: Summary of SOC tools weaknesses and strengths, and the participants
agreeing with each factor.
Section 7.3.2 and Section 7.4.2. This is due to humans’ ability to recognise patterns,
imagination, as well as the social ability of communication as discussed in Chapter 6.
Research is still far from integrating human cognitive and social abilities
into SOC solutions. One emerging field to attempt to tackle this challenge is
“Cognitive Security”. Cognitive security leverages multiple forms of AI, including
machine-learning and deep-learning networks to uncover the human cognitive ability.
Probably one of the leading research in this area is IBM Watson1 . Further research
is needed in developing human-machine cognition solutions in SOCs where humans
and machines work together rather than using full automation.
and systems governing them. We identified in Chapter 6 the factors that impact
SOC operations, and thus tool usage. Accordingly, although SOCs may deploy
the same technologies (e.g. SIEM), these technologies are configured and used
differently and should not be developed as one size fits all.
Machine Learning (ML) tools, although currently not used by our participants,
will be deployed in the near future. The success of such ML technologies relies on
their capacity to adapt the model to the monitored environment. For example, our
analysts stressed the importance of adding context to the alarms. Such contextual
knowledge could be incorporated in the ML models to produce more effective
alarms, reducing the number of benign triggers. In addition, incorporating the
analysts “feedback” on the generated alarms to the model itself can strengthen
its future predictions and its alarm prioritisation. In case of benign triggers, the
model can then learn that the prediction made, although true, is not suitable for
the organisation in hand, adjusting the model accordingly.
One of the main findings reported in Section 7.2, is that both vendors and the
research community should use false alarms or benign triggers as an alternative
metric to false positives, which is more general and vague. Specifically, when
evaluating the performance of a system deployed in a real-world setting, using the
term false positive gives the impression that the technology itself is fundamentally
flawed. Hence, when analysts’ report a 99% false positive, this may not be due to
the performance of the tool itself but due to the customer’s system and network
configurations generating benign triggers that analysts ignore.
7.9 Summary
In this chapter, we focused on understanding how SOC practitioners define false
positives, a term that is vague and general. We found that a clear distinction needs
to be made between false alarms and benign triggers when evaluating a SOC tool.
We also investigated the tools used by analysts, identifying their strengths
and weaknesses. We discovered that these tools produce alarms that lack the
context needed by analysts to filter false alarms from genuine ones. We identified
the knowledge sources that can support such contextual alarms and challenges
in deploying them.
Our interviews revealed research opportunities for improving SOC technologies,
which we summarised as lessons learned. One of the lessons derived are requirements
for network-based malware detection systems. Firstly, due to encryption, network-
based technologies need to be content-agnostic, relying on network header-level
features for detection. Moreover, such tools should be able to detect generic malware
activities and adapt to malware evolution. Furthermore, malware generated traffic
7. Malware Detection Tools in Security Operations Centres 127
Contents
8.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . 129
8.2 MalClassifier System Design . . . . . . . . . . . . . . . . 130
8.2.1 Design Goals and Requirements . . . . . . . . . . . . . . 132
8.2.2 Pre-processing . . . . . . . . . . . . . . . . . . . . . . . 133
8.2.3 Malware Family Profile Extraction . . . . . . . . . . . . 135
8.2.4 Building Malware Family Classification Models . . . . . 140
8.3 Evaluation . . . . . . . . . . . . . . . . . . . . . . . . . . . 140
8.3.1 Dataset . . . . . . . . . . . . . . . . . . . . . . . . . . . 141
8.3.2 Experiments . . . . . . . . . . . . . . . . . . . . . . . . 141
8.3.3 Classifier Design . . . . . . . . . . . . . . . . . . . . . . 142
8.4 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 144
8.4.1 Malware Family Profile Extraction . . . . . . . . . . . . 144
8.4.2 Classifier Performance . . . . . . . . . . . . . . . . . . . 145
8.4.3 Robustness to Evasion . . . . . . . . . . . . . . . . . . . 147
8.5 Discussion . . . . . . . . . . . . . . . . . . . . . . . . . . . 147
8.5.1 Meeting Design Goals . . . . . . . . . . . . . . . . . . . 148
8.5.2 Understanding Classifier Errors . . . . . . . . . . . . . . 148
8.5.3 Lessons Learned . . . . . . . . . . . . . . . . . . . . . . 149
8.5.4 Evasion and Limitations . . . . . . . . . . . . . . . . . . 150
8.6 Comparison to Related Work . . . . . . . . . . . . . . . 151
8.7 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . 152
128
8. Malware Family Classification Using Network Flow Sequence Behaviour 129
8.1 Introduction
We discussed in Chapter 6 the process analysts follow when investigating malware.
Although MSSP SOCs monitor the client’s systems and networks for signs of
malicious activity, they may not have direct access to the client’s hosts due to
privacy or policy reasons. When malware-infected hosts are the source of the
detected malicious traffic, the analysts need to negotiate access to the infected host
with the client. Even if access is possible, the time required to negotiate this and
running the executable in a sand-box for classification is not efficient, particularly
in situations where rapid detection and response is critical.
As identified in Chapter 6.4.3, the incident response procedure for ransomware
is different from a APT, and being able to determine the family can speed up the
remediation process. Therefore, analysts need on-the-wire malware classification
systems capable of matching detected malicious network traffic to a malware family,
thus not requiring to access the infected host. Moreover, as we found in Chapter 7,
with the increase in the use of encryption by malware for obfuscation and by hosts
for privacy (e.g. HTTPS), network-based monitoring has become challenging.
Considering the aforementioned requirements in mind, we propose MalClassifier,
a system for the automatic extraction of network behavioural characteristics for
malware family classification. As malware use some form of network communication
to propagate or contact their command and control (C&C) servers [17], MalClassifier
derives the network behavioural characteristics for a malware family, abstracting
the malware family’s network behaviour to a flow sequence profile. MalClassifier is
effective in classifying malicious network flow sequences to a malware family on-the-
wire, thus negating the need for access and sand-box execution of the malware binary.
MalClassifier uses non-identifiable network traffic features derived from network
connection logs for training the classifiers. Such features are visible even when
network connections are encrypted, making it resilient to adversary obfuscation
through encryption. This also makes the data required to train the classifiers
accessible and easier to share than full PCAP traces due to privacy concerns. In
designing MalClassifier, we make our approach IP-agnostic, as sophisticated malware,
such as exploit kits, apply dynamic DNS and Domain Generation Algorithms (DGA)
to change their communications destination [198].
Text mining approaches such as n-grams have been successfully applied to
malware family classification [100, 101]. Similar to [101], we apply n-grams to sort
network flows into groups of n consecutive flows (i.e. n-flows). Such a representation
provides granularity to the extracted malware behaviour. For example, a single
8. Malware Family Classification Using Network Flow Sequence Behaviour 130
failed SMTP flow might be benign, but multiple consecutive ones exhibit the
behaviour of a bot sending spam. MalClassifier mines n-flows that are distinctive
for each malware family. Such distinctive n-flows are used as features to train a
supervised model capable of classifying unseen n-flows to the malware family. The
models learn re-occurring network flow patterns that capture the characteristics
of a malware family’s network behaviour.
The contributions of this chapter are three-fold:
Family
131
3. Training and Building Models: Using the profiles, the supervised machine
learning model is trained for classifying unseen n-flows to a malware family.
system model should still be able to classify to the correct malware family. In
addition, the system should be robust against malware evasion through flow field
manipulation by using tamper-resistant features [199]. Malware may attempt to
change the order sequence of the flows to avoid detection. Therefore, the system
should consider flow order deception, and be able to still classify manipulated
sequences with high accuracy.
High classification accuracy— The classifier must aim to provide acceptable
classification accuracy using only sub-sequences of network traffic, thus not requiring
a malware’s full packet captures.
Considering the aforementioned design goals, we discuss the design of each
module of the MalClassifier system in detail in the following sections.
8.2.2 Pre-processing
In order to convert the network flows into a format that can be applied to sequence
mining methods, we first need to pre-process the data by reassembling and encoding
the network flows.
Table 8.1: Description of the fields in conn.log generated by Zeek network monitoring
framework and used in MalClassifier.
Field Description
resp_port Destination port
proto Transport layer protocol (TCP, UDP)
service Application protocol being sent over the connection.
orig_bytes Number of payload bytes the originator sent
resp_bytes Number of payload bytes the responder sent
conn_state State of the Connection.
13 different states (e.g. connection attempt rejected)
history State history of connections as a string of letters.
orig_pkts Number of packets that the originator sent
orig_ip_bytes Number of IP level bytes that the originator sent
resp_pkts Number of packets that the responder sent
resp_ip_bytes Number of IP level bytes that the responder sent
generates a number of logs for each malware sample. These logs include information
that is useful in understanding malware behaviour, such as C&C communication
statistics, DNS queries and fast fluxing, unusual communications (e.g. unknown
protocols) and port-host scanning.
In MalClassifier, we leverage the Zeek conn.log file that shows non-identifiable
network flow header information of TCP/UDP/ICMP connections. Each row in the
log represents an individual flow f and is described by 20 attributes representing
the column fields. We use 11 of the attributes derived from the conn.log and
described in Table 8.1.
Flow Encoding
For each log x ∈ X, where X is the set of all Zeek conn.log logs for samples of
a malware family y, we encode the log x to a long sequence of network flows,
E(x) = f1 → f2 → .. → fi , where fi represents a single flow in the log x. Formally,
we define a flow fi as:
fi :=< resp_port, proto, service, orig_bytes, resp_bytes, conn_state, history,
orig_pkts, orig_ip_bytes, resp_pkts, resp_ip_bytes >
As we show in Figure 8.1, the result of the Flow Encoding module is a
textual sequence representation of the flows in the conn.log of malware samples
of a malware y.
8. Malware Family Classification Using Network Flow Sequence Behaviour 135
Similarity measures are defined as “functions that quantify the extent to which
objects resemble each other, taking as an argument object pair and return numerical
values that are higher as the objects are alike” [200]. Choosing the similarity
measure relies on the data nature itself, whether binary, numerical or structured
data (e.g. sequences, trees, graphs).
Similarity measures for binary data are used to determine the presence or
absence of characteristics in the object pair (i.e. pair of flows), thus take a value
1 if the flow possesses the characteristic, and 0 otherwise. For example, the
authors in [100] applied a binary similarity approach, by counting the number of
times the exact n-gram occurs. We argue that network flows are structured data,
and thus binary similarity are not suitable as they do not capture the malware
underlying network flow semantic.
Instead, MalClassifier applies a fuzzy Value Similarity measure. Thus, instead
of determining is a sub-sequence for malware A and a sub-sequence for malware
8. Malware Family Classification Using Network Flow Sequence Behaviour 136
Assuming the cost of insertion, deletion, and modification is the same (= 1),
then the Levenshtein Distance of making ShADdF a into ShADadR is 3. The
trivial implementation has a runtime and space complexity of O(nm). The distance
value ranges from [0,100], where 0 indicates a low distance thus higher similarity
and 100 indicates a low similarity. We scale the distance value to be in the
range [0,1] instead of [0,100].
Using our example, the Levenshtein Distance of (f0 A ∈ Seq1 : SF |ShADdF a, f0 B ∈
Seq2 : SF |ShADadf F ) and (f1 A ∈ Seq1 : SF |ShADdF a, f1 B ∈ Seq2 : S0|ShAdDaf F )
is (2, 4) respectively, resulting in an average Levenshtein Distance of 3, scaled to 0.03.
Cosine Similarity— The numeric attributes of the two flows are represented
as two vectors. Cosine Similarity measures the similarity of two non-zero vectors
by calculating the cosine of the angle between them. Given the vectors x and y of
length n = 6 (number of numeric attributes), the cosine similarity is represented as:
n
xy i=1 xi yi
P
cos(x, y) = = qP qP (8.1)
kxkkyk i=1 (xi )
n 2
i=1 (yi )
n 2
The Cosine Similarity of x = [367, 3547, 6, 615, 7, 3835] ∈ f0 A , y = [1801, 15606, 14,
2369, 18, 16334] ∈ f0 B is 0.00025.
Similarly, the Cosine Similarity of x = [322, 464, 5, 530, 4, 632] ∈ f1 A , y =
[5274, 1370, 18, 5975, 28, 456] ∈ f1 B is 0.284. Thus, the average Cosine Similarity
for Seq1 and Seq2 is 0.142.
Inter-flow Distance— (resp_port): Inter-flow distance calculates the distance
between the resp_port in every two consecutive flows of a sub-sequence. This
helps identify malware network behavioural attributes such as performing a port
scan (e.g. when the difference of resp_port of two consecutive flows is 1 ). To
calculate the inter-flow similarity, we first calculate the distance of resp_port in
each consecutive flows in a sub-sequence. For example, the resp_port distance
between f0 A , f1 A ∈ Seq1 is 0, as the resp_port in both flows is the same. However,
the resp_port distance between f0 B , f1 B ∈ Seq1 is 55, which is the difference
between port 80 and port 25. The inter-flow similarity of Seq1 and Seq2 is the
distance of (0,55), thus is 55.
The Inter-flow distance is normalised to obtain a value in [0, 1] to define a
dissimilarity. The simplest and most common normalisation method uses a linear
transformation, known as feature scaling where η(d) = d−d_min
d_max−d_min
. However,
the drawback of applying this method is that it is sensitive to outliers [200].
Therefore, to normalise the distance value, we apply another normalisation method
that overcomes this drawback by defining parameter Z = (m, M ), where m and
8. Malware Family Classification Using Network Flow Sequence Behaviour 138
To extract the network flow sequence behaviour for a malware family y, we derive
the set S of n-flows of all the network flows in the log x that belong to the malware
family y. We calculate the similarity sim(i, s) of each n-flow i in a malware sample
mj to each n-flow s in the set S. To calculate the value sim(i, s), we apply the
value similarity measure defined in Section 8.2.3.
To reduce memory and processing complexity, we pre-compute the similarity
of each pair of flows in S and store as a key-value store. Consequently, when
determining the similarity of two n-flows, the similarity of each flow is pre-computed
and is fetched from the key-value store. Therefore, only unique flow pairs are stored
and only the similarity of new flow pairs are computed.
We are interested in mining the malware family’s network traffic for the highly
frequent n-flows, i.e. flows that occur in all or the majority of the samples of
8. Malware Family Classification Using Network Flow Sequence Behaviour 139
Profile Selection
In order to extract the malware family profiles, we select the most relevant n-flows
(profiles) that represent that malware family’s network behaviour. For each malware
family, we mine the network flows for all samples of that malware family, then
select n-flows that are significant. We apply a modified version of class-wise feature
selection approach as proposed in [202]. The class-wise document frequency is the
number of network flows in a malware family y that contain an n-flow s. We assign
each n-flow a class-dependent weight according to its coverage in the malware family
network traffic (tendency). As we apply a fuzzy similarity approach, we multiply
8. Malware Family Classification Using Network Flow Sequence Behaviour 140
8.3 Evaluation
To evaluate our approach, we implement the system as a multi-threaded Python
application. To accelerate the analysis, the application uses Python Multiprocessing
with separate threads to calculate the similarity scores for each malware sample. The
experiments were conducted on a 40-core processor with 126GB RAM, Centos OS.
8. Malware Family Classification Using Network Flow Sequence Behaviour 141
8.3.1 Dataset
Botnets are known to follow a certain infection life-cycle, sharing similarities in
network flow characteristics in each infection phase [82]. Thus, we evaluate the
effectiveness of our system in determining the unique network flow behaviour of
each botnet family in our dataset, and its accuracy in classifying botnet network
traffic to its family despite the botnets’ network behaviour similarities. To train
our classifier we used (1) the CTU-13 botnet traffic dataset [183] provided by
the Malware Capture Facility project and (2) current botnets and ransomware
(Miuref, WannaCry, Conficker, Sality, Notpetya) provided by Stratosphere IPS
Project2 , both described in Chapter 4.
We applied network flows of scenarios 1−13 to train and evaluate the classification
models. Scenario 7 was excluded as it did not contain a sufficient number of flows.
We applied only the malicious flows (C&C and botnet flows) to train the classifier
for malware family classification. We provide an overview of the malware families
in our datasets in Table 8.2.
8.3.2 Experiments
The aim of the experiments is to (1) determine how accurately we can classify
an n-flow to its malware family using the extracted profiles as features; and (2)
determine the classifiers robustness to malware evasion. In particular, we plan
to explore the following:
2
https://www.stratosphereips.org/
8. Malware Family Classification Using Network Flow Sequence Behaviour 142
Table 8.3: Example of the selected 2-flows for each malware family in our dataset.
transition probabilities of all the samples, PCA evaluates the variance of each
feature to linearly transform the feature space into a new one where the information
(evaluated as a percentage of variance) is contained in as few components as
possible. Thus, the 220 features are transformed to a new set of features, known
as the principal components. By reducing the number of features (dimensionality)
used in the classifier, the computational cost is reduced.
We set P CA = 10, thus reducing the features space to 10 principle components.
We evaluated different values for PCA, and the classifier performance when we
apply P CA = 10 performs almost as well as when we use all features. The
advantage of using PCA over using all features is the low memory complexity
required to run the machine learning algorithms due to reducing the dimensions
of the features space used by the classifier.
Classifier Performance Measures— To assess the performance of the clas-
sifiers, we apply 10-fold cross-validation. Cross-validation removes any bias in the
data while maximizing the number of score computations from a given dataset. As
8. Malware Family Classification Using Network Flow Sequence Behaviour 144
the CTU-13 datasets consist of only a few PCAPs (executions) per family, when
performing cross-validation we compare the different executions of the same family.
We employ evaluation measures such as Precision, Recall and F-measure to
evaluate the classifiers’ performance. A low precision can indicate a large number
of False Positives (F P ). A low recall indicates many False Negatives. Therefore,
we seek an analysis setup that maximizes the precision and recall. We also consider
F-measure (F 1) an aggregated performance score, that combines both precision
and recall. A perfect discovery of classes yields F 1 = 1, where either a low precision
or recall result in a low F-measure.
Although these metrics are defined for a binary classifier, they can be extended
for multi-class problems. Each class is represented by a binary classifier (e.g. One-
vs-rest approach), and we average the binary metric across the set of classes. We
also apply a macro-averaged Precision-Recall curve as an evaluation metric, which
gives equal weight to the classification of each label compared to micro-averaging
which gives equal weight to each per-family classification decision.
8.4 Results
We discuss in the following our results from the experiments outlined in Sec-
tion 8.3.2 and the impact of the various system configurations and approaches
on the classification performance.
100.0
97.5
95.0
Classifier 3erfRrPanFe (%)
92.5
90.0
87.5
85.0
Figure 8.2: Average F-measure for Random Forest and KNN classifiers, n = 1 − 7 for
n-flows.
email spam. Although, Neris 2-flows show HTTP flows with connection state set to
RST0. This means that the connection was established but the originator aborted.
However, we do see other successful HTTP connections with connection state SF,
meaning normal establishment and termination.
WannaCry is a ransomware known for sending SMB flows to other victim hosts
on the network. This is captured by the selected 2-flows of TCP flows with a
destination port of 445. Miuref’s 2−flows sequences were mostly HTTPS flows
(port = 443), followed by DSN requests or DSN requests followed by an ICMP flow.
Figure 8.3: Macro-averaged Precision Recall Curves for each malware family (Random
Forest classifier, n = 5).
8.5 Discussion
We discuss how MalClassifier meets our design goals and identify potential lim-
itations, suggesting approaches to address them.
8. Malware Family Classification Using Network Flow Sequence Behaviour 148
Classification Performance
100
80
60
Similarity metrics
40 Binary Simialrity
Levenshtein Distance
20 Cosine Similarity
Inter-flow Distance
0
Figure 8.4: The malware familys’ classification F-measure of the four Random Forest
Classifiers (n = 5), each using one of the four similarity measures.
Figure 8.5: Normalised confusion matrix showing actual classes vs. predicted classes
for the Random Forest Classifier (n = 5).
Table 8.4: Comparison of related work on malware family classification and MalClassifier
against our design goals.
Evasion Resiliency
Content-agnostic
Use sand-box
Use n-grams
IP-agnostic
Accuracy
Data
Rieck et al. [56] Behavioural reports N N N N Y 80.7%
Rieck et al. [57] Behavioural reports N N N N Y 95%
Perdisci et al. [103] HTTP traffic N Y N N N N/A
FIRMA [104] Network traffic N N N N N 98.8%
CHATTER [100] Network traffic N Y N Y Y 90%
MalClassifier Zeek conn logs Y Y Y Y N 96%
8.7 Summary
In this chapter, we presented a novel approach for analysing and classifying
network traffic of malware variants based on their network flow sequence behaviour.
Considering the limitations of existing approaches, we proposed a system that is
privacy-preserving, time efficient, and resilient to malware evasion. We showed
8. Malware Family Classification Using Network Flow Sequence Behaviour 153
Contents
9.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . 155
9.2 BOTection System Design . . . . . . . . . . . . . . . . . 155
9.2.1 Network Flow Reassembly . . . . . . . . . . . . . . . . . 157
9.2.2 Connection States Extraction . . . . . . . . . . . . . . . 157
9.2.3 Markov Chain Modeling . . . . . . . . . . . . . . . . . . 159
9.2.4 Classification . . . . . . . . . . . . . . . . . . . . . . . . 159
9.3 Evaluation . . . . . . . . . . . . . . . . . . . . . . . . . . . 161
9.3.1 Experiments . . . . . . . . . . . . . . . . . . . . . . . . 161
9.3.2 Datasets . . . . . . . . . . . . . . . . . . . . . . . . . . . 161
9.3.3 Classifier Design . . . . . . . . . . . . . . . . . . . . . . 163
9.4 Results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 164
9.4.1 Bot and Benign Communication Patterns . . . . . . . . 165
9.4.2 Performance of Binary Classifier . . . . . . . . . . . . . 167
9.4.3 Classifier Performance Over Time . . . . . . . . . . . . 168
9.4.4 Performance of Multi-Class Classifier . . . . . . . . . . . 169
9.5 Discussion . . . . . . . . . . . . . . . . . . . . . . . . . . . 171
9.5.1 Understanding Classifiers’ Errors . . . . . . . . . . . . . 171
9.5.2 Lessons Learned . . . . . . . . . . . . . . . . . . . . . . 173
9.5.3 Evasion and Limitations . . . . . . . . . . . . . . . . . . 175
9.6 Comparison to Related Work . . . . . . . . . . . . . . . 176
9.7 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . 177
154
9. Bot Detection by Building Markov Chain Models of Bots Network Behavior155
9.1 Introduction
In Chapter 8, we applied malware network behavioural features for malware family
classification. In our results, we found that most of the 2-flows selected for the
malware families are a sequence of two identical flows. For example, Notpetya traffic
contained a sequence of 5 flows of 445-tcp-0-0-0-S0-S-4-192-0-0. Moreover, we
found that the conn_state to be a lightweight and useful feature in understanding
the outcome of this connection. Mariconti et al. [207, 208] showed the effectiveness
of building Markov Chains from the sequence of Android API calls for malware
detection that can maintain its detection capabilities over time. In this chapter,
we explore if Markov Chains can be applied to a sequence of connection states
(conn_state) to provide insights on bot behaviour.
Malware tends to launch their attacks in bursts [209, 210], meaning they send
multiple network connections in a short amount of time. Hence, we first explore the
bursty nature of malware, to determine how frequently bots send network traffic. We
then investigate the discrepancies between malware network connections and those
produced by benign applications. Malware network traffic is then modelled using
Markov Chains to explore its strength in capturing bots’ network communication
behaviour during its C&C interactions, propagation stage, and attack operations.
To evaluate the effectiveness of applying Markov Chains for malware network
connections, we propose BOTection, a novel system that detects malware-infected
hosts (bots) on a network and classifies it to its malware family by monitoring
their network communication. We consider all types of malware-infected hosts that
communicate with a command and control server (e.g. ransomware).
In Chapter 7.8.7, the analysts reported that one of the weaknesses in existing
tools is that signatures are “loosely written”. Meaning, signatures are written to
capture a general malware behaviour (e.g. network scanning) but are incapable
of then identifying its malware family. This poses a challenge to analysts that
need to determine the cause of that activity to determine the best course of action.
Hence, we explore Markov Chains capability in modelling general malware behaviour
capable of detecting unseen malware. We then evaluate the system’s performance
in attributing that activity to a particular malware family.
real world. For example, organisations may deploy middle-boxes that intercept
their network communications, thus allowing the deployment of content-based
detection mechanisms. Alternatively, some organisations outsource their security
monitoring to MSSPs, due to lack of cybersecurity expertise or to avoid high-
security investments [24]. As we identified in Chapter 6, one of the factors that
impacts SOC operations, specifically outsourced SOCs, is security clearance. Some
organisations operate under restricted security permissions, requiring SOC analysts
to have certain level of clearance to view network traffic payload. For example,
maintaining the confidentiality of email communications is critical in such restricted
organisations. However, as we found also in Chapter 6, recruitment of SOC
analysts that have security clearance is a challenge. Hence, content inspection
technologies are discouraged and using non-privacy invasive features is crucial to
foster BOTection’s adoption in scale.
Considering the limitations of existing approaches and aforementioned moni-
toring restrictions, we identify five main design goals that the BOTection system
has to fulfill: 1 resilience to obfuscation/encryption by avoiding deep packet
inspection (Section 9.2.1), 2 privacy preservation (Section 9.2.2), 3 independence
of network topology and protocol (Section 9.2.2), 4 high detection and family
classification accuracy (Section 9.4.2, 9.4.1), and 5 capability of detecting unseen
bots by capturing general bot network behavior (Section 9.4.2).
BOTection operates in four phases as illustrated in Figure 9.1.
1. Network Flow Reassembly: The malicious and benign network traces are
converted to logs using Zeek. These logs are then split to sub-logs of n flows.
2. Connection State Extraction: The features are (e.g. conn_state) from the
logs and produce a key-value store of state transitions and their frequency.
3. Markov Chain Modelling: The state transition frequency are used to build
Markov Chain models and produce a feature vector for each sub-log.
We show the BOTection system in Figure 9.1, describing in this section its
modules and how design goals were considered.
9. Bot Detection by Building Markov Chain Models of Bots Network Behavior157
(OTH,OTH)
(Using Zeek) State Extraction
(REJ,REJ)
(S0,S0)
(S0,S1)
Class
...
REJ S0 .8 0 .4 .. 0
m1 B
m2 p p p .. p B
(REJ,REJ) 4
Zeek S0 S1 SF REJ S2 .. OTH m3 p p p .. p M
PCAP conn.log (REJ, S0) 6
... p p p .. p M
(S0, REJ) 2 S0 .8 0 0 .2 0 0 0
.. .. .. .. .. ..
F1 80 TCP HTTP S0 (S0,S0) 8 S1 0S0 0.8 00 00 0.2 00 00
F2 53 UDP DNS REJ mi p p p .. p M
... SF 0S1 00
S0 .8
00 00 00
0 0 .2
00 00
0 0
F3 53 UDP DNS REJ
(OTH,OTH) 0 REJ .6SF 00 00 .40 00 00 00
F4 53 UDP DNS REJ
…
n .. ..S1 ..0 0 0 0
..0 ..0 ...4
0 0
.. 0 ..0
REJ 0.6
SF 0 0 0 0 0
Fn 80 UDP HTTP S0 OTH ..
0 0.. 0.. 0.. 0.4.. 0.. 0.. ..
REJ .6 0 0 0 0
Extract conn_state and ..OTH ..0 ..0 ..0 ..0 ..0 ..0 ..0 Markov Chain State
Flows in conn.log split to windows calculate the state OTH Transitions Dataset
0 0 0 0 0 0 0
of length n transition frequencies
State Transition Matrix
0.65 0.24
udp|dns|SF
0.33
0.58 0.2 tcp|smtp|S1 tcp|smtp|SF
OTH S0 0.22
SF REJ 0.98
0.59 tcp|ssl|SF 0.48
0.44
Figure 9.2: Markov Chain for Neris ICMP port scanning (left) and Zeus C&C (right)
9.2.4 Classification
Using feature vectors of Markov Chain state transitions belonging to bot and benign
samples, we train supervised machine learning models. Particularly, we use an
ensemble classifier known to overcome over-fitting issues. Since our system carries
9. Bot Detection by Building Markov Chain Models of Bots Network Behavior160
Table 9.1: Description of the flow connection state (conn_state) feature obtained from
Zeek conn.log logs.
State Description
S0 Connection attempt seen, no reply.
S1 Connection established, not terminated.
SF Normal establishment and termination.
REJ Connection attempt rejected.
S2 Connection established and close attempt by originator
seen (but no reply from responder).
S3 Connection established and close attempt by responder
seen (but no reply from originator).
RSTO Connection established, originator aborted (sent a RST).
RSTR Responder sent a RST.
RSTOS0 Originator sent a SYN followed by a RST,
but did not receive a SYN-ACK from the responder.
RSTRH Responder sent a SYN ACK followed by a RST,
we never saw a SYN from the originator.
SH Originator sent a SYN followed by a FIN,
we never saw a SYN ACK from the responder.
SHR Responder sent a SYN ACK followed by a FIN,
we never saw a SYN from the originator.
OTH No SYN seen, just midstream traffic.
out two subsequent tasks (i.e. bot detection and bot family classification), we adopt
two different classification approaches discussed in the following.
Binary Classifier (Bot Detection Model) — Binary classification aims at
partitioning the samples in the feature space in two distinct and complementary
sub-spaces. The partition is done in such a way that a sub-space includes samples
of a class while the complementary sub-space includes the samples of the other
class. In our system, the classifier is trained using n-flow samples of known malware
families and benign traffic. The trained model is then used to classify an n-flow
into either the bot class (i.e. Malicious) or the benign class (i.e. not Malicious).
Multi-Class Classifier (Bot Family Classification Model) — Since our
family classification task aims at discriminating between malicious flows of different
bot families, we rely on a classification approach that partitions the feature space
in multiple non-overlapping sub-spaces. The multi-class classifier in our system
is trained using malicious n-flows belonging to multiple families. Thus, once an
n-flow is classified as malicious by the Bot Detection Model, it is then attributed
to a malware family by the Bot Family Classification Model.
9. Bot Detection by Building Markov Chain Models of Bots Network Behavior161
9.3 Evaluation
We implement the system as a Python 2.7 application using the Scikit-learn library. 1
The experiments were conducted on a 40-core processor with 128GB RAM and 20G
disk space, Centos OS. We open-sourced the implementation of our system as
a contribution. 2
9.3.1 Experiments
We design the following experiments to determine BOTection’s capability in
modelling bot network communication, that can be used for bot detection and
family classification.
Bot and Benign Communication Patterns (results in Section 9.4.1) — We
model the bots’ and benign communications as a Markov Chain to identify bots’
network behaviour and explore the discrepancies in bot and benign state transitions.
Bot Detection (results in Section 9.4.2 and Section 9.4.3) — We build and
evaluate two bot detection classifiers to: 1 detect known bots and 2 detect
previously unseen bots. The objective of the unseen bot classifier is to assess the
classifiers’ generalisability in detecting communication for bot families that were
not considered in classifier training. We also attempt to answer the question ‘What
if a few wrong flows end up within the window n?’ Hence, we study the effect of
injecting benign flows into the malicious n-flows on the system detection.
Bot Family Classification (results in Section 9.4.4) — We identify network
communications that are unique to each bot family. Thus, once an n-flow is identified
as malicious by the bot detection classifier, we evaluate the multi-class classifier’s
accuracy in classifying these malicious n-flows to a bot family.
9.3.2 Datasets
We use two datasets to evaluate the binary classifier (bot detection) and multi-
class classifier (bot family classification). In Table 9.2, we show the number of
malicious flows (both datasets) and benign flows (ISCX only) and the percentage
of flows belonging to each family. In Table 9.3, we also show the bot families
attack network communications in the dataset used to train the multi-class classifier.
These datasets were described in more detail in Chapter 4.
ISCX Botnet Dataset — for Binary Classifier— This dataset contains
malicious and benign traffic and is split into training and testing sets with 5 and 13
1
Scikit-learn: machine learning in Python - https://scikit-learn.org/
2
https://github.com/balahmadi-Ox/botection
9. Bot Detection by Building Markov Chain Models of Bots Network Behavior162
Table 9.2: Datasets used for evaluation - ISCX contains bot and benign flows used in
binary classifier, MCFP contains bot traffic only for multi-class classifier.
Dataset #Flows Bot Families
Training: Neris (12%), Rbot (22%), Virut (0.94%),
Zeus (0.01%), Zeus C&C (0.01%)
Training
- 6,925,812 flows Testing:
ISCX - 34.97% Malicious Known Bots — Neris (5.67%), Rbot (0.018%),
Botnet Virut (12.80%), Zeus (0.109%), Zeus C&C (0.006%)
Dataset Testing
- 5,040,068 flows Unseen Bots — Menti (0.62%), Sogou (0.019%),
- 44.97% Malicious Murlo (0.16%), Tbot (0.283%), Zero Access (0.221%),
Weasel (9.25%), Smoke Bot (0.017%),
ISCX IRC Bot(0.387%)
Miuref (2.8%), Zeus (3.5%), Geodo (3.4%) NSIS (0.3%),
Bunitu (2.7%), WannaCry (0.9%),Conficker (0.8%),
Zeus C&C (8.6%), Neris (2.8%), Rbot (4.5%),
MCFP 7,283,060 Bot flows
Virut (1.23%), Sality (20.7%), TrickBot (0.5%),
Stlrat (5.3%), Donbot (0.07%), Papras (41.8%),
Qvod (0.1%)
bots respectively, where eight bot families in the testing set were not used to train the
classifier (unseen bots). We use this dataset to train and evaluate the binary classifier
using malicious and benign network traces, in detecting unseen bots. This is due to
the testing set containing bot families that were not present in the training set.
Stratosphere IPS Project — for Binary Classifier— This dataset is used
to verify the binary classifier’s performance over time, thus determining the date
the samples were first in the wild was essential. The samples date of "first seen
in the wild" was verified using VirusTotal3 reports.
Malware Capture Facility Project (MCFP) and Stratosphere IPS
Project — for Multi-class Classifier— To train our multi-class bot family
classifier, we used the CTU-13 botnet traffic dataset provided by the Malware
Capture Facility Project [183]. In addition, we use the samples of newer bots
(e.g. Trickbot - 2016 ) and ransomware (e.g. WannaCry - 2017 ) collected by The
Stratosphere IPS Project. The multi-class classifier classifies detected malicious
traffic to a family; therefore, only malicious traffic (C&C communication and bot
communication) is used to train and evaluate the classifier. The botnet families
represented in the dataset employed various protocols (e.g. IRC, P2P, HTTP) and
various techniques (e.g. spam, DDoS), as shown in Table 9.3.
3
www.virustotal.com
9. Bot Detection by Building Markov Chain Models of Bots Network Behavior163
Table 9.3: Bot Network communication used to train the multi-class classifier(PS: Port
Scanning, NS: Network Scanning, FF: Fast-Fluxing, CF: ClickFraud).
Propagation C&C Operational
Family
PS NS DGA FF CC DDoS Spam CF
Miuref
Zeus
Geodo
Bunitu
WannaCry
Conficker
Zeus C&C
Neris
Rbot
Virut
Sality
Stlrat
Donbot
Papras
Qvod
Figure 9.4: Average connection state transitions for network communications of bot
(red-left) and benign (blue-right) samples in ISCX dataset.
each second. This confirms our system design hypothesis that bots send traffic in
bursts. To explore the ideal window size, we experiment with various configurations
for the window n, where n = 10, 15, 20, 25, 30, 35.
Classifier Performance— To assess the performance of the known bots binary
classifier and multi-class classifier, we apply stratified 10-fold cross-validation.
For the unseen bot classifier, we use the training set to train the classifier and
evaluate using the testing set containing samples of bot families not considered in
training. We employ the evaluation measures of Precision, Recall and F-measure
to evaluate classifiers’ performance. For the binary classifier, we also measure
the False Positive Rate (F P R = F P/(F P + T N )) and the False Negative Rate
(F N R = F N/(F N + T P )).
9.4 Results
We present in the following sections the results for the experiments identified
in Section 9.3.
9. Bot Detection by Building Markov Chain Models of Bots Network Behavior165
BlackHole
ZeroAccess
Cumulative Number of flows (log)
103 Menti
Murlo
Neris
102 Osx_trojan
RBot
SmokeBot
Sogou
101 TBot
Virut
Weasel
100 Zeus
20 40 60 80 100 120 140 160 180 200 220 240 260 280 300 320 340 360 380 400 420 440 460 480 500 520 540 560 580 600 620 640 660 680 700 720 740 760 780
Time (seconds)
Figure 9.5: Number of flows generated by each malware family every 20 seconds.
Bot Traffic Over Time— We show the overall amount of traffic by each bot
family in Table 9.2 over time in Figure 9.5. In particular, the y-axis has a logarithmic
scale, and each data-point consists of the traffic in an interval of 20 seconds. We
can notice that the majority of the bot families constantly produce a significant
amount of network flows. This means that a bot is more likely to be detected
in different phases of its malicious operations since our method has available an
abundant quantity of n-flows for the classification.
To understand the various communications a bot performs during propagation,
C&C communication, and other attack operations, we modelled the communication
as a Markov Chain. We discuss the most prominent bot network communication
found in the MCFP dataset. To simplify the Markov Chain for visualisation, we
show only the state transitions with probability p > 0.1.
Table 9.4: False Positive Rate (FPR), False Negative Rate (FNR) of the known and
unseen classifiers for each value n.
conn_state PSC
Classifier n
FPR FNR FPR FNR
10 0.004 0.111 0.353 0.256
15 0.060 0.096 0.386 0.055
20 0.100 0.074 0.407 0.056
Unseen Bots
25 0.093 0.073 0.175 0.043
30 0.079 0.065 0.142 0.030
35 0.097 0.067 0.142 0.026
10 0.003 0.000 0.002 0.003
15 0.003 0.001 0.001 0.002
20 0.004 0.001 0.001 0.008
Known Bots
25 0.008 0.001 0.003 0.001
30 0.004 0.001 0.000 0.000
35 0.001 0.000 0.002 0.009
message [213]. This can be seen in the Markov Chain with the HTTP connection with
state SF followed by an SSL connection with p = 0.44. Following the SSL state, the
bot establishes a connection with the C&C (p = 0.98) that is terminated by the bot.
SMTP Spam— Multiple bots in our dataset send email SPAM, specifically
Neris, Sality and Geodo. For brevity, we show the average of all the Markov chains
for the Geodo bot performing email SPAM in Figure 9.12. We found 29, 575 Simple
Mail Transfer Protocol (SMTP) flows in our dataset for Geodo bot. Overall, there
are 2 types of SMTP flows: (1) f = conn_state:S1, history:ShAdDa, (2) f =
conn_state:SF, history:ShAdDafF. Meaning, the first was not terminated, whilst
the second was terminated by a F IN bit sent by both ends. These two types were
captured in the Markov Chain model, showing the next state after a completed
SMTP connection is a DNS connection (P = 1), whilst an unterminated connection
will (1) either remain in the same state (P = 0.41), (2) receive a SYN Ack followed
by a FIN from destination (tcp | − | SHR)(P = 0.27), (3) make a DNS request
(P = 0.32). Other bot families had similar state transitions, although Neris had
some SMTP connection with the conn_state set to RSTR and history ShAdDaTfrr,
indicating the destination sent a RST.
Port Scanning— We were also able to identify ICMP port scanning (Figure
9.2), where an ICMP connection flow with conn_state OT H is followed by a UDP
flow with states S0, or a TCP connection with conn_state SF or REJ. Bot
families Rbot, Neris, Virut, Qvod, and Donbot all perform port scanning and have
shown to have similar state transitions.
9. Bot Detection by Building Markov Chain Models of Bots Network Behavior167
1.0
0.8
0.7
FonnBstate, .nown %ots
0.6
FonnBstate, 8nseen %ots
0.5 pUoto-seUviFe-FonnBstate, .nown %ots
pUoto-seUviFe-FonnBstate, 8nseen %ots
0.4
10 15 20 25 30 35
1uPbeU of )lows
Figure 9.6: Binary classifier F-measure for detecting known/unseen bots for each state
type.
DDoS— We are also able to observe two types of DDoS attack, TCP and
ICMP DDoS in Rbot. DDoS was observed as multiple connection requests sent
to a single host, where these requests could be an ICMP request or TCP request.
For ICMP DDoS, the bot establishes an ICMP request which has a conn_state
OT H, whilst TCP DDoS connections have conn_state REJ.
1.0
0.8
0.7
0.6
0.5
% detected
0.4
1 2 3 4 5 6 7 8 9 10 11 12 13 14
No. Flows Injected
Figure 9.7: Percentage of bot 15-flows detected when we inject b benign flows in the
sequence.
(F-measure= 93.03%) compared to PSC where detection efficiency was high for
unseen bots only when the burst window size was above 20.
Classifier Performance When Injecting Benign Flows— We explore the
effect of injecting random benign flows b into the bot 15-flow (n = 15) at random
locations on the detection performance. In practice, this means that we inject states
that are randomly selected among the most frequent conn_states in benign traffic,
as reported in Section 9.4.1 and shown in Figure 9.4. For example, REJ conn_state
was not injected in this experiment as it rarely occurs in benign traffic.
In this experiment, we injected benign flows (i.e. b random benign conn_states),
in the malicious 15-flows. For example, for 15-flow of malicious flows, we can inject
2 benign flows (b = 2), meaning that the other 13 flows are malicious (m = 13),
where (n = b + m). We conducted experiments with various configurations of b and
m. For each configuration, we repeated the experiment 40 times and obtained the
average of the number of detected malicious flows by the known bot classifier.
We show the results of this experiment in Figure 9.7. The x-axis represents the
number of benign flows injected b. For example, when less than four benign flow
was injected (b = 4, m = 11), the classifier was capable of detecting 80% of the
15-flows even after injecting a random state into the sequence. This may be due
to the diversity of the states in a bot n-flow, as discussed in 9.4.1.
Table 9.5: Binary Classifier Performance: F1 (F-measure), Precision (P) and Recall (R)
over time with Temporal Training Consistency. We train the classifier using bot samples
first seen in the years precedent of the samples in the testing set.
Testing set
Training
2015 2016 2017 2018
set
F1 P R F1 P R F1 P R F1 P R
Pre 2015 .86 .89 .87 .87 .89 .87 .86 .89 .87 .86 .89 .87
Pre 2016 - - - .987 .987 .987 .983 .983 .983 .982 .982 .982
Pre 2017 - - - - - - .981 .982 .981 .981 .981 .981
Pre 2018 - - - - - - - - - .989 .989 .989
In training and testing the classifier, we take into consideration the Temporal
Training Consistency, where all samples in the training are temporally precedent to
samples in the testing set [214]. Thus we split the test set, depending on the year
the malware was “first seen”, which was verified from VirusTotal reports.
We show the results of training the binary classifier (n = 15) in Table 9.5.
Interestingly, the classifier’s performance over time is 86% when trained with
samples appearing before 2015. However, the classifier performance increases to
98% when trained with samples appearing after 2015. These results show that
the binary classifier generalises well enough not to be affected by the temporal
bias problem defined in [214]. This means that our classifier (i) can detect traffic
from previously unseen bot families, and (ii) is robust over time; thus, it does
not need to be frequently re-trained.
0.8 0.8
0.7 0.7
0.6 0.6
F-Measure F-Measure
0.5 Precision 0.5 Precision
Recall Recall
0.4 0.4
10 15 20 25 30 35 10 15 20 25 30 35
No. Flows No. Flows
Figure 9.8: Multi-Class classifiers’ performance for each Markov Chain state feature
type.
Figure 9.9: Markov Chain Model Transition Probabilities of the conn_state feature for
each botnet family. For brevity, we show only transition probabilities greater than 0.6.
For example, Virut had a high transition probability between states S1 and REJ.
Classification Performance
1.0
0.8
0.6
9.5 Discussion
We discuss in the following sections the reasons behind classifiers’ misclassifications,
lessons learnt from our experiments, potential methods malware can deploy to
evade detection and system limitations.
1.0
0.8
0.6
FPR
0.4
0.2
0.0
10.0.0.254
172.16.2.11
172.16.2.112
172.16.2.113
203.69.42.35
203.73.24.75
111.89.134.93
192.168.1.101
192.168.2.106
192.168.2.107
192.168.2.108
192.168.2.111
192.168.3.114
192.168.3.115
192.168.3.116
192.168.3.117
192.168.4.119
192.168.4.121
195.228.245.1
202.43.195.13
208.111.160.6
66.225.226.11
66.225.226.27
68.71.209.206
72.11.132.253
74.86.158.244
91.190.170.71
97.74.104.201
192.150.18.200
198.105.193.76
206.220.42.181
209.85.135.104
209.85.135.147
64.214.232.121
202.235.196.117
208.111.161.254
Figure 9.11: Host IPs that produced the highest FPR for the binary classifier (FPR >
0.1%).
Table 9.6: Number of flows, 15 window length samples, and number of bots per family
in the ISCX testing dataset, and the number of False Negatives (FN) for each bot family
and benign traffic.
Family Total # Flows # 15-flow window # Hosts # FN
IRC Attack 773370 51558 8 107
Menti 13024 1628 1 8
Murlo 81135 5409 1 68
Neris 402000 2680 1 28
Osx_trojan 375 25 1 0
Rbot 116505 7767 1 10
SmokeBot 1755 117 1 0
Sogou 20625 1375 1 1
TBot 5205 347 4 5
Virut 437550 29170 1 7726
Weasel 508080 33872 2 165
ZeroAccess 19440 1296 2 9
Zeus 15660 1044 4 9
Benign 2,948,550 196570 1164 16100
the testing set contains unsuccessful IRC connections (S0) to the server followed
by a RST RH response from the server (meaning the connection was timed out).
Hence, this is not a result of Virut changing its behaviour but was a result of the
server delay in its reply and connection timing out.
Multi-Class Classifier— We attempt to investigate the misclassification
(when n = 20) per bot family. Geodo had the lowest F-Measure score of 84%,
with some 20-flows classified to Bunitu or Sltrat. Most flows misclassified to
Bunitu are connections with only state transition SF → SF . This explains the
misclassification, as the classifier learned that 20 consecutive SF flows is a behavior
of Bunitu, resulting in the low classification accuracy for Geodo. However, as
SF → SF state transition represents a normal network connection, n-flows with
only connection state SF should be white-listed. Similarly, Geodo flows with only
S0 → S0 transitions are classified as Sltrat, which has the same flows.
Figure 9.12: Markov Chain for Sality (right) and Geodo (left) spam behavior (only
state transition probabilities, where p > 0.1).
Although bots share similarities in their Markov Chains, there are still differences
that make it possible to distinguish between the network traffic of different bot
families. BOTection was able to classify a n-flow to its bot family with 99% accuracy
— fulfilling design goal 4 . Each bot family had a set of uniquely identifiable Markov
Chain state transitions (Figure 9.9). In fact, bots may operate the same attack in
different ways. For example, Figure 9.12 shows the behavior of two spam families,
Geodo and Sality. They both successfully use the TCP SMTP protocol (tcp|smtp|SF )
9. Bot Detection by Building Markov Chain Models of Bots Network Behavior174
and perform DNS queries through UDP (udp|dns|SF ), but they use them in different
ways. When Geodo successfully sends an email, it always goes back to the DNS
queries (P = 1). However, Sality is often sending many emails in a row as state
tcp|smtp|SF has 0.64 probability of returning to itself in the next step of the chain.
Another example is the Markov Chains for Papras that shows TCP DNS requests,
while normally DNS is sent over UDP. This may be due to the DNS response or
request data size exceeding the size of a single packet. Papras flows in our dataset
contain DGA communication; this unique DNS request behaviour is peculiar of
Papras DGA behaviour that is captured by our system.
We explored two types of states used in the Markov Chain models. Both
conn_state and PSC are effective in detecting bot communications, but PSC
provides a higher level of granularity. For example, a flow with conn_state SF
might be sent using either UDP or TCP, and might be an SSL communication,
a DNS request, or an HTTP request, each representing a different flow behavior.
However, PSC has 404 state transitions compared to conn_state (132 = 169)
increasing the dimensionality thus the computation needed for classifier training.
We highlight the importance of evaluating the system performance in detecting
bots not used when training the classifier. Our system was able to detect known
bots with an F-measure of 99% and new bots with an F-measure of 93% — fulfilling
design goals 4 and 5 .
Bot Markov Chain models are different from benign ones as bots have more
diverse state transitions, as seen in Figure 9.4. Hence, BOTection was able to capture
bot behaviour and recognise it in flows of families not included in the training
set. Markov Chain models are robust to bot evolution due to the abstraction of
network communication to finite states, thus less susceptible to the introduction
of new states. Applying conn_state in the Markov Chain models ensures that
no additional states could be introduced.
We evaluated our binary classifier’s performance over time. Following recom-
mendations provided in [214] to overcome the temporal bias problem, we trained the
classifier with samples first seen in the years preceding the test samples, complying
with the temporal training consistency. We also ensured that samples in the
test sets had been seen for the first time in the same year. As we reported
in Table 9.5, our method maintained its performance over time with a 98% F-
measure. This ensures that a classifier can detect new bots without frequent
and time-consuming re-training.
For binary classification, we used the ISCX dataset that contains network flows
labelled (malicious/benign) according to the host generating them. We studied the
9. Bot Detection by Building Markov Chain Models of Bots Network Behavior175
Table 9.7: Comparison of previous Bot Detection Approaches and proposed system
BOTection
Type of Encryption Single Bot Unseen
Approach
Detector Resiliency Detection Bots
BotSniffer [78] C&C N N N
BotFinder [84] C&C Y Y N
DISCLOSURE [79] C&C Y Y N
BotHunter [82] Operation N N N
BotMiner [81] Operation Y N Y
Abaid et. al [85] C&C N N Y
Abaid et. al [216] Operation N Y N
C&C
BOTection Y Y Y
Operation
using TCP, UDP is also used in rare benign applications so can be easily detected
by a rule-based system. Bots are known to have similar attack objectives, thus
attempting to change its network behaviour will be detected as its new behaviour
may be similar to another family’s attack behaviour.
In evaluating our system, we considered its performance in detecting traffic
generated by bot families that were not included in the training set. Our system
was able to detect such traffic with a 93% F-measure due to these bot families
exhibiting similar attacks (e.g. DDoS) to families that are present in the training
set. Hence, unseen bots that use novel techniques (i.e. new types of attacks that use
protocols in a unique way) may not be detectable in our system (e.g. zero-days).
As with other machine learning-based approaches, a malware might try to evade
the proposed systems via adversarial learning techniques [218]. It might also try to
mimic the patterns of multiple families to impact the classification. However, the
main objective of malware family classification is to identify the one that exhibits the
most similar network behaviour to assist the defender in deploying proper mitigation.
In the case of a new malware family, one solution is to classify such malicious traffic
to a “New Family” class using the confidence of multi-class Random Forest classifiers
applied previously in [219]. Thus, classifications with a confidence level below a
threshold will not be attributed to a known family but as “New Family”.
9.7 Summary
In this chapter, we presented BOTection, a novel topology and protocol-independent
system capable of detecting and classifying flows to a bot family with a 99% accuracy.
Importantly, BOTection models capture similar network behaviour, thus are capable
of detecting unseen bot network behaviour with 93% accuracy without re-training
the classifier. We evaluated our approach using the network communication of
various bots performing propagation, C&C communication and attacks. Our results
demonstrate the effectiveness of using Markov Chains to understand bot network
behaviour for better detection.
Our research on analysing malware behaviour revealed that malware utilise
similar techniques to launch their attacks. We capture this phenomenon when
9. Bot Detection by Building Markov Chain Models of Bots Network Behavior178
analysing the behavioural features extracted from their network traffic. Hence,
modelling malware common behaviour allows detection systems to identify new
malware variants that use similar attack techniques. As detecting the malware
is essential, attributing that activity to a malware family is also critical in SOCs
to determine the best remediation. It is crucial when designing network-based
monitoring tools to consider real-world deployment, where content-inspection is no
longer possible due to privacy and malware obfuscation through encryption. Our
proposed systems recognise these design goals, improving over existing state-of-the-
art malware detection tools. To foster further research and real-world deployment
evaluation, we open-source our system.4
In the following chapter, we will conclude by providing an overview of this thesis
contributions and identifying opportunities for future research directions.
4
https://github.com/balahmadi-Ox/botection
Conclusions and Future Work
10
Contents
10.1 Research Contributions . . . . . . . . . . . . . . . . . . . 179
10.2 Validation of Research Results . . . . . . . . . . . . . . 181
10.3 Future Work . . . . . . . . . . . . . . . . . . . . . . . . . 182
10.4 Final Remarks . . . . . . . . . . . . . . . . . . . . . . . . 184
179
10. Conclusions and Future Work 180
One of the main findings of our SOC study is the importance of designing systems
that are human-machine collaborative, where technology works together with SOC
practitioners for the detection of threats. The SOC operations are highly human
reliant, where technology is there to present the information for the practitioners
for them to make the intelligent decision. Moreover, most technologies in SOCs rely
on humans to configure and tune false positives or benign triggers according to the
networks and systems ecosystem being monitored. Hence, when designing systems
to be used in SOC environments it is important to not only validate its performance
standalone but evaluate its performance when integrated into a SOCs ecosystem of
various technologies, people and processes. In this work, we proposed a malware
detection system based on requirements identified in our qualitative research. In
future work, we need to validate the applicability of the proposed system in SOCs
and its integration to existing technological methods and operations. Such research
can be in the form of a user study, where SOCs integrate the proposed system in
their SOC technological pipeline and analysts evaluate its effectiveness in assisting
them to recognise and detect possible malware attacks. Similarly, our SOC research
has highlighted possible directions of designing technological solutions for malware
detection in SOCs, that should also be validated in a SOC setting. We discuss
future work relating to some of these technological solutions in the following section.
alarm. This supports the fast processing of alarms, allowing analysts to focus
their efforts on more serious threats.
To facilitate malware human-machine monitoring and detection in SOCs, mal-
ware network threat discovery can be modeled as a graph problem. Network traffic
logs, in particular, have unique properties, that make them different to other
logs. For example, the sequence of network traffic, as we found in Chapter 8 and
Chapter 9, provides insights on the malware behaviour. Therefore, representing
the sequence in the graph is useful to infer information about a malware’s network
communication behaviour. Although such modeling of malware network behaviour
proved effective for detection, still, real-world deployment requires incorporating
contextual knowledge. As we found in Chapter 7, machine learning models need
to incorporate contextual features in model training and include analysts feedback
on the predictions to reduce the number of benign triggers. Hence, to facilitate a
more accurate inference, contextual information about the malware (e.g. malware
family name, stage of attack, CVE) and context about the infected host (e.g. OS)
and subnet (e.g. business sector) can be incorporated in the graph to be used
by analysts to contextualise an alarm.
SDN-based SIEM
In Chapter 7, our participants noted they are not able to monitor the internal
network traffic due to storage and computational constraints. Instead, to understand
network activity, security analysts turn to a semi-manual investigation of network
data (e.g. firewall logs, packet capture data, network flows, system logs). As we
identified in Chapter 6 a host’s network connections are only looked at in the later
stages of an investigation, missing out on meaningful earlier network connections
that may have been stronger Indicators of Compromise (IOC).
As future work, we propose the idea of reducing the logs collected by the SIEM,
hence decreasing the hay in the “haystack”. Security analysts in the SOC define the
SIEM alarms (e.g. use-cases) that would initiate a further investigation. Once that
alarm is flagged, they then manually look at the logs to determine if the alarm is a
legitimate threat or a false alarm. Since the raw data is only viewed by analysts
when a SIEM alarm is generated, why not initiate network collection for a host
or subnet when an alarm is triggered for that host/subnet. Hence, we propose
collecting network traffic at the early stages of the investigation with the level of
granularity needed to make a decision of its maliciousness.
Software-Defined Networking (SDN) is a new network paradigm that allows
greater control over network entities by decoupling the control plane and the
10. Conclusions and Future Work 184
186
References 187
[32] Paul Cichonski et al. “Computer security incident handling guide”. In: NIST
Special Publication 800.61 (2012), pp. 1–147.
[33] Jeff Bollinger, Brandon Enright, and Matthew Valites. Crafting the InfoSec
Playbook: Security Monitoring and Incident Response Master Plan. " O’Reilly
Media, Inc.", 2015.
[34] Sheharbano Khattak et al. “A taxonomy of botnet behavior, detection, and
defense”. In: IEEE COMST 16.2 (2014), pp. 898–924.
[35] Sandeep Yadav et al. “Detecting Algorithmically Generated Malicious Domain
Names”. In: ACM IMC. 2010.
[36] Gernot Vormayr, Tanja Zseby, and Joachim Fabini. “Botnet Communication
Patterns”. In: IEEE COMST 19.4 (2017), pp. 2768–2796.
[37] GD Kurundkar, NA Naik, and SD Khamitkar. “Network intrusion detection using
Snort”. In: International Journal of Engineering Research and Applications 2.2
(2012), pp. 1288–1296.
[38] Suman Rani and Vikram Singh. “SNORT: an open source network security tool
for intrusion detection in campus network environment”. In: International Journal
of Computer Technology and Electronics Engineering 2.1 (2012), pp. 137–142.
[39] Farzaneh Izak Shiri, Bharanidharan Shanmugam, and Norbik Bashah Idris. “A
parallel technique for improving the performance of signature-based network
intrusion detection system”. In: 2011 IEEE 3rd International Conference on
Communication Software and Networks. IEEE. 2011, pp. 692–696.
[40] Yong Tang and Shigang Chen. “Defending against internet worms: A
signature-based approach”. In: Proceedings IEEE 24th Annual Joint Conference
of the IEEE Computer and Communications Societies. Vol. 2. IEEE. 2005,
pp. 1384–1394.
[41] Nattawat Khamphakdee, Nunnapus Benjamas, and Saiyan Saiyod. “Improving
intrusion detection system based on snort rules for network probe attack
detection”. In: 2014 2nd International Conference on Information and
Communication Technology (ICoICT). IEEE. 2014, pp. 69–74.
[42] Eugene Albin and Neil C Rowe. “A realistic experimental comparison of the
Suricata and Snort intrusion-detection systems”. In: 2012 26th International
Conference on Advanced Information Networking and Applications Workshops.
IEEE. 2012, pp. 122–127.
[43] Prasanta Gogoi, Bhogeswar Borah, and Dhruba K Bhattacharyya. “Anomaly
detection analysis of intrusion data using supervised & unsupervised approach”.
In: Journal of Convergence Information Technology 5.1 (2010), pp. 95–110.
[44] Dhruba Kumar Bhattacharyya and Jugal Kumar Kalita. Network anomaly
detection: A machine learning perspective. Chapman and Hall/CRC, 2013.
[45] Levent Koc, Thomas A Mazzuchi, and Shahram Sarkani. “A network intrusion
detection system based on a Hidden Naıve Bayes multiclass classifier”. In: Expert
Systems with Applications 39.18 (2012), pp. 13492–13500.
[46] Yan Qiao et al. “Anomaly intrusion detection method based on HMM”. In:
Electronics letters 38.13 (2002), pp. 663–664.
References 189
[63] Martin Roesch et al. “Snort: Lightweight Intrusion Detection for Networks.” In:
LISA. Vol. 99. 1. 1999, pp. 229–238.
[64] Koral Ilgun, Richard Kemmerer, Phillip Porras, et al. “State transition analysis: A
rule-based intrusion detection approach”. In: Software Engineering, IEEE
Transactions on 21.3 (1995), pp. 181–199.
[65] Jake Ryan, Meng-Jang Lin, and Risto Miikkulainen. “Intrusion detection with
neural networks”. In: Advances in neural information processing systems (1998),
pp. 943–949.
[66] James Cannady. “Artificial neural networks for misuse detection”. In: National
information systems security conference. 1998, pp. 368–81.
[67] Majid Ghonji Feshki, Omid Sojoodi, and Minoo Deljavan Anvary. “Managing
Intrusion Detection Alerts Using Support Vector Machines”. In: International
Journal of Computer Science and Security (IJCSS) 9.5 (2015), p. 266.
[68] Shi-Jinn Horng et al. “A novel intrusion detection system based on hierarchical
clustering and support vector machines”. In: Expert systems with Applications
38.1 (2011), pp. 306–313.
[69] Mohammad Sazzadul Hoque et al. “An implementation of intrusion detection
system using genetic algorithm”. In: arXiv preprint arXiv:1204.1336 (2012).
[70] Arman Tajbakhsh, Mohammad Rahmati, and Abdolreza Mirzaei. “Intrusion
detection using fuzzy association rules”. In: Applied Soft Computing 9.2 (2009),
pp. 462–469.
[71] Wenke Lee and Salvatore J Stolfo. “Data mining approaches for intrusion
detection”. In: Usenix Security. 1998.
[72] M Zubair Shafiq, S Momina Tabish, and Muddassar Farooq. “Are evolutionary
rule learning algorithms appropriate for malware detection?” In: Proceedings of
the 11th Annual conference on Genetic and evolutionary computation. ACM. 2009,
pp. 1915–1916.
[73] Ulrich Bayer et al. “A view on current malware behaviors”. In: USENIX workshop
on large-scale exploits and emergent threats (LEET). 2009.
[74] Carsten Willems, Thorsten Holz, and Felix Freiling. “Toward automated dynamic
malware analysis using cwsandbox”. In: IEEE Security & Privacy 2 (2007),
pp. 32–39.
[75] L Xue and G Sun. “Design and implementation of a malware detection system
based on network behavior”. In: Security and Communication Networks 8.3
(2015), pp. 459–470.
[76] Christian Rossow et al. “Sandnet: Network traffic analysis of malicious software”.
In: Proceedings of the First Workshop on Building Analysis Datasets and
Gathering Experience Returns for Security. ACM. 2011, pp. 78–88.
[77] Konrad Rieck et al. “Botzilla: Detecting the phoning home of malicious software”.
In: proceedings of the 2010 ACM Symposium on Applied Computing. ACM. 2010,
pp. 1978–1984.
[78] Guofei Gu, Junjie Zhang, and Wenke Lee. “BotSniffer: Detecting botnet
command and control channels in network traffic”. In: (2008).
References 191
[79] Leyla Bilge et al. “Disclosure: detecting botnet command and control servers
through large-scale netflow analysis”. In: Proc. of ACM ACSAC. 2012.
[80] Christian Rossow and Christian J Dietrich. “Provex: Detecting botnets with
encrypted command and control channels”. In: DIMVA. 2013, pp. 21–40.
[81] Guofei Gu et al. “BotMiner: Clustering Analysis of Network Traffic for
Protocol-and Structure-Independent Botnet Detection.” In: USENIX Security
Symposium. Vol. 5. 2. 2008, pp. 139–154.
[82] Guofei Gu et al. “BotHunter: Detecting Malware Infection Through IDS-Driven
Dialog Correlation.” In: Usenix Security. Vol. 7. 2007, pp. 1–16.
[83] Ayesha Binte Ashfaq et al. “Diagnosing bot infections using Bayesian inference”.
In: Journal of Computer Virology and Hacking Techniques (2016).
[84] Florian Tegeler et al. “BotFinder: Finding Bots in Network Traffic Without Deep
Packet Inspection”. In: Proceedings of the 8th International Conference on
Emerging Networking Experiments and Technologies. CoNEXT ’12. Nice, France:
ACM, 2012, pp. 349–360. url:
http://doi.acm.org/10.1145/2413176.2413217.
[85] Zainab Abaid, Mohamed Ali Kaafar, and Sanjay Jha. “Early Detection of
In-the-Wild Botnet Attacks by Exploiting Network Communication Uniformity:
An Empirical Study”. In: IFIP Networking (2017).
[86] Zhaoyan Xu et al. “PeerPress: utilizing enemies’ P2P strength against them”. In:
Proceedings of the 2012 ACM conference on Computer and communications
security. ACM. 2012, pp. 581–592.
[87] Antonio Nappa et al. “Cyberprobe: Towards internet-scale active detection of
malicious servers”. In: 2014.
[88] Zhaoyan Xu et al. “AUTOPROBE: Towards Automatic Active Malicious Server
Probing Using Dynamic Binary Analysis”. In: Proceedings of the 2014 ACM
SIGSAC Conference on Computer and Communications Security. CCS ’14.
Scottsdale, Arizona, USA: ACM, 2014, pp. 179–190. url:
http://doi.acm.org/10.1145/2660267.2660352.
[89] Guofei Gu et al. “Active botnet probing to identify obscure command and control
channels”. In: Computer Security Applications Conference, 2009. ACSAC’09.
Annual. IEEE. 2009, pp. 241–253.
[90] Leyla Bilge et al. “EXPOSURE: Finding Malicious Domains Using Passive DNS
Analysis.” In: NDSS. 2011.
[91] Manos Antonakakis et al. “Building a Dynamic Reputation System for DNS.” In:
USENIX security symposium. 2010, pp. 273–290.
[92] Lei Zhang et al. “A survey on latest botnet attack and defense”. In: Trust,
Security and Privacy in Computing and Communications (TrustCom), 2011 IEEE
10th International Conference on. IEEE. 2011, pp. 53–60.
[93] Xin Hu, Matthew Knysz, and Kang G Shin. “RB-Seeker: Auto-detection of
Redirection Botnets.” In: NDSS. 2009.
[94] Roberto Perdisci et al. “Detecting malicious flux service networks through passive
analysis of recursive dns traces”. In: Computer Security Applications Conference,
2009. ACSAC’09. Annual. IEEE. 2009, pp. 311–320.
References 192
[95] Etienne Stalmans and Barry Irwin. “A framework for DNS based detection and
mitigation of malware infections on a network”. In: Information Security South
Africa (ISSA), 2011. IEEE. 2011, pp. 1–8.
[96] Thorsten Holz et al. “Measuring and Detecting Fast-Flux Service Networks.” In:
NDSS. 2008.
[97] Emanuele Passerini et al. “Fluxor: Detecting and monitoring fast-flux service
networks”. In: Detection of intrusions and malware, and vulnerability assessment.
Springer, 2008, pp. 186–206.
[98] Jose Nazario and Thorsten Holz. “As the net churns: Fast-flux botnet
observations”. In: Malicious and Unwanted Software, 2008. MALWARE 2008. 3rd
International Conference on. IEEE. 2008, pp. 24–31.
[99] Pratyusa K Manadhata et al. “Detecting malicious domains via graph inference”.
In: Computer Security-ESORICS 2014. Springer, 2014, pp. 1–18.
[100] Aziz Mohaisen et al. “Chatter: Classifying malware families using system event
ordering”. In: Proc. of IEEE CNS. 2014.
[101] Hesham Mekky, Aziz Mohaisen, and Zhi-Li Zhang. “Separation of benign and
malicious network events for accurate malware family classification”. In: IEEE
CNS. 2015.
[102] Christian Wressnegger et al. “A close look on n-grams in intrusion detection:
anomaly detection vs. classification”. In: ACM AISec. 2013.
[103] Roberto Perdisci, Wenke Lee, and Nick Feamster. “Behavioral Clustering of
HTTP-Based Malware and Signature Generation Using Malicious Network
Traces.” In: Proc. of USENIX NSDI. 2010.
[104] M Zubair Rafique and Juan Caballero. “Firma: Malware clustering and network
signature generation with mixed network behaviors”. In: Proc. of RAID. 2013.
[105] Igor Santos et al. “N-grams-based File Signatures for Malware Detection.” In:
2009.
[106] Emmanouil Vasilomanolakis et al. “Taxonomy and Survey of Collaborative
Intrusion Detection”. In: ACM Computing Surveys (CSUR) 47.4 (2015), p. 55.
[107] Tim Bass. “Intrusion detection systems and multisensor data fusion”. In:
Communications of the ACM 43.4 (2000), pp. 99–105.
[108] Fang Lan, Wang Chunlei, and Ma Guoqing. “A framework for network security
situation awareness based on knowledge discovery”. In: Computer Engineering
and Technology (ICCET), 2010 2nd international conference on. Vol. 1. IEEE.
2010, pp. V1–226.
[109] Richard Zuech, Taghi M Khoshgoftaar, and Randall Wald. “Intrusion detection
and Big Heterogeneous Data: a Survey”. In: Journal of Big Data 2.1 (2015),
pp. 1–41.
[110] BA Fessi et al. “Data collection for information security system”. In: Engineering
Systems Management and Its Applications (ICESMA), 2010 second international
conference on. IEEE. 2010, pp. 1–8.
References 193
[111] Abdoul Karim Ganame et al. “A global security architecture for intrusion
detection on computer networks”. In: Computers and Security 27.1-2 (2008),
pp. 30–47.
[112] David L Hall and James Llinas. “An introduction to multisensor data fusion”. In:
Proceedings of the IEEE 85.1 (1997), pp. 6–23.
[113] Ting-Fang Yen et al. “Beehive: Largescale log analysis for detecting suspicious
activity in enterprise networks”. In: Proceedings of the 29th Annual Computer
Security Applications Conference. ACM. 2013, pp. 199–208.
[114] E. Bocchi et al. “Network Connectivity Graph for Malicious Traffic Dissection”.
In: Computer Communication and Networks (ICCCN), 2015 24th International
Conference on. Aug. 2015, pp. 1–9.
[115] Alina Oprea et al. “Detection of early-stage enterprise infection by mining
large-scale log data”. In: arXiv preprint arXiv:1411.5005 (2014).
[116] Ting-Fang Yen and Michael K Reiter. “Traffic aggregation for malware detection”.
In: Detection of Intrusions and Malware, and Vulnerability Assessment. Springer,
2008, pp. 207–227.
[117] C. Zhong et al. “Automate Cybersecurity Data Triage by Leveraging Human
Analysts’ Cognitive Process”. In: 2016 IEEE 2nd International Conference on Big
Data Security on Cloud (BigDataSecurity), IEEE International Conference on
High Performance and Smart Computing (HPSC), and IEEE International
Conference on Intelligent Data and Security (IDS). Apr. 2016, pp. 357–363.
[118] Chen Zhong et al. “A cyber security data triage operation retrieval system”. In:
Computers & Security 76 (2018), pp. 12–31.
[119] Elias Raftopoulos, Matthias Egli, and Xenofontas Dimitropoulos. “Shedding light
on log correlation in network forensics analysis”. In: International Conference on
Detection of Intrusions and Malware, and Vulnerability Assessment. Springer.
2012, pp. 232–241.
[120] Antonio Pecchia et al. “Filtering security alerts for the analysis of a production
saas cloud”. In: Utility and Cloud Computing (UCC), 2014 IEEE/ACM 7th
International Conference on. IEEE. 2014, pp. 233–241.
[121] Xiaokui Shu et al. “Threat Intelligence Computing”. In: Proceedings of the 2018
ACM SIGSAC Conference on Computer and Communications Security. CCS ’18.
Toronto, Canada: ACM, 2018, pp. 1883–1898. url:
http://doi.acm.org/10.1145/3243734.3243829.
[122] Rodrigo Werlinger et al. “Preparation, detection, and analysis: the diagnostic
work of IT security incident response”. In: Information Management & Computer
Security 18.1 (2010), pp. 26–42.
[123] Rodrigo Werlinger, Kirstie Hawkey, and Konstantin Beznosov. “An integrated
view of human, organizational, and technological challenges of IT security
management”. In: Information Management & Computer Security 17.1 (2009),
pp. 4–19.
[124] Rodrigo Werlinger et al. “Security practitioners in context: Their activities and
interactions with other stakeholders within organizations”. In: International
Journal of Human-Computer Studies 67.7 (2009), pp. 584–606.
References 194
[125] Rodrigo Werlinger et al. “The challenges of using an intrusion detection system: is
it worth the effort?” In: Proceedings of the 4th symposium on Usable privacy and
security. ACM. 2008, pp. 107–118.
[126] David Botta et al. “Toward understanding distributed cognition in IT security
management: the role of cues and norms”. In: Cognition, Technology & Work 13.2
(2011), pp. 121–134.
[127] Kirstie Hawkey et al. “Human, organizational, and technological factors of IT
security”. In: CHI’08 extended abstracts on Human factors in computing systems.
ACM. 2008, pp. 3639–3644.
[128] Sathya Chandran Sundaramurthy et al. “Turning contradictions into innovations
or: How we learned to stop whining and improve security operations”. In: Proc.
12th Symp. Usable Privacy and Security. 2016.
[129] Andrew M’manga et al. “Folk risk analysis: Factors influencing security analysts’
interpretation of risk”. In: Proc. of the 13th Symposium on Usable Privacy and
Security, ser. SOUPS. Vol. 17. 2017.
[130] Anita D’Amico et al. “Achieving Cyber Defense Situational Awareness: A
Cognitive Task Analysis of Information Assurance Analysts”. In: Proceedings of
the Human Factors and Ergonomics Society Annual Meeting 49.3 (2005),
pp. 229–233.
[131] Anita D’Amico and Kirsten Whitley. “The real work of computer network defense
analysts”. In: VizSEC 2007. Springer, 2008, pp. 19–37.
[132] Dorothy E Denning. “An intrusion-detection model”. In: Software Engineering,
IEEE Transactions on 2 (1987), pp. 222–232.
[133] Bazara IA Barry and H Anthony Chan. “Intrusion detection systems”. In:
Handbook of Information and Communication Security. Springer, 2010,
pp. 193–205.
[134] Zhi-cai SHI and Yong-xiang XIA. “Survey on intrusion detection techniques for
high-speed networks”. In: Application Research of Computers 5 (2010), p. 004.
[135] Chenfeng Vincent Zhou, Christopher Leckie, and Shanika Karunasekera. “A
survey of coordinated attacks and collaborative intrusion detection”. In:
Computers & Security 29.1 (2010), pp. 124–140.
[136] Ahmed Patel, Qais Qassim, and Christopher Wills. “A survey of intrusion
detection and prevention systems”. In: Information Management & Computer
Security 18.4 (2010), pp. 277–290.
[137] Niva Das and Tanmoy Sarkar. “Survey on Host and Network Based Intrusion
Detection System”. In: Int. J. Advanced Networking and Applications 6.2 (2014),
pp. 2266–2269.
[138] Pedro Garcia-Teodoro et al. “Anomaly-based network intrusion detection:
Techniques, systems and challenges”. In: computers & security 28.1 (2009),
pp. 18–28.
[139] Animesh Patcha and Jung-Min Park. “An overview of anomaly detection
techniques: Existing solutions and latest technological trends”. In: Computer
Networks: The International Journal of Computer and Telecommunications
Networking 51.12 (2007), pp. 3448–3470.
References 195
[188] ISO Guide. “73: 2009”. In: Risk management—Vocabulary 551 (2009), p. 49.
[189] Anton Chuvakin and Augusto Barros. How to Develop and Maintain Security
Monitoring Use Cases. url:
https://www.gartner.com/en/documents/3844970.
[190] Calvin Nobles. “The Cyber Talent Gap and Cybersecurity Professionalizing”. In:
International Journal of Hyperconnectivity and the Internet of Things 2.1 (2018),
pp. 42–51.
[191] Rob Van Os. “SOC-CMM: Designing and Evaluating a Tool for Measurement of
Capability Maturity in Security Operations Centers”. MA thesis. Luleå University
of Technology, Computer Science, 2016, p. 74.
[192] Christopher Kruegel and William Robertson. “Alert verification determining the
success of intrusion attempts”. In: DIMVA 2004, July 6-7, Dortmund, Germany
(2004).
[193] Fredrik Valeur et al. “Comprehensive approach to intrusion detection alert
correlation”. In: IEEE Transactions on dependable and secure computing 1.3
(2004), pp. 146–169.
[194] André Årnes et al. “Using hidden markov models to evaluate the risks of
intrusions”. In: International Workshop on Recent Advances in Intrusion
Detection. Springer. 2006, pp. 145–164.
[195] Sandeep Bhatt, Pratyusa K Manadhata, and Loai Zomlot. “The operational role
of security information and event management systems”. In: IEEE security &
Privacy 12.5 (2014), pp. 35–41.
[196] Chadni Islam, Muhammad Ali Babar, and Surya Nepal. “A Multi-Vocal Review of
Security Orchestration”. In: ACM Computing Surveys (CSUR) 52.2 (2019), p. 37.
[197] Xiaokui Shu et al. “Threat intelligence computing”. In: Proceedings of the 2018
ACM SIGSAC Conference on Computer and Communications Security. ACM.
2018, pp. 1883–1898.
[198] Fraser Howard. A closer look at the Angler exploit kit. July 2015.
[199] Z Berkay Celik et al. “Malware traffic detection using tamper resistant features”.
In: IEEE MILCOM. 2015.
[200] Marie-Jeanne Lesot, Maria Rifqi, and Hamid Benhadda. “Similarity Measures for
Binary and Numerical Data: a Survey”. In: Int. J. Knowl. Eng. Soft Data
Paradigm. 1.1 (Dec. 2009), pp. 63–84.
[201] Vladimir I Levenshtein. “Binary codes capable of correcting deletions, insertions
and reversals”. In: Soviet physics doklady. Vol. 10. 1966, p. 707.
[202] D Krishna Sandeep Reddy and Arun K Pujari. “N-gram analysis for computer
virus detection”. In: Journal in Computer Virology 2.3 (2006), pp. 231–239.
[203] Surendra K Singhi and Huan Liu. “Feature subset selection bias for classification
learning”. In: Proceedings of the 23rd international conference on Machine
learning. ACM. 2006, pp. 849–856.
[204] Svante Wold, Kim Esbensen, and Paul Geladi. “Principal component analysis”. In:
Chemometrics and intelligent laboratory systems 2.1-3 (1987), pp. 37–52.
References 199
[205] Roberto Perdisci et al. “Misleading worm signature generators using deliberate
noise injection”. In: 2006 IEEE Symposium on Security and Privacy (S&P’06).
IEEE. 2006, 15–pp.
[206] Elizabeth Stinson and John C Mitchell. “Towards Systematic Evaluation of the
Evadability of Bot/Botnet Detection Methods.” In: Proc. of USENIX WOOT
(2008).
[207] Enrico Mariconti et al. “MaMaDroid: Detecting Android Malware by Building
Markov Chains of Behavioral Models”. In: Proc. of NDSS. 2017.
[208] Lucky Onwuzurike et al. “MaMaDroid: Detecting android malware by building
markov chains of behavioral models (extended version)”. In: ACM TOPS 22.2
(2019), p. 14.
[209] Elias Raftopoulos and Xenofontas Dimitropoulos. “Detecting, validating and
characterizing computer infections in the wild”. In: ACM IMC. 2011.
[210] Paul Dokas et al. “Data mining for network intrusion detection”. In: Proc. of NSF
NGDM. 2002.
[211] Michal Piskozub, Riccardo Spolaor, and Ivan Martinovic. “MalAlert: Detecting
Malware in Large-Scale Network Traffic Using Statistical Features”. In: ACM
SIGMETRICS Performance Evaluation Review 46.3 (2019), pp. 151–154.
[212] Z Berkay Celik et al. “Malware traffic detection using tamper resistant features”.
In: MILCOM 2015-2015 IEEE Military Communications Conference. IEEE. 2015,
pp. 330–335.
[213] Hamad Binsalleeh et al. “On the analysis of the Zeus botnet crimeware toolkit”.
In: Proc. of ACM PST. 2010.
[214] Feargus Pendlebury et al. “TESSERACT: Eliminating Experimental Bias in
Malware Classification across Space and Time”. In: Proc. of USENIX Security.
2019.
[215] Moheeb Abu Rajab et al. “A Multifaceted Approach to Understanding the Botnet
Phenomenon”. In: Proc. of ACM IMC. 2006.
[216] Zainab Abaid et al. “The Early Bird Gets the Botnet: A Markov Chain Based
Early Warning System for Botnet Attacks”. In: Proc. of IEEE LCN. 2016.
[217] Sebastian Zander, Grenville Armitage, and Philip Branch. “A survey of covert
channels and countermeasures in computer network protocols”. In: IEEE COMST
9.3 (2007), pp. 44–57.
[218] Daniel Lowd and Christopher Meek. “Adversarial learning”. In: Proc. of ACM
SIGKDD. 2005.
[219] Vincent F Taylor et al. “Robust smartphone app identification via encrypted
network traffic analysis”. In: IEEE TIFS 13.1 (2018), pp. 63–78.
[220] David Zhao et al. “Botnet detection based on traffic behavior analysis and flow
intervals”. In: Computers & Security 39 (2013), pp. 2–16.
Appendices
200
Semi-Structured Interview Questions
A
A.1 Interview Part 1
The goal for this interview is to capture the analysts views on the current network
security tools used in the SOC.
3. Looking back on past events, have there been times during your use of
network-monitoring systems when they performed particularly well?
5. Looking back on past events, have there been times during your use of network-
monitoring systems when they could have performed better? Can you describe
an incident that was not detected as well as it should have been?
6. What is your view on the strengths and weaknesses of the network-monitoring
systems you use?
201
A. Semi-Structured Interview Questions 202
A.2.1 Scenarios
1. You are monitoring the organisation’s network traffic, and observe an increase
in the traffic from a server to an outside entity.
2. You are monitoring the organisation’s network traffic, and observe an increase
in the traffic from a server to another device on the network.
A. Semi-Structured Interview Questions 203
3. You are monitoring the organisation’s network traffic, and observe an increase
in the network outbound traffic (how does the analyst find the malicious
internal host?).
4. You receive an alert from the SIEM of unauthorised privilege escalation
attempt.
5. Scanning the network, you find an IOC such as a malware MD5 hash.
6. Your web server is receiving SYN flood requests.
B
Online Questionnaire - Assertions Results
204
Table B.1: Online Survey Results: Responses to Assertions (Resp, Ordered from "Strongly Disagree"(=1) - "Strongly Agree"(=5)) Mode,
Median, and Comparison of non-neutral scores - Disagree (1-2):Agree (4-5)(CNNS: D:A)
207
C. Template Analysis: Template 208
210
Security Operations Centre (SOC)
Monitoring Practice
Page 1: Introduction
We appreciate your interest in participating in our study on Security Operations Centre
Monitoring Tools. This survey is intended for security analysts that work in a Security Operations
Centre (SOC). The survey should take approximately 15-20 minutes to complete. Please read
through these terms before agreeing to participate by continuing to the next page.
2. To identify the thought processes of analysts in security monitoring, and data features to
be included in security tool development.
3. To compare working practice and security tool use across the SOCs of different
organisation types.
We will gather information for the above three parts through questions on SOC working
practice, security tool use, attack indicators, and security monitoring and detection techniques
applied by security analysts in SOCs in organisations.
The researchers are Louise Axon and Bushra Alahmadi, who are doctoral students in the
Centre for Doctoral Training in Cybersecurity. The research is being conducted under the
supervision of Professor Sadie Creese, Professor Michael Goldsmith and Professor Ivan
Martinovic.
You have been invited to take part because of your experience working as a security analyst in
a SOC.
1 / 37
2. Do I have to take part?
You can ask questions about the study before deciding whether to participate. You can choose
whether you participate and, if you agree to participate, you may withdraw yourself and your
data from the study without penalty at any time, and without giving a reason, by advising the
researchers of this decision.
There are no known risks or disadvantages of taking part and no specific preparatory
requirements, as we strive to protect your confidentiality, unless you explicitly agree that the
type or nature of your company can be mentioned in publications arising from the research.
The survey responses are recorded anonymously, and all participants in this study will remain
anonymous. The data will be stored in a password-protected file on the researcher’s computer,
and will be destroyed at the end of the project. Only the researchers will have access to the
personal data collected in the survey, and this data will be kept separately from the interview
responses, if you later take part in an interview.
The results of this study may be published in conference proceedings, peer-reviewed journals
and online in University archives. The results of the study will also contribute towards the DPhil
theses of the two researchers, which will be published online. No personal data from this study
will be published in any of the above; all study participants will remain anonymous and the data
collected will be stored in a password-protected file on the researchers’ computers, accessible
only by the researchers.
The University of Oxford is committed to the dissemination of its research for the benefit of
society and the economy and, in support of this commitment, has established an online archive
of research materials. This archive includes digital copies of student theses successfully
submitted as part of a University of Oxford postgraduate degree programme. Holding the
archive online gives easy access for researchers to the full text of freely available theses,
thereby increasing the likely impact and use of that research.
If you agree to participate in this project, the research will be written up as a thesis. On
successful submission of the thesis, it will be deposited both in print and online in the
University archives, to facilitate its use in future research. The thesis will be published with
open access, meaning it will be available to every internet user.
2 / 37
This project has been reviewed by, and received ethics clearance through, the University of
Oxford Central University Research Ethics Committee.
This research is funded by the Engineering and Physical Sciences Research Council
(EPSRC). The results of the study may be published in academic conferences or journals, and
will be published in the researchers' theses.
If you have a concern about any aspect of this project, please speak to the researchers, Louise
Axon and Bushra Alahmadi (louise.axon@cs.ox.ac.uk, bushra.alahmadi@cs.ox.ac.uk) or their
supervisors, Professor Sadie Creese (sadie.creese@cs.ox.ac.uk), Professor Michael
Goldsmith (michael.goldsmith@cs.ox.ac.uk), and Professor Ivan Martinovic
(ivan.martinovic@cs.ox.ac.uk) who will do their best to answer your query. The researcher
should acknowledge your concern within 10 working days and give you an indication of how
he/she intends to deal with it. If you remain unhappy or wish to make a formal complaint, please
contact the chair of the Research Ethics Committee at the University of Oxford (using the
contact details below) who will seek to resolve the matter in a reasonably expeditious manner:
Chair, Social Sciences & Humanities Inter-Divisional Research Ethics Committee; Email:
ethics@socsci.ox.ac.uk; Address: Research Services, University of Oxford, Wellington Square,
Oxford OX1 2JD
Please note that you may only participate in this survey if you are 18 years of age. If you
agree to participate and have read the terms above, please continue to the next page to
get started.
3 / 37
Page 2: Preliminary Questions
0 - 3
3 - 5
5 - 7
7 - 10
10 - 15
15 +
I prefer not to say
Very low
Low
Medium
High
Very high
Government
Financial Services
Healthcare Providers
4 / 37
Higher Education
K-12
Technology (Non-security focus)
Technology (security focus)
Manufacturing
Media and Entertainment
Travel, Hospitality, and Transportation
Retail
Other
Small enterprise
Medium enterprise
Large enterprise
I don't know
Other
How many network monitoring devices (e.g. a single IDS, firewall) are you personally responsible for
monitoring in the SOC you work in? Required
1
2 - 4
5 - 6
7 - 8
9 - 10
11 - 12
13- 14
15 - 16
17 - 18
19 - 20
21+
I don't know
How many network elements does a single monitoring device observe? Optional
6 / 37
In which country is the SOC you work in located? Required
Which tasks are you required to carry out in your role? Required
Monitor IDS
Monitoring network and systems logs
Track and trace intruders
Perform forensic evidence collection
Produce technical documents
Perform artifact analysis
Incident handling
Perform security policy development
Publish advisories or alerts
Perform virus handling
Provide and answer a hotline
Provide training and security awareness
Pursue legal investigations
Vulnerability handling
Security product development
Vulnerability scanning
Vulnerability assessments
Security configuration administration
Penetration testing
Other
7 / 37
If you selected Other, please specify:
How many network security alerts does your organisation receive on average daily? Required
Less Than 5K
5K–10K
10K–50K
50K–100K
100K–150K
Over 150K
I don't know
Other
How many network security alerts does your organisation investigate on average daily?
Optional
Which of the following does your SOC mostly focus on? You may select more than one if
necessary. Required
Threat prevention
Threat detection
Threat remediation
Recovery
Other
9 / 37
Page 3: Network Monitoring
What are the main network security threats your organisation faces? Required
10 / 37
Viruses, Worms, Trojans
Abnormal network activity
Ransomware
Advanced Persistent Threats (APTs)
Next-generation malware (e.g. Bots)
Phishing emails (including spear-phishing)
Brute-force attacks
Insider threats
Web defacement
Policy violation (e.g. gaming, streaming)
Other
Which of the following network security incidents do you believe you could do better at
addressing if you had the tools? Required
11 / 37
Insider threats
Web defacement
Policy violation (e.g. gaming, streaming)
Other
Intrusion Detection System (IDS) - signature-based (e.g Snort, Cisco Secure IDS )
Intrusion Detection System (IDS) - anomaly based (e.g. Spade)
Intrusion Detection System (IDS) - using machine learning
Data collection/ log aggregation (e.g. Splunk)
Security visualisations
Network monitoring (e.g. Argus, Bro)
Text-based data presentation (e.g. Wireshark/tcpdump, Nmap)
Information and Event Management (SIEM)
Policy management/profiling/posture assessment tool (e.g. Cisco ISE)
DNS Enforcer/Intelligent Proxy (e.g. Cisco Umbrella)
Other
If you selected Intrusion Detection System (IDS) - using machine learning or security
12 / 37
visualisations in the previous question, please specify the tools used.
Which network security data sources do you monitor in your work? For each source you
monitor, please indicate whether you monitor the source using a Security Incident and Event
Management (SIEM) tool or individually. Required
Monitor
Monitor Do not I don't
using a
individually monitor know
SIEM
Firewall logs
Network packet captures
Netflow
Host logs
Server logs
IDS logs
IPS logs
Web proxy logs
Website logs
Antivirus logs
Anomaly detection alerts
Domain Name System (DNS) logs
Authentication logs
Are there any other network security data sources you monitor in your work that were not
addressed in the question above? If so, please specify.
13 / 37
How important are the following features in choosing an ideal network monitoring system?
Required
14 / 37
Sophisticated alert-
management
capabilities
Do you have any further comments on any of the topics covered on this page?
15 / 37
Page 4: Network Monitoring Features
How important are the following data sources in detecting malicious activity on the network?
Required
16 / 37
Antivirus
software that
logs the results
of malware
scans on end
hosts.
Network firewall
Data server logs
Anomaly
detection logs
DNS logs
Security
vulnerability
mailing lists/blog
announcements
Do you rely on other data sources we have missed? If yes, please specify them below.
How important are the following network traffic features in detecting malicious activity on the
network? Required
17 / 37
Byte size
Packet content
Destination
domain URL
HTTP status
code
Web referrer
Domain
reputation
Domain category
User-agent
string
Sent and
received bytes
DNS response
Do you rely on other network features we have missed? If yes, please specify them below.
How important are the following Indicators of Compromise in detecting malicious activity on the
network? Required
18 / 37
Anomalies in
privileged user
account activity
Geographical
irregularities
Log-in red flags
Swells In
database read
volume
HTML response
sizes
Large numbers
of requests for
the same file
Mismatched
port-application
traffic
Suspicious
registry or
system file
changes
DNS request
anomalies
Unexpected
patching of
systems
Mobile device
profile changes
Bundles of data
in the wrong
places
Web traffic with
inhuman
behavior
Signs of DDoS
activity
19 / 37
Users' complaint
of slow system
Do you rely on other Indicators of Compromise we have missed? If yes, please specify them
below.
How much do you rely on your experience in analysing and aggregating data sources to detect
malicious activity as apposed to relying on security monitoring system alerts? Required
In what ways (if any) does your experience aid in network security monitoring tasks?
Do you have any further comments on any of the topics covered on this page?
21 / 37
Page 5: Assertions: Accuracy and Usability of Network
Monitoring Tools
Network Monitoring Tools: Please indicate your level of agreement with the following
assertions. Required
Strongly Strongly
Disagree Neutral Agree
disagree agree
The monitoring tools I
use frequently
produce false
positive results (they
detect a security
event when there
was not actually a
security event)
The monitoring tools I
use frequently
produce false
negative results (they
fail to detect a
security event that
occurs)
I sometimes rely on
my experience and
intuition to detect
attacks rather than
monitoring system
alerts
I rely on my custom
scripts to aggregate
the data I need to
analyse an incident
To detect malicious
activity on the
network I deploy
custom created
scripts
22 / 37
Intrusion Detection Systems: Please indicate your level of agreement with the following
assertions. Required
Strongly Strongly
Disagree Neutral Agree
disagree agree
I believe that current
IDS are inadequate
in detecting attacks
Current IDS solutions
are incapable of
coping with the high
data rates
Current IDS solutions
lack in effective and
speedy threat
detection and
response
Current IDS solutions
are built on the
assumption that
threats are observed
as they enter the
network in specific
perimeter points at
the Internet edge
The number of alerts
generated by most
IDS are
overwhelming
To reduce the
number of false
positives, I limit the
IDS signature set to
focus on important
attacks
I pick IDS alerts of
interest based on
problems we have
been having lately
23 / 37
Data Presentation Tools: Please indicate your level of agreement with the following
assertions. Required
Strongly Strongly
Disagree Neutral Agree
disagree agree
Data presentation
tools, such as
visualisations, are
important in my work
Data presentation
tools such as
visualisations can
help me to detect
incidents that are
missed by the
automated systems,
or that do not fit the
automated attack
detection profile
Visualisations are
useful for finding
interesting patterns in
raw network packet
data
Visual distractions
sometimes mean I
miss important
information that is
conveyed visually
Do you have any further comments on any of the topics covered on this page?
24 / 37
Page 6: Assertions: Working Practice in SOCs
The "Human in the Loop": Please indicate your level of agreement with the following
assertions. Required
Strongly Strongly
Disagree Neutral Agree
disagree agree
For my monitoring
work, it is important
that I maintain a
continuous
awareness of the
network security state
It is important to have
a "human in the loop"
for the detection and
preliminary analysis
of potential security
events - this process
cannot be carried out
by automated
systems alone
Human analysts
monitoring the
network are capable
of detecting network
anomalies that are
missed by automated
systems
The monitoring setup
I use enables me to
detect network
anomalies that are
missed by automated
systems
I am often required to
make decisions on
the accuracy of the
alerts produced by
automated systems
25 / 37
Maintaining
awareness of the
network security state
is important in
enabling me to make
decisions on the
accuracy of alerts
produced by
automated monitoring
systems
Data Fusion: Please indicate your level of agreement with the following assertions.
Required
Strongly Strongly
Disagree Neutral Agree
disagree agree
In monitoring , I am
required to watch
multiple monitors
depicting different
data at one time
I am required to
watch multiple
dashboards on the
same monitor
depicting different
data at one time
I am required to
simultaneously
monitor information
from multiple
monitoring sources
(eg. network traffic,
IDS logs, firewall
logs, etc)
26 / 37
Monitoring the status
of the network
involves interacting
with an IDS and other
monitoring tools in
addition to following
external information
resources for
vulnerabilities
I am required to
simultaneously
monitor the state of
multiple network
elements (e.g.
individual machines,
database server, web
server, etc.)
I am required to
monitor the network,
while carrying out
other tasks
simultaneously (e.g.
responding to emails,
carrying out incident
response)
SIEMS require
collecting and storing
billions of logs every
day, we have limited
resources and are not
able to keep these
logs for long periods
of time
Aggregation of
billions of log alerts a
day presented in
heterogeneous data
structures makes
analysis a challenge
27 / 37
Devices may
generate logs that are
either incomplete or
inconsistent (e.g.
different time-stamps)
making analysis an
even bigger
challenge
The tools I use are
effective at
supporting data
fusion and correlation
across incidents and
data sources
Network Configurations: Please indicate your level of agreement with the following assertions.
Required
Strongly Strongly
Disagree Neutral Agree
disagree agree
Keeping up with
changing
configurations in the
network is difficult,
but necessary to
provide the context
needed to analyse
and diagnose an alert
Keeping track and
knowing the network
environment is
important to detect
attacks
I track network
configurations using
personal memory
28 / 37
I track network
configurations by
keeping an updated
database of network
devices and
configuration
changes
Do you have any further comments on any of the topics covered on this page?
29 / 37
Page 7: Any further comments?
Do you have any further comments you would like to make about aspects of network security
monitoring in SOCs that have not been covered in this survey?
30 / 37
Page 8: Thank you for participating.
Thank you for completing this survey. Your answers have been recorded. If you have a concern
about any aspect of this project, please feel free to contact the researchers, Louise Axon and Bushra
Alahmadi (louise.axon@cs.ox.ac.uk, bushra.alahmadi@cs.ox.ac.uk).
32 / 37
FR - France
FX - France, Metropolitan
GA - Gabon
GB - United Kingdom
GD - Grenada
GE - Georgia
GF - French Guiana
GG - Guernsey
GH - Ghana
GI - Gibraltar
GL - Greenland
GM - Gambia, The
GN - Guinea
GP - Guadeloupe
GQ - Equatorial Guinea
GR - Greece
GS - South Georgia and the Islands
GT - Guatemala
GU - Guam
GW - Guinea-Bissau
GY - Guyana
HK - Hong Kong
HM - Heard Island and McDonald Islands
HN - Honduras
HR - Croatia
HT - Haiti
HU - Hungary
ID - Indonesia
IE - Ireland
IL - Israel
IM - Isle of Man
IN - India
IO - British Indian Ocean Territory
IQ - Iraq
IR - Iran
IS - Iceland
IT - Italy
JE - Jersey
JM - Jamaica
JO - Jordan
JP - Japan
KE - Kenya
33 / 37
KG - Kyrgyzstan
KH - Cambodia
KI - Kiribati
KM - Comoros
KN - Saint Kitts and Nevis
KP - Korea, North
KR - Korea, South
KW - Kuwait
KY - Cayman Islands
KZ - Kazakhstan
LA - Laos
LB - Lebanon
LC - Saint Lucia
LI - Liechtenstein
LK - Sri Lanka
LR - Liberia
LS - Lesotho
LT - Lithuania
LU - Luxembourg
LV - Latvia
LY - Libya
MA - Morocco
MC - Monaco
MD - Moldova
ME - Montenegro
MF - Saint Martin
MG - Madagascar
MH - Marshall Islands
MK - Macedonia
ML - Mali
MM - Burma
MN - Mongolia
MO - Macau
MP - Northern Mariana Islands
MQ - Martinique
MR - Mauritania
MS - Montserrat
MT - Malta
MU - Mauritius
MV - Maldives
MW - Malawi
MX - Mexico
34 / 37
MY - Malaysia
MZ - Mozambique
NA - Namibia
NC - New Caledonia
NE - Niger
NF - Norfolk Island
NG - Nigeria
NI - Nicaragua
NL - Netherlands
NO - Norway
NP - Nepal
NR - Nauru
NU - Niue
NZ - New Zealand
OM - Oman
PA - Panama
PE - Peru
PF - French Polynesia
PG - Papua New Guinea
PH - Philippines
PK - Pakistan
PL - Poland
PM - Saint Pierre and Miquelon
PN - Pitcairn Islands
PR - Puerto Rico
PS - Gaza Strip
PS - West Bank
PT - Portugal
PW - Palau
PY - Paraguay
QA - Qatar
RE - Reunion
RO - Romania
RS - Serbia
RU - Russia
RW - Rwanda
SA - Saudi Arabia
SB - Solomon Islands
SC - Seychelles
SD - Sudan
SE - Sweden
SG - Singapore
35 / 37
SH - Saint Helena, Ascension, and Tristan da Cunha
SI - Slovenia
SJ - Svalbard
SK - Slovakia
SL - Sierra Leone
SM - San Marino
SN - Senegal
SO - Somalia
SR - Suriname
SS - South Sudan
ST - Sao Tome and Principe
SV - El Salvador
SX - Sint Maarten
SY - Syria
SZ - Swaziland
TC - Turks and Caicos Islands
TD - Chad
TF - French Southern and Antarctic Lands
TG - Togo
TH - Thailand
TJ - Tajikistan
TK - Tokelau
TL - Timor-Leste
TM - Turkmenistan
TN - Tunisia
TO - Tonga
TR - Turkey
TT - Trinidad and Tobago
TV - Tuvalu
TW - Taiwan
TZ - Tanzania
UA - Ukraine
UG - Uganda
UM - United States Minor Outlying Islands
US - United States
UY - Uruguay
UZ - Uzbekistan
VA - Holy See (Vatican City)
VC - Saint Vincent and the Grenadines
VE - Venezuela
VG - British Virgin Islands
VI - Virgin Islands
36 / 37
VN - Vietnam
VU - Vanuatu
WF - Wallis and Futuna
WS - Samoa
XK - Kosovo
YE - Yemen
YT - Mayotte
ZA - South Africa
ZM - Zambia
ZW - Zimbabwe
37 / 37