- Sponsor:
- sigsac
It is our pleasure to welcome you to the 9th ACM Workshop on Artificial Intelligence and Security --- AISec 2016. AISec, having been annually co-located with CCS for nine consecutive years, is the premier meeting place for researchers interested in the intersection of security, privacy, AI, and machine learning. Its role as a venue has been to merge practical security problems with advances in AI and machine learning. In doing so, researchers also have been developing theory and analytics unique to this domain and have explored diverse topics such as learning in game-theoretic adversarial environments, privacy-preserving learning, and applications to spam and intrusion detection.
AISec 2016 drew a record 38 submissions, of which 12 (32%) were selected for publication and presentation. Submissions arrived from researchers in 16 countries, from a wide variety of institutions both academic and corporate. The accepted papers were organized into the following thematic groups:
Security Data Sets: Collection and analysis of data that can serve as a baseline for AI/ML research in security.
Machine Learning and Security in Practice: Systems that use machine learning to solve a particular security problem.
Foundations: Theoretical constructs and best practices for applying machine learning to security.
Privacy: Attacks on user privacy or anonymity, and privacy-preserving constructions of machine learning systems.
The keynote address will be given by Elie Bursztein of Google, Inc., whose talk is entitled, "Why is applying machine learning to anti-abuse so hard?" In this talk, Dr. Bursztein will discuss challenges in the reproducibility of scientific results from machine learning algorithms and what we can do about it. Dr. Bursztein's talk will touch on issues arising from proprietary hardware, dataset availability, adversarial machine learning, and the ethics of data. He will also consider several privacy questions related to machine learning models.
Proceeding Downloads
SherLock vs Moriarty: A Smartphone Dataset for Cybersecurity Research
In this paper we describe and share with the research community, a significant smartphone dataset obtained from an ongoing long-term data collection experiment. The dataset currently contains 10 billion data records from 30 users collected over a period ...
DeepDGA: Adversarially-Tuned Domain Generation and Detection
Many malware families utilize domain generation algorithms (DGAs) to establish command and control (C&C) connections. While there are many methods to pseudorandomly generate domains, we focus in this paper on detecting (and generating) domains on a per-...
Tracked Without a Trace: Linking Sessions of Users by Unsupervised Learning of Patterns in Their DNS Traffic
Behavior-based tracking is an unobtrusive technique that allows observers to monitor user activities on the Internet over long periods of time -- in spite of changing IP addresses. Previous work has employed supervised classifiers in order to link the ...
Identifying Encrypted Malware Traffic with Contextual Flow Data
Identifying threats contained within encrypted network traffic poses a unique set of challenges. It is important to monitor this traffic for threats and malware, but do so in a way that maintains the integrity of the encryption. Because pattern matching ...
Causality-based Sensemaking of Network Traffic for Android Application Security
Malicious Android applications pose serious threats to mobile security. They threaten the data confidentiality and system integrity on Android devices. Monitoring runtime activities serves as an important technique for analyzing dynamic app behaviors. ...
Secure Kernel Machines against Evasion Attacks
Machine learning is widely used in security-sensitive settings like spam and malware detection, although it has been shown that malicious data can be carefully modified at test time to evade detection. To overcome this limitation, adversary-aware ...
Prescience: Probabilistic Guidance on the Retraining Conundrum for Malware Detection
Malware evolves perpetually and relies on increasingly so- phisticated attacks to supersede defense strategies. Data-driven approaches to malware detection run the risk of becoming rapidly antiquated. Keeping pace with malware requires models that are ...
Discriminative Models for Multi-instance Problems with Tree Structure
Modelling network traffic is gaining importance to counter modern security threats of ever increasing sophistication. It is though surprisingly difficult and costly to construct reliable classifiers on top of telemetry data due to the variety and ...
True Friends Let You Down: Benchmarking Social Graph Anonymization Schemes
Greater demand for social graph data among researchers and analysts has fueled an increase in such datasets being published. Consequently, concerns about privacy breach have also risen steadily. To mitigate privacy risks a myriad of social graph ...
Change of Guard: The Next Generation of Social Graph De-anonymization Attacks
Past decade has seen active research in social graph de-anonymization with a variety of algorithms proposed. Previous algorithms used handcrafted tricks and were locked in a co-evolution of attack and defense with design of anonymization systems. We ...
Differentially Private Online Active Learning with Applications to Anomaly Detection
In settings where data instances are generated sequentially or in streaming fashion, online learning algorithms can learn predictors using incremental training algorithms such as stochastic gradient descent. In some security applications such as ...
A Dual Perturbation Approach for Differential Private ADMM-Based Distributed Empirical Risk Minimization
The rapid growth of data has raised the importance of privacy-preserving techniques in distributed machine learning. In this paper, we develop a privacy-preserving method to a class of regularized empirical risk minimization (ERM) machine learning ...
Index Terms
- Proceedings of the 2016 ACM Workshop on Artificial Intelligence and Security