Download presentation
Presentation is loading. Please wait.
1
Enabling Worm and Malware Investigation Using Virtualization (Demo and poster this afternoon) Dongyan Xu, Xuxian Jiang CERIAS and Department of Computer Science Purdue University
2
The Team Lab FRIENDS Xuxian Jiang (Ph.D. student) Paul Ruth (Ph.D. student) Dongyan Xu (faculty) CERIAS Eugene H. Spafford External Collaboration Microsoft Research
4
Our Goal In-depth understanding of increasingly sophisticated worm/malware behavior
5
Outline Motivation An integrated approach Front-end : Collapsar (Part I) Back-end : vGround (Part II) Bringing them together On-going work
6
The Big Picture Proxy ARP Domain A Domain B GRE Worm Analysis Worm Capture
7
Front-End: Collapsar Enabling Worm/Malware Capture * X. Jiang, D. Xu, “Collapsar: a VM-Based Architecture for Network Attack Detention Center”, 13 th USENIX Security Symposium (Security’04), 2004. Part I
8
General Approach Promise of honeypots Providing insights into intruders’ motivations, tactics, and tools Highly concentrated datasets w/ low noise Low false-positive and false negative rate Discovering unknown vulnerabilities/exploitations Example: CERT advisory CA-2002-01 (solaris CDE subprocess control daemon – dtspcd)
9
Current Honeypot Operation Individual honeypots Limited local view of attacks Federation of distributed honeypots Deploying honeypots in different networks Exchanging logs and alerts Problems Difficulties in distributed management Lack of honeypot expertise Inconsistency in security and management policies Example: log format, sharing policy, exchange frequency
10
Our Approach: Collapsar Based on the HoneyFarm idea of Lance Spitzner Achieving two (seemingly) conflicting goals Distributed honeypot presence Centralized honeypot operation Key ideas Leveraging unused IP addresses in each network Diverting corresponding traffic to a “detention” center (transparently) Creating VM-based honeypots in the center
11
VM-based Honeypot Collapsar Architecture Redirector Correlation Engine Management Station Production Network Collapsar Center Attacker Front-End
12
Comparison with Current Approaches Overlay-based approach (e.g., NetBait, Domino overlay) Honeypots deployed in different sites Logs aggregated from distributed honeypots Data mining performed on aggregated log information Key difference: where the attacks take place (on-site vs. off-site)
13
Comparison with Current Approaches Sinkhole networking approach (e.g., iSink ) “Dark” space to monitor Internet abnormality and commotion (e.g. msblaster worms) Limited interaction for better scalability Key difference: contiguous large address blocks (vs. scattered addresses)
14
Comparison with Current Approaches Low-interaction approach (e.g., honeyd, iSink ) Highly scalable deployment Low security risks Key difference: emulated services (vs. real things) Less effective to reveal unknown vulnerabilities Less effective to capture 0-day worms
15
Collapsar Design Functional components Redirector Collapsar Front-End Virtual honeypots Assurance modules Logging module Tarpitting module Correlation module
16
Collapsar Deployment Deployed in a local environment for a two-month period in 2003 Traffic redirected from five networks Three wired LANs One wireless LAN One DSL network ~ 50 honeypots analyzed so far Internet worms (MSBlaster, Enbiei, Nachi ) Interactive intrusions (Apache, Samba) OS: Windows, Linux, Solaris, FreeBSD
17
Incident: Apache Honeypot/VMware Vulnerabilities Vul 1: Apache (CERT® CA-2002-17) Vul 2: Ptrace (CERT® VU-6288429) Time-line Deployed: 23:44:03pm, 11/24/03 Compromised: 09:33:55am, 11/25/03 Attack monitoring Detailed log http://www.cs.purdue.edu/homes/jiangx/collapsar
18
Incident: Windows XP Honeypot/VMware Vulnerability RPC DCOM Vul. (Microsoft Security Bulletin MS03-026) Time-line Deployed: 22:10:00pm, 11/26/03 MSBlaster: 00:36:47am, 11/27/03 Enbiei: 01:48:57am, 11/27/03 Nachi: 07:03:55am, 11/27/03
19
Summary (Front-End) A novel front-end for worm/malware capture Distributed presence and centralized operation of honeypots Good potential in attack correlation and log mining Unique features Aggregation of Scattered unused (dark) IP addresses Off-site (relative to participating networks) attack occurrences and monitoring Real services for unknown vulnerability revelation
20
Back-End: vGround Enabling Worm/Malware Analysis Part II * X. Jiang, D. Xu, H. J. Wang, E. H. Spafford, “Virtual Playgrounds for Worm Behavior Investigation”, 8 th International Symposium on Recent Advances in Intrusion Detection (RAID’05), 2005.
21
Basic Approach A dedicated testbed Internet-inna-box (IBM), Blended Threat Lab (Symantec) DETER Goal: understanding worm behavior Static analysis/ execution trace Reverse Engineering ( IDA Pro, GDB, … ) Worm experiment within a limited scale Result: Only enabling relatively static analysis within a small scale
22
The Reality – Worm Threats Speed, Virulence, & Sophistication of Worms Flash/Warhol Worms Polymorphic/Metamorphic Appearances Zombie Networks (DDoS Attacks, Spam) What we also need A high-fidelity, large-scale, live but safe worm playground
23
Picture by Peter Szor, Symantec Corp. A Worm Playground
24
Requirements Cost & Scalability How about a topology with 2000+ nodes? Confinement In-house private use? Management & user convenience Diverse environment requirement Recovery from damages from a worm experiment re-installation, re-configuration, and reboot …
25
Our Approach vGround A virtualization-based approach Virtual Entities: Leveraging current virtual machine techniques Designing new virtual networking techniques User Configurability Customizing every node (end-hosts/routers) Enabling flexible experimental topologies
26
An Example Run: Internet Worms A shared infrastructure (e.g. PlanetLab) A worm playground Virtual Physical
27
Key Virtualization Techniques Full-System Virtualization Network Virtualization
28
Full-System Virtualization Emerging and New VM Techniques VMware, Xen, Denali, UML Supporting for real-world services DNS, Sendmail, Apache w/ “native” vulnerabilities Adopted technique: UML Deployability Convenience/Resource Efficiency
29
User-Mode Linux ( http://user-mode-linux.sf.net ) System-Call Virtualization User-Level Implementation Host OS Kernel Device Drivers Hardware Device DriversMMU Guest OS Kernel UM User Process 1 ptraceptrace UM User Process 2
30
New Network Virtualization Link Layer Virtualization User-Level Implementation Host OS Virtual Node 1Virtual Node 2 Virtual Switch 1 IP-IP
31
User Configurability Node Customization System Template End Node ( BIND, Apach, Sendmail, … ) Router ( RIP, OSPF, BGP, … ) Firewall ( iptables ) Sniffer/IDS ( bro, snort ) Topology Customization Language Network, Node Toolkits
32
Project Planetlab-Worm template slapper { image slapper.ext2 cow enabled startup { /etc/rc.d/init.d/httpd start } template router { image router.ext2 routing ospf startup { /etc/rc.d/init.d/ospfd start } router R1 { superclass router network eth0 { switch AS1_lan1 address 128.10.1.250/24 } network eth1 { switch AS1_AS2 address 128.8.1.1/24 } switch AS1_lan1 { unix_sock sock/as1_lan1 host planetlab6.millennium. berkeley.edu } switch AS1_AS2 { udp_sock 1500 host planetlab6.millennium. berkeley.edu } node AS1_H1 { superclass slapper network eth0 { switch AS1_lan1 address 128.10.1.1/24 gateway 128.10.1.250 } node AS1_H2 { superclass slapper network eth0 { switch AS1_lan1 address 128.10.1.2/24 gateway 128.10.1.250 } switch AS2_lan1 { unix_sock sock/as2_lan1 host planetlab1.cs.purdue.edu } switch AS2_AS3 { udp_sock 1500 host planetlab1.cs.purdue.edu } node AS2_H1 { superclass slapper network eth0 { switch AS2_lan1 address 128.11.1.5/24 gateway 128.11.1.250 } node AS2_H2 { superclass slapper network eth0 { switch AS2_lan1 address 128.11.1.6/24 gateway 128.11.1.250 } switch AS3_lan1 { unix_sock sock/as3_lan1 host planetlab8.lcs.mit.edu } router R2 { superclass router network eth0 { switch AS2_lan1 address 128.11.1.250/24 } network eth1 { switch AS1_AS2 address 128.8.1.2/24 } network eth2 { switch AS2_AS3 address 128.9.1.2/24 } node AS3_H1 { superclass slapper network eth0 { switch AS3_lan1 address 128.12.1.5/24 gateway 128.12.1.250 } node AS3_H2 { superclass slapper network eth0 { switch AS3_lan1 address 128.12.1.6/24 gateway 128.12.1.250 } router R3 { superclass router network eth0 { switch AS3_lan1 address 128.12.1.250/24 } network eth1 { switch AS2_AS3 address 128.9.1.1/24 } Networked Node Network System Template AS1_H1R1 AS1_H2 AS2_H1AS2_H2 R2R3 AS3_H1 AS3_H2
33
Features Scalability 3000 virtual hosts in 10 physical nodes Iterative Experiment Convenience Virtual node generation time: 60 seconds Boot-strap time: 90 seconds Tear-down time: 10 seconds Strict Confinement High Fidelity
34
Evaluation Current Focus Worm behavior reproduction Experiments Probing, exploitation, payloads, and propagation Further Potentials – on-going work Routing worms / Stealthy worms Infrastructure security (BGP)
35
Experiment Setup Two Real-World Worms Lion, Slapper, and their variants LionSlapper A vGround Topology 10 virtual networks 1500 virtual Nodes 10 physical machines in an ITaP cluster
36
Evaluation Target Host Distribution Detailed Exploitation Steps Malicious Payloads Propagation Pattern
37
Probing: Target Network Selection Lion Worms Slapper Worms 13243 80,81 http://www.iana.org/assignments/ipv4-address-space.
38
Exploitation (Lion) 1: Probing 2: Exploitation! 3: Propagation!
39
Exploitation (Slapper) 1: Probing 2: Exploitation! 3: Propagation!
40
Malicious Payload (Lion)
41
Propagation Pattern and Strategy Address-Sweeping Randomly choose a Class B address (a.b.0.0) Sequentially scan hosts a.b.0.0 – a.b.255.255 Island-Hopping Local subnet preference
42
Propagation Pattern and Strategy Address-Sweeping (Slapper Worm) Infected Hosts: 2% Infected Hosts: 5% Infected Hosts: 10% 192.168.a.b
43
Propagation Pattern and Strategy Island-Hopping Infected Hosts: 2% Infected Hosts: 5% Infected Hosts: 10%
44
Summary (Back-End) vGround – the back-end A Virtualization-Based Worm Playground Properties: High Fidelity Strict Confinement Good Scalability 3000 Virtual Hosts in 10 Physical Nodes High Resource Efficiency Flexible and Efficient Worm Experiment Control
45
Combining Collapsar and vGround Domain A Domain B GRE Worm Analysis Worm Capture
46
Conclusions An integrated virtualization-based platform for worm and malware investigation Front-end : Collapsar Back-end : vGround Great potential for automatic Characterization of unknown service vulnerabilities Generation of 0-day worm signatures Tracking of worm contaminations
47
On-going Work More real-world evaluation Stealthy worms Polymorphic worms Additional capabilities Collapsar center federation On-demand honeypot customization Worm/malware contamination tracking Automated signature generation
48
Thank you. Stop by our poster and demo this afternoon! For more information: Email: dxu@cs.purdue.eduxu@cs.purdue.edu URL: http://www.cs.purdue.edu/~dxuhttp://www.cs.purdue.edu/~dxu Google: “Purdue Collapsar Friends”
Similar presentations
© 2025 SlidePlayer.com. Inc.
All rights reserved.