results of the darpa 1998 offline intrusion detection...

29
MIT Lincoln Laboratory 1 RichardLippmann 5/99 Results of the DARPA 1998 Offline Intrusion Detection Evaluation Richard P. Lippmann, Robert K. Cunningham, David J. Fried, Isaac Graf, Kris R. Kendall, Seth E. Webster, Marc A. Zissman [email protected] MIT Lincoln Laboratory Room S4-121 244 Wood Street Lexington, MA 02173-0073 Presented at the Recent Advances in Intrusion Detection, RAID 99 Conference, 7-9 September West Lafayette, Indiana, USA

Upload: others

Post on 12-Aug-2020

1 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory1

Richard Lippmann 5/99

Results of the DARPA 1998 Offline Intrusion Detection Evaluation

Richard P. Lippmann, Robert K. Cunningham, David J. Fried, Isaac Graf, Kris R. Kendall, Seth E. Webster, Marc A. Zissman

[email protected]

MIT Lincoln LaboratoryRoom S4-121

244 Wood StreetLexington, MA 02173-0073

Presented at the Recent Advances in Intrusion Detection, RAID 99 Conference, 7-9 September

West Lafayette, Indiana, USA

Page 2: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory2

Richard Lippmann 8/99

Outline

• Background and Introduction• Analysis/Synthesis Approach to Generate

Normal Background Traffic• Attacks• Results• Summary and Conclusions

Page 3: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory3

Richard Lippmann 8/99

Goal of DARPA 1998 Intrusion Detection Evaluation

COLLECT NORMAL DATA

GENERATE ATTACKS

IMPROVEALGORITHMS

EVALUATEALGORITHMS

• Evaluations Required to Determine Current System Capabilities

• Lead to Iterative Performance Improvements

• Difficult Because No Standard Comparison Metrics, No Existing Attack or Background Traffic Collections, Privacy/Security Restrictions

Page 4: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory4

Richard Lippmann 8/99

Desired Receiver Operating Characteristic Curve (ROC) Performance

• Goal is to Reduce False Alarm Rates by Two to Three Orders of Magnitude and Improve Attack Detection Accuracy

0

20

40

60

80

100

0.1 1 10 100 1000 10000FALSE ALARMS PER DAY

NEXTGENERATION

SYSTEMS

VERYLOW

LOW HIGH

CURRENTSYSTEMS

Page 5: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory5

Richard Lippmann 8/99

Major Tasks and Timeline

7/97 7/981/98

July 6 - Sep 14

Nov 9- Dec 12

Dec15-17

IMPORTANT DATES

1/99

Oct 26

DELIVER 7 WEEKS OF TRAINING DATA

DELIVER 2 WEEKS OF TEST DATA

ANALYZE RETURNED DATA

EVALUATION W O R K S H O P- PI MEETING

GENERATE ATTACKSAND BACKGROUND

TRAFFIC

Page 6: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory6

Richard Lippmann 8/99

Data Types and Evaluation Overview

• Focus on UNIX, Outsider Attacks• Generate More than Two Months of Data with Attacks

– Network Sniffing Data (All Packets In/Out of Simulated Base)– Host Audit Data (Solaris Host BSM Audit Records)– Host File System Dumps (Solaris)

• Analyze and Compare False Alarm and Detection Rates

ROC CURVEANALYSES

INTRUSION DETECTOR

OUTPUTSCORES

DETECTIONVERSUS

FALSE ALARMS

AUDITDATA

SNIFFED

DATA

DISKDUMPS

Page 7: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory7

Richard Lippmann 8/99

Outline

• Background and Introduction• Analysis/Synthesis Approach to Generate

Normal Background Traffic• Attacks• Results• Summary and Conclusions

Page 8: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory8

Richard Lippmann 8/99

Corpus Generation Options

• Option I: Sniff/Audit Real Operational Data and Attack Base– Real-World, but Can’t Attack Operational Base and Can’t Release

Private Email, Passwords, Userid’s, ...

• Option II: Sanitize Operational Data, Mix in Attacks– Too Difficult to Sanitize All Data Types, Mixing in Attacks Would

Introduce Artifacts

• Option III - Synthesize Both Normal and Attack Sessions on a Private Network

– Generate Non-Sensitive Traffic Similar to That Seen on a Base Using Public Domain and Randomly Generated Data Sources

– Automate Normal Traffic Generation and Attacks Using Same Network Software (e.g. sendmail, ftp, telnet ) Used on Base

– Distribute Sniffing and Audit Data for Training and Testing Without Security or Privacy Concerns

Page 9: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory9

Richard Lippmann 8/99

Analysis/Synthesis Approach

• Examine 4 Months of Data From Hanscom Air Force Base and More than 50 Other Bases, and Add Attacks

• Recreate Traffic on Simulation Network

ROUTER

SIMULATION

N E T W O R K

SNIFFED

DATA

AUDIT

DATA

SNIFFED

DATA

RECENT

ATTACKSANALYZE

Page 10: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory10

Richard Lippmann 8/99

Simulation Network OverviewSimulation Network Overview

OutsideInternet

Eyrie AF Base

•http

•smtp•pop3

•FTP

•IRC

•Telnet

•X•SQL/Telnet

•DNS

•finger

•snmp•t ime

Services/Protocols

Inside

Router

Simulated UNIX Hosts (1000’s)

Sniffer

•Secretaries•Programmers

•Workers

•Managers

•System Administrators

•Attackers

Simulated Users (100’s)

Page 11: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory11

Richard Lippmann 8/99

Simulation Network DetailsSimulation Network Details

INSIDEGATEWAY

P 2

“ INSIDE” “ INSIDE” (172.16 - eyr ie.af .mil)

“OUTSIDE” “OUTSIDE” (Many IP Addresses)

PC

PC

PC

Web Server

Web Server

Web Server

Sniffer

Linux S u n O S Solar is Solar isSolar is

OUTSIDEWS

GATEWAY

OUTSIDEWEB

GATEWAY

Audit Host

Work Station

Work Station

Work Station

Work Station

Work Station

Work Station

S p a r c S p a r c

U l t raU l t ra

4 8 6

CISCOROUTER

AUDIT

DATA

SNIFFED

DATADISK

DUMPS

P 2 P 2

Page 12: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory12

Richard Lippmann 8/99

Outline

• Background and Introduction• Analysis/Synthesis Approach to Generate

Normal Background Traffic• Attacks• Results• Summary and Conclusions

Page 13: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory13

Richard Lippmann 8/99

Attackers and Victims in SimulationAttackers and Victims in Simulation

INSIDEGATEWAY

P2

“ INSIDE” “INSIDE” “OUTSIDE” “OUTSIDE”

PC

PC

PC

Web Server

Web Server

Web Server

Sniffer

Linux S u n O S Solar is Solar is

P2

Solar is

OUTSIDEWS

GATEWAY

OUTSIDEWEB

GATEWAY

P2

BSM Audit Host

Work Station

Work Station

Work Station

Work Station

Work Station

Work Station

S p a r c S p a r c

U l t raU l t ra

4 8 6

Victims

CISCOROUTER

Attackers

• Attackers Develop Novel, Never-Before-Seen Attacks

Page 14: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory14

Richard Lippmann 8/99

38 Attack Types in 1998 Test Data

SunOSinternal

Solaris Server(audited)

Linuxinternal

CiscoRouter

• d i c t i o n a r y

• f t p - wr i te

• g u e s t

• p h f

• h t t p t u n n e l

• x l o c k

• x s n o o p

• s n m p g e t a t t a c kDENIAL

OF SERVICE(11 Types,

43 Instances)

R E M O T E

TO LOCAL(14 Types,

17 Instances)

• l o a d m o d u l e

• ps

• p e r l

• x t e r m

• e j e c t

• f f b c o n f i g

• f d f o r m a t

• ps

USER TO ROOT(7 Types,

38 Instances)

• 120 Attacks in 2 Weeks of Test Data

SURVEILLANCE

/PROBE(6 Types,

22 Instances)

• ip s w e e p

• n m a p

• p o r t s w e e p

• s a t a n

• m s c a n

• s a i n t

• b a c k

• N e p t u n e

• P i n g o f d e a t h

• S m u r f

• l a n d

• A p a c h e 2

• M a i l b o m b

• P r o c e s s T a b l e

• U D P S t o r m

•back•Neptune•Ping of death

•Smurf•Teardrop• land

•Apache2•Mailbomb•Process Table

•UDP Storm

• d i c t i o n a r y

• f t p - wr i te

• g u e s t

• p h f

• h t t p t u n n e l

• x l o c k

• x s n o o p

• d i c t i o n a r y

• f t p - wr i te

• g u e s t

• i m a p

• p h f

• ip s w e e p

• n m a p

• p o r t s w e e p

• s a t a n

• m s c a n

• s a i n t

• ip s w e e p

• n m a p

• p o r t s w e e p

• s a t a n

• m s c a n

• s a i n t

• ip s w e e p

• n m a p

• p o r t s w e e p

• s a t a n

• m s c a n

• s a i n t

• h t t p t u n n e l

• n a m e d

• s e n d m a i l

• x l o c k

• x s n o o p

•back

•Neptune•Ping of death•Smurf

•Syslogd• land•Apache2

•Mailbomb•Process Table•UDP Storm

≡ test only

Page 15: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory15

Richard Lippmann 8/99

Novel Sendmail Remote to User Attack

Attacker Victim1 sendmail (port 25) /e tc /p a s s w d :

roo t : * :0 :0 :admin : /b in / sh

joe : * :1 :2 :use r : /b in /s h

m a r y: * :2 :2 :use r : /b in /s h

al ice: * : 3 :2 :use r : / b i n / sh

bob : * :4 :2 : /b in /sh

w00t::0:0:gotcha!!: /bin/sh

2

3 telnet (port 23)

• Novel Attack Code Developed for this Evaluation

• To Our Knowledge No One Else has Attack Code that Exploits

this Vulnerability

• An Attacker Sends One Email message to the Victim with a MIME

header field that Causes a Buffer Overflow and Modifies the

Password Fi le

• After this the Attacker Has Free Access to the Victim Machine as

Root using Telnet

Page 16: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory16

Richard Lippmann 8/99

Training Data Traffic, Week 5, Friday

PORTSWEEP

LOADMODULE

FORMAT - 1

EJECTFORMAT - 2

NEPTUNE (synflood)

SMURF

ROOTKIT (tftp)

Page 17: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory17

Richard Lippmann 8/99

Outline

• Background and Introduction• Analysis/Synthesis Approach to Generate

Normal Background Traffic• Attacks• Results

– Participants– Generating Receiver Operating Characteristic

(ROC) Curves– Overall ROC of Best Composite System– ROCs With Network Sniffing Data for Four Attack

Categories (Denial of Service, Probes, User to Root, Remote to Local)

– ROC with Host Audit Data for User to Root Attacks

• Summary and Conclusions

Page 18: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory18

Richard Lippmann 8/99

Participants and Systems

• Six Participants Submitted Seven Systems

– Network Sniffer Inputs Only (3)

– Host Audit BSM Inputs Only (2)

– Both Host Audit and Sniffer Inputs (1)

– File System Dumps (1)

• All Participants Followed the Blind Test Procedures

• System Types

– Finite-State Machine or Rule-Based Signature Detection

– Expert Systems

– Pattern Classification/Data Mining Trained System

Page 19: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory19

Richard Lippmann 8/99

Generating A Receiver Operating Characteristic (ROC) Curve

• Vary Threshold to Obtain Different False Alarm and Miss Values and Trace out ROC Curve

DeclaredIntrusion

WarningValue

> Threshold ?

NORMALCONNECTION

CORRECTREJECTION

=

INTRUSION = MISS

NORMALCONNECTION

FALSEALARM

=

INTRUSION = DETECTION

NO

YES

DeclaredNormal Session

TRANSCRIPTS

WarningValue

INTRUSION DETECTION

SYSTEM

Page 20: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory20

Richard Lippmann 8/99

Best Composite ROC Across All Systems for All Attacks

• Roughly 65% Detection at 5 False Alarms Per Day• Low False Alarm Rate, But Poor Detection Accuracy• Most Systems Miss New and Novel Attacks

0

20

40

60

80

100

0 20 40 60 80 100FALSE ALARMS PER DAY

BEST COMPOSITESYSTEM

Attacks: 120Normal: 660,049

2 σσ

Page 21: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory21

Richard Lippmann 8/99

ROC’s for Probe Attacks Using Network Sniffing Data

•Good Performance for Old and New Probes•Some Research Systems Find Almost all Probe Attacks at Low (1 False Alarm Per Day) False Alarm Rates•Old and New Probes are Similar (Satan, IP Sweeps, NMAP)

0

20

40

60

80

100

0 20 40 60 80 100FALSE ALARMS PER DAY

Attacks: 17Normal: 660,049

2 σσ

Page 22: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory22

Richard Lippmann 8/99

ROC’s for Denial of Service (DoS) Attacks Using Network Sniffing Data

•Research Systems Don’t Find all DoS Attacks•Systems Find Old Attacks but Miss New Attacks (Process Table Exhaustion, Mail Bomb, Chargen/Echo Storm)

0

20

40

60

80

100

0 20 40 60 80 100FALSE ALARMS PER DAY

Attacks: 43

Normal: 660,049

Page 23: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory23

Richard Lippmann 8/99

ROC’s for User to Root (u2r) Attacks Using Network Sniffing Data

•Research Systems Don’t Find all User to Root Attacks•Research Systems Perform Substantially Better than Baseline Keyword Reference System Which is Similar to Many Commercial and Government Systems

0

20

40

60

80

100

0 20 40 60 80 100FALSE ALARMS PER DAY

Attacks: 38Normal: 660,049

KEYWORDBASELINE

Page 24: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory24

Richard Lippmann 8/99

ROC’s for Remote to Local (r2l) Attacks Using Network Sniffing Data

• All Systems Have Low Detection Rates• Many New Attacks, Highly Varied Attack Mechanisms (imap, dictionary, http tunnel, named,sendmail, xlock, phf, ftp-write)

0

20

40

60

80

100

0 20 40 60 80 100FALSE ALARMS PER DAY

Attacks: 22Normal: 660,049

KEYWORD BASELINE

Page 25: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory25

Richard Lippmann 8/99

ROC’s for User to Root (u2r) Attacks Using Host Audit Data

• Excellent Performance Using Host Auditing to Detect Local Users Illegally Becoming Root• But This Requires Auditing on Each Host and is Only for User to Root Attacks

0

20

40

60

80

100

0 20 40 60 80 100FALSE ALARMS PER DAY

Attacks: 22

Normal: 4,600

Page 26: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory26

Richard Lippmann 8/99

Outline

• Background and Introduction• Analysis/Synthesis Approach to Generate

Normal Background Traffic• Attacks• Results• Summary and Conclusions

Page 27: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory27

Richard Lippmann 8/99

Best Combination System from This Evaluation Compared to Keyword Baseline

• False Alarm Rate Is More Than 100 Times Lower• Detection Rate Is Significantly Better• Keyword Baseline Performance Similar to Commercial and

Government Keyword-based Systems

0

20

40

60

80

100

0.001 0.01 0.1 1 10 100FALSE ALARMS (%)

KEYWORDBASELINE

BESTCOMBINATION

Page 28: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory28

Richard Lippmann 8/99

Best Systems in This Evaluation Don’t Accurately Detect New Attacks

• Systems Generalize Well to New Probe and User to Root Attacks, but Miss New Denial of Service and Remote to Local Attacks

• Basic Detection Accuracy for Old Attacks Must Also Improve

PROBE(14,3)

DOS(34,9)

U2R(27,11)

R2L(5,17)

0

20

40

60

80

100

CATEGORY

OLD

NEW

Page 29: Results of the DARPA 1998 Offline Intrusion Detection ...archive.ll.mit.edu/ideval/files/RAID_1999a.pdf– Network Sniffing Data (All Packets In/Out of Simulated Base) ... • An Attacker

MIT Lincoln Laboratory29

Richard Lippmann 8/99

Summary and Future Plans

• We Have Developed an Intrusion Detection Test Network Which Simulates a Typical Air Force Base

– Generate Realistic Background Traffic With 1000’s of Simulated Hosts and 100’s of Simulated Users

– Insert More Than 35 Types of Automated Attacks

– Measure Both Detection and False Alarm Rates

• The 1998 DARPA Evaluation Successfully Demonstrated

1) Research Intrusion Detection Systems Improve Dramatically Over Existing Keyword Systems

2) Research Systems, However, Miss New Denial-of-service and Remote-to-local Attacks and Do Not Perfectly Detect Old Attacks

• The 1999 DARPA Evaluation Will Add Windows NT Hosts and Many New Attacks

– Focus in on Detecting New Attacks and Maintaining Low False Alarm Rates