evilcohort: detecting communities of malicious accounts on

19
EvilCohort: Detecting Communities of Malicious Accounts on Online Services Gianluca Stringhini 1 , Pierre Mourlanne 2 , Gregoire Jacob 3 , Manuel Egele 4 , Christopher Kruegel 2,3 , and Giovanni Vigna 2,3 University College London 1 UC Santa Barbara 2 Lastline Inc. 3 Boston University 4

Upload: others

Post on 08-May-2022

4 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: EvilCohort: Detecting Communities of Malicious Accounts on

EvilCohort: Detecting Communities of Malicious Accounts on

Online Services

Gianluca Stringhini1, Pierre Mourlanne2, Gregoire Jacob3,

Manuel Egele4, Christopher Kruegel2,3, and Giovanni Vigna2,3

University College London1 UC Santa Barbara2 Lastline Inc. 3 Boston University4

Page 2: EvilCohort: Detecting Communities of Malicious Accounts on

Online services are abused by cybercriminals

EvilCohort: Detecting Communities of Malicious Accounts on Online Services 2

• Spam• Crawling sensitive information / documents• Storing illegal content• Running DoS attacks / hosting C&C servers

Page 3: EvilCohort: Detecting Communities of Malicious Accounts on

State of the art in malicious account detectionCurrent techniques leverage domain-specific elements to detect malicious activity on one type of service

Examples:

• Forums [Niu2007]

• Blogs [Thomason2007]

• Youtube [Benevenuto2009]

• Social Networks [Mittal2009], [Grier2010], [Stringhini2010]

• Webmail accounts [Taylor2006], [Stringhini2015]

EvilCohort: Detecting Communities of Malicious Accounts on Online Services 3

There are elements that are common to malicious activity on all online services!

Page 4: EvilCohort: Detecting Communities of Malicious Accounts on

Botnets accessing online accounts

EvilCohort: Detecting Communities of Malicious Accounts on Online Services 4

• Performance reasons• Resilience reasons

Communities of accounts

Page 5: EvilCohort: Detecting Communities of Malicious Accounts on

Advantages of community detection

EvilCohort: Detecting Communities of Malicious Accounts on Online Services 5

Service-agnosticCan be done on any service that uses accounts

Activity-agnosticWe only look at how accounts are accessed

Different types of cybercriminal operations• Crawl the online service• Use the service as C&C channel• Use the service as a “drop” service

Page 6: EvilCohort: Detecting Communities of Malicious Accounts on

Distributed access is prevalent

Web-based email service logs, 1 day period

72M emails sent by 21M distinct accounts

170k vetted spam accounts for ground truth

• 66k accounts accessed by a single IP address

• 104k accounts accessed by multiple IP addresses

Just looking at accounts that are accessed by many IPaddresses does not work (32% FPs for accountsaccessed by 10+ IPs)

EvilCohort: Detecting Communities of Malicious Accounts on Online Services 6

Page 7: EvilCohort: Detecting Communities of Malicious Accounts on

Our system: EvilCohort

• Phase I: data collection

• Phase II: building the graph representation

• Phase III: finding communities

• Phase IV (optional): characterizing communities

EvilCohort: Detecting Communities of Malicious Accounts on Online Services 7

Page 8: EvilCohort: Detecting Communities of Malicious Accounts on

Phase I: data collection

EvilCohort: Detecting Communities of Malicious Accounts on Online Services 8

Timestamp_1, IP_address_1, Account_1Timestamp_2, IP_address_2, Account_2Timestamp_3, IP_address_3, Account_3Timestamp_4, IP_address_4, Account_1…

Page 9: EvilCohort: Detecting Communities of Malicious Accounts on

Phase II: building graph representation

EvilCohort: Detecting Communities of Malicious Accounts on Online Services 9

5

2

3

4

3

7

4

• Vertices are online accounts• Edges’ weight is number of shared IP addresses

Page 10: EvilCohort: Detecting Communities of Malicious Accounts on

Phase III: finding communities

We apply the ``Louvain’’ method for clustering:

• Iterative method

• Based on modularity optimization

We can prune edges with low weight to improve precision (threshold s)

EvilCohort: Detecting Communities of Malicious Accounts on Online Services 10

Page 11: EvilCohort: Detecting Communities of Malicious Accounts on

Phase IV (optional): characterizing communities

• User agent correlation

• Event-based time series

• IP address and account usage

These filters can be used to further prune false positives

EvilCohort: Detecting Communities of Malicious Accounts on Online Services 11

Page 12: EvilCohort: Detecting Communities of Malicious Accounts on

s = 65 1,331 accounts

1,247 known (~1.2%)

0 false positives

Grown knowledge: 84 accounts (~0.08%)

s = 10 25,490 accounts

16,626 known (~16%)

433 false positives(~1.7%)

Grown knowledge: 8,864 accounts (~7.7%)

s = 5 77,910 accounts

51,868 known (~49.9%)

2,337 false positives(~3%)

Grown knowledge: 26,042 accounts (~25%)Grown knowledge: 40,728 accounts (~39%)

s = 2 135,602 accounts

94,874 known (~91%)

12,350 false positives(~9.1%)

Selection of sGround truth: 103k spam accounts accessed by 2+ IPs

False positive if ≤ 10% of the accounts sent spam

EvilCohort: Detecting Communities of Malicious Accounts on Online Services 12

We decided to set s to 10 for our experiments

Page 13: EvilCohort: Detecting Communities of Malicious Accounts on

Results in the wild

Webmail activity dataset: email events

5 month period, 1.2B emails

1.2M malicious accounts, 500k unknown, 23k FP (1.9%)

EvilCohort: Detecting Communities of Malicious Accounts on Online Services 13

Online social network dataset: login events

8 day period, 14M events, 4 social networks

111k malicious accounts

Page 14: EvilCohort: Detecting Communities of Malicious Accounts on

Analysis of the results

111k accounts formed 83 communities

5 communities showed characteristics that are typical of legitimate accounts (according to at least one of the postprocessing filters)

EvilCohort: Detecting Communities of Malicious Accounts on Online Services 14

Very small communities (< 8 accounts)

Huge community (5,272 accounts)

What is Charlie?

Charlie

Page 15: EvilCohort: Detecting Communities of Malicious Accounts on

Event-based time series

EvilCohort: Detecting Communities of Malicious Accounts on Online Services 15

Regular accounts show diurnal patterns

Malicious accounts show bursts in activity

Charlie shows a weird behavior

Page 16: EvilCohort: Detecting Communities of Malicious Accounts on

Account usage over time

EvilCohort: Detecting Communities of Malicious Accounts on Online Services 16

Legitimate accounts: still diurnal patterns

Malicious accounts: synchronized access

Charlie: regular + synchronized access

We conclude that Charlie is likely composed of compromised accounts

Page 17: EvilCohort: Detecting Communities of Malicious Accounts on

EvilCohort: discussion

Service and activity independentAccounts do not need to perform malicious activity to be detected

EvilCohort: Detecting Communities of Malicious Accounts on Online Services 17

Limitations• Only works on accounts accessed by multiple IP addresses• Does not distinguish between fake and compromised accounts

Our system detects botnet-like activity, legitimate accounts are unlikely to form communities

Page 18: EvilCohort: Detecting Communities of Malicious Accounts on

Conclusions

I presented EvilCohort, a system that detects malicious accounts on online services by identifying communities of accounts that are accessed by a common set of IP addresses

We ran EvilCohort on two real-world datasets, and detected more than one million malicious accounts

EvilCohort: Detecting Communities of Malicious Accounts on Online Services 18

Page 19: EvilCohort: Detecting Communities of Malicious Accounts on

[email protected]

@gianluca_string