detecting spam zombies by monitoring outgoing messages.bak

13
Detecting Spam Zombies by Monitoring Outgoing Messages Zhenhai Duan, Senior Member, IEEE, Peng Chen, Fernando Sanchez, Yingfei Dong, Member, IEEE, Mary Stephenson, and James Michael Barker Abstract—Compromised machines are one of the key security threats on the Internet; they are often used to launch various security attacks such as spamming and spreading malware, DDoS, and identity theft. Given that spamming provides a key economic incentive for attackers to recruit the large number of compromised machines, we focus on the detection of the compromised machines in a network that are involved in the spamming activities, commonly known as spam zombies. We develop an effective spam zombie detection system named SPOT by monitoring outgoing messages of a network. SPOT is designed based on a powerful statistical tool called Sequential Probability Ratio Test, which has bounded false positive and false negative error rates. In addition, we also evaluate the performance of the developed SPOT system using a two-month e-mail trace collected in a large US campus network. Our evaluation studies show that SPOT is an effective and efficient system in automatically detecting compromised machines in a network. For example, among the 440 internal IP addresses observed in the e-mail trace, SPOT identifies 132 of them as being associated with compromised machines. Out of the 132 IP addresses identified by SPOT, 126 can be either independently confirmed (110) or highly likely (16) to be compromised. Moreover, only seven internal IP addresses associated with compromised machines in the trace are missed by SPOT. In addition, we also compare the performance of SPOT with two other spam zombie detection algorithms based on the number and percentage of spam messages originated or forwarded by internal machines, respectively, and show that SPOT outperforms these two detection algorithms. Index Terms—Compromised machines, spam zombies, compromised machine detection algorithms. Ç 1 INTRODUCTION A major security challenge on the Internet is the existence of the large number of compromised machines. Such machines have been increasingly used to launch various security attacks including spamming and spreading mal- ware, DDoS, and identity theft [1], [10], [14]. Two natures of the compromised machines on the Internet—sheer volume and widespread—render many existing security counter- measures less effective and defending attacks involving compromised machines extremely hard. On the other hand, identifying and cleaning compromised machines in a network remain a significant challenge for system adminis- trators of networks of all sizes. In this paper, we focus on the detection of the compro- mised machines in a network that are used for sending spam messages, which are commonly referred to as spam zombies. Given that spamming provides a critical economic incentive for the controllers of the compromised machines to recruit these machines, it has been widely observed that many compromised machines are involved in spamming [17], [19], [25]. A number of recent research efforts have studied the aggregate global characteristics of spamming botnets (net- works of compromised machines involved in spamming) such as the size of botnets and the spamming patterns of botnets, based on the sampled spam messages received at a large e-mail service provider [25], [26]. Rather than the aggregate global characteristics of spam- ming botnets, we aim to develop a tool for system administrators to automatically detect the compromised machines in their networks in an online manner. We consider ourselves situated in a network and ask the following question: How can we automatically identify the compro- mised machines in the network as outgoing messages pass the monitoring point sequentially? The approaches devel- oped in the previous work [25], [26] cannot be applied here. The locally generated outgoing messages in a network normally cannot provide the aggregate large-scale spam view required by these approaches. Moreover, these ap- proaches cannot support the online detection requirement in the environment we consider. The nature of sequentially observing outgoing messages gives rise to the sequential detection problem. In this paper, we will develop a spam zombie detection system, named SPOT, by monitoring outgoing messages. SPOT is designed based on a statistical method called Sequential Probability Ratio Test (SPRT), developed by Wald in his seminal work [21]. SPRT is a powerful statistical method that can be used to test between two hypotheses (in our case, a machine is 198 IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, VOL. 9, NO. 2, MARCH/APRIL 2012 . Z. Duan and F. Sanchez are with the Department of Computer Science, Florida State University, Tallahassee, FL 32306-4530. E-mail: {duan, sanchez}@cs.fsu.edu. . P. Chen is with Juniper Networks, Sunnyvale, CA 94089. E-mail: [email protected]. . Y. Dong is with the Department of Electrical Engineering, University of Hawaii, 2540 Dole Street, Holmes Hall 483, Honolulu, HI 96822. E-mail: [email protected]. . M. Stephenson is with Information Technology Services, Florida State University, 2035 E. Paul Dirac Drive, 200 BFS, Tallahassee, FL 32310. E-mail: [email protected]. . J.M. Barker is with Information Technology Services, The University of North Carolina at Chapel Hill, 211 Manning Drive, Chapel Hill, NC 27599-3420. E-mail: [email protected]. Manuscript received 27 Aug. 2010; revised 9 May 2011; accepted 9 June 2011; published online 30 Sept. 2011. For information on obtaining reprints of this article, please send e-mail to: [email protected], and reference IEEECS Log Number TDSC-2010-08-0149. Digital Object Identifier no. 10.1109/TDSC.2011.49. 1545-5971/12/$31.00 ß 2012 IEEE Published by the IEEE Computer Society http://ieeexploreprojects.blogspot.com

Upload: ieeexploreprojects

Post on 13-May-2015

1.927 views

Category:

Documents


2 download

TRANSCRIPT

Page 1: Detecting spam zombies by monitoring outgoing messages.bak

Detecting Spam Zombies byMonitoring Outgoing Messages

Zhenhai Duan, Senior Member, IEEE, Peng Chen, Fernando Sanchez,

Yingfei Dong, Member, IEEE, Mary Stephenson, and James Michael Barker

Abstract—Compromised machines are one of the key security threats on the Internet; they are often used to launch various security

attacks such as spamming and spreading malware, DDoS, and identity theft. Given that spamming provides a key economic incentive

for attackers to recruit the large number of compromised machines, we focus on the detection of the compromised machines in a

network that are involved in the spamming activities, commonly known as spam zombies. We develop an effective spam zombie

detection system named SPOT by monitoring outgoing messages of a network. SPOT is designed based on a powerful statistical tool

called Sequential Probability Ratio Test, which has bounded false positive and false negative error rates. In addition, we also evaluate

the performance of the developed SPOT system using a two-month e-mail trace collected in a large US campus network. Our

evaluation studies show that SPOT is an effective and efficient system in automatically detecting compromised machines in a network.

For example, among the 440 internal IP addresses observed in the e-mail trace, SPOT identifies 132 of them as being associated with

compromised machines. Out of the 132 IP addresses identified by SPOT, 126 can be either independently confirmed (110) or highly

likely (16) to be compromised. Moreover, only seven internal IP addresses associated with compromised machines in the trace are

missed by SPOT. In addition, we also compare the performance of SPOT with two other spam zombie detection algorithms based on

the number and percentage of spam messages originated or forwarded by internal machines, respectively, and show that SPOT

outperforms these two detection algorithms.

Index Terms—Compromised machines, spam zombies, compromised machine detection algorithms.

Ç

1 INTRODUCTION

Amajor security challenge on the Internet is the existenceof the large number of compromised machines. Such

machines have been increasingly used to launch varioussecurity attacks including spamming and spreading mal-ware, DDoS, and identity theft [1], [10], [14]. Two natures ofthe compromised machines on the Internet—sheer volumeand widespread—render many existing security counter-measures less effective and defending attacks involvingcompromised machines extremely hard. On the other hand,identifying and cleaning compromised machines in anetwork remain a significant challenge for system adminis-trators of networks of all sizes.

In this paper, we focus on the detection of the compro-mised machines in a network that are used for sending spammessages, which are commonly referred to as spam zombies.

Given that spamming provides a critical economic incentivefor the controllers of the compromised machines to recruitthese machines, it has been widely observed that manycompromised machines are involved in spamming [17], [19],[25]. A number of recent research efforts have studied theaggregate global characteristics of spamming botnets (net-works of compromised machines involved in spamming)such as the size of botnets and the spamming patterns ofbotnets, based on the sampled spam messages received at alarge e-mail service provider [25], [26].

Rather than the aggregate global characteristics of spam-ming botnets, we aim to develop a tool for systemadministrators to automatically detect the compromisedmachines in their networks in an online manner. We considerourselves situated in a network and ask the followingquestion: How can we automatically identify the compro-mised machines in the network as outgoing messages passthe monitoring point sequentially? The approaches devel-oped in the previous work [25], [26] cannot be applied here.The locally generated outgoing messages in a networknormally cannot provide the aggregate large-scale spamview required by these approaches. Moreover, these ap-proaches cannot support the online detection requirement inthe environment we consider.

The nature of sequentially observing outgoing messagesgives rise to the sequential detection problem. In this paper,we will develop a spam zombie detection system, namedSPOT, by monitoring outgoing messages. SPOT is designedbased on a statistical method called Sequential ProbabilityRatio Test (SPRT), developed by Wald in his seminal work[21]. SPRT is a powerful statistical method that can be usedto test between two hypotheses (in our case, a machine is

198 IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, VOL. 9, NO. 2, MARCH/APRIL 2012

. Z. Duan and F. Sanchez are with the Department of Computer Science,Florida State University, Tallahassee, FL 32306-4530.E-mail: {duan, sanchez}@cs.fsu.edu.

. P. Chen is with Juniper Networks, Sunnyvale, CA 94089.E-mail: [email protected].

. Y. Dong is with the Department of Electrical Engineering, University ofHawaii, 2540 Dole Street, Holmes Hall 483, Honolulu, HI 96822.E-mail: [email protected].

. M. Stephenson is with Information Technology Services, Florida StateUniversity, 2035 E. Paul Dirac Drive, 200 BFS, Tallahassee, FL 32310.E-mail: [email protected].

. J.M. Barker is with Information Technology Services, The University ofNorth Carolina at Chapel Hill, 211 Manning Drive, Chapel Hill, NC27599-3420. E-mail: [email protected].

Manuscript received 27 Aug. 2010; revised 9 May 2011; accepted 9 June 2011;published online 30 Sept. 2011.For information on obtaining reprints of this article, please send e-mail to:[email protected], and reference IEEECS Log Number TDSC-2010-08-0149.Digital Object Identifier no. 10.1109/TDSC.2011.49.

1545-5971/12/$31.00 � 2012 IEEE Published by the IEEE Computer Society

http://ieeexploreprojects.blogspot.com

Page 2: Detecting spam zombies by monitoring outgoing messages.bak

compromised versus the machine is not compromised), asthe events (in our case, outgoing messages) occur sequen-tially. As a simple and powerful statistical method, SPRThas a number of desirable features. It minimizes theexpected number of observations required to reach adecision among all the sequential and nonsequentialstatistical tests with no greater error rates. This means thatthe SPOT detection system can identify a compromisedmachine quickly. Moreover, both the false positive andfalse negative probabilities of SPRT can be bounded byuser-defined thresholds. Consequently, users of the SPOTsystem can select the desired thresholds to control thefalse positive and false negative rates of the system.

In this paper, we develop the SPOT detection system toassist system administrators in automatically identifyingthe compromised machines in their networks. We alsoevaluate the performance of the SPOT system based on atwo-month e-mail trace collected in a large US campusnetwork. Our evaluation studies show that SPOT is aneffective and efficient system in automatically detectingcompromised machines in a network. For example, amongthe 440 internal IP addresses observed in the e-mail trace,SPOT identifies 132 of them as being associated withcompromised machines. Out of the 132 IP addressesidentified by SPOT, 126 can be either independentlyconfirmed (110) or are highly likely (16) to be compromised.Moreover, only seven internal IP addresses associated withcompromised machines in the trace are missed by SPOT. Inaddition, SPOT only needs a small number of observationsto detect a compromised machine. The majority of spamzombies are detected with as little as three spam messages.For comparison, we also design and study two other spamzombie detection algorithms based on the number of spammessages and the percentage of spam messages originatedor forwarded by internal machines, respectively. Wecompare the performance of SPOT with the two otherdetection algorithms to illustrate the advantages of theSPOT system.

The remainder of the paper is organized as follows: inSection 2, we discuss related work in the area of botnetdetection (focusing on spam zombie detection schemes). Weformulate the spam zombie detection problem in Section 3.Section 4 provides the necessary background on SPRT fordeveloping the SPOT spam zombie detection system. InSection 5, we provide the detailed design of SPOT and thetwo other detection algorithms. Section 6 evaluates theSPOT detection system based on the two-month e-mailtrace, and contrasts its performance with the two otherdetection algorithms. We briefly discuss the practicaldeployment issues and potential evasion techniques inSection 7, and conclude the paper in Section 8.

2 RELATED WORK

In this section, we discuss related work in detectingcompromised machines. We first focus on the studies thatutilize spamming activities to detect bots and then brieflydiscuss a number of efforts in detecting general botnets.

Based on e-mail messages received at a large e-mailservice provider, two recent studies [25], [26] investigatedthe aggregate global characteristics of spamming botnetsincluding the size of botnets and the spamming patterns ofbotnets. These studies provided important insights into the

aggregate global characteristics of spamming botnets byclustering spam messages received at the provider intospam campaigns using embedded URLs and near-duplicatecontent clustering, respectively. However, their approachesare better suited for large e-mail service providers tounderstand the aggregate global characteristics of spam-ming botnets instead of being deployed by individualnetworks to detect internal compromised machines. More-over, their approaches cannot support the online detectionrequirement in the network environment considered in thispaper. We aim to develop a tool to assist system adminis-trators in automatically detecting compromised machines intheir networks in an online manner.

Xie et al. developed an effective tool named DBSpam todetect proxy-based spamming activities in a networkrelying on the packet symmetry property of such activities[23]. We intend to identify all types of compromisedmachines involved in spamming, not only the spam proxiesthat translate and forward upstream non-SMTP packets (forexample, HTTP) into SMTP commands to downstream mailservers as in [23].

In the following, we discuss a few schemes on detectinggeneral botnets. BotHunter [8], developed by Gu et al.,detects compromised machines by correlating the IDSdialog trace in a network. It was developed based on theobservation that a complete malware infection process has anumber of well-defined stages including inbound scanning,exploit usage, egg downloading, outbound bot coordinationdialog, and outbound attack propagation. By correlatinginbound intrusion alarms with outbound communicationspatterns, BotHunter can detect the potential infectedmachines in a network. Unlike BotHunter which relies onthe specifics of the malware infection process, SPOT focuseson the economic incentive behind many compromisedmachines and their involvement in spamming.

An anomaly-based detection system named BotSniffer [9]identifies botnets by exploring the spatial-temporal beha-vioral similarity commonly observed in botnets. It focuseson IRC-based and HTTP-based botnets. In BotSniffer, flowsare classified into groups based on the common server thatthey connect to. If the flows within a group exhibitbehavioral similarity, the corresponding hosts involved aredetected as being compromised. BotMiner [7] is one of thefirst botnet detection systems that are both protocol andstructure independent. In BotMiner, flows are classified intogroups based on similar communication patterns andsimilar malicious activity patterns, respectively. The inter-section of the two groups is considered to be compromisedmachines. Compared to general botnet detection systemssuch as BotHunter, BotSniffer, and BotMiner, SPOT is alightweight compromised machine detection scheme, byexploring the economic incentives for attackers to recruit thelarge number of compromised machines.

As a simple and powerful statistical method, SequentialProbability Ratio Test has been successfully applied in manyareas [22]. In the area of networking security, SPRT has beenused to detect portscan activities [12], proxy-based spam-ming activities [23], anomaly-based botnet detection [9], andMAC protocol misbehavior in wireless networks [16].

DUAN ET AL.: DETECTING SPAM ZOMBIES BY MONITORING OUTGOING MESSAGES 199

http://ieeexploreprojects.blogspot.com

Page 3: Detecting spam zombies by monitoring outgoing messages.bak

3 PROBLEM FORMULATION AND ASSUMPTIONS

In this section, we formulate the spam zombie detection

problem in a network. In particular, we discuss the network

model and assumptions we make in the detection problem.Fig. 1 illustrates the logical view of the network model.

We assume that messages originated from machines inside

the network will pass the deployed spam zombie detection

system. This assumption can be achieved in a few different

scenarios. For example, the outgoing e-mail traffic (with

destination port number of 25) can be replicated and

redirected to the spam zombie detection system.A machine in the network is assumed to be either

compromised or normal (that is, not compromised). In this

paper, we only focus on the compromised machines that are

involved in spamming. Therefore, we use the term a

compromised machine to denote a spam zombie, and use the

two terms interchangeably. Let Xi for i ¼ 1; 2; . . . denote the

successive observations of a random variable X correspond-

ing to the sequence of messages originated from machine m

inside the network. We let Xi ¼ 1 if message i from the

machine is a spam, and Xi ¼ 0 otherwise. The detection

system assumes that the behavior of a compromised

machine is different from that of a normal machine in

terms of the messages they send. Specifically, a compro-

mised machine will with a higher probability generate a

spam message than a normal machine. Formally,

PrðXi ¼ 1jH1Þ > PrðXi ¼ 1jH0Þ; ð1Þ

where H1 denotes that machine m is compromised and H0

that the machine is normal. The spam zombie detection

problem can be formally stated as follows: as Xi arrives

sequentially at the detection system, the system determines

with a high probability if machine m has been compro-

mised. Once a decision is reached, the detection system

reports the result, and further actions can be taken, e.g., to

clean the machine.We assume that a (content-based) spam filter is deployed

at the detection system so that an outgoing message can be

classified as either a spam or nonspam [20]. None of existing

spam filters can achieve perfect spam detection accuracy,

and they all suffer from both false positive and false negative

errors. The false negative rate of a spam filter measures

the percentage of spam messages that are misclassified, and

the false positive rate measures the percentage of nonspam

messages that are misclassified. We note that all widely

deployed spam filters have very low false negative and false

positive rates, and such small spam classification errors will

only have a marginal impact on the performance of the spamzombie detection algorithms (see Section 6).

We also assume that a sending machine m as observedby the spam zombie detection system is an end-user clientmachine. It is not a mail relay server. This assumption is justfor the convenience of our exposition. The proposed SPOTsystem can handle the case where an outgoing message isforwarded by a few internal mail relay servers beforeleaving the network. We discuss practical deploymentissues in Section 7.

In addition, we assume that an IP address corresponds toa unique machine and ignores the potential impacts ofdynamic IP addresses on the detection algorithms in thepresentation of the algorithms [3], [24]. We will investigatethe potential impacts of dynamic IP addresses on thedetection algorithms in Sections 5 and 6.

4 BACKGROUND ON SEQUENTIAL PROBABILITY

RATIO TEST

In this section, we provide the necessary background onthe Sequential Probability Ratio Test for understanding theproposed spam zombie detection system. Interested read-ers are directed to [21] for a detailed discussion on thetopic of SPRT.

In its simplest form, SPRT is a statistical method fortesting a simple null hypothesis against a single alternativehypothesis. Intuitively, SPRT can be considered as an one-dimensional random walk with two user-specified bound-aries corresponding to the two hypotheses. As the samplesof the concerned random variable arrive sequentially, thewalk moves either upward or downward one step,depending on the value of the observed sample. Whenthe walk hits or crosses either of the boundaries for the firsttime, the walk terminates and the corresponding hypoth-esis is selected.

As a simple and powerful statistical tool, SPRT has anumber of compelling and desirable features that lead tothe widespread applications of the technique in many areas[22]. First, both the actual false positive and false negativeprobabilities of SPRT can be bounded by the user-specifiederror rates. A smaller error rate tends to require a largernumber of observations before SPRT terminates. Thus,users can balance the performance (in terms of false positiveand false negative rates) and cost (in terms of number ofrequired observations) of an SPRT test. Second, it has beenproved that SPRT minimizes the average number of therequired observations for reaching a decision for a givenerror rate, among all sequential and nonsequential statis-tical tests. In the following, we present the formal definitionand a number of important properties of SPRT. The detailedderivations of the properties can be found in [21].

Let X denote a Bernoulli random variable underconsideration with an unknown parameter �, and X1;X2; . . . the successive observations on X. As discussedabove, SPRT is used for testing a simple hypothesis H0 that� ¼ �0 against a single alternative H1 that � ¼ �1. That is,

PrðXi ¼ 1jH0Þ ¼ 1� PrðXi ¼ 0jH0Þ ¼ �0

PrðXi ¼ 1jH1Þ ¼ 1� PrðXi ¼ 0jH1Þ ¼ �1:

200 IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, VOL. 9, NO. 2, MARCH/APRIL 2012

Fig. 1. Network model.

http://ieeexploreprojects.blogspot.com

Page 4: Detecting spam zombies by monitoring outgoing messages.bak

To ease exposition and practical computation, we computethe logarithm of the probability ratio instead of theprobability ratio in the description of SPRT. For any positiveinteger n ¼ 1; 2; . . . , define

�n ¼ lnPrðX1; X2; . . . ; XnjH1ÞPrðX1; X2; . . . ; XnjH0Þ

: ð2Þ

Assume that Xi’s are independent (and identically dis-tributed), we have

�n ¼ lnQn

1 PrðXijH1ÞQn1 PrðXijH0Þ

¼Xn

i¼1

lnPrðXijH1ÞPrðXijH0Þ

¼Xn

i¼1

Zi; ð3Þ

where Zi ¼ ln PrðXijH1ÞPrðXijH0Þ , which can be considered as the step

in the random walk represented by �. When the observa-

tion is 1 (Xi ¼ 1), the constant ln �1

�0is added to the

preceding value of �. When the observation is 0 (Xi ¼ 0),

the constant ln 1��1

1��0is added.

The Sequential Probability Ratio Test for testing H0

against H1 is then defined as follows: given two user-specified constants A and B where A < B, at each stage n ofthe Bernoulli experiment, the value of �n is computed as in(3), then

�n � A ¼) accept H0 and terminate test;

�n � B ¼) accept H1 and terminate test;

A < �n < B ¼) take an additional observation

and continue experiment:

ð4Þ

In the following, we describe a number of importantproperties of SPRT. If we consider H1 as a detection and H0

as a normality, an SPRT process may result in two types oferrors: false positive where H0 is true but SPRT accepts H1

and false negative where H1 is true but SPRT accepts H0.We let � and � denote the user-desired false positive andfalse negative probabilities, respectively. There exist somefundamental relations among �, �, A, and B [21],

A � ln �

1� � ; B � ln 1� ��

;

for most practical purposes, we can take the equality, that is,

A ¼ ln �

1� � ; B ¼ ln 1� ��

: ð5Þ

This will only slightly affect the actual error rates. Formally,let �0 and �0 represent the actual false positive rate and theactual false negative rate, respectively, and let A and B becomputed using (5), then the following relations hold:

�0 � �

1� � ; �0 � �

1� � ; ð6Þ

and

�0 þ �0 � �þ �: ð7Þ

Equations (6) and (7) provide important bounds for �0

and �0. In all practical applications, the desired falsepositive and false negative rates will be small, for example,in the range from 0.01 to 0.05. In these cases, �

1�� and �1��

very closely equal the desired � and �, respectively. In

addition, (7) specifies that the actual false positive rate andthe false negative rate cannot be both larger than thecorresponding desired error rate in a given experiment.Therefore, in all practical applications, we can compute theboundaries A and B using (5), given the user-specified falsepositive and false negative rates. This will provide at leastthe same protection against errors as if we use the precisevalues of A and B for a given pair of desired error rates.The precise values of A and B are hard to obtain.

Another important property of SPRT is the number ofobservations, N , required before SPRT reaches a decision.The following two equations approximate the averagenumber of observations required when H1 and H0 aretrue, respectively.

E½NjH1� ¼�ln �

1��þ ð1� �Þln1���

�1ln �1

�0þ ð1� �1Þln 1��1

1��0

; ð8Þ

E½N jH0� ¼ð1� �Þln �

1��þ �ln 1���

�1ln �1

�0þ ð1� �1Þln 1��1

1��0

: ð9Þ

From the above equations, we can see that the averagenumber of required observations when H1 or H0 is truedepends on four parameters: the desired false positive andnegative rates (� and �), and the distribution parameters �1

and �0 for hypotheses H1 and H0, respectively. We note thatSPRT does not require the precise knowledge of thedistribution parameters �1 and �0. As long as the truedistribution of the underlying random variable is suffi-ciently close to one of hypotheses compared to another (thatis, � is closer to either �1 or �0), SPRT will terminate with thebounded error rates. An imprecise knowledge of �1 and �0

will only affect the number of required observations forSPRT to reach a decision.

5 SPAM ZOMBIE DETECTION ALGORITHMS

In this section, we will develop three spam zombiedetection algorithms. The first one is SPOT, which utilizesthe Sequential Probability Ratio Test presented in the lastsection. We discuss the impacts of SPRT parameters onSPOT in the context of spam zombie detection. The othertwo spam zombie detection algorithms are developed basedon the number of spam messages and the percentage ofspam messages sent from an internal machine, respectively.

5.1 SPOT Detection Algorithm

SPOT is designed based on the statistical tool SPRT wediscussed in the last section. In the context of detecting spamzombies in SPOT, we consider H1 as a detection and H0 as anormality. That is, H1 is true if the concerned machine iscompromised, and H0 is true if it is not compromised. Inaddition, we let Xi ¼ 1 if the ith message from theconcerned machine in the network is a spam, and Xi ¼ 0otherwise. Recall that SPRT requires four configurableparameters from users, namely, the desired false positiveprobability �, the desired false negative probability �, theprobability that a message is a spam when H1 is true (�1),and the probability that a message is a spam whenH0 is true(�0). We discuss how users configure the values of the four

DUAN ET AL.: DETECTING SPAM ZOMBIES BY MONITORING OUTGOING MESSAGES 201

http://ieeexploreprojects.blogspot.com

Page 5: Detecting spam zombies by monitoring outgoing messages.bak

parameters after we present the SPOT algorithm. Based on

the user-specified values of � and �, the values of the two

boundaries A and B of SPRT are computed using (5).In the following, we describe the SPOT detection

algorithm. Algorithm 1 outlines the steps of the algorithm.

When an outgoing message arrives at the SPOT detection

system, the sending machine’s IP address is recorded, and

the message is classified as either spam or nonspam by the

(content-based) spam filter. For each observed IP address,

SPOT maintains the logarithm value of the corresponding

probability ratio �n, whose value is updated according to

(3) as message n arrives from the IP address (lines 6 to 12 in

Algorithm 1). Based on the relation between �n and A and

B, the algorithm determines if the corresponding machine is

compromised, normal, or a decision cannot be reached and

additional observations are needed (lines 13 to 21).

Algorithm 1. SPOT spam zombie detection system

1: An outgoing message arrives at SPOT

2: Get IP address of sending machine m

3: // all following parameters specific to machine m4: Let n be the message index

5: Let Xn ¼ 1 if message is spam, Xn ¼ 0 otherwise

6: if (Xn ¼¼ 1) then

7: // spam, 3

8: �nþ ¼ ln �1

�0

9: else

10: // nonspam

11: �nþ ¼ ln 1��1

1��0

12: end if

13: if (�n � B) then

14: Machine m is compromised. Test terminates for m.

15: else if (�n � A) then

16: Machine m is normal. Test is reset for m.

17: �n ¼ 0

18: Test continues with new observations

19: else

20: Test continues with an additional observation

21: end if

We note that in the context of spam zombie detection,

from the viewpoint of network monitoring, it is more

important to identify the machines that have been compro-

mised than the machines that are normal. After a machine is

identified as being compromised (lines 13 and 14), it is

added into the list of potentially compromised machinesthat system administrators can go after to clean. Themessage-sending behavior of the machine is also recordedshould further analysis be required. Before the machine iscleaned and removed from the list, the SPOT detectionsystem does not need to further monitor the message-sending behavior of the machine.

On the other hand, a machine that is currently normalmay get compromised at a later time. Therefore, we need tocontinuously monitor machines that are determined to benormal by SPOT. Once such a machine is identified bySPOT, the records of the machine in SPOT are reset, inparticular, the value of �n is set to zero, so that a newmonitoring phase starts for the machine (lines 15 to 18).

5.2 Parameters of SPOT Algorithm

SPOT requires four user-defined parameters: �, �, �1, and �0.In the following, we discuss how a user of the SPOTalgorithm configures these parameters, and how theseparameters may affect the performance of SPOT. Asdiscussed in the previous section, � and � are the desiredfalse positive and false negative rates. They are normallysmall values in the range from 0.01 to 0.05, which users ofSPOT can easily specify independent of the behaviors of thecompromised and normal machines in the network. As wehave shown in Section 4, the values of � and � will affect thecost of the SPOT algorithm, that is, the number ofobservations needed for the algorithm to reach a conclusion.In general, a smaller value of � and � will require a largernumber of observations for SPOT to reach a detection.

Ideally, �1 and �0 should indicate the true probability of amessage being spam from a compromised machine and anormal machine, respectively, which are hard to obtain. Apractical way to assign values to �1 and �0 is to use thedetection rate and the false positive rate of the spam filterdeployed together with the spam zombie detection system,respectively. Given that all the widely used spam filtershave a high detection rate and low false positive rate [20],values of �1 and �0 assigned in this way should be very closeto the true probabilities.

To get some intuitive understanding of the averagenumber of required observations for SPRT to reach adecision, Figs. 2a and 2b show the value of E½N jH1� as afunction of �0 and �1, respectively, for different desired falsepositive rates. In the figures, we set the false negative

202 IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, VOL. 9, NO. 2, MARCH/APRIL 2012

Fig. 2. Average number of required observations when H1 is true (� ¼ 0:01).

http://ieeexploreprojects.blogspot.com

Page 6: Detecting spam zombies by monitoring outgoing messages.bak

rate � ¼ 0:01. In Fig. 2a, we assume the probability of amessage being spam when H1 is true to be 0.9 (�1 ¼ 0:9).From the figure, we can see that it only takes a smallnumber of observations for SPRT to reach a decision. Forexample, when �0 ¼ 0:2, SPRT requires about three ob-servations to detect that the machine is compromised if thedesired false positive rate is 0.01. As the behavior of anormal machine gets closer to that of compromisedmachine, i.e., �0 increases, a slightly higher number ofobservations are required for SPRT to reach a detection.

In Fig. 2b, we assume the probability of a message beingspam from a normal machine to be 0.2 (�0 ¼ 0:2). From thefigure, we can see that it also only takes a small number ofobservations for SPRT to reach a decision. As the behaviorof a compromised machine gets closer to that of a normalmachine, i.e., �1 decreases, a higher number of observationsare required for SPRT to reach a detection.

From the figures, we can also see that, as the desired falsepositive rate decreases, SPRT needs a higher number ofobservations to reach a conclusion. The same observationapplies to the desired false negative rate. These observationsillustrate the trade-offs between the desired performance ofSPRT and the cost of the algorithm. In the above discussion,we only show the average number of required observationswhen H1 is true because we are more interested in thespeed of SPOT in detecting compromised machines. Thestudy on E½N jH0� shows a similar trend (not shown).

We note that the statistical tool SPRT assumes events (inour cases, outgoing messages) are independently andidentically distributed. However, it is well known thatspam messages belonging to the same campaign are likelygenerated using the same spam template and delivered inbatch; therefore, spam messages observed in time proxi-mity may not be independent with each other. This canadversely affect the performance of SPOT in detectingcompromised machines. As a consequence, both thedetection accuracy of SPOT and the number of observationsrequired to reach a detection decision can differ from thetheoretical bounds provided in the statistical tool SPRT. Westudy the performance of SPOT using a real-world e-mailtrace in Section 6.

5.3 Spam Count and Percentage-Based DetectionAlgorithms

For comparison, in this section, we present two differentalgorithms in detecting spam zombies, one based on thenumber of spam messages and another the percentage ofspam messages sent from an internal machine, respectively.For simplicity, we refer to them as the count-threshold (CT)detection algorithm and the percentage-threshold (PT)detection algorithm, respectively.

In CT, the time is partitioned into windows of fixed lengthT . A user-defined threshold parameter Cs specifies themaximum number of spam message that may be originatedfrom a normal machine in any time window. The systemmonitors the number of spam messages n originated from amachine in each window. If n > Cs, then the algorithmdeclares that the machine has been compromised.

Similarly, in the PT detection algorithm, the time ispartitioned into windows of fixed length T . PT monitorstwo e-mail sending properties of each internal machine in

each time window: one is the percentage of spam messagessent from a machine, another the total number of messages.Let N and n denote the total messages and spam messagesoriginated from a machine m within a time window,respectively, then PT declares machine m as beingcompromised if N � Ca and n

N > P , where Ca is theminimum number of messages that a machine must send,and P is the user-defined maximum spam percentage of anormal machine. The first condition is in place forpreventing high false positive rates when a machine onlygenerates a small number of messages. For example, in anextreme case, a machine may only send a single messageand it is a spam, which renders the machine to have a100 percent spam ratio. However, it does not make sense toclassify this machine as being compromised based on thissmall number of messages generated.

In the following, we briefly compare the two spamzombie detection algorithms CT and PT with the SPOTsystem. The three algorithms have the similar running timeand space complexities. They all need to maintain a recordfor each observed machine and update the correspondingrecord as messages arrive from the machine. However,unlike SPOT, which can provide a bounded false positiverate and false negative rate, and consequently, a confidencehow well SPOT works, the error rates of CT and PT cannotbe a priori specified.

In addition, choosing the proper values for the four user-defined parameters (�, �, �1, and �0) in SPOT is relativelystraightforward (see the related discussion in the previoussection). In contrast, selecting the “right” values for theparameters of CT and PT is much more challenging andtricky. The performance of the two algorithms is sensitive tothe parameters used in the algorithm. They require athorough understanding of the different behaviors of thecompromised and normal machines in the concernednetwork and a training based on the behavioral history ofthe two different types of machines in order for them towork reasonably well in the network. For example, it can bechallenging to select the “best” length of time windows inCT and PT to obtain the optimal false positive and falsenegative rates. We discuss how an attacker may try to evadeCT and PT (and SPOT) in Section 7.

5.4 Impact of Dynamic IP Addresses

In the above discussion of the spam zombie detectionalgorithms, we have for simplicity ignored the potentialimpact of dynamic IP addresses and assumed that anobserved IP corresponds to a unique machine. In thefollowing, we informally discuss how well the threealgorithms fair with dynamic IP addresses. We formallyevaluate the impacts of dynamic IP addresses on detectingspam zombies in the next section using a two-month e-mailtrace collected on a large US campus network.

SPOT can work extremely well in the environment ofdynamic IP addresses. To understand the reason we notethat SPOT can reach a decision with a small number ofobservations as illustrated in Fig. 2, which shows theaverage number of observations required for SPRT toterminate with a conclusion. In practice, we have notedthat three or four observations are sufficient for SPRT toreach a decision for the vast majority of cases (see the

DUAN ET AL.: DETECTING SPAM ZOMBIES BY MONITORING OUTGOING MESSAGES 203

http://ieeexploreprojects.blogspot.com

Page 7: Detecting spam zombies by monitoring outgoing messages.bak

performance evaluation of SPOT in the next section). If a

machine is compromised, it is likely that more than three or

four spam messages will be sent before the (unwitting) user

shutdowns the machine and the corresponding IP address

gets reassigned to a different machine. Therefore, dynamic

IP addresses will not have any significant impact on SPOT.Dynamic IP addresses can have a greater impact on the

other two detection algorithms CT and PT. First, both

require the continuous monitoring of the sending behavior

of a machine for at least a specified time window, which in

practice can be on the order of hours or days. Second, CT

also requires a relatively larger number of spam messages

to be observed from a machine before reaching a detection.

By properly selecting the values for the parameters of CT

and PT (for example, a shorter time window for machines

with dynamic IP addresses), they can also work reasonably

well in the environment of dynamic IP addresses.

6 PERFORMANCE EVALUATION

In this section, we evaluate the performance of the three

detection algorithms based on a two-month e-mail trace

collected on a large US campus network. We also study the

potential impact of dynamic IP addresses on detecting

spam zombies.

6.1 Overview of the E-Mail Trace and Methodology

The e-mail trace was collected at a mail relay server deployed

in the Florida State University (FSU) campus network

between 8/25/2005 and 10/24/2005, excluding 9/11/2005

(we do not have trace on this date). During the course of the

e-mail trace collection, the mail server relayed messages

destined for 53 subdomains in the FSU campus network. The

mail relay server ran SpamAssassin [20] to detect spam

messages. The e-mail trace contains the following informa-

tion for each incoming message: the local arrival time, the IP

address of the sending machine (i.e., the upstream mail

server that delivered the message to the FSU mail relay

server), and whether or not the message is spam. In addition,

if a message has a known virus/worm attachment, it was so

indicated in the trace by an antivirus software. The antivirus

software and SpamAssassin were two independent compo-

nents deployed on the mail relay server. Due to privacy

issues, we do not have access to the content of the messagesin the trace.

Ideally, we should have collected all the outgoingmessages in order to evaluate the performance of thedetection algorithms. However, due to logistical constraints,we were not able to collect all such messages. Instead, weidentified the messages in the e-mail trace that have beenforwarded or originated by the FSU internal machines, thatis, the messages forwarded or originated by an FSU internalmachine and destined to an FSU account. We refer to thisset of messages as the FSU e-mails and perform ourevaluation of the detection algorithms based on the FSUe-mails. We note the set of FSU e-mails does not contain allthe outgoing messages originated from inside FSU, and thecompromised machines identified by the detection algo-rithms based on the FSU e-mails will likely be a lowerbound on the true number of compromised machines insideFSU campus network.

An e-mail message in the trace is classified as either spamor nonspam by SpamAssassin [20] deployed in the FSU mailrelay server. For ease of exposition, we refer to the set of allmessages as the aggregate e-mails including both spam andnonspam. If a message has a known virus/worm attach-ment, we refer to such a message as an infected message. Werefer to an IP address of a sending machine as a spam-only IPaddress if only spam messages are received from the IPaddress. Similarly, we refer to an IP address as nonspam onlyand mixed if we only receive nonspam messages, or wereceive both spam and nonspam messages, respectively,from the IP address.

Table 1 shows a summary of the e-mail trace. As shownin the table, the trace contains more than 25 M e-mails, ofwhich more than 18 M, or about 73 percent, are spam.About half of the messages in the e-mail trace wereoriginated or forwarded by FSU internal machines, i.e.,contained in the set of FSU e-mails. Table 2 shows theclassifications of the observed IP addresses. As shown in thetable, during the course of the trace collection, we observedmore than 2 M IP addresses (2;461;114) of sendingmachines, of which more than 95 percent sent at least onespam message. During the same course, we observed 440FSU internal IP addresses.

Table 3 shows the classification of the observed IPaddresses that sent at least one message carrying a virus/worm attachment. We note that a higher proportion of FSUinternal IP addresses sent e-mails with a virus/worm

204 IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, VOL. 9, NO. 2, MARCH/APRIL 2012

TABLE 1Summary of the E-Mail Trace

TABLE 2Summary of Sending IP Addresses

TABLE 3Summary of IP Addresses Sending Virus/Worm

http://ieeexploreprojects.blogspot.com

Page 8: Detecting spam zombies by monitoring outgoing messages.bak

attachment than the overall IP addresses observed (alle-mails were destined to FSU accounts). This could becaused by a few factors. First, a (compromised) e-mailaccount in general maintains more e-mail addresses offriends in the same domain than other remote domains.Second, an (e-mail-propagated) virus/worm may adopt aspreading strategy concentrating more on local targets [2].More detailed analysis of the e-mail trace can be found in[5] and [6], including the daily message arrival patterns,and the behaviors of spammers at both the mail-server leveland the network level.

In order to study the potential impacts of dynamicIP addresses on the detection algorithms, we obtain thesubset of FSU IP addresses in the trace whose domainnames contain “wireless,” which normally have dynami-cally allocated IP addresses. For each of the IP addresses,we group the messages sent from the IP address intoclusters, where the messages in each cluster are likely to befrom the same machine (before the IP address is reassignedto a different machine). We group messages according tothe interarrival times between consecutive messages, asdiscussed below. Let mi for i ¼ 1; 2; . . . denote the messagessent from an IP address, and ti denote the time whenmessage i is received. Then, messages mi for i ¼ 1; 2; . . . ; kbelong to the same cluster if jti � ti�1j � T for i ¼ 2; 3; . . . ; k,and jtkþ1 � tkj > T , where T is a user-defined time interval.We repeat the same process to group other messages. Let mi

for i ¼ j; jþ 1; . . . ; k be the sequence of messages in acluster, arriving in that order. Then, jtk � tjj is referred to asthe duration of the cluster, and jtkþ1 � tkj is referred to as thetime interval between two clusters.

Fig. 3 illustrates the message clustering process. Theintuition is that, if two messages come closely in time froman IP address (within a time interval T ), it is unlikely thatthe IP address has been assigned to two different machineswithin the short time interval.

In the evaluation studies, we whitelist the known mailservers deployed on the FSU campus network, given thatthey are unlikely to be compromised. If a deployed mailserver forwards a large number of spam messages, it ismore likely that machines behind the mail server arecompromised. However, just based on the informationavailable in the e-mail trace, we cannot decide whichmachines are responsible for the large number of spammessages, and consequently, determine the compromisedmachines. Section 7 discusses how we can handle this casein practical deployment.

6.2 Performance of SPOT

In this section, we evaluate the performance of SPOT basedon the collected FSU e-mails. In all the studies, we set� ¼ 0:01, � ¼ 0:01, �1 ¼ 0:9, and �0 ¼ 0:2.

Table 4 shows the performance of the SPOT spamzombie detection system. As discussed above, there are

440 FSU internal IP addresses observed in the e-mail trace.SPOT identifies 132 of them to be associated withcompromised machines. In order to understand theperformance of SPOT in terms of the false positive andfalse negative rates, we rely on a number of ways to verify ifa machine is indeed compromised. First, we check if anymessage sent from an IP address carries a known virus/worm attachment. If this is the case, we say we have aconfirmation. Out of the 132 IP addresses identified bySPOT, we can confirm 110 of them to be compromised inthis way. For the remaining 22 IP addresses, we manuallyexamine the spam sending patterns from the IP addressesand the domain names of the corresponding machines. Ifthe fraction of the spam messages from an IP address ishigh (greater than 98 percent), we also claim that thecorresponding machine has been confirmed to be compro-mised. We can confirm 16 of them to be compromised inthis way. We note that the majority (62.5 percent) of theIP addresses confirmed by the spam percentage aredynamic IP addresses, which further indicates the like-lihood of the machines to be compromised.

For the remaining six IP addresses that we cannotconfirm by either of the above means, we have alsomanually examined their sending patterns. We note that,they have a relatively overall low percentage of spammessages over the two month of the collection period.However, they sent substantially more spam messagestoward the end of the collection period. This indicates thatthey may get compromised toward the end of our collectionperiod. However, we cannot independently confirm if thisis the case.

Evaluating the false negative rate of SPOT is a bit trickyby noting that SPOT focuses on the machines that arepotentially compromised, but not the machines that arenormal (see Section 5). In order to have some intuitiveunderstanding of the false negative rate of the SPOT system,we consider the machines that SPOT does not identify asbeing compromised at the end of the e-mail collectionperiod, but for which SPOT has reset the records (lines 15 to18 in Algorithm 1). That is, such machines have beenclaimed as being normal by SPOT (but have continuouslybeen monitored). We also obtain the list of IP addresses thathave sent at least a message with a virus/worm attachment.Seven of such IP addresses have been claimed as beingnormal, i.e., missed, by SPOT.

We emphasize that the infected messages are only usedto confirm if a machine is compromised in order to studythe performance of SPOT. Infected messages are not usedby SPOT itself. SPOT relies on the spam messages insteadof infected messages to detect if a machine has beencompromised to produce the results in Table 4. We makethis decision by noting that, it is against the interest of aprofessional spammer to send spam messages with avirus/worm attachment. Such messages are more likely to

DUAN ET AL.: DETECTING SPAM ZOMBIES BY MONITORING OUTGOING MESSAGES 205

Fig. 3. Illustration of message clustering.

TABLE 4Performance of SPOT

http://ieeexploreprojects.blogspot.com

Page 9: Detecting spam zombies by monitoring outgoing messages.bak

be detected by antivirus softwares, and hence deletedbefore reaching the intended recipients. This is confirmedby the low percentage of infected messages in the overalle-mail trace shown in Table 1. Infected messages are morelikely to be observed during the spam zombie recruitmentphase instead of spamming phase. Infected messages canbe easily incorporated into the SPOT system to improveits performance.

We note that both the actual false positive rate and thefalse negative rate are higher than the specified falsepositive rate and false negative rate, respectively. A numberof factors can contribute to this observation. As we havediscussed in Section 5, the IID assumption of the underlyingstatistical tool SPRT may not be satisfied in the spamzombie detection problem. In addition, the spam filter usedto classify messages did not have a perfect spam detectionaccuracy, and as a consequence, events may be misclassi-fied. Moreover, the FSU e-mails can only provide a ratherlimited view of the outgoing messages originated frominside FSU, which will also affect the performance of SPOTin detection spam zombie machines.

Fig. 4 shows the distributions of the number of actualobservations that SPOT takes to detect the compromisedmachines. As we can see from the figure, the vast majorityof compromised machines can be detected with a smallnumber of observations. For example, more than 80 percentof the compromised machines are detected by SPOT withonly three observations. All the compromised machines aredetected with no more than 11 observations. This indicatesthat, SPOT can quickly detect the compromised machines.We note that SPOT does not need compromised machinesto send spam messages at a high rate in order to detectthem. Here, “quick” detection does not mean a shortduration, but rather a small number of observations. Acompromised machine can send spam messages at a lowrate (which, though, works against the interest of spam-mers), but it can still be detected once enough observationsare obtained by SPOT.

6.3 Performance of CT and PT

In this section, we evaluate the performance of CT and PTand compare their performance with that of SPOT, usingthe same two-month e-mail trace collected on the FSUcampus network. Recall that CT is a detection algorithmbased on the number of spam messages originated orforwarded by an internal machine, and PT based on thepercentage of spam messages originated or forwarded by

an internal machine (see Section 5.3). For comparison, wealso include a simple spam zombie detection algorithm thatidentifies any machine sending at least a single spammessage as a compromised machine. We note that such asimple scheme may not be deployed due to a potential highfalse positive rate; however, it provides us with the insightsinto the performance gain we can obtain by employing amore sophisticated spam zombie detection system.

In this evaluation study, we set the length of timewindows to be 1 hour, that is, T ¼ 1 hour, for both CT andPT. For CT, we set the maximum number of spam messagesthat a normal machine can send within a time window to be30 (Cs ¼ 3), that is, when a machine sends more than30 spam messages within any time windows, CT concludesthat the machine is compromised. In PT, we set theminimum number of (spam and nonspam) messages withina time window to be 6 (Ca ¼ 6), and the maximumpercentage of spam messages within a time window to be50 percent (P ¼ 50%). That is, if more than 50 percent of allmessages sent from a machine are spam in any timewindow with at least six messages in the window, PT willconclude that the machine is compromised. We choose thevalues for the parameters of PT in this way so that it isrelatively comparable with SPOT. Recall that based on ourempirical study in the last section, the minimum number ofobservations needed by SPOT to reach a detection is 3(when � ¼ 0:01, � ¼ 0:01, �0 ¼ 0:2, and �1 ¼ 0:9).

Table 5 shows the performance of CT and PT, whichincludes the number of compromised IP addresses de-tected, confirmed, and missed. We use the same methods toconfirm a detection or identify a missed IP address as wehave done with the SPOT detection algorithm. From thetable we can see that, CT and PT have a worse performancethan SPOT. For example, CT only detects 81 IP addresses asbeing compromised. Among the 81 IP addresses, 79 can beconfirmed to be associated with compromised machines.However, CT missed detecting 53 IP addresses associatedwith compromised machines. The detection rate and falsenegative rate of CT is 59.8 and 40.2 percent, respectively,much worse than that of SPOT, which are 94.7 and5.3 percent, respectively. We also note that all thecompromised IP addresses detected (confirmed) using CTor PT are also detected (confirmed) using the SPOTdetection algorithm. That is, the IP addresses detected(confirmed) using CT and PT are a subset of compromisedIP addresses detected (confirmed) using the SPOT detectionalgorithm. The IP addresses associated with compromisedmachines that are missed by SPOT are also missed by CTand PT. We conclude that SPOT outperforms both CT andPT in terms of both detection rate and miss rate.

In the table, we also show the performance of the simplespam zombie detection algorithm. We first note that themethods to confirm a detection or identify a missed

206 IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, VOL. 9, NO. 2, MARCH/APRIL 2012

Fig. 4. Number of actual observations.

TABLE 5Performance of CT and PT

http://ieeexploreprojects.blogspot.com

Page 10: Detecting spam zombies by monitoring outgoing messages.bak

IP address are different from the ones used in SPOT, CT,and PT, due to the special property of the simple algorithm.For this simple algorithm, we confirm the detection of acompromised machine if the machine sent at least oneoutgoing message carrying a virus/worm attachment.Similarly, if a machine sent out at least one outgoingmessage carrying a virus/worm attachment, but the simplealgorithm fails to detect it as a compromised machine, thenwe have identified a missed IP address. In this process, wehave similarly excluded all the whitelisted mail-servermachines. As we have discussed early that, the antivirussoftware and SpamAssassin were two independent compo-nents deployed at the FSU mail relay server, and a smallnumber of messages carrying virus/worm attachmentswere not detected as spam by the spam filter. Due to thedifference in the methods of confirming a detection oridentifying a missed IP address, the four detection algo-rithms observe different number of confirmed and missedIP addresses.

From the table we can see that, the simple detectionalgorithm can detect more machines (210) as beingcompromised than SPOT, CT, and PT. It also has betterperformance than CT and PT in terms of both detection rate(89.7 percent) and false negative rate (10.3 percent). Thesimple algorithm has worse performance than the SPOTalgorithm (see Table 4). To a degree, this result is surprisingin that the simple detection algorithm can have betterperformance than both the CT and PT algorithms. How-ever, we caution that this observation could be caused bythe limitation of the FSU e-mails used in the study. Recallthat this e-mail trace only contains outgoing messagesdestined to internal FSU accounts, and as we havediscussed, a higher portion of FSU internal IP addressessent e-mails with a virus/worm attachment to FSU internalaccounts than the overall IP addresses observed, whichresults in a higher confirmation rate in the compromisedmachine detection. We expect the percentage of FSUmachines sending e-mails with a virus/worm attachmentwill drop if the e-mail trace contains all outgoing messages(instead of only the ones destined to internal FSU accounts),which will adversely affect the observed performance of thesimple detection algorithm.

6.4 Dynamic IP Addresses

In this section, we conduct studies to understand thepotential impacts of dynamic IP addresses on theperformance of the three detection algorithms. Given that

SPOT outperforms both CT and PT, our discussion willfocus on the impacts on SPOT; similar observations alsoapply to CT and PT.

In order to understand the potential impacts of dynamicIP addresses on the detection algorithms, we groupmessages from a dynamic IP address (with domain namescontaining “wireless”) into clusters with a time intervalthreshold of 30 minutes. Messages with a consecutiveinterarrival time no greater than 30 minutes are groupedinto the same cluster. Given the short interarrival durationof messages within a cluster, we consider all the messagesfrom the same IP address within each cluster as being sentfrom the same machine. That is, the correspondingIP address has not been reassigned to a different machinewithin the concerned cluster. (It is possible that messagesfrom multiple adjacent clusters are actually sent from thesame machine.)

Fig. 5 shows the cumulative distribution function (CDF)of the number of spam messages in each cluster. From thefigure, we can see that more than 90 percent of the clustershave no less than 10 spam messages, and more than96 percent no less than three spam messages. Given thelarge number of spam messages sent within each cluster, itis unlikely for SPOT to mistake one compromised machineas another when it tries to detect spam zombies. Indeed, wehave manually checked that, spam messages tend to be sentback to back in a batch fashion when a dynamic IP addressis observed in the trace. Fig. 6 shows the CDF of the numberof all messages (including both spam and nonspam) in eachcluster. Similar observations can be made to that in Fig. 5.

Fig. 7 shows the CDF of the durations of the clusters. Aswe can see from the figure, more than 75 and 58 percent ofthe clusters last no less than 30 minutes and 1 hour

DUAN ET AL.: DETECTING SPAM ZOMBIES BY MONITORING OUTGOING MESSAGES 207

Fig. 5. Distribution of spam messages in each cluster. Fig. 6. Distribution of total messages in each cluster.

Fig. 7. Distribution of the cluster duration.

http://ieeexploreprojects.blogspot.com

Page 11: Detecting spam zombies by monitoring outgoing messages.bak

(corresponding to the two vertical lines in the figure),respectively. The longest duration of a cluster we observe inthe trace is about 3.5 hours. Fig. 8 shows the CDF of the timeintervals between consecutive clusters. As we can see fromthe figure, the minimum time interval between twoconsecutive clusters is slightly more than 30 minutes(31.38 minutes), and the longest one is close to 13 days(18,649.38 minutes). Moreover, more than 88 percent of allintervals between clusters are longer than 1 hour.

Given the above observations, in particular, the largenumber of spam messages in each cluster, we conclude thatdynamic IP addresses will not have any important impact onthe performance of SPOT. SPOT can reach a decision withinthe vast majority (96 percent) of the clusters in the setting weused in the current performance study. It is unlikely forSPOT to mistake a compromised machine as another.

7 DISCUSSION

In this section, we discuss the practical deployment issuesand possible techniques that spammers may employ toevade the detection algorithms. Our discussions will focuson the SPOT detection algorithm.

7.1 Practical Deployment

To ease exposition, we have assumed that a sending machinem (Fig. 1) is an end-user client machine. It cannot be a mailrelay server deployed by the network. In practice, a networkmay have multiple subdomains and each has its own mailservers. A message may be forwarded by a number of mailrelay servers before leaving the network. SPOT can workwell in this kind of network environments. In the following,we outline two possible approaches. First, SPOT can bedeployed at the mail servers in each subdomain to monitorthe outgoing messages so as to detect the compromisedmachines in that subdomain.

Second, and possibly more practically, SPOT is onlydeployed at the designated mail servers, which forward alloutgoing messages (or SPOT gets a replicated stream of alloutgoing messages), as discussed in Section 3. SPOT relieson the Received header fields to identify the originatingmachine of a message in the network [13], [18]. Given thatthe Received header fields can be spoofed by spammers[19], SPOT should only use the Received header fieldsinserted by the known mail servers in the network. SPOTcan determine the reliable Received header fields bybacktracking from the last known mail server in the

network that forwards the message. It terminates andidentifies the originating machine when an IP address in theReceived header field is not associated with a known mailserver in the network. The similar practical deploymentmethods also apply to the CT and PT detection algorithms.

7.2 Possible Evasion Techniques

Given that the developed compromised machine detectionalgorithms rely on (content-based) spam filters to classifymessages into spam and nonspam, spammers may try toevade the detection algorithms by evading the deployedspam filters. They may send completely meaningless“nonspam” messages (as classified by spam filters). How-ever, this will reduce the real spamming rate, and hence, thefinancial gains, of the spammers [4]. More importantly, asshown in Fig. 2b, even if a spammer reduces the spampercentage to 50 percent, SPOT can still detect the spamzombie with a relatively small number of observations (25when � ¼ 0:01, � ¼ 0:01, and �0 ¼ 0:2). So, trying to sendnonspam messages will not help spammers to evade theSPOT system.

Moreover, in certain environment where user feedback isreliable, for example, feedback from users of the samenetwork in which SPOT is deployed, SPOT can rely onclassifications from end users (in addition to the spamfilter). Although completely meaningless messages mayevade the deployed spam filter, it is impossible for them toremain undetected by end users who receive such mes-sages. User feedbacks may be incorporated into SPOT toimprove the spam detection rate of the spam filter. As wehave discussed in the previous section, trying to send spamat a low rate will also not evade the SPOT system. SPOTrelies on the number of (spam) messages, not the sendingrate, to detect spam zombies.

The current performance evaluation of SPOT was carriedout using the FSU e-mails collected in the year of 2005.However, based on the above discussion, we expect thatSPOT will work equally well in today’s environment.Indeed, as long as the spam filer deployed together withSPOT can provide a reasonable spam detection rate, theSPOT system should be effective in identifying compro-mised machines. We also argue that a spam zombiedetection system such as SPOT is in an even greater need,given that a large percentage of spam messages wereoriginated from spamming bots in recent years. Variousstudies in recent years have shown that spam messages sentfrom botnets accounted for above 80 percent of all spammessages on the Internet [11], [15]. For example, theMessageLabs Intelligence 2010 annual security reportshowed that approximately 88.2 percent of all spam in2010 were sent from botnets.

As we have discussed in Section 5.3, selecting the “right”values for the parameters of CT and PT is much morechallenging and tricky than those of SPOT. In addition, theparameters directly control the detection decision of the twodetection algorithms. For example, in CT, we specify themaximum number of spam messages that a normalmachine can send. Once the parameters are learned by thespammers, they can send spam messages below theconfigured threshold parameters to evade the detectionalgorithms. One possible countermeasure is to configure the

208 IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, VOL. 9, NO. 2, MARCH/APRIL 2012

Fig. 8. Distribution of time intervals between clusters.

http://ieeexploreprojects.blogspot.com

Page 12: Detecting spam zombies by monitoring outgoing messages.bak

algorithms with small threshold values, which helps reducethe spam sending rate of spammers from compromisedmachines, and therefore, the financial gains of spammers.Spammers can also try to evade PT by sending meaningless“nonspam” messages. Similarly, user feedback can be usedto improve the spam detection rate of spam filters to defeatthis type of evasions.

8 CONCLUSION

In this paper, we developed an effective spam zombiedetection system named SPOT by monitoring outgoingmessages in a network. SPOT was designed based on asimple and powerful statistical tool named SequentialProbability Ratio Test to detect the compromised machinesthat are involved in the spamming activities. SPOT hasbounded false positive and false negative error rates. It alsominimizes the number of required observations to detect aspam zombie. Our evaluation studies based on a two-monthe-mail trace collected on the FSU campus network showedthat SPOT is an effective and efficient system in automati-cally detecting compromised machines in a network. Inaddition, we also showed that SPOT outperforms two otherdetection algorithms based on the number and percentage ofspam messages sent by an internal machine, respectively.

ACKNOWLEDGMENTS

The authors thank the anonymous reviewers of IEEEINFOCOM 2009 and the IEEE Transactions on Dependableand Secure Computing. Their insightful and constructivecomments helped improve both the technical details andpresentation of the paper. Zhenhai Duan and FernandoSanchez were supported in part by US National ScienceFoundation (NSF) Grant CNS-1041677. Yingfei Dong wassupported in part by NSF Grants CNS-1041739 and CNS-1018971. Any opinions, findings, and conclusions orrecommendations expressed in this paper are those of theauthors and do not necessarily reflect the views of the NSF.A preliminary version of this paper appeared in theProceedings of IEEE INFOCOM 2009 with the same title.This project was carried out while Peng Chen was agraduate student at the Florida State University. Thisproject was carried out while James Michael Barker waswith the Florida State University.

REFERENCES

[1] P. Bacher, T. Holz, M. Kotter, and G. Wicherski, “Know YourEnemy: Tracking Botnets,” http://www.honeynet.org/papers/bots, 2011.

[2] Z. Chen, C. Chen, and C. Ji, “Understanding Localized-ScanningWorms,” Proc. IEEE Int’l Performance, Computing, and Comm. Conf.(IPCCC ’07), 2007.

[3] R. Droms, “Dynamic Host Configuration Protocol,” IETF RFC2131, Mar. 1997.

[4] Z. Duan, Y. Dong, and K. Gopalan, “DMTP: Controlling Spamthrough Message Delivery Differentiation,” Computer Networks,vol. 51, pp. 2616-2630, July 2007.

[5] Z. Duan, K. Gopalan, and X. Yuan, “Behavioral Characteristics ofSpammers and Their Network Reachability Properties,” TechnicalReport TR-060602, Dept. of Computer Science, Florida State Univ.,June 2006.

[6] Z. Duan, K. Gopalan, and X. Yuan, “Behavioral Characteristics ofSpammers and Their Network Reachability Properties,” Proc.IEEE Int’l Conf. Comm. (ICC ’07), June 2007.

[7] G. Gu, R. Perdisci, J. Zhang, and W. Lee, “BotMiner: ClusteringAnalysis of Network Traffic for Protocol- and Structure-Indepen-dent Botnet Detection,” Proc. 17th USENIX Security Symp., July2008.

[8] G. Gu, P. Porras, V. Yegneswaran, M. Fong, and W. Lee,“BotHunter: Detecting Malware Infection through Ids-DrivenDialog Correlation,” Proc. 16th USENIX Security Symp., Aug. 2007.

[9] G. Gu, J. Zhang, and W. Lee, “BotSniffer: Detecting BotnetCommand and Control Channels in Network Traffic,” Proc. 15thAnn. Network and Distributed System Security Symp. (NDSS ’08),Feb. 2008.

[10] N. Ianelli and A. Hackworth, “Botnets as a Vehicle for OnlineCrime,” Proc. First Int’l Conf. Forensic Computer Science, 2006.

[11] J.P. John, A. Moshchuk, S.D. Gribble, and A. Krishnamurthy,“Studying Spamming Botnets Using Botlab,” Proc. Sixth Symp.Networked Systems Design and Implementation (NSDI ’09), Apr. 2009.

[12] J. Jung, V. Paxson, A. Berger, and H. Balakrishnan, “Fast PortscanDetection Using Sequential Hypothesis Testing,” Proc. IEEE Symp.Security and Privacy, May 2004.

[13] J. Klensin, “Simple Mail Transfer Protocol,” IETF RFC 2821, Apr.2001.

[14] J. Markoff, “Russian Gang Hijacking PCs in Vast Scheme,” The NewYork Times, http://www.nytimes.com/2008/08/06/technology/06hack.html, Aug. 2008.

[15] P. Wood et al., “MessageLabs Intelligence: 2010 Annual SecurityReport,” 2010.

[16] S. Radosavac, J.S. Baras, and I. Koutsopoulos, “A Framework forMAC Protocol Misbehavior Detection in Wireless Networks,”Proc. Fourth ACM Workshop Wireless Security, Sept. 2005.

[17] A. Ramachandran and N. Feamster, “Understanding the Net-work-Level Behavior of Spammers,” Proc. ACM SIGCOMM,pp. 291-302, Sept. 2006.

[18] P. Resnick, “Internet Message Format,” IETF RFC 2822, Apr. 2001.[19] F. Sanchez, Z. Duan, and Y. Dong, “Understanding Forgery

Properties of Spam Delivery Paths,” Proc. Seventh Ann. Collabora-tion, Electronic Messaging, Anti-Abuse and Spam Conf. (CEAS ’10),July 2010.

[20] SpamAssassin, “The Apache SpamAssassin Project,” http://spamassassin.apache.org, 2011.

[21] A. Wald, Sequential Analysis. John Wiley & Sons, 1947.[22] G.B. Wetherill and K.D. Glazebrook, Sequential Methods in

Statistics. Chapman and Hall, 1986.[23] M. Xie, H. Yin, and H. Wang, “An Effective Defense against Email

Spam Laundering,” Proc. ACM Conf. Computer and Comm. Security,Oct./Nov. 2006.

[24] Y. Xie, F. Xu, K. Achan, E. Gillum, M. Goldszmidt, and T. Wobber,“How Dynamic Are IP Addresses?” Proc. ACM SIGCOMM, Aug.2007.

[25] Y. Xie, F. Xu, K. Achan, R. Panigrahy, G. Hulten, and I. Osipkov,“Spamming Botnets: Signatures and Characteristics,” Proc. ACMSIGCOMM, Aug. 2008.

[26] L. Zhuang, J. Dunagan, D.R. Simon, H.J. Wang, I. Osipkov, G.Hulten, and J.D. Tygar, “Characterizing Botnets from Email SpamRecords,” Proc. First Usenix Workshop Large-Scale Exploits andEmergent Threats, Apr. 2008.

Zhenhai Duan received the BS degree fromShandong University, China, in 1994, the MSdegree from Beijing University, China, in 1997,and the PhD degree from the University ofMinnesota, in 2003, all in computer science. Hejoined the faculty of the Department of ComputerScience at the Florida State University in 2003,where he is now an associate professor. Hisresearch interests include computer networksand network security. He was a corecipient of

Best Paper Award of the 2002 IEEE International Conference onNetwork Protocols (ICNP), the 2006 IEEE International Conference onComputer Communications and Networks (ICCCN), and the IEEEGLOBECOM 2008. He has served as a TPC cochair for CEAS 2011, theIEEE GLOBECOM 2010 Next-Generation Networking Symposium, andthe IEEE ICCCN 2007 and 2008 Network Algorithms and PerformanceEvaluation Track. He is a member of the ACM and a senior member ofthe IEEE.

DUAN ET AL.: DETECTING SPAM ZOMBIES BY MONITORING OUTGOING MESSAGES 209

http://ieeexploreprojects.blogspot.com

Page 13: Detecting spam zombies by monitoring outgoing messages.bak

Peng Chen received the BEng degree inmanagement information systems from TianjinUniversity, China, in 1999, the MEng degree incomputer science from Beihang University,China, in 2006, and the MS degree incomputer science from the Florida State Uni-versity, in 2008. He is currently a softwareengineer at Juniper Networks, Sunnyvale,California. His research interests include com-puter networks, Internet routing protocols, andnetworking security.

Fernando Sanchez received the BS degreefrom the Universidad San Francisco de Quito,Ecuador, in 2002, and the MS degree fromFlorida State University in 2007, all in computerscience. Currently, he is working toward the PhDdegree in the Department of Computer Scienceat the Florida State University. His researchinterests include network security with interest innetworking protocols security and network ser-vices security.

Yingfei Dong received the BS and MS degreesin computer science from the Harbin Institute ofTechnology, P.R. China, in 1989 and 1992, thedoctor degree in engineering from TsinghuaUniversity in 1996, and the PhD degree incomputer and information science from theUniversity of Minnesota in 2003. He is currentlyan associate professor in the Department ofElectrical Engineering at the University of Hawaiiat Manoa. His current research mostly focuses

on computer networks, especially in network security, real-time net-works reliable communications, multimedia content delivery, Internetservices, and distributed systems. His work has been published in manyreferred journals and conferences. He has served as both organizer andprogram committee member for many IEEE/ACM/IFIP conferences. Hiscurrent research is supported by US National Science Foundation. He isa member of the IEEE.

Mary Stephenson received BS degrees in meteorology and inmathematics/computer science from the Florida State University in1983 and the MS degree in meteorology from the University ofOklahoma in 1987. She is currently an associate director in InformationTechnology Services at Florida State University, within the Infrastructureand Operations Services Group. She is currently responsible for theUniversity’s enterprise Unix and Linux systems, storage, backup, andvirtualization technologies and the University’s data center facility.

James Michael Barker received the BA degree in philosophy from theCollege of William and Mary and the MA and PhD degrees in philosophyfrom the Florida State University. He is currently employed at theUniversity of North Carolina at Chapel Hill (UNC Chapel Hill) as anassistant vice chancellor and chief technology officer. He is responsiblefor the Infrastructure and Operations Division as well as the Commu-nication Technologies Division. His areas of operational oversightinclude UNC Chapel Hill’s network, telephony, datacenters, messa-ging/communication services, operating systems hosting, identitymanagement, middleware services, storage systems, and workloadautomation. In this role, he is responsible for the core informationtechnology infrastructure services of UNC Chapel Hill’s “enterprise” aswell as building block services for campus units. He previously served asthe director of the University Computing Services at the Florida StateUniversity. His dominant research interests include Kant, the history ofmodern philosophy, and philosophy of technology.

. For more information on this or any other computing topic,please visit our Digital Library at www.computer.org/publications/dlib.

210 IEEE TRANSACTIONS ON DEPENDABLE AND SECURE COMPUTING, VOL. 9, NO. 2, MARCH/APRIL 2012

http://ieeexploreprojects.blogspot.com