Open access peer-reviewed chapter

Modeling Spammer Behavior: Artificial Neural Network vs. Naïve Bayesian Classifier

By Saiful Islam and Rafiqul Islam

Submitted: June 24th 2010Reviewed: October 4th 2010Published: April 11th 2011

DOI: 10.5772/16002

Downloaded: 2233

1. Introduction

The exponential growth of spam emails in recent years is a fact of life. Internet subscribers world-wide are unwittingly paying an estimated €10 billion a year in connection costs just to receive “junk” emails, according to a study undertaken for the European Commission. Though there is no universal definition of spam, unwanted and unsolicited commercial email as a mass mailing to a large number of recipients is basically known as the junk email or spam to the internet community. Spams are considered to be potential threat to Internet Security. Spam's direct effects include the consumption of computer and network resources and the cost in human time and attention of dismissing unwanted messages. More importantly, these ever increasing spams are taking various forms and finding home not only in mailboxes but also in newsgroups, discussion forums etc without the consent of the recipients. Overflowing mailboxes are overwhelming users, causing newsgroups and discussion forums to be flooded with irrelevant or inappropriate messages. As a consequence, users are getting discouraged not to use them anymore though these systems can provide numerous benefits to them.

Combating spam is a difficult job contrast to the spamming. Millions of spammers around the world are engaged in spreading spams with ever changing tricks and tactis to circumvent the filters deployed by the mailbox providers. As spammers are paid per volume for thier job, they invest thier best effort in reaching everyone by all possible ways. No antispamming technique is hundred percent accurate for spam problem. Antispamming techniques try to make a trade-off between rejecting legitimate e-mail vs. not rejecting all spam, and the associated costs in time and effort.

The simplest and most common approaches are to use filters that screen messages based upon the presence of common words or phrases common to junk e-mail. Other simplistic approaches include blacklistingand whitelisting.

  • Blacklistingtechnique automatically rejects messages received from the addresses of known spammers.

  • Whitelistingaccepts messages received from known and trusted correspondents only.

The major flaw in the first two approaches is that it relies upon complacence by the spammers by assuming that they are not likely to change (or forge) their identities or to alter the style and vocabularyof their sales pitches. Whitelistingrisks the possibility that the recipient will miss legitimate e-mail from a known or expected correspondent with a heretofore unknown address, such as correspondence from a long-lost friend, or a purchase confirmation pertaining to a transaction with an online retailer. A detail explanation of these techniques is given in (Islam & Chowdhury, 2005). Ramachandran et al. (2007) propose a new technique called behavioral blacklisting, which complements existing blacklists by categorizing spammers based on howthey send email, rather than the IP address (or address range) from which they are sending it. The intuition behind their idea is that, while IP addresses are ephemeral as identifiers, spam campaigns, spam lists, and spamming techniques are more persistent. If one can identify email-sending patterns that are characteristic of spamming behavior, then she can continue to classify IP addresses as spammers even as spammers change their IP addresses.

Machine learning algorithms namely Naïve Bayesian classifier, Decision Tree induction, Artificial Neural Network and Support Vector Machines, based on keywords or tokens extracted from the e-mail’s Subject, Content-Type Header and Message Body, have been used successfully in the past (Aery & Chakravarthy, 2005; Drucker et al., 1999; Eichler, 2005; Islam & Chowdhury, 2005). Very soon they fall short to filter out spam emails as the spammer changing themselves in the ways that are very difficult to model by simple keywords or tokens (Stuart et al., 2004). The tactics the spammer uses follow patterns and these behavioral patterns can be modeled to combat spam. Actually the more they try to hide, the easier it is to see them (Stuart et al., 2004). Now the question is:

  1. Ques 1. Are the patterns that the spammers follow common to all?

  2. Ques 2. If the spammers follow patterns to spread spams, is it possible to track those patterns?

  3. Ques 3. If one can track the common spammer patterns, is it possible to model them?

  4. Ques 4. Is it possible to model common spammer patterns by machine learning approaches?

  5. Ques 5. What level of accuracy is possible to achieve if one apply machine learning approaches?

Many researchers observe that spammers follow patterns. These patterns can be discovered from many different places: from email corpus (Stuart et al., 2004) by analysing their contents, network-level behavioral patterns (Ramachandran et al., 2006; Sperotto et al., 2009), transfer pattern during transfer sessions (Zhang et al., 2006), resource usage patterns (Xu et al., 2010) and spammer behavior in terms of the chain of machines they use to deliver their messages (Guerra et al., 2009).

This study investigates the possibilities of modeling spammer behavioral patterns instead of vocabulary as features for spam email categorization and these behavioral patterns are discovered by analysing email corpus Subject, Content-Type Headerand Message Body. The two machine learning algorithms Naïve Bayesian Classifierand Artificial Neural Networksare experimented to model common spammer patterns and both of them achieve a promising detection rate that can be considered as an improvement of performance compared to the keyword-based contemporary filtering approaches.


2. Methodology

The success of machine learning algorithms in text categorization(TC) has led researchers to investigate learning algorithms for filtering spam emails (Sebastiani, 2002). The central purpose of learning is to accurately predict unseen data. There are two types of learning strategy and those are: supervisedand unsupervised. In supervised learningboth the data objects and their labels are given, but in unsupervised learning only data objects are provided. Naïve Bayesian classifier, Decision Tree induction, Artificial Neural Networkand Support Vector Machinesare supervised learningalgorithms. Clustering is a good example of unsupervised learningalgorithm.

To learn the machine both training and test data are needed. Training data are used to build the model and then test data are used to measure the effectiveness of the model. In n-fold cross validationtechnique initial data are divided into nsubsets. From these nsubsets, n-1 subsets are used to train the model and the remaining subset is used to test the model. This process continues ntimes and average performance is taken to judge the effectiveness of the model.

The following two supervisedmachine learning algorithms are exploited to model spammer tricks and techniques in this study.

2.1 Artificial Neural Network

Artificial neural networks(ANN) are non-linear statistical data modeling tools that tries to simulate the functions of biological neural networks. It consists of interconnected collection of simple processing elements or artificial neurons and processes information in a connectionist approach to computation (Han & Kamber, 2001; Stuart et al., 2004). ANN is generally considered to be an adaptive system that changes its structure in response to external or internal information that flows through the network during the learning phase. Fig. 2 shows an example of multilayer feed forward neural network(MFFNN).

Fig. 1.

Learning model

Fig. 2

An example of a multilayer feed-forward artificial neural network

In MFFNN nodes represent neurons are denoted by circle, edges represent connections between neurons are directed edges and labelled with corresponding weight. Nodes in the input layerreceive inputs (xi) from external world and nodes in the output layergive output (yi) to the user. Nodes in the middle layers(generally called hidden layers) receive inputs from the previous layer and deliver it to the next layer. The output of a node in an artificial neural networkis computed by the equation given below:


where yj is the value that will be passed to the next layer from node j, nis the number of incoming edges to node j, xi’s are the inputs coming from previous layer to node jand θ jis the bias for node j.

The topologyof an artificial neural network is defined by the number of layers in it and the number of nodes that appear in each layer. Once the topology is given of an artificial neural network, the learning task is to assign proper weight in each connection or edge so that it can correctly identify unknown data. This can be done by the most popular method for learning in multilayer networks: backpropagation. In backpropagationneural network learning algorithm, initially the weights and the biases are assigned randomly in the range [0, 1]. Then one of the example cases is applied to the network and the network produces some output based on the current state of its synaptic weights. This output is compared to the known-good output and a mean-squared error is calculated. The error value is then propagated backwards through the network and changes are made to the weights in each layer. The whole process is repeated for each of the example cases, then back to the first case again, and so on. The cycle is repeated until the overall error value drops below some pre-determined threshold. After then the network assumed to learn the problem well enough. The pseudo code for backpropagation artificial neural network learning algorithm is given below:

  1. Step 1: Randomly initialize the weights and the biases in the network

  2. Step 2: For each example ein the training set

    1. // forward pass

    2. = neural-net-output (network, e)

    3. T= teacher output for e

    4. Calculate error (T- O) at the output units

    5. // backward pass

    6. Compute delta_wh for all weights from hidden layer to output layer

    7. Compute delta_wi for all weights from input layer to hidden layer

    8. Update the weights in the network

  3. Step 3: Continue step 3 until all examples classified correctly or stopping criterion satisfied

  4. Step 4: Return the network

The detail about backpropagationlearning can be found in (Han & Kamber, 2001; Rojas, 1996).

2.2 Naïve Bayesian classifier

Bayesian classifier, the simplest and most widely used for filtering spams, is based on the so-called Bayes’ theorem. For a training e-mail E, the classifier calculates for each category, the probability that the e-mail should be classified under Ci, where Ciis the ithcategory, making use of the law of the conditional probability:


where P(Ci) is the prior probability of hypothesis Ci; P(E) is the prior probability of training data E; P(Ci| E) is the probability of Cigiven Eand P(E| Ci) is the probability of Egiven Ci. Assuming class conditional independence, that is, the probability of each word in an e-mail is independent of the word’s context and its position in the e-mail, P(E| Ci) can be calculated as the product of each individual word Wj’s probabilities appearing in the category Ci(Wjbeing the jthof lwords in the e-mail):


The category maximizing P(Ci| E) is predicted by the classifier (Han & Kamber, 2001; Eichler, 2005).

Bayesian classifier does not require having lots of observations for each possible combination of the variables and are particularly suitable when the dimensionality of input data is high. The effect of a variable value on a given class in Bayesian Classifier is independent of the values of other variable. This assumption simplifies the computation and despite its simplicity, Naive Bayesian Classifiercan often outperform more sophisticated classification methods (Caruana & Niculescu-Mizil, 2006) makes it particularly popular in commercial and open-source spam filters (Metsis et al., 2006).

2.3 Spammer behavioral patterns

The keyword-based statistical analyzers mostly depend on tokenization of the email content and extracting feature from tokenized keywords to model spammer behavior. Tokenization can be misguided in many several ways as today’s email supports character sets other than ASCII, non-text attachments and bodies with multiple parts. For example, the following HTML tricks can be used to do this:

GET<!-- banana -->V<!-- 45-->I<!-- wumpus -->A<!-- dskfj -->G <!-- adf -- >R<!-- free -->A

Thus above nonsense HTML tags only split the special word “viagra” and disguise the tokenizer though it would be shown as “GET VIAGRA” to email client.

Even a word can be replaced with characters of other languages or like same character. For example, “V1DEO” can be sent instead of “VIDEO” and “Fántástìç” instead of “Fantastic”. A combination of special characters can used to produce alphabetical characters. For example, char “V” can be represented as the combination of right slash “\” and left slash “/”. A grouping or clustering of these techniques is given Table 1 for quick review.

Java ScriptImageCSSHTMLMIME/Others
Title CaseY
Sticky FingerY
Readable SpellY
Dot MatrixYY
HTML NumbersY
Invisible InkYY
Encoding of MSGY
Encrypted Message BodiesY
Invisible ImageY
Zero ImageYYY
Slice and DiceYYY
Cross WordYY
Honorary TitleY
Image ChoppingY
Big TagY
Fake TextY
Slick ClickY
False ClickY
Pump & DumpY
I’m Feeling LuckyY

Table 1.

Common spammer tricks

Table 1 has 30 different tricks and one can easily verify that HTML based tactics cover most of them (70%). It can also be shown that 75% of Cascading Style Sheet (CSS) and 50% of Image-based tricks are also covered by HTML-based tactics. It is evident from table 1 that Java Script and MIME (and/or others) based tricks do not overlap with HTML/CSS based tactics.

In this study, a model has been developed exploiting two machine learning algorithms to capture common spammer patterns instead of keyword analysis. The 21 handy crafted features from each e-mail message extracted from subject header, priority & content-type headers and body shown in Table 2 simulate all possible common spammer tricks. These features have also been optimized in their capability of classifying spam emails. The rationale of these features can be verified by their statistics both in spam and non-spam emails. For example, whether a content-type header appeared within the message headers or whether the content type had been set to “text/html” is a common feature of spam, as our investigation revealed. The corpus that has been used in our experimentation, we observed that 98% spam emails include this feature. Similarly, color element (both CSS and HTML format) is also a frequent feature of spam emails. Colorful images those are generally included in the email for X-rated and unwanted internet marketing groups send to catch users’ attention. The use of color elements in non-spam mails is very low. We found that 56% spam emails contain color elements whereas it exists only for 10% non-spam emails. The inclusion of this feature in our classification has improved performance considerably, which shows its practicality. We also added feature 19-21 as in Table 2, which are significant features of recent spams.

FeatureCategory 1: Features From the Message Subject Header
1Binary feature indicating 3 or more repeated characters
2Number of words with all letters in uppercase
3Number of words with at least 15 characters
4Number of words with at least two of letters J, K, Q, X, Z
5Number of words with no vowels
6Number of words with non-English characters, special characters
such as punctuation, or digits at beginning or middle of word
Category 2: Features From the Priority and Content-Type Headers
7Binary feature indicating whether the priority had been set to any
level besides normal or medium
8Binary feature indicating whether a content-type header appeared
within the message headers or whether the content type had been set to “text/html”
Category 3: Features From the Message Body
9Proportion of alphabetic words with no vowels and at least 7 characters
10Proportion of alphabetic words with at least two of letters J, K, Q, X, Z
11Proportion of alphabetic words at least 15 characters long
12Binary feature indicating whether the strings “From:” and “To:” were both present
13Number of HTML opening comment tags
14Number of hyperlinks (“href=“)
15Number of clickable images represented in HTML
16Binary feature indicating whether a text color was set to white
17Number of URLs in hyperlinks with digits or “&”, “%”, or “@”
18Number of color element (both CSS and HTML format)
19Binary feature indicating whether JavaScript has been used or not
20Binary feature indicating whether CSS has been used or not
21Binary feature indicating opening tag of table

Table 2.

Features extracted from each e-mail

2.4 Email corpus

Classification based spam filtering systems have two major drawbacks. Firstly, building a perfect data set free from noise or imperfection as noise adversely affect the classifier’s performance (Islam et al., 2009). The nature of spam email is very dynamic and the content of email is textually misleading due to obfuscation. This remains a continuous challenge for spam filtering techniques. Secondly, most training models of the classifier have limitations on their operations (Ranawana & Palade, 2006). Classifiers often produce uncorrelated training errors due to the dimension of feature space; a dissimilar output space is generated for changing feature space from small dimension to complex high dimension.

In this work a corpus of 1,000 emails received over a period of several months is used for experimentation. The distribution of both spam and non-spam emails in this collection is equal. The equal distribution is preferred to make the classifier to eliminate the biasness towards a particular category. That is, out of 1,000 emails 500 is spam and 500 is non-spam. The collection of this corpus is selected over a time and latest trend in spamming is kept in mind. Also the author’s experience with spam research and statistical selection methodology is applied to the selection, which made this email bank very much representative of current spamming.

2.5 Feature construction

Each email is parsed as text file to identify each header element to distinguish them from the body of the message. Every substring within the subject header and the message body that was delimited by white space was considered to be a token, and an alphabetic word was defined as a token delimited by white space that contains only English alphabetic characters (A-Z, a-z)or apostrophes.

The tokens were evaluated to create a set of 21 hand-crafted features from each e-mail message (Table 2) of which features 1-17 are proposed in (Stuart et al., 2004). In addition of these 17 features this study proposes other four features 18-21. The study investigates the suitability of these 21 features in classifying spam emails. To do this, each email is represented by a vector of dimension 21 in the vector space model (Salton et al., 1975). To learn the neural network the input layer of Multilayer Perceptronwill have 21 nodes. These nodes receive values from each dimension of the vector representing the email. For Naïve Bayesian Classifier, each dimension corresponds to spammer behavioral pattern instead of keyword.

2.6 Evaluation measures

Estimating classifier accuracy is important since it allows one to evaluate how accurately a given classifier will classify unknown samples on which the classifier has not been trained. The effectiveness of a classifier is usually measured in terms of accuracy, precision and recall (Makhoul et al., 1999). These measures are calculated using the confusion matrix given below:

Category CiCorrect
Predicted ↓YESNOTP=true positives
FP=false positives
FN=false negatives
TN=true negatives

Table 3.

Confusion matrix

Accuracy of a classifier is calculated by dividing the number of correctly classified samples by the total number of test samples and is defined as:


Precision measures the system’s ability to present only relevant items while recall measures system’s ability to present all relevant items. These two measures are widely used in TREC evaluation of document retrieval (Makhoul, 1999). Precision is calculated by dividing the number of samples that are true positives by the total number of samples classified as positives and is defined as:


Analogously, recall is calculated by dividing the number of samples that are true positives by the total number of samples that classifier should classified as positives and is defined as:


3. Experimental results

Table 4 summarizes the comparative results of Naïve Bayesian Classifier and Artificial neural Networks. These algorithms are tested on Weka 3.6.0 suite of machine learning software written in Java, developed at the University of Waikato (Holmes et al., 1994). Before simulating on Weka Toolwe transform the vectors obtained from our dataset into Attribute-Relation File Format(ARFF).

FeaturesNaïve Bayesian Claaivessifier (NaïveBayes)ANN (Multilayer Perceptron)
Category 1 Only56.5 %55.7%56.5%67.8%73.1%67.8%
Category 2 Only65.2%75.0%65.2%65.2%75.0%65.2%
Category 3 Only88. 7 %88.7%88.7%86.1%86.1%86.1%
Category 1+Category 266.9 %67.3%67.0%73.1%77.2%73.0%
Category 2+ Category392.2 %92.2%92.2%87.8%88.1%87.8%
Category 1+Category 380.8 %80.9%80.9%74.7%75.4%74.8%
Category1+ Category 2 + Category 386.9 %87.0%87.0%84.3%85.1%84.3%

Table 4.

Comparison results for Naïve Bayesian classifier and Artificial Neural Network

ARFF files have two distinct sections. The first section is the Headerinformation, which is followed the Datainformation. The Headerof the ARFF file contains the name of the relation, a list of the attributes (the columns in the data), and their types. An example header for our dataset is given below:

@RELATION Email_Classification
@ATTRIBUTEfeature01{yes, no}
@ATTRIBUTEfeature07{normal, medium}
@ATTRIBUTEfeature08{yes, no}
@ATTRIBUTEfeature12{yes, no}
@ATTRIBUTEfeature16{yes, no}
@ATTRIBUTEfeature19{yes, no}
@ATTRIBUTEfeature20{yes, no}
@ATTRIBUTEfeature21{yes, no}
@ATTRIBUTEisHam{yes, no}

The last attribute is the class column. The Dataof the ARFF file looks like the following (only few samples out of 1000 are given here):

  1. no, 0, 0, 0, 1, 2, medium, yes, 2, 0, 118, yes, 1, 13, 0, no, 0, 0, no, yes, no, yes

  2. no, 8, 0, 0, 1, 2, medium, yes, 1, 219, 223, no, 0, 0, 0, no, 0, 0, no, no, no, yes

  3. no, 0, 1, 0, 0, 1, medium, yes, 5, 4, 48, no, 0, 4, 1, no, 2, 0, no, yes, no,yes

  4. no, 0, 0, 0, 0, 1, medium, yes, 3, 0, 144, yes, 1, 69, 0, no, 0, 3, no, yes, no, yes

  5. no, 0, 1, 1, 0, 1, medium, yes, 4, 0, 7, yes, 1, 1, 0, no, 0, 0, no, yes, no,yes

  6. no, 0, 0, 0, 1, 1, medium, yes, 0, 265, 271, no, 0, 0, 0, no, 0, 0, no, no, no, yes

  7. no, 0, 1, 0, 0, 1, medium, no, 0, 0, 0, no, 0, 0, 0, no, 0, 0, no, no, no,yes

  8. no, 0, 0, 0, 1, 0, medium, yes, 0, 0, 1, no, 0, 1, 0, no, 1, 1, no, yes, no,yes

  9. yes, 0, 0, 0, 3, 5, medium, yes, 0, 0, 2, no, 0, 0, 0, no, 0, 0, no, no, no,yes

  10. yes, 0, 0, 0, 3, 6, medium, no, 0, 0, 0, no, 0, 0, 0, no, 0, 0, no, no, no,yes

  11. no, 0, 0, 0, 0, 0, medium, yes, 10, 0, 55, no, 0, 11, 1, yes, 0, 2, no, yes, yes, no

  12. no, 0, 0, 0, 0, 1, medium, yes, 12, 0, 45, no, 0, 0, 1, no, 0, 15, no, yes, yes, no

  13. no, 0, 0, 0, 0, 0, medium, yes, 6, 0, 43, no, 0, 11, 1, no, 0, 0, no, yes, yes, no

  14. no, 0, 0, 0, 0, 1, medium, yes, 2, 0, 3, no, 0, 0, 0, no, 0, 0, no, no, no,no

  15. no, 0, 0, 0, 0, 0, medium, yes, 2, 0, 21, no, 0, 0, 0, no, 0, 2, no, yes, yes, no

  16. no, 0, 0, 0, 0, 1, medium, yes, 8, 0, 12, no, 0, 29, 0, no, 0, 0, no, no, no,no

  17. no, 0, 0, 0, 0, 0, medium, yes, 14, 0, 57, no, 0, 16, 1, yes, 0, 2, no, yes, yes, no

  18. no, 0, 0, 0, 0, 1, medium, yes, 1, 0, 39, no, 0, 14, 1, no, 0, 0, no, yes, yes, no

  19. yes, 1, 0, 0, 4, 6, medium, yes, 2, 0, 17, no, 0, 2, 0, no, 0, 0, no, no, no, no

  20. yes, 4, 0, 0, 4, 5, medium, yes, 3, 0, 8, no, 0, 7, 0, no, 0, 0, no, no, no, no

The highest level of accuracy that can be achieved by Naïve Bayesian classifier is 92.2% (shown in Table 4) using features from category 2 and 3. The accuracy that can be achieved by any learning algorithms using features from category 1 is negligible. Features from category 2 and 3 contribute mostly in classifying spam emails from non-spam emails for all machine learning algorithm experimented in this study.

Highest number of features is always desirable only if their inclusion increase classifier’s accuracy significantly. Growing number of features not only hinders multidimensional indexing but also increases overall execution time. So, this study starves to find an optimal number of features that can be effectively used to learn a classifier without degrading the level of accuracy.

Applying best first forward attribute selection method the study gets only 10 features from category 2 and category 3 useful for classifying the spam and non-spam emails without sacrificing the accuracy as shown in Table 5. The set includes features 8, 9, 10, 12, 13, 14, 15, 16, 17, and 18 of which feature 18 is identified in this study. The Naïve Bayesian classifier again outperforms the Artificial Neural Networks. The optimal feature set obtained by applying best first forward attribute selection method for the features proposed in (Stuart et al., 2004) includes only features 8, 9, 10, 12, 13, 14, 15, 16 and 17, a total of 9 features. In this case ANN outperforms than Naïve Bayesian classification algorithm as shown in Table 5.

FeaturesNaïve Bayesian Claaivessifier (NaïveBayes)ANN (Multilayer Perceptron)
Best first: 8, 9, 10, 12, 13, 14, 15, 16, 17, and 18 [This study]92.2%92.2%92.2%90.4%90.6%90.4%
Best first: 8, 9, 10, 12, 13, 14, 15, 16, and 17 (Stuart et al., 2004)86.1 %87.4%86.1%91.3%91.4%91.3%

Table 5.

Comparison results for Naïve Bayesian classifier and Artificial Neural Network

The study presented in (Stuart et al., 2004) uses neural network for modeling spammer common patterns and achieved similar performance, but the limitation of neural network is its longer training time and inherent complexity of explaining its derivation (less comprehensibility). On the contrary, Bayesian Classifier has the advantage of incremental inclusion of features and beforehand calculation. Therefore, Naïve Bayes is suitable for adapting itself in modeling new spammer patterns.

4. Conclusion

This research studies the modeling of spammer behavior by Artificial Neural Networks and Naïve Bayesian Classifier algorithms for spam email classification. Based on examining different features and two different learning strategies, the following conclusions can be drawn from the study presented in this study:

  1. Lesson 1. Spammer behavior can be modeled using features extracted from Content-Type header and message Body only.

  2. Lesson 2. The contribution of features extracted from subject header in spam email detection is negligible or insignificant.

  3. Lesson 3. Naïve Bayesian classifier models the spammer behavior best than Artificial Neural Networks.

  4. Lesson 4. It is possible to get an optimal number of features that can be effectively

applied to learning algorithms to classify spam emails without sacrificing accuracy The preliminary result presented in this study seems promising in modeling spammer common behavioral patterns compared to similar research. The contribution of this study is threefold: it shows why keyword based spam email classifier may fail to model spammers’ altering tricks, common patterns adopted by spammers and the rationale of using these patterns against them to combat spam; suitability of modeling spammer common patterns using ANN and Naïve Bayes and finally, establishment of the four concluding remarks.

© 2011 The Author(s). Licensee IntechOpen. This chapter is distributed under the terms of the Creative Commons Attribution-NonCommercial-ShareAlike-3.0 License, which permits use, distribution and reproduction for non-commercial purposes, provided the original is properly cited and derivative works building on this content are distributed under the same license.

How to cite and reference

Link to this chapter Copy to clipboard

Cite this chapter Copy to clipboard

Saiful Islam and Rafiqul Islam (April 11th 2011). Modeling Spammer Behavior: Artificial Neural Network vs. Naïve Bayesian Classifier, Artificial Neural Networks - Application, Chi Leung Patrick Hui, IntechOpen, DOI: 10.5772/16002. Available from:

chapter statistics

2233total chapter downloads

More statistics for editors and authors

Login to your personal dashboard for more detailed statistics on your publications.

Access personal reporting

Related Content

This Book

Next chapter

Applying an Artificial Neural Network to Predicting Effort and Errors for Embedded Software Development Projects

By Kazunori Iwata, Toyoshiro Nakashima, Yoshiyuki Anan and Naohiro Ishii

Related Book

First chapter

Introduction to the Artificial Neural Networks

By Andrej Krenker, Janez Bešter and Andrej Kos

We are IntechOpen, the world's leading publisher of Open Access books. Built by scientists, for scientists. Our readership spans scientists, professors, researchers, librarians, and students, as well as business professionals. We share our knowledge and peer-reveiwed research papers with libraries, scientific and engineering societies, and also work with corporate R&D departments and government entities.

More About Us