Sie sind auf Seite 1von 5

e-ISSN (O): 2348-4470

Scientific Journal of Impact Factor (SJIF): 4.72


p-ISSN (P): 2348-6406

International Journal of Advance Engineering and Research


Development
Emerging Technologies in the Computer World, January -2017

KIDS: Keyed Anomaly Detection System

Rohan Bendale1, Kedar Gokhale2, Sanket Nikam3,Abhijeet Dhore4


(Department of Computer engineering, Aissmss Institute of Information Technology)

Abstract---Most anomaly detection systems take for machine learning rule to derive a model of normality is later
accustomed observe suspicious event. One or two of works directed throughout the last years have notable that such rule
is typically susceptible to deception, notably at intervals the kind of attacks fastidiously developed to evade detection.
Totally utterly totally different learning plans square measure planned to beat this weakness. One such framework is
Keyed IDS (KIDS), given at DIMVA "10. KIDS main thought is far an identical as results of the operating of some
science primitives, specifically to introduce a secret component (the key) into the theme therefore as that one or 2 of
operations unit unfeasible whereas not knowing it. In kids the learned model and so the inconsistency's computation
scores unit each key-dependent, a reality that apparently prevents degree offender from making dodging attacks. In this,
we tend to tend to tend to demonstrate that sick the secret's to a extremely simple as long as offender will work with kids
and procure feedback with reference to inquiring requests. We tend to gift wise attack for 2 totally utterly totally different
adversarial settings and demonstrate that sick the key desires entirely a bit quantity of queries, that shows that kids
doesn't meet the claimed security properties. We tend to tend to tend to finally come back to KIDS' central discovered
and supply heuristic arguments with reference to its quality and limitations.

Keywords--- Machine Learning rule, KIDS, Heuristic Arguments, IDS.

I. INTRODUCTION

Many computer security problems could also be primarily reduced to separating malicious from non-malicious
activities. This is, as an example, the case of spam filtering, intrusion detection, or the identification of dishonorable
behavior. But, in general, shaping in associate passing precise and computationally useful means that what is harmless
or what is offensive is often too advanced. to beat these difficulties, most solutions to Such
problems have traditionally adopted a machine-learning approach, notably through the use of classifiers
to automatically derive models of (good and/or bad) behavior that unit of measurement later used to acknowledge the
prevalence of likely dangerous events. Recent work has accurately determined that security problems disagree
from various application domains of machine learning in, at least, one basic feature: the presence of associate soul World
Health Organization can strategically play against the formula to accomplish his goals. thus as an example, one major
objective for the wrongdoer is to avoid detection. Evasion attacks exploit weaknesses among the underlying classifiers,
that unit of measurement sometimes unable to identify a malicious sample that has been handily modified thus on
look ancient. Samples of such attacks abound. As an example, spammers typically modify their emails in varied ways in
which to avoid detection, e.g., by modifying words that unit of measurement generally found in spam, or by along
with AN oversized vary of words that do not. Similarly, malware and different things of attack code could also
be strictly made-to-order thus on evade intrusion detection systems (IDS) whereas not compromising the usefulness of
the attack some detection schemes projected over the previous number of is not entirely new: Wang et al. introduced in
Anagram, another payload-based anomaly detection system that addresses the evasion draw back in quite similar
manner. We have a tendency to tend to tell apart here between a pair of broad classes of classifiers that use a key.
Inside the first cluster, That we have a tendency to tend to term irregular. Classifiers, the classifier is entirely public (or,
equivalently, is trained with public knowledge only). However, in detection mode some parameters (the key) area
unit indiscriminately chosen whenever associate instance should be classified, thus making unsure for the bad
person but the instance area unit about to be processed. Note that, throughout this case, the same instance is processed
otherwise whenever if the secret's indiscriminately chosen.

We emphasize that organization can also be applied at work time, although it's attending to exclusively be
sufficiently effective once used throughout testing, a minimum of as method as evasion attacks unit of measurement
concerned. kids belong to a second cluster that we tend to tend to call keyed classifiers. throughout this case, there is one
secret and years have tried to incorporate defenses against evasion attacks. One such system is keyed intrusion detection
system (KIDS) [12], introduced by Mrdovic and Drazenovic at DIMVA10. kids ar academic degree application-layer
network anomaly detection system that extracts kind of choices (words) from each payload. The system then builds a
model of normality based every on the frequency of discovered choices and their relative positions at intervals the
payload. KIDS persistent key that is used throughout associate quantity of some time, presumptively as a results
of slashing the key implies preparation the classifier. If Kerckhoffs principle is to be followed, it ought to be assumed
that the security of the theme depends solely on the secrecy of the key and additionally the
procedure accustomed generate it. Anagram is employed every as irregular and as a keyed classifier; counting on the
@IJAERD-2017, All rights Reserved 108
International Journal of Advance Engineering and Research Development (IJAERD)
E.T.C.W, January -2017, e-ISSN: 2348 - 4470, print-ISSN: 2348-6406.

variant used core decide to impede evasion attacks is to incorporate the notion of a key, this being a
secret element wont to verify but classification choices unit of measurement extracted from the payload.The
security argument here is simple: even though the coaching and testing algorithms unit of
measurement public, associate someone World Health Organization is not in possession of the key will not
apprehend exactly but letter of invite are processed and, consequently, will not be able to vogue attacks that thwart
detection. To be precise, KIDS set up of learning with a secret.

II. LITERATURE SURVEY

1)Can Machine Learning Be Secure?

AUTHORS: Marco Barreno Blaine Nelson Russell Sears Anthony

Machine learning systems offer unparalled flexibility in dealing with evolving input in a variety of applications, such as
intrusion detection systems and spam e-mail filtering. However,
machine learning algorithms themselves can be a target of attack by a malicious adversary. This paper provides a
framework for answering the question, Can machine learning be secure? Novel contributions of this paper include a
taxonomy of different types of attacks on machine learning techniques and systems, a variety of defenses against those
attacks, a discussion of ideas that are important to security for machine learning, an analytical model giving a lower
bound on attackers work function, and a list of open

2) The security of machine learning

AUTHORS: Marco Barreno Blaine Nelson Anthony D. Joseph

Machine learnings ability to rapidly evolve to changing and complex situations has helped it become a fundamental tool
for computer security. That adaptability is also a vulnerability: attackers can exploit machine learning systems. We
present a taxonomy identifying and analyzing attacks against machine learning systems. We show how these classes
influence the costs for the attacker and defender, and we give a formal structure defining their interaction. We use our
framework to survey and analyze the literature of attacks against machine learning systems. We also illustrate our
taxonomy by showing how it can guide attacks against SpamBayes, a popular statistical spam filter. Finally, we discuss
how our taxonomy suggests new lines of defenses.

3) Adversarial Pattern Classification Using Multiple Classifiers and Randomization


AUTHORS: Battista Biggio, Giorgio Fumera, and Fabio Roli

In many security applications a pattern recognition system faces an adversarial classification problem, in which an
intelligent, adaptive adversary modifies patterns to evade the classifier. Several strategies have been recently proposed to
make a classifier harder to evade, but they are based only on qualitative and intuitive arguments. In this work, we
consider a strategy consisting in hiding information about the classifier to the adversary through the introduction of some
randomness in the decision function. We focus on an implementation of this strategy in a multiple classifier system,
which is a classification architecture widely used in security applications. We provide a formal support to this strategy,
based on an analytical framework for adversarial classification problems recently proposed by other authors, and give an
experimental evaluation on a spam filtering task to illustrate our findings.

4) Support Vector Machine Under Adversarial Label Noise


AUTHORS:B. BIGGIO, B. NELSON, AND P. LASKOV.

In adversarial classification tasks like spam filtering and intrusion detection, malicious adversaries may manipulate data
to thwart the outcome of an automatic analysis. Thus, besides achieving good classification performances, machine
learning algorithms have to be robust against adversarial data manipulation to successfully operate in these tasks. While
support vector machines (SVMs) have shown to be a very successful approach in classification problems, their
effectiveness in adversarial classification tasks has not been extensively investigated yet. In this paper we present a
preliminary investigation of the robustness of SVMs against adversarial data manipulation. In particular, we assume that
the adversary has control over some training data, and aims to subvert the SVM learning process. Within this assumption,
we show that this is indeed possible, and propose a strategy to improve the robustness of SVMs to training data
manipulation based on a simple kernel matrix correction. Keywords: Support Vector Machines, Adversarial
Classification, Label Noise.
@IJAERD-2017, All rights Reserved 109
International Journal of Advance Engineering and Research Development (IJAERD)
E.T.C.W, January -2017, e-ISSN: 2348 - 4470, print-ISSN: 2348-6406.

5 )Polymorphic Blending Attacks,

AUTHORS: P. Fogla, M. Sharif, R. Perdisci, O. Kolesnikov, and W. Lee,

A very effective means to evade signature-based intrusion detection systems (IDS) is to employ polymorphic techniques
to generate attack instances that do not share a fixed signature. Anomaly-based intrusion detection systems provide good
defense because existing polymorphic techniques can make the attack instances look different from each other, but
cannot make them look like normal. In this paper we introduce a new class of polymorphic attacks, called polymorphic
blending attacks, that can effectively evade byte frequency-based network anomaly IDS by carefully matching the
statistics of the mutated attack instances to the normal profiles. The proposed polymorphic blending attacks can be
viewed as a subclass of the mimicry attacks. We take a systematic approach to the problem and formally describe the
algorithms and steps required to carry out such attacks. We not only show that such attacks are feasible but also analyze
the hardness of evasion under different circumstances. We present detailed techniques using PAYL, a byte frequency-
based anomaly IDS, as a case study and demonstrate that these attacks are indeed feasible. We also provide some insight
into possible countermeasures that can be used as defense.

III. PROPOSED SYSTEM

The heuristics planned throughout this paper for accuracy-constrained privacy-preserving access
management irrelevant at intervals the context of workload-aware anonymization. The anonymization for
continuous information publication has been studied in literature. throughout this paper the most target is on a static
relative table that is anonymized just the once. To exemplify our approach, role-based access management is assumed.
However, the thought of accuracy constraints for permissions is applied to any privacy-preserving security policy, e.g.,
discretionary access management.
3.1 Advantages of Planned System:
1. Accuracy-constrained privacy-preserving access.
2. It maintains datas in an exceedingly secure manner.

IV. SYSTEM ARCHITECTURE

Figure 1. Architecture Diagram of Proposed System

V. MATHEMATICAL MODEL

Let S is the Whole System Consists:


S = {U, NC, KD, KA, PA}.
1. U is the set of number users.
U= {U1, U2Un}.

@IJAERD-2017, All rights Reserved 110


International Journal of Advance Engineering and Research Development (IJAERD)
E.T.C.W, January -2017, e-ISSN: 2348 - 4470, print-ISSN: 2348-6406.

2. NC is the set node created by admin.


NC= {NC1, NC2.NCn}.

3. KD is set of key recovery attack.


KD= {KD1, KD2.KDn}.

4. KA is set of keyed anomaly detection.


KA= {KA1, KA2..KAn}.

5. PA is set of performance analysis


PA= {PA1, PA2..PAn}

Step 1: User or hacker request for data and get important information.
U= {U1, U2Un}.

Step 2: To recover information or key. We create node and use routing on it.
NC= {NC1, NC2..NCn}.
Step 3: Then key recovery attack apply on KIDS.
KD= {KD1, KD2.KDn}.
Step 4: After that key anomaly detection and adversarial model revisited
KD= {KD1, KD2.KDn}.
Step 5: Then performance analysis and result comparing is done.
PA= {PA1, PA2..PAn}

Output: We recover our key.

VI. CONCLUSION

In this paper we have analyzed the strength of kids against key-recovery attacks. In doing therefore, we have
tailored to the Anomaly detection context associate degree adversarial model borrowed from the connected field of
adversarial learning. we have bestowed Key-recovery attacks in step with a pair of adversarial settings, counting on the
feedback given by kids to searching queries. To the only of our information, our work is that the initial to demonstrate
key-recovery attacks on a keyed classifier. Amazingly, our attacks unit of measurement terribly economical, showing that
it's moderately simple for associate degree aggressor to recover the key in any of the two settings mentioned. Such a
deficiency of security would possibly reveal that schemes like kids were just not designed to forestall key-recovery
attacks. However, we have argued that resistance against such attacks is crucial to any classifier that creates an endeavor
to impede evasion by wishing on a secret piece of knowledge. Weve got provided discussion on this and various queries
among the hope of stimulating any analysis throughout this house.

The attacks here bestowed could be prevented by introducing style of surprising countermeasures to the system,
like limiting the utmost length of words and payloads, or still per se quantities as classification choices. We suspect,
however, that these variants ought to still be in danger of other attacks. Thus, our recommendation for future designs is to
base alternatives on durable principles rather than specific fixes. occurring the way facet kids, it remains to be seen
whether or not or not similar schemes unit of measurement secure against key recovery attacks. Our attacks (or variants
of them) unit of measurement targeted on keyed classifiers, which we tend to believe that they will not carry over
irregular classifiers.

We note that, in its gift kind, kids can't be merely irregular, as choosing a replacement key implies employment
the classifier yet again that's clearly impractical in real-world eventualities. among the case of Anagram, the authors
discuss one mode of operation where the secret's accustomed split the packet in varied elements therefore each of them is
checked against a special Bloom filter. This theme bears varied resemblances to kids and additionally the key might even
be recovered with attacks quite like those presented here. All identical, this wishes further investigation and might be
self-addressed in future work. Our focus throughout this work has been on convalescent the key through economical
procedures, demonstrating that the classification technique leaks information regarding it which will be leveraged by
Associate in Nursing aggressor. However, the last word goal is to evade the system, which we've got merely assumed
that knowing the secret's essential to craft Associate in nursing attack that evades detection or, at least, that significantly
facilitates the tactic. It remains to be seen whether or not or not a keyed classifier like kids are merely evaded whereas not
expressly convalescent the key. If the answer is among the affirmative, then the key does not guarantee resistance against
evasion.

@IJAERD-2017, All rights Reserved 111


International Journal of Advance Engineering and Research Development (IJAERD)
E.T.C.W, January -2017, e-ISSN: 2348 - 4470, print-ISSN: 2348-6406.

REFERENCES

[1] M. Barreno, B. Nelson, R. Sears, A.D. Joseph, and J.D. Tygar, Can Machine Learning be Secure? Proc. ACM
Symp. Information, Computer and Comm. Security (ASIACCS 06 ), pp. 16-25, 2006.

[2] M. Barreno, B. Nelson, A.D. Joseph, and J.D. Tygar, The Security of Machine Learning, Machine Learning,
vol. 81, no. 2, pp. 121- 148, 2010.

[3] B. Biggio, G. Fumera, and F. Roli, Adversarial Pattern Classification Using Multiple Classifiers and
Randomisation, Proc. IAPR Intl Workshop Structural, Syntactic, and Statistical Pattern Recognition, pp. 500-
509, 2008.

[4] B. Biggio, B. Nelson, and P. Laskov, Support Vector Machines Under Adversarial Label Noise, J. Machine
Learning Research, vol. 20, pp. 97-112, 2011.

[5] N. Dalvi, P. Domingos, Mausam, S. Sanghai, and D. Verma, Adversarial Classification, Proc. 10th ACM
SIGKDD Intl Conf. Knowledge Discovery and Data Mining (KDD 04), pp. 99-108, 2004.

[6] P. Fogla, M. Sharif, R. Perdisci, O. Kolesnikov, and W. Lee, Polymorphic Blending Attacks, Proc. 15th Conf.
USENIX Security Symp., 2006.

[7] C. Gates and C. Taylo, Challenging the Anomaly Detection Paradigm: A Provocative Discussion, Proc. New
Security Paradigms Workshop (NSPW), pp. 21-29, 2006.

[8] A. Kolcz and C.H. Teo, Feature Weighting for Improved Classifier Robustness, Proc. Sixth Conf. Email and
Anti-Spam (CEAS 09), 2009.

[9] O. Kolesnikov, D. Dagon, and W. Lee, Advanced Polymorphic Worms: Evading IDS by Blending in with
Normal Traffic, Proc. USENIX Security Symp., 2005.

[10] D. Lowd and C. Meek, Adversarial Learning, Proc. 11th ACM SIGKDD Intl Conf. Knowledge Discovery in
Data Mining (KDD 05), pp. 641-647, 2005.

[11] Metasploit Framework, www.metasploit.com, 2013.

[12] S. Mrdovic and B. Drazenovic, KIDS-Keyed Intrusion Detection System, Proc. Seventh Intl Conf. Detection
of Intrusions and Malware, and Vulnerability Assessment (DIMVA 10), pp. 173-182, 2010.

[13] B. Nelson, B.I.P. Rubinstein, L. Huang, A.D. Joseph, and J.D. Tygar, Classifier Evasion: Models and Open
Problems, Proc. Intl ECML/PKDD Conf. Privacy and Security Issues in Data Mining and Machine Learning
(PSDML 10), pp. 92-98, 2011.

[14] B. Nelson, A.D. Joseph, S.J. Lee, and S. Rao, Near-Optimal Evasion of Convex-Inducing Classifiers, J.
Machine Learning Research, vol. 9, pp. 549-556, 2010.

@IJAERD-2017, All rights Reserved 112

Das könnte Ihnen auch gefallen