Sie sind auf Seite 1von 8

International Journal of Computer Science Engineering and

Information Technology Research (IJCSEITR)


ISSN(P): 2249-6831; ISSN(E): 2249-7943
Vol. 6, Issue 5, Oct 2016, 85-92
TJPRC Pvt. Ltd.

DATA CLASSIFICATION USING MIN-MAX


NEURAL NETWORK WITH PRUNING
ALOK KUMAR MISHRA & ER. ARCHANA SINGH
Department of CS&IT, SHIATS, Allahabad, Uttar Pradesh, India
ABSTRACT
Classification is a standout among the most dynamic research and application territory in neural network.
Fuzzy Min-Max (FMM) is a neural network display utilized for arrangement reason. There are sure restrictions in
unique FMM neural system with respect to arrangement of hyperboxes. Enhanced Fuzzy Min-Max (EFMM) neural
system outcomes the restrictions with respect to hyperbox extension; cover test lead and hyperbox withdrawal. Because
of these new standards EFMM makes more hyperboxes than FMM. In this manner, organize intricacy of EFMM is
more than FMM. EFMM makes expansive number of hyperboxes which influences the system unpredictability. In this
way, in this paper, pruning of hyperboxes is done to minimize the hyperboxes and enhance the system multifaceted
nature of EFMM, yet the acknowledgment rate diminishes somewhat. The pruning of hyperboxes is done utilizing
forecast.
KEYWORDS: Classification, Fuzzy Min-Max Model, Hyperbox, Neural Network Learning, Pruning

Received: Sep 19, 2016; Accepted: Sep 30, 2016; Published: Oct 14, 2016; Paper Id.: IJCSEITROCT201610

INTRODUCTION

Original Article

certainty calculate. The certainty variable of each hyperbox depends on its recurrence of utilization and precision of

Artificial neural system (ANN) is a computational model that comprises of an interconnected gathering of
artificial neurons that reproduces the organic neural framework in our cerebrum [I]. Design order in current
situation is utilized for some building arrangements. Control, following, and expectation frameworks will
frequently utilize classifiers to decide input-yield connections. As of late, ANNs are utilized under various fields,
e.g., social insurance, control frameworks and blame discovery [II]. Design grouping is one of the dynamic ANN
application areas. For instance, ANN models have been effectively connected to grouping undertakings in business
and science and modern blame recognition and conclusion [III].

Figure 1: Showing the Layers in ANN

www.tjprc.org

editor@tjprc.org

86

Alok Kumar Mishra & Er. Archana Singh

As far as ANN learning, one of the principle issue identified with clump learning, is disastrous overlooking [IV].
Disastrous overlooking means the failure of learning framework to recollect, at whatever point some new data is given.
The disastrous overlooking is likewise named as dependability versatility difficulty. Solidness pliancy problem for the
most part worried with how a learning framework is sufficiently plastic to learn new data and in addition sufficiently stable
to hold the beforehand learned data. The issue of security versatility situation must be explained particularly when an ANN
needs to gain from information tests in one pass utilizing web learning.
The FMM neural system is one of the ANN models which conquer the steadiness versatility predicament. FMM
neural system utilizes hyperboxes for grouping. A hyperbox contains designs with full class enrollment. A hyperbox is
characterized by its min point and its maximum point, and an enrollment capacity is characterized as for these hyperbox
min-max focuses. Learning in FMM grouping is described by appropriately putting and modifying hyperboxes.
A Few Notable Learning Properties those are connected with FMM [IV].

Online Learning: The example classifier ought to learn new data without influencing the authentic data. Some
classifier in neural system utilizes disconnected learning as a part of which both the old information and new must
be retrained at whatever point some new information comes. Web learning is essential to understand the security
versatility predicament.

Nonlinear Separability: The example classifier ought to have the capacity to partition two distinct classes of any
shape and size. Non direct detachability is the capacity to fabricate a nonlinear choice limit to isolate particular
classes. There ought to be a limit between examples of two distinct classes.

Overlapping Classes: The capacity of the nonlinear choice limit to decrease misclassification by expelling the
covering areas of various classes. There ought not to be any covering of characterized areas of two unique classes.

Training Time: The capacity to learn and modify the nonlinear choice limit with one-go learning through the
preparation information inside a little preparing time. A portion of the calculations take vast number of goes to
learn through the information. The property of example arrangement is to take in the examples partitioned by
non-straight limits in brief time. In any case, when online adjustment is done some issue happens.

Soft and Hard Decisions: An example classifier must give delicate or hard choices. A hard choice or fresh choice
can be either 0 or 1, i.e. either the example has a place with the class or not. A delicate choice gives an esteem that
by how much degree the example will be of that class.
EFMM [IV] has been Created to Conquer the Confinements of FMM. Three Heuristic Standards in EFMM to

Defeat the Impediments Are:

To wipe out the issue of covering amid hyperbox extension, new covering rules has been recommended.

To find other covering cases the hyperbox test lead has been expanded.

To resolve the hyperbox covering cases, hyperbox withdrawal lead is given.


In any case, it has been watched that the system multifaceted nature of EFMM is more than FMM because of

development of expansive number of hyperboxes. Along these lines, to minimize the system intricacy of EFMM the idea
of pruning is presented in this paper. We do pruning of hyperboxes to expel the pointless hyperboxes containing less
Impact Factor (JCC): 7.1293

NAAS Rating: 3.63

Data Classification using Min-Max Neural Network with Pruning

87

exactness or having less certainty calculate identified with a class. The certainty element will rely on upon the late
employments of the hyperboxes and the forecast precision of the hyperboxes.
The customary frameworks depicts fresh occasions, i.e. it is possible that it will happen or don't. Fuzzy set was
acquainted by Zadeh [V] with speak to and control information that are not exact, rather fuzzy. After the fuzzy set has been
presented it has been utilized for tackling issues identified with arrangement and bunching. For more elevated amount of
basic leadership a fuzzy set approach for characterization of example is utilized. Initially, FMM (Classification) organize
[VI] proposed by P.K. Simpson, is utilized to serve for directed learning. In this [VI] relationship between the example
order and fuzzy sets has been depicted. It clarifies how the fuzzy min-max classifier neural system works utilizing learning
and reviewing calculation. A neural system classifier that utilizations min-max hyperboxes as fuzzy sets that are joined into
fuzzy set classes was presented.
P.K. Simpson additionally presents FMM (Clustering) arrange [VII] that is utilized to serve for unsupervised
learning. There were a few blames in the first fuzzy ART that has been adjusted in this. Gabris and Bargiela presents
General Fuzzy Min-Max (GFMM) calculation [VIII] which is the combination of both arrangement and grouping
identified with FMM. It is created on the premise of extension and withdrawal prepare. This calculation can be connected
as unadulterated characterization, immaculate bunching or cross breed grouping arrangement. In this [VIII] the
arrangement results can be fresh or fuzzy. Likewise to the first techniques, the GFMM uses min-max hyperboxes as fuzzy
sets. In GFMM, the preparation of information is quick, and the length of no indistinguishable information has a place with
various classes, the acknowledgment rate is 100%.
In Stochastic FMM [IX] proposed by A. Likas, the fuzzy min-max neural system [V] can be prepared
progressively by fittingly changing the quantity of hyperboxes and their relating values. In this [IX] the possibility of
irregular hyperbox and stochastic fuzzy min-max neural system has been proposed. Each hyperboxes is connected with
stochastic learning machine. In Adaptive determination min-max classifier [X] presented by Rizzi et al., two calculations
Adaptive Resolution Classifier (ARC) and pruned ARC is conceived. The automation level of the preparation system is
likewise an imperative variable with speculation ability and commotion heartiness. The speculation capacity of the first
min-max classifier depends generally on position and size of the hyperboxes created amid preparing. In this [X] the
characterization framework is programmed, since the preparation calculation does not rely on upon presentation request of
example and no basic parameter must be set by the client.
Consideration/Exclusion fuzzy hyperbox classifier [XI] is depicted in which, one or more fuzzy hyperbox
characterized by their relating least and most extreme vertices and the hyperbox enrollment capacity is utilized to portray
every class. Consideration hyperbox and rejection hyperbox are the two sorts of hyperboxes made. With these two sorts of
hyperboxes every class fuzzy set is spoken to as a blend of consideration hyperboxes of the same class short a mix of
avoidance hyperboxes. Another model called the Fuzzy Min-Max Neural Network Classifier with Compensatory Neurons
(FMCN) was proposed. To speak to the example classes, FMM with Compensatory Neuron (FMCN) hyperbox fuzzy sets
was utilized. The idea of compensatory neuron [XII] originates from how human mind functions at troublesome
conditions. The utilization of the constriction procedure is maintained a strategic distance from by FCMN, decreases
blunders brought about by it. Since the hyperboxes that are now made are not contracted, FMCN can hold the information
of the effectively learned examples more proficiently than FMNN and GFMN. It is [XII] taking into account the reflex
instrument of human cerebrum, takes in the information online in a solitary go of information, and looks after
www.tjprc.org

editor@tjprc.org

88

Alok Kumar Mishra & Er. Archana Singh

straightforwardness.
An information center based FMM neural system (DCFMN) display for example grouping was proposed [XIII].
For arranging the neurons of DCFMN, another participation work has been characterized in which the information center,
the commotion and the geometric focus of the hyperbox are considered. Another participation work [XIII] in light of the
information center is utilized rather than constriction procedure of the FMNN. To demonstrate the covering ranges of
hyperboxes of various classes, the enrollment capacity is added to neural system. Quteishat and Lim proposed new
calculation named Modified FMM. In endeavor to enhance the arrangement execution of FMM, few quantities of
expansive hyperboxes are shaped in the Modified FMM (MFMM) organize [XIV] and a few changes are finished.
Another info example is given; Euclidean separation measure is utilized for foreseeing the objective class
connected with the new information furthermore the fuzzy enrollment capacity of the info example to the hyperboxes
framed in FMM must be measured. In this [XIV] a control extraction calculation is additionally encased. For each FMM
hyperbox a certainty component is computed, and a client characterized edge is utilized to prune the hyperboxes with low
certainty elements. Altered FMM with Genetic Algorithm (MFMM-GA) [XV] is a two phase order of example and
extraction of administer process. The primary stage comprises of Modified Fuzzy min-max (MFMM) classifier and the
second stage depends on the hereditary calculation (GA) based classifier. To lessen the quantity of elements in the
separated standards, a "couldn't care less" approach is chosen by the GA manage extractor and fuzzy ifthen principles are
removed from the altered FMM classifier.
Offline and Online FMM and the classification and regression tree (CART), another approach [XVI, XVII] to
group and identify issues utilizing a half and half fuzzy min-max (FMM) neural system and arrangement and relapse tree
has been proposed. It [XVI] utilizes the idea of FMM with the end goal of arrangement and CART is utilized for control
extraction prepare. It likewise bolsters the disconnected and internet learning properties for blame location and conclusion
prepare.

METHOD: EFMM LEARNING ALGORITHM


The EFMM [IV] comprises of three stages i.e. hyperbox extension, cover test administer and hyperbox
constriction. Learning in FMM starts with datasets comprising of info examples and target classes. FMM makes
hyperboxes in view of the information designs. Hyperboxes are identified by least and most extreme focuses in a ndimensional space inside a unit hypercube (In). Hyperbox fuzzy sets are characterized as
(1)
Where, Bj is the hyperbox fuzzy set, Ah = (ah1, ah2,..., ahn) is the input pattern, and vj = (vj1, vj2,..., vjn) and wj
= (wj1,wj2,...,wjn) are the minimum and maximum points of Bj, respectively.
The hyperboxes will be framed according to the info given to the framework. Creation of hyperbox for new information
examples is rely on the participation work, given by

(2)
n
Where Bj is the membership function of the jth hyperbox, Ah = (ah1, ah2,..., ahn) I is the hth input pattern,
Impact Factor (JCC): 7.1293

NAAS Rating: 3.63

Data Classification using Min-Max Neural Network with Pruning

89

and y is a sensitivity parameter parameter that controls how quick the enrollment diminishes as the separation amongst Ah
and Bj increments. After the hyperboxes are created, then which hyperbox will have match with which class is chosen by
matrix U, as

(3)
Where bj is the jth hyperbox node and Ck is the kth class node. This can be represented as the following figure.

Figure 2: Three-Layer FMM Network


For each class node Ck can be calculated as
Ck = max bjujk

(4)

The FMM structure comprises of three layers. FA is the information layer, which contains the info designs from
the client. FB is the hyperbox layer, which demonstrates the hyperbox made amid learning process. The association
amongst FA and FB is appeared by the base (V) and most extreme (W) purposes of the hyperbox. FC is the yield layer
which speaks to the class. The association amongst FB and FC are paired values and are put away in grid U.
In EFMM, the information dataset is given by the client. Amid learning stage, hyperboxes are made or appointed to the
information dataset. The hyperbox expansion rule is provided, which uses expansion coefficient () for expansion.
Maxn(Wji , ahi ) Minn (Vji, ahi)

(5)

In light of the above condition it is watched that how much a hyperbox can be extended as for extension
coefficient (). In FMM, amid extension handle it processes the entirety of all measurements and contrasts and (n). In
contrast to this, EFMM considers every aspect exclusively and checks the differences amongst min and max points of
every aspect independently against .

Figure 3: 3-D Hyperbox

Hyperbox Overlap Test Rule: In FMM, the cases were not adequate to recognize all the covering cases. Thus, new
cases are included EFMM to defeat this issue. Covering exists when one of the accompanying cases given in [IV]

www.tjprc.org

editor@tjprc.org

90

Alok Kumar Mishra & Er. Archana Singh

is fulfilled. Because of the new covering rules in EFMM, the misclassification diminishes furthermore causes era
of more number of hyperboxes.

Hyperbox Contraction Rule: The constriction control is given in view of the instances of the hyperbox cover test.
Here, all cases are tried to locate a legitimate change.

Network Pruning: After the EFMM system is prepared, pruning is utilized to decrease the quantity of hyperboxes.
As the quantity of hyperboxes gets diminished, the administer extraction process will be quicker. The pruning of
hyperboxes is done utilizing certainty figure. The certainty component of each hyperbox depends on its recurrence
of utilization and exactness of forecast.
CFj = 1 w Fj + wAj

(6)

Where, Fj is the frequency of use of hyperbox j, Aj the accuracy of prediction of hyperbox j, and w [0, 1] is a
weighing factor. The value of Fj is outlined because the variety of patterns within the prediction set classified by any
hyperbox j, divided by the utmost variety of patterns within the prediction set classified by any hyperbox with constant
classification class. On the opposite hand, the worth of Aj is outlined because the variety of properly foretold set of
patterns classified by any hyperbox j, divided by the utmost properly classified patterns with constant classification class.
The arrogance issue finds the nice hyperboxes that area unit oftentimes used and offers high classification accuracy,
and conjointly the hyperboxes that area unit seldom used however still offers the high classification accuracy.

RESULTS
The dataset utilized as a part of this paper is the IRIS dataset. IRIS dataset is accessible openly and can be gotten
to from UCI machine learning storehouse. The IRIS dataset has 4 characteristics and 3 classes. There are 150 cases
accessible to utilize. Every class contains 50 occurrences. This finish dataset is partitioned into preparing dataset, forecast
dataset and testing dataset for order.
We have actualized the EFMM neural system for learning and grouping of information. To test the execution of
FMM with EFMM and EFMM with pruning we have utilized Iris information set from UCI machine learning archive. For
preparing of neural system 60% information is utilized, 20% information is utilized as the forecast set. At that point for
testing complete dataset is utilized. To check the aggregate number of hyperboxes shaped, the estimation of development
parameter is fluctuated from 0.05 to 0.95. The affectability parameter for the participation capacity is set to 0.5. For
pruning of hyperboxes the slice of significant worth is set to 0.5 and measuring component is set to 1. Table 1
demonstrates the aggregate number of hyperboxes got when is to 0.3, 0.4, 0.5 and 0.6.
Table 1: Comparison between FMM, EFMM and EFMM after
Pruning in Terms of Average Number of Hyperboxes

0.3
0.4
0.5
0.6

Impact Factor (JCC): 7.1293

FMM
39
28
20
16

Number of Hyperboxes
EFMM EFMM after Pruning
53
5
38
7
28
5
24
4

NAAS Rating: 3.63

Data Classification using Min-Max Neural Network with Pruning

91

Figure 4: Number of Hyperboxes Formed for Different Algorithm Using IRIS Dataset

Figure 5: Comparison of Recognition rate


It has been found that the quantity of hyperboxes made in EFMM is more than that of FMM. This is a result of the
new guidelines presented in EFMM [IV]. Be that as it may, EFMM has less misclassification rate than that of FMM.
Figure 5 demonstrates the acknowledgment rate of FMM, EFMM and EFMM in the wake of pruning for Iris information
set. Table 2 demonstrates the normal of acknowledgment rate.
Table 2: Average of Recognition Rate
Dataset FMM
EFMM EFMM after Pruning
IRIS 98.9 + 1.1 99.2 + 1.7
97.7 + 2.15

CONCLUSIONS
EFMM neural system is utilized for grouping reason. EFMM utilizes the idea of hyperbox for grouping. There are
four stages followed in EFMM neural system with pruning. To start with, the hyperbox is extended to contain the sources
of info. Second, cover of hyperboxes is checked. Third, Contraction is done by covers. Finally, EFMM makes more
number of hyperboxes when contrasted with FMM neural system. As the quantity of hyperboxes is more, the system
many-sided quality of EFMM in additional. In this way, in this venture, pruning of hyperboxes is done over EFMM neural
www.tjprc.org

editor@tjprc.org

92

Alok Kumar Mishra & Er. Archana Singh

system. Pruning of hyperboxes is done utilizing the idea of certainty component. The certainty component of each
hyperbox depends on its recurrence of utilization and precision of forecast. As the quantity of hyperboxes lessened the
system unpredictability of EFMM likewise decreases, however it influences the acknowledgment proportion by little edge.
REFERENCES
1.

D. Graupe. (1997). Principles of Artificial Neural Networks. Singapore: World Scientific.

2.

A. Quteishat et al. (2009). A neural network based multi-agent classifier system, Neurocomputing, 72, 7,16391647.

3.

G. P. Zhang. (2000). Neural networks for classification: A survey, IEEE Trans. Syst., Man, Cybern. C, Appl. Rev., 30, 4, 451
462.

4.

Mohammed Mohammed Falah et al.(2015). An Enhanced Fuzzy Min--Max Neural Network for Pattern Classification, IEEE
Transactions on Neural Networks and Learning Systems, 26, 417-429.

5.
6.

L. A. Zadeh. (1965). Fuzzy sets, Inf. Control, 8, 3, 338353.


Simpson, Patrick K. (1992). Fuzzy min-max neural networks-I: Classification. Neural Networks, IEEE Transactions on 3.5,
776-786.

7.

Simpson, Patrick K. (1993). Fuzzy min-max neural networks-Part 2: Clustering. IEEE Transactions on Fuzzy systems 1.1, 32.

8.

B. Gabrys et al. (2000). General fuzzy min-max neural network for clustering and classification, IEEE Trans. Neural Netw.,
vol. 11, no. 3, pp. 769783, May 2000.

9.

A. Likas. (2001). Reinforcement learning using the stochastic fuzzy min-max neural network, Neural Process. Lett., 13, 3,
213220.

10. A. Rizzi, M. Panella et al. (2002). Adaptive resolution min-max classifiers, IEEE Trans. Neural Netw., 13, 2, 402414.
11. A. Bargiela et al. (2004). An inclusion/exclusion fuzzy hyperbox classifier, Int. J. Knowl.-Based Intell. Eng. Syst., vol. 8, no. 2,
pp. 9198, Aug. 2004.
12. A. V. Nandedkar et al. (2007). A fuzzy min-max neural network classifier with compensatory neuron architecture, IEEE Trans.
Neural Netw., 18, 1, 4254.
13. H. Zhang et al. (2011). Data-core-based fuzzy min-max neural network for pattern classification, IEEE Trans. Neural Netw.,
22, 12, 23392352.
14. A. Quteishat et al. (2008). A modified fuzzy min-max neural network with rule extraction and its application to fault detection
and classification, Appl. Soft Comput., 8, 2, 985995.
15. A. Quteishat et al. (2010). A modified fuzzy min-max neural network with a genetic-algorithm-based rule extractor for pattern
classification, IEEE Trans. Syst., Man, Cybern. A, Syst., Humans, 40, 3, 641650.
16. M. Seera et al. (2012). Fault detection and diagnosis of induction motors using motor current signature analysis and a hybrid
FMM-CART model, IEEE Trans. Neural Netw. Learn.Syst., 23, 1, 97108.
17. M. Seera wt al. (2014). Online motor fault detection and diagnosis using a hybrid FMM-CART model, IEEE Trans. Neural
Netw. Learn. Syst., 25, 4, 17.

Impact Factor (JCC): 7.1293

NAAS Rating: 3.63

Das könnte Ihnen auch gefallen