SlideShare a Scribd company logo
Natarajan Meghanathan, et al. (Eds): SIPM, FCST, ITCA, WSE, ACSIT, CS & IT 06, pp. 349–356, 2012.
© CS & IT-CSCP 2012 DOI : 10.5121/csit.2012.2334
DESIGN AND IMPLEMENTATION OF BINARY
NEURAL NETWORK LEARNING WITH FUZZY
CLUSTERING
Sachin Bhandari1
and Dr. Aruna Tiwari2
1
Department of Computer Engineering, SGSITS, Indore, India
er.bhandari04@gmail.com
2
Department of Computer Engineering, SGSITS, Indore, India
atiwari@sgsits.ac.in
ABSTRACT
In this paper, Design and Implementation of Binary Neural Network Learning with Fuzzy
Clustering (DIBNNFC), is proposed to classify semisupervised data, it is based on the
concept of binary neural network and geometrical expansion. Parameters are updated
according to the geometrical location of the training samples in the input space, and each
sample in the training set is learned only once. It’s a semisupervised based approach, the
training samples are semi-labelled i.e. for some samples, labels are known and for some
samples data labels are not known. The method starts with classification, which is done by
using the concept of ETL algorithm. In classification process various classes are formed.
These classes classify samples in to two classes after that considers each class as a region
and calculates the average of the entire region separately. This average is centres of the
region which is used for the purpose of clustering by using FCM algorithm. Once clustering
process over labelling of semi supervised data is done, then whole samples would be classify
by (DIBNNFC). The method proposes here is exhaustively tested with different benchmark
datasets and it is found that, on increasing value of training parameters number of hidden
neurons and training time both are getting decrease. The result reported, using real character
recognition data set and result will compare with existing semi-supervised classifier, the
proposed approach learned with semi-supervised leads to higher classification accuracy.
KEYWORDS
Semisupervised classification, Geometrical Expansion, Binary Neural Network, Fuzzy C-
means algorithm, ETL algorithm.
1. INTRODUCTION
Recently, the back propagation learning (BPL) algorithm has been applied to many binary-to-
binary mapping problems [6], [2]. However, since the BPL algorithm searches the solution in
continuous space, the BPL algorithm applied to binary-to-binary mapping problems results in
long training time and inefficient performance. Typically, the BLTA algorithm require an
extremely high number of iterations to obtain even a simple binary-to-binary mapping [3]. Also,
in the BLTA algorithm, the number of neurons in the hidden layer required to solve a given
problem is not known a priori. Since the numbers of neurons in the input and the output layer are
determined by the dimensions of the input and output vectors, respectively, the abilities of three-
layer neural networks depend on the number of neurons in the hidden layer. Therefore, one of the
most important problems in application of three-layer neural networks is to determine the
necessary number of neurons in the hidden layer. It has been widely recognized that Stone-
Weierstrass’s theorem does not give a practical guideline in determining the required number of
350 Computer Science & Information Technology ( CS & IT )
neurons [4]. In this paper, the geometrical learning algorithm called expand-and-truncate learning
(ETL) is proposed with fuzzy c-means algorithm to train a three-layer binary neural network
(BNN) for the generation of a semisupervised classifier.
In this paper “Design and implementation of Binary Neural Network learning with fuzzy
clustering (DIBNNFC)” an approach to classify semisupervised data is proposed. The work
present in this paper is based on binary neural network framework proposed by Jung H. Kim and
Sung-Kwon Park, Member, IEEE in “The Geometrical learning of binary neural networks and
weiling cai, Sangaon Chen, Daoqiang Zhang in Robust Fuzzy Relational classifier Incorporating
the Soft class Labels [5]. ETL algorithm is an approach to cover all true vertices in a neural
network. By using this algorithm a linearly inseparable problems is divided into a set of linearly
separable problems. In this firstly used ETL algorithm for classify the semi supervised data in to
classes, after that these classes considered as a region then for calculating the centre of these
region, taking average of the region then this average is considered as a centre of the region
which is used in FCM algorithm for Clustering. The second method is clustering which is done by
FCM. FCM done labelling of all samples and show that which sample is belong to which region
through this the classification of semi supervised data is completed. Existing experimental results
character data set is proposed to compare with semi-supervised classifier to be implemented.
2. OVERVIEW OF DIBNNFC
2.1 Basic Concept
ETL algorithm is a geometrical approach and it is a supervised learning algorithm works for
labelled dataset. In this paper, the learning algorithms called expand and-truncate learning (ETL)
is proposed to train multilayer binary neural networks (BNN) with guaranteed convergence for
any binary-to-binary mapping. By using this algorithm we make a classifier which classifies a
semi supervised data, with the help of using FCM algorithm which is an unsupervised learning
algorithm.
Boolean functions have the geometrical property which makes it possible to transform non linear
representation to linear representation for each hidden neuron. We consider a Boolean function
with n input and one output, y = f (x1, x2… xn), where y Є (0, 1) and xi Є (0, 1), i= (1.....n).
These 2n binary patterns (0, 1)n
can be considered as a n – 1 dimensional unit hypercube.
net (X, T)=(w1x1 + w2x2 + . . . . . + wnxn – T) = 0
2.2 Working of DIBNNFC
In the system given below the input data is labelled and unlabelled both. Labelled data are those
in which label is defined for each sample and unlabelled data are those in which label is not
defined for samples. The datasets from the different domain like Iris, Balance Scale, Bupa, and
Wine etc are taken for system. As we seen in fig.1 an overview of the system is given, in this the
semi labelled data is behave like an input and firstly train by ETL algorithm which is a supervised
learning technique and the clustering process is done by FCM algorithm which is an unsupervised
technique by the help of these two, an algorithm is proposed to make a binary neural network
with fuzzy clustering which work as a semi supervised classifier. This is an overview of the
proposed problem of systemThe goal of semi-supervised learning is to understand how
combining labelled and unlabeled data may change the learning behaviour, and design algorithms
that take advantage of such a combination
Computer Science & Information Technology ( CS & IT ) 351
Figure 1. An Overview of the System
3. PROPOSED METHOD: DIBNNFC
Suppose that a set of n-bit training input vectors is given and a binary desired output is assigned
to each training input vector. An n-bit input vector can be considered as a vertex of an n-
dimensional hypercube. The two classes of training input vectors (i.e., vertices) can be separated
by an (n - 1)-dimensional hyperplane which is expressed as a net function,
net (X, T) =
where the wi’s and T are constant. The set of training inputs is said to be linearly separable (LS),
and the (n - 1)-dimensional hyperplane is the separating hyperplane. The (n - 1)-dimensional
separating hyperplanes can be established by an n-input neuron with a hard-limiter activation
function as below,
where, is output of the neuron, is connection weight between the input and neuron,
input to the neuron, is threshold of neuron.
For learning the hidden layer the geometrical learning is used to decompose an arbitrarily linearly
inseparable function into multiple LS functions [1]. For any binary-to-binary mapping, ETL will
determine the required LS functions, each of which is realized by a neuron in the hidden layer [8].
Check if there are true and false vertices are
at hamming distance equal to one and If only true vertices at hamming distance less than‘d’ [7].
The hyperplane can be represented by the equation below.
where
Pre
processing
Apply
ETL
Classify
Labelled
Data
Apply
FCM
algorithm
Cluster the
unlabelled
data
Semi
labelled data
352 Computer Science & Information Technology ( CS & IT )
i.e.
The first hyperplane is found. The SITV includes the core vertex and the true vertices which are
separated by the first hyperplane [12]. Consider a function: {0, 1}n
=> {0, 1}. Denotes the n-bit
input vector and {0, 1} denotes the desired output corresponding to the n-bit input vector. The
value of f divides the 2n
points of n-tuple (i.e., 2n
vertices of n-cube) into two classes, those for
which f is zero and those for which it is one. A function f is linearly separable if and only if there
exists a hypersphere such that all true vertices lie inside or on the hypersphere and all false
vertices lie outside or vice versa. Consider a reference hypersphere and an n-dimensional
hypersphere which has its centre at and its radius r.
= number of elements in SITV including trial vertex.
where, is an element in SITV
is ith
bit of vertex .
The point in the n-dimensional space represents the centre of
gravity of all elements in SITV. The separating hyperplane can be representated as
where T is a constant called threshold. That is, if there exists a separating hyperplane
For learning an output layer if required separating hyperplanes are finds with only one core
vertex, the weights and threshold of one output neuron are set as follows. The weight of the link
from the odd-numbered hidden neuron to the output neuron is set to 1. The weight of the link
from the even-numbered neuron to the output neuron is set to -1. For calculating the center of the
class average method is used. Taking an average of entire region of classes, this existing center is
used for clustering as an input centers in FCM algorithm.
Fuzzy c-means (FCM) [9], [10], [11] is a data clustering technique in which a data set is grouped
into n clusters with every data point in the dataset belonging to every cluster will have a high
degree of belonging or membership to that cluster and another data point that lies far away from
the center of a cluster will have a low degree of belonging or membership to that cluster. It is
based on the concept of fuzzy C-partition which was introduced by Ruspini and developed Dunn
and generalized by Jim Bezdek in 1981 as an improvement over earlier clustering methods. It is
based on minimization of the following objective function:
Computer Science & Information Technology ( CS & IT ) 353
where m is any real number greater than 1, is the degree of membership of xi in the cluster j, xi
is the ith
of d-dimensional measured data, cj is the d-dimension center of the cluster, and ||*|| is any
norm expressing the similarity between any measured data and the center. The FCM algorithm is
composed of following steps:
Step 1: The FCM algorithm starts with a set of randomly initializing cluster membership matrix
and with a set of predefined number of cluster center.
Step 2: Using the randomly initialize cluster membership matrix, the cluster center are calculated.
Step 3: Using the obtained cluster center, the cluster membership matrix is updated by taking the
fractional distance from the point to the cluster center.
The fuzzy c-means algorithm imposes a direct constraint on the fuzzy membership function i.e.
the total membership for a point in sample or decision space must add to 1.The Fuzzy partitioning
is carried out through an iterative optimization of the objective function shown above, with the
update of membership uij and the cluster centers cj .This iteration will stop when
1
|| ||k k
U U ε+
− < where ε is a termination criterion between 0 and 1, whereas k are the iteration
steps. This procedure converges to a local minimum or a saddle point of Jm.
3.1 Steps for proposed approach as follows:
1. Firstly select a core vertex initially
2. Calculate number of bit difference between core vertex and rest of the vertex
3. Select shortest distance vertex as final core vertex
4. Core vertex tends two SITVand rest of vertex are put in to rest
5. Calculate distance between SITV and rest vertex
6. Select trial vertex whose output is 1 from rest whose distance from core vertex is less
then other vertex in rest
7. Find hyperplane equation by using this formula:-
here If
354 Computer Science & Information Technology ( CS & IT )
where
8. Calculate tmin and fmax
T = tmin + fmax / 2
9. If tmin > fmax hyperplane is exist then SITV = trial vertex, SITV removed trial vertex from
rest save equation & calculate threshold
else select new trial vertex until all true output vertices are covered Goto 6 until all true
vertices are covered.
10. If rest contain vertex whose output is 1then convert true in to false & false into true
Goto step 6.
else find weight of output layer by putting the weight of the link from the odd
numbered hidden neuron to the output neuron is set to 1. The weight of the link
from the even-numbered neuron to the output neuron is set to -1
11. Calculate the average of the classes to find the centers of the region, and these centers are
used in FCM
12. Initialize U= [uij] matrix, U (0)
13. At kth
-step: calculate the centers vectors C (k)
= [cj] with U (k)
14. Update U (k)
, U (k+1)
15. If then STOP; otherwise return to step 13.
4. EXPERIMENT AND RESULTS
The proposed approach is implemented in Matlab 7.8.0(R2009a) and applied on various data sets.
Results of these experiments are summarized in Table 1. In this section the proposed DIBNNFC
neural network have been tested on dataset and the results are obtained. Datasets used for training
are Iris, Balance Scale, Bupa, and Wine data set. The input data sets files are pre processed,
which provide require input set for the learning. The input data set files is converted into Binary
data by performing Normalisation. Proposed approach determines number of iterations in
obtaining cluster center. Furthermore we use Output of ETL to decide that which cluster goes to
which class after classification and calculate the accuracy of the classifier.
Computer Science & Information Technology ( CS & IT ) 355
Table 1. showing results of DIBNNFC classifier for various Datasets
Dataset Number of
Instances
Number
of
Classes
Number
of
Features
Iterations Training
time in
seconds
Classification
Accuracy
Iris 150 3 4 33 0.175633 96%
Balance Scale 625 3 4 8 0.238572 44.55%
BUPA 345 2 6 166 1.007675 58.14%
Wine 178 3 13 56 0.69163 68.54%
5. CONCLUSIONS
This paper, “Design and implementation of Binary Neural Network learning with fuzzy clustering
(DIBNNFC)” is proposed to classify semi supervised data and train multilayer BNN for any
binary-to-binary mapping. It improves significant performance as compared to the previous
framework. We have shown that for any generation of binary-to-binary mapping, the proposed
DIBNNFC approach always converges and finds the three-layer BNN, by automatically
determining a required number of neurons in the hidden layer. Fcm done Clustering for semi
labeled data and provide iterations and classification accuracy as per Input data. Any semi
supervised data can be classified with this classifier. The neuron in the proposed BNN employs a
hard-limiter activation function, only integer weights and integer thresholds. It leads to a higher
accuracy of classification. DIBNNFC is better approach to classify semi supervised data then
others.
REFERENCES
[1] Jung H. Kim and Sung-Kwon Park, “The Geometrical Learning of Binary Neural Networks,” IEEE
T~mn. Neural Networks’, Vol. 6, No.1. Pp.237-247, Jun. 1995.
[2] J. M. Zurada and W. Shen, “Sufficient condition for convergence of a relaxation algorithm in actual
single-layer neural networks,” pp. 3W303, Dec. 1990
[3] D. L. Gray and A. N. Michel, “A training algorithm for binary feed forward neural network,” IEEE
Trans. Neural Networks, pp. 176194, Mar. 1992
[4] N. E. Cotter, “The Stone-Weierstrass theorem and its application to neural networks,” IEEE Trans.
Neural Networks, Dec. 1990.
[5] Weiling Cai, Songcan Chen *, Daoqiang Zhang, “Robust fuzzy relational classifier incorp- orating the
soft class labels” Received 29 July 2006; received in revised form 3 June 2007 June 2007 Available
online 6 August 2007
[6] S. Oh and R. J. Mark 11, “Dispersive propagation skew effect in iteration neural networks
‘IEEE Trans. Neural Networks, pp. 16C-162, Jan. 1991
[7] D. Wang and Narendra S. Chaudhari, “A Multi-core Learning algorithm for Boolean Neural
Networks”, School of Computing Engineering, Block N4-2a-32, Nanyang Technological University,
Singapore 639798-2003 IEEE
[8] Narendra S. Chaudhari and (Mrs) Aruna Tiwari, “Extending ETL for multi-class output”, Nanyang
Technological University, Singapore 639798, Neural Information Processing (ICONIP'OZ), Vol. 4
[9] M. Setnes and R. Babuska, “Fuzzy relational classifier trained by fuzzy clustering,” IEEE Trans.
Syst. Man Cybern. B, Cybern., vol. 29, no. 5,pp. 619–625, Oct. 1999.
[10] Maedeh zirak Javanmard,“Fuzzy C-means algorithm and its application in case of public database of
automobile property information. 2010[Online]. Available: http:// www.authorstream.com//
Presentation / aSGuest 80031-738704-fuzzy-c-means-clustering.
356 Computer Science & Information Technology ( CS & IT )
[11] W. Pedrycz and G.Vukovich, “Fuzzy clustering with supervision,” Pattern Recognition., vol. 37, pp.
1229–1349, 2004.
[12] Atsushi Yamamoto, Toshimichi Saito, “An Improved Expand-and-Truncate Learning”, EEE Dept.,
HOSEI University, Aug 1997 IEEE.
Authors
Sachin Bhandari received her Bachelor of Engineering degree in Computer
Engineering from RGPV University, India in 2010. He is currently pursuing
Master of Engineering in Computer Engineering from SGSITS, Indore, India.
His research interests include Data mining, and Soft Computing.
Dr. Aruna Tiwari She is currently working as Associate Professors in
Computer Engineering Department at SGSITS Indore, India. Her research
interest areas are Data mining, Computational Learning and Soft Computing.

More Related Content

PDF
Investigations on Hybrid Learning in ANFIS
PDF
Self Organizing Maps: Fundamentals
PPT
Neural networks Self Organizing Map by Engr. Edgar Carrillo II
PPT
Sefl Organizing Map
PDF
Introduction to Applied Machine Learning
PPT
Supervised Learning
PDF
A survey research summary on neural networks
PDF
Data clustering using kernel based
Investigations on Hybrid Learning in ANFIS
Self Organizing Maps: Fundamentals
Neural networks Self Organizing Map by Engr. Edgar Carrillo II
Sefl Organizing Map
Introduction to Applied Machine Learning
Supervised Learning
A survey research summary on neural networks
Data clustering using kernel based

What's hot (20)

PDF
Machine Learning: Introduction to Neural Networks
PDF
A Mixed Binary-Real NSGA II Algorithm Ensuring Both Accuracy and Interpretabi...
PDF
The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...
PDF
Objective Evaluation of a Deep Neural Network Approach for Single-Channel Spe...
PDF
Offline Character Recognition Using Monte Carlo Method and Neural Network
PDF
B42010712
PDF
An efficient technique for color image classification based on lower feature ...
PDF
International Journal of Engineering Research and Development (IJERD)
PPTX
Associative memory network
PPT
Character Recognition using Artificial Neural Networks
PDF
Neural Networks: Self-Organizing Maps (SOM)
PDF
GENERAL REGRESSION NEURAL NETWORK BASED POS TAGGING FOR NEPALI TEXT
PDF
Artificial Neural Networks Lect3: Neural Network Learning rules
PPTX
Supervised learning network
PDF
Convolution Neural Networks
PPTX
Neural networks
PDF
Self Organization Map
PDF
Artificial neural network for machine learning
PDF
Hand Written Digit Classification
PPTX
Unsupervised learning
Machine Learning: Introduction to Neural Networks
A Mixed Binary-Real NSGA II Algorithm Ensuring Both Accuracy and Interpretabi...
The Positive Effects of Fuzzy C-Means Clustering on Supervised Learning Class...
Objective Evaluation of a Deep Neural Network Approach for Single-Channel Spe...
Offline Character Recognition Using Monte Carlo Method and Neural Network
B42010712
An efficient technique for color image classification based on lower feature ...
International Journal of Engineering Research and Development (IJERD)
Associative memory network
Character Recognition using Artificial Neural Networks
Neural Networks: Self-Organizing Maps (SOM)
GENERAL REGRESSION NEURAL NETWORK BASED POS TAGGING FOR NEPALI TEXT
Artificial Neural Networks Lect3: Neural Network Learning rules
Supervised learning network
Convolution Neural Networks
Neural networks
Self Organization Map
Artificial neural network for machine learning
Hand Written Digit Classification
Unsupervised learning
Ad

Similar to DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTERING (20)

PDF
Electricity Demand Forecasting Using Fuzzy-Neural Network
PDF
Adaptive Training of Radial Basis Function Networks Based on Cooperative
PDF
PDF
F017533540
PDF
2. 2017_IEEE_INTERCON_paper_109xxxxx.pdf
 
PDF
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
PDF
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
PDF
Recognition of handwritten digits using rbf neural network
PDF
Recognition of handwritten digits using rbf neural network
PDF
NEURALNETWORKS_DM_SOWMYAJYOTHI.pdf
PDF
Web spam classification using supervised artificial neural network algorithms
PDF
X-TREPAN: A MULTI CLASS REGRESSION AND ADAPTED EXTRACTION OF COMPREHENSIBLE D...
PDF
X-TREPAN : A Multi Class Regression and Adapted Extraction of Comprehensible ...
PDF
APPLYING NEURAL NETWORKS FOR SUPERVISED LEARNING OF MEDICAL DATA
PDF
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...
PDF
Classification of Iris Data using Kernel Radial Basis Probabilistic Neural N...
PDF
Classification of Iris Data using Kernel Radial Basis Probabilistic Neural Ne...
PDF
An efficient fuzzy classifier with feature selection based
PDF
H017376369
PDF
A New Classifier Based onRecurrent Neural Network Using Multiple Binary-Outpu...
Electricity Demand Forecasting Using Fuzzy-Neural Network
Adaptive Training of Radial Basis Function Networks Based on Cooperative
F017533540
2. 2017_IEEE_INTERCON_paper_109xxxxx.pdf
 
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
Web Spam Classification Using Supervised Artificial Neural Network Algorithms
Recognition of handwritten digits using rbf neural network
Recognition of handwritten digits using rbf neural network
NEURALNETWORKS_DM_SOWMYAJYOTHI.pdf
Web spam classification using supervised artificial neural network algorithms
X-TREPAN: A MULTI CLASS REGRESSION AND ADAPTED EXTRACTION OF COMPREHENSIBLE D...
X-TREPAN : A Multi Class Regression and Adapted Extraction of Comprehensible ...
APPLYING NEURAL NETWORKS FOR SUPERVISED LEARNING OF MEDICAL DATA
Machine Learning Algorithms for Image Classification of Hand Digits and Face ...
Classification of Iris Data using Kernel Radial Basis Probabilistic Neural N...
Classification of Iris Data using Kernel Radial Basis Probabilistic Neural Ne...
An efficient fuzzy classifier with feature selection based
H017376369
A New Classifier Based onRecurrent Neural Network Using Multiple Binary-Outpu...
Ad

More from cscpconf (20)

PDF
ANALYSIS OF LAND SURFACE DEFORMATION GRADIENT BY DINSAR
PDF
4D AUTOMATIC LIP-READING FOR SPEAKER'S FACE IDENTIFCATION
PDF
MOVING FROM WATERFALL TO AGILE PROCESS IN SOFTWARE ENGINEERING CAPSTONE PROJE...
PDF
PROMOTING STUDENT ENGAGEMENT USING SOCIAL MEDIA TECHNOLOGIES
PDF
A SURVEY ON QUESTION ANSWERING SYSTEMS: THE ADVANCES OF FUZZY LOGIC
PDF
DYNAMIC PHONE WARPING – A METHOD TO MEASURE THE DISTANCE BETWEEN PRONUNCIATIONS
PDF
INTELLIGENT ELECTRONIC ASSESSMENT FOR SUBJECTIVE EXAMS
PDF
TWO DISCRETE BINARY VERSIONS OF AFRICAN BUFFALO OPTIMIZATION METAHEURISTIC
PDF
DETECTION OF ALGORITHMICALLY GENERATED MALICIOUS DOMAIN
PDF
GLOBAL MUSIC ASSET ASSURANCE DIGITAL CURRENCY: A DRM SOLUTION FOR STREAMING C...
PDF
IMPORTANCE OF VERB SUFFIX MAPPING IN DISCOURSE TRANSLATION SYSTEM
PDF
EXACT SOLUTIONS OF A FAMILY OF HIGHER-DIMENSIONAL SPACE-TIME FRACTIONAL KDV-T...
PDF
AUTOMATED PENETRATION TESTING: AN OVERVIEW
PDF
CLASSIFICATION OF ALZHEIMER USING fMRI DATA AND BRAIN NETWORK
PDF
VALIDATION METHOD OF FUZZY ASSOCIATION RULES BASED ON FUZZY FORMAL CONCEPT AN...
PDF
PROBABILITY BASED CLUSTER EXPANSION OVERSAMPLING TECHNIQUE FOR IMBALANCED DATA
PDF
CHARACTER AND IMAGE RECOGNITION FOR DATA CATALOGING IN ECOLOGICAL RESEARCH
PDF
SOCIAL MEDIA ANALYTICS FOR SENTIMENT ANALYSIS AND EVENT DETECTION IN SMART CI...
PDF
SOCIAL NETWORK HATE SPEECH DETECTION FOR AMHARIC LANGUAGE
PDF
APPLYING DISTRIBUTIONAL SEMANTICS TO ENHANCE CLASSIFYING EMOTIONS IN ARABIC T...
ANALYSIS OF LAND SURFACE DEFORMATION GRADIENT BY DINSAR
4D AUTOMATIC LIP-READING FOR SPEAKER'S FACE IDENTIFCATION
MOVING FROM WATERFALL TO AGILE PROCESS IN SOFTWARE ENGINEERING CAPSTONE PROJE...
PROMOTING STUDENT ENGAGEMENT USING SOCIAL MEDIA TECHNOLOGIES
A SURVEY ON QUESTION ANSWERING SYSTEMS: THE ADVANCES OF FUZZY LOGIC
DYNAMIC PHONE WARPING – A METHOD TO MEASURE THE DISTANCE BETWEEN PRONUNCIATIONS
INTELLIGENT ELECTRONIC ASSESSMENT FOR SUBJECTIVE EXAMS
TWO DISCRETE BINARY VERSIONS OF AFRICAN BUFFALO OPTIMIZATION METAHEURISTIC
DETECTION OF ALGORITHMICALLY GENERATED MALICIOUS DOMAIN
GLOBAL MUSIC ASSET ASSURANCE DIGITAL CURRENCY: A DRM SOLUTION FOR STREAMING C...
IMPORTANCE OF VERB SUFFIX MAPPING IN DISCOURSE TRANSLATION SYSTEM
EXACT SOLUTIONS OF A FAMILY OF HIGHER-DIMENSIONAL SPACE-TIME FRACTIONAL KDV-T...
AUTOMATED PENETRATION TESTING: AN OVERVIEW
CLASSIFICATION OF ALZHEIMER USING fMRI DATA AND BRAIN NETWORK
VALIDATION METHOD OF FUZZY ASSOCIATION RULES BASED ON FUZZY FORMAL CONCEPT AN...
PROBABILITY BASED CLUSTER EXPANSION OVERSAMPLING TECHNIQUE FOR IMBALANCED DATA
CHARACTER AND IMAGE RECOGNITION FOR DATA CATALOGING IN ECOLOGICAL RESEARCH
SOCIAL MEDIA ANALYTICS FOR SENTIMENT ANALYSIS AND EVENT DETECTION IN SMART CI...
SOCIAL NETWORK HATE SPEECH DETECTION FOR AMHARIC LANGUAGE
APPLYING DISTRIBUTIONAL SEMANTICS TO ENHANCE CLASSIFYING EMOTIONS IN ARABIC T...

Recently uploaded (20)

PPTX
Pharmacology of Heart Failure /Pharmacotherapy of CHF
PDF
Saundersa Comprehensive Review for the NCLEX-RN Examination.pdf
PDF
102 student loan defaulters named and shamed – Is someone you know on the list?
PDF
Origin of periodic table-Mendeleev’s Periodic-Modern Periodic table
PPTX
Introduction to Child Health Nursing – Unit I | Child Health Nursing I | B.Sc...
PDF
BÀI TẬP BỔ TRỢ 4 KỸ NĂNG TIẾNG ANH 9 GLOBAL SUCCESS - CẢ NĂM - BÁM SÁT FORM Đ...
PDF
English Language Teaching from Post-.pdf
PDF
Physiotherapy_for_Respiratory_and_Cardiac_Problems WEBBER.pdf
PPTX
NOI Hackathon - Summer Edition - GreenThumber.pptx
PDF
Open folder Downloads.pdf yes yes ges yes
PPTX
Onica Farming 24rsclub profitable farm business
DOCX
UPPER GASTRO INTESTINAL DISORDER.docx
PPTX
Pharma ospi slides which help in ospi learning
PPTX
Introduction_to_Human_Anatomy_and_Physiology_for_B.Pharm.pptx
PDF
From loneliness to social connection charting
PPTX
Revamp in MTO Odoo 18 Inventory - Odoo Slides
PDF
Anesthesia in Laparoscopic Surgery in India
PDF
Abdominal Access Techniques with Prof. Dr. R K Mishra
PDF
ANTIBIOTICS.pptx.pdf………………… xxxxxxxxxxxxx
PPTX
IMMUNITY IMMUNITY refers to protection against infection, and the immune syst...
Pharmacology of Heart Failure /Pharmacotherapy of CHF
Saundersa Comprehensive Review for the NCLEX-RN Examination.pdf
102 student loan defaulters named and shamed – Is someone you know on the list?
Origin of periodic table-Mendeleev’s Periodic-Modern Periodic table
Introduction to Child Health Nursing – Unit I | Child Health Nursing I | B.Sc...
BÀI TẬP BỔ TRỢ 4 KỸ NĂNG TIẾNG ANH 9 GLOBAL SUCCESS - CẢ NĂM - BÁM SÁT FORM Đ...
English Language Teaching from Post-.pdf
Physiotherapy_for_Respiratory_and_Cardiac_Problems WEBBER.pdf
NOI Hackathon - Summer Edition - GreenThumber.pptx
Open folder Downloads.pdf yes yes ges yes
Onica Farming 24rsclub profitable farm business
UPPER GASTRO INTESTINAL DISORDER.docx
Pharma ospi slides which help in ospi learning
Introduction_to_Human_Anatomy_and_Physiology_for_B.Pharm.pptx
From loneliness to social connection charting
Revamp in MTO Odoo 18 Inventory - Odoo Slides
Anesthesia in Laparoscopic Surgery in India
Abdominal Access Techniques with Prof. Dr. R K Mishra
ANTIBIOTICS.pptx.pdf………………… xxxxxxxxxxxxx
IMMUNITY IMMUNITY refers to protection against infection, and the immune syst...

DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTERING

  • 1. Natarajan Meghanathan, et al. (Eds): SIPM, FCST, ITCA, WSE, ACSIT, CS & IT 06, pp. 349–356, 2012. © CS & IT-CSCP 2012 DOI : 10.5121/csit.2012.2334 DESIGN AND IMPLEMENTATION OF BINARY NEURAL NETWORK LEARNING WITH FUZZY CLUSTERING Sachin Bhandari1 and Dr. Aruna Tiwari2 1 Department of Computer Engineering, SGSITS, Indore, India [email protected] 2 Department of Computer Engineering, SGSITS, Indore, India [email protected] ABSTRACT In this paper, Design and Implementation of Binary Neural Network Learning with Fuzzy Clustering (DIBNNFC), is proposed to classify semisupervised data, it is based on the concept of binary neural network and geometrical expansion. Parameters are updated according to the geometrical location of the training samples in the input space, and each sample in the training set is learned only once. It’s a semisupervised based approach, the training samples are semi-labelled i.e. for some samples, labels are known and for some samples data labels are not known. The method starts with classification, which is done by using the concept of ETL algorithm. In classification process various classes are formed. These classes classify samples in to two classes after that considers each class as a region and calculates the average of the entire region separately. This average is centres of the region which is used for the purpose of clustering by using FCM algorithm. Once clustering process over labelling of semi supervised data is done, then whole samples would be classify by (DIBNNFC). The method proposes here is exhaustively tested with different benchmark datasets and it is found that, on increasing value of training parameters number of hidden neurons and training time both are getting decrease. The result reported, using real character recognition data set and result will compare with existing semi-supervised classifier, the proposed approach learned with semi-supervised leads to higher classification accuracy. KEYWORDS Semisupervised classification, Geometrical Expansion, Binary Neural Network, Fuzzy C- means algorithm, ETL algorithm. 1. INTRODUCTION Recently, the back propagation learning (BPL) algorithm has been applied to many binary-to- binary mapping problems [6], [2]. However, since the BPL algorithm searches the solution in continuous space, the BPL algorithm applied to binary-to-binary mapping problems results in long training time and inefficient performance. Typically, the BLTA algorithm require an extremely high number of iterations to obtain even a simple binary-to-binary mapping [3]. Also, in the BLTA algorithm, the number of neurons in the hidden layer required to solve a given problem is not known a priori. Since the numbers of neurons in the input and the output layer are determined by the dimensions of the input and output vectors, respectively, the abilities of three- layer neural networks depend on the number of neurons in the hidden layer. Therefore, one of the most important problems in application of three-layer neural networks is to determine the necessary number of neurons in the hidden layer. It has been widely recognized that Stone- Weierstrass’s theorem does not give a practical guideline in determining the required number of
  • 2. 350 Computer Science & Information Technology ( CS & IT ) neurons [4]. In this paper, the geometrical learning algorithm called expand-and-truncate learning (ETL) is proposed with fuzzy c-means algorithm to train a three-layer binary neural network (BNN) for the generation of a semisupervised classifier. In this paper “Design and implementation of Binary Neural Network learning with fuzzy clustering (DIBNNFC)” an approach to classify semisupervised data is proposed. The work present in this paper is based on binary neural network framework proposed by Jung H. Kim and Sung-Kwon Park, Member, IEEE in “The Geometrical learning of binary neural networks and weiling cai, Sangaon Chen, Daoqiang Zhang in Robust Fuzzy Relational classifier Incorporating the Soft class Labels [5]. ETL algorithm is an approach to cover all true vertices in a neural network. By using this algorithm a linearly inseparable problems is divided into a set of linearly separable problems. In this firstly used ETL algorithm for classify the semi supervised data in to classes, after that these classes considered as a region then for calculating the centre of these region, taking average of the region then this average is considered as a centre of the region which is used in FCM algorithm for Clustering. The second method is clustering which is done by FCM. FCM done labelling of all samples and show that which sample is belong to which region through this the classification of semi supervised data is completed. Existing experimental results character data set is proposed to compare with semi-supervised classifier to be implemented. 2. OVERVIEW OF DIBNNFC 2.1 Basic Concept ETL algorithm is a geometrical approach and it is a supervised learning algorithm works for labelled dataset. In this paper, the learning algorithms called expand and-truncate learning (ETL) is proposed to train multilayer binary neural networks (BNN) with guaranteed convergence for any binary-to-binary mapping. By using this algorithm we make a classifier which classifies a semi supervised data, with the help of using FCM algorithm which is an unsupervised learning algorithm. Boolean functions have the geometrical property which makes it possible to transform non linear representation to linear representation for each hidden neuron. We consider a Boolean function with n input and one output, y = f (x1, x2… xn), where y Є (0, 1) and xi Є (0, 1), i= (1.....n). These 2n binary patterns (0, 1)n can be considered as a n – 1 dimensional unit hypercube. net (X, T)=(w1x1 + w2x2 + . . . . . + wnxn – T) = 0 2.2 Working of DIBNNFC In the system given below the input data is labelled and unlabelled both. Labelled data are those in which label is defined for each sample and unlabelled data are those in which label is not defined for samples. The datasets from the different domain like Iris, Balance Scale, Bupa, and Wine etc are taken for system. As we seen in fig.1 an overview of the system is given, in this the semi labelled data is behave like an input and firstly train by ETL algorithm which is a supervised learning technique and the clustering process is done by FCM algorithm which is an unsupervised technique by the help of these two, an algorithm is proposed to make a binary neural network with fuzzy clustering which work as a semi supervised classifier. This is an overview of the proposed problem of systemThe goal of semi-supervised learning is to understand how combining labelled and unlabeled data may change the learning behaviour, and design algorithms that take advantage of such a combination
  • 3. Computer Science & Information Technology ( CS & IT ) 351 Figure 1. An Overview of the System 3. PROPOSED METHOD: DIBNNFC Suppose that a set of n-bit training input vectors is given and a binary desired output is assigned to each training input vector. An n-bit input vector can be considered as a vertex of an n- dimensional hypercube. The two classes of training input vectors (i.e., vertices) can be separated by an (n - 1)-dimensional hyperplane which is expressed as a net function, net (X, T) = where the wi’s and T are constant. The set of training inputs is said to be linearly separable (LS), and the (n - 1)-dimensional hyperplane is the separating hyperplane. The (n - 1)-dimensional separating hyperplanes can be established by an n-input neuron with a hard-limiter activation function as below, where, is output of the neuron, is connection weight between the input and neuron, input to the neuron, is threshold of neuron. For learning the hidden layer the geometrical learning is used to decompose an arbitrarily linearly inseparable function into multiple LS functions [1]. For any binary-to-binary mapping, ETL will determine the required LS functions, each of which is realized by a neuron in the hidden layer [8]. Check if there are true and false vertices are at hamming distance equal to one and If only true vertices at hamming distance less than‘d’ [7]. The hyperplane can be represented by the equation below. where Pre processing Apply ETL Classify Labelled Data Apply FCM algorithm Cluster the unlabelled data Semi labelled data
  • 4. 352 Computer Science & Information Technology ( CS & IT ) i.e. The first hyperplane is found. The SITV includes the core vertex and the true vertices which are separated by the first hyperplane [12]. Consider a function: {0, 1}n => {0, 1}. Denotes the n-bit input vector and {0, 1} denotes the desired output corresponding to the n-bit input vector. The value of f divides the 2n points of n-tuple (i.e., 2n vertices of n-cube) into two classes, those for which f is zero and those for which it is one. A function f is linearly separable if and only if there exists a hypersphere such that all true vertices lie inside or on the hypersphere and all false vertices lie outside or vice versa. Consider a reference hypersphere and an n-dimensional hypersphere which has its centre at and its radius r. = number of elements in SITV including trial vertex. where, is an element in SITV is ith bit of vertex . The point in the n-dimensional space represents the centre of gravity of all elements in SITV. The separating hyperplane can be representated as where T is a constant called threshold. That is, if there exists a separating hyperplane For learning an output layer if required separating hyperplanes are finds with only one core vertex, the weights and threshold of one output neuron are set as follows. The weight of the link from the odd-numbered hidden neuron to the output neuron is set to 1. The weight of the link from the even-numbered neuron to the output neuron is set to -1. For calculating the center of the class average method is used. Taking an average of entire region of classes, this existing center is used for clustering as an input centers in FCM algorithm. Fuzzy c-means (FCM) [9], [10], [11] is a data clustering technique in which a data set is grouped into n clusters with every data point in the dataset belonging to every cluster will have a high degree of belonging or membership to that cluster and another data point that lies far away from the center of a cluster will have a low degree of belonging or membership to that cluster. It is based on the concept of fuzzy C-partition which was introduced by Ruspini and developed Dunn and generalized by Jim Bezdek in 1981 as an improvement over earlier clustering methods. It is based on minimization of the following objective function:
  • 5. Computer Science & Information Technology ( CS & IT ) 353 where m is any real number greater than 1, is the degree of membership of xi in the cluster j, xi is the ith of d-dimensional measured data, cj is the d-dimension center of the cluster, and ||*|| is any norm expressing the similarity between any measured data and the center. The FCM algorithm is composed of following steps: Step 1: The FCM algorithm starts with a set of randomly initializing cluster membership matrix and with a set of predefined number of cluster center. Step 2: Using the randomly initialize cluster membership matrix, the cluster center are calculated. Step 3: Using the obtained cluster center, the cluster membership matrix is updated by taking the fractional distance from the point to the cluster center. The fuzzy c-means algorithm imposes a direct constraint on the fuzzy membership function i.e. the total membership for a point in sample or decision space must add to 1.The Fuzzy partitioning is carried out through an iterative optimization of the objective function shown above, with the update of membership uij and the cluster centers cj .This iteration will stop when 1 || ||k k U U ε+ − < where ε is a termination criterion between 0 and 1, whereas k are the iteration steps. This procedure converges to a local minimum or a saddle point of Jm. 3.1 Steps for proposed approach as follows: 1. Firstly select a core vertex initially 2. Calculate number of bit difference between core vertex and rest of the vertex 3. Select shortest distance vertex as final core vertex 4. Core vertex tends two SITVand rest of vertex are put in to rest 5. Calculate distance between SITV and rest vertex 6. Select trial vertex whose output is 1 from rest whose distance from core vertex is less then other vertex in rest 7. Find hyperplane equation by using this formula:- here If
  • 6. 354 Computer Science & Information Technology ( CS & IT ) where 8. Calculate tmin and fmax T = tmin + fmax / 2 9. If tmin > fmax hyperplane is exist then SITV = trial vertex, SITV removed trial vertex from rest save equation & calculate threshold else select new trial vertex until all true output vertices are covered Goto 6 until all true vertices are covered. 10. If rest contain vertex whose output is 1then convert true in to false & false into true Goto step 6. else find weight of output layer by putting the weight of the link from the odd numbered hidden neuron to the output neuron is set to 1. The weight of the link from the even-numbered neuron to the output neuron is set to -1 11. Calculate the average of the classes to find the centers of the region, and these centers are used in FCM 12. Initialize U= [uij] matrix, U (0) 13. At kth -step: calculate the centers vectors C (k) = [cj] with U (k) 14. Update U (k) , U (k+1) 15. If then STOP; otherwise return to step 13. 4. EXPERIMENT AND RESULTS The proposed approach is implemented in Matlab 7.8.0(R2009a) and applied on various data sets. Results of these experiments are summarized in Table 1. In this section the proposed DIBNNFC neural network have been tested on dataset and the results are obtained. Datasets used for training are Iris, Balance Scale, Bupa, and Wine data set. The input data sets files are pre processed, which provide require input set for the learning. The input data set files is converted into Binary data by performing Normalisation. Proposed approach determines number of iterations in obtaining cluster center. Furthermore we use Output of ETL to decide that which cluster goes to which class after classification and calculate the accuracy of the classifier.
  • 7. Computer Science & Information Technology ( CS & IT ) 355 Table 1. showing results of DIBNNFC classifier for various Datasets Dataset Number of Instances Number of Classes Number of Features Iterations Training time in seconds Classification Accuracy Iris 150 3 4 33 0.175633 96% Balance Scale 625 3 4 8 0.238572 44.55% BUPA 345 2 6 166 1.007675 58.14% Wine 178 3 13 56 0.69163 68.54% 5. CONCLUSIONS This paper, “Design and implementation of Binary Neural Network learning with fuzzy clustering (DIBNNFC)” is proposed to classify semi supervised data and train multilayer BNN for any binary-to-binary mapping. It improves significant performance as compared to the previous framework. We have shown that for any generation of binary-to-binary mapping, the proposed DIBNNFC approach always converges and finds the three-layer BNN, by automatically determining a required number of neurons in the hidden layer. Fcm done Clustering for semi labeled data and provide iterations and classification accuracy as per Input data. Any semi supervised data can be classified with this classifier. The neuron in the proposed BNN employs a hard-limiter activation function, only integer weights and integer thresholds. It leads to a higher accuracy of classification. DIBNNFC is better approach to classify semi supervised data then others. REFERENCES [1] Jung H. Kim and Sung-Kwon Park, “The Geometrical Learning of Binary Neural Networks,” IEEE T~mn. Neural Networks’, Vol. 6, No.1. Pp.237-247, Jun. 1995. [2] J. M. Zurada and W. Shen, “Sufficient condition for convergence of a relaxation algorithm in actual single-layer neural networks,” pp. 3W303, Dec. 1990 [3] D. L. Gray and A. N. Michel, “A training algorithm for binary feed forward neural network,” IEEE Trans. Neural Networks, pp. 176194, Mar. 1992 [4] N. E. Cotter, “The Stone-Weierstrass theorem and its application to neural networks,” IEEE Trans. Neural Networks, Dec. 1990. [5] Weiling Cai, Songcan Chen *, Daoqiang Zhang, “Robust fuzzy relational classifier incorp- orating the soft class labels” Received 29 July 2006; received in revised form 3 June 2007 June 2007 Available online 6 August 2007 [6] S. Oh and R. J. Mark 11, “Dispersive propagation skew effect in iteration neural networks ‘IEEE Trans. Neural Networks, pp. 16C-162, Jan. 1991 [7] D. Wang and Narendra S. Chaudhari, “A Multi-core Learning algorithm for Boolean Neural Networks”, School of Computing Engineering, Block N4-2a-32, Nanyang Technological University, Singapore 639798-2003 IEEE [8] Narendra S. Chaudhari and (Mrs) Aruna Tiwari, “Extending ETL for multi-class output”, Nanyang Technological University, Singapore 639798, Neural Information Processing (ICONIP'OZ), Vol. 4 [9] M. Setnes and R. Babuska, “Fuzzy relational classifier trained by fuzzy clustering,” IEEE Trans. Syst. Man Cybern. B, Cybern., vol. 29, no. 5,pp. 619–625, Oct. 1999. [10] Maedeh zirak Javanmard,“Fuzzy C-means algorithm and its application in case of public database of automobile property information. 2010[Online]. Available: http:// www.authorstream.com// Presentation / aSGuest 80031-738704-fuzzy-c-means-clustering.
  • 8. 356 Computer Science & Information Technology ( CS & IT ) [11] W. Pedrycz and G.Vukovich, “Fuzzy clustering with supervision,” Pattern Recognition., vol. 37, pp. 1229–1349, 2004. [12] Atsushi Yamamoto, Toshimichi Saito, “An Improved Expand-and-Truncate Learning”, EEE Dept., HOSEI University, Aug 1997 IEEE. Authors Sachin Bhandari received her Bachelor of Engineering degree in Computer Engineering from RGPV University, India in 2010. He is currently pursuing Master of Engineering in Computer Engineering from SGSITS, Indore, India. His research interests include Data mining, and Soft Computing. Dr. Aruna Tiwari She is currently working as Associate Professors in Computer Engineering Department at SGSITS Indore, India. Her research interest areas are Data mining, Computational Learning and Soft Computing.