SlideShare a Scribd company logo
2
Most read
5
Most read
6
Most read
M.Rajshree
M.SC(IT)
Nadar saraswathi college of
arts&science
Lazy learners
 lazy learning is a learning method in which
generalization of the training data is, in
theory, delayed until a query is made to the
system, as opposed to in eager learning,
where the system tries to generalize the
training data before receiving queries.
 Lazy learners do less work while training
data is given and more work when
classification of a test tuple is given.
 The classification methods discussed so far
in this chapter—decision tree induction,
Bayesian classification, rule-based
classification, classification by
backpropagation, support vector machines,
and classification based on association rule
mining—are all examples of eager learners
 A lazy learner simply stores the training data
and only when it sees a test tuple starts
generalization to classify the tuple based on
its similarity to the stored training tuple
 Building a model from a given set of training
data
 Applying the model to a given set of testing
data
 Eager Learners like Bayesian Classification,
Rule-based classification, support vector
machines, etc. will construct a classification
model before receiving new tuple when a set
of training tuple is given
k-Nearest-Neighbor
Classifiers
 The k-nearest-neighbor method was first
described in the early 1950s.
 Nearest-neighbor classifiers are based on
learning by analogy, that is, by comparing a
given test tuple with training tuples that are
similar to it.
 The training tuples are described
by n attributes. Each tuple represents a point
in an n-dimensional space.
 In this way, all of the training tuples are
stored in an n-dimensional pattern space.
When given an unknown tuple, a k-nearest-
neighbor classifier searches the pattern
space for the k training tuples that are
closest to the unknown tuple
 distance between two points or tuples,
say, X1 = (x11, x12…. x1n) and X2 =
(x21, x22…x2n)
 When given a test tuple, a k-nearest
neighbor classifier searches the pattern
space for the k training tuples that are
closest to the test tuple.
 These k training tuples are the k “nearest
neighbors” of the test tuple
lazy learners and other classication methods
Case-Based Reasoning
 Base-based reasoning is the process of
solving new problems based on the solutions
of similar past problems.
 These classifiers use a database of problem
solutions to solve new problems.
 The case-based reasoner tries to combine
the solutions of the neighboring training
cases in order to propose a solution for the
new case
 Case-based reasoning (CBR) classifiers use
a database of problem solutions to solve new
problems.
 Unlike nearest-neighbor classifiers, which
store training tuples as points in Euclidean
space, CBR stores the tuples or cases‖ for
problem solving as complex symbolic
descriptions.
 Business applications of CBR include
problem resolution for customer service help
desks, where cases describe product-related
diagnostic problems.
 CBR has also been applied to areas such as
engineering and law, where cases are either
technical designs or legal rulings, respectively.
 Medical education is another area for CBR,
where patient case histories and treatments are
used to help diagnose and treat new patients.
 The case-based reasoner tries to combine the
solutions of the neighboring training cases in
order to propose a solution for the new case.
 The case-based reasoner may employ
background knowledge and problem-solving
strategies in order to propose a feasible
combined solution.
Other classification methods
 Data mining involves six common classes of
tasks. Anomaly detection, Association rule
learning, Clustering, Classification,
Regression,
Summarization. Classification is a
major technique in data mining and widely
used in various fields.
 Classification is a technique where we
categorize data into a given number of
classes
 Binary Classification: Classification task
with two possible outcomes Eg: Gender
classification (Male / Female)
 Multi class classification: Classification
with more than two classes. In multi class
classification each sample is assigned to one
and only one target label Eg: An animal can
be cat or dog but not both at the same time
 Multi label classification: Classification
task where each sample is mapped to a set
of target labels (more than one class). Eg: A
news article can be about sports, a person,
and location at the same time.
Naïve Bayes
 Naive Bayes algorithm based on Bayes’
theorem with the assumption of
independence between every pair of
features. Naive Bayes classifiers work well in
many real-world situations such as document
classification and spam filtering.
 This algorithm requires a small amount of
training data to estimate the necessary
parameters. Naive Bayes classifiers are
extremely fast compared to more
sophisticated methods.
Fuzzy Set Approaches
 Fuzzy Set Theory is also called Possibility
Theory. This theory was proposed by Lotfi
Zadeh in 1965 as an alternative the two-value
logic and probability theory
 This theory allows us to work at a high level of
abstraction. It also provides us the means for
dealing with imprecise measurement of data.
 fuzzy set approach an important consideration
is the treatment of data from a linguistic view
point from this has developed an approach that
uses linguistically quantified propositions to
summarize the content of a data base by
providing a general characterization of the
analyzed data

More Related Content

PDF
3D Printing
PPT
2.5 backpropagation
PPTX
3D Printing: Endless Opportunities
PDF
Data Mining: Association Rules Basics
PPTX
3D Printing Technology & Its Applications
PPTX
Decision Support System(DSS)
3D Printing
2.5 backpropagation
3D Printing: Endless Opportunities
Data Mining: Association Rules Basics
3D Printing Technology & Its Applications
Decision Support System(DSS)

What's hot (20)

PPT
Conceptual dependency
PPT
Heuristic Search Techniques Unit -II.ppt
PPTX
Homomorphic filtering
PPT
3. mining frequent patterns
PPT
3.7 outlier analysis
PPTX
Register allocation and assignment
PPT
01 Data Mining: Concepts and Techniques, 2nd ed.
PPTX
Smoothing in Digital Image Processing
PPTX
Histogram Processing
PPT
Association rule mining
PDF
I. AO* SEARCH ALGORITHM
PPTX
Inductive bias
PDF
Address in the target code in Compiler Construction
PPTX
Data mining tasks
PPTX
Semantic nets in artificial intelligence
PPTX
Data Integration and Transformation in Data mining
PPTX
Density based methods
PPT
4.2 spatial data mining
PPTX
04 Classification in Data Mining
PPT
2.4 rule based classification
Conceptual dependency
Heuristic Search Techniques Unit -II.ppt
Homomorphic filtering
3. mining frequent patterns
3.7 outlier analysis
Register allocation and assignment
01 Data Mining: Concepts and Techniques, 2nd ed.
Smoothing in Digital Image Processing
Histogram Processing
Association rule mining
I. AO* SEARCH ALGORITHM
Inductive bias
Address in the target code in Compiler Construction
Data mining tasks
Semantic nets in artificial intelligence
Data Integration and Transformation in Data mining
Density based methods
4.2 spatial data mining
04 Classification in Data Mining
2.4 rule based classification
Ad

Similar to lazy learners and other classication methods (20)

PDF
18 ijcse-01232
PDF
A novel methodology for constructing rule based naïve bayesian classifiers
PPT
activelearning.ppt
PDF
IJCSI-10-6-1-288-292
PDF
Classifiers
PPTX
MACHINE LEARNING Unit -2 Algorithm.pptx
PPTX
SVM - Functional Verification
PDF
A Decision Tree Based Classifier for Classification & Prediction of Diseases
PPTX
Presentation on Text Classification
PDF
A New Active Learning Technique Using Furthest Nearest Neighbour Criterion fo...
PPT
Machine Learning: Decision Trees Chapter 18.1-18.3
PDF
MACHINE LEARNING TOOLBOX
PPT
Classification Of Web Documents
PPT
Introduction to Machine Learning.
PDF
Data mining chapter04and5-best
PPT
UNIT2_NaiveBayes algorithms used in machine learning
PPT
Text categorization
PPTX
Mis End Term Exam Theory Concepts
PPT
Textmining Predictive Models
PPT
Textmining Predictive Models
18 ijcse-01232
A novel methodology for constructing rule based naïve bayesian classifiers
activelearning.ppt
IJCSI-10-6-1-288-292
Classifiers
MACHINE LEARNING Unit -2 Algorithm.pptx
SVM - Functional Verification
A Decision Tree Based Classifier for Classification & Prediction of Diseases
Presentation on Text Classification
A New Active Learning Technique Using Furthest Nearest Neighbour Criterion fo...
Machine Learning: Decision Trees Chapter 18.1-18.3
MACHINE LEARNING TOOLBOX
Classification Of Web Documents
Introduction to Machine Learning.
Data mining chapter04and5-best
UNIT2_NaiveBayes algorithms used in machine learning
Text categorization
Mis End Term Exam Theory Concepts
Textmining Predictive Models
Textmining Predictive Models
Ad

More from rajshreemuthiah (20)

PPTX
PPTX
PPTX
PPTX
polymorphism
PPTX
solutions and understanding text analytics
PPTX
interface
PPTX
Testing &ampdebugging
PPTX
concurrency control
PPTX
Education
PPTX
Formal verification
PPTX
Transaction management
PPTX
Multi thread
PPTX
System testing
PPTX
software maintenance
PPTX
exception handling
PPTX
e governance
PPTX
recovery management
PPTX
Implementing polymorphism
PPSX
Buffer managements
PPTX
os linux
polymorphism
solutions and understanding text analytics
interface
Testing &ampdebugging
concurrency control
Education
Formal verification
Transaction management
Multi thread
System testing
software maintenance
exception handling
e governance
recovery management
Implementing polymorphism
Buffer managements
os linux

Recently uploaded (20)

PDF
Review of recent advances in non-invasive hemoglobin estimation
PPT
“AI and Expert System Decision Support & Business Intelligence Systems”
PPTX
A Presentation on Artificial Intelligence
PPTX
Big Data Technologies - Introduction.pptx
PDF
Unlocking AI with Model Context Protocol (MCP)
PPTX
Digital-Transformation-Roadmap-for-Companies.pptx
PDF
Architecting across the Boundaries of two Complex Domains - Healthcare & Tech...
DOCX
The AUB Centre for AI in Media Proposal.docx
PDF
Machine learning based COVID-19 study performance prediction
PDF
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
PDF
Optimiser vos workloads AI/ML sur Amazon EC2 et AWS Graviton
PDF
Encapsulation_ Review paper, used for researhc scholars
PPTX
sap open course for s4hana steps from ECC to s4
PDF
Electronic commerce courselecture one. Pdf
PDF
Peak of Data & AI Encore- AI for Metadata and Smarter Workflows
PPTX
Machine Learning_overview_presentation.pptx
PDF
Dropbox Q2 2025 Financial Results & Investor Presentation
PDF
Agricultural_Statistics_at_a_Glance_2022_0.pdf
PDF
A comparative analysis of optical character recognition models for extracting...
PDF
Diabetes mellitus diagnosis method based random forest with bat algorithm
Review of recent advances in non-invasive hemoglobin estimation
“AI and Expert System Decision Support & Business Intelligence Systems”
A Presentation on Artificial Intelligence
Big Data Technologies - Introduction.pptx
Unlocking AI with Model Context Protocol (MCP)
Digital-Transformation-Roadmap-for-Companies.pptx
Architecting across the Boundaries of two Complex Domains - Healthcare & Tech...
The AUB Centre for AI in Media Proposal.docx
Machine learning based COVID-19 study performance prediction
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
Optimiser vos workloads AI/ML sur Amazon EC2 et AWS Graviton
Encapsulation_ Review paper, used for researhc scholars
sap open course for s4hana steps from ECC to s4
Electronic commerce courselecture one. Pdf
Peak of Data & AI Encore- AI for Metadata and Smarter Workflows
Machine Learning_overview_presentation.pptx
Dropbox Q2 2025 Financial Results & Investor Presentation
Agricultural_Statistics_at_a_Glance_2022_0.pdf
A comparative analysis of optical character recognition models for extracting...
Diabetes mellitus diagnosis method based random forest with bat algorithm

lazy learners and other classication methods

  • 2. Lazy learners  lazy learning is a learning method in which generalization of the training data is, in theory, delayed until a query is made to the system, as opposed to in eager learning, where the system tries to generalize the training data before receiving queries.  Lazy learners do less work while training data is given and more work when classification of a test tuple is given.
  • 3.  The classification methods discussed so far in this chapter—decision tree induction, Bayesian classification, rule-based classification, classification by backpropagation, support vector machines, and classification based on association rule mining—are all examples of eager learners  A lazy learner simply stores the training data and only when it sees a test tuple starts generalization to classify the tuple based on its similarity to the stored training tuple
  • 4.  Building a model from a given set of training data  Applying the model to a given set of testing data  Eager Learners like Bayesian Classification, Rule-based classification, support vector machines, etc. will construct a classification model before receiving new tuple when a set of training tuple is given
  • 5. k-Nearest-Neighbor Classifiers  The k-nearest-neighbor method was first described in the early 1950s.  Nearest-neighbor classifiers are based on learning by analogy, that is, by comparing a given test tuple with training tuples that are similar to it.  The training tuples are described by n attributes. Each tuple represents a point in an n-dimensional space.
  • 6.  In this way, all of the training tuples are stored in an n-dimensional pattern space. When given an unknown tuple, a k-nearest- neighbor classifier searches the pattern space for the k training tuples that are closest to the unknown tuple  distance between two points or tuples, say, X1 = (x11, x12…. x1n) and X2 = (x21, x22…x2n)  When given a test tuple, a k-nearest neighbor classifier searches the pattern space for the k training tuples that are closest to the test tuple.  These k training tuples are the k “nearest neighbors” of the test tuple
  • 8. Case-Based Reasoning  Base-based reasoning is the process of solving new problems based on the solutions of similar past problems.  These classifiers use a database of problem solutions to solve new problems.  The case-based reasoner tries to combine the solutions of the neighboring training cases in order to propose a solution for the new case
  • 9.  Case-based reasoning (CBR) classifiers use a database of problem solutions to solve new problems.  Unlike nearest-neighbor classifiers, which store training tuples as points in Euclidean space, CBR stores the tuples or cases‖ for problem solving as complex symbolic descriptions.  Business applications of CBR include problem resolution for customer service help desks, where cases describe product-related diagnostic problems.
  • 10.  CBR has also been applied to areas such as engineering and law, where cases are either technical designs or legal rulings, respectively.  Medical education is another area for CBR, where patient case histories and treatments are used to help diagnose and treat new patients.  The case-based reasoner tries to combine the solutions of the neighboring training cases in order to propose a solution for the new case.  The case-based reasoner may employ background knowledge and problem-solving strategies in order to propose a feasible combined solution.
  • 11. Other classification methods  Data mining involves six common classes of tasks. Anomaly detection, Association rule learning, Clustering, Classification, Regression, Summarization. Classification is a major technique in data mining and widely used in various fields.  Classification is a technique where we categorize data into a given number of classes
  • 12.  Binary Classification: Classification task with two possible outcomes Eg: Gender classification (Male / Female)  Multi class classification: Classification with more than two classes. In multi class classification each sample is assigned to one and only one target label Eg: An animal can be cat or dog but not both at the same time  Multi label classification: Classification task where each sample is mapped to a set of target labels (more than one class). Eg: A news article can be about sports, a person, and location at the same time.
  • 13. Naïve Bayes  Naive Bayes algorithm based on Bayes’ theorem with the assumption of independence between every pair of features. Naive Bayes classifiers work well in many real-world situations such as document classification and spam filtering.  This algorithm requires a small amount of training data to estimate the necessary parameters. Naive Bayes classifiers are extremely fast compared to more sophisticated methods.
  • 14. Fuzzy Set Approaches  Fuzzy Set Theory is also called Possibility Theory. This theory was proposed by Lotfi Zadeh in 1965 as an alternative the two-value logic and probability theory  This theory allows us to work at a high level of abstraction. It also provides us the means for dealing with imprecise measurement of data.  fuzzy set approach an important consideration is the treatment of data from a linguistic view point from this has developed an approach that uses linguistically quantified propositions to summarize the content of a data base by providing a general characterization of the analyzed data