SlideShare a Scribd company logo
INTRODUCTION TO
Advanced Machine Learning
By
R. Sanghavi
Asst Professor
CSE(DS)
MALLA REDDY ENGINEERING COLLEGE (Autonomous)
Module 1: Learning
Syllabus
Learning – Types of Machine Learning – Supervised
Learning – Unsupervised Learning- semi supervised
learning - The Brain and the Neuron – Design a Learning
System – Perspectives and Issues in Machine Learning –
Concept Learning Task – Concept Learning as Search –
Finding a Maximally Specific Hypothesis – Version
Spaces and the Candidate Elimination Algorithm
Course Outcomes:
Upon completion of the course, the students will be able to: Distinguish
between, supervised, unsupervised and semi-supervised learning
1. Apply the apt machine learning strategy for any given problem
2. Suggest supervised, unsupervised or semi-supervised learning algorithms for
any given problem
3. Design systems that uses the appropriate Trees in Probabilities Models of
machine learning
4. Modify existing machine learning algorithms to improve classification
efficiency
5. Design systems that uses the appropriate graph models of machine learning
Learning – Types of Machine Learning – Supervised Learning – Unsupervised UNIT - 1.pptx
Learning – Types of Machine Learning – Supervised Learning – Unsupervised UNIT - 1.pptx
Learning – Types of Machine Learning – Supervised Learning – Unsupervised UNIT - 1.pptx
Learning – Types of Machine Learning – Supervised Learning – Unsupervised UNIT - 1.pptx
Learning – Types of Machine Learning – Supervised Learning – Unsupervised UNIT - 1.pptx
Learning – Types of Machine Learning – Supervised Learning – Unsupervised UNIT - 1.pptx
C. Optimization
C Optimization
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Supervised Learning
 Supervised learning is an ML method in which a model learns from a
labeled dataset containing input-output pairs. Each input in the
dataset has a corresponding correct output (the label), and the
model's task is to learn the relationship between the inputs and
outputs. This enables the model to make predictions on new, unseen
data by applying the learned mapping.
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Categories of Supervised Learning
 Regression: When dealing with real-valued output variables like
"price" or "temperature," several popular Regression algorithms come
into play, such as the Simple Linear Regression Algorithm,
Multivariate Regression Algorithm, Decision Tree Algorithm, and
Lasso Regression.
 Classification: In instances where the output variable is a category,
like distinguishing between 'spam' and 'not spam' in email filtering,
several widely-used classification algorithms come into play. These
encompass the following algorithms: Random Forest, Decision Tree,
Logistic Regression, and Support Vector Machine.
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Advantages of Supervised Learning
 Effectiveness: Supervised learning can predict outcomes based on
past data.
 Simplicity: It's relatively easy to understand and implement.
 Performance Evaluation: It is easy to measure the performance of a
supervised learning model since the ground truth (labels) is known.
 Applications: Can be used in various fields like finance, healthcare,
marketing, etc.
 Feature Importance: It allows an understanding of which features are
most important in making predictions.
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Unsupervised Learning
 Unsupervised Learning is a type of ML that uses input data without
labeled responses to uncover hidden structures from the data itself.
Unlike supervised learning, where the training data includes both
input vectors and corresponding target labels, unsupervised learning
algorithms try to learn patterns and relationships directly from the
input data.
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Categories of Unsupervised Learning
 Clustering: Grouping similar instances into clusters (e.g., k-means,
hierarchical clustering). Some popular clustering algorithms are the K-
Means Clustering algorithm, Mean-shift algorithm, DBSCAN
Algorithm, Principal Component Analysis, and Independent
Component Analysis.
 Association: Discovering rules that capture interesting relationships
between variables in large databases (e.g., market basket analysis).
Some popular algorithms of Association are the Apriori Algorithm,
Eclat, and FP-growth algorithm.
 Dimensionality Reduction: Reducing the number of random variables
under consideration (e.g., PCA, t-SNE), which helps to simplify the
data without losing important information.
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Semi-supervised learning
 Semi-supervised learning is an ML approach that trains models using
a combination of a small amount of labeled data and a large amount
of unlabeled data. This method lies between supervised learning
(where all data is labeled) and unsupervised learning (where no data is
labeled). The main goal of semi-supervised learning is to leverage the
large pool of unlabeled data to understand the underlying structure
of the data better and improve learning accuracy with the limited
labeled data.
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
 Example of Semi-Supervised Learning
 A classic example of semi-supervised learning is classifying web
pages. Consider a scenario where you have a small number of web
pages manually categorized into topics like sports, news, technology,
etc., and a much larger set of uncategorized pages. Semi-supervised
learning algorithms can use the labeled pages to learn about features
indicative of each category and apply this knowledge to categorize the
unlabeled pages.
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Advantages of Semi-Supervised Learning
 Efficiency: Reduces the need for labeled data, which is often expensive
and time-consuming.
 Improved accuracy: Combining labeled and unlabeled data can often
improve learning accuracy.
 Utilizes unlabeled data: Effectively uses the abundance of available
unlabeled data.
 Versatility: Useful in scenarios where obtaining a fully labeled dataset
is impractical.
 Better generalization: This can help by learning the underlying data
distribution more effectively.
The Brain and the
Neuron
 In machine learning, neural
networks are models that
mimic the structure and
function of the human
brain to help computers
process data and solve
problems.
Working of a Neural Network
 Neural networks are complex systems that mimic
some features of the functioning of the human brain.
 It is composed of an input layer, one or more hidden
layers, and an output layer made up of layers of
artificial neurons that are coupled.
 The two stages of the basic process are called
backpropagation and forward propagation.
Working
of a
Neural
Network
Forward Propagation
• input Layer: Each feature in the input layer is represented by a node on
the network, which receives input data.
• Weights and Connections: The weight of each neuronal connection
indicates how strong the connection is. Throughout training, these
weights are changed.
• Hidden Layers: Each hidden layer neuron processes inputs by
multiplying them by weights, adding them up, and then passing them
through an activation function. By doing this, non-linearity is introduced,
enabling the network to recognize intricate patterns.
• Output: The final result is produced by repeating the process until the
output layer is reached.
Design a Learning System in Machine Learning
 According to Arthur Samuel “Machine Learning enables a Machine to
Automatically learn from Data, Improve performance from an
Experience and predict things without explicitly programmed.”
 In Simple Words, When we fed the Training Data to Machine Learning
Algorithm, this algorithm will produce a mathematical model and with
the help of the mathematical model, the machine will make a prediction
and take a decision without being explicitly programmed. Also, during
training data, the more machine will work with it. The more it will
get experience and the more efficient result is produced.
Design a Learning System in Machine Learning
Design a Learning System in Machine Learning
 Example : In Driverless Car, the training data is fed to Algorithm
like how to Drive Car in Highway, Busy and Narrow Street with
factors like speed limit, parking, stop at signal etc.
 After that, a Logical and Mathematical model is created on the basis
of that and after that, the car will work according to the logical
model.
 Also, the more data the data is fed the more efficient output is
produced
Design a Learning System in Machine Learning
 According to Tom Mitchell, “A computer program is said to be
learning from experience (E), with respect to some task (T). Thus,
the performance measure (P) is the performance at task T, which is
measured by P, and it improves with experience E.”
 Example: In Spam E-Mail detection,
• Task, T: To classify mails into Spam or Not Spam.
• Performance measure, P: Total percent of mails being correctly
classified as being “Spam” or “Not Spam”.
• Experience, E: Set of Mails with label “Spam”
Design a Learning System in Machine Learning
 Steps for Designing Learning System are: Step 4 choosing a learning algorithm
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Step 1) Choosing the Training Experience
 The very important and first task is to choose the training data or
training experience which will be fed to the Machine Learning
Algorithm. It is important to note that the data or experience that we
fed to the algorithm must have a significant impact on the Success or
Failure of the Model. So Training data or experience should be chosen
wisely.
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Step 2- Choosing target function
 The next important step is choosing the target function. It means
according to the knowledge fed to the algorithm the machine learning
will choose NextMove function which will describe what type of legal
moves should be taken. For example : While playing chess with the
opponent, when opponent will play then the machine learning
algorithm will decide what be the number of possible legal moves
taken in order to get success.
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Step 3- Choosing Representation for Target function
 When the machine algorithm will know all the possible legal moves
the next step is to choose the optimized move using any
representation i.e. using linear Equations, Hierarchical Graph
Representation, Tabular form etc. The NextMove function will move
the Target move like out of these move which will provide more
success rate. For Example : while playing chess machine have 4
possible moves, so the machine will choose that optimized move
which will provide success to it.
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Step 4- Choosing Function Approximation Algorithm
 An optimized move cannot be chosen just with the training data. The
training data had to go through with set of example and through
these examples the training data will approximates which steps are
chosen and after that machine will provide feedback on it. For
Example : When a training data of Playing chess is fed to algorithm so
at that time it is not machine algorithm will fail or get success and
again from that failure or success it will measure while next move
what step should be chosen and what is its success rate.
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Step 5- Final Design
 The final design is created at last when system goes from number of
examples , failures and success , correct and incorrect decision and
what will be the next step etc. Example: DeepBlue is an intelligent
computer which is ML-based won chess game against the chess expert
Garry Kasparov, and it became the first computer which had beaten a
human chess expert.
Perspectives and Issues in Machine Learning
 Perspective in Machine Learning:
• A crucial aspect of machine learning is considering different viewpoints
and approaches. These perspectives influence model development,
training, and application.
• Diverse viewpoints lead to improved model accuracy, fairness, and
generalization.
• Human bias and ethical considerations play a significant role in
shaping machine learning outcomes.
• Interdisciplinary collaboration enhances the field by incorporating i
nsights from various domains
Issues in Machine Learning
• Bias and Fairness: Models can inherit biases from training data, leading to
unfair predictions. Addressing bias is essential for ethical and equitable ML.
• Data Quality: Garbage in, garbage out! High-quality data is crucial for robust
models.
• Interpretability: Understanding why a model makes certain predictions is
challenging.
• Scalability: Handling large datasets and distributed learning remains a challenge.
• Generalization: Ensuring models generalize well to unseen examples.
• Security and Privacy: Protecting sensitive information in ML systems.
• Model Selection: Choosing the right algorithm and hyperparameters.
AML
Overfitting: Models may perform well on training data but poorly on unseen data due
to overfitting.
Methods to reduce overfitting:
 Increase training data in a dataset.
 Reduce model complexity by simplifying the model by selecting one with fewer
parameters
 Ridge Regularization and Lasso Regularization
 Early stopping during the training phase
 Reduce the noise
 Reduce the number of attributes in training data.
 Constraining the model.
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Methods to reduce Underfitting:
 Increase model complexity
 Remove noise from the data
 Trained on increased and better features
 Reduce the constraints
 Increase the number of epochs to get better results.
Concept Learning Task
 Concept learning in machine learning refers to the
process of inferring a general rule or concept from
specific examples.
 It's a fundamental aspect of supervised learning, where
the goal is to classify unseen instances based on a set of
training data.
 The training data includes labeled examples, with each
example consisting of an input (features) and the correct
output (label).
Key Aspects of Concept Learning:
1. Hypothesis Space (H):The set of all possible concepts (or rules) that the
learning algorithm could potentially choose to explain the training data.
A concept is a function that maps inputs to outputs (e.g., binary
classification, where the concept assigns a positive or negative label to
each input).
2. Target Concept (C):The actual, unknown function or rule that perfectly
explains the training data. The goal of concept learning is to find a
hypothesis in the hypothesis space that closely matches this target
concept.
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Concept Learning as Search
 Concept learning can be viewed as the task of searching through a
large space of hypothesis implicitly defined by the hypothesis
representation. The goal of the concept learning search is to find the
hypothesis that best fits the training examples.
Key Aspects of Concept Learning:
 Generalization vs. Specialization: The learning algorithm explores the
hypothesis space by generalizing from the specific examples it has seen.
The challenge is to find a hypothesis that is specific enough to explain
the training data but general enough to classify unseen examples
correctly.
 Inductive Bias: This refers to the assumptions the learning algorithm
makes to choose between different hypotheses when multiple hypotheses
are consistent with the training data. These biases are necessary to
generalize from limited data.
Examples of Concept Learning Algorithms:
 Version Space Learning: The version space is the set of all hypotheses that
are consistent with the training data. As the algorithm receives more examples,
it refines the version space by eliminating hypotheses that do not explain the
data.
 Decision Trees: A decision tree is a form of concept learning where the
internal nodes represent tests on features, and the leaves represent the output
labels (concepts). The tree generalizes a rule by recursively partitioning the
data based on feature values.
 Inductive Logic Programming (ILP):This is a type of concept learning that
uses logic-based representations to induce hypotheses. ILP combines machine
learning with symbolic reasoning to learn concepts expressed as logical rules
Applications of Concept Learning:
 Spam Detection: Learning the concept of "spam" from labeled
email data to classify future emails.
 Medical Diagnosis: Learning to identify a disease (the
concept) from patient data to predict future diagnoses.
 Image Classification: Learning to recognize objects or scenes
by identifying visual patterns from labeled images.
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Find S Algorithm
 It is a basic concept learning algorithm in machine learning.
 it finds the most specific hypothesis that fits all the positive examples.
 We have to note here that the algorithm considers only those positive training
example.
 The find-S algorithm starts with the most specific hypothesis and generalizes
this hypothesis each time it fails to classify an observed positive training data.
 Hence, the Find-S algorithm moves from the most specific hypothesis to the
most general hypothesis.
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Important Representation :
 ? indicates that any value is acceptable for the attribute.
 specify a single required value ( e.g., Cold ) for the attribute.
 ϕ indicates that no value is acceptable.
 The most general hypothesis is represented by: {?, ?, ?, ?, ?, ?}
 The most specific hypothesis is represented by: { , , , , , }
ϕ ϕ ϕ ϕ ϕ ϕ
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Steps Involved In Find-S :
Start with the most specific hypothesis.
 h = { , , , , , }
ϕ ϕ ϕ ϕ ϕ ϕ
 Take the next example and if it is negative, then no changes occur to the
hypothesis.
 If the example is positive and we find that our initial hypothesis is too
specific then we update our current hypothesis to a general condition.
 Keep repeating the above steps till all the training examples are complete.
 After we have completed all the training examples we will have the final
hypothesis when can use to classify the new examples.
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Algorithm
1. Initialize h to the most specific hypothesis in H
2. For each positive training instance x
For each attribute constraint a, in h
If the constraint a, is satisfied by x
Then do nothing
Else replace a, in h by the next more general constraint that is
satisfied by x
3. Output hypothesis h
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Example problem
Consider the following data set
having the data about which
particular seeds are poisonous.
First, we consider the hypothesis
to be a more specific hypothesis.
Hence, our hypothesis would be :
h = { , , , , , }
ϕ ϕ ϕ ϕ ϕ ϕ
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Consider example 1 :
 The data in example 1 is { GREEN, HARD, NO, WRINKLED }. We see that our initial hypothesis
is more specific and we have to generalize it for this example. Hence, the hypothesis
becomes :
 h = { GREEN, HARD, NO, WRINKLED }
Consider example 2 :
 Here we see that this example has a negative outcome. Hence we neglect this example and
our hypothesis remains the same.
 h = { GREEN, HARD, NO, WRINKLED }
Consider example 3 :
 Here we see that this example has a negative outcome. Hence we neglect this example and
our hypothesis remains the same.
 h = { GREEN, HARD, NO, WRINKLED }
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Consider example 4 :
The data present in example 4 is { ORANGE, HARD, NO, WRINKLED }. We compare every single
attribute with the initial data and if any mismatch is found we replace that particular attribute
with a general case ( ” ? ” ). After doing the process the hypothesis becomes :
 h = { ?, HARD, NO, WRINKLED }
Consider example 5 : The data present in example 5 is { GREEN, SOFT, YES, SMOOTH }. We
compare every single attribute with the initial data and if any mismatch is found we replace that
particular attribute with a general case ( ” ? ” ). After doing the process the hypothesis becomes :
 h = { ?, ?, ?, ? }
Since we have reached a point where all the attributes in our hypothesis have the general
condition, example 6 and example 7 would result in the same hypothesizes with all general
attributes.
 h = { ?, ?, ?, ? }
Hence, for the given data the final hypothesis would be :
 Final Hyposthesis: h = { ?, ?, ?, ? }
Candidate Elimination Algorithm
 FIND-S outputs a hypothesis from H, that is consistent with the
training examples, this is just one of many hypotheses from H that
might fit the training data equally well. •
 The key idea in the Candidate-Elimination algorithm is to output a
description of the set of all hypotheses consistent with the training
examples.
 Candidate-Elimination algorithm computes the description of this
set without explicitly enumerating all of its members. – This is
accomplished by using the more-general-than partial ordering and
maintaining a compact representation of the set of consistent
hypotheses.
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
 You can consider this as an extended form of the Find-S algorithm.
 Consider both positive and negative examples.
 Actually, positive examples are used here as the Find-S algorithm
(Basically they are generalizing from the specification).
 While the negative example is specified in the generalizing form.
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Algorithm
 Step1: Load Data set
 Step2: Initialize General Hypothesis and Specific Hypothesis.
 Step3: For each training example
 Step4: If example is positive example
 if attribute_value == hypothesis_value:
 Do nothing
 else:
 replace attribute value with '?' (Basically generalizing it)
 Step5: If example is Negative example
 Make generalize hypothesis more specific.
 finally, Remove hypotheses from G that are more specific than S.
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Terms Used:
 Concept learning: Concept learning is basically the learning task of the machine
(Learn by Train data)
 General Hypothesis: Not Specifying features to learn the machine.
 G = {‘?’, ‘?’,’?’,’?’…}: Number of attributes
 Specific Hypothesis: Specifying features to learn machine (Specific feature)
 S= {‘pi’,’pi’,’pi’…}: The number of pi depends on a number of attributes.
Version Space
It is an intermediate of general hypothesis and Specific hypothesis. It not only just
writes one hypothesis but a set of all possible hypotheses based on training data-set.
(or)
It is the set of all hypotheses consistent with training examples. CEA refines this space
to find the best fit.
Candidate Elimination Algorithm
Learning – Types of Machine Learning – Supervised Learning – Unsupervised UNIT - 1.pptx
Example Version Space
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Example 2
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
eg
 Initially : G = [[?, ?, ?, ?, ?, ?], [?, ?, ?, ?, ?, ?], [?, ?, ?, ?, ?, ?],
 [?, ?, ?, ?, ?, ?], [?, ?, ?, ?, ?, ?], [?, ?, ?, ?, ?, ?]]
 S = [Null, Null, Null, Null, Null, Null]

 For instance 1 : <'sunny','warm','normal','strong','warm ','same'> and positive output.
 G1 = G
 S1 = ['sunny','warm','normal','strong','warm ','same']

 For instance 2 : <'sunny','warm','high','strong','warm ','same'> and positive output.
 G2 = G
 S2 = ['sunny','warm',?,'strong','warm ','same']
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
 For instance 3 : <'rainy','cold','high','strong','warm ','change'> and negative output.
G3 = [['sunny', ?, ?, ?, ?, ?], [?, 'warm', ?, ?, ?, ?], [?, ?, ?, ?, ?, ?],
[?, ?, ?, ?, ?, ?], [?, ?, ?, ?, ?, ?], [?, ?, ?, ?, ?, 'same']]
S3 = S2
 For instance 4 : <'sunny','warm','high','strong','cool','change'> and positive output.
G4 = G3
S4 = ['sunny','warm',?,'strong', ?, ?]
At last, by synchronizing the G4 and S4 algorithm produce the output.
G = [['sunny', ?, ?, ?, ?, ?], [?, 'warm', ?, ?, ?, ?]]
S = ['sunny','warm',?,'strong', ?, ?]
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Advantages of CEA over Find-S:
 Improved accuracy: CEA considers both positive and negative examples to generate
the hypothesis, which can result in higher accuracy when dealing with noisy or
incomplete data.
 Flexibility: CEA can handle more complex classification tasks, such as those with
multiple classes or non-linear decision boundaries.
 More efficient: CEA reduces the number of hypotheses by generating a set of
general hypotheses and then eliminating them one by one. This can result in faster
processing and improved efficiency.
 Better handling of continuous attributes: CEA can handle continuous attributes by
creating boundaries for each attribute, which makes it more suitable for a wider
range of datasets.
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Disadvantages of CEA in comparison with Find-S
 More complex: CEA is a more complex algorithm than Find-S, which may make it
more difficult for beginners or those without a strong background in machine
learning to use and understand.
 Higher memory requirements: CEA requires more memory to store the set of
hypotheses and boundaries, which may make it less suitable for memory-
constrained environments.
 Slower processing for large datasets: CEA may become slower for larger datasets
due to the increased number of hypotheses generated.
 Higher potential for overfitting: The increased complexity of CEA may make it more
prone to overfitting on the training data, especially if the dataset is small or has a
high degree of noise.
Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)
Reference
 https://www.simplilearn.com/tutorials/machine-learning-tutorial/typ
es-of-machine-learning
 https://www.geeksforgeeks.org/design-a-learning-system-in-machine-l
earning/

More Related Content

PPTX
Graph representation
PPTX
Breadth First Search & Depth First Search
PDF
Minimum spanning tree
PPT
15. Transactions in DBMS
PPT
Spanning trees
PPTX
Graphs in data structure
PPTX
Graph Traversal Algorithm
DOC
Generalized transition graphs
Graph representation
Breadth First Search & Depth First Search
Minimum spanning tree
15. Transactions in DBMS
Spanning trees
Graphs in data structure
Graph Traversal Algorithm
Generalized transition graphs

What's hot (20)

PDF
UNIT III NON LINEAR DATA STRUCTURES – TREES
PDF
Lecture Note-1: Algorithm and Its Properties
PPTX
Dijkstra's Algorithm
PPTX
prim's and kruskal's algorithm
PPTX
Naïve Bayes Classifier Algorithm.pptx
PDF
Measures of query cost
PDF
Turing machines
PPTX
serializability in dbms
PPTX
Tree and graph
PPTX
Implementation of lexical analyser
PPT
Game playing (tic tac-toe), andor graph
PPTX
Top down parsing
PPTX
Decision Tree Learning
PPTX
Bfs and Dfs
PPTX
1.8. equivalence of finite automaton and regular expressions
PPTX
Analysis vs reporting
PPTX
Ensemble learning
PDF
Algorithms Lecture 7: Graph Algorithms
PPTX
Data structure power point presentation
PPT
recursive transition_networks
UNIT III NON LINEAR DATA STRUCTURES – TREES
Lecture Note-1: Algorithm and Its Properties
Dijkstra's Algorithm
prim's and kruskal's algorithm
Naïve Bayes Classifier Algorithm.pptx
Measures of query cost
Turing machines
serializability in dbms
Tree and graph
Implementation of lexical analyser
Game playing (tic tac-toe), andor graph
Top down parsing
Decision Tree Learning
Bfs and Dfs
1.8. equivalence of finite automaton and regular expressions
Analysis vs reporting
Ensemble learning
Algorithms Lecture 7: Graph Algorithms
Data structure power point presentation
recursive transition_networks
Ad

Similar to Learning – Types of Machine Learning – Supervised Learning – Unsupervised UNIT - 1.pptx (20)

PPTX
ML-Chapter_one.pptx
PPTX
Intro/Overview on Machine Learning Presentation
PDF
Introduction Machine Learning Syllabus
PPTX
Machine learning basics
PPTX
Machine learning basics by akanksha bali
PDF
machinelearning_slide note this is repdf
PPTX
ML PPT-1.pptx
PDF
Introduction AI ML& Mathematicals of ML.pdf
PPTX
3171617_introduction_applied machine learning.pptx
PPTX
introduction to machin learning
PPTX
i2ml3e-chap1.pptx
PPT
Introduction to Machine Learning
PDF
Week 1.pdf
PPT
Introduction to MACHINE LEARNING for beginners.ppt
PPTX
WEEK 4 - Beginning With Machine Learning_020418.pptx
PDF
Machine learning
PDF
27332020002_PC-CS601_Robotics_Debjit Doira.pdf
PPTX
Machine Learning an Exploratory Tool: Key Concepts
PPTX
Lecture 1.pptxgggggggggggggggggggggggggggggggggggggggggggg
PPTX
Unit - 1 - Introduction of the machine learning
ML-Chapter_one.pptx
Intro/Overview on Machine Learning Presentation
Introduction Machine Learning Syllabus
Machine learning basics
Machine learning basics by akanksha bali
machinelearning_slide note this is repdf
ML PPT-1.pptx
Introduction AI ML& Mathematicals of ML.pdf
3171617_introduction_applied machine learning.pptx
introduction to machin learning
i2ml3e-chap1.pptx
Introduction to Machine Learning
Week 1.pdf
Introduction to MACHINE LEARNING for beginners.ppt
WEEK 4 - Beginning With Machine Learning_020418.pptx
Machine learning
27332020002_PC-CS601_Robotics_Debjit Doira.pdf
Machine Learning an Exploratory Tool: Key Concepts
Lecture 1.pptxgggggggggggggggggggggggggggggggggggggggggggg
Unit - 1 - Introduction of the machine learning
Ad

More from 23Q95A6706 (8)

PPTX
R22 Machine learning jntuh UNIT- 5.pptx
PPTX
house price prediction w2_batc_4[1].pptx
PPTX
PPT Depression Detection from Text, Image & Speech using Deep Learning Algori...
PDF
evolution of machine learning algorithms for the detection.pdf
PPTX
BHAVANI internet of things internet .pptx
PPTX
DATA MINING OF DATA 1 DATA MINING OF DATA .pptx
PPTX
The Apache Hadoop software library is a framework that allows for the distrib...
PPTX
IoT- Evolution of Internet of Things, Enabling. Technologies, M2M Communicati...
R22 Machine learning jntuh UNIT- 5.pptx
house price prediction w2_batc_4[1].pptx
PPT Depression Detection from Text, Image & Speech using Deep Learning Algori...
evolution of machine learning algorithms for the detection.pdf
BHAVANI internet of things internet .pptx
DATA MINING OF DATA 1 DATA MINING OF DATA .pptx
The Apache Hadoop software library is a framework that allows for the distrib...
IoT- Evolution of Internet of Things, Enabling. Technologies, M2M Communicati...

Recently uploaded (20)

PDF
Well-logging-methods_new................
PDF
Enhancing Cyber Defense Against Zero-Day Attacks using Ensemble Neural Networks
PDF
BIO-INSPIRED HORMONAL MODULATION AND ADAPTIVE ORCHESTRATION IN S-AI-GPT
PPTX
M Tech Sem 1 Civil Engineering Environmental Sciences.pptx
PDF
Mohammad Mahdi Farshadian CV - Prospective PhD Student 2026
PDF
R24 SURVEYING LAB MANUAL for civil enggi
PPTX
MET 305 2019 SCHEME MODULE 2 COMPLETE.pptx
PPTX
Sustainable Sites - Green Building Construction
PDF
Mitigating Risks through Effective Management for Enhancing Organizational Pe...
PPTX
CYBER-CRIMES AND SECURITY A guide to understanding
PDF
Human-AI Collaboration: Balancing Agentic AI and Autonomy in Hybrid Systems
PPTX
Current and future trends in Computer Vision.pptx
PPTX
CARTOGRAPHY AND GEOINFORMATION VISUALIZATION chapter1 NPTE (2).pptx
PPTX
Geodesy 1.pptx...............................................
PPTX
Infosys Presentation by1.Riyan Bagwan 2.Samadhan Naiknavare 3.Gaurav Shinde 4...
PPT
Introduction, IoT Design Methodology, Case Study on IoT System for Weather Mo...
DOCX
573137875-Attendance-Management-System-original
PDF
PPT on Performance Review to get promotions
PDF
The CXO Playbook 2025 – Future-Ready Strategies for C-Suite Leaders Cerebrai...
PPTX
Foundation to blockchain - A guide to Blockchain Tech
Well-logging-methods_new................
Enhancing Cyber Defense Against Zero-Day Attacks using Ensemble Neural Networks
BIO-INSPIRED HORMONAL MODULATION AND ADAPTIVE ORCHESTRATION IN S-AI-GPT
M Tech Sem 1 Civil Engineering Environmental Sciences.pptx
Mohammad Mahdi Farshadian CV - Prospective PhD Student 2026
R24 SURVEYING LAB MANUAL for civil enggi
MET 305 2019 SCHEME MODULE 2 COMPLETE.pptx
Sustainable Sites - Green Building Construction
Mitigating Risks through Effective Management for Enhancing Organizational Pe...
CYBER-CRIMES AND SECURITY A guide to understanding
Human-AI Collaboration: Balancing Agentic AI and Autonomy in Hybrid Systems
Current and future trends in Computer Vision.pptx
CARTOGRAPHY AND GEOINFORMATION VISUALIZATION chapter1 NPTE (2).pptx
Geodesy 1.pptx...............................................
Infosys Presentation by1.Riyan Bagwan 2.Samadhan Naiknavare 3.Gaurav Shinde 4...
Introduction, IoT Design Methodology, Case Study on IoT System for Weather Mo...
573137875-Attendance-Management-System-original
PPT on Performance Review to get promotions
The CXO Playbook 2025 – Future-Ready Strategies for C-Suite Leaders Cerebrai...
Foundation to blockchain - A guide to Blockchain Tech

Learning – Types of Machine Learning – Supervised Learning – Unsupervised UNIT - 1.pptx

  • 1. INTRODUCTION TO Advanced Machine Learning By R. Sanghavi Asst Professor CSE(DS) MALLA REDDY ENGINEERING COLLEGE (Autonomous)
  • 3. Syllabus Learning – Types of Machine Learning – Supervised Learning – Unsupervised Learning- semi supervised learning - The Brain and the Neuron – Design a Learning System – Perspectives and Issues in Machine Learning – Concept Learning Task – Concept Learning as Search – Finding a Maximally Specific Hypothesis – Version Spaces and the Candidate Elimination Algorithm
  • 4. Course Outcomes: Upon completion of the course, the students will be able to: Distinguish between, supervised, unsupervised and semi-supervised learning 1. Apply the apt machine learning strategy for any given problem 2. Suggest supervised, unsupervised or semi-supervised learning algorithms for any given problem 3. Design systems that uses the appropriate Trees in Probabilities Models of machine learning 4. Modify existing machine learning algorithms to improve classification efficiency 5. Design systems that uses the appropriate graph models of machine learning
  • 12. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Supervised Learning  Supervised learning is an ML method in which a model learns from a labeled dataset containing input-output pairs. Each input in the dataset has a corresponding correct output (the label), and the model's task is to learn the relationship between the inputs and outputs. This enables the model to make predictions on new, unseen data by applying the learned mapping.
  • 13. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Categories of Supervised Learning  Regression: When dealing with real-valued output variables like "price" or "temperature," several popular Regression algorithms come into play, such as the Simple Linear Regression Algorithm, Multivariate Regression Algorithm, Decision Tree Algorithm, and Lasso Regression.  Classification: In instances where the output variable is a category, like distinguishing between 'spam' and 'not spam' in email filtering, several widely-used classification algorithms come into play. These encompass the following algorithms: Random Forest, Decision Tree, Logistic Regression, and Support Vector Machine.
  • 14. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Advantages of Supervised Learning  Effectiveness: Supervised learning can predict outcomes based on past data.  Simplicity: It's relatively easy to understand and implement.  Performance Evaluation: It is easy to measure the performance of a supervised learning model since the ground truth (labels) is known.  Applications: Can be used in various fields like finance, healthcare, marketing, etc.  Feature Importance: It allows an understanding of which features are most important in making predictions.
  • 15. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Unsupervised Learning  Unsupervised Learning is a type of ML that uses input data without labeled responses to uncover hidden structures from the data itself. Unlike supervised learning, where the training data includes both input vectors and corresponding target labels, unsupervised learning algorithms try to learn patterns and relationships directly from the input data.
  • 16. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Categories of Unsupervised Learning  Clustering: Grouping similar instances into clusters (e.g., k-means, hierarchical clustering). Some popular clustering algorithms are the K- Means Clustering algorithm, Mean-shift algorithm, DBSCAN Algorithm, Principal Component Analysis, and Independent Component Analysis.  Association: Discovering rules that capture interesting relationships between variables in large databases (e.g., market basket analysis). Some popular algorithms of Association are the Apriori Algorithm, Eclat, and FP-growth algorithm.  Dimensionality Reduction: Reducing the number of random variables under consideration (e.g., PCA, t-SNE), which helps to simplify the data without losing important information.
  • 17. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Semi-supervised learning  Semi-supervised learning is an ML approach that trains models using a combination of a small amount of labeled data and a large amount of unlabeled data. This method lies between supervised learning (where all data is labeled) and unsupervised learning (where no data is labeled). The main goal of semi-supervised learning is to leverage the large pool of unlabeled data to understand the underlying structure of the data better and improve learning accuracy with the limited labeled data.
  • 18. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)  Example of Semi-Supervised Learning  A classic example of semi-supervised learning is classifying web pages. Consider a scenario where you have a small number of web pages manually categorized into topics like sports, news, technology, etc., and a much larger set of uncategorized pages. Semi-supervised learning algorithms can use the labeled pages to learn about features indicative of each category and apply this knowledge to categorize the unlabeled pages.
  • 19. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Advantages of Semi-Supervised Learning  Efficiency: Reduces the need for labeled data, which is often expensive and time-consuming.  Improved accuracy: Combining labeled and unlabeled data can often improve learning accuracy.  Utilizes unlabeled data: Effectively uses the abundance of available unlabeled data.  Versatility: Useful in scenarios where obtaining a fully labeled dataset is impractical.  Better generalization: This can help by learning the underlying data distribution more effectively.
  • 20. The Brain and the Neuron  In machine learning, neural networks are models that mimic the structure and function of the human brain to help computers process data and solve problems.
  • 21. Working of a Neural Network  Neural networks are complex systems that mimic some features of the functioning of the human brain.  It is composed of an input layer, one or more hidden layers, and an output layer made up of layers of artificial neurons that are coupled.  The two stages of the basic process are called backpropagation and forward propagation.
  • 23. Forward Propagation • input Layer: Each feature in the input layer is represented by a node on the network, which receives input data. • Weights and Connections: The weight of each neuronal connection indicates how strong the connection is. Throughout training, these weights are changed. • Hidden Layers: Each hidden layer neuron processes inputs by multiplying them by weights, adding them up, and then passing them through an activation function. By doing this, non-linearity is introduced, enabling the network to recognize intricate patterns. • Output: The final result is produced by repeating the process until the output layer is reached.
  • 24. Design a Learning System in Machine Learning  According to Arthur Samuel “Machine Learning enables a Machine to Automatically learn from Data, Improve performance from an Experience and predict things without explicitly programmed.”  In Simple Words, When we fed the Training Data to Machine Learning Algorithm, this algorithm will produce a mathematical model and with the help of the mathematical model, the machine will make a prediction and take a decision without being explicitly programmed. Also, during training data, the more machine will work with it. The more it will get experience and the more efficient result is produced.
  • 25. Design a Learning System in Machine Learning
  • 26. Design a Learning System in Machine Learning  Example : In Driverless Car, the training data is fed to Algorithm like how to Drive Car in Highway, Busy and Narrow Street with factors like speed limit, parking, stop at signal etc.  After that, a Logical and Mathematical model is created on the basis of that and after that, the car will work according to the logical model.  Also, the more data the data is fed the more efficient output is produced
  • 27. Design a Learning System in Machine Learning  According to Tom Mitchell, “A computer program is said to be learning from experience (E), with respect to some task (T). Thus, the performance measure (P) is the performance at task T, which is measured by P, and it improves with experience E.”  Example: In Spam E-Mail detection, • Task, T: To classify mails into Spam or Not Spam. • Performance measure, P: Total percent of mails being correctly classified as being “Spam” or “Not Spam”. • Experience, E: Set of Mails with label “Spam”
  • 28. Design a Learning System in Machine Learning  Steps for Designing Learning System are: Step 4 choosing a learning algorithm
  • 29. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Step 1) Choosing the Training Experience  The very important and first task is to choose the training data or training experience which will be fed to the Machine Learning Algorithm. It is important to note that the data or experience that we fed to the algorithm must have a significant impact on the Success or Failure of the Model. So Training data or experience should be chosen wisely.
  • 30. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Step 2- Choosing target function  The next important step is choosing the target function. It means according to the knowledge fed to the algorithm the machine learning will choose NextMove function which will describe what type of legal moves should be taken. For example : While playing chess with the opponent, when opponent will play then the machine learning algorithm will decide what be the number of possible legal moves taken in order to get success.
  • 31. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Step 3- Choosing Representation for Target function  When the machine algorithm will know all the possible legal moves the next step is to choose the optimized move using any representation i.e. using linear Equations, Hierarchical Graph Representation, Tabular form etc. The NextMove function will move the Target move like out of these move which will provide more success rate. For Example : while playing chess machine have 4 possible moves, so the machine will choose that optimized move which will provide success to it.
  • 32. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Step 4- Choosing Function Approximation Algorithm  An optimized move cannot be chosen just with the training data. The training data had to go through with set of example and through these examples the training data will approximates which steps are chosen and after that machine will provide feedback on it. For Example : When a training data of Playing chess is fed to algorithm so at that time it is not machine algorithm will fail or get success and again from that failure or success it will measure while next move what step should be chosen and what is its success rate.
  • 33. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Step 5- Final Design  The final design is created at last when system goes from number of examples , failures and success , correct and incorrect decision and what will be the next step etc. Example: DeepBlue is an intelligent computer which is ML-based won chess game against the chess expert Garry Kasparov, and it became the first computer which had beaten a human chess expert.
  • 34. Perspectives and Issues in Machine Learning  Perspective in Machine Learning: • A crucial aspect of machine learning is considering different viewpoints and approaches. These perspectives influence model development, training, and application. • Diverse viewpoints lead to improved model accuracy, fairness, and generalization. • Human bias and ethical considerations play a significant role in shaping machine learning outcomes. • Interdisciplinary collaboration enhances the field by incorporating i nsights from various domains
  • 35. Issues in Machine Learning • Bias and Fairness: Models can inherit biases from training data, leading to unfair predictions. Addressing bias is essential for ethical and equitable ML. • Data Quality: Garbage in, garbage out! High-quality data is crucial for robust models. • Interpretability: Understanding why a model makes certain predictions is challenging. • Scalability: Handling large datasets and distributed learning remains a challenge. • Generalization: Ensuring models generalize well to unseen examples. • Security and Privacy: Protecting sensitive information in ML systems. • Model Selection: Choosing the right algorithm and hyperparameters.
  • 36. AML Overfitting: Models may perform well on training data but poorly on unseen data due to overfitting. Methods to reduce overfitting:  Increase training data in a dataset.  Reduce model complexity by simplifying the model by selecting one with fewer parameters  Ridge Regularization and Lasso Regularization  Early stopping during the training phase  Reduce the noise  Reduce the number of attributes in training data.  Constraining the model.
  • 37. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Methods to reduce Underfitting:  Increase model complexity  Remove noise from the data  Trained on increased and better features  Reduce the constraints  Increase the number of epochs to get better results.
  • 38. Concept Learning Task  Concept learning in machine learning refers to the process of inferring a general rule or concept from specific examples.  It's a fundamental aspect of supervised learning, where the goal is to classify unseen instances based on a set of training data.  The training data includes labeled examples, with each example consisting of an input (features) and the correct output (label).
  • 39. Key Aspects of Concept Learning: 1. Hypothesis Space (H):The set of all possible concepts (or rules) that the learning algorithm could potentially choose to explain the training data. A concept is a function that maps inputs to outputs (e.g., binary classification, where the concept assigns a positive or negative label to each input). 2. Target Concept (C):The actual, unknown function or rule that perfectly explains the training data. The goal of concept learning is to find a hypothesis in the hypothesis space that closely matches this target concept.
  • 40. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Concept Learning as Search  Concept learning can be viewed as the task of searching through a large space of hypothesis implicitly defined by the hypothesis representation. The goal of the concept learning search is to find the hypothesis that best fits the training examples.
  • 41. Key Aspects of Concept Learning:  Generalization vs. Specialization: The learning algorithm explores the hypothesis space by generalizing from the specific examples it has seen. The challenge is to find a hypothesis that is specific enough to explain the training data but general enough to classify unseen examples correctly.  Inductive Bias: This refers to the assumptions the learning algorithm makes to choose between different hypotheses when multiple hypotheses are consistent with the training data. These biases are necessary to generalize from limited data.
  • 42. Examples of Concept Learning Algorithms:  Version Space Learning: The version space is the set of all hypotheses that are consistent with the training data. As the algorithm receives more examples, it refines the version space by eliminating hypotheses that do not explain the data.  Decision Trees: A decision tree is a form of concept learning where the internal nodes represent tests on features, and the leaves represent the output labels (concepts). The tree generalizes a rule by recursively partitioning the data based on feature values.  Inductive Logic Programming (ILP):This is a type of concept learning that uses logic-based representations to induce hypotheses. ILP combines machine learning with symbolic reasoning to learn concepts expressed as logical rules
  • 43. Applications of Concept Learning:  Spam Detection: Learning the concept of "spam" from labeled email data to classify future emails.  Medical Diagnosis: Learning to identify a disease (the concept) from patient data to predict future diagnoses.  Image Classification: Learning to recognize objects or scenes by identifying visual patterns from labeled images.
  • 44. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Find S Algorithm  It is a basic concept learning algorithm in machine learning.  it finds the most specific hypothesis that fits all the positive examples.  We have to note here that the algorithm considers only those positive training example.  The find-S algorithm starts with the most specific hypothesis and generalizes this hypothesis each time it fails to classify an observed positive training data.  Hence, the Find-S algorithm moves from the most specific hypothesis to the most general hypothesis.
  • 45. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Important Representation :  ? indicates that any value is acceptable for the attribute.  specify a single required value ( e.g., Cold ) for the attribute.  ϕ indicates that no value is acceptable.  The most general hypothesis is represented by: {?, ?, ?, ?, ?, ?}  The most specific hypothesis is represented by: { , , , , , } ϕ ϕ ϕ ϕ ϕ ϕ
  • 46. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Steps Involved In Find-S : Start with the most specific hypothesis.  h = { , , , , , } ϕ ϕ ϕ ϕ ϕ ϕ  Take the next example and if it is negative, then no changes occur to the hypothesis.  If the example is positive and we find that our initial hypothesis is too specific then we update our current hypothesis to a general condition.  Keep repeating the above steps till all the training examples are complete.  After we have completed all the training examples we will have the final hypothesis when can use to classify the new examples.
  • 47. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Algorithm 1. Initialize h to the most specific hypothesis in H 2. For each positive training instance x For each attribute constraint a, in h If the constraint a, is satisfied by x Then do nothing Else replace a, in h by the next more general constraint that is satisfied by x 3. Output hypothesis h
  • 48. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Example problem Consider the following data set having the data about which particular seeds are poisonous. First, we consider the hypothesis to be a more specific hypothesis. Hence, our hypothesis would be : h = { , , , , , } ϕ ϕ ϕ ϕ ϕ ϕ
  • 49. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Consider example 1 :  The data in example 1 is { GREEN, HARD, NO, WRINKLED }. We see that our initial hypothesis is more specific and we have to generalize it for this example. Hence, the hypothesis becomes :  h = { GREEN, HARD, NO, WRINKLED } Consider example 2 :  Here we see that this example has a negative outcome. Hence we neglect this example and our hypothesis remains the same.  h = { GREEN, HARD, NO, WRINKLED } Consider example 3 :  Here we see that this example has a negative outcome. Hence we neglect this example and our hypothesis remains the same.  h = { GREEN, HARD, NO, WRINKLED }
  • 50. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Consider example 4 : The data present in example 4 is { ORANGE, HARD, NO, WRINKLED }. We compare every single attribute with the initial data and if any mismatch is found we replace that particular attribute with a general case ( ” ? ” ). After doing the process the hypothesis becomes :  h = { ?, HARD, NO, WRINKLED } Consider example 5 : The data present in example 5 is { GREEN, SOFT, YES, SMOOTH }. We compare every single attribute with the initial data and if any mismatch is found we replace that particular attribute with a general case ( ” ? ” ). After doing the process the hypothesis becomes :  h = { ?, ?, ?, ? } Since we have reached a point where all the attributes in our hypothesis have the general condition, example 6 and example 7 would result in the same hypothesizes with all general attributes.  h = { ?, ?, ?, ? } Hence, for the given data the final hypothesis would be :  Final Hyposthesis: h = { ?, ?, ?, ? }
  • 51. Candidate Elimination Algorithm  FIND-S outputs a hypothesis from H, that is consistent with the training examples, this is just one of many hypotheses from H that might fit the training data equally well. •  The key idea in the Candidate-Elimination algorithm is to output a description of the set of all hypotheses consistent with the training examples.  Candidate-Elimination algorithm computes the description of this set without explicitly enumerating all of its members. – This is accomplished by using the more-general-than partial ordering and maintaining a compact representation of the set of consistent hypotheses.
  • 52. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)  You can consider this as an extended form of the Find-S algorithm.  Consider both positive and negative examples.  Actually, positive examples are used here as the Find-S algorithm (Basically they are generalizing from the specification).  While the negative example is specified in the generalizing form.
  • 53. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Algorithm  Step1: Load Data set  Step2: Initialize General Hypothesis and Specific Hypothesis.  Step3: For each training example  Step4: If example is positive example  if attribute_value == hypothesis_value:  Do nothing  else:  replace attribute value with '?' (Basically generalizing it)  Step5: If example is Negative example  Make generalize hypothesis more specific.  finally, Remove hypotheses from G that are more specific than S.
  • 54. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Terms Used:  Concept learning: Concept learning is basically the learning task of the machine (Learn by Train data)  General Hypothesis: Not Specifying features to learn the machine.  G = {‘?’, ‘?’,’?’,’?’…}: Number of attributes  Specific Hypothesis: Specifying features to learn machine (Specific feature)  S= {‘pi’,’pi’,’pi’…}: The number of pi depends on a number of attributes. Version Space It is an intermediate of general hypothesis and Specific hypothesis. It not only just writes one hypothesis but a set of all possible hypotheses based on training data-set. (or) It is the set of all hypotheses consistent with training examples. CEA refines this space to find the best fit.
  • 58. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Example 2
  • 59. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) eg  Initially : G = [[?, ?, ?, ?, ?, ?], [?, ?, ?, ?, ?, ?], [?, ?, ?, ?, ?, ?],  [?, ?, ?, ?, ?, ?], [?, ?, ?, ?, ?, ?], [?, ?, ?, ?, ?, ?]]  S = [Null, Null, Null, Null, Null, Null]   For instance 1 : <'sunny','warm','normal','strong','warm ','same'> and positive output.  G1 = G  S1 = ['sunny','warm','normal','strong','warm ','same']   For instance 2 : <'sunny','warm','high','strong','warm ','same'> and positive output.  G2 = G  S2 = ['sunny','warm',?,'strong','warm ','same']
  • 60. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1)  For instance 3 : <'rainy','cold','high','strong','warm ','change'> and negative output. G3 = [['sunny', ?, ?, ?, ?, ?], [?, 'warm', ?, ?, ?, ?], [?, ?, ?, ?, ?, ?], [?, ?, ?, ?, ?, ?], [?, ?, ?, ?, ?, ?], [?, ?, ?, ?, ?, 'same']] S3 = S2  For instance 4 : <'sunny','warm','high','strong','cool','change'> and positive output. G4 = G3 S4 = ['sunny','warm',?,'strong', ?, ?] At last, by synchronizing the G4 and S4 algorithm produce the output. G = [['sunny', ?, ?, ?, ?, ?], [?, 'warm', ?, ?, ?, ?]] S = ['sunny','warm',?,'strong', ?, ?]
  • 61. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Advantages of CEA over Find-S:  Improved accuracy: CEA considers both positive and negative examples to generate the hypothesis, which can result in higher accuracy when dealing with noisy or incomplete data.  Flexibility: CEA can handle more complex classification tasks, such as those with multiple classes or non-linear decision boundaries.  More efficient: CEA reduces the number of hypotheses by generating a set of general hypotheses and then eliminating them one by one. This can result in faster processing and improved efficiency.  Better handling of continuous attributes: CEA can handle continuous attributes by creating boundaries for each attribute, which makes it more suitable for a wider range of datasets.
  • 62. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Disadvantages of CEA in comparison with Find-S  More complex: CEA is a more complex algorithm than Find-S, which may make it more difficult for beginners or those without a strong background in machine learning to use and understand.  Higher memory requirements: CEA requires more memory to store the set of hypotheses and boundaries, which may make it less suitable for memory- constrained environments.  Slower processing for large datasets: CEA may become slower for larger datasets due to the increased number of hypotheses generated.  Higher potential for overfitting: The increased complexity of CEA may make it more prone to overfitting on the training data, especially if the dataset is small or has a high degree of noise.
  • 63. Lecture Notes for E Alpaydın 2004 Introduction to Machine Learning © The MIT Press (V1.1) Reference  https://www.simplilearn.com/tutorials/machine-learning-tutorial/typ es-of-machine-learning  https://www.geeksforgeeks.org/design-a-learning-system-in-machine-l earning/