SlideShare a Scribd company logo
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
Methods for Meta-Learning in AutoML
Learning how to Learn
1
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
Motivational Example
● Alex starts to learn a Maths course of 10 tests for
the first time in his life. (Problem)
● Alex wants to get a grade A in most of the
course tests. (Target)
● Alex thought that attending all lectures would
easily help him to get grade A like what he
always does in history courses. (Approach 1)
2
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
Motivational Example
● Alex got grade D in his first test. (Result 1)
● Alex decided to switch to reading the reference
book instead of attending all lectures only.
(Approach 2)
● Alex got grade C in his second test. (Result 2)
3
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
Motivational Example
● After that, Alex decided to switch to solving practice
problems instead. (Approach 3)
● Alex got grade B in his third test. (Result 3)
● So, Alex decided to summarize each lesson and
teach it to his colleagues too. (Approach 4)
● Alex got grade A in his fourth test. (Result 4)
4
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
Motivational Example
● Now, the question is how will Alex study for his 5th
test in the course ?
5
Alex has already learnt how to learn.
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
Back to Machine Learning
6
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
Motivation
7
Data
Collection
1. Data
Preprocessing
2. Feature
Extraction
3. Feature
Selection
4.
Algorithm
Selection
Deploym
ent
5.
Parameter
Tuning
Prediction
Real-World
Data Feature Engineering Model Building
Typical Supervised Machine Learning Pipeline
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
8
Model Building
4.
Algorithm
Selection
5.
Parameter
Tuning
Examples:
- Linear Classification: (Simple Linear Classification, Ridge, Lasso, Simple Perceptron, ….)
- Support Vector Machines
- Decision Tree (ID3, C4.5, C5.0, CART, ….)
- Nearest Neighbors
- Gaussian Processes
- Naive Bayes (Gaussian, Bernoulli, Complement, ….)
- Ensembling: (Random Forest, GBM, AdaBoost, ….)
Motivation: Model Building
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
9
Model Building
4.
Algorithm
Selection
5.
Parameter
Tuning
Kernel
Linear RBF Polynomial
Gamma
[2^-15, 2^3]
Degree
2,3,....
C - Penalty
[2^-5, 2^15]
Example: Support Vector Machine
……..
Motivation: Model Building
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
10
Motivation: Dimensionality Reduction
Examples of Feature Extraction:
1. Principal Component Analysis
2. Linear Discriminant Analysis
3. Multiple Discriminant Analysis
4. Independent Component
Analysis
Examples of Multivariate Feature Selection:
1. Relief
2. Correlation Feature Selection
3. Branch and Bound
4. Sequential Forward Selection
5. Plus L - Minus R
Examples of Univariate Feature Selection:
1. Information Gain
2. Fisher Score
3. Correlation with Target
2. Feature
Extraction
3. Feature
Selection
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
11
Motivation: Data PreProcessing
Examples of Data Preprocessors:
1. Scaling
2. Normalization
3. Standardization
4. Binarization
5. Imputation
6. Deletion
7. One-Hot-Encoding
8. Hashing
9. Discretization
1. Data
Preprocessing
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
Solution: Meta-Learning
1. Science of systematically observing how different machine learning
approaches perform on a wide range of learning tasks and then
learning from this experience.
12
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
Solution: Meta-Learning
2. It also allows to replace hand-written rules and algorithms with
novel approaches that are data-driven.
13
1. Science of systematically observing how different machine learning
approaches perform on a wide range of learning tasks and then
learning from this experience.
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
HOW ? Collect MetaData
1. Model Configurations:
- Pipeline Composition: (Normalization → PCA → SVM)
- Hyperparameter Settings: (PCA = 2 components, SVM = gamma: 1e-9, C = 1e2)
- Network Architectures: (2 Hidden Layers, 100 Neurons per layer)
2. Resulting Model Evaluations:
- Different Metrics: Accuracy, error rate, F1-Score.
- Training Time.
3. Task Itself (Meta-Features):
- Description of the data
14
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
HOW ? Use Meta-Data
1. Knowledge Transfer.
Use the same model as an initial
point and start to tune it.
2. Guided Search.
If Classifier X is worth than Classifier
Y by 10% then there is no need to
tune classifier X
15
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
HOW ? Use Meta-Data
Remember that Alex starts with the same approach that
succeeds in History courses.
Meta-Learning won’t be effective and may affect performance
badly in case of:
- Tasks with random noise, and unrelated phenomena.
“Tasks that are Never Seen Before”
16
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
Meta-Learning Methodologies:
1. Learning from Task Properties.
2. Learning from Model Evaluations.
3. Learning from Prior Models.
17
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
1-Learning from Task Properties:
● Represent task as a meta-feature vector.
● Studies show that optimal set of meta-features depends on application
type.[2]
● Different studies used various feature selection and extraction techniques
to reduce set of meta-features.[2][3]
18
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
1-Learning from Task Properties:
● What are Task Properties? = Types of Meta-features:
1. Simple
2. Statistical
3. Information Theoretic
4. Complexity
5. LandMarkers
19
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
Meta-Features Types: (Simple)
● Examples:
1. Number of Instances
2. Number of Features
3. Number of Classes
4. Number of Missing Values
5. Number of Outliers
20
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
Meta-Features Types: (Statistical)
● Examples:
1. Skewness of Numerical Features.
2. Kurtosis of Numerical Features.
3. Correlation Covariance between features.
4. Variance in first PCA.
5. Skewness and Kurtosis of first PCA.
6. Class probability distribution.
7. Concentration, Sparsity, Gravity of Features
(Measurements of independence and
dispersion of values.)
21
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
Meta-Features Types: (Information theoretic)
● Examples:
1. Class Entropy.
2. Mutual Information between feature and
Class.
3. Equivalent number of features (2/1)
4. Noise to Signal ratio.
22
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
Meta-Features Types: (Task Complexity)
● Examples:
1. Fisher discriminant (Measure separability
between classes).
23
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
Meta-Features Types: (Landmarkers)
● Examples:
1. LandMarker 1NN.
2. LandMarker Decision Tree.
3. LandMarker Naive Bayes.
4. LandMarker Linear Discriminant Analysis.
24
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
How to use Meta-Features?
25
● Different Similarity Measurements (Unsupervised) and warm starting optimization of
similar tasks for recommendation of candidate configurations:
Examples:
1. Rank of different configurations.
- Tasks A, B are twin tasks.
- SVM and KNN are the best for Task A.
- Then, SVM and KNN are the best for Task B.
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
How to use Meta-Features?
26
● Different Similarity Measurements (Unsupervised) and warm starting optimization of
similar tasks for recommendation of candidate configurations:
Examples:
2. Collaborative Filtering
Use results of few configurations on Task A to
predicts results of all other configurations based on
configurations results on a similar Task B
Knowledge Base Needs almost full
configurations results to be updated.
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
1-Learning from Task Properties:
27
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
How to use Meta-Features?
28
● Learning High Level Meta-Features
Low Level Features High Level Features
NEEDS BIG KNOWLEDGE BASE
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
How to use Meta-Features?
29
● Meta-Models (Supervised): Learn the complex relationship between
meta-features and useful configurations in this large space.
Example:
- Ranking of Top N Promising Configurations:
Literature suggests Boosting, and Bagging Models [4][5].
+
Approximate Ranking Tree Forests [6] (Auto Meta-Feature Selection
based on some initial results).
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
How to use Meta-Features?
30
● Pipeline Synthesis:
1. Meta-Model to predict which preprocessor with improve
performance of a specific classifier in that particular task. [7] [8]
2. Reinforcement Learning to construct pipeline by addition, deletion,
replacement of pipeline blocks. [9] (Alpha D3M - Evolutionary
Approach)
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
How to use Meta-Features?
31
● Tune or Not to Tune:
Meta Models to predict:
1. How much improvement we can expect from tuning this particular
classifier on that particular task [10].
2. How much improvement VS additional time investment? [11].
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
2-Learning from Model Evaluations:
● Using Current configuration evaluations as a prior to suggesting the
coming candidate outperforming configuration in an iterative way.
32
Example:
1. Evaluate Px on Task 1
2. Suggest new Ps
3. Select most candidate outperforming P
4. Set Px = P
5. GO TO 1
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
How it is used?
● Task Independent Recommendation:
1. Discretize the search space into a set of configurations.
2. Apply over many datasets.
3. Aggregate single task rankings into a global ranking.
● Example: Scikit Learn Cheat Sheet Algorithm.
33
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
34
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
How it is used?
● Search Space Design:
1. Learn hyperparameter default values (Best configuration over all
tasks).
2. Learn different hyperparameters importance:
- Measure variance of algorithm performance by keeping all
hyperparameters fixed and change only one.
35
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
How it is used?
● Learning Curves: (Example: 1. Apply SVM Over 100 Training Datasets)
36
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
How it is used?
● Learning Curves: (Example: 2. Apply SVM Over New Dataset)
37
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
How it is used?
● Learning Curves: (Example: 3. Measure Similarity between training curves
and testing curve)
38
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
How it is used?
● Configuration Transfer:
- Surrogate Models: usually suitable with Gaussian Processes Bayesian Optimization like the
SMAC algorithm.
- We can define task similarity based on Learning Distribution Similarity between tasks too.
39
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
How it is used?
● Configuration Transfer:
- Surrogate Models: usually suitable with Gaussian Processes like the SMAC algorithm.
We can define task similarity based of accuracy of predictions for
40
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
How it is used?
● Configuration Transfer:
- Multi-armed bandits:
1. Start with small data portion and apply multiple
configurations on these small portions.
2. Drop lowest performing configurations and increase
portion size for other configurations.
41
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
How it is used?
● Configuration Transfer:
- Multi-armed bandits:
42
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
3-Learning from Prior Models:
● Take already trained Models (Model HUB) to use for similar tasks.
● Suitable for few classifiers (Eg: Kernel Classifiers - Bayesian Networks)
BUT very good with Neural Networks. WHY?
Both Structure and Network Parameters can be a good initialization
for the target model.
43
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
References:
[1] Hutter Frank and Kotthoff Lars and Vanschoren Joaquin - Automated Machine Learning: Methods, Systems,
Challenges - (2019), Springer
[2] Bilalli, B., Abell´o, A., Aluja-Banet, T.: On the predictive power of metafeatures in OpenML. International Journal of
Applied Mathematics and Computer Science 27(4), 697 – 712 (2017)
[3] Todorovski, L., Brazdil, P., Soares, C.: Report on the experiments with feature selection in meta-level learning.
PKDD 2000 Workshop on Data mining, Decision support, Meta-learning and ILP pp. 27–39 (2000)
[4] Pinto, F., Cerqueira, V., Soares, C., Mendes-Moreira, J.: autoBagging: Learning to rank bagging workflows with
metalearning. arXiv 1706.09367 (2017)
[5] Lorena, A.C., Maciel, A.I., de Miranda, P.B.C., Costa, I.G., Prudˆencio, R.B.C.: Data complexity meta-features for
regression problems. Machine Learning 107(1), 209–246 (2018)
[6] Sun, Q., Pfahringer, B.: Pairwise meta-rules for better meta-learning based algorithm ranking. Machine Learning
93(1), 141–161 (2013)
[7] Bilalli, B., Abell´o, A., Aluja-Banet, T., Wrembel, R.: Intelligent assistance for data pre-processing. Computer
Standards & Interf. 57, 101 – 109 (2018)
[8] Schoenfeld, B., Giraud-Carrier, C., Poggeman, M., Christensen, J., Seppi, K.: Feature selection for high-dimensional
data: A fast correlation-based filter solution. In: AutoML Workshop at ICML (2018)
[9] Drori, I., Krishnamurthy, Y., Rampin, R., de Paula Lourenco, R., Ono, J.P., Cho, K., Silva, C., Freire, J.: AlphaD3M:
Machine learning pipeline synthesis. In: AutoML Workshop at ICML (2018)
[10] Ridd, P., Giraud-Carrier, C.: Using metalearning to predict when parameter optimization is likely to improve
classification accuracy. In: ECAI Workshop on Meta-learning and Algorithm Selection. pp. 18–23 (2014)
[11] Sanders, S., Giraud-Carrier, C.: Informing the use of hyperparameter optimization through metalearning. In: Proc.
ICDM. pp. 1051–1056 (2017) 44
Mohamed Maher - University of Tartu - 2019 - mohamed.abdelrahman@ut.ee
45

More Related Content

PDF
AutoML lectures (ACDL 2019)
PDF
Automated Machine Learning
PDF
Automatic machine learning (AutoML) 101
PDF
Micro project list dms- 22319
PPTX
MACHINE LEARNING PRESENTATION (ARTIFICIAL INTELLIGENCE)
PDF
Machine Learning: Applications, Process and Techniques
PDF
General Tips for participating Kaggle Competitions
PDF
Automatic Machine Learning, AutoML
AutoML lectures (ACDL 2019)
Automated Machine Learning
Automatic machine learning (AutoML) 101
Micro project list dms- 22319
MACHINE LEARNING PRESENTATION (ARTIFICIAL INTELLIGENCE)
Machine Learning: Applications, Process and Techniques
General Tips for participating Kaggle Competitions
Automatic Machine Learning, AutoML

What's hot (8)

PPT
Web Scraping and Data Extraction Service
PPTX
A Solution to Bus Bunching.pptx
PDF
AutoML - The Future of AI
PPTX
AWS Forcecast: DeepAR Predictor Time-series
PPTX
Web based online shopping system Presentation slide
PDF
Market sizing TAM SAM SOM Target Market
PPTX
Introduction to Artificial Intteligence
PDF
Scalable Automatic Machine Learning in H2O
Web Scraping and Data Extraction Service
A Solution to Bus Bunching.pptx
AutoML - The Future of AI
AWS Forcecast: DeepAR Predictor Time-series
Web based online shopping system Presentation slide
Market sizing TAM SAM SOM Target Market
Introduction to Artificial Intteligence
Scalable Automatic Machine Learning in H2O
Ad

Similar to Methods for meta learning in AutoML (20)

PPTX
Data science seminar - University of Tartu - SmartML
PPTX
Automated Machine Learning and SmartML
PDF
Scheme g third semester (co,cm,cd,if, cw)
PDF
To get the notes of python of MCA batchs
PDF
M.sc cs annual_2019-20
PDF
A SIMPLE PROCESS TO SPEED UP MACHINE LEARNING METHODS: APPLICATION TO HIDDEN ...
PDF
Te computer syllabus 2015 course 3-4-17 3-5-17
PDF
06522405
PDF
Te computer-syllabus-2015-course-3-4-17
PPTX
22PCOAM16 P_UNIT 1_ Session 2 Brain and the Neuron.pptx
PDF
Se be information technology rev 2016
PDF
Se be information technology rev 2016
PDF
[update] Introductory Parts of the Book "Dive into Deep Learning"
PPTX
Lecture 1 Introduction to Engineering Analysis.pptx
PPTX
22PCOAM16_UNIT 1_Session 8 Multi layer Perceptrons.pptx
PDF
Big Data & Text Analytics - Lesson Schedule
PPTX
META-LEARNING.pptx
PDF
4.74 s.e. computer engineering (1)
PPT
CS8082_MachineLearnigTechniques _Unit-1.ppt
PPTX
Numerical Methods
Data science seminar - University of Tartu - SmartML
Automated Machine Learning and SmartML
Scheme g third semester (co,cm,cd,if, cw)
To get the notes of python of MCA batchs
M.sc cs annual_2019-20
A SIMPLE PROCESS TO SPEED UP MACHINE LEARNING METHODS: APPLICATION TO HIDDEN ...
Te computer syllabus 2015 course 3-4-17 3-5-17
06522405
Te computer-syllabus-2015-course-3-4-17
22PCOAM16 P_UNIT 1_ Session 2 Brain and the Neuron.pptx
Se be information technology rev 2016
Se be information technology rev 2016
[update] Introductory Parts of the Book "Dive into Deep Learning"
Lecture 1 Introduction to Engineering Analysis.pptx
22PCOAM16_UNIT 1_Session 8 Multi layer Perceptrons.pptx
Big Data & Text Analytics - Lesson Schedule
META-LEARNING.pptx
4.74 s.e. computer engineering (1)
CS8082_MachineLearnigTechniques _Unit-1.ppt
Numerical Methods
Ad

Recently uploaded (20)

PPT
ISS -ESG Data flows What is ESG and HowHow
PPTX
STERILIZATION AND DISINFECTION-1.ppthhhbx
PPTX
Data_Analytics_and_PowerBI_Presentation.pptx
PDF
Business Analytics and business intelligence.pdf
PDF
Transcultural that can help you someday.
PDF
How to run a consulting project- client discovery
PPT
DATA COLLECTION METHODS-ppt for nursing research
PDF
Mega Projects Data Mega Projects Data
PPTX
Modelling in Business Intelligence , information system
PDF
Optimise Shopper Experiences with a Strong Data Estate.pdf
PDF
Oracle OFSAA_ The Complete Guide to Transforming Financial Risk Management an...
PPTX
Database Infoormation System (DBIS).pptx
PPTX
Pilar Kemerdekaan dan Identi Bangsa.pptx
PPTX
QUANTUM_COMPUTING_AND_ITS_POTENTIAL_APPLICATIONS[2].pptx
PPTX
Topic 5 Presentation 5 Lesson 5 Corporate Fin
PPTX
CEE 2 REPORT G7.pptxbdbshjdgsgjgsjfiuhsd
PDF
Microsoft Core Cloud Services powerpoint
PDF
Galatica Smart Energy Infrastructure Startup Pitch Deck
PDF
Introduction to Data Science and Data Analysis
PPTX
Managing Community Partner Relationships
ISS -ESG Data flows What is ESG and HowHow
STERILIZATION AND DISINFECTION-1.ppthhhbx
Data_Analytics_and_PowerBI_Presentation.pptx
Business Analytics and business intelligence.pdf
Transcultural that can help you someday.
How to run a consulting project- client discovery
DATA COLLECTION METHODS-ppt for nursing research
Mega Projects Data Mega Projects Data
Modelling in Business Intelligence , information system
Optimise Shopper Experiences with a Strong Data Estate.pdf
Oracle OFSAA_ The Complete Guide to Transforming Financial Risk Management an...
Database Infoormation System (DBIS).pptx
Pilar Kemerdekaan dan Identi Bangsa.pptx
QUANTUM_COMPUTING_AND_ITS_POTENTIAL_APPLICATIONS[2].pptx
Topic 5 Presentation 5 Lesson 5 Corporate Fin
CEE 2 REPORT G7.pptxbdbshjdgsgjgsjfiuhsd
Microsoft Core Cloud Services powerpoint
Galatica Smart Energy Infrastructure Startup Pitch Deck
Introduction to Data Science and Data Analysis
Managing Community Partner Relationships

Methods for meta learning in AutoML

  • 1. Mohamed Maher - University of Tartu - 2019 - [email protected] Methods for Meta-Learning in AutoML Learning how to Learn 1
  • 2. Mohamed Maher - University of Tartu - 2019 - [email protected] Motivational Example ● Alex starts to learn a Maths course of 10 tests for the first time in his life. (Problem) ● Alex wants to get a grade A in most of the course tests. (Target) ● Alex thought that attending all lectures would easily help him to get grade A like what he always does in history courses. (Approach 1) 2
  • 3. Mohamed Maher - University of Tartu - 2019 - [email protected] Motivational Example ● Alex got grade D in his first test. (Result 1) ● Alex decided to switch to reading the reference book instead of attending all lectures only. (Approach 2) ● Alex got grade C in his second test. (Result 2) 3
  • 4. Mohamed Maher - University of Tartu - 2019 - [email protected] Motivational Example ● After that, Alex decided to switch to solving practice problems instead. (Approach 3) ● Alex got grade B in his third test. (Result 3) ● So, Alex decided to summarize each lesson and teach it to his colleagues too. (Approach 4) ● Alex got grade A in his fourth test. (Result 4) 4
  • 5. Mohamed Maher - University of Tartu - 2019 - [email protected] Motivational Example ● Now, the question is how will Alex study for his 5th test in the course ? 5 Alex has already learnt how to learn.
  • 6. Mohamed Maher - University of Tartu - 2019 - [email protected] Back to Machine Learning 6
  • 7. Mohamed Maher - University of Tartu - 2019 - [email protected] Motivation 7 Data Collection 1. Data Preprocessing 2. Feature Extraction 3. Feature Selection 4. Algorithm Selection Deploym ent 5. Parameter Tuning Prediction Real-World Data Feature Engineering Model Building Typical Supervised Machine Learning Pipeline
  • 8. Mohamed Maher - University of Tartu - 2019 - [email protected] 8 Model Building 4. Algorithm Selection 5. Parameter Tuning Examples: - Linear Classification: (Simple Linear Classification, Ridge, Lasso, Simple Perceptron, ….) - Support Vector Machines - Decision Tree (ID3, C4.5, C5.0, CART, ….) - Nearest Neighbors - Gaussian Processes - Naive Bayes (Gaussian, Bernoulli, Complement, ….) - Ensembling: (Random Forest, GBM, AdaBoost, ….) Motivation: Model Building
  • 9. Mohamed Maher - University of Tartu - 2019 - [email protected] 9 Model Building 4. Algorithm Selection 5. Parameter Tuning Kernel Linear RBF Polynomial Gamma [2^-15, 2^3] Degree 2,3,.... C - Penalty [2^-5, 2^15] Example: Support Vector Machine …….. Motivation: Model Building
  • 10. Mohamed Maher - University of Tartu - 2019 - [email protected] 10 Motivation: Dimensionality Reduction Examples of Feature Extraction: 1. Principal Component Analysis 2. Linear Discriminant Analysis 3. Multiple Discriminant Analysis 4. Independent Component Analysis Examples of Multivariate Feature Selection: 1. Relief 2. Correlation Feature Selection 3. Branch and Bound 4. Sequential Forward Selection 5. Plus L - Minus R Examples of Univariate Feature Selection: 1. Information Gain 2. Fisher Score 3. Correlation with Target 2. Feature Extraction 3. Feature Selection
  • 11. Mohamed Maher - University of Tartu - 2019 - [email protected] 11 Motivation: Data PreProcessing Examples of Data Preprocessors: 1. Scaling 2. Normalization 3. Standardization 4. Binarization 5. Imputation 6. Deletion 7. One-Hot-Encoding 8. Hashing 9. Discretization 1. Data Preprocessing
  • 12. Mohamed Maher - University of Tartu - 2019 - [email protected] Solution: Meta-Learning 1. Science of systematically observing how different machine learning approaches perform on a wide range of learning tasks and then learning from this experience. 12
  • 13. Mohamed Maher - University of Tartu - 2019 - [email protected] Solution: Meta-Learning 2. It also allows to replace hand-written rules and algorithms with novel approaches that are data-driven. 13 1. Science of systematically observing how different machine learning approaches perform on a wide range of learning tasks and then learning from this experience.
  • 14. Mohamed Maher - University of Tartu - 2019 - [email protected] HOW ? Collect MetaData 1. Model Configurations: - Pipeline Composition: (Normalization → PCA → SVM) - Hyperparameter Settings: (PCA = 2 components, SVM = gamma: 1e-9, C = 1e2) - Network Architectures: (2 Hidden Layers, 100 Neurons per layer) 2. Resulting Model Evaluations: - Different Metrics: Accuracy, error rate, F1-Score. - Training Time. 3. Task Itself (Meta-Features): - Description of the data 14
  • 15. Mohamed Maher - University of Tartu - 2019 - [email protected] HOW ? Use Meta-Data 1. Knowledge Transfer. Use the same model as an initial point and start to tune it. 2. Guided Search. If Classifier X is worth than Classifier Y by 10% then there is no need to tune classifier X 15
  • 16. Mohamed Maher - University of Tartu - 2019 - [email protected] HOW ? Use Meta-Data Remember that Alex starts with the same approach that succeeds in History courses. Meta-Learning won’t be effective and may affect performance badly in case of: - Tasks with random noise, and unrelated phenomena. “Tasks that are Never Seen Before” 16
  • 17. Mohamed Maher - University of Tartu - 2019 - [email protected] Meta-Learning Methodologies: 1. Learning from Task Properties. 2. Learning from Model Evaluations. 3. Learning from Prior Models. 17
  • 18. Mohamed Maher - University of Tartu - 2019 - [email protected] 1-Learning from Task Properties: ● Represent task as a meta-feature vector. ● Studies show that optimal set of meta-features depends on application type.[2] ● Different studies used various feature selection and extraction techniques to reduce set of meta-features.[2][3] 18
  • 19. Mohamed Maher - University of Tartu - 2019 - [email protected] 1-Learning from Task Properties: ● What are Task Properties? = Types of Meta-features: 1. Simple 2. Statistical 3. Information Theoretic 4. Complexity 5. LandMarkers 19
  • 20. Mohamed Maher - University of Tartu - 2019 - [email protected] Meta-Features Types: (Simple) ● Examples: 1. Number of Instances 2. Number of Features 3. Number of Classes 4. Number of Missing Values 5. Number of Outliers 20
  • 21. Mohamed Maher - University of Tartu - 2019 - [email protected] Meta-Features Types: (Statistical) ● Examples: 1. Skewness of Numerical Features. 2. Kurtosis of Numerical Features. 3. Correlation Covariance between features. 4. Variance in first PCA. 5. Skewness and Kurtosis of first PCA. 6. Class probability distribution. 7. Concentration, Sparsity, Gravity of Features (Measurements of independence and dispersion of values.) 21
  • 22. Mohamed Maher - University of Tartu - 2019 - [email protected] Meta-Features Types: (Information theoretic) ● Examples: 1. Class Entropy. 2. Mutual Information between feature and Class. 3. Equivalent number of features (2/1) 4. Noise to Signal ratio. 22
  • 23. Mohamed Maher - University of Tartu - 2019 - [email protected] Meta-Features Types: (Task Complexity) ● Examples: 1. Fisher discriminant (Measure separability between classes). 23
  • 24. Mohamed Maher - University of Tartu - 2019 - [email protected] Meta-Features Types: (Landmarkers) ● Examples: 1. LandMarker 1NN. 2. LandMarker Decision Tree. 3. LandMarker Naive Bayes. 4. LandMarker Linear Discriminant Analysis. 24
  • 25. Mohamed Maher - University of Tartu - 2019 - [email protected] How to use Meta-Features? 25 ● Different Similarity Measurements (Unsupervised) and warm starting optimization of similar tasks for recommendation of candidate configurations: Examples: 1. Rank of different configurations. - Tasks A, B are twin tasks. - SVM and KNN are the best for Task A. - Then, SVM and KNN are the best for Task B.
  • 26. Mohamed Maher - University of Tartu - 2019 - [email protected] How to use Meta-Features? 26 ● Different Similarity Measurements (Unsupervised) and warm starting optimization of similar tasks for recommendation of candidate configurations: Examples: 2. Collaborative Filtering Use results of few configurations on Task A to predicts results of all other configurations based on configurations results on a similar Task B Knowledge Base Needs almost full configurations results to be updated.
  • 27. Mohamed Maher - University of Tartu - 2019 - [email protected] 1-Learning from Task Properties: 27
  • 28. Mohamed Maher - University of Tartu - 2019 - [email protected] How to use Meta-Features? 28 ● Learning High Level Meta-Features Low Level Features High Level Features NEEDS BIG KNOWLEDGE BASE
  • 29. Mohamed Maher - University of Tartu - 2019 - [email protected] How to use Meta-Features? 29 ● Meta-Models (Supervised): Learn the complex relationship between meta-features and useful configurations in this large space. Example: - Ranking of Top N Promising Configurations: Literature suggests Boosting, and Bagging Models [4][5]. + Approximate Ranking Tree Forests [6] (Auto Meta-Feature Selection based on some initial results).
  • 30. Mohamed Maher - University of Tartu - 2019 - [email protected] How to use Meta-Features? 30 ● Pipeline Synthesis: 1. Meta-Model to predict which preprocessor with improve performance of a specific classifier in that particular task. [7] [8] 2. Reinforcement Learning to construct pipeline by addition, deletion, replacement of pipeline blocks. [9] (Alpha D3M - Evolutionary Approach)
  • 31. Mohamed Maher - University of Tartu - 2019 - [email protected] How to use Meta-Features? 31 ● Tune or Not to Tune: Meta Models to predict: 1. How much improvement we can expect from tuning this particular classifier on that particular task [10]. 2. How much improvement VS additional time investment? [11].
  • 32. Mohamed Maher - University of Tartu - 2019 - [email protected] 2-Learning from Model Evaluations: ● Using Current configuration evaluations as a prior to suggesting the coming candidate outperforming configuration in an iterative way. 32 Example: 1. Evaluate Px on Task 1 2. Suggest new Ps 3. Select most candidate outperforming P 4. Set Px = P 5. GO TO 1
  • 33. Mohamed Maher - University of Tartu - 2019 - [email protected] How it is used? ● Task Independent Recommendation: 1. Discretize the search space into a set of configurations. 2. Apply over many datasets. 3. Aggregate single task rankings into a global ranking. ● Example: Scikit Learn Cheat Sheet Algorithm. 33
  • 34. Mohamed Maher - University of Tartu - 2019 - [email protected] 34
  • 35. Mohamed Maher - University of Tartu - 2019 - [email protected] How it is used? ● Search Space Design: 1. Learn hyperparameter default values (Best configuration over all tasks). 2. Learn different hyperparameters importance: - Measure variance of algorithm performance by keeping all hyperparameters fixed and change only one. 35
  • 36. Mohamed Maher - University of Tartu - 2019 - [email protected] How it is used? ● Learning Curves: (Example: 1. Apply SVM Over 100 Training Datasets) 36
  • 37. Mohamed Maher - University of Tartu - 2019 - [email protected] How it is used? ● Learning Curves: (Example: 2. Apply SVM Over New Dataset) 37
  • 38. Mohamed Maher - University of Tartu - 2019 - [email protected] How it is used? ● Learning Curves: (Example: 3. Measure Similarity between training curves and testing curve) 38
  • 39. Mohamed Maher - University of Tartu - 2019 - [email protected] How it is used? ● Configuration Transfer: - Surrogate Models: usually suitable with Gaussian Processes Bayesian Optimization like the SMAC algorithm. - We can define task similarity based on Learning Distribution Similarity between tasks too. 39
  • 40. Mohamed Maher - University of Tartu - 2019 - [email protected] How it is used? ● Configuration Transfer: - Surrogate Models: usually suitable with Gaussian Processes like the SMAC algorithm. We can define task similarity based of accuracy of predictions for 40
  • 41. Mohamed Maher - University of Tartu - 2019 - [email protected] How it is used? ● Configuration Transfer: - Multi-armed bandits: 1. Start with small data portion and apply multiple configurations on these small portions. 2. Drop lowest performing configurations and increase portion size for other configurations. 41
  • 42. Mohamed Maher - University of Tartu - 2019 - [email protected] How it is used? ● Configuration Transfer: - Multi-armed bandits: 42
  • 43. Mohamed Maher - University of Tartu - 2019 - [email protected] 3-Learning from Prior Models: ● Take already trained Models (Model HUB) to use for similar tasks. ● Suitable for few classifiers (Eg: Kernel Classifiers - Bayesian Networks) BUT very good with Neural Networks. WHY? Both Structure and Network Parameters can be a good initialization for the target model. 43
  • 44. Mohamed Maher - University of Tartu - 2019 - [email protected] References: [1] Hutter Frank and Kotthoff Lars and Vanschoren Joaquin - Automated Machine Learning: Methods, Systems, Challenges - (2019), Springer [2] Bilalli, B., Abell´o, A., Aluja-Banet, T.: On the predictive power of metafeatures in OpenML. International Journal of Applied Mathematics and Computer Science 27(4), 697 – 712 (2017) [3] Todorovski, L., Brazdil, P., Soares, C.: Report on the experiments with feature selection in meta-level learning. PKDD 2000 Workshop on Data mining, Decision support, Meta-learning and ILP pp. 27–39 (2000) [4] Pinto, F., Cerqueira, V., Soares, C., Mendes-Moreira, J.: autoBagging: Learning to rank bagging workflows with metalearning. arXiv 1706.09367 (2017) [5] Lorena, A.C., Maciel, A.I., de Miranda, P.B.C., Costa, I.G., Prudˆencio, R.B.C.: Data complexity meta-features for regression problems. Machine Learning 107(1), 209–246 (2018) [6] Sun, Q., Pfahringer, B.: Pairwise meta-rules for better meta-learning based algorithm ranking. Machine Learning 93(1), 141–161 (2013) [7] Bilalli, B., Abell´o, A., Aluja-Banet, T., Wrembel, R.: Intelligent assistance for data pre-processing. Computer Standards & Interf. 57, 101 – 109 (2018) [8] Schoenfeld, B., Giraud-Carrier, C., Poggeman, M., Christensen, J., Seppi, K.: Feature selection for high-dimensional data: A fast correlation-based filter solution. In: AutoML Workshop at ICML (2018) [9] Drori, I., Krishnamurthy, Y., Rampin, R., de Paula Lourenco, R., Ono, J.P., Cho, K., Silva, C., Freire, J.: AlphaD3M: Machine learning pipeline synthesis. In: AutoML Workshop at ICML (2018) [10] Ridd, P., Giraud-Carrier, C.: Using metalearning to predict when parameter optimization is likely to improve classification accuracy. In: ECAI Workshop on Meta-learning and Algorithm Selection. pp. 18–23 (2014) [11] Sanders, S., Giraud-Carrier, C.: Informing the use of hyperparameter optimization through metalearning. In: Proc. ICDM. pp. 1051–1056 (2017) 44
  • 45. Mohamed Maher - University of Tartu - 2019 - [email protected] 45