SlideShare a Scribd company logo
Automatic Gain Tuning
based on
Gaussian Process Global Optimization
(= Bayesian Optimization)
https://is.tuebingen.mpg.de/publications/marco_icra_2016
=
A simple PD control example
Global optimal gains, θ to get a minimum cost J ?
A simple PD control example
Procedure of Bayesian Optimization
1. GP prior before observing any data
2. GP posterior, after five noisy evaluations
3. The next parameters θnext are chosen
at the maximum of Acquisition function
Repeat until you can find
a globally optimal θ
Acquisition function
Information gain, : Mutual information for an observed data
 Reduction of uncertainty in the location ∗ by selecting points
that are expected to cause the largest reduction in entropy of
distribution ∗
Acquisition function and Entropy Search
Control design problem
propose the use of Entropy Search, a recent algorithm for global
Bayesian optimization, as the minimizer for the LQR tuning problem.
ES employs a Gaussian process (GP) as a non-parametric model
capturing the knowledge about the unknown cost function.
consider a system that follows a discrete-time nonlinear dynamic
model
system states xk, control input uk, and zero-mean process noise wk at
time instant k
common way to measure the performance of a control system is
through a quadratic cost function such as
cost captures a trade-off between control performance (keeping xk
small) and control effort(keeping uk small)
Nonlinear control design problem is intractable in general.
linear model is often sufficient for control design
Control design problem
Linear Quadratic Regulator (LQR)
static gain matrix F can readily be computed by solving the
discrete-time infinite-horizon LQR problem for the nominal model
(An, Bn) and the weights (Q, R).
LQR tuning problem
goal of the automatic LQR tuning is to vary the parameters θ
such as to minimize the cost
Optimization problem
Consider the approximate one of quadratic cost function
uncertainty over the objective function J is represented by a
probability measure p(J), typically a Gaussian process (GP)
LQR TUNING WITH ENTROPY SEARCH
prior knowledge about the cost function J as the GP
with mean function and covariance function ∗
squared exponential (SE) covariance function
noisy evaluations of cost function can be modeled as
Conditioning the GP on the data then yields another GP with
posterior mean and a posterior variance ∗ .
 shape of the mean is adjusted to fit the data points, and the
uncertainty (standard deviation) is reduced around the evaluations
points.
LQR TUNING WITH ENTROPY SEARCH
LQR TUNING WITH ENTROPY SEARCH
ES is one out of several popular formulations of Bayesian Optimization
Aims to reduce the uncertainty in the location ∗
expected to cause the largest reduction in entropy( uncertainty)
of distribution, J(θ)
Bayesian
Optimization
Bayesian Optimization and How to Scale It Up
Zi Wang
CS Colloquium, US
Automatic Gain Tuning based on Gaussian Process Global Optimization (= Bayesian Optimization)
Automatic Gain Tuning based on Gaussian Process Global Optimization (= Bayesian Optimization)
Automatic Gain Tuning based on Gaussian Process Global Optimization (= Bayesian Optimization)
Automatic Gain Tuning based on Gaussian Process Global Optimization (= Bayesian Optimization)
Automatic Gain Tuning based on Gaussian Process Global Optimization (= Bayesian Optimization)
Automatic Gain Tuning based on Gaussian Process Global Optimization (= Bayesian Optimization)
Information gain, : Mutual information for an observed data
 Reduction of uncertainty in the location ∗ by selecting points
that are expected to cause the largest reduction in entropy of
distribution ∗
Probability of Improvement
 
: Gaussian Cumulative Density Ft’n
Automatic Gain Tuning based on Gaussian Process Global Optimization (= Bayesian Optimization)
Automatic Gain Tuning based on Gaussian Process Global Optimization (= Bayesian Optimization)
Automatic Gain Tuning based on Gaussian Process Global Optimization (= Bayesian Optimization)
Automatic Gain Tuning based on Gaussian Process Global Optimization (= Bayesian Optimization)
Automatic Gain Tuning based on Gaussian Process Global Optimization (= Bayesian Optimization)
Automatic Gain Tuning based on Gaussian Process Global Optimization (= Bayesian Optimization)
Evolution of an example Gaussian process for three successive
function evaluations (orange dots)
Approximated probability distribution over the location of the
minimum pmin(θ) in green

More Related Content

PPTX
Gaussian processing
PDF
proposal_pura
PDF
Hyperparameter optimization with approximate gradient
PDF
Clustering:k-means, expect-maximization and gaussian mixture model
PDF
Radial Basis Function Interpolation
PPTX
Machine learning applications in aerospace domain
PDF
K-means, EM and Mixture models
PDF
Lecture 6: Convolutional Neural Networks
Gaussian processing
proposal_pura
Hyperparameter optimization with approximate gradient
Clustering:k-means, expect-maximization and gaussian mixture model
Radial Basis Function Interpolation
Machine learning applications in aerospace domain
K-means, EM and Mixture models
Lecture 6: Convolutional Neural Networks

What's hot (20)

PDF
2012 mdsp pr04 monte carlo
PDF
Lecture 3 image sampling and quantization
PDF
Gradient Estimation Using Stochastic Computation Graphs
PDF
2012 mdsp pr08 nonparametric approach
PDF
Iclr2016 vaeまとめ
PDF
MLHEP 2015: Introductory Lecture #4
PDF
机器学习Adaboost
PPT
Section5 Rbf
PDF
K-means and GMM
PDF
"Automatic Variational Inference in Stan" NIPS2015_yomi2016-01-20
PDF
Accelerating Random Forests in Scikit-Learn
PDF
Neural Networks: Radial Bases Functions (RBF)
PPTX
Detailed Description on Cross Entropy Loss Function
PPTX
The world of loss function
PDF
L1 intro2 supervised_learning
PDF
VAE-type Deep Generative Models
PDF
Minimax optimal alternating minimization \\ for kernel nonparametric tensor l...
PDF
ICML2013読み会 Large-Scale Learning with Less RAM via Randomization
PDF
Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)
PDF
MLHEP 2015: Introductory Lecture #3
2012 mdsp pr04 monte carlo
Lecture 3 image sampling and quantization
Gradient Estimation Using Stochastic Computation Graphs
2012 mdsp pr08 nonparametric approach
Iclr2016 vaeまとめ
MLHEP 2015: Introductory Lecture #4
机器学习Adaboost
Section5 Rbf
K-means and GMM
"Automatic Variational Inference in Stan" NIPS2015_yomi2016-01-20
Accelerating Random Forests in Scikit-Learn
Neural Networks: Radial Bases Functions (RBF)
Detailed Description on Cross Entropy Loss Function
The world of loss function
L1 intro2 supervised_learning
VAE-type Deep Generative Models
Minimax optimal alternating minimization \\ for kernel nonparametric tensor l...
ICML2013読み会 Large-Scale Learning with Less RAM via Randomization
Backpropagation (DLAI D3L1 2017 UPC Deep Learning for Artificial Intelligence)
MLHEP 2015: Introductory Lecture #3
Ad

Similar to Automatic Gain Tuning based on Gaussian Process Global Optimization (= Bayesian Optimization) (20)

PDF
Efficient Solution of Two-Stage Stochastic Linear Programs Using Interior Poi...
PDF
Research on Fuzzy C- Clustering Recursive Genetic Algorithm based on Cloud Co...
PDF
RESEARCH ON FUZZY C- CLUSTERING RECURSIVE GENETIC ALGORITHM BASED ON CLOUD CO...
PDF
Research on Fuzzy C- Clustering Recursive Genetic Algorithm based on Cloud Co...
PDF
Research on Fuzzy C- Clustering Recursive Genetic Algorithm based on Cloud Co...
PPTX
A machine learning method for efficient design optimization in nano-optics
PDF
Global analysis of nonlinear dynamics
PPTX
ngboost.pptx
PDF
An efficient hardware logarithm generator with modified quasi-symmetrical app...
PDF
Scalable trust-region method for deep reinforcement learning using Kronecker-...
PDF
Session 4 .pdf
PPTX
A machine learning method for efficient design optimization in nano-optics
PDF
THE RESEARCH OF QUANTUM PHASE ESTIMATION ALGORITHM
PPT
Data Structures- Part2 analysis tools
PDF
CLIM Program: Remote Sensing Workshop, Optimization for Distributed Data Syst...
PPTX
Optimization of mathematical function using gradient descent algorithm.pptx
PDF
cs.ds-2211.13454.pdf
PDF
A011130109
PPTX
Discrete_Mathematics 6 algorithm complexity.pptx
PPTX
Techniques in Deep Learning
Efficient Solution of Two-Stage Stochastic Linear Programs Using Interior Poi...
Research on Fuzzy C- Clustering Recursive Genetic Algorithm based on Cloud Co...
RESEARCH ON FUZZY C- CLUSTERING RECURSIVE GENETIC ALGORITHM BASED ON CLOUD CO...
Research on Fuzzy C- Clustering Recursive Genetic Algorithm based on Cloud Co...
Research on Fuzzy C- Clustering Recursive Genetic Algorithm based on Cloud Co...
A machine learning method for efficient design optimization in nano-optics
Global analysis of nonlinear dynamics
ngboost.pptx
An efficient hardware logarithm generator with modified quasi-symmetrical app...
Scalable trust-region method for deep reinforcement learning using Kronecker-...
Session 4 .pdf
A machine learning method for efficient design optimization in nano-optics
THE RESEARCH OF QUANTUM PHASE ESTIMATION ALGORITHM
Data Structures- Part2 analysis tools
CLIM Program: Remote Sensing Workshop, Optimization for Distributed Data Syst...
Optimization of mathematical function using gradient descent algorithm.pptx
cs.ds-2211.13454.pdf
A011130109
Discrete_Mathematics 6 algorithm complexity.pptx
Techniques in Deep Learning
Ad

More from 홍배 김 (20)

PPTX
Lecture Summary : Camera Projection
PPTX
Learning agile and dynamic motor skills for legged robots
PPTX
Robotics of Quadruped Robot
PPTX
Basics of Robotics
PPTX
Recurrent Neural Net의 이론과 설명
PPTX
Convolutional neural networks 이론과 응용
PPTX
Anomaly detection using deep one class classifier
PPTX
Optimal real-time landing using DNN
PPTX
Anomaly Detection and Localization Using GAN and One-Class Classifier
PPTX
ARCHITECTURAL CONDITIONING FOR DISENTANGLEMENT OF OBJECT IDENTITY AND POSTURE...
PPTX
Brief intro : Invariance and Equivariance
PPTX
Anomaly Detection with GANs
PPTX
Focal loss의 응용(Detection & Classification)
PPTX
Convolution 종류 설명
PPTX
Learning by association
PPTX
알기쉬운 Variational autoencoder
PPTX
Binarized CNN on FPGA
PPTX
Visualizing data using t-SNE
PPTX
Normalization 방법
PPTX
Learning to remember rare events
Lecture Summary : Camera Projection
Learning agile and dynamic motor skills for legged robots
Robotics of Quadruped Robot
Basics of Robotics
Recurrent Neural Net의 이론과 설명
Convolutional neural networks 이론과 응용
Anomaly detection using deep one class classifier
Optimal real-time landing using DNN
Anomaly Detection and Localization Using GAN and One-Class Classifier
ARCHITECTURAL CONDITIONING FOR DISENTANGLEMENT OF OBJECT IDENTITY AND POSTURE...
Brief intro : Invariance and Equivariance
Anomaly Detection with GANs
Focal loss의 응용(Detection & Classification)
Convolution 종류 설명
Learning by association
알기쉬운 Variational autoencoder
Binarized CNN on FPGA
Visualizing data using t-SNE
Normalization 방법
Learning to remember rare events

Recently uploaded (20)

PPTX
KOM of Painting work and Equipment Insulation REV00 update 25-dec.pptx
PPTX
sap open course for s4hana steps from ECC to s4
PPTX
VMware vSphere Foundation How to Sell Presentation-Ver1.4-2-14-2024.pptx
PDF
KodekX | Application Modernization Development
PPTX
Understanding_Digital_Forensics_Presentation.pptx
PDF
Empathic Computing: Creating Shared Understanding
PDF
Approach and Philosophy of On baking technology
PPTX
Digital-Transformation-Roadmap-for-Companies.pptx
PDF
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
PDF
Per capita expenditure prediction using model stacking based on satellite ima...
PDF
Network Security Unit 5.pdf for BCA BBA.
PPTX
Effective Security Operations Center (SOC) A Modern, Strategic, and Threat-In...
PDF
cuic standard and advanced reporting.pdf
PPTX
MYSQL Presentation for SQL database connectivity
PDF
Diabetes mellitus diagnosis method based random forest with bat algorithm
PDF
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
PDF
Electronic commerce courselecture one. Pdf
PDF
Unlocking AI with Model Context Protocol (MCP)
PDF
Encapsulation theory and applications.pdf
PPTX
Cloud computing and distributed systems.
KOM of Painting work and Equipment Insulation REV00 update 25-dec.pptx
sap open course for s4hana steps from ECC to s4
VMware vSphere Foundation How to Sell Presentation-Ver1.4-2-14-2024.pptx
KodekX | Application Modernization Development
Understanding_Digital_Forensics_Presentation.pptx
Empathic Computing: Creating Shared Understanding
Approach and Philosophy of On baking technology
Digital-Transformation-Roadmap-for-Companies.pptx
TokAI - TikTok AI Agent : The First AI Application That Analyzes 10,000+ Vira...
Per capita expenditure prediction using model stacking based on satellite ima...
Network Security Unit 5.pdf for BCA BBA.
Effective Security Operations Center (SOC) A Modern, Strategic, and Threat-In...
cuic standard and advanced reporting.pdf
MYSQL Presentation for SQL database connectivity
Diabetes mellitus diagnosis method based random forest with bat algorithm
Blue Purple Modern Animated Computer Science Presentation.pdf.pdf
Electronic commerce courselecture one. Pdf
Unlocking AI with Model Context Protocol (MCP)
Encapsulation theory and applications.pdf
Cloud computing and distributed systems.

Automatic Gain Tuning based on Gaussian Process Global Optimization (= Bayesian Optimization)

  • 1. Automatic Gain Tuning based on Gaussian Process Global Optimization (= Bayesian Optimization) https://is.tuebingen.mpg.de/publications/marco_icra_2016
  • 2. = A simple PD control example Global optimal gains, θ to get a minimum cost J ?
  • 3. A simple PD control example Procedure of Bayesian Optimization 1. GP prior before observing any data 2. GP posterior, after five noisy evaluations 3. The next parameters θnext are chosen at the maximum of Acquisition function Repeat until you can find a globally optimal θ
  • 5. Information gain, : Mutual information for an observed data  Reduction of uncertainty in the location ∗ by selecting points that are expected to cause the largest reduction in entropy of distribution ∗ Acquisition function and Entropy Search
  • 6. Control design problem propose the use of Entropy Search, a recent algorithm for global Bayesian optimization, as the minimizer for the LQR tuning problem. ES employs a Gaussian process (GP) as a non-parametric model capturing the knowledge about the unknown cost function. consider a system that follows a discrete-time nonlinear dynamic model system states xk, control input uk, and zero-mean process noise wk at time instant k
  • 7. common way to measure the performance of a control system is through a quadratic cost function such as cost captures a trade-off between control performance (keeping xk small) and control effort(keeping uk small) Nonlinear control design problem is intractable in general. linear model is often sufficient for control design Control design problem
  • 8. Linear Quadratic Regulator (LQR) static gain matrix F can readily be computed by solving the discrete-time infinite-horizon LQR problem for the nominal model (An, Bn) and the weights (Q, R). LQR tuning problem goal of the automatic LQR tuning is to vary the parameters θ such as to minimize the cost
  • 9. Optimization problem Consider the approximate one of quadratic cost function
  • 10. uncertainty over the objective function J is represented by a probability measure p(J), typically a Gaussian process (GP) LQR TUNING WITH ENTROPY SEARCH prior knowledge about the cost function J as the GP with mean function and covariance function ∗ squared exponential (SE) covariance function
  • 11. noisy evaluations of cost function can be modeled as Conditioning the GP on the data then yields another GP with posterior mean and a posterior variance ∗ .  shape of the mean is adjusted to fit the data points, and the uncertainty (standard deviation) is reduced around the evaluations points. LQR TUNING WITH ENTROPY SEARCH
  • 12. LQR TUNING WITH ENTROPY SEARCH ES is one out of several popular formulations of Bayesian Optimization Aims to reduce the uncertainty in the location ∗ expected to cause the largest reduction in entropy( uncertainty) of distribution, J(θ)
  • 13. Bayesian Optimization Bayesian Optimization and How to Scale It Up Zi Wang CS Colloquium, US
  • 20. Information gain, : Mutual information for an observed data  Reduction of uncertainty in the location ∗ by selecting points that are expected to cause the largest reduction in entropy of distribution ∗
  • 21. Probability of Improvement   : Gaussian Cumulative Density Ft’n
  • 28. Evolution of an example Gaussian process for three successive function evaluations (orange dots) Approximated probability distribution over the location of the minimum pmin(θ) in green