Menu
Home Explore People Places Arts History Plants & Animals Science Life & Culture Technology
On this page
Bayesian optimization
Optimization technique for undifferentiable, black-box functions

Bayesian optimization is a sequential design strategy for global optimization of black-box functions, that does not assume any functional forms. It is usually employed to optimize expensive-to-evaluate functions. With the rise of artificial intelligence innovation in the 21st century, Bayesian optimizations have found prominent use in machine learning problems for optimizing hyperparameter values.

We don't have any images related to Bayesian optimization yet.
We don't have any YouTube videos related to Bayesian optimization yet.
We don't have any PDF documents related to Bayesian optimization yet.
We don't have any Books related to Bayesian optimization yet.
We don't have any archived web articles related to Bayesian optimization yet.

History

The term is generally attributed to Jonas Mockus [lt] and is coined in his work from a series of publications on global optimization in the 1970s and 1980s.678

Early Mathematics Foundations

From 1960s to 1980s

The earliest idea of Bayesian optimization 9sprang in 1964, from a paper by American applied mathematician Harold J. Kushner,10 “A New Method of Locating the Maximum Point of an Arbitrary Multipeak Curve in the Presence of Noise”. Although not directly proposing Bayesian optimization, in this paper, he first proposed a new method of locating the maximum point of an arbitrary multipeak curve in a noisy environment. This method provided an important theoretical foundation for subsequent Bayesian optimization.

By the 1980s, the framework we now use for Bayesian optimization was explicitly established. In 1978, the Lithuanian scientist Jonas Mockus,11 in his paper “The Application of Bayesian Methods for Seeking the Extremum”, discussed how to use Bayesian methods to find the extreme value of a function under various uncertain conditions. In his paper, Mockus first proposed the Expected Improvement principle (EI), which is one of the core sampling strategies of Bayesian optimization. This criterion balances exploration while optimizing the function efficiently by maximizing the expected improvement. Because of the usefulness and profound impact of this principle, Jonas Mockus is widely regarded as the founder of Bayesian optimization. Although Expected Improvement principle (IE) is one of the earliest proposed core sampling strategies for Bayesian optimization, it is not the only one, with the development of modern society, we also have Probability of Improvement (PI), or Upper Confidence Bound (UCB)12 and so on.

From Theory to Practice

In the 1990s, Bayesian optimization began to gradually transition from pure theory to real-world applications. In 1998, Donald R. Jones13 and his coworkers published a paper titled “Gaussian Optimization14”. In this paper, they proposed the Gaussian Process(GP) and elaborated on the Expected Improvement principle(EI) proposed by Jonas Mockus in 1978. Through the efforts of Donald R. Jones and his colleagues, Bayesian Optimization began to shine in the fields like computers science and engineering. However, the computational complexity of Bayesian optimization for the computing power at that time still affected its development to a large extent.

In the 21st century, with the gradual rise of artificial intelligence and bionic robots, Bayesian optimization has been widely used in machine learning and deep learning, and has become an important tool for Hyperparameter Tuning.15 Companies such as Google, Facebook and OpenAI have added Bayesian optimization to their deep learning frameworks to improve search efficiency. However, Bayesian optimization still faces many challenges, for example, because of the use of Gaussian Process16 as a proxy model for optimization, when there is a lot of data, the training of Gaussian Process will be very slow and the computational cost is very high. This makes it difficult for this optimization method to work well in more complex drug development and medical experiments.

Strategy

Bayesian optimization is typically used on problems of the form max x ∈ A f ( x ) {\textstyle \max _{x\in A}f(x)} , where A {\textstyle A} is a set of points, x {\textstyle x} , which rely upon less (or equal to) than 20 dimensions ( R d , d ≤ 20 {\textstyle \mathbb {R} ^{d},d\leq 20} ), and whose membership can easily be evaluated. Bayesian optimization is particularly advantageous for problems where f ( x ) {\textstyle f(x)} is difficult to evaluate due to its computational cost. The objective function, f {\textstyle f} , is continuous and takes the form of some unknown structure, referred to as a "black box". Upon its evaluation, only f ( x ) {\textstyle f(x)} is observed and its derivatives are not evaluated.17

Since the objective function is unknown, the Bayesian strategy is to treat it as a random function and place a prior over it. The prior captures beliefs about the behavior of the function. After gathering the function evaluations, which are treated as data, the prior is updated to form the posterior distribution over the objective function. The posterior distribution, in turn, is used to construct an acquisition function (often also referred to as infill sampling criteria) that determines the next query point.

There are several methods used to define the prior/posterior distribution over the objective function. The most common two methods use Gaussian processes in a method called kriging. Another less expensive method uses the Parzen-Tree Estimator to construct two distributions for 'high' and 'low' points, and then finds the location that maximizes the expected improvement.18

Standard Bayesian optimization relies upon each x ∈ A {\displaystyle x\in A} being easy to evaluate, and problems that deviate from this assumption are known as exotic Bayesian optimization problems. Optimization problems can become exotic if it is known that there is noise, the evaluations are being done in parallel, the quality of evaluations relies upon a tradeoff between difficulty and accuracy, the presence of random environmental conditions, or if the evaluation involves derivatives.19

Acquisition functions

Examples of acquisition functions include

  • probability of improvement
  • expected improvement
  • Bayesian expected losses
  • upper confidence bounds (UCB) or lower confidence bounds
  • Thompson sampling

and hybrids of these.20 They all trade-off exploration and exploitation so as to minimize the number of function queries. As such, Bayesian optimization is well suited for functions that are expensive to evaluate.

Solution methods

The maximum of the acquisition function is typically found by resorting to discretization or by means of an auxiliary optimizer. Acquisition functions are maximized using a numerical optimization technique, such as Newton's method or quasi-Newton methods like the Broyden–Fletcher–Goldfarb–Shanno algorithm.

Applications

The approach has been applied to solve a wide range of problems,21 including learning to rank,22 computer graphics and visual design,232425 robotics,26272829 sensor networks,3031 automatic algorithm configuration,3233 automatic machine learning toolboxes,343536 reinforcement learning,37 planning, visual attention, architecture configuration in deep learning, static program analysis, experimental particle physics,3839 quality-diversity optimization,404142 chemistry, material design, and drug development.434445

Bayesian optimization has been applied in the field of facial recognition.46 The performance of the Histogram of Oriented Gradients (HOG) algorithm, a popular feature extraction method, heavily relies on its parameter settings. Optimizing these parameters can be challenging but crucial for achieving high accuracy.47 A novel approach to optimize the HOG algorithm parameters and image size for facial recognition using a Tree-structured Parzen Estimator (TPE) based Bayesian optimization technique has been proposed.48 This optimized approach has the potential to be adapted for other computer vision applications and contributes to the ongoing development of hand-crafted parameter-based feature extraction algorithms in computer vision.49

See also

References

  1. Močkus, J. (1989). Bayesian Approach to Global Optimization. Dordrecht: Kluwer Academic. ISBN 0-7923-0115-3. 0-7923-0115-3

  2. Garnett, Roman (2023). Bayesian Optimization. Cambridge University Press. ISBN 978-1-108-42578-0. 978-1-108-42578-0

  3. Hennig, P.; Osborne, M. A.; Kersting, H. P. (2022). Probabilistic Numerics (PDF). Cambridge University Press. pp. 243–278. ISBN 978-1107163447. 978-1107163447

  4. Snoek, Jasper (2012). "Practical Bayesian Optimization of Machine Learning Algorithms". Advances in Neural Information Processing Systems 25 (NIPS 2012). 25. arXiv:1206.2944. https://proceedings.neurips.cc/paper/2012/hash/05311655a15b75fab86956663e1819cd-Abstract.html

  5. Klein, Aaron (2017). "Fast bayesian optimization of machine learning hyperparameters on large datasets". Proceedings of the 20th International Conference on Artificial Intelligence and Statistics, PMLR: 528–536. arXiv:1605.07079. https://proceedings.mlr.press/v54/klein17a.html

  6. Močkus, Jonas (1975). "On bayesian methods for seeking the extremum". Optimization Techniques IFIP Technical Conference Novosibirsk, July 1–7, 1974. Lecture Notes in Computer Science. Vol. 27. pp. 400–404. doi:10.1007/3-540-07165-2_55. ISBN 978-3-540-07165-5. 978-3-540-07165-5

  7. Močkus, Jonas (1977). "On Bayesian Methods for Seeking the Extremum and their Application". IFIP Congress: 195–200.

  8. Močkus, J. (1989). Bayesian Approach to Global Optimization. Dordrecht: Kluwer Academic. ISBN 0-7923-0115-3. 0-7923-0115-3

  9. GARNETT, ROMAN (2023). BAYESIAN OPTIMIZATION (First published 2023 ed.). Cambridge University Press. ISBN 978-1-108-42578-0. 978-1-108-42578-0

  10. "Kushner, Harold". vivo.brown.edu. https://vivo.brown.edu/display/hkushner

  11. "Jonas Mockus". Kaunas University of Technology. Retrieved 2025-03-06. https://en.ktu.edu/people/jonas-mockus/

  12. Kaufmann, Emilie; Cappe, Olivier; Garivier, Aurelien (2012-03-21). "On Bayesian Upper Confidence Bounds for Bandit Problems". Proceedings of the Fifteenth International Conference on Artificial Intelligence and Statistics. PMLR: 592–600. https://proceedings.mlr.press/v22/kaufmann12.html

  13. "Donald R. Jones". scholar.google.com. Retrieved 2025-02-25. https://scholar.google.com/citations?user=CZhZ4MYAAAAJ&hl=en

  14. Grcar, Joseph F. Mathematicians of Gaussian Elimination.

  15. T. T. Joy, S. Rana, S. Gupta and S. Venkatesh, "Hyperparameter tuning for big data using Bayesian optimisation," 2016 23rd International Conference on Pattern Recognition (ICPR), Cancun, Mexico, 2016, pp. 2574-2579, doi: 10.1109/ICPR.2016.7900023. keywords: {Big Data;Bayes methods;Optimization;Tuning;Data models;Gaussian processes;Noise measurement},

  16. "Wayback Machine". citeseerx.ist.psu.edu. Archived from the original on 2024-04-23. Retrieved 2025-03-06. http://web.archive.org/web/20240423144014/https://citeseerx.ist.psu.edu/document?repid=rep1&type=pdf&doi=e045b76dc5daf9f4656ac10b456c5d1d9de5bc84

  17. Frazier, Peter I. (2018-07-08). "A Tutorial on Bayesian Optimization". arXiv:1807.02811 [stat.ML]. /wiki/ArXiv_(identifier)

  18. J. S. Bergstra, R. Bardenet, Y. Bengio, B. Kégl: Algorithms for Hyper-Parameter Optimization. Advances in Neural Information Processing Systems: 2546–2554 (2011) http://papers.nips.cc/paper/4443-algorithms-for-hyper-parameter-optimization.pdf

  19. Frazier, Peter I. (2018-07-08). "A Tutorial on Bayesian Optimization". arXiv:1807.02811 [stat.ML]. /wiki/ArXiv_(identifier)

  20. Matthew W. Hoffman, Eric Brochu, Nando de Freitas: Portfolio Allocation for Bayesian Optimization. Uncertainty in Artificial Intelligence: 327–336 (2011) /wiki/Nando_de_Freitas

  21. Eric Brochu, Vlad M. Cora, Nando de Freitas: A Tutorial on Bayesian Optimization of Expensive Cost Functions, with Application to Active User Modeling and Hierarchical Reinforcement Learning. CoRR abs/1012.2599 (2010) https://arxiv.org/abs/1012.2599

  22. Eric Brochu, Nando de Freitas, Abhijeet Ghosh: Active Preference Learning with Discrete Choice Data. Advances in Neural Information Processing Systems: 409-416 (2007) http://papers.nips.cc/paper/3219-active-preference-learning-with-discrete-choice-data.pdf

  23. Eric Brochu, Tyson Brochu, Nando de Freitas: A Bayesian Interactive Optimization Approach to Procedural Animation Design. Symposium on Computer Animation 2010: 103–112 http://haikufactory.com/files/sca2010.pdf

  24. Yuki Koyama, Issei Sato, Daisuke Sakamoto, Takeo Igarashi: Sequential Line Search for Efficient Visual Design Optimization by Crowds. ACM Transactions on Graphics, Volume 36, Issue 4, pp.48:1–48:11 (2017). DOI: https://doi.org/10.1145/3072959.3073598 https://koyama.xyz/project/sequential_line_search/download/preprint.pdf

  25. Yuki Koyama, Issei Sato, Masataka Goto: Sequential Gallery for Interactive Visual Design Optimization. ACM Transactions on Graphics, Volume 39, Issue 4, pp.88:1–88:12 (2020). DOI: https://doi.org/10.1145/3386569.3392444 https://arxiv.org/abs/2005.04107

  26. Daniel J. Lizotte, Tao Wang, Michael H. Bowling, Dale Schuurmans: Automatic Gait Optimization with Gaussian Process Regression Archived 2017-08-12 at the Wayback Machine. International Joint Conference on Artificial Intelligence: 944–949 (2007) https://www.aaai.org/Papers/IJCAI/2007/IJCAI07-152.pdf

  27. Ruben Martinez-Cantin, Nando de Freitas, Eric Brochu, Jose Castellanos and Arnaud Doucet. A Bayesian exploration-exploitation approach for optimal online sensing and planning with a visually guided mobile robot. Autonomous Robots. Volume 27, Issue 2, pp 93–103 (2009) https://link.springer.com/article/10.1007%2Fs10514-009-9130-2#

  28. Scott Kuindersma, Roderic Grupen, and Andrew Barto. Variable Risk Control via Stochastic Optimization. International Journal of Robotics Research, volume 32, number 7, pp 806–825 (2013) http://ijr.sagepub.com/content/32/7/806.abstract#

  29. Roberto Calandra, André Seyfarth, Jan Peters, and Marc P. Deisenroth Bayesian optimization for learning gaits under uncertainty. Ann. Math. Artif. Intell. Volume 76, Issue 1, pp 5-23 (2016) DOI:10.1007/s10472-015-9463-9 https://link.springer.com/article/10.1007%2Fs10472-015-9463-9

  30. Niranjan Srinivas, Andreas Krause, Sham M. Kakade, Matthias W. Seeger: Information-Theoretic Regret Bounds for Gaussian Process Optimization in the Bandit Setting. IEEE Transactions on Information Theory 58(5):3250–3265 (2012) https://infoscience.epfl.ch/record/177246/files/srinivas_ieeeit2012.pdf

  31. Garnett, Roman; Osborne, Michael A.; Roberts, Stephen J. (2010). "Bayesian optimization for sensor set selection". In Abdelzaher, Tarek F.; Voigt, Thiemo; Wolisz, Adam (eds.). Proceedings of the 9th International Conference on Information Processing in Sensor Networks, IPSN 2010, April 12–16, 2010, Stockholm, Sweden. ACM. pp. 209–219. doi:10.1145/1791212.1791238. ISBN 978-1-60558-988-6. 978-1-60558-988-6

  32. Frank Hutter, Holger Hoos, and Kevin Leyton-Brown (2011). Sequential model-based optimization for general algorithm configuration, Learning and Intelligent Optimization http://www.cs.ubc.ca/labs/beta/Projects/SMAC/papers/11-LION5-SMAC.pdf

  33. J. Snoek, H. Larochelle, R. P. Adams Practical Bayesian Optimization of Machine Learning Algorithms. Advances in Neural Information Processing Systems: 2951-2959 (2012) https://papers.nips.cc/paper/4522-practical-bayesian-optimization-of-machine-learning-algorithms.pdf

  34. J. Bergstra, D. Yamins, D. D. Cox (2013). Hyperopt: A Python Library for Optimizing the Hyperparameters of Machine Learning Algorithms. Proc. SciPy 2013. http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.704.3494&rep=rep1&type=pdf

  35. Chris Thornton, Frank Hutter, Holger H. Hoos, Kevin Leyton-Brown: Auto-WEKA: combined selection and hyperparameter optimization of classification algorithms. KDD 2013: 847–855 https://arxiv.org/abs/1208.3719

  36. Jasper Snoek, Hugo Larochelle and Ryan Prescott Adams. Practical Bayesian Optimization of Machine Learning Algorithms. Advances in Neural Information Processing Systems, 2012 https://papers.nips.cc/paper/4522-practical-bayesian-optimization-of-machine-learning-algorithms.pdf

  37. Berkenkamp, Felix (2019). Safe Exploration in Reinforcement Learning: Theory and Applications in Robotics (Doctoral Thesis thesis). ETH Zurich. doi:10.3929/ethz-b-000370833. hdl:20.500.11850/370833. https://www.research-collection.ethz.ch/handle/20.500.11850/370833

  38. Philip Ilten, Mike Williams, Yunjie Yang. Event generator tuning using Bayesian optimization. 2017 JINST 12 P04028. DOI: 10.1088/1748-0221/12/04/P04028 https://arxiv.org/abs/1610.08328

  39. Evaristo Cisbani et al. AI-optimized detector design for the future Electron-Ion Collider: the dual-radiator RICH case 2020 JINST 15 P05009. DOI: 10.1088/1748-0221/15/05/P05009 https://iopscience.iop.org/article/10.1088/1748-0221/15/05/P05009

  40. Kent, Paul; Gaier, Adam; Mouret, Jean-Baptiste; Branke, Juergen (2023-07-19). "BOP-Elites, a Bayesian Optimisation Approach to Quality Diversity Search with Black-Box descriptor functions". arXiv:2307.09326 [math.OC]. Preprint: Arxiv. /wiki/ArXiv_(identifier)

  41. Kent, Paul; Branke, Juergen (2023-07-12). "Bayesian Quality Diversity Search with Interactive Illumination". Proceedings of the Genetic and Evolutionary Computation Conference (PDF). GECCO '23. New York, NY, USA: Association for Computing Machinery. pp. 1019–1026. doi:10.1145/3583131.3590486. ISBN 979-8-4007-0119-1. S2CID 259833672. 979-8-4007-0119-1

  42. Gaier, Adam; Asteroth, Alexander; Mouret, Jean-Baptiste (2018-09-01). "Data-Efficient Design Exploration through Surrogate-Assisted Illumination". Evolutionary Computation. 26 (3): 381–410. arXiv:1806.05865. doi:10.1162/evco_a_00231. ISSN 1063-6560. PMID 29883202. S2CID 47003986. https://doi.org/10.1162%2Fevco_a_00231

  43. Frazier, Peter I. (2018-07-08). "A Tutorial on Bayesian Optimization". arXiv:1807.02811 [stat.ML]. /wiki/ArXiv_(identifier)

  44. Gomez-Bombarelli et al. Automatic Chemical Design using a Data-Driven Continuous Representation of Molecules. ACS Central Science, Volume 4, Issue 2, 268-276 (2018) https://pubs.acs.org/doi/10.1021/acscentsci.7b00572

  45. Griffiths et al. Constrained Bayesian Optimization for Automatic Chemical Design using Variational Autoencoders Chemical Science: 11, 577-586 (2020) https://pubs.rsc.org/en/content/articlehtml/2020/sc/c9sc04026a

  46. Mohammed Mehdi Bouchene: Bayesian Optimization of Histogram of Oriented Gradients (Hog) Parameters for Facial Recognition. SSRN (2023)

  47. Mohammed Mehdi Bouchene: Bayesian Optimization of Histogram of Oriented Gradients (Hog) Parameters for Facial Recognition. SSRN (2023)

  48. Mohammed Mehdi Bouchene: Bayesian Optimization of Histogram of Oriented Gradients (Hog) Parameters for Facial Recognition. SSRN (2023)

  49. Mohammed Mehdi Bouchene: Bayesian Optimization of Histogram of Oriented Gradients (Hog) Parameters for Facial Recognition. SSRN (2023)