325
Views
1
CrossRef citations to date
0
Altmetric
Research Articles

A Global-Local Approximation Framework for Large-Scale Gaussian Process Modeling

&
Pages 295-305 | Received 15 May 2023, Accepted 30 Nov 2023, Published online: 16 Jan 2024

References

  • Ba, S., and Joseph, V. R. (2012), “Composite Gaussian Process Models for Emulating Expensive Functions,” The Annals of Applied Statistics, 6, 1838–1860. DOI: 10.1214/12-AOAS570.
  • Basak, S., Petit, S., Bect, J., and Vazquez, E. (2022), “Numerical Issues in Maximum Likelihood Parameter Estimation for Gaussian Process Interpolation,” in Machine Learning, Optimization, and Data Science: 7th International Conference, LOD 2021, Grasmere, UK, October 4–8, 2021, Revised Selected Papers, Part II, pp. 116–131, Springer.
  • Chalupka, K., Williams, C. K. I., and Murray, I. (2013), “A Framework for Evaluating Approximation Methods for Gaussian Process Regression,” Journal of Machine Learning Research, 14, 333–350.
  • Chen, T., and Ren, J. (2009), “Bagging for Gaussian Process Regression,” Neurocomputing, 72, 1605–1610. DOI: 10.1016/j.neucom.2008.09.002.
  • Cole, D. A., Christianson, R. B., and Gramacy, R. B. (2021), “Locally Induced Gaussian Processes for Large-Scale Simulation Experiments,” Statistics and Computing, 31, 1–21. DOI: 10.1007/s11222-021-10007-9.
  • Dancik, G. M., and Dorman, K. S. (2008), “mlegp: Statistical Analysis for Computer Models of Biological Systems Using R,” Bioinformatics, 24, 1966–1967. DOI: 10.1093/bioinformatics/btn329.
  • Das, K., and Srivastava, A. N. (2010), “Block-GP: Scalable Gaussian Process Regression for Multimodal Data,” in 2010 IEEE International Conference on Data Mining, pp. 791–796.
  • Deisenroth, M., and Ng, J. W. (2015), “Distributed Gaussian Processes,” in International Conference on Machine Learning, pp. 1481–1490, PMLR.
  • Dette, H., and Pepelyshev, A. (2010), “Generalized Latin Hypercube Design for Computer Experiments,” Technometrics, 52, 421–429. DOI: 10.1198/TECH.2010.09157.
  • Emery, X. (2009), “The kriging Update Equations and their Application to the Selection of Neighboring Data,” Computational Geosciences, 13, 269–280. DOI: 10.1007/s10596-008-9116-8.
  • Fasshauer, G. E. (2007), Meshfree Approximation Methods with MATLAB, New Jersey, NJ: World Scientific.
  • Gardner, J. R., Pleiss, G., Bindel, D., Weinberger, K. Q., and Wilson, A. G. (2018), “GPyTorch: Blackbox Matrix-Matrix Gaussian Process Inference with GPU Acceleration,” in Advances in Neural Information Processing Systems.
  • Gneiting, T. (2002), “Compactly Supported Correlation Functions,” Journal of Multivariate Analysis, 83, 493–508. DOI: 10.1006/jmva.2001.2056.
  • Gramacy, R. B. (2016), “laGP: Large-Scale Spatial Modeling via Local Approximate Gaussian Processes in R,” Journal of Statistical Software, 72, 1–46. DOI: 10.18637/jss.v072.i01.
  • ———(2020), Surrogates: Gaussian Process Modeling, Design and Optimization for the Applied Sciences, Boca Raton, FL: Chapman Hall/CRC.
  • Gramacy, R. B., and Apley, D. W. (2015), “Local Gaussian Process Approximation for Large Computer Experiments,” Journal of Computational and Graphical Statistics, 24, 561–578. DOI: 10.1080/10618600.2014.914442.
  • Gramacy, R. B., and Lee, H. K. (2012), “Cases for the Nugget in Modeling Computer Experiments,” Statistics and Computing, 22, 713–722. DOI: 10.1007/s11222-010-9224-x.
  • Gramacy, R. B., and Lee, H. K. H. (2008), “Bayesian Treed Gaussian Process Models with An Application to Computer Modeling,” Journal of the American Statistical Association, 103, 1119–1130. DOI: 10.1198/016214508000000689.
  • Harari, O., and Steinberg, D. M. (2014), “Convex Combination of Gaussian Processes for Bayesian Analysis of Deterministic Computer Experiments,” Technometrics, 56, 443–454. DOI: 10.1080/00401706.2013.861629.
  • Hensman, J., Matthews, A., and Ghahramani, Z. (2015), “Scalable Variational Gaussian Process Classification,” in Artificial Intelligence and Statistics, pp. 351–360, PMLR.
  • Johnson, S. G. (2007), “The NLopt Nonlinear-Optimization Package,” available at https://github.com/stevengj/nlopt.
  • Joseph, V. R., and Vakayil, A. (2022), “SPlit: An Optimal Method for Data Splitting,” Technometrics, 64, 166–176. DOI: 10.1080/00401706.2021.1921037.
  • Kaufman, C. G., Bingham, D., Habib, S., Heitmann, K., and Frieman, J. A. (2011), “Efficient Emulators of Computer Experiments Using Compactly Supported Correlation Functions, with an Application to Cosmology,” The Annals of Applied Statistics, 5, 2470–2492. DOI: 10.1214/11-AOAS489.
  • Keerthi, S., and Chu, W. (2005), “A Matching Pursuit Approach to Sparse Gaussian Process Regression,” in Advances in Neural Information Processing Systems (Vol. 18), Cambridge, MA: MIT Press.
  • Kenett, R. S., and Zacks, S. (2021), Modern Industrial Statistics: With Applications in R, MINITAB, and JMP, Hoboken, NJ: Wiley.
  • Lawrence, N., Seeger, M., and Herbrich, R. (2002), “Fast Sparse Gaussian Process Methods: The Informative Vector Machine,” in Advances in Neural Information Processing Systems (Vol. 15), MIT Press.
  • Lin, L., Bingham, D., Broekgaarden, F., and Mandel, I. (2021), “Uncertainty Quantification of a Computer Model for Binary Black Hole Formation,” The Annals of Applied Statistics, 15, 1604–1627. DOI: 10.1214/21-AOAS1484.
  • Mak, S., and Joseph, V. R. (2018), “Support Points,” The Annals of Statistics, 46, 2562–2592. DOI: 10.1214/17-AOS1629.
  • Meng, Q., and Ng, S. H. (2015), “An Additive Global and Local Gaussian Process Model for Large Data Sets,” in 2015 Winter Simulation Conference (WSC), pp. 505–516. DOI: 10.1109/WSC.2015.7408191.
  • Morris, M. D., Mitchell, T. J., and Ylvisaker, D. (1993), “Bayesian Design and Analysis of Computer Experiments: Use of Derivatives in Surface Prediction,” Technometrics, 35, 243–255. DOI: 10.1080/00401706.1993.10485320.
  • Park, C., and Apley, D. (2018), “Patchwork kriging for Large-Scale Gaussian Process Regression,” The Journal of Machine Learning Research, 19, 269–311.
  • Park, C., and Huang, J. Z. (2016), “Efficient Computation of Gaussian Process Regression for Large Spatial Data Sets by Patching Local Gaussian Processes,” Journal of Machine Learning Research, 17, 1–29.
  • Quiñonero-Candela, J., and Rasmussen, C. E. (2005), “A Unifying View of Sparse Approximate Gaussian Process Regression,” Journal of Machine Learning Research, 6, 1939–1959.
  • Rasmussen, C., and Ghahramani, Z. (2001), “Infinite Mixtures of Gaussian Process Experts,” in Advances in Neural Information Processing Systems (Vol. 14), Cambridge, MA: MIT Press.
  • Rasmussen, C. E., and Williams, C. K. I. (2005), Gaussian Processes for Machine Learning, Cambridge, MA: The MIT Press.
  • Rowan, T. H. (1990), “Functional Stability Analysis of Numerical Algorithms,” PhD thesis, Department of Computer Science, University of Texas at Austin, Austin, TX.
  • Sacks, J., Welch, W. J., Mitchell, T. J., and Wynn, H. P. (1989), “Design and Analysis of Computer Experiments,” Statistical Science, 4, 409–423. DOI: 10.1214/ss/1177012413.
  • Santner, T. J., Williams, B. J., and Notz, W. I. (2003), The Design and Analysis of Computer Experiments, Springer series in statistics. New York: Springer-Verlag.
  • Snelson, E., and Ghahramani, Z. (2005), “Sparse Gaussian Processes Using Pseudo-Inputs,” in Advances in Neural Information Processing Systems (Vol. 18), MIT Press.
  • ———(2007), “Local and Global Sparse Gaussian Process Approximations,” in Proceedings of the Eleventh International Conference on Artificial Intelligence and Statistics, volume 2 of Proceedings of Machine Learning Research, pp. 524–531, San Juan, Puerto Rico. PMLR.
  • Székely, G. J., and Rizzo, M. L. (2013), “Energy Statistics: A Class of Statistics based on Distances,” Journal of Statistical Planning and Inference, 143, 1249–1272. DOI: 10.1016/j.jspi.2013.03.018.
  • Titsias, M. (2009), “Variational Learning of Inducing Variables in Sparse Gaussian Processes,” in Proceedings of the Twelth International Conference on Artificial Intelligence and Statistics, volume 5 of Proceedings of Machine Learning Research, eds. D. van Dyk and M. Welling, pp. 567–574, Hilton Clearwater Beach Resort, Clearwater Beach, Florida USA. PMLR.
  • Tresp, V. (2000), “A Bayesian Committee Machine,” Neural Computation, 12, 2719–2741. DOI: 10.1162/089976600300014908.
  • Vakayil, A., and Joseph, V. R. (2022), “Data Twinning,” Statistical Analysis and Data Mining: The ASA Data Science Journal, 15, 598–610. DOI: 10.1002/sam.11574.
  • ———(2023), “twingp: A Fast Global-Local Gaussian Process Approximation,” available at https://github.com/avkl/twingp.
  • Vanhatalo, J., and Vehtari, A. (2008), “Modelling Local and Global Phenomena with Sparse Gaussian Processes,” in Proceedings of the Twenty-Fourth Conference on Uncertainty in Artificial Intelligence, UAI’08, pp. 571–578, Arlington, VA: AUAI Press.
  • Vijayakumar, S., and Schaal, S. (2000), “Locally Weighted Projection Regression: An O(n) Algorithm for Incremental Real Time Learning in High Dimensional Space,” in Proceedings of the Seventeenth International Conference on Machine Learning (ICML 2000) (Vol. 1), pp. 288–293, Morgan Kaufmann.
  • Wendland, H. (2004). Scattered Data Approximation. Cambridge Monographs on Applied and Computational Mathematics. Cambridge, UK: Cambridge University Press.

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.