668
Views
3
CrossRef citations to date
0
Altmetric
Statistical Computing and Graphics

Semi-Structured Distributional Regression

ORCID Icon, & ORCID Icon
Pages 88-99 | Received 18 Jul 2022, Accepted 21 Dec 2022, Published online: 10 Feb 2023

References

  • Agarwal, R., Melnick, L., Frosst, N., Zhang, X., Lengerich, B., Caruana, R., and Hinton, G. E. (2021), “Neural Additive Models: Interpretable Machine Learning with Neural Nets,” in Advances in Neural Information Processing Systems (Vol. 34), pp. 4699–4711.
  • Arora, S., Cohen, N., Hu, W., and Luo, Y. (2019), “Implicit Regularization in Deep Matrix Factorization,” in Advances in Neural Information Processing Systems, pp. 7411–7422.
  • Bishop, C. M. (1994), “Mixture Density Networks.” Neural Computing Research Group Report: NCRG/94/004. Birmingham: Aston University.
  • Blei, D. M., Kucukelbir, A., and McAuliffe, J. D. (2017), “Variational Inference: A Review for Statisticians,” Journal of the American Statistical Association, 112, 859–877. DOI: 10.1080/01621459.2017.1285773.
  • Brás-Geraldes, C., Papoila, A., and Xufre, P. (2019), “Generalized Additive Neural Network with Flexible Parametric Link Function: Model Estimation Using Simulated and Real Clinical Data,” Neural Computing and Applications, 31, 719–736. DOI: 10.1007/s00521-017-3105-6.
  • Buja, A., Hastie, T., and Tibshirani, R. (1989), “Linear Smoothers and Additive Models,” The Annals of Statistics, 17, 453–510. DOI: 10.1214/aos/1176347115.
  • Chen, M. L., Doddi, A., Royer, J., Freschi, L., Schito, M., Ezewudo, M., Kohane, I. S., Beam, A., and Farhat, M. (2018), “Deep Learning Predicts Tuberculosis Drug Resistance Status from Whole-Genome Sequencing Data,” BioRxiv, 275628.
  • Cheng, H.-T., Koc, L., Harmsen, J., Shaked, T., Chandra, T., Aradhye, H., Anderson, G., Corrado, G., Chai, W., Ispir, M., Anil, R., Haque, Z., Hong, L., Jain, V., Liu, X., and Shah, H. (2016), “Wide & Deep Learning for Recommender Systems,” in Proceedings of the 1st workshop on Deep Learning for Recommender Systems, pp. 7–10. ACM. DOI: 10.1145/2988450.2988454.
  • Cornwell, C., and Rupert, P. (1988), “Efficient Estimation with Panel Data: An Empirical Comparison of Instrumental Variables Estimators,” Journal of Applied Econometrics, 3, 149–155. DOI: 10.1002/jae.3950030206.
  • Cybenko, G. (1989), “Approximation by Superpositions of a Sigmoidal Function,” Mathematics of Control, Signals and Systems, 2, 303–314. DOI: 10.1007/BF02551274.
  • Dabney, W., Rowland, M., Bellemare, M., and Munos, R. (2018), “Distributional Reinforcement Learning with Quantile Regression,” in Proceedings of the AAAI Conference on Artificial Intelligence (Vol. 32). DOI: 10.1609/aaai.v32i1.11791.
  • Daxberger, E., Kristiadi, A., Immer, A., Eschenhagen, R., Bauer, M., and Hennig, P. (2021), “Laplace Redux – Effortless Bayesian Deep Learning,” in 35th Conference on Neural Information Processing Systems (NeurIPS 2021).
  • De Waal, D. A., and Du Toit, J. V. (2011), “Automation of Generalized Additive Neural Networks for Predictive Data Mining,” Applied Artificial Intelligence, 25, 380–425. DOI: 10.1080/08839514.2011.570156.
  • Erichson, N. B., Voronin, S., Brunton, S. L., and Kutz, J. N. (2019), “Randomized Matrix Decompositions Using r,” Journal of Statistical Software, 89, 1–48. DOI: 10.18637/jss.v089.i11.
  • Groll, A., Hambuckers, J., Kneib, T., and Umlauf, N. (2019), “Lasso-Type Penalization in the Framework of Generalized Additive Models for Location, Scale and Shape,” Computational Statistics & Data Analysis, 140, 59–73. DOI: 10.1016/j.csda.2019.06.005.
  • Hastie, T., and Tibshirani, R. (1986), “Generalized Additive Models,” Statistical Science, 1, 297–318. DOI: 10.1214/ss/1177013604.
  • Hornik, K., Stinchcombe, M., and White, H. (1989), “Multilayer Feedforward Networks are Universal Approximators,” Neural Networks, 2, 359–366. DOI: 10.1016/0893-6080(89)90020-8.
  • Hubin, A., Storvik, G., and Frommlet, F. (2018), “Deep Bayesian Regression Models.” DOI: 10.48550/arXiv.1806.02160.
  • Hüllermeier, E., and Waegeman, W. (2021), “Aleatoric and Epistemic Uncertainty in Machine Learning: An Introduction to Concepts and Methods,” Machine Learning, 110, 457–506. DOI: 10.1007/s10994-021-05946-3.
  • Immer, A., Korzepa, M., and Bauer, M. (2021), “Improving Predictions of Bayesian Neural Nets via Local Linearization,” in International Conference on Artificial Intelligence and Statistics, pp. 703–711. PMLR.
  • Kingma, D. P., and Ba, J. (2014), “Adam: A Method for Stochastic Optimization,” arXiv preprint arXiv:1412.6980.
  • Klein, N., Kneib, T., and Lang, S. (2015), “Bayesian Generalized Additive Models for Location, Scale, and Shape for Zero-Inflated and Overdispersed Count Data,” Journal of the American Statistical Association, 110, 405–419. DOI: 10.1080/01621459.2014.912955.
  • Klein, N., Kneib, T., Lang, S., and Sohn, A. (2015), “Bayesian Structured Additive Distributional Regression with an Application to Regional Income Inequality in Germany,” Annals of Applied Statistics, 9, 1024–1052.
  • Koenker, R. (2005), Quantile Regression, Volume Economic Society Monographs, Cambridge: Cambridge University Press.
  • Koenker, R., and Bassett Jr., G. (1978), “Regression Quantiles,” Econometrica: journal of the Econometric Society, 46, 33–50. DOI: 10.2307/1913643.
  • Kuleshov, V., Fenner, N., and Ermon, S. (2018), “Accurate Uncertainties for Deep Learning Using Calibrated Regression,” in Proceedings of the 35th International Conference on Machine Learning (Vol. 80), pp. 2796–2804.
  • Lakshminarayanan, B., Pritzel, A., and Blundell, C. (2017), “Simple and Scalable Predictive Uncertainty Estimation using Deep Ensembles,” in Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017).
  • Li, R., Reich, B. J., and Bondell, H. D. (2021), “Deep Distribution Regression,” Computational Statistics & Data Analysis, 159, 107203. DOI: 10.1016/j.csda.2021.107203.
  • Maddox, W. J., Izmailov, P., Garipov, T., Vetrov, D. P., and Wilson, A. G. (2019), “A Simple Baseline for Bayesian Uncertainty in Deep Learning,” in Advances in Neural Information Processing Systems (Vol. 32).
  • Mayr, A., Fenske, N., Hofner, B., Kneib, T., and Schmid, M. (2012), “Generalized Additive Models for Location, Scale and Shape for High-Dimensional Data - A Flexible Approach based on Boosting,” Journal of the Royal Statistical Society, Series C, 61, 403–427. DOI: 10.1111/j.1467-9876.2011.01033.x.
  • Nelder, J. A., and Wedderburn, R. W. (1972), “Generalized Linear Models,” Journal of the Royal Statistical Society, Series A, 135, 370–384. DOI: 10.2307/2344614.
  • Pölsterl, S., Sarasua, I., Gutiérrez-Becker, B., and Wachinger, C. (2020), “A Wide and Deep Neural Network for Survival Analysis from Anatomical Shape and Tabular Clinical Data,” in Communications in Computer and Information Science, 453–464. Cham: Springer.
  • Rigby, R. A., and Stasinopoulos, D. M. (2005), “Generalized Additive Models for Location, Scale and Shape,” Journal of the Royal Statistical Society, Series C, 54, 507–554. DOI: 10.1111/j.1467-9876.2005.00510.x.
  • Roberts, D. A. O., and Roberts, L. R. (2020), “Qr and lq Decomposition Matrix Backpropagation Algorithms for Square, Wide, and Deep – Real or Complex – Matrices and their Software Implementation.” DOI: 10.48550/ARXIV.2009.10071.
  • Rügamer, D., Bender, A., Wiegrebe, S., Racek, D., Bischl, B., Müller, C., and Stachl, C. (2022), “Factorized Structured Regression for Large-Scale Varying Coefficient Models,” Accepted at ECML-PKDD 2022.
  • Rügamer, D., Brockhaus, S., Gentsch, K., Scherer, K., and Greven, S. (2018), “Boosting Factor-Specific Functional Historical Models for the Detection of Synchronization in Bioelectrical Signals,” Journal of the Royal Statistical Society, Series C, 67, 621–642. DOI: 10.1111/rssc.12241.
  • Rügamer, D., Pfisterer, F., Bischl, B., and Grün, B. (2022), “Mixture of Experts Distributional Regression: Implementation Using Robust Estimation with Adaptive First-Order Methods,” arXiv:2211.09875.
  • Ruppert, D., Wand, M. P., and Carroll, R. J. (2003), Semiparametric Regression, Cambridge and New York: Cambridge University Press.
  • Sarle, W. S. (1994), “Neural Networks and Statistical Models,” in Proceedings of the Nineteenth Annual SAS Users Group International Conference, Citeseer.
  • Song, H., Diethe, T., Kull, M., and Flach, P. (2019), “Distribution Calibration for Regression,” in Proceedings of the 36th International Conference on Machine Learning, Volume 97 of Proceedings of Machine Learning Research, pp. 5897–5906, PMLR.
  • Tran, M.-N., Nguyen, N., Nott, D., and Kohn, R. (2020), “Bayesian Deep Net GLM and GLMM,” Journal of Computational and Graphical Statistics, 29, 97–113. DOI: 10.1080/10618600.2019.1637747.
  • Umlauf, N., Klein, N., and Zeileis, A. (2018), “Bamlss: Bayesian Additive Models for Location, Scale, and Shape (and Beyond),” Journal of Computational and Graphical Statistics, 27, 612–627. DOI: 10.1080/10618600.2017.1407325.
  • Wood, S. N. (2017), Generalized Additive Models: An Introduction with R, Boca Raton, FL: Chapman and Hall/CRC.

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.