References
Aden-Buie, G., Sievert, C., Cheng, J., & Schloerke, B. (2024).
bslib: Custom Bootstrap
Sass themes for shiny and rmarkdown. https://rstudio.github.io/bslib/
Bates, D., & Maechler, M. (2010). Matrix: Sparse and
dense matrix classes and methods. R Package. https://cran.r-project.org/package=Matrix
Beaumont, M. A. (2010). Approximate Bayesian computation in
evolution and ecology. Annual Review of Ecology, Evolution, and
Systematics, 41, 379–406.
Bengtsson, H. (2021). A unifying framework for parallel and distributed
processing in R using futures. The R Journal,
13(2), 208–227.
Betancourt, M. (2017). A conceptual introduction to
Hamiltonian Monte Carlo.
arXiv:1701.02434.
Booth, J. G., & Hobert, J. P. (1999). Maximizing generalized linear
mixed model likelihoods with an automated Monte Carlo
EM algorithm. Journal of the Royal Statistical Society,
Series B, 61(1), 265–285.
Boyd, S., Parikh, N., Chu, E., Peleato, B., & Eckstein, J. (2011).
Distributed optimization and statistical learning via the alternating
direction method of multipliers. Foundations and Trends in Machine
Learning, 3(1), 1–122.
Boyd, S., & Vandenberghe, L. (2004). Convex optimization.
Cambridge University Press. https://web.stanford.edu/~boyd/cvxbook
Bürkner, P.-C. (2017). brms: An
R package for Bayesian multilevel models using
Stan. In Journal of Statistical Software (No. 1;
Vol. 80).
Candès, E., Fan, Y., Janson, L., & Lv, J. (2018). Panning for gold:
Model-X knockoffs for high dimensional controlled variable
selection. Journal of the Royal Statistical Society, Series B,
80(3), 551–577.
Carpenter, B., Gelman, A., Hoffman, M. D., Lee, D., Goodrich, B.,
Betancourt, M., Brubaker, M., Guo, J., Li, P., & Riddell, A. (2017).
Stan: A probabilistic programming language. Journal of
Statistical Software, 76(1).
Chang, W., Cheng, J., Allaire, J., Sievert, C., Schloerke, B., Xie, Y.,
Allen, J., McPherson, J., Dipert, A., & Borges, B. (2024). shiny: Web application framework for
R. https://shiny.posit.co/
Chang, W., Luraschi, J., & Mastny, T. (2024). profvis: Interactive visualizations for profiling
R code. https://rstudio.github.io/profvis/
Chen, T., & Guestrin, C. (2016). XGBoost: A scalable
tree boosting system. Proceedings of the 22nd ACM SIGKDD
International Conference on Knowledge Discovery and Data Mining
(KDD), 785–794.
Dempster, A. P., Laird, N. M., & Rubin, D. B. (1977). Maximum
likelihood from incomplete data via the EM algorithm.
Journal of the Royal Statistical Society, Series B,
39(1), 1–38.
Dipert, A. (2024). shinyloadtest: Load
test Shiny applications. https://rstudio.github.io/shinyloadtest/
Eddelbuettel, D. (2013). Seamless R and
C++ integration with Rcpp. Springer.
Eddelbuettel, D., & François, R. (2011). Rcpp: Seamless
R and C++ integration. Journal of
Statistical Software, 40(8), 1–18.
Eddelbuettel, D., & Sanderson, C. (2014).
RcppArmadillo: Accelerating R with
high-performance C++ linear algebra. Computational
Statistics and Data Analysis, 71, 1054–1063.
El Ghaoui, L., Viallon, V., & Rabbani, T. (2012). Safe feature
elimination for the LASSO and sparse supervised learning
problems. Pacific Journal of Optimization, 8(4),
667–698.
Fan, J., & Li, R. (2001). Variable selection via nonconcave
penalized likelihood and its oracle properties. Journal of the
American Statistical Association, 96(456), 1348–1360.
Few, S. (2013). Information dashboard design: Displaying data for
at-a-glance monitoring (2nd ed.). Analytics Press.
Ge, H., Xu, K., & Ghahramani, Z. (2018). Turing: A
language for flexible probabilistic inference. AISTATS.
Geer, S. van de, Bühlmann, P., Ritov, Y., & Dezeure, R. (2014). On
asymptotically optimal confidence regions and tests for high-dimensional
models. Annals of Statistics, 42(3), 1166–1202.
Gelman, A., Vehtari, A., Simpson, D., Margossian, C. C., Carpenter, B.,
Yao, Y., Kennedy, L., Gabry, J., Bürkner, P.-C., & Modrák, M.
(2020). Bayesian workflow. arXiv:2011.01808.
Goldberg, D. (1991). What every computer scientist should know about
floating-point arithmetic. ACM Computing Surveys,
23(1), 5–48.
Golub, G. H., & Van Loan, C. F. (2013). Matrix computations
(4th ed.). Johns Hopkins University Press.
Goodrich, B., Gabry, J., Ali, I., & Brilleman, S. (2024). rstanarm: Bayesian applied regression
modeling via Stan. https://mc-stan.org/rstanarm/
Hastie, T., Tibshirani, R., & Friedman, J. (2009). The elements
of statistical learning (2nd ed.). Springer.
Hastie, T., Tibshirani, R., & Wainwright, M. (2015). Statistical
learning with sparsity: The Lasso and generalizations.
Chapman; Hall/CRC.
Higham, N. J. (2002). Accuracy and stability of numerical
algorithms (2nd ed.). SIAM.
Hoffman, M. D., & Gelman, A. (2014). The No-U-Turn
sampler: Adaptively setting path lengths in Hamiltonian
Monte Carlo. Journal of Machine Learning
Research, 15, 1593–1623.
Kucukelbir, A., Tran, D., Ranganath, R., Gelman, A., & Blei, D. M.
(2017). Automatic differentiation variational inference. Journal of
Machine Learning Research, 18(14), 1–45.
Kuhn, M., & Silge, J. (2022). Tidy modeling with
R. O’Reilly Media. https://www.tmwr.org
Lee, J. D., Sun, D. L., Sun, Y., & Taylor, J. E. (2016). Exact
post-selection inference, with application to the lasso. Annals of
Statistics, 44(3), 907–927.
Lundberg, S. M., & Lee, S.-I. (2017). A unified approach to
interpreting model predictions. Advances in Neural Information
Processing Systems (NeurIPS), 4765–4774.
McElreath, R. (2020). Statistical rethinking: A
Bayesian course with examples in R and
Stan (2nd ed.). Chapman; Hall/CRC.
McLachlan, G. J., & Krishnan, T. (2008). The EM
algorithm and extensions (2nd ed.). Wiley.
Meinshausen, N., & Bühlmann, P. (2010). Stability selection.
Journal of the Royal Statistical Society, Series B,
72(4), 417–473.
Meinshausen, N., Meier, L., & Bühlmann, P. (2009).
P-values for high-dimensional regression. Journal of
the American Statistical Association, 104(488), 1671–1681.
Meng, X.-L., & Rubin, D. B. (1993). Maximum likelihood estimation
via the ECM algorithm: A general framework.
Biometrika, 80(2), 267–278.
Munzner, T. (2014). Visualization analysis and design. CRC
Press.
Nocedal, J., & Wright, S. J. (2006). Numerical optimization
(2nd ed.). Springer.
Owen, A. B. (2013). Monte carlo theory, methods and examples.
https://artowen.su.domains/mc/
Parikh, N., & Boyd, S. (2014). Proximal algorithms. Foundations
and Trends in Optimization, 1(3), 127–239.
Posit Software, PBC. (2024). Quarto dashboards. https://quarto.org/docs/dashboards/
Raasveldt, M., & Mühleisen, H. (2019). DuckDB: An
embeddable analytical database. Proceedings of the 2019
International Conference on Management of Data (SIGMOD), 1981–1984.
Robert, C. P., & Casella, G. (2004). Monte carlo statistical
methods (2nd ed.). Springer.
Rue, H., Martino, S., & Chopin, N. (2009). Approximate
Bayesian inference for latent Gaussian models
by using integrated nested Laplace approximations.
Journal of the Royal Statistical Society, Series B,
71(2), 319–392.
Saad, Y. (2003). Iterative methods for sparse linear systems
(2nd ed.). SIAM.
Salvatier, J., Wiecki, T. V., & Fonnesbeck, C. (2016). Probabilistic
programming in Python using PyMC3. PeerJ
Computer Science, 2, e55.
Taylor, J., & Tibshirani, R. (2018). Post-selection inference for
ℓ1-
penalized likelihood models. Canadian Journal of Statistics,
46(1), 41–61.
Tibshirani, R. (1996). Regression shrinkage and selection via the lasso.
Journal of the Royal Statistical Society, Series B,
58(1), 267–288.
Tibshirani, R., Bien, J., Friedman, J., Hastie, T., Simon, N., Taylor,
J., & Tibshirani, R. J. (2012). Strong rules for discarding
predictors in lasso-type problems. Journal of the Royal Statistical
Society, Series B, 74(2), 245–266.
Trefethen, L. N., & Bau III, D. (1997). Numerical linear
algebra. SIAM.
Van Calster, B., McLernon, D. J., Smeden, M. van, Wynants, L.,
Steyerberg, E. W., & Topic Group ‘Evaluating diagnostic tests and
prediction models’ of the STRATOS initiative, on behalf of. (2019).
Calibration: The Achilles heel of predictive analytics.
BMC Medicine, 17(1), 230.
Vehtari, A., Gelman, A., & Gabry, J. (2017). Practical
Bayesian model evaluation using leave-one-out
cross-validation and WAIC. Statistics and
Computing, 27, 1413–1432.
Vehtari, A., Gelman, A., Simpson, D., Carpenter, B., & Bürkner,
P.-C. (2021). Rank-normalization, folding, and localization: An improved
R̂ for assessing
convergence of MCMC. Bayesian Analysis,
16(2), 667–718.
Wang, J., Wonka, P., & Ye, J. (2015). Lasso screening rules via dual
polytope projection. Journal of Machine Learning Research,
16, 1063–1101.
Watanabe, S. (2010). Asymptotic equivalence of Bayes cross
validation and widely applicable information criterion in singular
learning theory. Journal of Machine Learning Research,
11, 3571–3594.
Wickham, H. (2021). Mastering Shiny. O’Reilly
Media. https://mastering-shiny.org
Wickham, H., & Bryan, J. (2023). R packages
(2nd ed.). O’Reilly Media. https://r-pkgs.org
Yuan, M., & Lin, Y. (2006). Model selection and estimation in
regression with grouped variables. Journal of the Royal Statistical
Society, Series B, 68(1), 49–67.
Zeng, Y., & Breheny, P. (2017). The biglasso package: A memory- and
computation-efficient solver for lasso model fitting with big data in
R. arXiv:1701.05936.
Zhang, C.-H. (2010). Nearly unbiased variable selection under minimax
concave penalty. Annals of Statistics, 38(2), 894–942.
Zhang, C.-H., & Zhang, S. S. (2014). Confidence intervals for low
dimensional parameters in high dimensional linear models. Journal of
the Royal Statistical Society, Series B, 76(1), 217–242.
Zhang, L., Carpenter, B., Gelman, A., & Vehtari, A. (2022).
Pathfinder: Parallel quasi-Newton variational
inference. Journal of Machine Learning Research,
23(306), 1–49.
Zou, H. (2006). The adaptive lasso and its oracle properties.
Journal of the American Statistical Association,
101(476), 1418–1429.
Zou, H., & Hastie, T. (2005). Regularization and variable selection
via the elastic net. Journal of the Royal Statistical Society,
Series B, 67(2), 301–320.