glmbayes provides independent and identically distributed (iid) samples for Bayesian Generalized Linear Models (GLMs). Its primary interface, glmb(), serves as a Bayesian analogue to R’s glm() function, supporting Gaussian, Poisson, Binomial, and Gamma families under log-concave likelihoods. Sampling for most models is performed using accept-reject methods based on likelihood subgradients (Nygren and Nygren, 2006). For Gaussian models, the package also includes lmb(), a Bayesian counterpart to R’s lm().
The package includes a rich set of supporting tools for prior specification, model diagnostics, and method functions that mirror those for lm() and glm(). Most functions are extensively documented, and a comprehensive set of vignettes are available to guide users through the package’s capabilities.
The package is currently available on R-Universe, with plans for a future CRAN submission. For recent updates and planned enhancements, see https://github.com/knygren/glmbayes/blob/main/NEWS.md
To install the current development version (excluding OpenCL functionality):
install.packages(“glmbayes”, repos = c(“https://cloud.r-project.org”, “https://knygren.r-universe.dev”))
To install a version suitable for large models with GPU acceleration, follow the instructions from
Chapter 12 - Large Models: GPU Acceleration using OpenCL https://knygren.r-universe.dev/articles/glmbayes/Chapter-12.html
library(glmbayes)
# Dobson (1990), p. 93: Randomized Controlled Trial
counts <- c(18,17,15,20,10,20,25,13,12)
outcome <- gl(3,1,9)
treatment <- gl(3,3)
print(d.AD <- data.frame(treatment, outcome, counts))
## Classical glm
glm.D93 <- glm(counts ~ outcome + treatment,
family = poisson())
## Bayesian glmb
# Step 1: Set up prior
ps <- Prior_Setup(counts ~ outcome + treatment, family = poisson())
mu <- ps$mu
V <- ps$Sigma
# Step 2: Fit using glmb
glmb.D93 <- glmb(counts ~ outcome + treatment,
family = poisson(),
pfamily = dNormal(mu = mu, Sigma = V))
summary(glmb.D93)
As with glm(), models are defined by a formula for the
linear predictor and a family() describing the likelihood
and link. In addition, glmb() requires a
pfamily object specifying the prior.
The supported likelihood families, link functions, and compatible pfamilies are:
| Likelihood family | Link functions | Compatible pfamilies |
|---|---|---|
| Gaussian | identity | dNormal, dGamma, dNormal_Gamma, dIndependent_Normal_Gamma |
| Poisson / Quasi-Poisson | log | dNormal |
| Binomial / Quasi-Binomial | logit, probit, cloglog | dNormal |
| Gamma | log | dNormal, dGamma |
For a default, data‑aligned prior using the same formula and family
as glm(), call
Prior_Setup(formula, family, data = ..., ...). The returned
list includes default settings for the following:
dispersion for use with the dNormal()
prior (gaussian and Gamma families)Sigma_0, shape and rate for
use with the dNormal_Gamma() priorshape_ING and rate for use with
dIndependent_Normal_Gamma() priorshape, rate_gamma and
coefficients for use with the dGamma()
priorOptional arguments adjust prior weight, centering, and related settings (see the function help and vignette Chapter 03).
Assuming ps <- Prior_Setup(...):
All non‑Gaussian families:
Use dNormal(mu = ps$mu, Sigma = ps$Sigma).
(For Gamma GLMs, also supply dispersion from the fitted GLM
or from ps; see example("glmb").)
Gaussian — normal prior with known
dispersion:
Use
dNormal(mu = ps$mu, Sigma = ps$Sigma, dispersion = ps$dispersion).
Gaussian — conjugate Normal–Gamma:
Use
dNormal_Gamma(mu = ps$mu, Sigma_0 = ps$Sigma_0, shape = ps$shape, rate = ps$rate).
Gaussian — independent Normal–Gamma:
Use
dIndependent_Normal_Gamma(mu = ps$mu, Sigma = ps$Sigma, shape = ps$shape_ING, rate = ps$rate).
Gaussian — dispersion via dGamma (coefficients
fixed):
With
rate_dg <- if (!is.null(ps$rate_gamma)) ps$rate_gamma else ps$rate,
use
dGamma(shape = ps$shape, rate = rate_dg, beta = ps$coefficients).
The default priors have limiting behaviors that produce estimates resembling classical estimates as priors get weak (see documentation and vignettes for details).
All supported models have log‑concave likelihoods, enabling efficient iid sampling via enveloping functions and subgradient‑based accept–reject algorithms, especially for models lacking standard iid samplers.
Use example() and demo() to explore
built-in examples and demos for supported families and links:
## Bayesian linear regression
example("lmb")
## Bayesian generalized linear models
example("glmb")
## Predictions for fitted glmb objects (newdata, type, etc.)
example("predict.glmb")
## Deviance residuals and simulate() for posterior predictive checks (menarche)
example("residuals.glmb")
## Two-block Gibbs sampler compared with iid sampling (linear model)
example("rlmb")
## Default prior specification using Prior_Setup
example("Prior_Setup")
## Matrix-input GLM example with an informative prior
example("rglmb")
## Two-step Boston example: estimates and summarizes models with unknown
## dispersion using dGamma priors via rGamma_reg, rglmb, rlmb, glmb, and lmb
example("summary.rGamma_reg")
## High-dimensional Gaussian model (14 predictors) with GPU acceleration (requires OpenCL)
example("Boston_centered")
## High-dimensional binomial model (14 predictors) with GPU acceleration (requires OpenCL)
example("Cleveland")
## Hierarchical linear model (Rubin/Gelman 8-schools) via rlmb
demo("Ex_07_Schools")
## Hierarchical generalized linear model (Poisson BikeSharing) via rglmb
demo("Ex_09_BikeSharingPoisson")
## Detailed simulation pipeline for rNormalGLM models (JASA 2006; Vignette Chapter A05)
example("rNormalGLM_std")
## Detailed simulation pipeline for rIndepNormalGammaReg models (Vignette Chapter A07)
example("rIndepNormalGammaReg_std")
For generalized linear models where well known sampling methods are
unavailable, sampling follows the framework from Nygren and Nygren
(2006), using likelihood subgradients to construct enveloping functions
for the posterior distribution. When the posterior is approximately
normal, the expected number of draws per acceptance is bounded as per
that paper and as discussed in our vignettes. Dispersion can be sampled
via rGamma_reg() (standalone) or jointly with coefficients
via rNormalGamma_reg() and
rindepNormalGamma_reg().
The implemented algorithms tend to have acceptable performance on CPUs up to around 10-14 dimensions. For larger models, the envelope construction is embarrassingly parallel. To accelerate envelope construction in such cases, the package provides optional GPU acceleration using OpenCL. This requires that users have GPU enabled machines and an OpenCL installation. These features are discussed in more detail in two of our vignettes.
The glmbayes package includes a comprehensive set of vignettes organized into five major parts. These vignettes guide users from introductory material through applied modeling, advanced topics, and the underlying simulation methods that support the package.
Overview of the package, its design philosophy, and the basic workflow for fitting Bayesian linear and generalized linear models. It introduces the core functions, model objects, and the structure of the modeling interface.
Chapter 00 - Introduction
https://knygren.r-universe.dev/articles/glmbayes/Chapter-00.html
Chapter 01 - Getting Started with glmbayes
https://knygren.r-universe.dev/articles/glmbayes/Chapter-01.html
These chapters focus on Bayesian linear regression using the Gaussian family. Topics include model fitting, prior construction, posterior summaries, predictions, and deviance residuals. This part establishes the foundation for understanding the Bayesian GLM framework used throughout the package.
Chapter 02 - Estimating Bayesian Linear
Models
https://knygren.r-universe.dev/articles/glmbayes/Chapter-02.html
Chapter 03 - Tailoring Priors - Leveraging the
Prior_Setup Function
https://knygren.r-universe.dev/articles/glmbayes/Chapter-03.html
Chapter 04 - Reviewing Model Predictions, Deviance
Residuals and Model Statistics
https://knygren.r-universe.dev/articles/glmbayes/Chapter-04.html
This part presents Bayesian GLMs across the major likelihood families, including binomial, quasi-binomial, Poisson, quasi-Poisson, and Gamma models. It covers model specification, link functions, log-concavity, diagnostics, and interpretation of posterior results.
Chapter 05 - Foundations of GLMs - Families, Links, and
Log-Concave Likelihoods
https://knygren.r-universe.dev/articles/glmbayes/Chapter-05.html
Chapter 06 - Estimating Bayesian Generalized Linear
Models
https://knygren.r-universe.dev/articles/glmbayes/Chapter-06.html
Chapter 07 - Models for the Binomial
Family
https://knygren.r-universe.dev/articles/glmbayes/Chapter-07.html
Chapter 08 - Models for the Poisson Family
https://knygren.r-universe.dev/articles/glmbayes/Chapter-08.html
Chapter 09 - Models for the Gamma Family
https://knygren.r-universe.dev/articles/glmbayes/Chapter-09.html
These chapters explore more complex modeling scenarios and computational strategies, such as informative priors, two-block Gibbs sampling, hierarchical linear and generalized linear models, models with unknown dispersion parameters, and large-scale model fitting using GPU acceleration using OpenCL.
Chapter 10 - Informative Priors: Centering and priors
with differential prior weights
https://knygren.r-universe.dev/articles/glmbayes/Chapter-10.html
Chapter 11 - Estimating Models with unknown dispersion
parameters
https://knygren.r-universe.dev/articles/glmbayes/Chapter-11.html
Chapter 12 - Large Models: GPU Acceleration using
OpenCL
https://knygren.r-universe.dev/articles/glmbayes/Chapter-12.html
Chapter 13 - Hierarchical Linear Models
https://knygren.r-universe.dev/articles/glmbayes/Chapter-13.html
Chapter 14 - Hierarchical Generalized Linear
Models
https://knygren.r-universe.dev/articles/glmbayes/Chapter-14.html
This part documents the mathematical and algorithmic foundations of the package. Topics include estimation procedures, likelihood subgradient densities, envelope construction, accept-reject sampling, and technical reports on sampler design including implementation aspects for GPU acceleration using OpenCL.
Chapter A01 - A detailed overview of the glmbayes
package
https://knygren.r-universe.dev/articles/glmbayes/Chapter-A01.html
Chapter A02 - Overview of Estimation
Procedures
https://knygren.r-universe.dev/articles/glmbayes/Chapter-A02.html
Chapter A03 - Methods Available in
glmbayes
https://knygren.r-universe.dev/articles/glmbayes/Chapter-A03.html
Chapter A04 - Directional Tail Diagnostics for
Prior-Posterior Disagreement
https://knygren.r-universe.dev/articles/glmbayes/Chapter-A04.html
Chapter A05 - Simulation Methods - Likelihood Subgradient
Densities
https://knygren.r-universe.dev/articles/glmbayes/Chapter-A05.html
Chapter A06 - Accept-Reject Sampling for Dispersion in
Gamma Regression
https://knygren.r-universe.dev/articles/glmbayes/Chapter-A06.html
Chapter A07 - Accept-Reject Sampling for gaussian
Regression models with independent normal-gamma priors
https://knygren.r-universe.dev/articles/glmbayes/Chapter-A07.html
Chapter A08 - Overview of Envelope Related
Functions
https://knygren.r-universe.dev/articles/glmbayes/Chapter-A08.html
Chapter A09 - Parallel Sampling Implementation using
RcppParallel
https://knygren.r-universe.dev/articles/glmbayes/Chapter-A09.html
Chapter A10 - Accelerated EnvelopeBuild Implementation
using OpenCL
https://knygren.r-universe.dev/articles/glmbayes/Chapter-A10.html
Chapter A11 - Implementation Companion for Independent
Normal-Gamma
https://knygren.r-universe.dev/articles/glmbayes/Chapter-A11.html
Chapter A12 - Technical Derivations for Priors Returned
by Prior_Setup()
https://knygren.r-universe.dev/articles/glmbayes/Chapter-A12.html
Together, these vignettes form a comprehensive reference that supports users at all levels, from first-time Bayesian GLM users to researchers interested in the mathematical and computational details behind the samplers.
log(y!) once per observation and reuse it in both
OpenCL envelope construction and accept-reject simulation, since it
depends only on the response, to reduce redundant lgamma
evaluation and improve performance for large Poisson models.