ssp.softmax
:
Subsampling for Softmax (Multinomial) Regression Modelvignettes/ssp-softmax.Rmd
ssp-softmax.Rmd
This vignette introduces the usage of ssp.softmax
, which
draws optimal subsample from full data and fit softmax (multinomial)
regression on the subsample. The statistical theory and algorithms in
this implementation can be found in the relevant reference papers.
Denote \(y\) as multi-category response variable and \(K+1\) is the number of categories. \(N\) is the number of observations in the full dataset. \(X\) is the \(N \times d\) covariates matrix. Softmax regression model assumes that \[ P(y_{i,k} = 1 \mid \mathbf{x}_i) = \frac{\exp(\mathbf{x}_i^\top \boldsymbol{\beta}_k)}{\sum_{l=0}^{K} \exp(\mathbf{x}_i^\top \boldsymbol{\beta}_l)} \] for \(i = 1, \ldots, N\) and \(k = 0, 1, \ldots, K\), where \(\boldsymbol{\beta}_k\)’s are \(d\)-dimensional unknown coefficients.
The log-likelihood function of softmax regression is
\[ \max_{\beta} L(\beta) = \frac{1}{N} \sum_{i=1}^{N} \left[ \sum_{k=0}^{K} y_{i,k} \mathbf{x}_i^\top \boldsymbol{\beta}_k - \ln \left\{ \sum_{l=0}^{K} \exp(\mathbf{x}_i^\top \boldsymbol{\beta}_l) \right\} \right]. \]
The idea of subsampling methods is as follows: instead of fitting the model on the size \(N\) full dataset, a subsampling probability is assigned to each observation and a smaller, informative subsample is drawn. The model is then fitted on the subsample to obtain an estimator with reduced computational cost.
Full dataset: The whole dataset used as input.
Full data estimator: The estimator obtained by fitting the model on the full dataset.
Subsample: A subset of observations drawn from the full dataset.
Subsample estimator: The estimator obtained by fitting the model on the subsample.
Subsampling probability (\(\pi\)): The probability assigned to each observation for inclusion in the subsample.
We introduce the usage of ssp.softmax
with simulated
data. \(X\) contains \(d=3\) covariates drawn from multinormal
distribution and \(Y\) is the
multicategory response variable with \(K+1=3\) categories. The full data size is
\(N = 1 \times
10^4\).
set.seed(1)
d <- 3
K <- 2
G <- rbind(rep(-1/(K+1), K), diag(K) - 1/(K+1)) %x% diag(d)
N <- 1e4
beta.true.baseline <- cbind(rep(0, d), matrix(-1.5, d, K))
beta.true.summation <- cbind(rep(1, d), 0.5 * matrix(-1, d, K))
mu <- rep(0, d)
sigma <- matrix(0.5, nrow = d, ncol = d)
diag(sigma) <- rep(1, d)
X <- MASS::mvrnorm(N, mu, sigma)
prob <- exp(X %*% beta.true.summation)
prob <- prob / rowSums(prob)
Y <- apply(prob, 1, function(row) sample(0:K, size = 1, prob = row))
data <- as.data.frame(cbind(Y, X))
colnames(data) <- c("Y", paste("V", 1:ncol(X), sep=""))
head(data)
#> Y V1 V2 V3
#> 1 2 -0.3756189 -0.17727086 -0.9816025
#> 2 2 0.2912939 0.60753208 -0.4489936
#> 3 2 -1.0530547 0.02079337 -1.0146024
#> 4 0 0.1854791 2.45385260 1.2682922
#> 5 0 0.8687332 0.21941612 -0.2810234
#> 6 1 -0.8336174 -0.32556141 -0.8505501
The function usage is
ssp.softmax(
formula,
data,
subset,
n.plt,
n.ssp,
criterion = "MSPE",
sampling.method = "poisson",
likelihood = "MSCLE",
constraint = "summation",
control = list(...),
contrasts = NULL,
...
)
The core functionality of ssp.softmax
revolves around
three key questions:
How are subsampling probabilities computed? (Controlled by the
criterion
argument)
How is the subsample drawn? (Controlled by the
sampling.method
argument)
How is the likelihood adjusted to correct for bias? (Controlled
by the likelihood
argument)
criterion
The choices of criterion
include optA
,
optL
, ,MSPE
(default), LUC
and
uniform
. The default criterion MSPE
minimizes
the mean squared prediction error between subsample estimator and full
data estimator. Criterion optA
and optL
are
derived by minimizing the asymptotic covariance of subsample estimator.
LUC
and uniform
are baseline methods. See
Yao, Zou, and Wang (2023) and Wang and Kim (2022) for details.
sampling.method
The options for sampling.method
include
withReplacement
and poisson
(default).
withReplacement.
stands for drawing \(n.ssp\) subsamples from full dataset with
replacement, using the specified subsampling probability.
poisson
stands for drawing subsamples one by one by
comparing the subsampling probability with a realization of uniform
random variable \(U(0,1)\). The
expected number of drawed samples are \(n.ssp\).
likelihood
The available choices for likelihood
include
weighted
and MSCLE
(default).
MSCLE
stands for maximum sampled conditional likelihood.
Both of these likelihood functions can derive an unbiased optimal
subsample estimator. See Wang and Kim
(2022) for details about MSCLE
.
constraint
Softmax model needs constraint on unknown coefficients for
identifiability. The options for constraint
include
summation
and baseline
(default). The baseline
constraint assumes the coefficient for the baseline category are \(0\). Without loss of generality,
ssp.softmax
sets the category \(Y=0\) as the baseline category so that
\(\boldsymbol{\beta}_0=0\). The
summation constraint \(\sum_{k=0}^{K}
\boldsymbol{\beta}_k\) can also used in the subsampling method
for the purpose of calculating optimal subsampling probability. These
two constraints lead to different interpretation of coefficients but are
equal for computing \(P(y_{i,k} = 1 \mid
\mathbf{x}_i)\). The estimation of coefficients returned by
ssp.softmax()
is under baseline constraint.
After drawing subsample, ssp.softmax
utilizes
nnet::multinom
to fit the model on the subsample. Arguments
accepted by nnet::multinom
can be passed through
...
in ssp.softmax
.
Below are two examples demonstrating the use of
ssp.softmax
with different configurations.
n.plt <- 200
n.ssp <- 600
formula <- Y ~ . -1
ssp.results1 <- ssp.softmax(formula = formula,
data = data,
n.plt = n.plt,
n.ssp = n.ssp,
criterion = 'MSPE',
sampling.method = 'withReplacement',
likelihood = 'weighted',
constraint = 'baseline'
)
#> [1] "Message from nnet::multinom: "
#> # weights: 12 (6 variable)
#> initial value 219.722458
#> iter 10 value 143.304799
#> final value 143.304778
#> converged
#> [1] "Message from nnet::multinom: "
#> # weights: 12 (6 variable)
#> initial value 5789394.068697
#> iter 10 value 4219770.981211
#> final value 4219003.340996
#> converged
summary(ssp.results1)
#> Model Summary
#>
#>
#> Call:
#>
#> ssp.softmax(formula = formula, data = data, n.plt = n.plt, n.ssp = n.ssp,
#> criterion = "MSPE", sampling.method = "withReplacement",
#> likelihood = "weighted", constraint = "baseline")
#>
#> Subsample Size:
#>
#> 1 Total Sample Size 10000
#> 2 Expected Subsample Size 600
#> 3 Actual Subsample Size 600
#> 4 Unique Subsample Size 563
#> 5 Expected Subample Rate 6%
#> 6 Actual Subample Rate 6%
#> 7 Unique Subample Rate 5.63%
#>
#> Coefficients:
#>
#> [,1] [,2]
#> V1 -1.175907 -1.281935
#> V2 -1.619482 -1.502412
#> V3 -1.231662 -1.414723
#>
#> Std. Errors:
#>
#> [,1] [,2]
#> V1 0.1457377 0.1459830
#> V2 0.1442445 0.1461619
#> V3 0.1498036 0.1492071
summary(ssp.results1)
shows that it draws 600
observations out of 10000, where the number of unique indices is less
than 600 since we use sampling.method = 'withReplacement'
.
After fitting softmax model on subsample using the choosen
weighted
likelihood function, we get coefficients
estimation and standard errors as above.
ssp.results2 <- ssp.softmax(formula = formula,
data = data,
n.plt = n.plt,
n.ssp = n.ssp,
criterion = 'MSPE',
sampling.method = 'poisson',
likelihood = 'MSCLE',
constraint = 'baseline'
)
#> [1] "Message from nnet::multinom: "
#> # weights: 12 (6 variable)
#> initial value 219.722458
#> iter 10 value 126.651711
#> final value 126.637171
#> converged
#> [1] "Message from nnet::multinom: "
#> # weights: 21 (6 variable)
#> initial value 1155.595717
#> iter 10 value 694.729584
#> final value 694.727136
#> converged
summary(ssp.results2)
#> Model Summary
#>
#>
#> Call:
#>
#> ssp.softmax(formula = formula, data = data, n.plt = n.plt, n.ssp = n.ssp,
#> criterion = "MSPE", sampling.method = "poisson", likelihood = "MSCLE",
#> constraint = "baseline")
#>
#> Subsample Size:
#>
#> 1 Total Sample Size 10000
#> 2 Expected Subsample Size 600
#> 3 Actual Subsample Size 702
#> 4 Unique Subsample Size 702
#> 5 Expected Subample Rate 6%
#> 6 Actual Subample Rate 7.02%
#> 7 Unique Subample Rate 7.02%
#>
#> Coefficients:
#>
#> [,1] [,2]
#> V1 -1.451076 -1.433182
#> V2 -1.780664 -1.735468
#> V3 -1.445315 -1.632805
#>
#> Std. Errors:
#>
#> [,1] [,2]
#> V1 0.1191177 0.1229054
#> V2 0.1320770 0.1283766
#> V3 0.1243176 0.1266856
The returned object contains estimation results and index of drawn subsamples in the full dataset.
names(ssp.results1)
#> [1] "model.call" "coef.plt" "coef.ssp"
#> [4] "coef" "coef.plt.sum" "coef.ssp.sum"
#> [7] "coef.sum" "cov.plt" "cov.ssp"
#> [10] "cov" "cov.plt.sum" "cov.sum"
#> [13] "cov.ssp.sum" "index.plt" "index.ssp"
#> [16] "N" "subsample.size.expect" "terms"
Some key returned variables:
index.plt
and index
are the row indices
of drawn pilot subsamples and optimal subsamples in the full
data.
coef.ssp
is the subsample estimator for \(\beta\) and coef
is the linear
combination of coef.plt
(pilot estimator) and
coef.ssp
.
cov.ssp
and cov
are estimated
covariance matrices of coef.ssp
and
coef
.