Title: | Covariance Inference and Decompositions for Tensor Datasets |
---|---|
Description: | A collection of functions for Kronecker structured covariance estimation and testing under the array normal model. For estimation, maximum likelihood and Bayesian equivariant estimation procedures are implemented. For testing, a likelihood ratio testing procedure is available. This package also contains additional functions for manipulating and decomposing tensor data sets. This work was partially supported by NSF grant DMS-1505136. Details of the methods are described in Gerard and Hoff (2015) <doi:10.1016/j.jmva.2015.01.020> and Gerard and Hoff (2016) <doi:10.1016/j.laa.2016.04.033>. |
Authors: | David Gerard [aut, cre] , Peter Hoff [aut] |
Maintainer: | David Gerard <[email protected]> |
License: | GPL-3 |
Version: | 1.0.2 |
Built: | 2024-10-23 03:14:11 UTC |
Source: | https://github.com/dcgerard/tensr |
-mode product.amprod
returns the -mode product of an array with a
matrix.
amprod(A, M, k)
amprod(A, M, k)
A |
A real valued array. |
M |
A real matrix. |
k |
An integer. The mode along which |
The -mode product of a tensor
with a matrix
results in a tensor whose
-mode unfolding is
times
the
-mode unfolding of
. That is
mat(amprod(A,M,k)) = M %*% mat(A,k)
. More details of the
-mode product can be found in
Kolda and
Bader (2009).
An array whose -mode unfolding is
M %*%
mat(A,k)
.
Peter Hoff.
Kolda, T. G., & Bader, B. W. (2009). Tensor decompositions and applications. SIAM review, 51(3), 455-500.
atrans
for applying multiple -mode
products.
A <- array(1:8, dim = c(2,2,2)) M <- matrix(1:4, nrow = 2, ncol = 2) Y <- amprod(A, M, 2) Y identical(M %*% mat(A,2), mat(Y,2))
A <- array(1:8, dim = c(2,2,2)) M <- matrix(1:4, nrow = 2, ncol = 2) Y <- amprod(A, M, 2) Y identical(M %*% mat(A,2), mat(Y,2))
Conditional mean and variance of a subarray.
anorm_cd(Y, M, S, saidx)
anorm_cd(Y, M, S, saidx)
Y |
A real valued array. |
M |
Mean of |
S |
List of mode-specific covariance matrices of |
saidx |
List of indices for indexing sub-array for which the conditional
mean and variance should be computed. For example, |
This function calculates the conditional mean and variance in the array
normal model. Let be array normal and let
be a subarray of
. Then this function will calculate the conditional means and
variances of
, conditional on every other element in
.
Peter Hoff.
Hoff, P. D. (2011). Separable covariance arrays via the Tucker product, with applications to multivariate relational data. Bayesian Analysis, 6(2), 179-196.
p <- c(4, 4, 4) Y <- array(stats::rnorm(prod(p)), dim = p) saidx <- list(1:2, 1:2, 1:2) true_cov <- tensr::start_ident(p) true_mean <- array(0, dim = p) cond_params <- anorm_cd(Y = Y, M = true_mean, S = true_cov, saidx = saidx) ## Since data are independent standard normals, conditional mean is 0 and ## conditional covariance matrices are identities. cond_params$Mab cond_params$Sab
p <- c(4, 4, 4) Y <- array(stats::rnorm(prod(p)), dim = p) saidx <- list(1:2, 1:2, 1:2) true_cov <- tensr::start_ident(p) true_mean <- array(0, dim = p) cond_params <- anorm_cd(Y = Y, M = true_mean, S = true_cov, saidx = saidx) ## Since data are independent standard normals, conditional mean is 0 and ## conditional covariance matrices are identities. cond_params$Mab cond_params$Sab
Calculate the AIC and BIC for Kronecker structured covariance models, assuming the array normal distribution.
array_bic_aic( sig_squared, p, mode_ident = NULL, mode_diag = NULL, mode_unstructured = NULL )
array_bic_aic( sig_squared, p, mode_ident = NULL, mode_diag = NULL, mode_unstructured = NULL )
sig_squared |
A numeric. The MLE of sigma^2 in the array normal model (the 'variance' form of the total variation parameter). |
p |
A vector of integers. The dimension of the data array (including replication modes). |
mode_ident |
A vector of integers. The modes assumed to have identity covariances. |
mode_diag |
A vector of integers. The modes assumed to have diagional covariances. |
mode_unstructured |
A vector of integers. The modes of assumed to have unstructured covariances. |
The AIC and BIC depend only on the data through the MLE of the total variation parameter. Given this, the dimension of the array, and a specification of which modes are the identity and which are unstructured, this function will calculate the AIC and BIC.
AIC
A numeric. The AIC of the model.
BIC
A numeric. The BIC of the model.
David Gerard.
holq
for obtaining sig_squared
.
# Generate random array data with first mode having unstructured covariance # second having diagonal covariance structure and third mode having identity # covariance structure. set.seed(857) p <- c(4, 4, 4) Z <- array(stats::rnorm(prod(p)), dim = p) Y <- atrans(Z, list(tensr:::rwish(diag(p[1])), diag(1:p[2]), diag(p[3]))) # Use holq() to fit various models. false_fit1 <- holq(Y, mode_rep = 1:3) ## identity for all modes false_fit2 <- holq(Y, mode_rep = 2:3) ## unstructured first mode true_fit <- holq(Y, mode_rep = 3, mode_diag = 2) ## correct model # Get AIC and BIC values. false_aic1 <- array_bic_aic(false_fit1$sig ^ 2, p, mode_ident = 1:length(p)) false_aic2 <- array_bic_aic(false_fit2$sig ^ 2, p, mode_ident = 2:length(p), mode_unstructured = 1) true_aic <- array_bic_aic(true_fit$sig ^ 2, p, mode_ident = 2:length(p), mode_diag = 1) # Plot the results. plot(c(false_aic1$AIC, false_aic2$AIC, true_aic$AIC), type = "l", xaxt = "n", xlab = "Model", ylab = "AIC", main = "AIC") axis(side = 1, at = 1:3, labels = c("Wrong Model 1", "Wrong Model 2", "Right Model")) plot(c(false_aic1$BIC, false_aic2$BIC, true_aic$BIC), type = "l", xaxt = "n", xlab = "Model", ylab = "BIC", main = "BIC") axis(side = 1, at = 1:3, labels = c("Wrong Model 1", "Wrong Model 2", "Right Model"))
# Generate random array data with first mode having unstructured covariance # second having diagonal covariance structure and third mode having identity # covariance structure. set.seed(857) p <- c(4, 4, 4) Z <- array(stats::rnorm(prod(p)), dim = p) Y <- atrans(Z, list(tensr:::rwish(diag(p[1])), diag(1:p[2]), diag(p[3]))) # Use holq() to fit various models. false_fit1 <- holq(Y, mode_rep = 1:3) ## identity for all modes false_fit2 <- holq(Y, mode_rep = 2:3) ## unstructured first mode true_fit <- holq(Y, mode_rep = 3, mode_diag = 2) ## correct model # Get AIC and BIC values. false_aic1 <- array_bic_aic(false_fit1$sig ^ 2, p, mode_ident = 1:length(p)) false_aic2 <- array_bic_aic(false_fit2$sig ^ 2, p, mode_ident = 2:length(p), mode_unstructured = 1) true_aic <- array_bic_aic(true_fit$sig ^ 2, p, mode_ident = 2:length(p), mode_diag = 1) # Plot the results. plot(c(false_aic1$AIC, false_aic2$AIC, true_aic$AIC), type = "l", xaxt = "n", xlab = "Model", ylab = "AIC", main = "AIC") axis(side = 1, at = 1:3, labels = c("Wrong Model 1", "Wrong Model 2", "Right Model")) plot(c(false_aic1$BIC, false_aic2$BIC, true_aic$BIC), type = "l", xaxt = "n", xlab = "Model", ylab = "BIC", main = "BIC") axis(side = 1, at = 1:3, labels = c("Wrong Model 1", "Wrong Model 2", "Right Model"))
Generates indices corresponding to subarrays.
arrIndices(saidx)
arrIndices(saidx)
saidx |
either a vector of the dimensions of a potential array, or a list of the indices in the subarray. |
This function generates a matrix corresponding to all combinations of a list of indices, to be used in subsetting arrays.
Peter Hoff.
# all indices of an array arrIndices(c(4, 3, 2)) # indices of a subarray arrIndices(list(c(1, 3), c(4, 5), c(2, 3, 6)))
# all indices of an array arrIndices(c(4, 3, 2)) # indices of a subarray arrIndices(list(c(1, 3), c(4, 5), c(2, 3, 6)))
Performs the Tucker product between an array and a list of matrices.
atrans(A, B)
atrans(A, B)
A |
An array of dimension |
B |
A list of matrices of length |
The Tucker product between a list of matrices B
and an array A
is formally equivalent to performing the -mode product between
A
and each list element in B
. For example, if the dimension of
A
is three, then atrans(A,B) =
amprod(amprod(amprod(A,B[[1]],1),B[[2]],2),B[[3]],3)
. The ordering of this
-mode product does not matter. See
Kolda and Bader
(2009) for details.
Peter Hoff.
Kolda, T. G., & Bader, B. W. (2009). Tensor decompositions and applications. SIAM review, 51(3), 455-500.
amprod
for multiplying one matrix along one mode of an
array.
A <- array(1:8, dim = c(2,2,2)) B <- list() B[[1]] <-matrix(1:4, nrow = 2) B[[2]] <- matrix(1:6, nrow = 3) B[[3]] <- matrix(1:2, nrow = 1) atrans(A,B)
A <- array(1:8, dim = c(2,2,2)) B <- list() B[[1]] <-matrix(1:4, nrow = 2) B[[2]] <- matrix(1:6, nrow = 3) B[[3]] <- matrix(1:2, nrow = 1) atrans(A,B)
Given an array X
and a vector of integers m
,
collapse_mode
returns an array of lower order where the
first mode indexes the modes indicated in m
.
collapse_mode(X, m)
collapse_mode(X, m)
X |
An array whose modes we are collapsing. |
m |
A vector of integers giving the modes to collapse. |
Transforms an array into another array where the provided modes are
collapsed into one mode. The indexing along this new mode is in
lexicographical order of the indices of the collapsed modes. The
collapsed mode is the first mode unless length(m) == 1
, then
collapse_mode
simply returns X
.
If is of order
and
length(m) = q
,
then returns an array of order
, where
the modes indicated in
m
are combined to be the first
mode in .
David Gerard.
X <- array(rep(c(1, 2), 8), dim = c(2, 2, 2, 2)) X #mode 1 is now mode 2, modes 2, 3, and 4 are combined to be mode 1. collapse_mode(X, c(2, 3, 4)) collapse_mode(X, c(2, 4)) ## another example. collapse_mode(X, 4) #returns X
X <- array(rep(c(1, 2), 8), dim = c(2, 2, 2, 2)) X #mode 1 is now mode 2, modes 2, 3, and 4 are combined to be mode 1. collapse_mode(X, c(2, 3, 4)) collapse_mode(X, c(2, 4)) ## another example. collapse_mode(X, 4) #returns X
equi_mcmc
to component covariance matrices.This takes the output from equi_mcmc
, which are the inverses of the
lower-triangular Cholesky square roots of the component covariance matrices,
and returns the component covariance matrices. These are the more useful
posterior draws to use in actual data analysis.
convert_cov(equi_mcmc_obj)
convert_cov(equi_mcmc_obj)
equi_mcmc_obj |
The output from |
The output from equi_mcmc
is the inverse of the lower-triangular
Cholesky square root of each component covariance matrix. This output is
convenient for calculating the Bayes rule under multiway-Stein's loss (see
get_equi_bayes
). Call one of these outputs from
equi_mcmc
. Then this function calculates
, which are the posterior draws of the component covariance
matrices. These component covariance matrices are constrained to have
determinant one, hence there is a total variation parameter
.
cov_post
A list containing the posterior draws of each
component covariance matrix.
sig2_post
A vector containing the posterior draws of the total
variation parameter.
David Gerard.
Gerard, D., & Hoff, P. (2015). Equivariant minimax dominators of the MLE in the array normal model. Journal of Multivariate Analysis, 137, 32-49. https://doi.org/10.1016/j.jmva.2015.01.020 http://arxiv.org/pdf/1408.0424.pdf
#Generate data whose true covariance is just the identity. p <- c(4,4,4) X <- array(stats::rnorm(prod(p)),dim = p) #Then run the Gibbs sampler. mcmc_out <- equi_mcmc(X) cov_out <- convert_cov(mcmc_out) # Some trace plots. plot(cov_out[[2]], type = 'l', xlab = 'Iteration', ylab = expression(sigma ^ 2), main = 'Trace Plot') abline(h = 1, col = 2, lty = 2) legend('topleft', 'True Value', col = 2, lty = 2, bty = 'n') k <- sample(1:length(p), size = 1) i <- sample(1:p[k], size = 1) j <- sample(1:p[k], size = 1) plot(cov_out[[1]][[k]][i, j, ], type = 'l', xlab = 'Iteration', main = 'Trace Plot', ylab = substitute(Sigma[k][group('[', list(i, j), ']')], list(k = k, i = i, j = j))) abline(h = 1 * (i == j), lty = 2, col = 2) legend('topleft', 'True Value', col = 2, lty = 2, bty = 'n')
#Generate data whose true covariance is just the identity. p <- c(4,4,4) X <- array(stats::rnorm(prod(p)),dim = p) #Then run the Gibbs sampler. mcmc_out <- equi_mcmc(X) cov_out <- convert_cov(mcmc_out) # Some trace plots. plot(cov_out[[2]], type = 'l', xlab = 'Iteration', ylab = expression(sigma ^ 2), main = 'Trace Plot') abline(h = 1, col = 2, lty = 2) legend('topleft', 'True Value', col = 2, lty = 2, bty = 'n') k <- sample(1:length(p), size = 1) i <- sample(1:p[k], size = 1) j <- sample(1:p[k], size = 1) plot(cov_out[[1]][[k]][i, j, ], type = 'l', xlab = 'Iteration', main = 'Trace Plot', ylab = substitute(Sigma[k][group('[', list(i, j), ']')], list(k = k, i = i, j = j))) abline(h = 1 * (i == j), lty = 2, col = 2) legend('topleft', 'True Value', col = 2, lty = 2, bty = 'n')
Rotates an array into two parts, one of which has mean zero.
demean_tensor(X, mode_reps)
demean_tensor(X, mode_reps)
X |
An array, one of whose modes is assumed to be samples from the array normal model. |
mode_reps |
The mode(s) that contain(s) the samples, or repetitions, from the array normal model. |
If one mode contains samples (or repetitions), then this function will rotate the array into two parts, a mean part and a covariance part. The 'covariance part' has mean zero and the rest of the methods in this package apply. The 'mean part' is simply the sample mean. If the data are array normal, then the 'covariance part' will also be array normal with the exact same covariance structure as the original tensor, except that there are one fewer samples.
Y
An array that has the same dimensions as X
except that the mode mode_reps
has dimension one
smaller. This array is mean 0 array normal with the same
covariance structure as X
.
X_bar
The sample mean of X
. Under the array normal
model, X
and Y
are statistically independent.
David Gerard.
Gerard, D., & Hoff, P. (2015). Equivariant minimax dominators of the MLE in the array normal model. Journal of Multivariate Analysis, 137, 32-49. https://doi.org/10.1016/j.jmva.2015.01.020 http://arxiv.org/pdf/1408.0424.pdf
equi_mcmc
obtains posterior draws that are useful in optimal
equivariant estimation under the array normal model.
equi_mcmc( X, itermax = 1000, start_identity = FALSE, print_iter = FALSE, mode_rep = NULL )
equi_mcmc( X, itermax = 1000, start_identity = FALSE, print_iter = FALSE, mode_rep = NULL )
X |
A tensor. |
itermax |
The number of iterations in the Gibb's sampler. |
start_identity |
Should we start the component covariance matrices at the identity (TRUE) or the sample covariance matrices (FALSE)? |
print_iter |
Should we print the iteration number at each iteration? |
mode_rep |
The mode that contains samples. I.e., the mode whose component covariance matrix is the identity. If NULL then no modes are assumed to have identity covariance. |
equi_mcmc
obtains posterior samples of the component
covariance matrices from the array normal model. This is with
respect to using the right Haar measure over a product group of
lower triangular matrices as the prior.
This returns only the upper triangular Cholesky square root of the inverses of the component covariance matrices. Equivalently, these are the inverses of the lower triangular Cholesky square roots of the component covariance matrices. This is because sampling the inverse is faster computationally and the Bayes rules (based on multiway Stein's loss) only depend on the inverse.
Phi_inv
List of posterior draws of the inverse of
the cholesky square roots of each component covariance
matrix. Phi_inv[[i]][,,j]
provides the th sample
of the
th component.
sigma
Vector of posterior samples of the overall scale
paramater.
David Gerard.
Gerard, D., & Hoff, P. (2015). Equivariant minimax dominators of the MLE in the array normal model. Journal of Multivariate Analysis, 137, 32-49. https://doi.org/10.1016/j.jmva.2015.01.020 http://arxiv.org/pdf/1408.0424.pdf
sample_right_wishart
and
sample_sig
for the Gibbs
updates. convert_cov
and
get_equi_bayes
for getting posterior summaries
based on the output of
equi_mcmc
. multiway_takemura
for an
improvement on this procedure.
#Generate data whose true covariance is just the identity. p <- c(2,2,2) X <- array(stats::rnorm(prod(p)),dim = p) #Then run the Gibbs sampler. mcmc_out <- equi_mcmc(X) plot(mcmc_out$sigma, type = 'l', lwd = 2, ylab = expression(sigma), xlab = 'Iteration', main = 'Trace Plot') abline(h = 1,col = 2,lty = 2)
#Generate data whose true covariance is just the identity. p <- c(2,2,2) X <- array(stats::rnorm(prod(p)),dim = p) #Then run the Gibbs sampler. mcmc_out <- equi_mcmc(X) plot(mcmc_out$sigma, type = 'l', lwd = 2, ylab = expression(sigma), xlab = 'Iteration', main = 'Trace Plot') abline(h = 1,col = 2,lty = 2)
Calculates the Frobenius norm of an array.
fnorm(X)
fnorm(X)
X |
An array, a matrix, or a vector. |
The Frobenius norm of an array is the square root of the sum of its squared elements. This function works for vector and matrix arguments as well.
The square root of the sum of the squared elements of
X
.
David Gerard.
X <- c(1:8) Y <- matrix(1:8, nrow = 2) Z <- array(1:8, dim = c(2, 2, 2)) fnorm(X) fnorm(Y) fnorm(Z)
X <- c(1:8) Y <- matrix(1:8, nrow = 2) Z <- array(1:8, dim = c(2, 2, 2)) fnorm(X) fnorm(Y) fnorm(Z)
Given the output of equi_mcmc
, this function will calculate the Bayes
rule under multiway Stein's loss.
get_equi_bayes(psi_inv, sigma, burnin = NULL)
get_equi_bayes(psi_inv, sigma, burnin = NULL)
psi_inv |
A list of arrays where |
sigma |
A vector of posteior draws of the total variation parameter.
This is just |
burnin |
A numeric between 0 and 1. What proportion of the posterior samples do you want to discard as burnin? The default is 0.25. |
Multiway Stein's loss is a generalization of Stein's loss to more than two
dimensions. The Bayes rule under this loss is simply represented in terms of
the posterior moments of the component precision matrices. These moments can
be approximated by using the output of equi_mcmc
. When using the
invariant prior that is used in equi_mcmc
, the resulting Bayes rule is
the uniformly minimum risk equivariant estimator.
More details on multiway Stein's loss and the Bayes rules under it can be found in Gerard and Hoff (2015).
Sig_hat
A list of the Bayes rules of the component covariance
matrices under multiway Stein's loss.
B
A list of the lower-triangular Cholesky square roots of the Bayes
rules of the component covariance matrices under multiway Stein's loss. We
have that Sig_hat[[i]]
is equal to B[[i]] %*% t(B[[i]])
.
b
A numeric. This is the bayes rule of the total variation
parameter. This is the 'standard deviation' version. That is, the b ^
2
would be used to calculate the overall covariance matrix.
David Gerard.
Gerard, D., & Hoff, P. (2015). Equivariant minimax dominators of the MLE in the array normal model. Journal of Multivariate Analysis, 137, 32-49. https://doi.org/10.1016/j.jmva.2015.01.020 http://arxiv.org/pdf/1408.0424.pdf
#Generate data whose true covariance is just the identity. p <- c(4,4,4) X <- array(stats::rnorm(prod(p)),dim = p) #Then run the Gibbs sampler. mcmc_out <- equi_mcmc(X) bayes_rules <- get_equi_bayes(mcmc_out$Phi_inv, mcmc_out$sigma) bayes_rules$Sig_hat[[1]]
#Generate data whose true covariance is just the identity. p <- c(4,4,4) X <- array(stats::rnorm(prod(p)),dim = p) #Then run the Gibbs sampler. mcmc_out <- equi_mcmc(X) bayes_rules <- get_equi_bayes(mcmc_out$Phi_inv, mcmc_out$sigma) bayes_rules$Sig_hat[[1]]
The ISVD is a generalization of the SVD to tensors. It is derived from the incredible HOLQ.
get_isvd(x_holq)
get_isvd(x_holq)
x_holq |
The output from |
Let sig * atrans(Z, L)
be the HOLQ of X
. Then the ISVD
calculates the SVD of each L[[i]]
, call it U[[i]] %*% D[[i]]
%*% t(W[[i]])
. It then returns l = sig
, U
, D
, and
V = atrans(Z, W)
. These values have the property that X
is
equal to l * atrans(atrans(V, D), U)
, up to numerical precision.
V
is also scaled all-orthonormal.
For more details on the ISVD, see Gerard and Hoff (2016).
l A numeric.
U A list of orthogonal matrices.
D A list of diagonal matrices with positive diagonal entries and unit determinant. The diagonal entries are in descending order.
V A scaled all-orthonormal array.
David Gerard.
Gerard, D., & Hoff, P. (2016). A higher-order LQ decomposition for separable covariance models. Linear Algebra and its Applications, 505, 57-84. https://doi.org/10.1016/j.laa.2016.04.033 http://arxiv.org/pdf/1410.1094v1.pdf
#Generate random data. p <- c(4,4,4) X <- array(stats::rnorm(prod(p)), dim = p) #Calculate HOLQ, then ISVD holq_x <- holq(X) isvd_x <- get_isvd(holq_x) l <- isvd_x$l U <- isvd_x$U D <- isvd_x$D V <- isvd_x$V #Recover X trim(X - l * atrans(atrans(V, D), U)) #V is scaled all-orthonormal trim(mat(V, 1) %*% t(mat(V, 1)), epsilon = 10^-5) trim(mat(V, 2) %*% t(mat(V, 2)), epsilon = 10^-5) trim(mat(V, 3) %*% t(mat(V, 3)), epsilon = 10^-5)
#Generate random data. p <- c(4,4,4) X <- array(stats::rnorm(prod(p)), dim = p) #Calculate HOLQ, then ISVD holq_x <- holq(X) isvd_x <- get_isvd(holq_x) l <- isvd_x$l U <- isvd_x$U D <- isvd_x$D V <- isvd_x$V #Recover X trim(X - l * atrans(atrans(V, D), U)) #V is scaled all-orthonormal trim(mat(V, 1) %*% t(mat(V, 1)), epsilon = 10^-5) trim(mat(V, 2) %*% t(mat(V, 2)), epsilon = 10^-5) trim(mat(V, 3) %*% t(mat(V, 3)), epsilon = 10^-5)
This function calculates a generalization of the LQ decomposition to tensors. This decomposition has a close connection to likelihood inference in Kronecker structured covariande models.
holq( X, tol = 10^-9, itermax = 1000, mode_rep = NULL, mode_diag = NULL, mode_ldu = NULL, print_diff = TRUE, start_vals = "identity", use_sig = TRUE )
holq( X, tol = 10^-9, itermax = 1000, mode_rep = NULL, mode_diag = NULL, mode_ldu = NULL, print_diff = TRUE, start_vals = "identity", use_sig = TRUE )
X |
An array of numerics. |
tol |
A numeric. The maximum difference in frobenius norm between two
successive core arrays before stopping. Or maximum difference of the ratio
of sigs from 1 before stopping (which depends on the value of
|
itermax |
An integer. The maximum number of iterations of the LQ decomposition to do before stopping. |
mode_rep |
A vector of integers. The optional mode(s) to be considered identity matrices. |
mode_diag |
A vector of integers. The optional mode(s) to be considered as independent but heteroscedastic. |
mode_ldu |
A vector of integers. The optional modes(s) to be considered as having unit diagonal. |
print_diff |
A logical. Should the updates be printed to the screen each iteration? |
start_vals |
Determines how to start the optimization algorithm. Either 'identity' (default), or 'residuals', which results in using the cholesky square roots of the sample covariance matrices along each mode scaled to have unit determinant. You can also use your own start values. |
use_sig |
A logical. What stopping criterion should we use? Frobenius
norm of difference of cores (FALSE) or absolute value of difference of
ratio of |
Given an array X
, the default version of this function will calculate
(1) L
a list of lower triangular matricies with positive diagonal
elements and unit determinant, Z
an array of the same dimensions as
X
that has special orthogonal properties, and (3) sig
a numeric
such that X
is the same as sig * atrans(Z,L)
up to numeric
precision.
This output (1) can be considered a generalization of the LQ decomposition to tensors, (2) solves an optimization problem which the matrix LQ decomposition solves, and (3) has a special connection to likelihood inference in the array normal model.
There are options to constrain the matrices in L
to either be
diagonal, lower triangular with unit diagonal, or the identity matrix. Each
of these correspond to submodels in Kronecker structured covariance models.
The core array corresponding to each of these options has different
properities (see Gerard and Hoff
(2016)). These more constrained tensor decompositions are called HOLQ
juniors.
The MLE of the th component covariance matrix under any
elliptically contoured Kronecker structured covariance model is given by
L[[i]] %*% t(L[[i]])
. The MLE for the total variation pamarameter
will be different depending on the distribution of the array, but for the
array normal it is sig ^ 2 / prod(p)
(the "variance" form for the
total variation parameter).
The likelihood ratio test statistic depends only on sig
and can be
implemented in lrt_stat
.
The algorithm used to fit the HOLQ iteratively repeats the LQ decomposition along each mode.
For more details on the incredible HOLQ, see Gerard and Hoff (2016).
Z
The core array with scaled all-orthonormality property.
A
A list of matrices.
sig
A numeric. The total variation parameter. This is the "standard
deviation" form.
David Gerard.
Gerard, D., & Hoff, P. (2016). A higher-order LQ decomposition for separable covariance models. Linear Algebra and its Applications, 505, 57-84. https://doi.org/10.1016/j.laa.2016.04.033 http://arxiv.org/pdf/1410.1094v1.pdf
array_bic_aic
for using the output of holq
to
calculate AIC and BIC,
get_isvd
for using the output of holq
to calculate a
tensor generalization of the singular value decomposition.
lq
for the matrix LQ decomposition.
lrt_stat
for using the output of holq
to calculate
likelihood ratio test statistics.
mle_from_holq
for using the output of holq
to
calculate the maximum likelihood estimates of the component covariance
matrices under the array normal model.
#Genrate random data. p <- c(2, 3, 4) X <- array(stats::rnorm(prod(p)), dim = p) #Calculate HOLQ with unit diagonal on 2nd mode, # and diagonal along 3rd mode. holq_x <- holq(X, mode_ldu = 2, mode_diag = 3) Z <- holq_x$Z A <- holq_x$A sig <- holq_x$sig #Reconstruct X trim(X - sig * atrans(Z, A), 10^-5) #Properties of core #First mode has orthonormal rows. trim(mat(Z, 1) %*% t(mat(Z, 1)), 10^-5) #Second mode has orthogonal rows. trim(mat(Z, 2) %*% t(mat(Z, 2)), 10^-5) #Third mode has unit diagonal (up to scale). diag(mat(Z, 3) %*% t(mat(Z, 3)))
#Genrate random data. p <- c(2, 3, 4) X <- array(stats::rnorm(prod(p)), dim = p) #Calculate HOLQ with unit diagonal on 2nd mode, # and diagonal along 3rd mode. holq_x <- holq(X, mode_ldu = 2, mode_diag = 3) Z <- holq_x$Z A <- holq_x$A sig <- holq_x$sig #Reconstruct X trim(X - sig * atrans(Z, A), 10^-5) #Properties of core #First mode has orthonormal rows. trim(mat(Z, 1) %*% t(mat(Z, 1)), 10^-5) #Second mode has orthogonal rows. trim(mat(Z, 2) %*% t(mat(Z, 2)), 10^-5) #Third mode has unit diagonal (up to scale). diag(mat(Z, 3) %*% t(mat(Z, 3)))
This function will calculate the best rank r
(where r
is a
vector) approximation (in terms of sum of squared differences) to a given
data array.
hooi(X, r, tol = 10^-6, print_fnorm = FALSE, itermax = 500)
hooi(X, r, tol = 10^-6, print_fnorm = FALSE, itermax = 500)
X |
An array of numerics. |
r |
A vector of integers. This is the given low multilinear rank of the approximation. |
tol |
A numeric. Stopping criterion. |
print_fnorm |
Should updates of the optimization procedure be printed? This number should get larger during the optimizaton procedure. |
itermax |
The maximum number of iterations to run the optimization procedure. |
Given an array X
, this code will find a core array G
and a list
of matrices with orthonormal columns U
that minimizes fnorm(X -
atrans(G, U))
. If r
is equal to the dimension of X
, then it
returns the HOSVD (see hosvd
).
For details on the HOOI see Lathauwer et al (2000).
G
An all-orthogonal core array.
U
A vector of matrices with orthonormal columns.
David Gerard.
De Lathauwer, L., De Moor, B., & Vandewalle, J. (2000).
On the best
rank-1 and rank-() approximation of higher-order tensors.
SIAM Journal on Matrix Analysis and Applications, 21(4), 1324-1342.
## Generate random data. p <- c(2, 3, 4) X <- array(stats::rnorm(prod(p)), dim = p) ## Calculate HOOI r <- c(2, 2, 2) hooi_x <- hooi(X, r = r) G <- hooi_x$G U <- hooi_x$U ## Reconstruct the hooi approximation. X_approx <- atrans(G, U) fnorm(X - X_approx)
## Generate random data. p <- c(2, 3, 4) X <- array(stats::rnorm(prod(p)), dim = p) ## Calculate HOOI r <- c(2, 2, 2) hooi_x <- hooi(X, r = r) G <- hooi_x$G U <- hooi_x$U ## Reconstruct the hooi approximation. X_approx <- atrans(G, U) fnorm(X - X_approx)
Calculates the left singular vectors of each matrix unfolding of an array, then calculates the core array. The resulting output is a Tucker decomposition.
hosvd(Y, r = NULL)
hosvd(Y, r = NULL)
Y |
An array of numerics. |
r |
A vector of integers. The rank of the truncated HOSVD. |
If r
is equal to the rank of Y
, then Y
is equal to
atrans(S, U)
, up to numerical accuracy.
More details on the HOSVD can be found in De Lathauwer et. al. (2000).
U
A list of matrices with orthonormal columns. Each matrix
contains the mode-specific singular vectors of its mode.
S
An all-orthogonal array. This is the core array from the HOSVD.
Peter Hoff.
De Lathauwer, L., De Moor, B., & Vandewalle, J. (2000). A multilinear singular value decomposition. SIAM journal on Matrix Analysis and Applications, 21(4), 1253-1278.
#Generate random data. p <- c(2, 3, 4) X <- array(stats::rnorm(prod(p)), dim = p) #Calculate HOSVD. hosvd_x <- hosvd(X) S <- hosvd_x$S U <- hosvd_x$U #Recover X. trim(X - atrans(S, U)) #S is all-orthogonal. trim(mat(S, 1) %*% t(mat(S, 1))) trim(mat(S, 2) %*% t(mat(S, 2))) trim(mat(S, 3) %*% t(mat(S, 3)))
#Generate random data. p <- c(2, 3, 4) X <- array(stats::rnorm(prod(p)), dim = p) #Calculate HOSVD. hosvd_x <- hosvd(X) S <- hosvd_x$S U <- hosvd_x$U #Recover X. trim(X - atrans(S, U)) #S is all-orthogonal. trim(mat(S, 1) %*% t(mat(S, 1))) trim(mat(S, 2) %*% t(mat(S, 2))) trim(mat(S, 3) %*% t(mat(S, 3)))
Mmm, pancakes.
ihop( X, itermax = 100, tol = 10^-9, print_diff = TRUE, mode_rep = NULL, use_sig = TRUE )
ihop( X, itermax = 100, tol = 10^-9, print_diff = TRUE, mode_rep = NULL, use_sig = TRUE )
X |
An array of numerics. |
itermax |
An integer. The maximum number of iterations to perform during the optimization procedure. |
tol |
A numeric. The algorithm will stop when the Frobenius norm of the
difference of core arrays between subsequent iterations is below |
print_diff |
A logical. Should we print the updates of the algorithm? |
mode_rep |
A vector. Which component matrices should be set to be the identity? |
use_sig |
A logical. See |
This function will calculate the higher-order polar decomposition, a generalization of the polar decomposition to tensors. It generalizes a minimization formulation of the polar decomposition.
Given an array X
, ihop
will output L
a list of lower
triangular matrices with positive diagonal elements and unit Frobenius norm,
R
a core array with certain orthogonality properties, and sig
a
total variation parameter. We have that X
is equal to sig *
atrans(R, L)
up to numerical precision.
t(solve(L[[i]])) %*% mat(R, i)
will have orthonormal rows for all
i
.
For more details on the IHOP, see Gerard and Hoff (2016).
R
A core array which, in combination with L
, has
certain orthogonality properties.
L
A list of lower triangular matrices with unit Frobenius norm.
sig
A numeric.
David Gerard.
Gerard, D., & Hoff, P. (2016). A higher-order LQ decomposition for separable covariance models. Linear Algebra and its Applications, 505, 57-84. https://doi.org/10.1016/j.laa.2016.04.033 http://arxiv.org/pdf/1410.1094v1.pdf
#Generate random data. p <- c(2, 3, 4) X <- array(stats::rnorm(prod(p)), dim = p) #Calculate IHOP. ihop_x <- ihop(X) R <- ihop_x$R L <- ihop_x$L sig <- ihop_x$sig #Reconstruct X trim(X - sig * atrans(R, L)) #Orthogonality properties ortho_1 <- t(solve(L[[1]])) %*% mat(R, 1) trim(ortho_1 %*% t(ortho_1)) ortho_2 <- t(solve(L[[2]])) %*% mat(R, 2) trim(ortho_2 %*% t(ortho_2)) ortho_3 <- t(solve(L[[3]])) %*% mat(R, 3) trim(ortho_3 %*% t(ortho_3))
#Generate random data. p <- c(2, 3, 4) X <- array(stats::rnorm(prod(p)), dim = p) #Calculate IHOP. ihop_x <- ihop(X) R <- ihop_x$R L <- ihop_x$L sig <- ihop_x$sig #Reconstruct X trim(X - sig * atrans(R, L)) #Orthogonality properties ortho_1 <- t(solve(L[[1]])) %*% mat(R, 1) trim(ortho_1 %*% t(ortho_1)) ortho_2 <- t(solve(L[[2]])) %*% mat(R, 2) trim(ortho_2 %*% t(ortho_2)) ortho_3 <- t(solve(L[[3]])) %*% mat(R, 3) trim(ortho_3 %*% t(ortho_3))
This function provides a Monte Carlo approximation to Kendall's tau measure of association.
kendalltau(x, y, nmc = 1e+05)
kendalltau(x, y, nmc = 1e+05)
x |
a vector. |
y |
a vector. |
nmc |
an integer number of Monte Carlo simulations. |
A Monte Carlo approximation to Kendall's tau measure of association.
Peter Hoff.
mu <- rexp(30) tensr:::kendalltau(rpois(30, mu), rpois(30, mu))
mu <- rexp(30) tensr:::kendalltau(rpois(30, mu), rpois(30, mu))
Construct the communtation matrix.
Kom(m, n)
Kom(m, n)
m |
a natural number. |
n |
another natural number. |
This function constructs the commutation matrix K
, which maps
c(A)
to c(t(A))
for an by
matrix
A
.
K
The m * n
by m * n
commutation
matrix.
Peter Hoff.
Magnus, J. R., & Neudecker, H. (1979). The commutation matrix: some properties and applications. The Annals of Statistics, 381-394.
Tracy, D. S., & Dwyer, P. S. (1969). Multivariate maxima and minima with matrix derivatives. Journal of the American Statistical Association, 64(328), 1576-1594.
m <- 5 ; n <- 4 A <- matrix(stats::rnorm(m * n), m, n) Kom(5, 4) %*% c(A) - c(t(A))
m <- 5 ; n <- 4 A <- matrix(stats::rnorm(m * n), m, n) Kom(5, 4) %*% c(A) - c(t(A))
ldan
calculates the log-likelihood of the array normal
model, minus a constant.
ldan(E, Sig)
ldan(E, Sig)
E |
An array. This is the data. |
Sig |
A list of symmetric positive definite matrices. These are the component covariance matrices. |
David Gerard.
Given two lists of matrices with conformable dimensions,
listprod
returns a list whose elements are the matrix
products of the elements of these two lists.
listprod(A, B)
listprod(A, B)
A |
A list of matrices. |
B |
A second list of matrices. |
A list C
such that C[[i]] = A[[i]] %*% B[[i]]
.
David Gerard.
Computes the LQ decomposition of a matrix.
lq(X)
lq(X)
X |
A |
If is an
by
matrix with
, then
lq
computes the LQ decomposition of . That is,
where
is
by
with orthonormal columns
and
is
by
lower triangular with positive
diaognal entries.
L
An by
lower triangular matrix with
positive diagonal entries.
Q
An by
matrix with orthonormal columns.
The returned values satisfy X = L %*% t(Q)
, up to
numerical precision.
David Gerard.
qr2
for the related QR decomposition.
X <- matrix(stats::rnorm(12), nrow = 3) lq_X <- lq(X) L <- lq_X$L Q <- lq_X$Q L Q trim(t(Q) %*% Q) trim(X - L%*%t(Q))
X <- matrix(stats::rnorm(12), nrow = 3) lq_X <- lq(X) L <- lq_X$L Q <- lq_X$Q L Q trim(t(Q) %*% Q) trim(X - L%*%t(Q))
When testing for the covariance structure of modes, this function may be used to draw a sample from the null distribution of the likelihood ratio test stistics, whose distribution doesn't depend on any unknown parameters under the null.
lrt_null_dist_dim_same( p, null_ident = NULL, alt_ident = NULL, null_diag = NULL, alt_diag = NULL, reference_dist = "normal", t_df = NULL, itermax = 100, holq_itermax = 100, holq_tol = 10^-9 )
lrt_null_dist_dim_same( p, null_ident = NULL, alt_ident = NULL, null_diag = NULL, alt_diag = NULL, reference_dist = "normal", t_df = NULL, itermax = 100, holq_itermax = 100, holq_tol = 10^-9 )
p |
A vector of integers. The dimensions of the array. |
null_ident |
A vector of integers. The modes that under the null have identity covariance. |
alt_ident |
A vector of integers. The modes that under the alternative have the identity covariance. |
null_diag |
A vector of integers. The modes that under the null have diagonal covariance. |
alt_diag |
A vector of integers. The modes that under the alternative have diagonal covariance. |
reference_dist |
Two options are supported, 'normal' and 't'. If 't' is
specified, you have to specify |
t_df |
A numeric. If |
itermax |
An integer. The number of draws from the null distribution of the likelihood ratio test statistic that is to be performed. |
holq_itermax |
An integer. The maximum number of block coordinate ascent iterations to perform when calculating the MLE at each step. |
holq_tol |
A numeric. The stopping criterion when calculating the MLE. |
Let be
. Given two nested hypotheses,
versus
this function will draw from the null
distribution of the likelihood ratio test statistic. The possible options are
that or
are the identity matrix, a diagonal
matrix, or any positive definite matrix. By default, it's assumed that these
matrices are any positive definite matrix.
Unfortunately, this fuction does not support testing for the hypothesis of modeling the covariance between two modes with a single covariance matrix. I might code this up in later versions.
A vector of draws from the null distribution of the likelihood ratio test statistic.
David Gerard.
Gerard, D., & Hoff, P. (2016). A higher-order LQ decomposition for separable covariance models. Linear Algebra and its Applications, 505, 57-84. https://doi.org/10.1016/j.laa.2016.04.033 http://arxiv.org/pdf/1410.1094v1.pdf
lrt_stat
for calculating the likelihood ratio test
statistic.
#Test for all identity versus all unconstrained. p = c(4,4,4) null1 <- lrt_null_dist_dim_same(p,null_ident = 1:3) #Generate Null Data X <- array(stats::rnorm(prod(p)), dim = p) sig_null <- holq(X, mode_rep = 1:3)$sig sig_alt <- holq(X)$sig lrt_x <- lrt_stat(sig_null, sig_alt, p = p) p_value <- mean(null1 > lrt_x) hist(null1,main = 'Null Distribution of LRT', xlab = 'LRT Statistic') abline(v = lrt_x, lty = 2, col = 2, lwd = 2) legend('topleft', 'Observed LRT Statistic', lty = 2, col = 2, lwd = 2) mtext(side = 1, paste('P-value = ', round(p_value, digits = 2), sep = ''), line = 2) #------------------------------------------------------------------------- #Test for all identity versus all mode 1 identity, # mode 2 diagonal, mode 3 unconstrained. p = c(4,4,4) null2 <- lrt_null_dist_dim_same(p,null_ident = 1:3, alt_ident = 1, alt_diag = 2) #Generate Null Data X <- array(stats::rnorm(prod(p)), dim = p) sig_null <- holq(X, mode_rep = 1:3)$sig sig_alt <- holq(X, mode_rep = 1, mode_diag = 2)$sig lrt_x <- lrt_stat(sig_null, sig_alt, p = p) p_value <- mean(null2 > lrt_x) hist(null2,main = 'Null Distribution of LRT', xlab = 'LRT Statistic') abline(v = lrt_x, lty = 2, col = 2, lwd = 2) legend('topleft', 'Observed LRT Statistic', lty = 2, col = 2, lwd = 2) mtext(side = 1, paste('P-value = ', round(p_value, digits = 2), sep = ''), line = 2)
#Test for all identity versus all unconstrained. p = c(4,4,4) null1 <- lrt_null_dist_dim_same(p,null_ident = 1:3) #Generate Null Data X <- array(stats::rnorm(prod(p)), dim = p) sig_null <- holq(X, mode_rep = 1:3)$sig sig_alt <- holq(X)$sig lrt_x <- lrt_stat(sig_null, sig_alt, p = p) p_value <- mean(null1 > lrt_x) hist(null1,main = 'Null Distribution of LRT', xlab = 'LRT Statistic') abline(v = lrt_x, lty = 2, col = 2, lwd = 2) legend('topleft', 'Observed LRT Statistic', lty = 2, col = 2, lwd = 2) mtext(side = 1, paste('P-value = ', round(p_value, digits = 2), sep = ''), line = 2) #------------------------------------------------------------------------- #Test for all identity versus all mode 1 identity, # mode 2 diagonal, mode 3 unconstrained. p = c(4,4,4) null2 <- lrt_null_dist_dim_same(p,null_ident = 1:3, alt_ident = 1, alt_diag = 2) #Generate Null Data X <- array(stats::rnorm(prod(p)), dim = p) sig_null <- holq(X, mode_rep = 1:3)$sig sig_alt <- holq(X, mode_rep = 1, mode_diag = 2)$sig lrt_x <- lrt_stat(sig_null, sig_alt, p = p) p_value <- mean(null2 > lrt_x) hist(null2,main = 'Null Distribution of LRT', xlab = 'LRT Statistic') abline(v = lrt_x, lty = 2, col = 2, lwd = 2) legend('topleft', 'Observed LRT Statistic', lty = 2, col = 2, lwd = 2) mtext(side = 1, paste('P-value = ', round(p_value, digits = 2), sep = ''), line = 2)
Calulate the likelihood ratio test statistic for Kronecker structured covariance models.
lrt_stat(sig_null, sig_alt, p)
lrt_stat(sig_null, sig_alt, p)
sig_null |
A numeric. The MLE of the total variation parameter under the null (the standard deviation version). |
sig_alt |
A numeric. The MLE of the total variation parameter under the alternative (the standard deviation version). |
p |
A vector of integers. The dimension of the array. |
The LRT statistic is the exact same for all elliptically distributed Kronecker structured covariance models (not just the normal). The distribution of the likelihood ratio test statistic does change.
A numeric. The likelihood ratio test statistic.
David Gerard.
Gerard, D., & Hoff, P. (2016). A higher-order LQ decomposition for separable covariance models. Linear Algebra and its Applications, 505, 57-84. https://doi.org/10.1016/j.laa.2016.04.033 http://arxiv.org/pdf/1410.1094v1.pdf
holq
for obtaining the MLE of the total variation
parameter.
lrt_null_dist_dim_same
for getting the null distribution of
the likelihood ratio test statistic.
mat
returns a matrix version of a provided tensor.
mat(A, k)
mat(A, k)
A |
An array to be unfolded. |
k |
The mode, or dimension, along which the unfolding is to be applied. |
Applies the matrix unfolding operator (also called 'matricization' or 'matrix flattening' operator) on a provided tensor. There are multiple ways one could do this. This function performs the matrix unfolding described in Kolda and Bader (2009).
A matrix whose rows index the th mode and whose columns
index every other mode. The ordering of the columns is in lexicographical
order of the indices of the array
.
Peter Hoff.
Kolda, T. G., & Bader, B. W. (2009). Tensor decompositions and applications. SIAM review, 51(3), 455-500.
A <- array(1:8, dim = c(2,2,2)) mat(A, 1) mat(A, 2) mat(A, 3)
A <- array(1:8, dim = c(2,2,2)) mat(A, 1) mat(A, 2) mat(A, 3)
Returns the unique symmetric positive definite square root matrix of a provided symmetric positive definite matrix.
mhalf(M)
mhalf(M)
M |
A symmetric positive definite matrix. |
The unique symmetric positive definite matrix such
that
.
Peter Hoff.
Y <- matrix(stats::rnorm(4), nrow = 2) M <- Y %*% t(Y) X <- mhalf(M) X identical(M, X %*% X)
Y <- matrix(stats::rnorm(4), nrow = 2) M <- Y %*% t(Y) X <- mhalf(M) X identical(M, X %*% X)
holq
.From the output of holq
, this function will calculate the
MLEs for the component covariance matrices and for the total
variation parameter.
mle_from_holq(holq_obj)
mle_from_holq(holq_obj)
holq_obj |
The output returned from |
The function simply takes the A[[i]]
output of holq
and returs A[[i]] %*% t(A[[i]])
. The estimate of the total
variation parameter is sqrt(sig ^ 2 / prod{p})
, whre p
is the
vector of dimensions of the data array and sig
is the output
from holq
.
cov_mle
A list of positive definite matrices. These
are the MLEs for the component covariance matrices.
sig_mle
A numeric. This is an estimate of the "standard
deviation" form of the total variation parameter.
David Gerard.
Gerard, D., & Hoff, P. (2016). A higher-order LQ decomposition for separable covariance models. Linear Algebra and its Applications, 505, 57-84. https://doi.org/10.1016/j.laa.2016.04.033 http://arxiv.org/pdf/1410.1094v1.pdf
holq
.
Given a list of estimates of the lower-triangular Cholesky square roots of
component covariance matrices, a list of true lower-triangular Cholesky
square roots of component covariance matrices, an estimate of the total
variation, and the true total variation, multi_stein_loss
will
calculate multiway Stein's loss between the estimates and the truth.
multi_stein_loss(B, Psi, b, psi)
multi_stein_loss(B, Psi, b, psi)
B |
A list of lower triangular matrices. These are the 'estimates' of the lower-triangular Cholesky square roots of the component covariance matrices. |
Psi |
A list of lower triangular matrices. These are the 'true' lower-triangular Cholesky square roots of the component covariance matrices. |
b |
A numeric. This is an 'estimate' of the total variation parameter, the 'standard devation' version of it. |
psi |
A numeric. This is the 'true' total variation parameter, the 'standard devation' version of it. |
Multiway Stein's loss is a generalization of Stein's loss. More details on multiway Stein's loss and the Bayes rules under it can be found in Gerard and Hoff (2015).
The function multi_stien_loss_cov
also calculates multiway Stein's
loss, but uses the component covariance matrices (not the Cholesky roots) as
input.
A numeric, the multiway Stein's loss between the 'truth' and the 'estimates'.
David Gerard.
Gerard, D., & Hoff, P. (2015). Equivariant minimax dominators of the MLE in the array normal model. Journal of Multivariate Analysis, 137, 32-49. https://doi.org/10.1016/j.jmva.2015.01.020 http://arxiv.org/pdf/1408.0424.pdf
multi_stein_loss_cov
, get_equi_bayes
.
Given a list of estimated component covariance matrices, a list of true
component covariance matrices, an estimate of the total variation, and the
true total variation, multi_stein_loss_cov
will calculate multiway
Stein's loss between the estimates and the truth.
multi_stein_loss_cov(B, Sigma, b, sigma)
multi_stein_loss_cov(B, Sigma, b, sigma)
B |
A list of positive definite matrices. These are the 'estimates' of the component covariance matrices. |
Sigma |
A list of positive definite matrices. These are the 'true' component covariance matrices. |
b |
A numeric. This is an 'estimate' of the total variation parameter, the 'standard devation' version of it. |
sigma |
A numeric. This is the 'true' total variation parameter, the 'standard devation' version of it. |
Multiway Stein's loss is a generalization of Stein's loss. More details on multiway Stein's loss and the Bayes rules under it can be found in Gerard and Hoff (2015).
The function multi_stien_loss
also calculates multiway Stein's loss,
but uses the lower-triangular Cholesky square roots of the component
covariance matrices as input.
A numeric, the multiway Stein's loss between the 'truth' and the 'estimates'.
David Gerard.
Gerard, D., & Hoff, P. (2015). Equivariant minimax dominators of the MLE in the array normal model. Journal of Multivariate Analysis, 137, 32-49. https://doi.org/10.1016/j.jmva.2015.01.020 http://arxiv.org/pdf/1408.0424.pdf
multi_stein_loss
, get_equi_bayes
.
This function will 'average' Bayes rules given random rotations of the data array. This 'averaged' estimator has lower risk than the uniformly minimum risk equivariant estimator under a product group of lower triangular matrices. Truncated multiway Takemura's estimator is not equivariant with respect to this product group of lower triangular matrices, but it is an equivariant randomized estimator with respect to a product group of orthogonal matrices.
multiway_takemura( X, ortho_max = 2, mcmc_itermax = 1000, start_identity = FALSE, print_mcmc = FALSE, mode_rep = NULL )
multiway_takemura( X, ortho_max = 2, mcmc_itermax = 1000, start_identity = FALSE, print_mcmc = FALSE, mode_rep = NULL )
X |
An array. This is the data array. |
ortho_max |
An integer. The number of 'averagings' to perform. |
mcmc_itermax |
An integer. The number of iterations each MCMC should
perform using |
start_identity |
Should each MCMC start their covariance matrices at the identity (TRUE) or at the sample covariance matrices (FALSE)? |
print_mcmc |
Should the output of the MCMC be printed to the screen (TRUE) or not (FALSE)? |
mode_rep |
A vector of integers. Which mode(s) are considered iid observations? Default is none. |
This function will (1) randomly rotate X
along every mode, then (2) it
will calculate the uniformly minimum risk equivariant estimator using
equi_mcmc
, then (3) it will 'average' these estimates.
B
A list of the truncated multiway Takemura's estimators for
each component covariance matrix. Not their Cholesky square roots.
b
Truncated multiway Takemura's estimator for the total variation
parameter. The 'variance' form, not the 'standard devation' form.
David Gerard.
Gerard, D., & Hoff, P. (2015). Equivariant minimax dominators of the MLE in the array normal model. Journal of Multivariate Analysis, 137, 32-49. https://doi.org/10.1016/j.jmva.2015.01.020 http://arxiv.org/pdf/1408.0424.pdf
# Simulate data. p <- c(5, 5, 5) X <- array(stats::rnorm(prod(p)), dim = p) multi_out <- multiway_takemura(X, mode_rep = 3) multi_out$b trim(multi_out$B[[1]]) trim(multi_out$B[[2]]) trim(multi_out$B[[3]])
# Simulate data. p <- c(5, 5, 5) X <- array(stats::rnorm(prod(p)), dim = p) multi_out <- multiway_takemura(X, mode_rep = 3) multi_out$b trim(multi_out$B[[1]]) trim(multi_out$B[[2]]) trim(multi_out$B[[3]])
polar
calculates the left polar decomposition of a matrix.
polar(X)
polar(X)
X |
A matrix. |
polar
Takes a matrix , of dimensions
by
, and returns two matrices
and
such that
.
is a symmetric positive definite matrix of
dimension
by
and
is an
by
matrix with orthonormal rows.
P
A by
symmetric positive definite
matrix.
Z
A by
matrix with orthonormal rows.
Note that X == P %*% Z
, up to numerical precision.
David Gerard.
X <- matrix(1:6, nrow = 2) polar_x <- polar(X) P <- polar_x$P Z <- polar_x$Z P Z trim(Z %*% t(Z)) trim(X - P %*% Z)
X <- matrix(1:6, nrow = 2) polar_x <- polar(X) P <- polar_x$P Z <- polar_x$Z P Z trim(Z %*% t(Z)) trim(X - P %*% Z)
QR decomposition, constraining the R matrix to have non-negative diagonal entries.
qr2(X)
qr2(X)
X |
A matrix of dimension |
This function is almost a wrapper for qr()
, qr.R()
, and
qr.Q()
, except it constrains the diagonal elements of R
to be
non-negative. If X
is full rank with fewer columns than rows, then
this is sufficient to gaurantee uniqueness of the QR decomposition
(Proposition 5.2 of
Eaton (1983)).
Q
An by
matrix with orthonormal columns.
R
A by
upper-triangular matrix with non-negative
diagonal elements.
David Gerard.
qr
, qr.Q
, and
qr.R
for the base methods on the obtaining the QR
decomposition. lq
for the related LQ decomposition.
Given a vector p
, random_ortho
will generate a list
ortho_list
such that ortho_list[[i]]
is a matrix with row and
column dimensions p[[i]]
and is drawn from the uniform (Haar)
distribution over the space of orthogonal matrices.
random_ortho(p)
random_ortho(p)
p |
A vector of dimensions for the matrices. |
This function is primarily used by multiway_takemura
in its
averaging over uniformly minimum risk equivariant estimators under rotations
of the data array.
ortho_list
A list of orthogonal matrices whose dimensions are
given in p
.
David Gerard.
Given scale matrix Phi
and degrees of freedom nu
,
rmirror_wishart
will sample from the mirror-Wishart distribution.
rmirror_wishart(nu, Phi)
rmirror_wishart(nu, Phi)
nu |
An integer. The degrees of freedom in the mirror-Wishart. |
Phi |
A matrix. The scale matrix of the mirror-Wishart. |
is mirror-Wishart(
) if
where
is the lower triangular Cholesky decomposition of a
Wishart(
)-distributed random matrix and
is the upper
triangular Cholesky decomposition of
. That is,
is lower
triangular and
is upper triangular. For details on its applications,
see
Gerard and Hoff (2015).
A matrix drawn from the mirror-Wishart distribution with nu
degrees of freedom and scale matrix Phi
.
David Gerard.
Gerard, D., & Hoff, P. (2015). Equivariant minimax dominators of the MLE in the array normal model. Journal of Multivariate Analysis, 137, 32-49. https://doi.org/10.1016/j.jmva.2015.01.020 http://arxiv.org/pdf/1408.0424.pdf
Simulate a multivariate normal random matrix.
rmvnorm(n, mu, Sigma, Sigma.chol = chol(Sigma))
rmvnorm(n, mu, Sigma, Sigma.chol = chol(Sigma))
n |
number of mvnormal vectors to simulate. |
mu |
mean vector. |
Sigma |
covariance matrix. |
Sigma.chol |
Cholesky decomposition of |
This function simulates multivariate normal random vectors.
Peter Hoff.
# Simulate several matrices and compute the mean. Y <- tensr:::rmvnorm(100, c(1, 2, 3), matrix(c(3, 0, 1, 0, 1, -1, 1, -1, 2), 3, 3)) colMeans(Y) cov(Y)
# Simulate several matrices and compute the mean. Y <- tensr:::rmvnorm(100, c(1, 2, 3), matrix(c(3, 0, 1, 0, 1, -1, 1, -1, 2), 3, 3)) colMeans(Y) cov(Y)
Generate an array of iid standard normal variables.
rsan(dim)
rsan(dim)
dim |
a vector of positive integers. |
This functions generates an array of dimension dim
filled
with iid standard normal variables.
Peter Hoff.
tensr:::rsan(c(5,4,3))
tensr:::rsan(c(5,4,3))
Simulate a Wishart-distributed random matrix.
rwish(S0, nu = dim(as.matrix(S0))[1] + 1)
rwish(S0, nu = dim(as.matrix(S0))[1] + 1)
S0 |
a positive definite matrix. |
nu |
a positive scalar. |
This function simulates a Wishart random matrix using Bartletts decomposition, as described in Everson and Morris (2000).
Peter Hoff.
# simulate several matrices and compute the mean. SS <- matrix(0, 5, 5) for(s in 1:1000) { SS <- SS + tensr:::rwish(diag(5), 3) } SS / s
# simulate several matrices and compute the mean. SS <- matrix(0, 5, 5) for(s in 1:1000) { SS <- SS + tensr:::rwish(diag(5), 3) } SS / s
Phi_inv
.Samples an upper triangular Cholesky square root of a mirror-Wishart distributed random variable.
sample_right_wishart(nu, V)
sample_right_wishart(nu, V)
nu |
A numeric. The degrees of freedom in the mirror-Wishart. |
V |
A matrix. The inverse of the scale matrix in the mirror-Wishart. |
Let be mirror-Wishart(
,
). Then This code
returns an upper triangular
where
. This
function is used primarily during the Gibbs updates of the inverse
of the lower triangular Cholesky square root of the component
covariance matrices in
equi_mcmc
.
C
An upper triangular matrix such that C %*% t(C)
is
a sample from the mirror-Wishart(nu
, V ^ -1
) distribution.
David Gerard.
Gerard, D., & Hoff, P. (2015). Equivariant minimax dominators of the MLE in the array normal model. Journal of Multivariate Analysis, 137, 32-49. https://doi.org/10.1016/j.jmva.2015.01.020 http://arxiv.org/pdf/1408.0424.pdf
equi_mcmc
.Samples from the square root of an inverse-gamma.
sample_sig(X, phi_inv)
sample_sig(X, phi_inv)
X |
An array. The tensor data. |
phi_inv |
A list of the current values of inverse of the lower-triangular Cholesky square root of the the component covariance matrices. This is equivalent to the transpose of the upper-triangular Cholesky square root of the inverse component covariance matrices.
|
This function provides a Gibbs update for the total variation parameter from
the MCMC implemented in equi_mcmc
. This corresponds to the square root
of an inverse-gamma distributed random variable whose parameters depend on
the data and the component covariance matrices. Roughly, this is the update
for the standard deviation, not the variance.
A numeric. The update for the total variation parameter in the MCMC
implemented in equi_bayes
.
David Gerard.
Gerard, D., & Hoff, P. (2015). Equivariant minimax dominators of the MLE in the array normal model. Journal of Multivariate Analysis, 137, 32-49. https://doi.org/10.1016/j.jmva.2015.01.020 http://arxiv.org/pdf/1408.0424.pdf
equi_mcmc
for a Gibbs sampler where this function is
used.
Will provide a list of identity matrices for the specified modes.
start_ident(p, modes = NULL)
start_ident(p, modes = NULL)
p |
A vector of integers. This is the dimension of the array and the length of the list to be created. |
modes |
A vector of integers. These are the indices in the list to be given an identity matrix. |
Given a vector of dimensions p
and a vector indicating which
modes will get an identity matrix modes
, this function will
return a list start_vals
where start_vals[[i]]
is the
identity matrix of dimensions p[i]
if i
is in
modes
and will be NULL
otherwise.
This is primarily used when getting starting values in equi_mcmc
.
start_vals
A list of identity matrices and NULL values.
David Gerard.
Scaled Cholesky square roots of the sample covariance matrix and its inverse.
start_resids(Y, mode_rep = NULL)
start_resids(Y, mode_rep = NULL)
Y |
An array of numeric data. |
mode_rep |
A vector of integers. The modes specified by
|
This function will take the sample covariance matrix of the
th matricization of an input array
and will return
(1) its lower-triangular Cholesky square root scaled down to have
determinant 1 and (2) the inverse of its lower-triangular Cholesky
square root scaled down to have determinant 1. This function is
primarily used to obtain starting values for the Gibbs sampler
implemented in
equi_mcmc
.
Sig
A list where Sig[[i]]
is the
lower-triangular Cholesky square root of the sample covariance
matrix of the th mode, scaled down to have determinant
1.
Sig_inv
A list where Sig_inv[[i]]
is the inverse of the
lower-triangular Cholesky square root of the sample covariance matrix of
the th mode, scaled down to have determinant 1.
If mode_rep
is not NULL
, then the list elements in Sig
and Sig_inv
specified in mode_rep
will be the identity matrix
instead of sample-based matrices.
David Gerard.
This package provides a collection of functions for likelihood and equivariant inference for covariance matrices under the array normal model. Also included are functions for calculating tensor decompositions that are related to likelihood inference in the array normal model.
Let be a multidimensional array
(also called a tensor) of K dimensions. This package provides a
series of functions to perform statistical inference when
where is assumed to
be Kronecker structured. That is,
is the Kronecker
product of
covariance matrices, each of which has the
interpretation of being the covariance of
along its
th mode, or dimension.
Pay particular attention to the zero mean assumption. That is,
you need to de-mean your data prior to applying these
functions. If you have more than one sample, for
, then you can concatenate these tensors along a
th mode to form a new tensor
and apply the
demean_tensor()
function to Y which will return a tensor
that satisfies the mean-zero assumption.
The details of the methods in this package can be found in Gerard and Hoff (2015) and Gerard and Hoff (2016).
amprod
-mode product.
anorm_cd
Array normal conditional distributions.
array_bic_aic
Calculate the AIC and BIC.
arrIndices
Array indices.
atrans
Tucker product.
collapse_mode
Collapse multiple modes into one mode.
convert_cov
Convert the output from equi_mcmc
to
component covariance matrices.
demean_tensor
Demeans array data.
equi_mcmc
Gibbs sampler using an invariant prior.
fnorm
Frobenius norm of an array.
get_equi_bayes
Get the Bayes rule under multiway Stein's
loss.
get_isvd
Calculate the incredible SVD (ISVD).
holq
Calculate the incredible higher-order LQ decomposition
(HOLQ).
hooi
Calculate the higher-order orthogonal iteration (HOOI).
hosvd
Calculate the (truncated) higher-order SVD (HOSVD).
Kom
Commutation matrix.
ihop
The incredible higher-order polar decomposition (IHOP).
ldan
Log-likelihood of array normal model.
listprod
Element-wise matrix products between two lists.
lq
LQ decomposition.
lrt_null_dist_dim_same
Draw from null distribution of
likelihood ratio test statistic.
lrt_stat
Calculate the likelihood ratio test statistic.
mat
Unfold a matrix.
mhalf
The symmetric square root of a positive definite
matrix.
mle_from_holq
Get MLE from output of holq
.
multi_stein_loss
Calculate multiway Stein's loss from square
root matrices.
multi_stein_loss_cov
Calculate multiway Stein's loss from
component covariance matrices.
multiway_takemura
Calculate a truncated multiway Takemura
estimator.
polar
The left polar decomposition.
qr2
QR Decomposition.
random_ortho
Generate a list of orthogonal matrices drawn
from Haar distribution.
rmirror_wishart
Sample from the mirror-Wishart distribution.
sample_sig
Update for total variation parameter in
equi_mcmc
.
sample_right_wishart
Gibbs update of Phi_inv
.
start_ident
Get list of identity matrices.
start_resids
Sample covariance matrices for each mode.
tsum
Tucker sum.
tr
Trace of a matrix.
trim
Truncates small numbers to 0.
Gerard, D., & Hoff, P. (2016). A higher-order LQ decomposition for separable covariance models. Linear Algebra and its Applications, 505, 57-84. https://doi.org/10.1016/j.laa.2016.04.033 http://arxiv.org/pdf/1410.1094v1.pdf
Gerard, D., & Hoff, P. (2015). Equivariant minimax dominators of the MLE in the array normal model. Journal of Multivariate Analysis, 137, 32-49. https://doi.org/10.1016/j.jmva.2015.01.020 http://arxiv.org/pdf/1408.0424.pdf
Identify top K elements of a vector.
topK(x, K = 1, ignoreties = TRUE)
topK(x, K = 1, ignoreties = TRUE)
x |
The vector. |
K |
The number of indices to return. |
ignoreties |
If |
This function returns the indices corresponding to the top elements of a vector.
Peter Hoff.
x <- c(3, 6, 2, 4, 1) tensr:::topK(x, 3)
x <- c(3, 6, 2, 4, 1) tensr:::topK(x, 3)
Returns the sum of the diagonal elements of a matrix.
tr(X)
tr(X)
X |
A matrix whose diagonal elements will be added together. |
This returns the trace of a matrix, which is just the sum of its diagonal elements.
The sum of the diagonal elements of X.
Peter Hoff.
X <- matrix(1:4, nrow = 2, ncol = 2) X tr(X)
X <- matrix(1:4, nrow = 2, ncol = 2) X tr(X)
Given an array, matrix, or vector, trim
will truncate all
elements smaller than epsilon
(in absolute value) to zero.
trim(X, epsilon = 10^-6)
trim(X, epsilon = 10^-6)
X |
An array, a matrix, or a vector. |
epsilon |
A numeric. |
All elements in X
that are smaller than epsilon
(in
absolute value) will be set to zero then returned.
David Gerard.
X <- c(0, 1, 10^-7, -1, -10^-7) X trim(X)
X <- c(0, 1, 10^-7, -1, -10^-7) X trim(X)
Computes the Tucker sum of an array and a list of matrices.
tsum(X, A)
tsum(X, A)
X |
A real array. |
A |
A list of real matrices. |
This function applies a quantile-quantile transformation to the data, resulting in a distribution that is approximately normal but has the same ranks as the original data.
zscores(y, ties.method = "average")
zscores(y, ties.method = "average")
y |
A vector. |
ties.method |
The option |
A vector of the same length as y
.
Peter Hoff.
y <- rexp(100) z <- tensr:::zscores(y) par(mfrow = c(1, 3)) hist(y) hist(z) plot(y,z)
y <- rexp(100) z <- tensr:::zscores(y) par(mfrow = c(1, 3)) hist(y) hist(z) plot(y,z)