Search results for "Matrix"
showing 10 items of 3205 documents
Tests of multinormality based on location vectors and scatter matrices
2007
Classical univariate measures of asymmetry such as Pearson’s (mean-median)/σ or (mean-mode)/σ often measure the standardized distance between two separate location parameters and have been widely used in assessing univariate normality. Similarly, measures of univariate kurtosis are often just ratios of two scale measures. The classical standardized fourth moment and the ratio of the mean deviation to the standard deviation serve as examples. In this paper we consider tests of multinormality which are based on the Mahalanobis distance between two multivariate location vector estimates or on the (matrix) distance between two scatter matrix estimates, respectively. Asymptotic theory is develop…
Adaptive Metropolis algorithm using variational Bayesian adaptive Kalman filter
2013
Markov chain Monte Carlo (MCMC) methods are powerful computational tools for analysis of complex statistical problems. However, their computational efficiency is highly dependent on the chosen proposal distribution, which is generally difficult to find. One way to solve this problem is to use adaptive MCMC algorithms which automatically tune the statistics of a proposal distribution during the MCMC run. A new adaptive MCMC algorithm, called the variational Bayesian adaptive Metropolis (VBAM) algorithm, is developed. The VBAM algorithm updates the proposal covariance matrix using the variational Bayesian adaptive Kalman filter (VB-AKF). A strong law of large numbers for the VBAM algorithm is…
Linear Recursive Equations, Covariance Selection, and Path Analysis
1980
Abstract By defining a reducible zero pattern and by using the concept of multiplicative models, we relate linear recursive equations that have been introduced by econometrician Herman Wold (1954) and path analysis as it was proposed by geneticist Sewall Wright (1923) to the statistical theory of covariance selection formulated by Arthur Dempster (1972). We show that a reducible zero pattern is the condition under which parameters as well as least squares estimates in recursive equations are one-to-one transformations of parameters and of maximum likelihood estimates, respectively, in a decomposable covariance selection model. As a consequence, (a) we can give a closed-form expression for t…
Robustifying principal component analysis with spatial sign vectors
2012
Abstract In this paper, we apply orthogonally equivariant spatial sign covariance matrices as well as their affine equivariant counterparts in principal component analysis. The influence functions and asymptotic covariance matrices of eigenvectors based on robust covariance estimators are derived in order to compare the robustness and efficiency properties. We show in particular that the estimators that use pairwise differences of the observed data have very good efficiency properties, providing practical robust alternatives to classical sample covariance matrix based methods.
Solving type-2 assembly line balancing problem with fuzzy binary linear programming
2013
This paper deals with the use of fuzzy set theory as a viable alternative method for modelling and solving the stochastic assembly line balancing problem. This paper presents a fuzzy extension of the simple assembly line balancing problem of type 2 SALBP-2 with fuzzy job processing times since uncertainty, variability, and imprecision are often occurred in real-world production systems. The job processing times are formulated by triangular fuzzy membership functions using their statistical distributions. This study proposes to solve a Fuzzy Binary Linear Problem FBLP with fuzzy coefficients in the objective function and in a constraint. Finally, the effect of the unbalancing of a station in…
A more efficient second order blind identification method for separation of uncorrelated stationary time series
2016
The classical second order source separation methods use approximate joint diagonalization of autocovariance matrices with several lags to estimate the unmixing matrix. Based on recent asymptotic results, we propose a novel unmixing matrix estimator which selects the best lag set from a finite set of candidate sets specified by the user. The theory is illustrated by a simulation study.
Central Limit Theorem for Linear Eigenvalue Statistics for a Tensor Product Version of Sample Covariance Matrices
2017
For $$k,m,n\in {\mathbb {N}}$$ , we consider $$n^k\times n^k$$ random matrices of the form $$\begin{aligned} {\mathcal {M}}_{n,m,k}({\mathbf {y}})=\sum _{\alpha =1}^m\tau _\alpha {Y_\alpha }Y_\alpha ^T,\quad {Y}_\alpha ={\mathbf {y}}_\alpha ^{(1)}\otimes \cdots \otimes {\mathbf {y}}_\alpha ^{(k)}, \end{aligned}$$ where $$\tau _{\alpha }$$ , $$\alpha \in [m]$$ , are real numbers and $${\mathbf {y}}_\alpha ^{(j)}$$ , $$\alpha \in [m]$$ , $$j\in [k]$$ , are i.i.d. copies of a normalized isotropic random vector $${\mathbf {y}}\in {\mathbb {R}}^n$$ . For every fixed $$k\ge 1$$ , if the Normalized Counting Measures of $$\{\tau _{\alpha }\}_{\alpha }$$ converge weakly as $$m,n\rightarrow \infty $$…
Algorithm AS 105: Fitting a Covariance Selection Model to a Matrix
1977
Statistical properties of a blind source separation estimator for stationary time series
2012
Abstract In this paper, we assume that the observed p time series are linear combinations of p latent uncorrelated weakly stationary time series. The problem is then, using the observed p -variate time series, to find an estimate for a mixing or unmixing matrix for the combinations. The estimated uncorrelated time series may then have nice interpretations and can be used in a further analysis. The popular AMUSE algorithm finds an estimate of an unmixing matrix using covariances and autocovariances of the observed time series. In this paper, we derive the limiting distribution of the AMUSE estimator under general conditions, and show how the results can be used for the comparison of estimate…
Parallel Construction and Query of Index Data Structures for Pattern Matching on Square Matrices
1999
AbstractWe describe fast parallel algorithms for building index data structures that can be used to gather various statistics on square matrices. The main data structure is the Lsuffix tree, which is a generalization of the classical suffix tree for strings. Given ann×ntext matrixA, we build our data structures inO(logn) time withn2processors on a CRCW PRAM, so that we can quickly processAin parallel as follows: (i) report some statistical information aboutA, e.g., find the largest repeated square submatrices that appear at least twice inAor determine, for each position inA, the smallest submatrix that occurs only there; (ii) given, on-line, anm×mpattern matrixPAT, check whether it occurs i…