Search results for "ESTIMATOR"
showing 10 items of 313 documents
2018
The retrieval of ancient DNA from osteological material provides direct evidence of human genetic diversity in the past. Ancient DNA samples are often used to investigate whether there was population continuity in the settlement history of an area. Methods based on the serial coalescent algorithm have been developed to test whether the population continuity hypothesis can be statistically rejected by analysing DNA samples from the same region but of different ages. Rejection of this hypothesis is indicative of a large genetic shift, possibly due to immigration occurring between two sampling times. However, this approach is only able to reject a model of full continuity model (a total absenc…
L1-Penalized Censored Gaussian Graphical Model
2018
Graphical lasso is one of the most used estimators for inferring genetic networks. Despite its diffusion, there are several fields in applied research where the limits of detection of modern measurement technologies make the use of this estimator theoretically unfounded, even when the assumption of a multivariate Gaussian distribution is satisfied. Typical examples are data generated by polymerase chain reactions and flow cytometer. The combination of censoring and high-dimensionality make inference of the underlying genetic networks from these data very challenging. In this article, we propose an $\ell_1$-penalized Gaussian graphical model for censored data and derive two EM-like algorithm…
Probabilistic cross-validation estimators for Gaussian process regression
2018
Gaussian Processes (GPs) are state-of-the-art tools for regression. Inference of GP hyperparameters is typically done by maximizing the marginal log-likelihood (ML). If the data truly follows the GP model, using the ML approach is optimal and computationally efficient. Unfortunately very often this is not case and suboptimal results are obtained in terms of prediction error. Alternative procedures such as cross-validation (CV) schemes are often employed instead, but they usually incur in high computational costs. We propose a probabilistic version of CV (PCV) based on two different model pieces in order to reduce the dependence on a specific model choice. PCV presents the benefits from both…
ON THE CALCULATION OF THE HEAT CAPACITY IN PATH INTEGRAL MONTE CARLO SIMULATIONS
1992
In Path Integral Monte Carlo simulations the systems partition function is mapped to an equivalent classical one at the expense of a temperature-dependent Hamiltonian with an additional imaginary time dimension. As a consequence the standard relation linking the heat capacity Cv to the energy fluctuations, <E2>−<E>2, which is useful in standard classical problems with temperature-independent Hamiltonian, becomes invalid. Instead, it gets replaced by the general relation [Formula: see text] for the intensive heat capacity estimator; β being the inverse temperature and the subscript P indicates the P-fold discretization in the imaginary time direction. This heatcapacity estimator…
Cancer net survival on registry data: use of the new unbiased Pohar-Perme estimator and magnitude of the bias with the classical methods
2013
Net survival, the survival which might occur if cancer was the only cause of death, is a major epidemiological indicator required for international or temporal comparisons. Recent findings have shown that all classical methods used for routine estimation of net survival from cancer-registry data, sometimes called "relative-survival methods," provide biased estimates. Meanwhile, an unbiased estimator, the Pohar-Perme estimator (PPE), was recently proposed. Using real data, we investigated the magnitude of the errors made by four "relative-survival" methods (Ederer I, Hakulinen, Ederer II and a univariable regression model) vs. PPE as reference and examined the influence of time of follow-up,…
EMG, heart rate, and accelerometer as estimators of energy expenditure in locomotion.
2014
AB Purpose: Precise measures of energy expenditure (EE) during everyday activities are needed. This study assessed the validity of novel shorts measuring EMG and compared this method with HR and accelerometry (ACC) when estimating EE. Methods: Fifty-four volunteers (39.4 +/- 13.9 yr) performed a maximal treadmill test (3-min loads) including walking with different speeds uphill, downhill, and on level ground and one running load. The data were categorized into all, low, and level loads. EE was measured by indirect calorimetry, whereas HR, ACC, and EMG were measured continuously. EMG from quadriceps (Q) and hamstrings (H) was measured using shorts with textile electrodes. Validity of the met…
Testing Frequency-Domain Causality in Multivariate Time Series
2010
We introduce a new hypothesis-testing framework, based on surrogate data generation, to assess in the frequency domain, the concept of causality among multivariate (MV) time series. The approach extends the traditional Fourier transform (FT) method for generating surrogate data in a MV process and adapts it to the specific issue of causality. It generates causal FT (CFT) surrogates with FT modulus taken from the original series, and FT phase taken from a set of series with causal interactions set to zero over the direction of interest and preserved over all other directions. Two different zero-setting procedures, acting on the parameters of a MV autoregressive (MVAR) model fitted on the ori…
Design Issues and Sample Size when Exposure Measurement is Inaccurate
2001
AbstractMeasurement error often leads to biased estimates and incorrect tests in epidemiological studies. These problems can be corrected by design modifications which allow for refined statistical models, or in some situations by adjusted sample sizes to compensate a power reduction. The design options are mainly an additional replication or internal validation study. Sample size calculations for these designs are more complex, since usually there is no unique design solution to obtain a prespecified power. Thus, additionally to a power requirement, an optimal design should also fulfill the criteria of minimizing overall costs. In this review corresponding strategies and formulae are descr…
Ensuring High Performance of Consensus-Based Estimation by Lifetime Maximization in WSNs
2015
The estimation of a parameter corrupted by noise is a common tasks in wireless sensor networks, where the deployed nodes cooperate in order to improve their own inaccurate observations. This cooperation usually involves successive data exchanges and local information updates until a global consensus value is reached. The quality of the final estimator depends on the amount of collected observations, hence the number of active nodes. Moreover, the inherent iterative nature of the consensus process involves a certain energy consumption. Since the devices composing the network are usually battery powered, nodes becoming inactive due to battery depletion emerges as a serious problem. In this wo…
Error estimation and reduction with cross correlations
2010
Besides the well-known effect of autocorrelations in time series of Monte Carlo simulation data resulting from the underlying Markov process, using the same data pool for computing various estimates entails additional cross correlations. This effect, if not properly taken into account, leads to systematically wrong error estimates for combined quantities. Using a straightforward recipe of data analysis employing the jackknife or similar resampling techniques, such problems can be avoided. In addition, a covariance analysis allows for the formulation of optimal estimators with often significantly reduced variance as compared to more conventional averages.