Search results for "bayesian"
showing 10 items of 604 documents
Cluster priors in the Bayesian modelling of fMRI data
2001
Statistical Adhockeries Are No Criteria for Legal Decisions—The Case of the Expert Medical Report on the Assessment of Urine Specimens Collected Amon…
2019
On the origin and diversification of Podolian cattle breeds: testing scenarios of European colonization using genome-wide SNP data
2021
AbstractBackgroundDuring the Neolithic expansion, cattle accompanied humans and spread from their domestication centres to colonize the ancient world. In addition, European cattle occasionally intermingled with both indicine cattle and local aurochs resulting in an exclusive pattern of genetic diversity. Among the most ancient European cattle are breeds that belong to the so-called Podolian trunk, the history of which is still not well established. Here, we used genome-wide single nucleotide polymorphism (SNP) data on 806 individuals belonging to 36 breeds to reconstruct the origin and diversification of Podolian cattle and to provide a reliable scenario of the European colonization, throug…
贝叶斯因子及其在JASP中的实现
2018
Statistical inference plays a critical role in modern scientific research, however, the dominant method for statistical inference in science, null hypothesis significance testing (NHST), is often misunderstood and misused, which leads to unreproducible findings. To address this issue, researchers propose to adopt the Bayes factor as an alternative to NHST. The Bayes factor is a principled Bayesian tool for model selection and hypothesis testing, and can be interpreted as the strength for both the null hypothesis H0 and the alternative hypothesis H1 based on the current data. Compared to NHST, the Bayes factor has the following advantages: it quantifies the evidence that the data provide for…
Ensemble feature selection with the simple Bayesian classification
2003
Abstract A popular method for creating an accurate classifier from a set of training data is to build several classifiers, and then to combine their predictions. The ensembles of simple Bayesian classifiers have traditionally not been a focus of research. One way to generate an ensemble of accurate and diverse simple Bayesian classifiers is to use different feature subsets generated with the random subspace method. In this case, the ensemble consists of multiple classifiers constructed by randomly selecting feature subsets, that is, classifiers constructed in randomly chosen subspaces. In this paper, we present an algorithm for building ensembles of simple Bayesian classifiers in random sub…
Use of hierarchical Bayesian framework in MTS studies to model different causes and novel possible forms of acquired MTS
2015
Abstract: An integrative account of MTS could be cast in terms of hierarchical Bayesian inference. It may help to highlight a central role of sensory (tactile) precision could play in MTS. We suggest that anosognosic patients, with anesthetic hemisoma, can also be interpreted as a form of acquired MTS, providing additional data for the model.
A Survey of Bayesian Techniques in Computer Vision
2010
The Bayesian approach to classification is intended to solve questions concerning how to assign a class to an observed pattern using probability estimations. Red, green and blue (RGB) or hue, saturation and lightness (HSL) values of pixels in digital colour images can be considered as feature vectors to be classified, thus leading to Bayesian colour image segmentation. Bayesian classifiers are also used to sort objects but, in this case, reduction of the dimensionality of the feature vector is often required prior to the analysis. This chapter shows some applications of Bayesian learning techniques in computer vision in the agriculture and agri-food sectors. Inspection and classification of…
Text Classification Using Novel “Anti-Bayesian” Techniques
2015
This paper presents a non-traditional “Anti-Bayesian” solution for the traditional Text Classification (TC) problem. Historically, all the recorded TC schemes work using the fundamental paradigm that once the statistical features are inferred from the syntactic/semantic indicators, the classifiers themselves are the well-established statistical ones. In this paper, we shall demonstrate that by virtue of the skewed distributions of the features, one could advantageously work with information latent in certain “non-central” quantiles (i.e., those distant from the mean) of the distributions. We, indeed, demonstrate that such classifiers exist and are attainable, and show that the design and im…
Predictive and Contextual Feature Separation for Bayesian Metanetworks
2007
Bayesian Networks are proven to be a comprehensive model to describe causal relationships among domain attributes with probabilistic measure of conditional dependency. However, depending on a context, many attributes of the model might not be relevant. If a Bayesian Network has been learned across multiple contexts then all uncovered conditional dependencies are averaged over all contexts and cannot guarantee high predictive accuracy when applied to a concrete case. We are considering a context as a set of contextual attributes, which are not directly effect probability distribution of the target attributes, but they effect on "relevance" of the predictive attributes towards target attribut…
An adaptive probabilistic approach to goal-level imitation learning
2010
Imitation learning has been recognized as a promising technique to teach robots advanced skills. It is based on the idea that robots could learn new behaviors by observing and imitating the behaviors of other skilled actors. We propose an adaptive probabilistic graphical model which copes with three core issues of any imitative behavior: observation, representation and reproduction of skills. Our model, Growing Hierarchical Dynamic Bayesian Network (GHDBN), is hierarchical (i.e. able to characterize structured behaviors at different levels of abstraction), and growing (i.e. skills are learned or updated incrementally - and at each level of abstraction - every time a new observation sequence…