Search results for "Probability"
showing 10 items of 3417 documents
贝叶斯因子及其在JASP中的实现
2018
Statistical inference plays a critical role in modern scientific research, however, the dominant method for statistical inference in science, null hypothesis significance testing (NHST), is often misunderstood and misused, which leads to unreproducible findings. To address this issue, researchers propose to adopt the Bayes factor as an alternative to NHST. The Bayes factor is a principled Bayesian tool for model selection and hypothesis testing, and can be interpreted as the strength for both the null hypothesis H0 and the alternative hypothesis H1 based on the current data. Compared to NHST, the Bayes factor has the following advantages: it quantifies the evidence that the data provide for…
Ensemble feature selection with the simple Bayesian classification
2003
Abstract A popular method for creating an accurate classifier from a set of training data is to build several classifiers, and then to combine their predictions. The ensembles of simple Bayesian classifiers have traditionally not been a focus of research. One way to generate an ensemble of accurate and diverse simple Bayesian classifiers is to use different feature subsets generated with the random subspace method. In this case, the ensemble consists of multiple classifiers constructed by randomly selecting feature subsets, that is, classifiers constructed in randomly chosen subspaces. In this paper, we present an algorithm for building ensembles of simple Bayesian classifiers in random sub…
Fuzzy temporal random sets with an application to cell biology
2007
Total Internal Reflection Fluorescence Microscopy (TIRFM) greatly facilitates to imaging the first steps of endocytosis, a process whereby cells traffic cargo from the cell surface to endosomes. Using TIRFM, fluorescent-tagged endocytic proteins are observed as overlapped areas forming random clumps of different sizes, shapes and durations. A common procedure to segment these objects consists of thresholding the original gray-level images to produce binary sequences in which a pixel is covered or not by a given fluorescent-tagged protein. This binary logic is not appropriate because it leaves a free tuning parameter to be set by the user which can influence on the conclusions of the statist…
Psychophysical response to electrocutaneous stimulation.
1984
A method is presented to determine a reliable stimulus-sensation relationship particularly suitable for electrocutaneous stimulation. An experimental intensity-discrimination curve was obtained through simple psychophysical comparison tasks, and sensory response was inferred from integration of a JND's density function. The psychophysical response resembles a power law, although departures cannot be described in terms of a unique exponent. An estimate of binary information capacity per electrode is also given as a feature of a stimulation procedure that preserves a low value of the size-intensity product.
Modeling user preferences in content-based image retrieval: A novel attempt to bridge the semantic gap
2015
This paper is concerned with content-based image retrieval from a stochastic point of view. The semantic gap problem is addressed in two ways. First, a dimensional reduction is applied using the (pre-calculated) distances among images. The dimension of the reduced vector is the number of preferences that we allow the user to choose from, in this case, three levels. Second, the conditional probability distribution of the random user preference, given this reduced feature vector, is modeled using a proportional odds model. A new model is fitted at each iteration. The score used to rank the image database is based on the estimated probability function of the random preference. Additionally, so…
A principled approach to network-based classification and data representation
2013
Measures of similarity are fundamental in pattern recognition and data mining. Typically the Euclidean metric is used in this context, weighting all variables equally and therefore assuming equal relevance, which is very rare in real applications. In contrast, given an estimate of a conditional density function, the Fisher information calculated in primary data space implicitly measures the relevance of variables in a principled way by reference to auxiliary data such as class labels. This paper proposes a framework that uses a distance metric based on Fisher information to construct similarity networks that achieve a more informative and principled representation of data. The framework ena…
Automating statistical diagrammatic representations with data characterization
2017
The search for an efficient method to enhance data cognition is especially important when managing data from multidimensional databases. Open data policies have dramatically increased not only the volume of data available to the public, but also the need to automate the translation of data into efficient graphical representations. Graphic automation involves producing an algorithm that necessarily contains inputs derived from the type of data. A set of rules are then applied to combine the input variables and produce a graphical representation. Automated systems, however, fail to provide an efficient graphical representation because they only consider either a one-dimensional characterizat…
A Survey of Bayesian Techniques in Computer Vision
2010
The Bayesian approach to classification is intended to solve questions concerning how to assign a class to an observed pattern using probability estimations. Red, green and blue (RGB) or hue, saturation and lightness (HSL) values of pixels in digital colour images can be considered as feature vectors to be classified, thus leading to Bayesian colour image segmentation. Bayesian classifiers are also used to sort objects but, in this case, reduction of the dimensionality of the feature vector is often required prior to the analysis. This chapter shows some applications of Bayesian learning techniques in computer vision in the agriculture and agri-food sectors. Inspection and classification of…
Text Classification Using Novel “Anti-Bayesian” Techniques
2015
This paper presents a non-traditional “Anti-Bayesian” solution for the traditional Text Classification (TC) problem. Historically, all the recorded TC schemes work using the fundamental paradigm that once the statistical features are inferred from the syntactic/semantic indicators, the classifiers themselves are the well-established statistical ones. In this paper, we shall demonstrate that by virtue of the skewed distributions of the features, one could advantageously work with information latent in certain “non-central” quantiles (i.e., those distant from the mean) of the distributions. We, indeed, demonstrate that such classifiers exist and are attainable, and show that the design and im…
Predictive and Contextual Feature Separation for Bayesian Metanetworks
2007
Bayesian Networks are proven to be a comprehensive model to describe causal relationships among domain attributes with probabilistic measure of conditional dependency. However, depending on a context, many attributes of the model might not be relevant. If a Bayesian Network has been learned across multiple contexts then all uncovered conditional dependencies are averaged over all contexts and cannot guarantee high predictive accuracy when applied to a concrete case. We are considering a context as a set of contextual attributes, which are not directly effect probability distribution of the target attributes, but they effect on "relevance" of the predictive attributes towards target attribut…