Search results for "Statistics - Machine Learning"

showing 10 items of 90 documents

Visualization of Jacques Lacan's Registers of the Psychoanalytic Field, and Discovery of Metaphor and of Metonymy. Analytical Case Study of Edgar All…

2016

We start with a description of Lacan's work that we then take into our analytics methodology. In a first investigation, a Lacan-motivated template of the Poe story is fitted to the data. A segmentation of the storyline is used in order to map out the diachrony. Based on this, it will be shown how synchronous aspects, potentially related to Lacanian registers, can be sought. This demonstrates the effectiveness of an approach based on a model template of the storyline narrative. In a second and more comprehensive investigation, we develop an approach for revealing, that is, uncovering, Lacanian register relationships. Objectives of this work include the wide and general application of our met…

FOS: Computer and information sciencesI.2Computer Science - Computation and LanguageStatistics - Machine LearningI.5.3I.5.462H25 62H30G.3Machine Learning (stat.ML)G.2.2Computation and Language (cs.CL)I.5.3; I.5.4; I.2; G.2.2; G.3
researchProduct

Local Granger causality

2021

Granger causality is a statistical notion of causal influence based on prediction via vector autoregression. For Gaussian variables it is equivalent to transfer entropy, an information-theoretic measure of time-directed information transfer between jointly dependent processes. We exploit such equivalence and calculate exactly the 'local Granger causality', i.e. the profile of the information transfer at each discrete time point in Gaussian processes; in this frame Granger causality is the average of its local version. Our approach offers a robust and computationally fast method to follow the information transfer along the time history of linear stochastic processes, as well as of nonlinear …

FOS: Computer and information sciencesInformation transferGaussianFOS: Physical sciencestechniques; information theory; granger causalityMachine Learning (stat.ML)Quantitative Biology - Quantitative Methods01 natural sciences010305 fluids & plasmasVector autoregressionsymbols.namesakegranger causalityGranger causalityStatistics - Machine Learning0103 physical sciencesApplied mathematicstime serie010306 general physicsQuantitative Methods (q-bio.QM)Mathematicsinformation theoryStochastic processDisordered Systems and Neural Networks (cond-mat.dis-nn)Condensed Matter - Disordered Systems and Neural NetworksComputational Physics (physics.comp-ph)Discrete time and continuous timeAutoregressive modelFOS: Biological sciencesSettore ING-INF/06 - Bioingegneria Elettronica E InformaticasymbolsTransfer entropytechniquesPhysics - Computational Physics
researchProduct

CLEAR: Covariant LEAst-Square Refitting with Applications to Image Restoration

2017

International audience; In this paper, we propose a new framework to remove parts of the systematic errors affecting popular restoration algorithms, with a special focus for image processing tasks. Generalizing ideas that emerged for $\ell_1$ regularization, we develop an approach re-fitting the results of standard methods towards the input data. Total variation regularizations and non-local means are special cases of interest. We identify important covariant information that should be preserved by the re-fitting method, and emphasize the importance of preserving the Jacobian (w.r.t. the observed signal) of the original estimator. Then, we provide an approach that has a ``twicing'' flavor a…

FOS: Computer and information sciencesInverse problemsMathematical optimization[ INFO.INFO-TS ] Computer Science [cs]/Signal and Image ProcessingComputer Vision and Pattern Recognition (cs.CV)General MathematicsComputer Science - Computer Vision and Pattern RecognitionMachine Learning (stat.ML)Mathematics - Statistics TheoryImage processingStatistics Theory (math.ST)02 engineering and technologyDebiasing[ INFO.INFO-CV ] Computer Science [cs]/Computer Vision and Pattern Recognition [cs.CV]01 natural sciencesRegularization (mathematics)Boosting010104 statistics & probabilitysymbols.namesake[INFO.INFO-TS]Computer Science [cs]/Signal and Image Processing[STAT.ML]Statistics [stat]/Machine Learning [stat.ML]Variational methods[MATH.MATH-ST]Mathematics [math]/Statistics [math.ST]Statistics - Machine LearningRefittingMSC: 49N45 65K10 68U10[ INFO.INFO-TI ] Computer Science [cs]/Image ProcessingFOS: Mathematics0202 electrical engineering electronic engineering information engineeringCovariant transformation[ MATH.MATH-ST ] Mathematics [math]/Statistics [math.ST]0101 mathematicsImage restoration[ STAT.ML ] Statistics [stat]/Machine Learning [stat.ML]MathematicsApplied Mathematics[INFO.INFO-CV]Computer Science [cs]/Computer Vision and Pattern Recognition [cs.CV]EstimatorInverse problem[INFO.INFO-TI]Computer Science [cs]/Image Processing [eess.IV]Jacobian matrix and determinantsymbolsTwicing020201 artificial intelligence & image processingAffine transformationAlgorithm
researchProduct

Effectiveness of Data-Driven Induction of Semantic Spaces and Traditional Classifiers for Sarcasm Detection

2019

Irony and sarcasm are two complex linguistic phenomena that are widely used in everyday language and especially over the social media, but they represent two serious issues for automated text understanding. Many labeled corpora have been extracted from several sources to accomplish this task, and it seems that sarcasm is conveyed in different ways for different domains. Nonetheless, very little work has been done for comparing different methods among the available corpora. Furthermore, usually, each author collects and uses their own datasets to evaluate his own method. In this paper, we show that sarcasm detection can be tackled by applying classical machine learning algorithms to input te…

FOS: Computer and information sciencesLinguistics and LanguageComputer Science - Machine LearningComputer sciencemedia_common.quotation_subjectSemantic spaceMachine Learning (stat.ML)02 engineering and technologycomputer.software_genreLanguage and LinguisticsTask (project management)Data-drivenMachine Learning (cs.LG)Artificial IntelligenceStatistics - Machine Learning020204 information systemsEveryday language0202 electrical engineering electronic engineering information engineeringSocial medianatural language processingmedia_commonComputer Science - Computation and LanguageSarcasmSettore INF/01 - Informaticabusiness.industryirony detectionIronymachine learningsemantic spaces020201 artificial intelligence & image processingArtificial intelligencebusinessIrony detectionsemantic spacecomputerComputation and Language (cs.CL)SoftwareNatural language processingsarcasm detection
researchProduct

Metropolis Sampling

2017

Monte Carlo (MC) sampling methods are widely applied in Bayesian inference, system simulation and optimization problems. The Markov Chain Monte Carlo (MCMC) algorithms are a well-known class of MC methods which generate a Markov chain with the desired invariant distribution. In this document, we focus on the Metropolis-Hastings (MH) sampler, which can be considered as the atom of the MCMC techniques, introducing the basic notions and different properties. We describe in details all the elements involved in the MH algorithm and the most relevant variants. Several improvements and recent extensions proposed in the literature are also briefly discussed, providing a quick but exhaustive overvie…

FOS: Computer and information sciencesMachine Learning (stat.ML)020206 networking & telecommunications02 engineering and technologyStatistics - Computation01 natural sciencesStatistics::ComputationMethodology (stat.ME)010104 statistics & probabilityStatistics - Machine Learning0202 electrical engineering electronic engineering information engineering0101 mathematicsComputation (stat.CO)Statistics - MethodologyWiley StatsRef: Statistics Reference Online
researchProduct

Adaptive independent sticky MCMC algorithms

2018

In this work, we introduce a novel class of adaptive Monte Carlo methods, called adaptive independent sticky MCMC algorithms, for efficient sampling from a generic target probability density function (pdf). The new class of algorithms employs adaptive non-parametric proposal densities which become closer and closer to the target as the number of iterations increases. The proposal pdf is built using interpolation procedures based on a set of support points which is constructed iteratively based on previously drawn samples. The algorithm's efficiency is ensured by a test that controls the evolution of the set of support points. This extra stage controls the computational cost and the converge…

FOS: Computer and information sciencesMathematical optimizationAdaptive Markov chain Monte Carlo (MCMC)Monte Carlo methodBayesian inferenceHASettore SECS-P/05 - Econometrialcsh:TK7800-8360Machine Learning (stat.ML)02 engineering and technologyBayesian inference01 natural sciencesStatistics - Computationlcsh:Telecommunication010104 statistics & probabilitysymbols.namesakeAdaptive Markov chain Monte Carlo (MCMC); Adaptive rejection Metropolis sampling (ARMS); Bayesian inference; Gibbs sampling; Hit and run algorithm; Metropolis-within-Gibbs; Monte Carlo methods; Signal Processing; Hardware and Architecture; Electrical and Electronic EngineeringGibbs samplingStatistics - Machine Learninglcsh:TK5101-67200202 electrical engineering electronic engineering information engineeringComputational statisticsMetropolis-within-GibbsHit and run algorithm0101 mathematicsElectrical and Electronic EngineeringGaussian processComputation (stat.CO)MathematicsSignal processinglcsh:Electronics020206 networking & telecommunicationsMarkov chain Monte CarloMonte Carlo methodsHardware and ArchitectureSignal ProcessingSettore SECS-S/03 - Statistica EconomicasymbolsSettore SECS-S/01 - StatisticaStatistical signal processingGibbs samplingAdaptive rejection Metropolis sampling (ARMS)EURASIP Journal on Advances in Signal Processing
researchProduct

Consistent Regression of Biophysical Parameters with Kernel Methods

2020

This paper introduces a novel statistical regression framework that allows the incorporation of consistency constraints. A linear and nonlinear (kernel-based) formulation are introduced, and both imply closed-form analytical solutions. The models exploit all the information from a set of drivers while being maximally independent of a set of auxiliary, protected variables. We successfully illustrate the performance in the estimation of chlorophyll content.

FOS: Computer and information sciencesMathematical optimizationComputer Science - Machine Learning010504 meteorology & atmospheric sciences0211 other engineering and technologiesRegression analysisMachine Learning (stat.ML)02 engineering and technology01 natural sciencesRegressionData modelingMachine Learning (cs.LG)Set (abstract data type)Methodology (stat.ME)Nonlinear systemKernel methodConsistency (statistics)Statistics - Machine LearningKernel (statistics)Statistics - Methodology021101 geological & geomatics engineering0105 earth and related environmental sciencesMathematicsIGARSS 2018 - 2018 IEEE International Geoscience and Remote Sensing Symposium
researchProduct

An LP-based hyperparameter optimization model for language modeling

2018

In order to find hyperparameters for a machine learning model, algorithms such as grid search or random search are used over the space of possible values of the models hyperparameters. These search algorithms opt the solution that minimizes a specific cost function. In language models, perplexity is one of the most popular cost functions. In this study, we propose a fractional nonlinear programming model that finds the optimal perplexity value. The special structure of the model allows us to approximate it by a linear programming model that can be solved using the well-known simplex algorithm. To the best of our knowledge, this is the first attempt to use optimization techniques to find per…

FOS: Computer and information sciencesMathematical optimizationPerplexityLinear programmingComputer scienceMachine Learning (stat.ML)02 engineering and technology010501 environmental sciences01 natural sciencesTheoretical Computer ScienceNonlinear programmingMachine Learning (cs.LG)Random searchSimplex algorithmSearch algorithmStatistics - Machine Learning0202 electrical engineering electronic engineering information engineeringFOS: MathematicsMathematics - Optimization and Control0105 earth and related environmental sciencesHyperparameterComputer Science::Computation and Language (Computational Linguistics and Natural Language and Speech Processing)Computer Science - LearningHardware and ArchitectureOptimization and Control (math.OC)Hyperparameter optimization020201 artificial intelligence & image processingLanguage modelSoftwareInformation Systems
researchProduct

The Recycling Gibbs sampler for efficient learning

2018

Monte Carlo methods are essential tools for Bayesian inference. Gibbs sampling is a well-known Markov chain Monte Carlo (MCMC) algorithm, extensively used in signal processing, machine learning, and statistics, employed to draw samples from complicated high-dimensional posterior distributions. The key point for the successful application of the Gibbs sampler is the ability to draw efficiently samples from the full-conditional probability density functions. Since in the general case this is not possible, in order to speed up the convergence of the chain, it is required to generate auxiliary samples whose information is eventually disregarded. In this work, we show that these auxiliary sample…

FOS: Computer and information sciencesMonte Carlo methodSlice samplingInferenceMachine Learning (stat.ML)02 engineering and technologyBayesian inferenceStatistics - Computation01 natural sciencesMachine Learning (cs.LG)010104 statistics & probabilitysymbols.namesake[INFO.INFO-TS]Computer Science [cs]/Signal and Image ProcessingStatistics - Machine LearningArtificial IntelligenceStatistics0202 electrical engineering electronic engineering information engineering0101 mathematicsElectrical and Electronic EngineeringGaussian processComputation (stat.CO)ComputingMilieux_MISCELLANEOUSMathematicsChain rule (probability)Applied Mathematics020206 networking & telecommunicationsMarkov chain Monte CarloStatistics::ComputationComputer Science - LearningComputational Theory and MathematicsSignal ProcessingsymbolsComputer Vision and Pattern RecognitionStatistics Probability and UncertaintyAlgorithm[SPI.SIGNAL]Engineering Sciences [physics]/Signal and Image processingGibbs samplingDigital Signal Processing
researchProduct

Gaussianizing the Earth: Multidimensional Information Measures for Earth Data Analysis

2021

Information theory is an excellent framework for analyzing Earth system data because it allows us to characterize uncertainty and redundancy, and is universally interpretable. However, accurately estimating information content is challenging because spatio-temporal data is high-dimensional, heterogeneous and has non-linear characteristics. In this paper, we apply multivariate Gaussianization for probability density estimation which is robust to dimensionality, comes with statistical guarantees, and is easy to apply. In addition, this methodology allows us to estimate information-theoretic measures to characterize multivariate densities: information, entropy, total correlation, and mutual in…

FOS: Computer and information sciencesMultivariate statisticsGeneral Computer ScienceComputer scienceMachine Learning (stat.ML)Mutual informationInformation theorycomputer.software_genreStatistics - ApplicationsEarth system scienceRedundancy (information theory)13. Climate actionStatistics - Machine LearningGeneral Earth and Planetary SciencesEntropy (information theory)Applications (stat.AP)Total correlationData miningElectrical and Electronic EngineeringInstrumentationcomputerCurse of dimensionality
researchProduct