Search results for "Machine learning"

showing 10 items of 1464 documents

An LP-based hyperparameter optimization model for language modeling

2018

In order to find hyperparameters for a machine learning model, algorithms such as grid search or random search are used over the space of possible values of the models hyperparameters. These search algorithms opt the solution that minimizes a specific cost function. In language models, perplexity is one of the most popular cost functions. In this study, we propose a fractional nonlinear programming model that finds the optimal perplexity value. The special structure of the model allows us to approximate it by a linear programming model that can be solved using the well-known simplex algorithm. To the best of our knowledge, this is the first attempt to use optimization techniques to find per…

FOS: Computer and information sciencesMathematical optimizationPerplexityLinear programmingComputer scienceMachine Learning (stat.ML)02 engineering and technology010501 environmental sciences01 natural sciencesTheoretical Computer ScienceNonlinear programmingMachine Learning (cs.LG)Random searchSimplex algorithmSearch algorithmStatistics - Machine Learning0202 electrical engineering electronic engineering information engineeringFOS: MathematicsMathematics - Optimization and Control0105 earth and related environmental sciencesHyperparameterComputer Science::Computation and Language (Computational Linguistics and Natural Language and Speech Processing)Computer Science - LearningHardware and ArchitectureOptimization and Control (math.OC)Hyperparameter optimization020201 artificial intelligence & image processingLanguage modelSoftwareInformation Systems
researchProduct

The Recycling Gibbs sampler for efficient learning

2018

Monte Carlo methods are essential tools for Bayesian inference. Gibbs sampling is a well-known Markov chain Monte Carlo (MCMC) algorithm, extensively used in signal processing, machine learning, and statistics, employed to draw samples from complicated high-dimensional posterior distributions. The key point for the successful application of the Gibbs sampler is the ability to draw efficiently samples from the full-conditional probability density functions. Since in the general case this is not possible, in order to speed up the convergence of the chain, it is required to generate auxiliary samples whose information is eventually disregarded. In this work, we show that these auxiliary sample…

FOS: Computer and information sciencesMonte Carlo methodSlice samplingInferenceMachine Learning (stat.ML)02 engineering and technologyBayesian inferenceStatistics - Computation01 natural sciencesMachine Learning (cs.LG)010104 statistics & probabilitysymbols.namesake[INFO.INFO-TS]Computer Science [cs]/Signal and Image ProcessingStatistics - Machine LearningArtificial IntelligenceStatistics0202 electrical engineering electronic engineering information engineering0101 mathematicsElectrical and Electronic EngineeringGaussian processComputation (stat.CO)ComputingMilieux_MISCELLANEOUSMathematicsChain rule (probability)Applied Mathematics020206 networking & telecommunicationsMarkov chain Monte CarloStatistics::ComputationComputer Science - LearningComputational Theory and MathematicsSignal ProcessingsymbolsComputer Vision and Pattern RecognitionStatistics Probability and UncertaintyAlgorithm[SPI.SIGNAL]Engineering Sciences [physics]/Signal and Image processingGibbs samplingDigital Signal Processing
researchProduct

Gaussianizing the Earth: Multidimensional Information Measures for Earth Data Analysis

2021

Information theory is an excellent framework for analyzing Earth system data because it allows us to characterize uncertainty and redundancy, and is universally interpretable. However, accurately estimating information content is challenging because spatio-temporal data is high-dimensional, heterogeneous and has non-linear characteristics. In this paper, we apply multivariate Gaussianization for probability density estimation which is robust to dimensionality, comes with statistical guarantees, and is easy to apply. In addition, this methodology allows us to estimate information-theoretic measures to characterize multivariate densities: information, entropy, total correlation, and mutual in…

FOS: Computer and information sciencesMultivariate statisticsGeneral Computer ScienceComputer scienceMachine Learning (stat.ML)Mutual informationInformation theorycomputer.software_genreStatistics - ApplicationsEarth system scienceRedundancy (information theory)13. Climate actionStatistics - Machine LearningGeneral Earth and Planetary SciencesEntropy (information theory)Applications (stat.AP)Total correlationData miningElectrical and Electronic EngineeringInstrumentationcomputerCurse of dimensionality
researchProduct

Understanding Climate Impacts on Vegetation with Gaussian Processes in Granger Causality

2020

Global warming is leading to unprecedented changes in our planet, with great societal, economical and environmental implications, especially with the growing demand of biofuels and food. Assessing the impact of climate on vegetation is of pressing need. We approached the attribution problem with a novel nonlinear Granger causal (GC) methodology and used a large data archive of remote sensing satellite products, environmental and climatic variables spatio-temporally gridded over more than 30 years. We generalize kernel Granger causality by considering the variables cross-relations explicitly in Hilbert spaces, and use the covariance in Gaussian processes. The method generalizes the linear an…

FOS: Computer and information sciencesPhysics - Atmospheric and Oceanic PhysicsComputer Science - Machine LearningAtmospheric and Oceanic Physics (physics.ao-ph)FOS: Physical sciencesMachine Learning (cs.LG)
researchProduct

PRINCIPAL POLYNOMIAL ANALYSIS

2014

© 2014 World Scientific Publishing Company. This paper presents a new framework for manifold learning based on a sequence of principal polynomials that capture the possibly nonlinear nature of the data. The proposed Principal Polynomial Analysis (PPA) generalizes PCA by modeling the directions of maximal variance by means of curves instead of straight lines. Contrarily to previous approaches PPA reduces to performing simple univariate regressions which makes it computationally feasible and robust. Moreover PPA shows a number of interesting analytical properties. First PPA is a volume preserving map which in turn guarantees the existence of the inverse. Second such an inverse can be obtained…

FOS: Computer and information sciencesPolynomialComputer Networks and CommunicationsComputer scienceMachine Learning (stat.ML)02 engineering and technologyReduction (complexity)03 medical and health sciencessymbols.namesake0302 clinical medicineStatistics - Machine LearningArtificial Intelligence0202 electrical engineering electronic engineering information engineeringPrincipal Polynomial AnalysisPrincipal Component AnalysisMahalanobis distanceModels StatisticalCodingDimensionality reductionNonlinear dimensionality reductionGeneral MedicineClassificationDimensionality reductionManifold learningNonlinear DynamicsMetric (mathematics)Jacobian matrix and determinantsymbolsRegression Analysis020201 artificial intelligence & image processingNeural Networks ComputerAlgorithmAlgorithms030217 neurology & neurosurgeryCurse of dimensionalityInternational Journal of Neural Systems
researchProduct

Quantum pattern recognition in photonic circuits

2021

This paper proposes a machine learning method to characterize photonic states via a simple optical circuit and data processing of photon number distributions, such as photonic patterns. The input states consist of two coherent states used as references and a two-mode unknown state to be studied. We successfully trained supervised learning algorithms that can predict the degree of entanglement in the two-mode state as well as perform the full tomography of one photonic mode, obtaining satisfactory values in the considered regression metrics.

FOS: Computer and information sciencesQuantum PhysicsComputer Science - Machine LearningData processingPhotonCondensed Matter - Mesoscale and Nanoscale PhysicsPhysics and Astronomy (miscellaneous)business.industryComputer scienceMaterials Science (miscellaneous)FOS: Physical sciencesQuantum entanglementAtomic and Molecular Physics and OpticsMachine Learning (cs.LG)Pattern recognition (psychology)Mesoscale and Nanoscale Physics (cond-mat.mes-hall)Coherent statesElectrical and Electronic EngineeringPhotonicsbusinessQuantum Physics (quant-ph)AlgorithmQuantumElectronic circuitQuantum Science and Technology
researchProduct

Regression of high-dimensional angular momentum states of light

2023

The Orbital Angular Momentum (OAM) of light is an infinite-dimensional degree of freedom of light with several applications in both classical and quantum optics. However, to fully take advantage of the potential of OAM states, reliable detection platforms to characterize generated states in experimental conditions are needed. Here, we present an approach to reconstruct input OAM states from measurements of the spatial intensity distributions they produce. To obviate issues arising from intrinsic symmetry of Laguerre-Gauss modes, we employ a pair of intensity profiles per state projecting it only on two distinct bases, showing how this allows to uniquely recover input states from the collect…

FOS: Computer and information sciencesQuantum PhysicsComputer Science - Machine Learningphotonicquantum opticFOS: Physical sciencesGeneral Physics and Astronomyorbital angular momentum of lightSettore FIS/03 - Fisica Della MateriaMachine Learning (cs.LG)machine learningquantum informationQuantum Physics (quant-ph)Optics (physics.optics)Physics - OpticsPhysical Review Research
researchProduct

Supervised Quantum Learning without Measurements

2017

We propose a quantum machine learning algorithm for efficiently solving a class of problems encoded in quantum controlled unitary operations. The central physical mechanism of the protocol is the iteration of a quantum time-delayed equation that introduces feedback in the dynamics and eliminates the necessity of intermediate measurements. The performance of the quantum algorithm is analyzed by comparing the results obtained in numerical simulations with the outcome of classical machine learning methods for the same problem. The use of time-delayed equations enhances the toolbox of the field of quantum machine learning, which may enable unprecedented applications in quantum technologies. The…

FOS: Computer and information sciencesQuantum machine learningField (physics)Computer Science - Artificial IntelligenceComputer sciencelcsh:MedicineFOS: Physical sciencesMachine Learning (stat.ML)01 natural sciencesUnitary stateArticle010305 fluids & plasmasSuperconductivity (cond-mat.supr-con)Statistics - Machine Learning0103 physical sciencesMesoscale and Nanoscale Physics (cond-mat.mes-hall)lcsh:Science010306 general physicsQuantumProtocol (object-oriented programming)Quantum PhysicsClass (computer programming)MultidisciplinaryCondensed Matter - Mesoscale and Nanoscale PhysicsCondensed Matter - Superconductivitylcsh:RQuantum technologyArtificial Intelligence (cs.AI)ComputerSystemsOrganization_MISCELLANEOUSlcsh:QQuantum algorithmQuantum Physics (quant-ph)Algorithm
researchProduct

Progressive Stochastic Binarization of Deep Networks

2019

A plethora of recent research has focused on improving the memory footprint and inference speed of deep networks by reducing the complexity of (i) numerical representations (for example, by deterministic or stochastic quantization) and (ii) arithmetic operations (for example, by binarization of weights). We propose a stochastic binarization scheme for deep networks that allows for efficient inference on hardware by restricting itself to additions of small integers and fixed shifts. Unlike previous approaches, the underlying randomized approximation is progressive, thus permitting an adaptive control of the accuracy of each operation at run-time. In a low-precision setting, we match the accu…

FOS: Computer and information sciencesScheme (programming language)Computer Science - Machine LearningComputer scienceStochastic processScalar (physics)Sampling (statistics)Machine Learning (stat.ML)Machine Learning (cs.LG)Statistics - Machine LearningApproximation errorBounded functionReference implementationRepresentation (mathematics)computerAlgorithmcomputer.programming_language2019 Fifth Workshop on Energy Efficient Machine Learning and Cognitive Computing - NeurIPS Edition (EMC2-NIPS)
researchProduct

At Your Service: Coffee Beans Recommendation From a Robot Assistant

2020

With advances in the field of machine learning, precisely algorithms for recommendation systems, robot assistants are envisioned to become more present in the hospitality industry. Additionally, the COVID-19 pandemic has also highlighted the need to have more service robots in our everyday lives, to minimise the risk of human to-human transmission. One such example would be coffee shops, which have become intrinsic to our everyday lives. However, serving an excellent cup of coffee is not a trivial feat as a coffee blend typically comprises rich aromas, indulgent and unique flavours and a lingering aftertaste. Our work addresses this by proposing a computational model which recommends optima…

FOS: Computer and information sciencesService (systems architecture)business.industryComputer scienceFeature vectorSupervised learningComputer Science - Human-Computer InteractionComputingMilieux_PERSONALCOMPUTING02 engineering and technologyRecommender systemMachine learningcomputer.software_genreField (computer science)GeneralLiterature_MISCELLANEOUSComputer Science - Information RetrievalPersonalizationHuman-Computer Interaction (cs.HC)0202 electrical engineering electronic engineering information engineeringRobotUnsupervised learning020201 artificial intelligence & image processingArtificial intelligencebusinesscomputerInformation Retrieval (cs.IR)
researchProduct