Search results for "Vector"

showing 10 items of 2660 documents

Kernel methods and their derivatives: Concept and perspectives for the earth system sciences.

2020

Kernel methods are powerful machine learning techniques which implement generic non-linear functions to solve complex tasks in a simple way. They Have a solid mathematical background and exhibit excellent performance in practice. However, kernel machines are still considered black-box models as the feature mapping is not directly accessible and difficult to interpret.The aim of this work is to show that it is indeed possible to interpret the functions learned by various kernel methods is intuitive despite their complexity. Specifically, we show that derivatives of these functions have a simple mathematical formulation, are easy to compute, and can be applied to many different problems. We n…

FOS: Computer and information sciencesComputer Science - Machine LearningSupport Vector MachineTheoretical computer scienceComputer scienceEntropyKernel FunctionsNormal Distribution0211 other engineering and technologies02 engineering and technologyMachine Learning (cs.LG)Machine LearningStatistics - Machine LearningSimple (abstract algebra)0202 electrical engineering electronic engineering information engineeringOperator TheoryData ManagementMultidisciplinaryGeographyApplied MathematicsSimulation and ModelingQRDensity estimationKernel methodKernel (statistics)Physical SciencessymbolsMedicine020201 artificial intelligence & image processingAlgorithmsResearch ArticleComputer and Information SciencesScienceMachine Learning (stat.ML)Research and Analysis MethodsKernel MethodsKernel (linear algebra)symbols.namesakeArtificial IntelligenceSupport Vector MachinesHumansEntropy (information theory)Computer SimulationGaussian process021101 geological & geomatics engineeringData VisualizationCorrectionRandom VariablesFunction (mathematics)Probability TheorySupport vector machineAlgebraPhysical GeographyLinear AlgebraEarth SciencesEigenvectorsRandom variableMathematicsEarth SystemsPLoS ONE
researchProduct

Human experts vs. machines in taxa recognition

2020

The step of expert taxa recognition currently slows down the response time of many bioassessments. Shifting to quicker and cheaper state-of-the-art machine learning approaches is still met with expert scepticism towards the ability and logic of machines. In our study, we investigate both the differences in accuracy and in the identification logic of taxonomic experts and machines. We propose a systematic approach utilizing deep Convolutional Neural Nets with the transfer learning paradigm and extensively evaluate it over a multi-pose taxonomic dataset with hierarchical labels specifically created for this comparison. We also study the prediction accuracy on different ranks of taxonomic hier…

FOS: Computer and information sciencesComputer Science - Machine Learninghahmontunnistus (tietotekniikka)Computer scienceClassification approachTaxonomic expert02 engineering and technologyneuroverkotcomputer.software_genreConvolutional neural networkQuantitative Biology - Quantitative MethodsField (computer science)Machine Learning (cs.LG)Machine learning approachesStatistics - Machine LearningAutomated approachDeep neural networks0202 electrical engineering electronic engineering information engineeringTaxonomic rankQuantitative Methods (q-bio.QM)Classification (of information)Artificial neural networksystematiikka (biologia)Prediction accuracyIdentification (information)koneoppiminenMulti-image dataBenchmark (computing)020201 artificial intelligence & image processingConvolutional neural networksComputer Vision and Pattern RecognitionClassification errorsMachine Learning (stat.ML)Machine learningState of the artElectrical and Electronic EngineeringTaxonomySupport vector machinesLearning systemsbusiness.industryNode (networking)020206 networking & telecommunicationsComputer circuitsHierarchical classificationConvolutionSupport vector machineFOS: Biological sciencesTaxonomic hierarchySignal ProcessingBiomonitoringBenchmark datasetsArtificial intelligencebusinesscomputertaksonitSoftware
researchProduct

A survey of active learning algorithms for supervised remote sensing image classification

2011

Defining an efficient training set is one of the most delicate phases for the success of remote sensing image classification routines. The complexity of the problem, the limited temporal and financial resources, as well as the high intraclass variance can make an algorithm fail if it is trained with a suboptimal dataset. Active learning aims at building efficient training sets by iteratively improving the model performance through sampling. A user-defined heuristic ranks the unlabeled pixels according to a function of the uncertainty of their class membership and then the user is asked to provide labels for the most uncertain pixels. This paper reviews and tests the main families of active …

FOS: Computer and information sciencesComputer scienceComputer Vision and Pattern Recognition (cs.CV)Computer Science - Computer Vision and Pattern RecognitionMachine learningcomputer.software_genreactive learningHyperspectral image classificationEntropy (information theory)Electrical and Electronic EngineeringArchitectureRemote sensingvery high resolution (VHR)PixelContextual image classificationbusiness.industryHyperspectral imagingSupport vector machinehyperspectraltraining set definitionSignal Processingsupport vector machine (SVM)Artificial intelligenceHeuristicsbusinessAlgorithmcomputerimage classification
researchProduct

Anti-powers in infinite words

2018

In combinatorics of words, a concatenation of $k$ consecutive equal blocks is called a power of order $k$. In this paper we take a different point of view and define an anti-power of order $k$ as a concatenation of $k$ consecutive pairwise distinct blocks of the same length. As a main result, we show that every infinite word contains powers of any order or anti-powers of any order. That is, the existence of powers or anti-powers is an unavoidable regularity. Indeed, we prove a stronger result, which relates the density of anti-powers to the existence of a factor that occurs with arbitrary exponent. As a consequence, we show that in every aperiodic uniformly recurrent word, anti-powers of ev…

FOS: Computer and information sciencesDiscrete Mathematics (cs.DM)Formal Languages and Automata Theory (cs.FL)ConcatenationComputer Science - Formal Languages and Automata Theory68R150102 computer and information sciences01 natural sciencesTheoretical Computer ScienceCombinatoricsUnavoidable regularityPosition (vector)Infinite wordAvoidability[MATH.MATH-CO]Mathematics [math]/Combinatorics [math.CO]FOS: MathematicsMathematics - CombinatoricsDiscrete Mathematics and CombinatoricsOrder (group theory)Point (geometry)0101 mathematicsDiscrete Mathematics and CombinatoricMathematicsDiscrete mathematics000 Computer science knowledge general worksAnti-power010101 applied mathematicsComputational Theory and Mathematics010201 computation theory & mathematicsAperiodic graphComputer ScienceExponentPairwise comparisonCombinatorics (math.CO)SoftwareWord (group theory)Computer Science - Discrete Mathematics
researchProduct

Fast MATLAB assembly of FEM matrices in 2D and 3D: Edge elements

2014

We propose an effective and flexible way to assemble finite element stiffness and mass matrices in MATLAB. We apply this for problems discretized by edge finite elements. Typical edge finite elements are Raviart-Thomas elements used in discretizations of H(div) spaces and Nedelec elements in discretizations of H(curl) spaces. We explain vectorization ideas and comment on a freely available MATLAB code which is fast and scalable with respect to time.

FOS: Computer and information sciencesDiscretizationfinite element method97N80 65M60Matlab codeComputational scienceMathematics::Numerical AnalysisMATLAB code vectorizationmedicineFOS: MathematicsMathematics - Numerical AnalysisMATLABMathematicscomputer.programming_languageCurl (mathematics)ta113Nédélec elementApplied Mathematicsta111StiffnessRaviart–Thomas elementMixed finite element methodNumerical Analysis (math.NA)Finite element methodComputational Mathematicsedge elementScalabilityComputer Science - Mathematical Softwaremedicine.symptomcomputerMathematical Software (cs.MS)
researchProduct

On generalized Lyndon words

2018

Abstract A generalized lexicographical order on infinite words is defined by choosing for each position a total order on the alphabet. This allows to define generalized Lyndon words. Every word in the free monoid can be factorized in a unique way as a nonincreasing factorization of generalized Lyndon words. We give new characterizations of the first and the last factor in this factorization as well as new characterization of generalized Lyndon words. We also give more specific results on two special cases: the classical one and the one arising from the alternating lexicographical order.

FOS: Computer and information sciencesGeneral Computer ScienceDiscrete Mathematics (cs.DM)Computer Science::Computation and Language (Computational Linguistics and Natural Language and Speech Processing)68R15Characterization (mathematics)Lexicographical orderTheoretical Computer ScienceLyndon wordsCombinatoricsFactorizationPosition (vector)Free monoidFOS: MathematicsOrder (group theory)Mathematics - CombinatoricsCombinatorics (math.CO)Word (group theory)Computer Science::Formal Languages and Automata TheoryMathematicsComputer Science - Discrete Mathematics
researchProduct

Multiscale Information Decomposition: Exact Computation for Multivariate Gaussian Processes

2017

Exploiting the theory of state space models, we derive the exact expressions of the information transfer, as well as redundant and synergistic transfer, for coupled Gaussian processes observed at multiple temporal scales. All of the terms, constituting the frameworks known as interaction information decomposition and partial information decomposition, can thus be analytically obtained for different time scales from the parameters of the VAR model that fits the processes. We report the application of the proposed methodology firstly to benchmark Gaussian systems, showing that this class of systems may generate patterns of information decomposition characterized by prevalently redundant or sy…

FOS: Computer and information sciencesInformation transferComputer scienceGaussianSocial SciencesGeneral Physics and AstronomyInformation theory01 natural sciences010305 fluids & plasmasState spaceStatistical physicslcsh:Scienceinformation theorymultiscale entropylcsh:QC1-999Interaction informationMathematics and Statisticssymbolsinformation dynamicsInformation dynamics; Information transfer; Multiscale entropy; Multivariate time series analysis; Redundancy and synergy; State space models; Vector autoregressive models; Physics and Astronomy (all)information dynamics; information transfer; multiscale entropy; multivariate time series analysis; redundancy and synergy; state space models; vector autoregressive modelsMultivariate time series analysiMathematics - Statistics Theorylcsh:AstrophysicsStatistics Theory (math.ST)Statistics - ApplicationsMethodology (stat.ME)symbols.namesakePhysics and Astronomy (all)0103 physical scienceslcsh:QB460-466FOS: Mathematicsinformation transferRelevance (information retrieval)Applications (stat.AP)Transfer Entropy010306 general physicsGaussian processStatistics - MethodologyState space modelstate space modelsmultivariate time series analysisredundancy and synergyvector autoregressive modelsInformation dynamicVector autoregressive modelSettore ING-INF/06 - Bioingegneria Elettronica E InformaticaTransfer entropylcsh:Qlcsh:PhysicsEntropy
researchProduct

Local Granger causality

2021

Granger causality is a statistical notion of causal influence based on prediction via vector autoregression. For Gaussian variables it is equivalent to transfer entropy, an information-theoretic measure of time-directed information transfer between jointly dependent processes. We exploit such equivalence and calculate exactly the 'local Granger causality', i.e. the profile of the information transfer at each discrete time point in Gaussian processes; in this frame Granger causality is the average of its local version. Our approach offers a robust and computationally fast method to follow the information transfer along the time history of linear stochastic processes, as well as of nonlinear …

FOS: Computer and information sciencesInformation transferGaussianFOS: Physical sciencestechniques; information theory; granger causalityMachine Learning (stat.ML)Quantitative Biology - Quantitative Methods01 natural sciences010305 fluids & plasmasVector autoregressionsymbols.namesakegranger causalityGranger causalityStatistics - Machine Learning0103 physical sciencesApplied mathematicstime serie010306 general physicsQuantitative Methods (q-bio.QM)Mathematicsinformation theoryStochastic processDisordered Systems and Neural Networks (cond-mat.dis-nn)Condensed Matter - Disordered Systems and Neural NetworksComputational Physics (physics.comp-ph)Discrete time and continuous timeAutoregressive modelFOS: Biological sciencesSettore ING-INF/06 - Bioingegneria Elettronica E InformaticasymbolsTransfer entropytechniquesPhysics - Computational Physics
researchProduct

ALGORITHMS FOR JUMBLED PATTERN MATCHING IN STRINGS

2011

The Parikh vector p(s) of a string s is defined as the vector of multiplicities of the characters. Parikh vector q occurs in s if s has a substring t with p(t)=q. We present two novel algorithms for searching for a query q in a text s. One solves the decision problem over a binary text in constant time, using a linear size index of the text. The second algorithm, for a general finite alphabet, finds all occurrences of a given Parikh vector q and has sub-linear expected time complexity; we present two variants, which both use a linear size index of the text.

FOS: Computer and information sciencesJ.3average case analysis.Binary numberaverage case analysispermuted stringpermuted stringsComputer Science - Data Structures and AlgorithmsComputer Science (miscellaneous)Parikh vectorData Structures and Algorithms (cs.DS)Pattern matchingTime complexityMathematicsString (computer science)Parikh vectorsstring algorithmDecision problemstring algorithmsSubstringParikh vectors; permuted strings; pattern matching; string algorithms; average case analysisF.2.2; J.3Index (publishing)pattern matchingF.2.2Constant (mathematics)AlgorithmComputer Science::Formal Languages and Automata Theory
researchProduct

Randomized Block Frank–Wolfe for Convergent Large-Scale Learning

2017

Owing to their low-complexity iterations, Frank-Wolfe (FW) solvers are well suited for various large-scale learning tasks. When block-separable constraints are present, randomized block FW (RB-FW) has been shown to further reduce complexity by updating only a fraction of coordinate blocks per iteration. To circumvent the limitations of existing methods, the present work develops step sizes for RB-FW that enable a flexible selection of the number of blocks to update per iteration while ensuring convergence and feasibility of the iterates. To this end, convergence rates of RB-FW are established through computational bounds on a primal sub-optimality measure and on the duality gap. The novel b…

FOS: Computer and information sciencesMathematical optimization0102 computer and information sciences02 engineering and technology01 natural sciencesMeasure (mathematics)Machine Learning (cs.LG)Convergence (routing)FOS: Mathematics0202 electrical engineering electronic engineering information engineeringFraction (mathematics)Electrical and Electronic EngineeringMathematics - Optimization and ControlMathematicsSequenceDuality gapComputer Science - Numerical Analysis020206 networking & telecommunicationsNumerical Analysis (math.NA)Stationary pointSupport vector machineComputer Science - LearningOptimization and Control (math.OC)010201 computation theory & mathematicsIterated functionSignal ProcessingAlgorithmIEEE Transactions on Signal Processing
researchProduct