Search results for "computer.software_genre"

showing 10 items of 3858 documents

Using mathematical morphology for unsupervised classification of functional data

2011

This paper is concerned with the unsupervised classification of functional data by using mathematical morphology. Different morphological operators are used to extract relevant structures of the functions (considered as sets through their subgraph representations). These operators can be considered as preprocessing tools whose outputs are also functional data. We explore some dissimilarity measures and clustering methods for the classification of the transformed data. Our approach is illustrated through a detailed analysis of two data sets. These techniques, which have mainly been used in image processing, provide a flexible and robust toolbox for improving the results in unsupervised funct…

Statistics and ProbabilityApplied MathematicsData classificationImage processingMathematical morphologycomputer.software_genreToolboxComputingMethodologies_PATTERNRECOGNITIONModeling and SimulationPreprocessorData miningStatistics Probability and UncertaintyCluster analysisMorphological operatorscomputerMathematicsJournal of Statistical Computation and Simulation
researchProduct

An introduction to Bayesian reference analysis: inference on the ratio of multinomial parameters

1998

This paper offers an introduction to Bayesian reference analysis, often described as the more successful method to produce non-subjective, model-based, posterior distributions. The ideas are illustrated in detail with an interesting problem, the ratio of multinomial parameters, for which no model-based Bayesian analysis has been proposed. Signposts are provided to the huge related literature.

Statistics and ProbabilityBayesian probabilityPosterior probabilityInferenceBayesian inferencecomputer.software_genreStatistics::ComputationBayesian statisticsComputingMethodologies_PATTERNRECOGNITIONPrior probabilityEconometricsData miningBayesian linear regressionBayesian averagecomputerMathematicsJournal of the Royal Statistical Society: Series D (The Statistician)
researchProduct

A model-based approach to Spotify data analysis: a Beta GLMM

2020

Digital music distribution is increasingly powered by automated mechanisms that continuously capture, sort and analyze large amounts of Web-based data. This paper deals with the management of songs audio features from a statistical point of view. In particular, it explores the data catching mechanisms enabled by Spotify Web API and suggests statistical tools for the analysis of these data. Special attention is devoted to songs popularity and a Beta model, including random effects, is proposed in order to give the first answer to questions like: which are the determinants of popularity? The identification of a model able to describe this relationship, the determination within the set of char…

Statistics and ProbabilityBeta GLMMDistribution (number theory)Computer scienceApplication Notes0211 other engineering and technologies02 engineering and technologycomputer.software_genreWeb API01 natural sciencesSet (abstract data type)010104 statistics & probabilitySpotify Web API audio features Popularity Index Beta GLMMsortSpotify Web API0101 mathematicsDigital audio021103 operations researchPoint (typography)Random effects modelData sciencePopularityIdentification (information)Popularity IndexData miningStatistics Probability and Uncertaintycomputeraudio feature
researchProduct

Cluster-Localized Sparse Logistic Regression for SNP Data

2012

The task of analyzing high-dimensional single nucleotide polymorphism (SNP) data in a case-control design using multivariable techniques has only recently been tackled. While many available approaches investigate only main effects in a high-dimensional setting, we propose a more flexible technique, cluster-localized regression (CLR), based on localized logistic regression models, that allows different SNPs to have an effect for different groups of individuals. Separate multivariable regression models are fitted for the different groups of individuals by incorporating weights into componentwise boosting, which provides simultaneous variable selection, hence sparse fits. For model fitting, th…

Statistics and ProbabilityBoosting (machine learning)Computer scienceMultivariable calculusComputational BiologyHigh-Throughput Nucleotide SequencingFeature selectionRegression analysisModels TheoreticalLogistic regressioncomputer.software_genrePolymorphism Single NucleotideRegressionComputational MathematicsLogistic ModelsData Interpretation StatisticalGeneticsCluster AnalysisHumansData miningCluster analysisMolecular BiologyUnit-weighted regressioncomputerGenome-Wide Association StudyStatistical Applications in Genetics and Molecular Biology
researchProduct

Multiple testing in candidate gene situations: a comparison of classical, discrete, and resampling-based procedures.

2011

In candidate gene association studies, usually several elementary hypotheses are tested simultaneously using one particular set of data. The data normally consist of partly correlated SNP information. Every SNP can be tested for association with the disease, e.g., using the Cochran-Armitage test for trend. To account for the multiplicity of the test situation, different types of multiple testing procedures have been proposed. The question arises whether procedures taking into account the discreteness of the situation show a benefit especially in case of correlated data. We empirically evaluate several different multiple testing procedures via simulation studies using simulated correlated SN…

Statistics and ProbabilityCandidate geneContrast (statistics)computer.software_genrePolymorphism Single NucleotideSet (abstract data type)Computational MathematicsSample size determinationResamplingData Interpretation StatisticalSample SizeStatisticsMultiple comparisons problemGeneticsCochran–Armitage test for trendRange (statistics)HumansComputer SimulationDiseaseData miningMolecular BiologycomputerGenetic Association StudiesMathematicsStatistical applications in genetics and molecular biology
researchProduct

Opportunities and challenges of combined effect measures based on prioritized outcomes

2013

Many authors have proposed different approaches to combine multiple endpoints in a univariate outcome measure in the literature. In case of binary or time-to-event variables, composite endpoints, which combine several event types within a single event or time-to-first-event analysis are often used to assess the overall treatment effect. A main drawback of this approach is that the interpretation of the composite effect can be difficult as a negative effect in one component can be masked by a positive effect in another. Recently, some authors proposed more general approaches based on a priority ranking of outcomes, which moreover allow to combine outcome variables of different scale levels. …

Statistics and ProbabilityClinical Trials as TopicEpidemiologyUnivariatecomputer.software_genreOutcome (game theory)Treatment OutcomeRankingScale (social sciences)Component (UML)Outcome Assessment Health CareMultiple comparisons problemHumansComputer SimulationData miningcomputerProportional Hazards ModelsMathematicsStatistical hypothesis testingEvent (probability theory)Statistics in Medicine
researchProduct

Sparse relative risk regression models

2020

Summary Clinical studies where patients are routinely screened for many genomic features are becoming more routine. In principle, this holds the promise of being able to find genomic signatures for a particular disease. In particular, cancer survival is thought to be closely linked to the genomic constitution of the tumor. Discovering such signatures will be useful in the diagnosis of the patient, may be used for treatment decisions and, perhaps, even the development of new treatments. However, genomic data are typically noisy and high-dimensional, not rarely outstripping the number of patients included in the study. Regularized survival models have been proposed to deal with such scenarios…

Statistics and ProbabilityClustering high-dimensional dataComputer sciencedgLARSInferenceScale (descriptive set theory)BiostatisticsMachine learningcomputer.software_genreRisk Assessment01 natural sciencesRegularization (mathematics)Relative risk regression model010104 statistics & probability03 medical and health sciencesNeoplasmsCovariateHumansComputer Simulation0101 mathematicsOnline Only ArticlesSurvival analysis030304 developmental biology0303 health sciencesModels Statisticalbusiness.industryLeast-angle regressionRegression analysisGeneral MedicineSurvival AnalysisHigh-dimensional dataGene expression dataRegression AnalysisArtificial intelligenceStatistics Probability and UncertaintySettore SECS-S/01 - StatisticabusinessSparsitycomputerBiostatistics
researchProduct

An interest rates cluster analysis

2004

An empirical analysis of interest rates in money and capital markets is performed. We investigate a set of 34 different weekly interest rate time series during a time period of 16 years between 1982 and 1997. Our study is focused on the collective behavior of the stochastic fluctuations of these time-series which is investigated by using a clustering linkage procedure. Without any a priori assumption, we individuate a meaningful separation in 6 main clusters organized in a hierarchical structure.

Statistics and ProbabilityCollective behaviormedia_common.quotation_subjectFOS: Physical sciencesLinkage (mechanical)computer.software_genrelaw.inventionFOS: Economics and businesslawEconometricsCluster (physics)Cluster analysisCondensed Matter - Statistical Mechanicsmedia_commonStatistical Finance (q-fin.ST)Statistical Mechanics (cond-mat.stat-mech)EconophysicsSeries (mathematics)Quantitative Finance - Statistical FinanceCondensed Matter PhysicsInterest rateCondensed Matter - Other Condensed MatterData miningCapital marketcomputerOther Condensed Matter (cond-mat.other)
researchProduct

An overview of robust Bayesian analysis

1994

Robust Bayesian analysis is the study of the sensitivity of Bayesian answers to uncertain inputs. This paper seeks to provide an overview of the subject, one that is accessible to statisticians outside the field. Recent developments in the area are also reviewed, though with very uneven emphasis. © 1994 SEIO.

Statistics and ProbabilityComputer scienceBayesian probabilitycomputer.software_genreData scienceField (computer science)Bayesian robustnessN/ARobust Bayesian analysisPrior probabilityData miningSensitivity (control systems)Statistics Probability and Uncertaintycomputer
researchProduct

A new mathematical approach for the estimation of the AUC and its variability under different experimental designs in preclinical studies

2011

The aim of the present work was to develop a new mathematical method for estimating the area under the curve (AUC) and its variability that could be applied in different preclinical experimental designs and amenable to be implemented in standard calculation worksheets. In order to assess the usefulness of the new approach, different experimental scenarios were studied and the results were compared with those obtained with commonly used software: WinNonlin® and Phoenix WinNonlin®. The results do not show statistical differences among the AUC values obtained by both procedures, but the new method appears to be a better estimator of the AUC standard error, measured as the coverage of 95% confi…

Statistics and ProbabilityComputer scienceDrug Evaluation PreclinicalAdministration Oralcomputer.software_genreSoftwareCiprofloxacinArea under curveVariance estimationAnimalsPharmacology (medical)Rats WistarPharmacologyModels Statisticalbusiness.industryDesign of experimentsEstimatorModels TheoreticalConfidence intervalRatsStandard errorResearch DesignArea Under CurveData miningbusinesscomputerSoftwarePharmaceutical Statistics
researchProduct