Search results for "computer.software_genre"

showing 10 items of 3858 documents

Learning Improved Feature Rankings through Decremental Input Pruning for Support Vector Based Drug Activity Prediction

2010

The use of certain machine learning and pattern recognition tools for automated pharmacological drug design has been recently introduced. Different families of learning algorithms and Support Vector Machines in particular have been applied to the task of associating observed chemical properties and pharmacological activities to certain kinds of representations of the candidate compounds. The purpose of this work, is to select an appropriate feature ordering from a large set of molecular descriptors usually used in the domain of Drug Activity Characterization. To this end, a new input pruning method is introduced and assessed with respect to commonly used feature ranking algorithms.

Computer scienceActive learning (machine learning)business.industryFeature vectorPattern recognitionMachine learningcomputer.software_genreKernel methodComputational learning theoryRanking SVMFeature (machine learning)Artificial intelligencePruning (decision trees)businessFeature learningcomputer
researchProduct

Revisitation of Nonorthogonal Spin Adaptation in Coupled Cluster Theory.

2015

The benefits of what is alternatively called a nonorthogonally spin-adapted, spin-free, or orbital representation of the coupled cluster equations is discussed relative to orthogonally spin-adapted, spin-orbital, and spin-integrated theories. In particular, specific linear combinations of the orbital cluster amplitudes, denoted spin-summed amplitudes, are shown to reduce the number of contractions that must be explicitly performed and to simplify the expressions and their derivation. The computational efficiency of the spin-summed approach is discussed and compared to orthogonally spin-adapted and spin-integrated approaches. The spin-summed approach is shown to have significant computationa…

Computer scienceAdaptation (eye)computer.software_genreComputer Science ApplicationsAmplitudeCoupled clusterCluster (physics)Condensed Matter::Strongly Correlated ElectronsData miningStatistical physicsPhysical and Theoretical ChemistryRepresentation (mathematics)Linear combinationcomputerSpin-½Journal of chemical theory and computation
researchProduct

Hierarchical modeling for rare event detection and cell subset alignment across flow cytometry samples.

2013

Flow cytometry is the prototypical assay for multi-parameter single cell analysis, and is essential in vaccine and biomarker research for the enumeration of antigen-specific lymphocytes that are often found in extremely low frequencies (0.1% or less). Standard analysis of flow cytometry data relies on visual identification of cell subsets by experts, a process that is subjective and often difficult to reproduce. An alternative and more objective approach is the use of statistical models to identify cell subsets of interest in an automated fashion. Two specific challenges for automated analysis are to detect extremely low frequency event subsets without biasing the estimate by pre-processing…

Computer scienceAdaptive Immunitycomputer.software_genre0302 clinical medicineSingle-cell analysisEnumerationBiology (General)Immune ResponseEvent (probability theory)0303 health sciencesEcologymedicine.diagnostic_testT CellsStatisticsFlow Cytometry3. Good healthComputational Theory and MathematicsData modelModeling and SimulationMedicineData miningImmunotherapyResearch ArticleTumor ImmunologyQH301-705.5Immune CellsImmunologyContext (language use)BiostatisticsModels BiologicalFlow cytometry03 medical and health sciencesCellular and Molecular NeuroscienceGeneticsmedicineHumansSensitivity (control systems)Statistical MethodsImmunoassaysMolecular BiologyBiologyEcology Evolution Behavior and Systematics030304 developmental biologybusiness.industryImmunityReproducibility of ResultsPattern recognitionStatistical modelImmunologic SubspecialtiesLymphocyte SubsetsImmunologic TechniquesClinical ImmunologyArtificial intelligencebusinesscomputerMathematics030215 immunologyPLoS computational biology
researchProduct

Compression-based classification of biological sequences and structures via the Universal Similarity Metric: experimental assessment.

2007

Abstract Background Similarity of sequences is a key mathematical notion for Classification and Phylogenetic studies in Biology. It is currently primarily handled using alignments. However, the alignment methods seem inadequate for post-genomic studies since they do not scale well with data set size and they seem to be confined only to genomic and proteomic sequences. Therefore, alignment-free similarity measures are actively pursued. Among those, USM (Universal Similarity Metric) has gained prominence. It is based on the deep theory of Kolmogorov Complexity and universality is its most novel striking feature. Since it can only be approximated via data compression, USM is a methodology rath…

Computer scienceAlgorismesPrediction by partial matchingCompression dissimilaritycomputer.software_genreBiochemistryProtein Structure SecondaryPhylogenetic studiesStructural BiologySequence Analysis ProteinDatabases Proteinlcsh:QH301-705.5Biological dataNCDApplied MathematicsGenomicsClassificationCDComputer Science ApplicationsBenchmarking:Informàtica::Informàtica teòrica [Àrees temàtiques de la UPC]Universal compression dissimilarityArea Under CurveMetric (mathematics)lcsh:R858-859.7Data miningAlgorithmsData compressionResearch Article:Informàtica::Aplicacions de la informàtica::Bioinformàtica [Àrees temàtiques de la UPC]Normalization (statistics)lcsh:Computer applications to medicine. Medical informaticsBioinformatics Sequence Alignment AlgorithmsSet (abstract data type)Similarity (network science)Normalized compression sissimilarityData compression (Computer science)AnimalsHumansAmino Acid SequenceMolecular BiologyBiologyDades -- Compressió (Informàtica)USMUniversal similarity metricProteinsUCDProtein Structure TertiaryData setGenòmicaStatistical classificationlcsh:Biology (General)ROC CurvecomputerSequence AlignmentSoftwareBMC bioinformatics
researchProduct

Machine Learning Techniques for Intrusion Detection: A Comparative Analysis

2016

International audience; With the growth of internet world has transformed into a global market with all monetary and business exercises being carried online. Being the most imperative resource of the developing scene, it is the vulnerable object and hence needs to be secured from the users with dangerous personality set. Since the Internet does not have focal surveillance component, assailants once in a while, utilizing varied and advancing hacking topologies discover a path to bypass framework " s security and one such collection of assaults is Intrusion. An intrusion is a movement of breaking into the framework by compromising the security arrangements of the framework set up. The techniq…

Computer scienceAnomaly-based intrusion detection system02 engineering and technologyIntrusion detection systemIDSMachine learningcomputer.software_genre[ INFO.INFO-CV ] Computer Science [cs]/Computer Vision and Pattern Recognition [cs.CV][INFO.INFO-AI]Computer Science [cs]/Artificial Intelligence [cs.AI]Machine LearningResource (project management)Component (UML)0202 electrical engineering electronic engineering information engineeringROCSet (psychology)[ INFO.INFO-AI ] Computer Science [cs]/Artificial Intelligence [cs.AI]False Positivebusiness.industryACM[INFO.INFO-CV]Computer Science [cs]/Computer Vision and Pattern Recognition [cs.CV]020206 networking & telecommunicationsPrecisionObject (computer science)True PositiveOutlier020201 artificial intelligence & image processingThe InternetArtificial intelligenceData miningbusinesscomputer
researchProduct

Combining conjunctive rule extraction with diffusion maps for network intrusion detection

2013

Network security and intrusion detection are important in the modern world where communication happens via information networks. Traditional signature-based intrusion detection methods cannot find previously unknown attacks. On the other hand, algorithms used for anomaly detection often have black box qualities that are difficult to understand for people who are not algorithm experts. Rule extraction methods create interpretable rule sets that act as classifiers. They have mostly been combined with already labeled data sets. This paper aims to combine unsupervised anomaly detection with rule extraction techniques to create an online anomaly detection framework. Unsupervised anomaly detectio…

Computer scienceAnomaly-based intrusion detection systemNetwork securityintrusion detectiontunkeutumisen havaitseminenFeature extractionDiffusion mapdiffusion mapIntrusion detection systemMachine learningcomputer.software_genrepoikkeavuuden havaitseminenBlack boxtiedon louhintan-grammiCluster analysista113Training setrule extractionbusiness.industryn-gramanomaly detectiondiffuusiokarttakoneoppiminensääntöjen erottaminenAnomaly detectionArtificial intelligenceData miningtiedonlouhintabusinesscomputer2013 IEEE Symposium on Computers and Communications (ISCC)
researchProduct

UnipaBCI a novel general software framework for brain computer interface

2017

The increasing interest in Brain Computer Interface (BCI) requires new fast, reliable and scalable frameworks that can be used by researchers to develop BCI based high performance applications in efficient and fast ways. In this paper is presented "UnipaBCI", a general software framework for BCI applications based on electroencephalogra-phy (EEG) that can fulfill these new needs. A visual evoked potentials (VEP) application has also been developed using the proposed framework in order to test the modular architecture and the overall performance. Different types of users (beginners and experts in BCI) have been involved during the "UnipaBCI" experimental test and they have exhibited good and…

Computer scienceAugmentative communication02 engineering and technologyVisual evoked potentialsHumanoid robotElectroencephalographycomputer.software_genre03 medical and health sciences0302 clinical medicineInformationSystems_MODELSANDPRINCIPLESBrain-Computer Interface (BCI) Humanoid Robot Assistive technology Augmentative Communication rehabilitation BCI frameworkHuman–computer interaction0202 electrical engineering electronic engineering information engineeringmedicineOverall performanceBrain–computer interfaceSettore ING-INF/05 - Sistemi Di Elaborazione Delle Informazionimedicine.diagnostic_testRehabilitationModular architectureBCI frameworkSoftware frameworkAssistive technologyScalability020201 artificial intelligence & image processingcomputerBrain-Computer Interface (BCI)030217 neurology & neurosurgery
researchProduct

Enriqueciendo la investigación en humanidades digitales. Análisis de textos de claustros académicos de la Universidad de Valencia (1775-1779) con KH …

2020

[ES] La aplicación de métodos automatizados en cualquier investigación ha facilitado el trasvase de metodologías de una disciplina a otra, permitiendo realizar análisis cuantitativos a textos con estructura o semiestructurados. El objeto de este trabajo es aplicar a un dataset en lenguaje natural -castellano del siglo XVIII- métodos de análisis de la disciplina de documentación. Pretende establecer una metodología automática de análisis cuantitativo y cualitativo de textos, que permita enriquecer en el futuro las conclusiones procedentes del análisis histórico tradicional. Este estudio construye los procedimientos necesarios para poder aplicar análisis de frecuencia, extracción y clasificac…

Computer scienceBIBLIOTECONOMIA Y DOCUMENTACIONUniversitatsLibrary and Information Sciences050905 science studiescomputer.software_genresiglo xviiiBibliography. Library science. Information resourcesQualitative analysis5202.01 Metodología de InvestigaciónClaustros universitarios1209.03 Análisis de Datoscastellano antiguoKH CoderSiglo XVIIIbusiness.industry05 social sciencesAnàlisi del discursanálisis de lenguaje naturalAutomationHumanitats InformàticaAnálisis de lenguaje naturalkh coderclaustros universitariosWork (electrical)Quantitative analysis (finance)Natural language analysisArtificial intelligence0509 other social sciencesCastellano antiguo050904 information & library sciencesbusinesscomputerWord frequency analysisNatural language processingZRevista Española de Documentación Científica
researchProduct

Adaptive Importance Sampling: The past, the present, and the future

2017

A fundamental problem in signal processing is the estimation of unknown parameters or functions from noisy observations. Important examples include localization of objects in wireless sensor networks [1] and the Internet of Things [2]; multiple source reconstruction from electroencephalograms [3]; estimation of power spectral density for speech enhancement [4]; or inference in genomic signal processing [5]. Within the Bayesian signal processing framework, these problems are addressed by constructing posterior probability distributions of the unknowns. The posteriors combine optimally all of the information about the unknowns in the observations with the information that is present in their …

Computer scienceBayesian probabilityPosterior probabilityInference02 engineering and technologyMachine learningcomputer.software_genre01 natural sciences010104 statistics & probabilityMultidimensional signal processing[INFO.INFO-TS]Computer Science [cs]/Signal and Image ProcessingPrior probability0202 electrical engineering electronic engineering information engineering0101 mathematicsElectrical and Electronic EngineeringComputingMilieux_MISCELLANEOUSbusiness.industryApplied Mathematics020206 networking & telecommunicationsApproximate inferenceSignal ProcessingProbability distributionArtificial intelligencebusinessAlgorithmcomputer[SPI.SIGNAL]Engineering Sciences [physics]/Signal and Image processingImportance sampling
researchProduct

Manulex-infra: Distributional characteristics of grapheme—phoneme mappings, and infralexical and lexical units in child-directed written material

2007

It is well known that the statistical characteristics of a language, such as word frequency or the consistency of the relationships between orthography and phonology, influence literacy acquisition. Accordingly, linguistic databases play a central role by compiling quantitative and objective estimates about the principal variables that affect reading and writing acquisition. We describe a new set of Web-accessible databases of French orthography whose main characteristic is that they are based on frequency analyses of words occurring in reading books used in the elementary school grades. Quantitative estimates were made for several infralexical variables (syllable, grapheme-to-phoneme mappi…

Computer scienceBigrammedia_common.quotation_subjectExperimental and Cognitive PsychologyHomophonycomputer.software_genreVocabularyManuals as TopicArts and Humanities (miscellaneous)PhoneticsReading (process)Developmental and Educational PsychologyHumansChildGeneral Psychologymedia_commonPsycholinguisticsbusiness.industryPhonologyLinguisticsWord lists by frequencyWritten languagePsychology (miscellaneous)Artificial intelligenceSyllablebusinesscomputerNatural language processingOrthographyBehavior Research Methods
researchProduct