Search results for "sampling"

showing 10 items of 788 documents

Utilization of long duration high-volume sampling coupled to SPME-GC-MS/MS for the assessment of airborne pesticides variability in an urban area (St…

2016

Atmospheric samples have been collected between 14 March and 12 September 2012 on a 2-week basis (15 days of sampling and exchange of traps each 7 days) in Strasbourg (east of France) for the analysis of 43 pesticides. Samples (particle and gas phases) were separately extracted using Accelerated Solvent Extraction (ASE) and pre-concentrated by Solid Phase Micro-Extraction (SPME) before analysis by gas chromatography coupled to tandem mass spectrometry (GC-MS/MS). Four SPME consecutive injections at distinct temperatures were made in order to increase the sensitivity of detection for the all monitored pesticides. Currently used detected pesticides can be grouped in four classes; those used i…

Chromatography Gas010501 environmental sciences01 natural sciencesGas Chromatography-Mass Spectrometrychemistry.chemical_compoundTandem Mass SpectrometryAcetochlorCitiesPesticides0105 earth and related environmental sciencesFenpropimorphPesticide residueBromoxynilSolid Phase Extraction010401 analytical chemistryPesticide ResiduesSampling (statistics)General MedicinePesticidePollution0104 chemical scienceschemistryEnvironmental chemistryParticulate MatterFranceGas chromatographyGas chromatography–mass spectrometryAgrochemicalsEnvironmental MonitoringFood ScienceJournal of Environmental Science and Health, Part B
researchProduct

Incremental linear model trees on massive datasets

2013

The existence of massive datasets raises the need for algorithms that make efficient use of resources like memory and computation time. Besides well-known approaches such as sampling, online algorithms are being recognized as good alternatives, as they often process datasets faster using much less memory. The important class of algorithms learning linear model trees online (incremental linear model trees or ILMTs in the following) offers interesting options for regression tasks in this sense. However, surprisingly little is known about their performance, as there exists no large-scale evaluation on massive stationary datasets under equal conditions. Therefore, this paper shows their applica…

Class (computer programming)Computer scienceProcess (engineering)business.industryComputationLinear modelSampling (statistics)computer.software_genreMachine learningKISS principleData miningArtificial intelligenceOnline algorithmbusinesscomputerProceedings of the 28th Annual ACM Symposium on Applied Computing
researchProduct

Importance sampling for Lambda-coalescents in the infinitely many sites model

2011

We present and discuss new importance sampling schemes for the approximate computation of the sample probability of observed genetic types in the infinitely many sites model from population genetics. More specifically, we extend the 'classical framework', where genealogies are assumed to be governed by Kingman's coalescent, to the more general class of Lambda-coalescents and develop further Hobolth et. al.'s (2008) idea of deriving importance sampling schemes based on 'compressed genetrees'. The resulting schemes extend earlier work by Griffiths and Tavar\'e (1994), Stephens and Donnelly (2000), Birkner and Blath (2008) and Hobolth et. al. (2008). We conclude with a performance comparison o…

Class (set theory)ComputationSample (statistics)62F99 (Primary) 62P10 92D10 92D20 (Secondary)LambdaArticleSampling StudiesCoalescent theoryEvolution MolecularGene FrequencyFOS: MathematicsQuantitative Biology::Populations and EvolutionAnimalsQuantitative Biology - Populations and EvolutionEcology Evolution Behavior and Systematicscomputer.programming_languageMathematicsDiscrete mathematicsModels GeneticBETA (programming language)Probability (math.PR)Populations and Evolution (q-bio.PE)Markov ChainsGenetics PopulationPerformance comparisonFOS: Biological sciencesMutationcomputerMonte Carlo MethodMathematics - ProbabilityImportance sampling
researchProduct

Channel selection in Cognitive Radio Networks: A Switchable Bayesian Learning Automata approach

2013

We consider the problem of a user operating within a Cognitive Radio Network (CRN) which involves N channels each associated with a Primary User (PU). The problem consists of allocating a channel which, at any given time instant is not being used by a PU, to a Secondary User (SU). Within our study, we assume that a SU is allowed to perform “channel switching”, i.e., to choose an alternate channel S times (where S +1 ≤ N) if the previous choice does not lead to a channel which is vacant. The paper first presents a formal probabilistic model for the problem itself, referred to as the Formal Secondary Channel Selection (FSCS) problem, and the characteristics of the FSCS are then analyzed. Ther…

Cognitive radioTheoretical computer sciencebusiness.industryComputer scienceBayesian probabilitySampling (statistics)Statistical modelArtificial intelligenceBayesian inferencebusinessProbability vectorCommunication channelAutomaton2013 IEEE 24th Annual International Symposium on Personal, Indoor, and Mobile Radio Communications (PIMRC)
researchProduct

On the uniform sampling of CIELAB color space and the number of discernible colors

2013

This paper presents a useful algorithmic strategy to sample uniformly the CIELAB color space based on close packed hexagonal grid. This sampling scheme has been used successfully in different research works from computational color science to color image processing. The main objective of this paper is to demonstrate the relevance and the accuracy of the hexagonal grid sampling method applied to the CIELAB color space. The second objective of this paper is to show that the number of color samples computed depends on the application and on the color gamut boundary considered. As demonstration, we use this sampling to support a discussion on the number of discernible colors related to a JND.

Color histogram[ INFO.INFO-TS ] Computer Science [cs]/Signal and Image Processing[INFO.INFO-TS] Computer Science [cs]/Signal and Image ProcessingComputingMethodologies_IMAGEPROCESSINGANDCOMPUTERVISIONcomputational color imagingColor balance[ SPI.SIGNAL ] Engineering Sciences [physics]/Signal and Image processing02 engineering and technologyperceptually uniform color spaceColor space01 natural sciences010309 optics[INFO.INFO-TS]Computer Science [cs]/Signal and Image ProcessingICC profile0103 physical sciencesColor depth[ INFO.INFO-TI ] Computer Science [cs]/Image Processing0202 electrical engineering electronic engineering information engineering3D close packed hexagonal gridComputer visionSamplingComputingMilieux_MISCELLANEOUS[SPI.SIGNAL] Engineering Sciences [physics]/Signal and Image processingComputingMethodologies_COMPUTERGRAPHICSMathematicsColor differencebusiness.industry020207 software engineeringColor quantizationColor modelArtificial intelligencebusiness[SPI.SIGNAL]Engineering Sciences [physics]/Signal and Image processing
researchProduct

Population Monte Carlo Schemes with Reduced Path Degeneracy

2017

Population Monte Carlo (PMC) algorithms are versatile adaptive tools for approximating moments of complicated distributions. A common problem of PMC algorithms is the so-called path degeneracy; the diversity in the adaptation is endangered due to the resampling step. In this paper we focus on novel population Monte Carlo schemes that present enhanced diversity, compared to the standard approach, while keeping the same implementation structure (sample generation, weighting and resampling). The new schemes combine different weighting and resampling strategies to reduce the path degeneracy and achieve a higher performance at the cost of additional low computational complexity cost. Computer si…

Computational complexity theoryMonte Carlo methodApproximation algorithm020206 networking & telecommunications02 engineering and technology01 natural sciencesStatistics::ComputationWeighting010104 statistics & probabilitysymbols.namesake[INFO.INFO-TS]Computer Science [cs]/Signal and Image ProcessingGaussian noiseResamplingPath (graph theory)0202 electrical engineering electronic engineering information engineeringsymbols0101 mathematicsDegeneracy (mathematics)Algorithm[SPI.SIGNAL]Engineering Sciences [physics]/Signal and Image processingComputingMilieux_MISCELLANEOUS
researchProduct

Thompson Sampling for Dynamic Multi-armed Bandits

2011

The importance of multi-armed bandit (MAB) problems is on the rise due to their recent application in a large variety of areas such as online advertising, news article selection, wireless networks, and medicinal trials, to name a few. The most common assumption made when solving such MAB problems is that the unknown reward probability theta k of each bandit arm k is fixed. However, this assumption rarely holds in practice simply because real-life problems often involve underlying processes that are dynamically evolving. In this paper, we model problems where reward probabilities theta k are drifting, and introduce a new method called Dynamic Thompson Sampling (DTS) that facilitates Order St…

Computer Science::Machine LearningMathematical optimizationbusiness.industryComputer scienceOrder statisticBayesian probabilitySampling (statistics)RegretArtificial intelligencebusinessThompson samplingRandom variableSelection (genetic algorithm)2011 10th International Conference on Machine Learning and Applications and Workshops
researchProduct

Seismic evaluation of ordinary RC buildings retrofitted with externally bonded FRPs using a reliability-based approach

2020

International audience; Despite the extensive literature on reinforced concrete (RC) members retrofitted with fiberreinforced polymer (FRP) composites, few studies have employed a reliability-based approach to evaluate the seismic performance of RC buildings in terms of their collapse capacity and ductility. In this study, the performance of a poorly-confined RC building structure is investigated for different FRP retrofitting schemes using different configurations and combinations of wrapping and flange-bonded FRPs, as two well-established techniques. A nonlinear pushover analysis is then implemented with a computational reliability analysis based on Latin Hypercube Sampling (LHS) to deter…

Computer science02 engineering and technologyRetrofitting0203 mechanical engineeringRC buildings[PHYS.MECA.SOLID]Physics [physics]/Mechanics [physics]/Solid mechanics [physics.class-ph]RetrofittingCollapse capacityDuctilityReliability (statistics)Civil and Structural EngineeringDuctilitybusiness.industryProbabilistic logicFailure modeStructural engineeringFibre-reinforced plastic021001 nanoscience & nanotechnologyReliability020303 mechanical engineering & transportsLatin hypercube samplingCeramics and Composites0210 nano-technologybusinessMaterial propertiesFailure mode and effects analysisFRP
researchProduct

Adaptive Importance Sampling: The past, the present, and the future

2017

A fundamental problem in signal processing is the estimation of unknown parameters or functions from noisy observations. Important examples include localization of objects in wireless sensor networks [1] and the Internet of Things [2]; multiple source reconstruction from electroencephalograms [3]; estimation of power spectral density for speech enhancement [4]; or inference in genomic signal processing [5]. Within the Bayesian signal processing framework, these problems are addressed by constructing posterior probability distributions of the unknowns. The posteriors combine optimally all of the information about the unknowns in the observations with the information that is present in their …

Computer scienceBayesian probabilityPosterior probabilityInference02 engineering and technologyMachine learningcomputer.software_genre01 natural sciences010104 statistics & probabilityMultidimensional signal processing[INFO.INFO-TS]Computer Science [cs]/Signal and Image ProcessingPrior probability0202 electrical engineering electronic engineering information engineering0101 mathematicsElectrical and Electronic EngineeringComputingMilieux_MISCELLANEOUSbusiness.industryApplied Mathematics020206 networking & telecommunicationsApproximate inferenceSignal ProcessingProbability distributionArtificial intelligencebusinessAlgorithmcomputer[SPI.SIGNAL]Engineering Sciences [physics]/Signal and Image processingImportance sampling
researchProduct

A chirp-z transform-based synchronizer for power system measurements

2005

In the last few years, increased interest in power and voltage quality has forced international working groups to standardize testing and measurement techniques. IEC 61000-4-30, which defines the characteristics of instrumentation for the measurement of power quality, refers to IEC 61000-4-7 for the evaluation of harmonics and interharmonics. This standard, revised in 2002, requires a synchronous sampling of voltage or current signal, in order to limit errors and to ensure reproducible results even in the presence of nonstationary signals. Therefore, an accurate estimation of the fundamental frequency is required, even in the presence of disturbances. In this paper, an algorithm to detect t…

Computer scienceBluestein's FFT algorithmFast Fourier transformChirp-z transform power quality synchronizationFundamental frequencyPower (physics)Electric power systemSampling (signal processing)SynchronizerHarmonicsElectronic engineeringElectrical and Electronic EngineeringInstrumentationSettore ING-INF/07 - Misure Elettriche E ElettronicheInterpolation
researchProduct