Search results for "speech recognition"

showing 10 items of 357 documents

A comparison of methods for investigating the perceptual center of musical sounds

2019

In speech and music, the acoustic and perceptual onset(s) of a sound are usually not congruent with its perceived temporal location. Rather, these "P-centers" are heard some milliseconds after the acoustic onset, and a variety of techniques have been used in speech and music research to find them. Here we report on a comparative study that uses various forms of the method of adjustment (aligning a click or filtered noise in-phase or anti-phase to a repeated target sound), as well as tapping in synchrony with a repeated target sound. The advantages and disadvantages of each method and probe type are discussed, and then all methods are tested using a set of musical instrument sounds that syst…

MaleP-centerLinguistics and LanguageComputer scienceSpeech recognitionmedia_common.quotation_subjectmusiikkipsykologiaExperimental and Cognitive PsychologyMusical instrumentMusicalStimulus (physiology)negative mean asynchronyLanguage and Linguisticsalignment taskYoung AdultRhythmPerceptionmicrotimingtutkimusmenetelmätotorhinolaryngologic diseasesHumansCenter frequencymedia_commontapping taskAcousticsrytmiSensory SystemsSoundAcoustic StimulationRise timeTime PerceptionAuditory PerceptionTappingFemaleMusicAttention, Perception, & Psychophysics
researchProduct

miMic

2016

miMic, a sonic analogue of paper and pencil is proposed: An augmented microphone for vocal and gestural sonic sketching. Vocalizations are classified and interpreted as instances of sound models, which the user can play with by vocal and gestural control. The physical device is based on a modified microphone, with embedded inertial sensors and buttons. Sound models can be selected by vocal imitations that are automatically classified, and each model is mapped to vocal and gestural features for real-time control. With miMic, the sound designer can explore a vast sonic space and quickly produce expressive sonic sketches, which may be turned into sound prototypes by further adjustment of model…

sonic interaction design system architecture vocal sketching sound designSound (medical instrument)Settore INF/01 - InformaticaInformationSystems_INFORMATIONINTERFACESANDPRESENTATION(e.g.HCI)Computer scienceMicrophoneSound designSpeech recognition05 social sciencesModel parameterssystem architecturevocal sketchingsonic interaction designAugmented microphone050105 experimental psychologysound designGestureInertial measurement unitSonic interaction designSettore ICAR/13 - Disegno Industriale0501 psychology and cognitive sciences050107 human factorsPencil (mathematics)GestureProceedings of the TEI '16: Tenth International Conference on Tangible, Embedded, and Embodied Interaction
researchProduct

Head movements in Finnish Sign Language on the basis of Motion Capture data

2015

This paper reports a study of the forms and functions of head movements produced in the dimension of depth in Finnish Sign Language (FinSL). Specifically, the paper describes and analyzes the phonetic forms and prosodic, grammatical, communicative, and textual functions of nods, head thrusts, nodding, and head pulls occurring in FinSL data consisting of a continuous dialogue recorded with motion capture technology. The analysis yields a novel classification of the kinematic characteristics and functional properties of the four types of head movement. However, it also reveals that there is no perfect correspondence between form and function in the head movements investigated.

Linguistics and LanguageCommunicationComputer scienceMovement (music)Head (linguistics)business.industrymedia_common.quotation_subjectSpeech recognitionOf the formKinematicsSign languageMotion captureLanguage and LinguisticsProsodybusinessFunction (engineering)media_commonSign Language & Linguistics
researchProduct

Tonal Hierarchies in Jazz Improvisation

1995

Statistical methods were used to investigate 18 bebop-styled jazz improvisations based on the so- called Rhythm Changes chord progression. The data were compared with results obtained by C. L. Krumhansl and her colleagues in empirical tests investigating the perceived stability of the tones in the chromatic scale in various contexts. Comparisons were also made with data on the statistical distribution of the 12 chromatic tones in actual European art music. It was found that the chorus- level hierarchies (measured over a whole chorus) are remarkably similar to the rating profiles obtained in empirical tests and to the relative frequencies of the tones in European art music. The chord- level …

ImprovisationClassical musicHierarchyRhythmbiologySpeech recognitionChorusChord (music)Chromatic scaleJazzbiology.organism_classificationMusicMathematicsMusic Perception
researchProduct

Numerical simulation of glottal flow

2012

In cases of permanent immobility of both vocal folds patients have difficulties with breathing but rarely with voicing. However, clinical experience shows that the shape of the larynx (voice box) seems to have a significant influence on the degree of airflow and breathing pattern. In order to find an optimal geometry of the larynx in terms of easiness for breathing after the surgical change of vocal folds or false vocal cords (ventricular folds), a set of numerical simulations of glottal flow for weakly compressible Navier-Stokes equations has been performed. We compare airflow resistance and volumetric flow rate for several geometry concepts for inspiration as well as expiration. Finally, …

MaleLarynxComputer simulationRespirationSpeech recognitionAcousticsAirflowHealth InformaticsVocal Cordsrespiratory systemModels BiologicalComputer Science ApplicationsGlottal flowBreathing patternmedicine.anatomical_structureVocal Cord DysfunctionVocal foldsotorhinolaryngologic diseasesmedicineBreathingHumansVoiceFemaleMathematicsComputers in Biology and Medicine
researchProduct

Transposed-letter effects: Consonants, vowels and letter frequency

2008

There is now considerable evidence (e.g., Perea & Lupker, 2003a, 2003b) that transposed-letter nonword primes (e.g., jugde for JUDGE) are more effective primes than replacement-letter nonword primes (e.g., jupte for JUDGE). Recently, Perea and Lupker (2004) demonstrated that, in Spanish, this transposed-letter prime advantage exists only when the transposed letters are consonants (C-C transpositions) and not when they are vowels (V-V transpositions). This vowel-consonant difference causes problems even for models that can successfully explain transposed-letter effects (e.g., SOLAR, Davis, 1999). In Experiment 1 in the present paper, we demonstrated a parallel result in a language with a dif…

Linguistics and LanguagePrime (symbol)Speech recognitionLexical decision taskLetter frequencyExperimental and Cognitive PsychologySyllabic versePsychologyPriming (psychology)Language and LinguisticsLinguisticsEducationLanguage and Cognitive Processes
researchProduct

Do Grading Gray Stimuli Help to Encode Letter Position?

2021

Numerous experiments in the past decades recurrently showed that a transposed-letter pseudoword (e.g., JUGDE) is much more wordlike than a replacement-letter control (e.g., JUPTE). Critically, there is an ongoing debate as to whether this effect arises at a perceptual level (e.g., perceptual uncertainty at assigning letter position of an array of visual objects) or at an abstract language-specific level (e.g., via a level of “open bigrams” between the letter and word levels). Here, we designed an experiment to test the limits of perceptual accounts of letter position coding. The stimuli in a lexical decision task were presented either with a homogeneous letter intensity or with a graded gra…

Cognitive Neurosciencemedia_common.quotation_subjectBigramSpeech recognitionword recognition050105 experimental psychologyorthographic processingVisual ObjectsPerceptionperceptual factorsLexical decision task0501 psychology and cognitive sciencesControl (linguistics)lcsh:QH301-705.5computer.programming_languagemedia_commonlexical decisionCommunication05 social sciences050301 educationCell BiologySensory SystemsPseudowordOphthalmologylcsh:Biology (General)letter position codingWord recognitionPsychology0503 educationcomputerOptometryCoding (social sciences)Vision
researchProduct

Same–different discrepancy in an exhaustive matching task

1988

In this study, we investigated in a multistimulus matching task the size of the discrepancy between response times for “same” and response times for “different” judgments. Frequently, results have shown that “same” judgments are faster than “different” judgments. Krueger (1984) found inversion in the speed advantage when stimuli were presented simultaneously and concluded that a self-termination factor would explain this result. In the experiment reported here, the subject had to exhaustively scan the whole set of items in the stimulus string. The analysis shows no significant interaction of presentation and response type; that is, the advantage for same stimuli is not reduced for simultane…

Speech recognitionStatisticsResponse typeSame differentGeneral ChemistryStimulus (physiology)behavioral disciplines and activitieshumanitiesCatalysisMathematicsBulletin of the Psychonomic Society
researchProduct

Adaptive Vocabulary Learning Environment for Late Talkers

2016

The main aim of this research is to provide children who have an early language delay with an adaptive way to train their vocabulary taking into account individuality of the learner. The suggested system is a mobile game-based learning environment which provides simple tasks where the learner chooses a picture that corresponds to a played back sound from multiple pictures presented on the screen. Our basic assumption is that the more similar the concepts (in our case, words) are, the harder the recognition task is. The system chooses the pictures to be presented on the screen by calculating the distances between the concepts in different dimensions. The distances are considered to consist o…

VocabularySLIComputer scienceProcess (engineering)media_common.quotation_subjectSpeech recognitioncomputer.software_genre050105 experimental psychologyTask (project management)self adaptive learning03 medical and health sciences0302 clinical medicineSimple (abstract algebra)Factor (programming language)Similarity (psychology)ta5160501 psychology and cognitive scienceslate talkersmedia_commoncomputer.programming_languageta113business.industryLearning environment05 social sciencesadaptive learningvocabulary learninggame-based learningFacilitationArtificial intelligencebusinesscomputer030217 neurology & neurosurgeryNatural language processingProceedings of the 8th International Conference on Computer Supported Education
researchProduct

Perceptual Interactions in Complex Odor Mixtures

2014

The perception of everyday odors relies on elemental or configural processing of complex mixtures of odorants. Theoretically, the configural processing of a mixture could lead to the perception of a single specific odor for the mixture; however, such a type of perception has hardly been proven in human studies. Here, we report the results of a sorting task demonstrating that a six-component mixture carries an odor clearly distinct from the odors of its components. These results suggest a blending effect of individual components’ odors in mixtures containing more than three odorants.

genetic structuresOdorHuman studiesComputer sciencemusculoskeletal neural and ocular physiologySpeech recognitionPerceptionmedia_common.quotation_subjectSortingbehavioral disciplines and activitiespsychological phenomena and processesTask (project management)media_common
researchProduct