Search results for "speech"
showing 10 items of 1281 documents
UBFC-Phys: A Multimodal Database For Psychophysiological Studies of Social Stress
2021
As humans, we experience social stress in countless everyday-life situations. Giving a speech in front of an audience, passing a job interview, and similar experiences all lead us to go through stress states that impact both our psychological and physiological states. Therefore, studying the link between stress and physiological responses had become a critical societal issue, and recently, research in this field has grown in popularity. However, publicly available datasets have limitations. In this article, we propose a new dataset, UBFC-Phys, collected with and without contact from participants living social stress situations. A wristband was used to measure contact blood volume pulse (BVP…
Cuando el aprendizaje basado en la práctica profesional no provoca un interés laboral
2021
The link between the academic and the professional has become an essential element in promoting the educational quality of today’s University. This article shows an experience of teaching innovation in higher education based on the use of professional technologies brought to the educational field. Specifically, the computer application and methodology used professionally by CEAL-Spanish Association of Socio-Labour Auditors in the development of their work, and which has allowed the professionalization of teaching and bringing it closer to the reality of the professional collective. A questionnaire was used to find out the students’ perception of the methodological proposal and its potential…
Automatic screening of cardiac disorders using wavelet analysis of heart sound
2017
Body auscultation is a dominant method for physical examination of human heart using conventional stethoscope. This clinical method is non invasive and efficient but it requires a medical expert to interpret the heart sound for assessment of cardiac disorders. This paper presents analysis of heart sounds in wavelet domain for automated screening of cardiac disorders. Heart sound signal is transformed in wavelet domain to find out discrimination between heart sounds recorded from healthy and anomalous patients. Discriminatory features extracted from wavelet coefficients of heart sound are subjected to machine learning for screening of cardiac disorders automatically. The proposed method for …
On the effectiveness of vocal imitations and verbal descriptions of sounds
2014
cote interne IRCAM: Lemaitre14b; None / None; International audience; Describing unidentified sounds with words is a frustrating task and vocally imitating them is often a convenient way to address the issue. This article reports on a study that compared the effectiveness of vocal imitations and verbalizations to communicate different referent sounds. The stimuli included mechanical and synthesized sounds and were selected on the basis of participants' confidence in identifying the cause of the sounds, ranging from easy-to-identify to unidentifiable sounds. The study used a selection of vocal imitations and verbalizations deemed adequate descriptions of the referent sounds. These descriptio…
Advantages of binaural hearing provided through bimodal stimulation via a cochlear implant and a conventional hearing aid: a 6-month comparative stud…
2005
Our study data demonstrate the additional benefit derived from continued use of a contralateral hearing aid (HA) post-cochlear implantation for speech recognition ability in quiet and in noise. Postoperative bimodal stimulation is recommended for all subjects who show some speech recognition ability in the contralateral ear as it may offer binaural listening advantages in various listening situations encountered in everyday life.To assess the benefits derived from bimodal stimulation for experienced HA users implanted with a cochlear implant (CI) (score=20% in disyllabic test). The correlation between pre- and postoperative performance on speech perception measures was examined to determine…
Source localization of event-related potentials to pitch change mapped onto age-appropriate MRIs at 6 months-of-age
2010
Auditory event-related potentials (ERPs) have been used to understand how the brain processes auditory input, and to track developmental change in sensory systems. Localizing ERP generators can provide invaluable insights into how and where auditory information is processed. However, age-appropriate infant brain templates have not been available to aid such developmental mapping. In this study, auditory change detection responses of brain ERPs were examined in 6-month-old infants using discrete and distributed source localization methods mapped onto age-appropriate magnetic resonance images. Infants received a passive oddball paradigm using fast-rate non-linguistic auditory stimuli (tone do…
Cumulative-Sum-Based Localization of Sound Events in Low-Cost Wireless Acoustic Sensor Networks
2014
Wireless acoustic sensor networks (WASNs) are known for their potential applications in multiple areas, such as audio-based surveillance, binaural hearing aids or advanced acoustic monitoring. The knowledge of the spatial position of a source of interest is usually a requirement for many of these applications. Therefore, source localization is an important problem to be addressed in WASNs. Unfortunately, most localization algorithms need costly signal processing stages that prevent them from being implemented in low-cost sensor networks, requiring additional modules for signal acquisition and processing. This paper presents a low-complexity method for acoustic event detection and localizati…
An integrated dialect analysis tool using phonetics and acoustics
2019
This study aimed to verify a computational phonetic and acoustic analysis tool created in the MATLAB environment. A dataset was obtained containing 3 broad American dialects (Northern, Western and New England) from the TIMIT database using words that also appeared in the Swadesh list. Each dialect consisted of 20 speakers uttering 10 sentences. Verification using phonetic comparisons between dialects was made by calculating the Levenshtein distance in Gabmap and the proposed software tool. Agreement between the linguistic distances using each analysis method was found. Each tool showed increasing linguistic distance as a function of increasing geographic distance, in a similar shape to Segu…
Usage of HMM-Based Speech Recognition Methods for Automated Determination of a Similarity Level Between Languages
2019
The problem of automated determination of language similarity (or even defining of a distance on the space of languages) could be solved in different ways – working with phonetic transcriptions, with speech recordings or both of them. For the recordings, we propose and test a HMM-based one: in the first part of our article we successfully try language detection, afterwards we are trying to calculate distances between HMM-based models, using different metrics and divergences. The Kullback-Leibler divergence is the only one we got good results with – it means that the calculated distances between languages correspond to analytical understanding of similarity between them. Even if it does not …
Using privacy-transformed speech in the automatic speech recognition acoustic model training
2020
Automatic Speech Recognition (ASR) requires huge amounts of real user speech data to reach state-of-the-art performance. However, speech data conveys sensitive speaker attributes like identity that can be inferred and exploited for malicious purposes. Therefore, there is an interest in the collection of anonymized speech data that is processed by some voice conversion method. In this paper, we evaluate one of the voice conversion methods on Latvian speech data and also investigate if privacy-transformed data can be used to improve ASR acoustic models. Results show the effectiveness of voice conversion against state-of-the-art speaker verification models on Latvian speech and the effectivene…