keyword
MENU ▼
Read by QxMD icon Read
search

Speech acoustics

keyword
https://www.readbyqxmd.com/read/28238657/frogs-exploit-statistical-regularities-in-noisy-acoustic-scenes-to-solve-cocktail-party-like-problems
#1
Norman Lee, Jessica L Ward, Alejandro Vélez, Christophe Micheyl, Mark A Bee
Noise is a ubiquitous source of errors in all forms of communication [1]. Noise-induced errors in speech communication, for example, make it difficult for humans to converse in noisy social settings, a challenge aptly named the "cocktail party problem" [2]. Many nonhuman animals also communicate acoustically in noisy social groups and thus face biologically analogous problems [3]. However, we know little about how the perceptual systems of receivers are evolutionarily adapted to avoid the costs of noise-induced errors in communication...
February 17, 2017: Current Biology: CB
https://www.readbyqxmd.com/read/28237358/audio-visual-interaction-in-visual-motion-detection-synchrony-versus-asynchrony
#2
Stephanie Rosemann, Inga-Maria Wefel, Volkan Elis, Manfred Fahle
OBJECTIVE: Detection and identification of moving targets is of paramount importance in everyday life, even if it is not widely tested in optometric practice, mostly for technical reasons. There are clear indications in the literature that in perception of moving targets, vision and hearing interact, for example in noisy surrounds and in understanding speech. The main aim of visual perception, the ability that optometry aims to optimize, is the identification of objects, from everyday objects to letters, but also the spatial orientation of subjects in natural surrounds...
February 23, 2017: Journal of Optometry
https://www.readbyqxmd.com/read/28227737/development-of-bone-conduction-mobile-phones-assessment-of-hearing-mechanisms-by-measuring-psychological-characteristics-and-acoustical-properties-in-the-outer-ear-canal
#3
Seiji Nakagawa, Takuya Hotehama, Kazuhito Ito, Tomohiro Inagaki, Seiji Nakagawa, Takuya Hotehama, Kazuhito Ito, Tomohiro Inagaki, Seiji Nakagawa, Tomohiro Inagaki, Kazuhito Ito, Takuya Hotehama
We have been developing novel mobile phones using bone conduction, with flat-panel loudspeakers that convey speech sound by vibrating the pinna. In bone conduction via the pinna, i.e., pinna conduction, it is thought that speech sounds are conveyed via both air- and bone-conduction pathways. To obtain useful information for further development of bone-conduction mobile phones, peripheral mechanisms of the pinna conduction need to be clarified. In this study, hearing thresholds, sound field in the outer ear canals, and vibrations of the inner wall of the outer ear canals were measured while normal-hearing participants used pinna-conduction mobile phones...
August 2016: Conference Proceedings: Annual International Conference of the IEEE Engineering in Medicine and Biology Society
https://www.readbyqxmd.com/read/28227312/improving-quality-and-intelligibility-of-speech-using-single-microphone-for-the-broadband-fmri-noise-at-low-snr
#4
Chetan Vahanesa, Chandan K A Reddy, Issa M S Panahi, Chetan Vahanesa, Chandan K A Reddy, Issa M S Panahi, Chetan Vahanesa, Issa M S Panahi, Chandan K A Reddy
Functional Magnetic Resonance Imaging (fMRI) is used in many diagnostic procedures for neurological related disorders. Strong broadband acoustic noise generated during fMRI scan interferes with the speech communication between the physician and the patient. In this paper, we propose a single microphone Speech Enhancement (SE) technique which is based on the supervised machine learning technique and a statistical model based SE technique. The proposed algorithm is robust and computationally efficient and has capability to run in real-time...
August 2016: Conference Proceedings: Annual International Conference of the IEEE Engineering in Medicine and Biology Society
https://www.readbyqxmd.com/read/28227200/obstructive-sleep-apnea-severity-estimation-fusion-of-speech-based-systems
#5
D Ben Or, E Dafna, A Tarasiuk, Y Zigel, D Ben Or, E Dafna, A Tarasiuk, Y Zigel, A Tarasiuk, Y Zigel, E Dafna
Obstructive sleep apnea (OSA) is a common sleep-related breathing disorder. Previous studies associated OSA with anatomical abnormalities of the upper respiratory tract that may be reflected in the acoustic characteristics of speech. We tested the hypothesis that the speech signal carries essential information that can assist in early assessment of OSA severity by estimating apnea-hypopnea index (AHI). 198 men referred to routine polysomnography (PSG) were recorded shortly prior to sleep onset while reading a one-minute speech protocol...
August 2016: Conference Proceedings: Annual International Conference of the IEEE Engineering in Medicine and Biology Society
https://www.readbyqxmd.com/read/28226816/neural-representation-of-different-mandarin-tones-in-the-inferior-colliculus-of-the-guinea-pig
#6
Fei Peng, Nan Xia, Xing Wang, Xiao L Zheng, Yi Zhou, Xing Y Fan, Wen S Hou, Fei Peng, Nan Xia, Xing Wang, Xiao L Zheng, Yi Zhou, Xing Y Fan, Wen S Hou, Xing Y Fan, Nan Xia, Wen S Hou, Xing Wang, Yi Zhou, Fei Peng, Xiao L Zheng
Mandarin speech has four different tones and the coding mechanism underlying tone identification still remain unclear. Here in the inferior colliculus (IC) of anesthetized guinea pigs, we recorded single neuron activities to one word with four tones using tungsten electrode. Peri-stimulus time histograms (PSTHs) and inter-spike-interval (ISI) were used to evaluate the neural response. The results showed that PSTHs grouped into frequency band reflected the spectrotemporal patterns of different tones; average population PSTHs matched envelops of different tones; and the peaks of histogram of ISIs in three time segments exhibited a displacement which reflected the profile of fundamental frequency (F0)...
August 2016: Conference Proceedings: Annual International Conference of the IEEE Engineering in Medicine and Biology Society
https://www.readbyqxmd.com/read/28226442/adaptive-attention-driven-speech-enhancement-for-eeg-informed-hearing-prostheses
#7
Neetha Das, Simon Van Eyndhoven, Tom Francart, Alexander Bertrand, Neetha Das, Simon Van Eyndhoven, Tom Francart, Alexander Bertrand, Tom Francart, Neetha Das, Simon Van Eyndhoven, Alexander Bertrand
State-of-the-art hearing prostheses are equipped with acoustic noise reduction algorithms to improve speech intelligibility. Currently, one of the major challenges is to perform acoustic noise reduction in so-called cocktail party scenarios with multiple speakers, in particular because it is difficult-if not impossible-for the algorithm to determine which are the target speaker(s) that should be enhanced, and which speaker(s) should be treated as interfering sources. Recently, it has been shown that electroencephalography (EEG) can be used to perform auditory attention detection, i...
August 2016: Conference Proceedings: Annual International Conference of the IEEE Engineering in Medicine and Biology Society
https://www.readbyqxmd.com/read/28225849/effect-of-performance-time-of-the-high-pitched-blowing-vocal-exercise-in-the-voice-of-women
#8
Fabíola Santos Moreira, Ana Cristina Côrtes Gama
Purpose: To analyze the results of the runtimes of one, three, five, and seven minutes of the high-pitched blowing vocal exercise in women without voice complaints and with dysphonia and vocal nodules. Methods: This is an experimental study with a consecutive and convenience sample of 60 women divided into two groups: 30 participants with dysphonia caused by vocal fold nodules (study group - SG) and 30 participants without vocal complaints (control group - CG). All participants performed the high-pitched blowing vocal exercise for one, three, five, and seven minutes...
February 16, 2017: CoDAS
https://www.readbyqxmd.com/read/28225796/what-can-we-learn-about-beat-perception-by-comparing-brain-signals-and-stimulus-envelopes
#9
Molly J Henry, Björn Herrmann, Jessica A Grahn
Entrainment of neural oscillations on multiple time scales is important for the perception of speech. Musical rhythms, and in particular the perception of a regular beat in musical rhythms, is also likely to rely on entrainment of neural oscillations. One recently proposed approach to studying beat perception in the context of neural entrainment and resonance (the "frequency-tagging" approach) has received an enthusiastic response from the scientific community. A specific version of the approach involves comparing frequency-domain representations of acoustic rhythm stimuli to the frequency-domain representations of neural responses to those rhythms (measured by electroencephalography, EEG)...
2017: PloS One
https://www.readbyqxmd.com/read/28225734/auditory-performance-and-electrical-stimulation-measures-in-cochlear-implant-recipients-with-auditory-neuropathy-compared-with-severe-to-profound-sensorineural-hearing-loss
#10
Joseph Attias, Tally Greenstein, Miriam Peled, David Ulanovski, Jay Wohlgelernter, Eyal Raveh
OBJECTIVES: The aim of the study was to compare auditory and speech outcomes and electrical parameters on average 8 years after cochlear implantation between children with isolated auditory neuropathy (AN) and children with sensorineural hearing loss (SNHL). DESIGN: The study was conducted at a tertiary, university-affiliated pediatric medical center. The cohort included 16 patients with isolated AN with current age of 5 to 12.2 years who had been using a cochlear implant for at least 3...
March 2017: Ear and Hearing
https://www.readbyqxmd.com/read/28222332/acquisition-of-vowel-articulation-in-childhood-investigated-by-acoustic-to-articulatory-inversion
#11
Hiroki Oohashi, Hama Watanabe, Gentaro Taga
While the acoustical features of speech sounds in children have been extensively studied, limited information is available as to their articulation during speech production. Instead of directly measuring articulatory movements, this study used an acoustic-to-articulatory inversion model with scalable vocal tract size to estimate developmental changes in articulatory state during vowel production. Using a pseudo-inverse Jacobian matrix of a model mapping seven articulatory parameters to acoustic ones, the formant frequencies of each vowel produced by three Japanese children over time at ages between 6 and 60 months were transformed into articulatory parameters...
February 18, 2017: Infant Behavior & Development
https://www.readbyqxmd.com/read/28213999/application-of-the-acoustic-voice-quality-index-for-objective-measurement-of-dysphonia-severity
#12
Faustino Núñez-Batalla, Estefanía Díaz-Fresno, Andrea Álvarez-Fernández, Gabriela Muñoz Cordero, José Luis Llorente Pendás
INTRODUCTION: Over the past several decades, many acoustic parameters have been studied as sensitive to and to measure dysphonia. However, current acoustic measures might not be sensitive measures of perceived voice quality. A meta-analysis which evaluated the relationship between perceived overall voice quality and several acoustic-phonetic correlates, identified measures that do not rely on the extraction of the fundamental period, such the measures derived from the cepstrum, and that can be used in sustained vowel as well as continuous speech samples...
February 14, 2017: Acta Otorrinolaringológica Española
https://www.readbyqxmd.com/read/28212857/temporal-modulations-in-speech-and-music
#13
REVIEW
Nai Ding, Aniruddh D Patel, Lin Chen, Henry Butler, Cheng Luo, David Poeppel
Speech and music have structured rhythms. Here we discuss a major acoustic correlate of spoken and musical rhythms, the slow (0.25-32Hz) temporal modulations in sound intensity and compare the modulation properties of speech and music. We analyze these modulations using over 25h of speech and over 39h of recordings of Western music. We show that the speech modulation spectrum is highly consistent across 9 languages (including languages with typologically different rhythmic characteristics). A different, but similarly consistent modulation spectrum is observed for music, including classical music played by single instruments of different types, symphonic, jazz, and rock...
February 14, 2017: Neuroscience and Biobehavioral Reviews
https://www.readbyqxmd.com/read/28208129/stability-and-variability-in-slovak-prosodic-boundaries
#14
Štefan Benuš, Juraj Šimko
BACKGROUND/AIM: Encoding intended meanings in the type and strength of prosodic boundaries and strategies for communicating these meanings in ambient noise use similar prosodic cues. We analyze how increasing the level of ambient noise affects the realization of Slovak prosodic boundaries. METHODS: Five native speakers of Slovak read sentences, manipulating the boundary type (weak, rise, fall) and the location of pre-boundary pitch accent. Ambient noise of several levels was administered via headphones...
2016: Phonetica
https://www.readbyqxmd.com/read/28207576/relationship-between-peripheral-and-psychophysical-measures-of-amplitude-modulation-detection-in-cochlear-implant-users
#15
Viral D Tejani, Paul J Abbas, Carolyn J Brown
OBJECTIVE: This study investigates the relationship between electrophysiological and psychophysical measures of amplitude modulation (AM) detection. Prior studies have reported both measures of AM detection recorded separately from cochlear implant (CI) users and acutely deafened animals, but no study has made both measures in the same CI users. Animal studies suggest a progressive loss of high-frequency encoding as one ascends the auditory pathway from the auditory nerve to the cortex...
February 15, 2017: Ear and Hearing
https://www.readbyqxmd.com/read/28203463/the-relative-weight-of-temporal-envelope-cues-in-different-frequency-regions-for-mandarin-sentence-recognition
#16
Yang Guo, Yuanyuan Sun, Yanmei Feng, Yujun Zhang, Shankai Yin
Acoustic temporal envelope (E) cues containing speech information are distributed across the frequency spectrum. To investigate the relative weight of E cues in different frequency regions for Mandarin sentence recognition, E information was extracted from 30 contiguous bands across the range of 80-7,562 Hz using Hilbert decomposition and then allocated to five frequency regions. Recognition scores were obtained with acoustic E cues from 1 or 2 random regions from 40 normal-hearing listeners. While the recognition scores ranged from 8...
2017: Neural Plasticity
https://www.readbyqxmd.com/read/28203263/perception-of-emotional-valence-in-horse-whinnies
#17
Elodie F Briefer, Roi Mandel, Anne-Laure Maigrot, Sabrina Briefer Freymond, Iris Bachmann, Edna Hillmann
BACKGROUND: Non-human animals often produce different types of vocalisations in negative and positive contexts (i.e. different valence), similar to humans, in which crying is associated with negative emotions and laughter is associated with positive ones. However, some types of vocalisations (e.g. contact calls, human speech) can be produced in both negative and positive contexts, and changes in valence are only accompanied by slight structural differences. Although such acoustically graded signals associated with opposite valence have been highlighted in some species, it is not known if conspecifics discriminate them, and if contagion of emotional valence occurs as a result...
2017: Frontiers in Zoology
https://www.readbyqxmd.com/read/28203214/searching-high-and-low-prosodic-breaks-disambiguate-relative-clauses
#18
Lauren A Fromont, Salvador Soto-Faraco, Emmanuel Biau
During natural speech perception, listeners rely on a wide range of cues to support comprehension, from semantic context to prosodic information. There is a general consensus that prosody plays a role in syntactic parsing, but most studies focusing on ambiguous relative clauses (RC) show that prosodic cues, alone, are insufficient to reverse the preferred interpretation of sentence. These findings suggest that universally preferred structures (e.g., Late Closure principle) matter far more than prosodic cues in such cases...
2017: Frontiers in Psychology
https://www.readbyqxmd.com/read/28198405/an-acoustic-key-to-eight-languages-dialects-factor-analyses-of-critical-band-filtered-speech
#19
Kazuo Ueda, Yoshitaka Nakajima
The peripheral auditory system functions like a frequency analyser, often modelled as a bank of non-overlapping band-pass filters called critical bands; 20 bands are necessary for simulating frequency resolution of the ear within an ordinary frequency range of speech (up to 7,000 Hz). A far smaller number of filters seemed sufficient, however, to re-synthesise intelligible speech sentences with power fluctuations of the speech signals passing through them; nevertheless, the number and frequency ranges of the frequency bands for efficient speech communication are yet unknown...
February 15, 2017: Scientific Reports
https://www.readbyqxmd.com/read/28194222/intelligibility-evaluation-of-pathological-speech-through-multigranularity-feature-extraction-and-optimization
#20
Chunying Fang, Haifeng Li, Lin Ma, Mancai Zhang
Pathological speech usually refers to speech distortion resulting from illness or other biological insults. The assessment of pathological speech plays an important role in assisting the experts, while automatic evaluation of speech intelligibility is difficult because it is usually nonstationary and mutational. In this paper, we carry out an independent innovation of feature extraction and reduction, and we describe a multigranularity combined feature scheme which is optimized by the hierarchical visual method...
2017: Computational and Mathematical Methods in Medicine
keyword
keyword
59607
1
2
Fetch more papers »
Fetching more papers... Fetching...
Read by QxMD. Sign in or create an account to discover new knowledge that matter to you.
Remove bar
Read by QxMD icon Read
×

Search Tips

Use Boolean operators: AND/OR

diabetic AND foot
diabetes OR diabetic

Exclude a word using the 'minus' sign

Virchow -triad

Use Parentheses

water AND (cup OR glass)

Add an asterisk (*) at end of a word to include word stems

Neuro* will search for Neurology, Neuroscientist, Neurological, and so on

Use quotes to search for an exact phrase

"primary prevention of cancer"
(heart or cardiac or cardio*) AND arrest -"American Heart Association"