Humans are inherently emotional and to better understand them, robots need to recognize emotions from human speech. Due to the complexity of auditory perception models, however, emotion recognition is a challenging task. In a new study, researchers from Japan and China design a novel feature that captures temporal and contextual information and extracts temporal variation of emotion using a parallel neural network architecture, opening doors to future applications in more complex speech analysis tasks.
People's ability to perceive speech sounds has been deeply studied, specially during someone's first year of life, but what happens during the first hours after birth? Are babies born with innate abilities to perceive speech sounds, or do neural encoding processes need to age for some time?
Smart assistant devices often need to perform speech translation, which does not always produce the desired voice identity due to drawbacks in the conventional voice conversion (VC) model. In a new study, researchers from Japan Advanced Institute of Science and Technology designed a VC model that mimics and controls speaker voice identity during speech translation using two deep learning based training frameworks, opening doors to voice modification, voice restoring, and voice cloning applications.
Imaging techniques enable decisive step toward development of novel hearing prostheses
What The Study Did: This review of 125 U.S.-based clinical trials that investigated the management of hearing loss assessed representation in the trials by race/ethnicity and sex.
Plenty of people struggle to make sense of a multitude of converging voices in a crowded room. Commonly known as the "cocktail party effect," people with hearing loss find it's especially difficult to understand speech in a noisy environment. New research suggests that, for some listeners, this may have less to do with actually discerning sounds. Instead, it may be a processing problem in which two ears blend different sounds together - a condition known as binaural pitch fusion.
What The Study Did: The association between hearing loss and level of physical activity among U.S. adults ages 60 to 69 was analyzed in this study.
The brain's auditory system tracks the speed and location of moving sounds in the same way the visual system tracks moving objects. The study recently published in eNeuro lays the groundwork for more detailed research on how humans hear in dynamic environments.
Researchers from the HSE Center for Language and Brain have, for the first time, described the language abilities of Russian children with Autism Spectrum Disorder (ASD) at all linguistic levels (e.g., phonology, lexicon, morphosyntax, and discourse), using a language test that takes into account the psycholinguistic variables most relevant for Russians. The study was published in Journal of Autism and Developmental Disorders.
According to a linguistic survey report, people often confuse the pronunciation of /hi/ with that of /si/ in the dialect of Tokyo and the Tohoku region of Japan. A team of researchers at Toyohashi University of Technology and the National Institute for Japanese Language and Linguistics (NINJAL) found that the confusion is resulted from the articulation of the tongue varying in the transverse direction while the tongue tip is positioned at the same place of articulation.