期刊文献+
共找到16篇文章
< 1 >
每页显示 20 50 100
Speech perception in noise:Masking and unmasking 被引量:2
1
作者 Xianhui Wang Li Xu 《Journal of Otology》 CSCD 2021年第2期109-119,共11页
Speech perception is essential for daily communication.Background noise or concurrent talkers,on the other hand,can make it challenging for listeners to track the target speech(i.e.,cocktail party problem).The present... Speech perception is essential for daily communication.Background noise or concurrent talkers,on the other hand,can make it challenging for listeners to track the target speech(i.e.,cocktail party problem).The present study reviews and compares existing findings on speech perception and unmasking in cocktail party listening environments in English and Mandarin Chinese.The review starts with an introduction section followed by related concepts of auditory masking.The next two sections review factors that release speech perception from masking in English and Mandarin Chinese,respectively.The last section presents an overall summary of the findings with comparisons between the two languages.Future research directions with respect to the difference in literature on the reviewed topic between the two languages are also discussed. 展开更多
关键词 speech perception Auditory masking speech unmasking Cocktail party problems Mandarin Chinese
下载PDF
Subclinical hearing loss associated with aging
2
作者 Cristian Aedo-Sanchez Jose Oliveros +3 位作者 Constanza Aranguiz Camila Munoz Claudia Lazo-Maturana Enzo Aguilar-Vidal 《Journal of Otology》 CAS CSCD 2023年第3期111-117,共7页
Objective:Contribute to clarifying the existence of subclinical hearing deficits associated with aging.Design:In this work,we study and compare the auditory perceptual and electrophysiological performance of normal-he... Objective:Contribute to clarifying the existence of subclinical hearing deficits associated with aging.Design:In this work,we study and compare the auditory perceptual and electrophysiological performance of normal-hearing young and adult subjects(tonal audiometry,high-frequency tone threshold,a triplet of digits in noise,and click-evoked auditory brainstem response).Study sample:45 normal hearing volunteers were evaluated and divided into two groups according to age.27 subjects were included in the“young group”(mean 22.1 years),and 18 subjects(mean 42.22 years)were included in the“adult group.”Results:In the perceptual tests,the adult group presented significantly worse tonal thresholds in the high frequencies(12 and 16 kHz)and worse performance in the digit triplet tests in noise.In the electrophysiological test using the auditory brainstem response technique,the adult group presented significantly lower I and V wave amplitudes and higher V wave latencies at the supra-threshold level.At the threshold level,we observed a significantly higher latency in wave V in the adult group.In addition,in the partial correlation analysis,controlling for the hearing level,we observed a relationship(negative)between age and speech in noise performance and high-frequency thresholds.No significant association was observed between age and the auditory brainstem response.Conclusion:The results are compatible with subclinical hearing loss associated with aging. 展开更多
关键词 Age-related hearing loss Cochlear synaptopathy Auditory electrophysiology speech perception AGING
下载PDF
Interaction between auditory and motor systems in speech perception 被引量:2
3
作者 Zhe-Meng Wu Ming-Li Chen +1 位作者 Xi-Hong Wu Liang Li 《Neuroscience Bulletin》 SCIE CAS CSCD 2014年第3期490-496,共7页
Based on the Motor Theory of speech perception, the interaction between the auditory and motor systems plays an essential role in speech perception. Since the Motor Theory was proposed, it has received remarkable atte... Based on the Motor Theory of speech perception, the interaction between the auditory and motor systems plays an essential role in speech perception. Since the Motor Theory was proposed, it has received remarkable attention in the field. However, each of the three hypotheses of the theory still needs further verification. In this review, we focus on how the auditory-motor anatomical and functional associations play a role in speech perception and discuss why previous studies could not reach an agreement and particularly whether the motor system involvement in speech perception is task-load dependent. Finally, we suggest that the function of the auditory-motor link is particularly useful for speech perception under adverse listening conditions and the further revised Motor Theory is a potential solution to the "cocktail-party" problem. 展开更多
关键词 auditory-motor interaction Motor Theory of speech perception motor cortex "cocktail-party" problem.
原文传递
Communication Between Speech Production and Perception Within the Brain--Observation and Simulation 被引量:1
4
作者 党建武 赤木正人 本多清志 《Journal of Computer Science & Technology》 SCIE EI CSCD 2006年第1期95-105,共11页
Realization of an intelligent human-machine interface requires us to investigate human mechanisms and learn from them. This study focuses on communication between speech production and perception within human brain an... Realization of an intelligent human-machine interface requires us to investigate human mechanisms and learn from them. This study focuses on communication between speech production and perception within human brain and realizing it in an artificial system. A physiological research study based on electromyographic signals (Honda, 1996) suggested that speech communication in human brain might be based on a topological mapping between speech production and perception, according to an analogous topology between motor and sensory representations. Following this hypothesis, this study first investigated the topologies of the vowel system across the motor, kinematic, and acoustic spaces by means of a model simulation, and then examined the linkage between vowel production and perception in terms of a transformed auditory feedback (TAF) experiment. The model simulation indicated that there exists an invariant mapping from muscle activations (motor space) to articulations (kinematic space) via a coordinate consisting of force-dependent equilibrium positions, and the mapping from the motor space to kinematic space is unique. The motor-kinematic-acoustic deduction in the model simulation showed that the topologies were compatible from one space to another. In the TAF experiment, vowel production exhibited a compensatory response for a perturbation in the feedback sound. This implied that vowel production is controlled in reference to perception monitoring. 展开更多
关键词 speech communication human mechanism speech production speech perception
原文传递
Syntax Development in Language Education of Children Who Are Deaf or Hard-of-Hearing by Usage of Modem Technology 被引量:1
5
作者 Husnija Hasanbegovic Esad Mahmutovic 《Sino-US English Teaching》 2014年第5期307-315,共9页
The aim of the study was to determine the development of syntax in language development of children who are deaf or hard-of-hearing, who are taught to new dynamic linguistic features with the help of computers. The sa... The aim of the study was to determine the development of syntax in language development of children who are deaf or hard-of-hearing, who are taught to new dynamic linguistic features with the help of computers. The sample consisted of 70 children who are deaf or hard-of-hearing, aged 7-17 years. To assess language development were applied following variables: total number of words used, the total number of different words used, the correct and incorrect statements (sentences) of the respondents. We calculated the basic statistical parameters on which it was found that the experimental program computer teaching children who are deaf or hard-of-hearing gave better results in the development of syntax. Also, canonical discriminate analysis revealed a statistically significant difference in the applied variables between the control and experimental groups the level of statistical significance ofp = 0.000. The results showed a significant improvement of the experimental group and that dynamic computer programming activities, which were challenged participants of the experimental group, contribute to a better linguistic competence of children who are deaf or hard-of-hearing. 展开更多
关键词 SYNTAX speech and language vocabulary speech perception child who is deaf
下载PDF
Speech endpoint detection in low-SNRs environment based on perception spectrogram structure boundary parameter 被引量:9
6
作者 WU Di ZHAO Heming +4 位作者 HUANG Chengwei XIAO Zhongzhe ZHANG Xiaojun XU Yishen TAO Zhi 《Chinese Journal of Acoustics》 2014年第4期428-440,共13页
The Perception Spectrogram Structure Boundary(PSSB)parameter is proposed for speech endpoint detection as a preprocess of speech or speaker recognition.At first a hearing perception speech enhancement is carried out... The Perception Spectrogram Structure Boundary(PSSB)parameter is proposed for speech endpoint detection as a preprocess of speech or speaker recognition.At first a hearing perception speech enhancement is carried out.Then the two-dimensional enhancement is performed upon the sound spectrogram according to the difference between the determinacy distribution characteristic of speech and the random distribution characteristic of noise.Finally a decision for endpoint was made by the PSSB parameter.Experimental results show that,in a low SNR environment from-10 dB to 10 dB,the algorithm proposed in this paper may achieve higher accuracy than the extant endpoint detection algorithms.The detection accuracy of 75.2%can be reached even in the extremely low SNR at-10 dB.Therefore it is suitable for speech endpoint detection in low-SNRs environment. 展开更多
关键词 speech endpoint detection in low-SNRs environment based on perception spectrogram structure boundary parameter
原文传递
Grading the Severity of Mispronunciations in CAPT Based on Statistical Analysis and Computational Speech Perception
7
作者 贾珈 梁伟俭 +4 位作者 吴育昊 张秀龙 王昊 蔡莲红 蒙美玲 《Journal of Computer Science & Technology》 SCIE EI CSCD 2014年第5期751-761,共11页
Computer-aided pronunciation training(CAPT) technologies enable the use of automatic speech recognition to detect mispronunciations in second language(L2) learners' speech. In order to further facilitate learning... Computer-aided pronunciation training(CAPT) technologies enable the use of automatic speech recognition to detect mispronunciations in second language(L2) learners' speech. In order to further facilitate learning, we aim to develop a principle-based method for generating a gradation of the severity of mispronunciations. This paper presents an approach towards gradation that is motivated by auditory perception. We have developed a computational method for generating a perceptual distance(PD) between two spoken phonemes. This is used to compute the auditory confusion of native language(L1). PD is found to correlate well with the mispronunciations detected in CAPT system for Chinese learners of English,i.e., L1 being Chinese(Mandarin and Cantonese) and L2 being US English. The results show that auditory confusion is indicative of pronunciation confusions in L2 learning. PD can also be used to help us grade the severity of errors(i.e.,mispronunciations that confuse more distant phonemes are more severe) and accordingly prioritize the order of corrective feedback generated for the learners. 展开更多
关键词 second language learning computer-aided pronunciation training mispronunciation computational speech perception
原文传递
Acceptable noise level as a deciding factor for prescribing hearing aids for older adults with cochlear hearing loss-A scoping review
8
作者 Hemanth Narayan Shetty Swathi Subbanna 《Journal of Otology》 CSCD 2015年第3期93-98,共6页
Older adults often find it difficult to perceive speech, especially in noisy conditions. Though hearing aid is one of the rehabilitative devices available to older adults to alleviate hearing loss, some of them may ex... Older adults often find it difficult to perceive speech, especially in noisy conditions. Though hearing aid is one of the rehabilitative devices available to older adults to alleviate hearing loss, some of them may experience annoyance through hearing aid and hence reject it, may be due to circuitry noise and/or background noise. Acceptable noise level is a direct behavioural measure to estimate the extent of how much a person is able to put up with noise while simultaneously listening to speech. Acceptable noise level is a central auditory measure and it is not influenced by age, gender, presentation level or speaker. Using this measure, we can quantify the annoyance level experienced by an individual. This in-formation is of utmost importance and caution should be paid before setting the parameters in hearing aid, especially for those who are unable to accept noise. In this review article, an attempt has been made to document how to optimize the hearing aid program by setting parameters such as noise reduction circuit, microphone sensitivity and gain. These adjustments of parameters might help to reduce rejection rate of hearing aids, especially in those individuals who are annoyed by background noise. Copyright ? 2015 The Authors. Production & hosting by Elsevier (Singapore) Pte Ltd On behalf of PLA General Hospital Department of Otolaryngology Head and Neck Surgery. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/). 展开更多
关键词 ANNOYANCE AMPLIFICATION speech perception in noise Older adults Hearing impairment
下载PDF
Improved hearing in noise using new signal processing algorithms with the Cochlear^(TM) Nucleus~ 6 sound processor
9
作者 Jan Gilden Kristen Lewis +1 位作者 Ginger Grant Jillian Crosson 《Journal of Otology》 CSCD 2015年第2期51-56,共6页
Objective: To demonstrate the performance benefit of the Automatic Scene Classifier (SCAN) algorithm available in the Nucleus 6 (CP900 series) sound processor over the default processing algorithms of the previou... Objective: To demonstrate the performance benefit of the Automatic Scene Classifier (SCAN) algorithm available in the Nucleus 6 (CP900 series) sound processor over the default processing algorithms of the previous generation Nucleus 5 (CP810) and Freedom HybridTM sound processors. Methods: Eighty-two cochlear implant recipients (40 Nucleus 5 processor users and 42 Freedom Hybrid processor users) listened to and repeated AzBio sentences in noise with their current processor and with the Nucleus 6 processor. Results: The SCAN algorithm when enabled yielded statistically significant non-inferior and superior performance when compared to the Nucleus 5 and Freedom Hybrid sound processors programmed with ASC + ADRO. Conclusion: The results of these studies demonstrate the superior performance and clinical utility of the SCAN algorithm in the Nucleus 6 processor over the Nucleus 5 and Freedom Hybrid processors. 展开更多
关键词 Cochlear implant speech perception Nucleus 6 Automatic scene classifier (SCAN) SmartSound iQ
下载PDF
Auditory neuropathy in a patient with hemochromatosis
10
作者 Gary Rance Donella Chisari 《Journal of Otology》 CSCD 2016年第4期-,共7页
Objective:To evaluate the auditory function of an individual with genetically confirmed hemochromatosis. Methods: A 57 year old male with mildly impaired sound detection thresholds underwent a range of behavioural, el... Objective:To evaluate the auditory function of an individual with genetically confirmed hemochromatosis. Methods: A 57 year old male with mildly impaired sound detection thresholds underwent a range of behavioural, electroacoustic and elec-trophysiologic assessments. These included the recording of otoacoustic emissions and auditory brainstem responses, measurement of monaural temporal resolution and evaluation of binaural speech processing. Findings for this patient were subsequently compared with those of 80 healthy controls with similar audiometric thresholds. Results: The patient showed the three cardinal features of auditory neuropathy, presenting with evidence of normal cochlear outer hair cell function, disrupted neural activity in the auditory nerve/brainstem and impaired temporal processing. His functional hearing ability (speech perception) was significantly affected and suggested a reduced capacity to use localization cues to segregate signals in the presence of back-ground noise. Conclusion:We present the first case of an individual with hemochromatosis and auditory neuropathy. The findings for this patient highlight the need for careful evaluation of auditory function in individuals with the disorder. 展开更多
关键词 HEMOCHROMATOSIS Auditory neuropathy Temporal processing speech perception
下载PDF
Towards human-like and transhuman perception in AI 2.0:a review 被引量:3
11
作者 Yong-hong TIAN Xi-lin CHEN +9 位作者 Hong-kai XIONG Hong-liang LI Li-rong DAI Jing CHEN Jun-liang XING Xi-hong WU Wei-min HU Yu HU Tie-jun HUANG Wen GAO 《Frontiers of Information Technology & Electronic Engineering》 SCIE EI CSCD 2017年第1期58-67,共10页
Perception is the interaction interface between an intelligent system and the real world. Without sophisticated and flexible perceptual capabilities, it is impossible to create advanced artificial intelligence (AI) ... Perception is the interaction interface between an intelligent system and the real world. Without sophisticated and flexible perceptual capabilities, it is impossible to create advanced artificial intelligence (AI) systems. For the next-generation AI, called 'AI 2.0', one of the most significant features will be that AI is empowered with intelligent perceptual capabilities, which can simulate human brain's mechanisms and are likely to surpass human brain in terms of performance. In this paper, we briefly review the state-of-the-art advances across different areas of perception, including visual perception, auditory perception, speech perception, and perceptual information processing and learning engines. On this basis, we envision several R&D trends in intelligent perception for the forthcoming era of AI 2.0, including: (1) human-like and transhuman active vision; (2) auditory perception and computation in an actual auditory setting; (3) speech perception and computation in a natural interaction setting; (4) autonomous learning of perceptual information; (5) large-scale perceptual information processing and learning platforms; and (6) urban omnidirectional intelligent perception and reasoning engines. We believe these research directions should be highlighted in the future plans for AI 2.0. 展开更多
关键词 Intelligent perception Active vision Auditory perception speech perception Autonomous learning
原文传递
Benefits of spatial hearing to speech recognition in young people with normal hearing
12
作者 SONG Peng-long LI Hui-jun WANG Ning-yu 《Chinese Medical Journal》 SCIE CAS CSCD 2011年第24期4269-4274,共6页
Background Many factors interfering with a listener attempting to grasp speech in noisy environments. The spatial hearing by which speech and noise can be spatially separated may play a crucial role in speech recognit... Background Many factors interfering with a listener attempting to grasp speech in noisy environments. The spatial hearing by which speech and noise can be spatially separated may play a crucial role in speech recognition in the presence of competing noise. This study aimed to assess whether, and to what degree, spatial hearing benefit speech recognition in young normal-hearing participants in both quiet and noisy environments. Methods Twenty-eight young participants were tested by Mandarin Hearing In Noise Test (MHINT) in quiet and noisy environments. The assessment method used was characterized by modifications of speech and noise configurations, as well as by changes of speech presentation mode. The benefit of spatial hearing was measured by speech recognition threshold (SRT) variation between speech condition 1 (SC1) and speech condition 2 (SC2). Results There was no significant difference found in the SRT between SC1 and SC2 in quiet. SRT in SC1 was about 4.2 dB lower than that in SC2, both in speech-shaped and four-babble noise conditions. SRTs measured in both SC1 and SC2 were lower in the speech-shaped noise condition than in the four-babble noise condition. Conclusion Spatial hearing in young normal-hearing participants contribute to speech recognition in noisy environments, but provide no benefit to speech recognition in quiet environments, which may be due to the offset of auditory extrinsic redundancy against the lack of spatial hearing. 展开更多
关键词 speech recognition spatial hearing speech recognition threshold speech perception
原文传递
FREQUENCY-DIFFERENTIATED BRAIN NETWORK DYNAMICS IN REAL WORDS AND PSEUDOWORDS PERCEPTION
13
作者 Guo Taiyang ZHAO Bin +2 位作者 ZHANG Gaoyan Dang Jianwu UNOKI Masashi 《中国语音学报》 2020年第1期114-122,共9页
Brain mechanisms of lexical-semantic processing have been well researched using electroencephalography(EEG)technique with high temporal resolution.However,the detailed brain dynamics regarding spatial connectivity and... Brain mechanisms of lexical-semantic processing have been well researched using electroencephalography(EEG)technique with high temporal resolution.However,the detailed brain dynamics regarding spatial connectivity and the spectral characteristics remain to be clarified.For this reason,this study performed frequency-specific effective connectivity analysis for the EEG recordings during the processing of real and pseudowords.In addition,we introduced f MRI-based network templates into a representational similarity analysis to compare the functional differences between real and pseudowords in different frequency bands.Our results revealed that real words could rapidly activate the brain network for speech perception and complete its comprehension with efficiency,especially when the first syllable of the real word has clear categorical features.In contrast,the pseudowords were delayed in the initiation of speech perception and required a longer time span to retrieve its meaning.The frequency-specific analysis showed that the theta,alpha,and beta rhythms contribute more to semantic processing than the gamma oscillation.These results showed that semantic processing is frequency-specific and time-dependent on the word categories. 展开更多
关键词 speech perception EEG Brain network Dynamic brain network Frequency specified brain network
原文传递
PERCEPTION OF VELARIZED/l/AND VOWEL/?/CONTRAST BY CHINESE LEARNERS OF ENGLISH
14
作者 LIAO Sishi DING Hongwei 《中国语音学报》 2020年第2期104-113,共10页
This paper aims to examine the second language(L2)phonetic categorical perception(CP)pattern by Chinese learners of English,regarding the contrast of dark/l/and vowel/?/.Three perception experiments were carried out p... This paper aims to examine the second language(L2)phonetic categorical perception(CP)pattern by Chinese learners of English,regarding the contrast of dark/l/and vowel/?/.Three perception experiments were carried out progressively:a simple identification task,an AXB identification task,and a revised AX discrimination task.The study discovered a significant difference in vowel contexts in the perception of dark/l/and vowel/?/,in which high vowels stand out,and demonstrated that English proficiency evaluated by standard examinations cannot be reflected in L2 phonetic discrimination.The study also proved the validity of adding reference stimuli in enhancing CP performance,but this improvement only benefits the identification tasks.The study helps to fill in the current knowledge gap concerning Chinese L2 learners’difficulty in distinguishing dark/l/and vowel/?/.The new finding contributes to a deeper understanding of the vowel-context effect on CP performance,as well as implications in second language teaching in exploring the connections between L2 speech perception and production. 展开更多
关键词 L2 speech perception Velarized/l/ Categorical perception
原文传递
Excitation patterns and phonetic identities of Chinese vowels 被引量:2
15
作者 YU ShuiYuan 《Science in China(Series F)》 2009年第10期1936-1946,共11页
After entering the peripheral auditory system, a sound undergoes many significant changes. The excitation pattern describes these changes psychoacoustically as inner expression. This study investigates the relations b... After entering the peripheral auditory system, a sound undergoes many significant changes. The excitation pattern describes these changes psychoacoustically as inner expression. This study investigates the relations between excitation patterns and their phonetic qualities for Chinese steady-state vowels. First, the peak positions of the envelope of excitation patterns were measured on a database. The results demonstrated that each Chinese vowel has its own special position for the representative peak of the excitation pattern. Second, to examine the sufficiency of these results, a series of experiments that consisted of identification and evaluation tasks were conducted, in which spectral components of natural isolated vowels were manipulated to create certain excitation patterns. Subjects' responses of these stimuli show that the position of the representative peak of the excitation pattern of a vowel plays a crucial role on its phonetic identity. The results suggest that the phonetic identity of vowels is determined by the position of the representative peak of the excitation pattern evoked by it, and other peaks, if any, do not have phonetic meaning. Additionally, several phenomena about speech perception are discussed on the basis of this study. 展开更多
关键词 excitation pattern speech perception phonetic quality steady-state vowel
原文传递
Audiovisual bimodal mutual compensation of Chinese
16
作者 周治 杜利民 徐彦居 《Science China(Technological Sciences)》 SCIE EI CAS 2001年第1期19-26,共8页
The perception of human languages is inherently a multi-modalprocess, in which audio information can be compensated by visual information to improve the recognition performance. Such a phenomenon in English, German, S... The perception of human languages is inherently a multi-modalprocess, in which audio information can be compensated by visual information to improve the recognition performance. Such a phenomenon in English, German, Spanish and so on has been researched, but in Chinese it has not been reported yet. In our experiment, 14 syllables (/ba, bi, bian, biao, bin, de, di, dian, duo, dong, gai, gan, gen, gu/), extracted from Chinese audiovisual bimodal speech database CAVSR-1.0, were pronounced by 10 subjects. The audio-only stimuli, audiovisual stimuli, and visual-only stimuli were recognized by 20 observers. The audio-only stimuli and audiovisual stimuli both were presented under 5 conditions: no noise, SNR 0 dB, -8 dB, -12 dB, and -16 dB. The experimental result is studied and the following conclusions for Chinese speech are reached. Human beings can recognize visual-only stimuli rather well. The place of articulation determines the visual distinction. In noisy environment, audio information can remarkably be compensated by visual information and as a result the recognition performance is greatly improved. 展开更多
关键词 audiovisual bimodal speech recognition bimodal speech perception perception experiment audio-visual information mutual compensation
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部