期刊文献+
共找到1,829篇文章
< 1 2 92 >
每页显示 20 50 100
Values of macular ganglion cell-inner plexiform layer and 10-2 visual field measurements in detecting and evaluating glaucoma
1
作者 Hai-Jian Hu Ping Li +7 位作者 Bin Tong Yu-Lian Pang Hong-Dou Luo Fei-Fei Wang Chan Xiong Yu-Lin Yu Hai He Xu Zhang 《International Journal of Ophthalmology(English edition)》 SCIE CAS 2024年第5期852-860,共9页
AIM:To assess the performance of macular ganglion cell-inner plexiform layer thickness(mGCIPLT)and 10-2 visual field(VF)parameters in detecting early glaucoma and evaluating the severity of advanced glaucoma.METHODS:T... AIM:To assess the performance of macular ganglion cell-inner plexiform layer thickness(mGCIPLT)and 10-2 visual field(VF)parameters in detecting early glaucoma and evaluating the severity of advanced glaucoma.METHODS:Totally 127 eyes from 89 participants(36 eyes of 19 healthy participants,45 eyes of 31 early glaucoma patients and 46 eyes of 39 advanced glaucoma patients)were included.The relationships between the optical coherence tomography(OCT)-derived parameters and VF sensitivity were determined.Patients with early glaucoma were divided into eyes with or without central 10°of the VF damages(CVFDs),and the diagnostic performances of OCT-derived parameters were assessed.RESULTS:In early glaucoma,the mGCIPLT was significantly correlated with 10-2 VF pattern standard deviation(PSD;with average mGCIPLT:β=-0.046,95%CI,-0.067 to-0.024,P<0.001).In advanced glaucoma,the mGCIPLT was related to the 24-2 VF mean deviation(MD;with average mGCIPLT:β=0.397,95%CI,0.199 to 0.595,P<0.001),10-2 VF MD(with average mGCIPLT:β=0.762,95%CI,0.485 to 1.038,P<0.001)and 24-2 VF PSD(with average mGCIPLT:β=0.244,95%CI,0.124 to 0.364,P<0.001).Except for the minimum and superotemporal mGCIPLT,the decrease of mGCIPLT in early glaucomatous eyes with CVFDs was more severe than that of early glaucomatous eyes without CVFDs.The area under the curve(AUC)of the average mGCIPLT(AUC=0.949,95%CI,0.868 to 0.982)was greater than that of the average circumpapillary retinal nerve fiber layer thickness(cpRNFLT;AUC=0.827,95%CI,0.674 to 0.918)and rim area(AUC=0.799,95%CI,0.610 to 0.907)in early glaucomatous eyes with CVFDs versus normal eyes.CONCLUSION:The 10-2 VF and mGCIPLT parameters are complementary to 24-2 VF,cpRNFLT and ONH parameters,especially in detecting early glaucoma with CVFDs and evaluating the severity of advanced glaucoma in group level. 展开更多
关键词 10-2 visual field ganglion cell-inner plexiform layer retinal nerve fiber layer thickness GLAUCOMA
下载PDF
Integrating Audio-Visual Features and Text Information for Story Segmentation of News Video 被引量:1
2
作者 Liu Hua-yong, Zhou Dong-ru School of Computer,Wuhan University,Wuhan 430072, Hubei, China 《Wuhan University Journal of Natural Sciences》 CAS 2003年第04A期1070-1074,共5页
Video data are composed of multimodal information streams including visual, auditory and textual streams, so an approach of story segmentation for news video using multimodal analysis is described in this paper. The p... Video data are composed of multimodal information streams including visual, auditory and textual streams, so an approach of story segmentation for news video using multimodal analysis is described in this paper. The proposed approach detects the topic-caption frames, and integrates them with silence clips detection results, as well as shot segmentation results to locate the news story boundaries. The integration of audio-visual features and text information overcomes the weakness of the approach using only image analysis techniques. On test data with 135 400 frames, when the boundaries between news stories are detected, the accuracy rate 85.8% and the recall rate 97.5% are obtained. The experimental results show the approach is valid and robust. 展开更多
关键词 news video story segmentation audio-visual features analysis text detection
下载PDF
On‐device audio‐visual multi‐person wake word spotting
3
作者 Yidi Li Guoquan Wang +2 位作者 Zhan Chen Hao Tang Hong Liu 《CAAI Transactions on Intelligence Technology》 SCIE EI 2023年第4期1578-1589,共12页
Audio‐visual wake word spotting is a challenging multi‐modal task that exploits visual information of lip motion patterns to supplement acoustic speech to improve overall detection performance.However,most audio‐vi... Audio‐visual wake word spotting is a challenging multi‐modal task that exploits visual information of lip motion patterns to supplement acoustic speech to improve overall detection performance.However,most audio‐visual wake word spotting models are only suitable for simple single‐speaker scenarios and require high computational complexity.Further development is hindered by complex multi‐person scenarios and computational limitations in mobile environments.In this paper,a novel audio‐visual model is proposed for on‐device multi‐person wake word spotting.Firstly,an attention‐based audio‐visual voice activity detection module is presented,which generates an attention score matrix of audio and visual representations to derive active speaker representation.Secondly,the knowledge distillation method is introduced to transfer knowledge from the large model to the on‐device model to control the size of our model.Moreover,a new audio‐visual dataset,PKU‐KWS,is collected for sentence‐level multi‐person wake word spotting.Experimental results on the PKU‐KWS dataset show that this approach outperforms the previous state‐of‐the‐art methods. 展开更多
关键词 audiovisual fusion human‐computer interfacing speech processing
下载PDF
Securing Technique Using Pattern-Based LSB Audio Steganography and Intensity-Based Visual Cryptography 被引量:2
4
作者 Pranati Rakshit Sreeparna Ganguly +2 位作者 Souvik Pal Ayman AAly Dac-Nhuong Le 《Computers, Materials & Continua》 SCIE EI 2021年第4期1207-1224,共18页
With the increasing need of sensitive or secret data transmission through public network,security demands using cryptography and steganography are becoming a thirsty research area of last few years.These two technique... With the increasing need of sensitive or secret data transmission through public network,security demands using cryptography and steganography are becoming a thirsty research area of last few years.These two techniques can be merged and provide better security which is nowadays extremely required.The proposed system provides a novel method of information security using the techniques of audio steganography combined with visual cryptography.In this system,we take a secret image and divide it into several subparts to make more than one incomprehensible sub-images using the method of visual cryptography.Each of the sub-images is then hidden within individual cover audio files using audio steganographic techniques.The cover audios are then sent to the required destinations where reverse steganography schemes are applied to them to get the incomprehensible component images back.At last,all the sub-images are superimposed to get the actual secret image.This method is very secure as it uses a two-step security mechanism to maintain secrecy.The possibility of interception is less in this technique because one must have each piece of correct sub-image to regenerate the actual secret image.Without superimposing every one of the sub-images meaningful secret images cannot be formed.Audio files are composed of densely packed bits.The high density of data in audio makes it hard for a listener to detect the manipulation due to the proposed time-domain audio steganographic method. 展开更多
关键词 Information security visual cryptography audio steganography secret image reverse steganography
下载PDF
Application of Conversational Implicatures in Teaching English Audio-visual Course
5
作者 刘慧莹 《商情》 2014年第17期370-371,共2页
关键词 英语学习 学习方法 阅读知识 阅读材料
下载PDF
Audio-visual emotion recognition with multilayer boosted HMM
6
作者 吕坤 贾云得 张欣 《Journal of Beijing Institute of Technology》 EI CAS 2013年第1期89-93,共5页
Emotion recognition has become an important task of modern human-computer interac- tion. A multilayer boosted HMM ( MBHMM ) classifier for automatic audio-visual emotion recognition is presented in this paper. A mod... Emotion recognition has become an important task of modern human-computer interac- tion. A multilayer boosted HMM ( MBHMM ) classifier for automatic audio-visual emotion recognition is presented in this paper. A modified Baum-Welch algorithm is proposed for component HMM learn- ing and adaptive boosting (AdaBoost) is used to train ensemble classifiers for different layers (cues). Except for the first layer, the initial weights of training samples in current layer are decided by recognition results of the ensemble classifier in the upper layer. Thus the training procedure using current cue can focus more on the difficult samples according to the previous cue. Our MBHMM clas- sifier is combined by these ensemble classifiers and takes advantage of the complementary informa- tion from multiple cues and modalities. Experimental results on audio-visual emotion data collected in Wizard of Oz scenarios and labeled under two types of emotion category sets demonstrate that our approach is effective and promising. 展开更多
关键词 emotion recognition audio-visual fusion Baum-Welch algorithm multilayer boostedHMM Wizard of Oz scenario
下载PDF
Application of Task-based Teaching Method to College Audio-visual English Teaching
7
作者 Liguo Shi 《International Journal of Technology Management》 2015年第9期65-67,共3页
关键词 大学英语 教学法 视听 应用 教学过程 课堂教学 教学特点
下载PDF
Interpreting Synergism Based Upon Visual-Audio-Oral Course of EAP: A Case of CI Between English and Chinese
8
作者 KANG Zhi-feng 《Journal of Literature and Art Studies》 2015年第10期889-900,共12页
关键词 协同作用 英语专业 EAP 课程 视听 汉语 传输理论 目标语言
下载PDF
The Research on Audio-Visual-Oral Instructional theory in Foreign Language
9
作者 FENG Xiaowei 《Journal of Zhouyi Research》 2014年第3期4-6,共3页
关键词 口语教学 视听 外语 教学理论 语言学习 语言教学 教学设备 心理基础
下载PDF
Prioritized MPEG-4 Audio-Visual Objects Streaming over the DiffServ
10
作者 黄天云 郑婵 《Journal of Electronic Science and Technology of China》 2005年第4期314-320,共7页
The object-based scalable coding in MPEG-4 is investigated, and a prioritized transmission scheme of MPEG-4 audio-visual objects (AVOs) over the DiffServ network with the QoS guarantee is proposed. MPEG-4 AVOs are e... The object-based scalable coding in MPEG-4 is investigated, and a prioritized transmission scheme of MPEG-4 audio-visual objects (AVOs) over the DiffServ network with the QoS guarantee is proposed. MPEG-4 AVOs are extracted and classified into different groups according to their priority values and scalable layers (visual importance). These priority values are mapped to the 1P DiffServ per hop behaviors (PHB). This scheme can selectively discard packets with low importance, in order to avoid the network congestion. Simulation results show that the quality of received video can gracefully adapt to network state, as compared with the ‘best-effort' manner. Also, by allowing the content provider to define prioritization of each audio-visual object, the adaptive transmission of object-based scalable video can be customized based on the content. 展开更多
关键词 video streaming quality of service (QoS) MPEG-4 audio-visual objects (AVOs) DIFFSERV PRIORITIZATION
下载PDF
The Effect of Audio Visual Entrainment on Pre-Attentive Dysfunctional Processing to Stressful Events in Anxious Individuals
11
作者 Guadalupe Villarreal Trevino Ernesto Octavio Lopez Ramirez +2 位作者 Guadalupe Elizabeth Morales Martinez Claudia Castro Campos Maria Elena Urdiales Ibarra 《Open Journal of Medical Psychology》 2014年第5期364-372,共9页
Experimental single case studies on automatic processing of emotion were carried on a sample of people with an anxiety disorder. Participants were required to take three Audio Visual Entrainment (AVE) sessions to test... Experimental single case studies on automatic processing of emotion were carried on a sample of people with an anxiety disorder. Participants were required to take three Audio Visual Entrainment (AVE) sessions to test for anxiety reduction as proclaimed by some academic research. Explicit reports were measured as well as pre-attentive bias to stressing information by using affective priming studies before and after AVE intervention. Group analysis shows that indeed AVEs program applications do reduce anxiety producing significant changes over explicit reports on anxiety levels and automatic processing bias of emotion. However, case by case analysis of six anxious participants shows that even when all of the participants report emotional improvement after intervention, not all of them reduce or eliminate dysfunctional bias to stressing information. Rather, they show a variety of processing styles due to intervention and some of them show no change at all. Implications of this differential effect to clinical sets are discussed. 展开更多
关键词 audio visual Entrainment Anxiety Disorders Affective Priming Single Case Experimental Study
下载PDF
On the Development of a Maritime English Visual-Audio-Lingual Course Book
12
作者 GAO Song 《Journal of Shipping and Ocean Engineering》 2016年第4期212-220,共9页
下载PDF
基于自然环境的视-听-嗅感知交互体验对居民福祉之研究综述 被引量:2
13
作者 邱玲 齐莹 高天 《中国园林》 CSCD 北大核心 2023年第11期29-35,共7页
高速的城市化发展迫使人们需构建人与自然和谐的宜居城市来降低未来的风险,而掌握人类对自然感知裨益的信息是有效降低风险的基础。尽管前人面对自然环境的感知裨益做了一定的研究,但涉及多感官交互裨益的研究仍处于探索阶段,缺少相关... 高速的城市化发展迫使人们需构建人与自然和谐的宜居城市来降低未来的风险,而掌握人类对自然感知裨益的信息是有效降低风险的基础。尽管前人面对自然环境的感知裨益做了一定的研究,但涉及多感官交互裨益的研究仍处于探索阶段,缺少相关方法的总结与理论的概括,实践中难以真正实现人类与自然和谐的环境建设。因此,借助文献计量分析的CiteSpace软件平台,基于Web of Science核心数据库与中国知网数据库,系统分析国内外基于自然环境的视听嗅感知交互对居民福祉研究的进展及动向。结果表明:1)发文阶段性突出,近5年爆发性增长,主要集中于亚欧地区,发文学科与期刊成多学科交叉特点,且研究由单维感知向多维感知裨益转变;2)研究样地涉及自然要素、自然景观构成及土地覆盖类型3个层面,研究对象包含不同年龄阶段群体,且主要以大学生为主;3)应用的研究方法主要包括实际现场、室内模拟及社交媒体众包数据分析3类,且感知维度数与感知裨益效应呈非线性正相关,其中自然要素、环境一致性及自然认同感是影响感知裨益的关键因素;4)该领域涉及环境心理及景观复愈等多理论的作用途径,进而提出了“刺激-有机体-反应”(Stimulus-Organism-Response,SOR)的机理概念框架,以期为今后以人类福祉为导向的城市绿地规划与管理提供理论依据和实践方法。 展开更多
关键词 风景园林 自然环境 福祉 感官交互 视听嗅感知 绿地规划
下载PDF
结合光流算法与注意力机制的U-Net网络跨模态视听语音分离 被引量:1
14
作者 兰朝凤 蒋朋威 +2 位作者 陈欢 韩闯 郭小霞 《电子与信息学报》 EI CSCD 北大核心 2023年第10期3538-3546,共9页
目前多数的视听分离模型,大多是基于视频特征和音频特征简单拼接,没有充分考虑各个模态的相互关系,导致视觉信息未被充分利用,该文针对此问题提出了新的模型。该文充分考虑视觉特征、音频特征之间的相互联系,采用多头注意力机制,结合稠... 目前多数的视听分离模型,大多是基于视频特征和音频特征简单拼接,没有充分考虑各个模态的相互关系,导致视觉信息未被充分利用,该文针对此问题提出了新的模型。该文充分考虑视觉特征、音频特征之间的相互联系,采用多头注意力机制,结合稠密光流(Farneback)算法和U-Net网络,提出跨模态融合的光流-视听语音分离(Flow-AVSS)模型。该模型通过Farneback算法和轻量级网络ShuffleNet v2分别提取运动特征和唇部特征,然后将运动特征与唇部特征进行仿射变换,经过时间卷积模块(TCN)得到视觉特征。为充分利用到视觉信息,在进行特征融合时采用多头注意力机制,将视觉特征与音频特征进行跨模态融合,得到融合视听特征,最后融合视听特征经过U-Net分离网络得到分离语音。利用客观语音质量评估(PESQ)、短时客观可懂度(STOI)及源失真比(SDR)评价指标,在AVspeech数据集进行实验测试。研究表明,该文所提方法与纯语音分离网络和仅采用特征拼接的视听分离网络相比,性能上分别提高了2.23 dB和1.68 dB。由此表明,采用跨模态注意力进行特征融合,能更加充分利用各个模态相关性,增加的唇部运动特征,能有效提高视频特征的鲁棒性,提高分离效果。 展开更多
关键词 视听语音分离 视听融合 跨模态注意力 光流算法
下载PDF
基于Visual Modflow研究武隆-广杨深层岩溶水径流特征 被引量:1
15
作者 何浩 张强 +3 位作者 张金林 冯杰 李威龙 王志鹏 《甘肃水利水电技术》 2022年第2期22-27,34,共7页
近年来,在武隆地区经济快速发展的同时,对水资源的需求量大增,土地石漠化等水文地质问题频发,使岩溶地区生态安全受到巨大威胁。运用Visual Modflow软件对研究区的典型剖面进行数值模拟,研究武隆-广杨地区深层岩溶水的运动规律,研究该... 近年来,在武隆地区经济快速发展的同时,对水资源的需求量大增,土地石漠化等水文地质问题频发,使岩溶地区生态安全受到巨大威胁。运用Visual Modflow软件对研究区的典型剖面进行数值模拟,研究武隆-广杨地区深层岩溶水的运动规律,研究该地区深层岩溶水流系统循环特征、剖面上岩溶水的径流强度与滞留时间,为该地区工程建设提供技术支撑。研究结果表明:研究区地下存在局部、中间、区域三级地下水流动系统,其排泄区为十三沟、百笋溪和乌江,渗流速度分别为0.34 m/d、0.59 m/d和0.51 m/d。整体上看,研究区局部流动系统最发育,地下水滞留时间10~110年。由于百笋溪与乌江相距不远,垂直海拔相近,导致中间水流系统和区域水流系统发育程度相近,地下水滞留时间分别为约420年和约680年,并且岩溶水年龄沿岩溶水流动路径有逐渐衰老的趋势。总体而言,研究区的岩溶水年龄均小于700年,为现代水。 展开更多
关键词 visual Modflow 武隆-广杨 深层岩溶水 径流特征
下载PDF
Visual Design and Searching of Multimedia Percussion Theatre "The Call from Sigangli A Dialogue of Natural Character and Avant-garde"
16
作者 YuhuaiCheng 《International Journal of Technology Management》 2013年第5期24-25,共2页
关键词 多媒体图像 视觉设计 打击乐器 前卫 搜索 对话 自然 呼叫
下载PDF
神话与科技的“联姻”--动画电影《飞奔去月球》的艺术价值
17
作者 高娟娟 《新疆艺术学院学报》 2023年第3期83-89,共7页
动画电影《飞奔去月球》将中国古代嫦娥奔月的神话故事与现代科技进行了完美融合,借助歌舞、色彩等多种视听元素,以新颖的艺术形式叙述少女成长主题,从而表现爱的丰富性和多样性。该片带给国内外观众别样的审美体验,彰显了独特的艺术价... 动画电影《飞奔去月球》将中国古代嫦娥奔月的神话故事与现代科技进行了完美融合,借助歌舞、色彩等多种视听元素,以新颖的艺术形式叙述少女成长主题,从而表现爱的丰富性和多样性。该片带给国内外观众别样的审美体验,彰显了独特的艺术价值,取得了动画电影艺术上的成功。深入挖掘动画电影《飞奔去月球》的审美构成和艺术风格,剖析其中蕴含的文化特征,可以发现该片对中国动画电影市场的发展具有一定的借鉴意义。 展开更多
关键词 《飞奔去月球》 视听艺术 文化内涵
下载PDF
First Asia-Pacific Symposium on Visual Sciences November 2-5,1994 Guangzhou,PR China
18
《眼科学报》 1993年第3期161-,共1页
With the largest population in the world,the Asia-Pacific area is in a great need for the fundamentalresearch of visual sciences,the protection of vision and the prevention and treatment of visual diseases.The Symposi... With the largest population in the world,the Asia-Pacific area is in a great need for the fundamentalresearch of visual sciences,the protection of vision and the prevention and treatment of visual diseases.The Symposium will open a new era of the academic exchanges in the field of visual sciences in thisarea.It also will enhance the academic exchanges of visual sciences worldwide. 展开更多
关键词 PR Asia First Asia-Pacific Symposium on visual Sciences November 2-5 1994 Guangzhou PR China
下载PDF
基于ⅢF A/V规范和Avalon系统的大学图书馆视听数据库建设研究
19
作者 张毅 熊泽泉 +1 位作者 胡晓明 陈丹 《图书馆杂志》 北大核心 2024年第1期50-58,49,共10页
随着中国网络基础设施的不断改善,视听媒体在年轻一代中非常流行,给以文本资源为主的图书馆带来了挑战。本研究旨在探究国内外大学图书馆视听资源数据库建设的现状,借鉴ⅢF规范在图像资源管理方面的成功经验和各种视听保存社区的实践,... 随着中国网络基础设施的不断改善,视听媒体在年轻一代中非常流行,给以文本资源为主的图书馆带来了挑战。本研究旨在探究国内外大学图书馆视听资源数据库建设的现状,借鉴ⅢF规范在图像资源管理方面的成功经验和各种视听保存社区的实践,提出基于ⅢF A/V规范与开源软件的中国大学图书馆视听资源管理方法。通过分析华东师范大学图书馆在视听资源保存、流媒体发布、时间轴气泡注释、转录、视听结构化和开放共享方面的实践,进行实证研究。 展开更多
关键词 视听数据库 ⅢF A/V Avalon媒体系统 视听可视化
下载PDF
基于扩张卷积和Transformer的视听融合语音分离方法
20
作者 刘宏清 谢奇洲 +1 位作者 赵宇 周翊 《信号处理》 CSCD 北大核心 2024年第7期1208-1217,共10页
为了提高语音分离的效果,除了利用混合的语音信号,还可以借助视觉信号作为辅助信息。这种融合了视觉与音频信号的多模态建模方式,已被证实可以有效地提高语音分离的性能,为语音分离任务提供了新的可能性。为了更好地捕捉视觉与音频特征... 为了提高语音分离的效果,除了利用混合的语音信号,还可以借助视觉信号作为辅助信息。这种融合了视觉与音频信号的多模态建模方式,已被证实可以有效地提高语音分离的性能,为语音分离任务提供了新的可能性。为了更好地捕捉视觉与音频特征中的长期依赖关系,并强化网络对输入上下文信息的理解,本文提出了一种基于一维扩张卷积与Transformer的时域视听融合语音分离模型。将基于频域的传统视听融合语音分离方法应用到时域中,避免了时频变换带来的信息损失和相位重构问题。所提网络架构包含四个模块:一个视觉特征提取网络,用于从视频帧中提取唇部嵌入特征;一个音频编码器,用于将混合语音转换为特征表示;一个多模态分离网络,主要由音频子网络、视频子网络,以及Transformer网络组成,用于利用视觉和音频特征进行语音分离;以及一个音频解码器,用于将分离后的特征还原为干净的语音。本文使用LRS2数据集生成的包含两个说话者混合语音的数据集。实验结果表明,所提出的网络在尺度不变信噪比改进(Scale-Invariant Signal-to-Noise Ratio Improvement,SISNRi)与信号失真比改进(Signal-to-Distortion Ratio Improvement,SDRi)这两种指标上分别达到14.0 dB与14.3 dB,较纯音频分离模型和普适的视听融合分离模型有明显的性能提升。 展开更多
关键词 语音分离 视听融合 多头自注意力机制 扩张卷积
下载PDF
上一页 1 2 92 下一页 到第
使用帮助 返回顶部