期刊文献+
共找到7篇文章
< 1 >
每页显示 20 50 100
Efficient visual learning by bumble bees in virtual-reality conditions:Size does not matter
1
作者 Gregory Lafon Marco Paoli +4 位作者 Benjamin HPaffhausen Gabriela de Brito Sanchez Mathieu Lihoreau Aurore Avarguès-Weber Martin Giurfa 《Insect Science》 SCIE CSCD 2023年第6期1734-1748,共15页
Recent developments allowed establishing virtual-reality (VR) setups to study multiple aspects of visual learning in honey bees under controlled experimental conditions. Here, we adopted a VR environment to investigat... Recent developments allowed establishing virtual-reality (VR) setups to study multiple aspects of visual learning in honey bees under controlled experimental conditions. Here, we adopted a VR environment to investigate the visual learning in the buff-tailed bumble bee Bombus terrestris. Based on responses to appetitive and aversive reinforcements used for conditioning, we show that bumble bees had the proper appetitive motivation to engage in the VR experiments and that they learned efficiently elemental color discriminations. In doing so, they reduced the latency to make a choice, increased the proportion of direct paths toward the virtual stimuli and walked faster toward them. Performance in a short-term retention test showed that bumble bees chose and fixated longer on the correct stimulus in the absence of reinforcement. Body size and weight, although variable across individuals, did not affect cognitive performances and had a mild impact on motor performances. Overall, we show that bumble bees are suitable experimental subjects for experiments on visual learning under VR conditions, which opens important perspectives for invasive studies on the neural and molecular bases of such learning given the robustness of these insects and the accessibility of their brain. 展开更多
关键词 Bombus terrestris bumble bees interindividual size/weight differences reinforcement responses visual learning virtual reality
原文传递
Causal Reasoning Meets Visual Representation Learning: A Prospective Study 被引量:1
2
作者 Yang Liu Yu-Shen Wei +2 位作者 Hong Yan Guan-Bin Li Liang Lin 《Machine Intelligence Research》 EI CSCD 2022年第6期485-511,共27页
Visual representation learning is ubiquitous in various real-world applications,including visual comprehension,video understanding,multi-modal analysis,human-computer interaction,and urban computing.Due to the emergen... Visual representation learning is ubiquitous in various real-world applications,including visual comprehension,video understanding,multi-modal analysis,human-computer interaction,and urban computing.Due to the emergence of huge amounts of multimodal heterogeneous spatial/temporal/spatial-temporal data in the big data era,the lack of interpretability,robustness,and out-of-distribution generalization are becoming the challenges of the existing visual models.The majority of the existing methods tend to fit the original data/variable distributions and ignore the essential causal relations behind the multi-modal knowledge,which lacks unified guidance and analysis about why modern visual representation learning methods easily collapse into data bias and have limited generalization and cognitive abilities.Inspired by the strong inference ability of human-level agents,recent years have therefore witnessed great effort in developing causal reasoning paradigms to realize robust representation and model learning with good cognitive ability.In this paper,we conduct a comprehensive review of existing causal reasoning methods for visual representation learning,covering fundamental theories,models,and datasets.The limitations of current methods and datasets are also discussed.Moreover,we propose some prospective challenges,opportunities,and future research directions for benchmarking causal reasoning algorithms in visual representation learning.This paper aims to provide a comprehensive overview of this emerging field,attract attention,encourage discussions,bring to the forefront the urgency of developing novel causal reasoning methods,publicly available benchmarks,and consensus-building standards for reliable visual representation learning and related real-world applications more efficiently. 展开更多
关键词 Causal reasoning visual representation learning reliable artificial intelligence spatial-temporal data multi-modal analysis
原文传递
VISHIEN-MAAT:Scrollytelling visualization design for explaining Siamese Neural Network concept to non-technical users
3
作者 Noptanit Chotisarn Sarun Gulyanon +1 位作者 Tianye Zhang Wei Chen 《Visual Informatics》 EI 2023年第1期18-29,共12页
The past decade has witnessed rapid progress in AI research since the breakthrough in deep learning.AI technology has been applied in almost every field;therefore,technical and non-technical endusers must understand t... The past decade has witnessed rapid progress in AI research since the breakthrough in deep learning.AI technology has been applied in almost every field;therefore,technical and non-technical endusers must understand these technologies to exploit them.However existing materials are designed for experts,but non-technical users need appealing materials that deliver complex ideas in easy-tofollow steps.One notable tool that fits such a profile is scrollytelling,an approach to storytelling that provides readers with a natural and rich experience at the reader’s pace,along with in-depth interactive explanations of complex concepts.Hence,this work proposes a novel visualization design for creating a scrollytelling that can effectively explain an AI concept to non-technical users.As a demonstration of our design,we created a scrollytelling to explain the Siamese Neural Network for the visual similarity matching problem.Our approach helps create a visualization valuable for a shorttimeline situation like a sales pitch.The results show that the visualization based on our novel design helps improve non-technical users’perception and machine learning concept knowledge acquisition compared to traditional materials like online articles. 展开更多
关键词 Story synthesis Scrollytelling visual storytelling visualizing deep learning learning science
原文传递
Proteomic and transcriptomic analysis of visual long-term memory in Drosophila melanogaster 被引量:1
4
作者 Huoqing Jiang Qinlong Hou +1 位作者 Zhefeng Gong Li Liu 《Protein & Cell》 SCIE CSCD 2011年第3期215-222,共8页
The fruit fly,Drosophila melanogaster,is able to discriminate visual landmarks and form visual long-term memory in a flight simulator.Studies focused on the molecular mechanism of long-term memory have shown that memo... The fruit fly,Drosophila melanogaster,is able to discriminate visual landmarks and form visual long-term memory in a flight simulator.Studies focused on the molecular mechanism of long-term memory have shown that memory formation requires mRNA transcription and protein synthesis.However,little is known about the molecular mechanisms underlying the visual learning paradigm.The present study demonstrated that both spaced training procedure(STP)and consecutive training procedure(CTP)would induce long-term memory at 12 hour after training,and STP caused significantly higher 12-h memory scores compared with CTP.Labelfree quantification of liquid chromatography-tandem mass spectrometry(LC-MS/MS)and microarray were utilized to analyze proteomic and transcriptomic differences between the STP and CTP groups.Proteomic analysis revealed 30 up-regulated and 27 down-regulated proteins;Transcriptomic analysis revealed 145 up-regulated and 129 down-regulated genes.Among them,five candidate genes were verified by quantitative PCR,which revealed results similar to microarray.These results provide insight into the molecular components influencing visual long-term memory and facilitate further studies on the roles of identified genes in memory formation. 展开更多
关键词 visual learning and memory DROSOPHILA long-term memory MICROARRAY liquid chromatography-tandem mass spectrometry
原文传递
Masked Autoencoders as Single Object Tracking Learners
5
作者 Chunjuan Bo XinChen Junxing Zhang 《Computers, Materials & Continua》 SCIE EI 2024年第7期1105-1122,共18页
Significant advancements have beenwitnessed in visual tracking applications leveragingViT in recent years,mainly due to the formidablemodeling capabilities of Vision Transformer(ViT).However,the strong performance of ... Significant advancements have beenwitnessed in visual tracking applications leveragingViT in recent years,mainly due to the formidablemodeling capabilities of Vision Transformer(ViT).However,the strong performance of such trackers heavily relies on ViT models pretrained for long periods,limitingmore flexible model designs for tracking tasks.To address this issue,we propose an efficient unsupervised ViT pretraining method for the tracking task based on masked autoencoders,called TrackMAE.During pretraining,we employ two shared-parameter ViTs,serving as the appearance encoder and motion encoder,respectively.The appearance encoder encodes randomly masked image data,while the motion encoder encodes randomly masked pairs of video frames.Subsequently,an appearance decoder and a motion decoder separately reconstruct the original image data and video frame data at the pixel level.In this way,ViT learns to understand both the appearance of images and the motion between video frames simultaneously.Experimental results demonstrate that ViT-Base and ViT-Large models,pretrained with TrackMAE and combined with a simple tracking head,achieve state-of-the-art(SOTA)performance without additional design.Moreover,compared to the currently popular MAE pretraining methods,TrackMAE consumes only 1/5 of the training time,which will facilitate the customization of diverse models for tracking.For instance,we additionally customize a lightweight ViT-XS,which achieves SOTA efficient tracking performance. 展开更多
关键词 visual object tracking vision transformer masked autoencoder visual representation learning
下载PDF
Same or different?Abstract relational concept use in juvenile bamboo sharks and Malawi cichlids
6
作者 Theodora Fuss Leonie John Vera Schluessel 《Current Zoology》 SCIE CAS CSCD 2021年第3期279-292,共14页
Sorting objects and events into categories and concepts is an important cognitive prerequisite that spares an individual the learning of every object or situation encountered in its daily life.Accordingly,specific ite... Sorting objects and events into categories and concepts is an important cognitive prerequisite that spares an individual the learning of every object or situation encountered in its daily life.Accordingly,specific items are classified in general groups that allow fast responses to novel situations.The present study assessed whether bamboo sharks Chiloscyllium griseum and Malawi cichlids Pseudotropheus zebra can distinguish sets of stimuli(each stimulus consisting of two abstract,geometric objects)that meet two conceptual preconditions,i.e.,(1)"sameness"versus"difference"and(2)a certain spatial arrangement of both objects.In two alternative forced choice experiments,individuals were first trained to choose two different,vertically arranged objects from two different but horizontally arranged ones.Pair discriminations were followed by extensive transfer test experiments.Transfer tests using stimuli consisting of(a)black and gray circles and(b)squares with novel geometric patterns provided conflicting information with respect to the learnt rule"choose two different,vertically arranged objects",thereby investigating(1)the individuals'ability to transfer previously gained knowledge to novel stimuli and(2)the abstract relational concept(s)or rule(s)applied to categorize these novel objects.Present results suggest that the level of processing and usage of both abstract concepts differed considerably between bamboo sharks and Malawi cichlids.Bamboo sharks seemed to combine both concepts-although not with equal but hierarchical prominence-pointing to advanced cognitive capabilities.Conversely,Malawi cichlids had difficulties in discriminating between symbols and failed to apply the acquired training knowledge on new sets of geometric and,in particular,gray-level transfer stimuli. 展开更多
关键词 visual concept learning spatial arrangement relational abstract concept Malawi cichlid bamboo shark
原文传递
Human-Computer Interaction Lab(HCIL)in Seoul National University
7
作者 GuHyun Han Jaemin Jo +1 位作者 Han Joo Chae Jinwook Seo 《Visual Informatics》 EI 2020年第4期35-39,共5页
This article introduces Human–Computer Interaction Laboratory(HCIL)established at Seoul National University,Korea,in 2009.We first summarized the history of foundation,achievement,and collaboration for the last 10 ye... This article introduces Human–Computer Interaction Laboratory(HCIL)established at Seoul National University,Korea,in 2009.We first summarized the history of foundation,achievement,and collaboration for the last 10 years.Then,we delineated our current research directions related to information visualization.Finally,we presented our facilities and equipment to adequately support the research. 展开更多
关键词 Lab introduction Information visualization visual analytics Progressive visualization Machine learning for visualization MR and AR
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部