期刊文献+
共找到15篇文章
< 1 >
每页显示 20 50 100
甲状腺乳头状癌右侧喉返神经深层(Ⅵb区)淋巴结转移的风险分析 被引量:3
1
作者 程鸣鸣 柴芳 张晓明 《中国现代医学杂志》 CAS 北大核心 2022年第1期85-90,共6页
目的探讨甲状腺乳头状癌(PTC)患者发生右侧喉返神经深层(ⅥⅥb区)淋巴结转移的危险性。方法收集锦州医科大学附属第一医院2018年1月—2019年4月收治的175例PTC患者的临床资料,回顾性分析PTC患者右侧喉返神经浅层(Ⅵa区)及Ⅵb区发生淋巴... 目的探讨甲状腺乳头状癌(PTC)患者发生右侧喉返神经深层(ⅥⅥb区)淋巴结转移的危险性。方法收集锦州医科大学附属第一医院2018年1月—2019年4月收治的175例PTC患者的临床资料,回顾性分析PTC患者右侧喉返神经浅层(Ⅵa区)及Ⅵb区发生淋巴结转移在不同临床病理特征间的差异性。结果175例PTC患者中,发生Ⅵa区淋巴结转移67例,Ⅵb区淋巴结转移29例,兼有Ⅵa、Ⅵb区淋巴结转移23例。单因素分析和多因素Logistic回归分析结果显示,年龄、癌灶最大径、癌灶多发性、颈侧区淋巴结转移(LLNM)是Ⅵa区淋巴结转移的独立危险因素;癌灶最大径、右侧癌灶、LLNM、Ⅵa区淋巴结转移是Ⅵb区淋巴结转移的独立危险因素;列线图显示,癌灶最大径和癌灶位置对Ⅵb区淋巴结转移的影响最大,年龄、LLNM和Ⅵa区淋巴结转移的影响次之,癌灶多发性的影响最小。受试者工作特征曲线显示,Ⅵb区淋巴结转移的独立危险因素中癌灶最大径的诊断截断值为0.75 cm。结论PTC患者癌灶最大径、右侧癌灶、LLNM或Ⅵa区淋巴结转移为Ⅵb区淋巴结转移的独立危险因素。且可以根据列线图计算出PTC患者发生Ⅵb区淋巴结转移的概率。PTC患者存在这些危险因素且列线图评分较高时,发生Ⅵb区淋巴结转移的风险增加。 展开更多
关键词 甲状腺乳头状癌 右侧喉返神经深层 淋巴结转移 列线图 ROC曲线
下载PDF
Dual Frequency Transformer for Efficient SDR-to-HDR Translation
2
作者 Gang Xu Qibin Hou ming-ming cheng 《Machine Intelligence Research》 EI CSCD 2024年第3期538-548,共11页
The SDR-to-HDR translation technique can convert the abundant standard-dynamic-range (SDR) media resources to high-dynamic-range (HDR) ones, which can represent high-contrast scenes, providing more realistic visual ex... The SDR-to-HDR translation technique can convert the abundant standard-dynamic-range (SDR) media resources to high-dynamic-range (HDR) ones, which can represent high-contrast scenes, providing more realistic visual experiences. While recent vision Transformers have achieved promising performance in many low-level vision tasks, there are few works attempting to leverage Transformers for SDR-to-HDR translation. In this paper, we are among the first to investigate the performance of Transformers for SDR-to-HDR translation. We find that directly using the self-attention mechanism may involve artifacts in the results due to the inappropriate way to model long-range dependencies between the low-frequency and high-frequency components. Taking this into account, we advance the self-attention mechanism and present a dual frequency attention (DFA), which leverages the self-attention mechanism to separately encode the low-frequency structural information and high-frequency detail information. Based on the proposed DFA, we further design a multi-scale feature fusion network, named dual frequency Transformer (DFT), for efficient SDR-to-HDR translation. Extensive experiments on the HDRTV1K dataset demonstrate that our DFT can achieve better quantitative and qualitative performance than the recent state-of-the-art methods. The code of our DFT is made publicly available at https://github.com/CS-GangXu/DFT. 展开更多
关键词 Standard-dynamic-range to high-dynamic-range(SDR-to-HDR)translation TRANSFORMER dual frequency attention(DFA) frequency-aware feature decomposition efficient model
原文传递
Visual attention network 被引量:11
3
作者 Meng-Hao Guo cheng-Ze Lu +2 位作者 Zheng-Ning Liu ming-ming cheng Shi-Min Hu 《Computational Visual Media》 SCIE EI CSCD 2023年第4期733-752,共20页
While originally designed for natural language processing tasks,the self-attention mechanism has recently taken various computer vision areas by storm.However,the 2D nature of images brings three challenges for applyi... While originally designed for natural language processing tasks,the self-attention mechanism has recently taken various computer vision areas by storm.However,the 2D nature of images brings three challenges for applying self-attention in computer vision:(1)treating images as 1D sequences neglects their 2D structures;(2)the quadratic complexity is too expensive for high-resolution images;(3)it only captures spatial adaptability but ignores channel adaptability.In this paper,we propose a novel linear attention named large kernel attention(LKA)to enable self-adaptive and long-range correlations in self-attention while avoiding its shortcomings.Furthermore,we present a neural network based on LKA,namely Visual Attention Network(VAN).While extremely simple,VAN achieves comparable results with similar size convolutional neural networks(CNNs)and vision transformers(ViTs)in various tasks,including image classification,object detection,semantic segmentation,panoptic segmentation,pose estimation,etc.For example,VAN-B6 achieves 87.8%accuracy on ImageNet benchmark,and sets new state-of-the-art performance(58.2%PQ)for panoptic segmentation.Besides,VAN-B2 surpasses Swin-T 4%mloU(50.1%vs.46.1%)for semantic segmentation on ADE20K benchmark,2.6%AP(48.8%vs.46.2%)for object detection on COCO dataset.It provides a novel method and a simple yet strong baseline for the community.The code is available at https://github.com/Visual-Attention-Network. 展开更多
关键词 vision backbone deep learning ConvNets ATTENTION
原文传递
D2ANet:Difference-aware attention network for multi-level change detection from satellite imagery 被引量:1
4
作者 Jie Mei Yi-Bo Zheng ming-ming cheng 《Computational Visual Media》 SCIE EI CSCD 2023年第3期563-579,共17页
Recognizing dynamic variations on the ground,especially changes caused by various natural disasters,is critical for assessing the severity of the damage and directing the disaster response.However,current workflows fo... Recognizing dynamic variations on the ground,especially changes caused by various natural disasters,is critical for assessing the severity of the damage and directing the disaster response.However,current workflows for disaster assessment usually require human analysts to observe and identify damaged buildings,which is labor-intensive and unsuitable for large-scale disaster areas.In this paper,we propose a difference-aware attention network(D2ANet)for simultaneous building localization and multi-level change detection from the dual-temporal satellite imagery.Considering the differences in different channels in the features of pre-and post-disaster images,we develop a dual-temporal aggregation module using paired features to excite change-sensitive channels of the features and learn the global change pattern.Since the nature of building damage caused by disasters is diverse in complex environments,we design a difference-attention module to exploit local correlations among the multi-level changes,which improves the ability to identify damage on different scales.Extensive experiments on the large-scale building damage assessment dataset xBD demonstrate that our approach provides new state-of-the-art results.Source code is publicly available at https://github.com/mj129/D2ANet. 展开更多
关键词 change detection building localization sate-llite imagery dual-temporal aggregation difference attention
原文传递
Attention mechanisms in computer vision:A survey 被引量:71
5
作者 Meng-Hao Guo Tian-Xing Xu +7 位作者 Jiang-Jiang Liu Zheng-Ning Liu Peng-Tao Jiang Tai-Jiang Mu Song-Hai Zhang Ralph R.Martin ming-ming cheng Shi-Min Hu 《Computational Visual Media》 SCIE EI CSCD 2022年第3期331-368,共38页
Humans can naturally and effectively find salient regions in complex scenes.Motivated by this observation,attention mechanisms were introduced into computer vision with the aim of imitating this aspect of the human vi... Humans can naturally and effectively find salient regions in complex scenes.Motivated by this observation,attention mechanisms were introduced into computer vision with the aim of imitating this aspect of the human visual system.Such an attention mechanism can be regarded as a dynamic weight adjustment process based on features of the input image.Attention mechanisms have achieved great success in many visual tasks,including image classification,object detection,semantic segmentation,video understanding,image generation,3D vision,multimodal tasks,and self-supervised learning.In this survey,we provide a comprehensive review of various attention mechanisms in computer vision and categorize them according to approach,such as channel attention,spatial attention,temporal attention,and branch attention;a related repository https://github.com/MenghaoG uo/Awesome-Vision-Attentions is dedicated to collecting related work.We also suggest future directions for attention mechanism research. 展开更多
关键词 ATTENTION TRANSFORMER computer vision deep learning salience
原文传递
Salient object detection: A survey 被引量:45
6
作者 Ali Borji ming-ming cheng +2 位作者 Qibin Hou Huaizu Jiang Jia Li 《Computational Visual Media》 CSCD 2019年第2期117-150,共34页
Detecting and segmenting salient objects from natural scenes, often referred to as salient object detection, has attracted great interest in computer vision. While many models have been proposed and several applicatio... Detecting and segmenting salient objects from natural scenes, often referred to as salient object detection, has attracted great interest in computer vision. While many models have been proposed and several applications have emerged, a deep understanding of achievements and issues remains lacking. We aim to provide a comprehensive review of recent progress in salient object detection and situate this field among other closely related areas such as generic scene segmentation, object proposal generation, and saliency for fixation prediction. Covering 228 publications, we survey i) roots, key concepts, and tasks, ii) core techniques and main modeling trends, and iii) datasets and evaluation metrics for salient object detection. We also discuss open problems such as evaluation metrics and dataset bias in model performance, and suggest future research directions. 展开更多
关键词 salient OBJECT detection SALIENCY visual ATTENTION REGIONS of INTEREST
原文传递
Intelligent Visual Media Processing: When Graphics Meets Vision 被引量:12
7
作者 ming-ming cheng Qi-Bin Hou +1 位作者 Song-Hai Zhang Paul L. Rosin 《Journal of Computer Science & Technology》 SCIE EI CSCD 2017年第1期110-121,共12页
The computer graphics and computer vision communities have been working closely together in recent years and a variety of algorithms and applications have been developed to analyze and manipulate the visual media arou... The computer graphics and computer vision communities have been working closely together in recent years and a variety of algorithms and applications have been developed to analyze and manipulate the visual media around us. There are three major driving forces behind this phenomenon: 1) the availability of big data from the Internet has created a demand for dealing with the ever-increasing, vast amount of resources; 2) powerful processing tools, such as deep neural networks, provide effective ways for learning how to deal with heterogeneous visual data; 3) new data capture devices, such as the Kilxect, the bridge betweea algorithms for 2D image understanding and 3D model analysis. These driving forces have emerged only recently, and we believe that the computer graphics and computer vision communities are still in the beginning of their honeymoon phase. In this work we survey recent research on how computer vision techniques benefit computer graphics techniques and vice versa, and cover research on analysis, manipulation, synthesis, and interaction. We also discuss existing problems and suggest possible further research directions. 展开更多
关键词 computer graphics computer vision SURVEY scene understanding image manipulation
原文传递
BING: Binarized normed gradients for objectness estimation at 300fps 被引量:12
8
作者 ming-ming cheng Yun Liu +3 位作者 Wen-Yan Lin Ziming Zhang Paul L.Rosin Philip H.S.Torr 《Computational Visual Media》 CSCD 2019年第1期3-20,共18页
Training a generic objectness measure to produce object proposals has recently become of significant interest. We observe that generic objects with well-defined closed boundaries can be detected by looking at the norm... Training a generic objectness measure to produce object proposals has recently become of significant interest. We observe that generic objects with well-defined closed boundaries can be detected by looking at the norm of gradients, with a suitable resizing of their corresponding image windows to a small fixed size. Based on this observation and computational reasons, we propose to resize the window to 8 × 8 and use the norm of the gradients as a simple 64 D feature to describe it, for explicitly training a generic objectness measure. We further show how the binarized version of this feature, namely binarized normed gradients(BING), can be used for efficient objectness estimation, which requires only a few atomic operations(e.g., add, bitwise shift, etc.). To improve localization quality of the proposals while maintaining efficiency, we propose a novel fast segmentation method and demonstrate its effectiveness for improving BING's localization performance, when used in multithresholding straddling expansion(MTSE) postprocessing. On the challenging PASCAL VOC2007 dataset, using 1000 proposals per image and intersectionover-union threshold of 0.5, our proposal method achieves a 95.6% object detection rate and 78.6% mean average best overlap in less than 0.005 second per image. 展开更多
关键词 OBJECT proposals objectness VISUAL ATTENTION CATEGORY agnostic proposals
原文传递
FLIC: Fast linear iterative clustering with active search 被引量:12
9
作者 Jiaxing Zhao Ren Bo +2 位作者 Qibin Hou ming-ming cheng Paul Rosin 《Computational Visual Media》 CSCD 2018年第4期333-348,共16页
In this paper, we reconsider the clustering problem for image over-segmentation from a new perspective. We propose a novel search algorithm called"active search" which explicitly considers neighbor continuit... In this paper, we reconsider the clustering problem for image over-segmentation from a new perspective. We propose a novel search algorithm called"active search" which explicitly considers neighbor continuity. Based on this search method, we design a back-and-forth traversal strategy and a joint assignment and update step to speed up the algorithm. Compared to earlier methods, such as simple linear iterative clustering(SLIC) and its variants, which use fixed search regions and perform the assignment and the update steps separately, our novel scheme reduces the number of iterations required for convergence,and also provides better boundaries in the oversegmentation results. Extensive evaluation using the Berkeley segmentation benchmark verifies that our method outperforms competing methods under various evaluation metrics. In particular, our method is fastest,achieving approximately 30 fps for a 481 × 321 image on a single CPU core. To facilitate further research, our code is made publicly available. 展开更多
关键词 image OVER-SEGMENTATION SLIC NEIGHBOR continuity back-and-forth TRAVERSAL
原文传递
RGB-D salient object detection:A survey 被引量:16
10
作者 Tao Zhou Deng-Ping Fan +2 位作者 ming-ming cheng Jianbing Shen Ling Shao 《Computational Visual Media》 EI CSCD 2021年第1期37-69,共33页
Salient object detection,which simulates human visual perception in locating the most significant object(s)in a scene,has been widely applied to various computer vision tasks.Now,the advent of depth sensors means that... Salient object detection,which simulates human visual perception in locating the most significant object(s)in a scene,has been widely applied to various computer vision tasks.Now,the advent of depth sensors means that depth maps can easily be captured;this additional spatial information can boost the performance of salient object detection.Although various RGB-D based salient object detection models with promising performance have been proposed over the past several years,an in-depth understanding of these models and the challenges in this field remains lacking.In this paper,we provide a comprehensive survey of RGBD based salient object detection models from various perspectives,and review related benchmark datasets in detail.Further,as light fields can also provide depth maps,we review salient object detection models and popular benchmark datasets from this domain too.Moreover,to investigate the ability of existing models to detect salient objects,we have carried out a comprehensive attribute-based evaluation of several representative RGB-D based salient object detection models.Finally,we discuss several challenges and open directions of RGB-D based salient object detection for future research.All collected models,benchmark datasets,datasets constructed for attribute-based evaluation,and related code are publicly available at https://github.com/taozh2017/RGBD-SODsurvey. 展开更多
关键词 RGB-D SALIENCY light fields benchmarks
原文传递
Joint salient object detection and existence prediction 被引量:4
11
作者 Huaizu JIANG ming-ming cheng +2 位作者 Shi-Jie LI Ali BORJI Jingdong WANG 《Frontiers of Computer Science》 SCIE EI CSCD 2019年第4期778-788,共11页
Recent advances in supervised salient object detection modeling has resulted in significant performance improvements on benchmark datasets. However, most of the existing salient object detection models assume that at ... Recent advances in supervised salient object detection modeling has resulted in significant performance improvements on benchmark datasets. However, most of the existing salient object detection models assume that at least one salient object exists in the input image. Such an assumption often leads to less appealing saliency maps on the background images with no salient object at all. Therefore, handling those cases can reduce the false positive rate of a model. In this paper, we propose a supervised learning approach for jointly addressing the salient object detection and existence prediction problems. Given a set of background-only images and images with salient objects, as well as their salient object annotations, we adopt the structural SVM framework and formulate the two problems jointly in a single integrated objective function: saliency labels of superpixels are involved in a classification term conditioned on the salient object existence variable, which in turn depends on both global image and regional saliency features and saliency labels assignments. The loss function also considers both image-level and regionlevel mis-classifications. Extensive evaluation on benchmark datasets validate the effectiveness of our proposed joint approach compared to the baseline and state-of-the-art models. 展开更多
关键词 salient object DETECTION EXISTENCE PREDICTION JOINT INFERENCE SALIENCY DETECTION
原文传递
Saliency Rank:Two-stage manifold ranking for salient object detection 被引量:5
12
作者 Wei Qi ming-ming cheng +2 位作者 Ali Borji Huchuan Lu Lian-Fa Bai 《Computational Visual Media》 2015年第4期309-320,共12页
Salient object detection remains one of the most important and active research topics in computer vision,with wide-ranging applications to object recognition,scene understanding,image retrieval,context aware image edi... Salient object detection remains one of the most important and active research topics in computer vision,with wide-ranging applications to object recognition,scene understanding,image retrieval,context aware image editing,image compression,etc. Most existing methods directly determine salient objects by exploring various salient object features.Here,we propose a novel graph based ranking method to detect and segment the most salient object in a scene according to its relationship to image border(background) regions,i.e.,the background feature.Firstly,we use regions/super-pixels as graph nodes,which are fully connected to enable both long range and short range relations to be modeled. The relationship of each region to the image border(background) is evaluated in two stages:(i) ranking with hard background queries,and(ii) ranking with soft foreground queries. We experimentally show how this two-stage ranking based salient object detection method is complementary to traditional methods,and that integrated results outperform both. Our method allows the exploitation of intrinsic image structure to achieve high quality salient object determination using a quadratic optimization framework,with a closed form solution which can be easily computed.Extensive method evaluation and comparison using three challenging saliency datasets demonstrate that our method consistently outperforms 10 state-of-theart models by a big margin. 展开更多
关键词 salient object detection manifold ranking visual attention SALIENCY
原文传递
Geometry-Aware ICP for Scene Reconstruction from RGB-D Camera 被引量:2
13
作者 Bo Ren Jia-cheng Wu +2 位作者 Ya-Lei Lv ming-ming cheng Shao-Ping Lu 《Journal of Computer Science & Technology》 SCIE EI CSCD 2019年第3期581-593,共13页
The Iterative Closest Point (ICP) scheme has been widely used for the registration of surfaces and point clouds.However, when working on depth image sequences where there are large geometric planes with small (or even... The Iterative Closest Point (ICP) scheme has been widely used for the registration of surfaces and point clouds.However, when working on depth image sequences where there are large geometric planes with small (or even without) details,existing ICP algorithms are prone to tangential drifting and erroneous rotational estimations due to input device errors.In this paper, we propose a novel ICP algorithm that aims to overcome such drawbacks, and provides significantly stabler registration estimation for simultaneous localization and mapping (SLAM) tasks on RGB-D camera inputs. In our approach,the tangential drifting and the rotational estimation error are reduced by: 1) updating the conventional Euclidean distance term with the local geometry information, and 2) introducing a new camera stabilization term that prevents improper camera movement in the calculation. Our approach is simple, fast, effective, and is readily integratable with previous ICP algorithms. We test our new method with the TUM RGB-D SLAM dataset on state-of-the-art real-time 3D dense reconstruction platforms, i.e., ElasticFusion and Kintinuous. Experiments show that our new strategy outperforms all previous ones on various RGB-D data sequences under different combinations of registration systems and solutions. 展开更多
关键词 ICP(iterative closest point) RGB-D tangential DRIFTING ROTATIONAL estimation COVARIANCE matrix
原文传递
S4Net: Single stage salient-instance segmentation 被引量:2
14
作者 Ruochen Fan ming-ming cheng +3 位作者 Qibin Hou Tai-Jiang Mu Jingdong Wang Shi-Min Hu 《Computational Visual Media》 CSCD 2020年第2期191-204,共14页
In this paper, we consider salient instance segmentation. As well as producing bounding boxes,our network also outputs high-quality instance-level segments as initial selections to indicate the regions of interest. Ta... In this paper, we consider salient instance segmentation. As well as producing bounding boxes,our network also outputs high-quality instance-level segments as initial selections to indicate the regions of interest. Taking into account the category-independent property of each target, we design a single stage salient instance segmentation framework, with a novel segmentation branch. Our new branch regards not only local context inside each detection window but also the surrounding context, enabling us to distinguish instances in the same scope even with partial occlusion.Our network is end-to-end trainable and is fast(running at 40 fps for images with resolution 320 × 320). We evaluate our approach on a publicly available benchmark and show that it outperforms alternative solutions. We also provide a thorough analysis of our design choices to help readers better understand the function of each part of our network. Source code can be found at https://github.com/Ruochen Fan/S4 Net. 展开更多
关键词 salient-instance segmentation salient object detection single stage region-of-interest masking
原文传递
Editorial for Special Issue on Multi-modal Representation Learning
15
作者 Deng-Ping Fan Nick Barnes +1 位作者 ming-ming cheng Luc Van Gool 《Machine Intelligence Research》 EI 2024年第4期615-616,共2页
The past decade has witnessed the impressive and steady development of single-modal AI technologies in several fields,thanks to the emergence of deep learning.Less studied,however,is multi-modal AI-commonly considered... The past decade has witnessed the impressive and steady development of single-modal AI technologies in several fields,thanks to the emergence of deep learning.Less studied,however,is multi-modal AI-commonly considered the next generation of AI-which utilizes complementary context concealed in different-modality inputs to improve performance.Humans naturally learn to form a global concept from multiple modalities(i.e.,sight,hearing,touch,smell,and taste),even when some are incomplete or missing.Thus,in addition to the two popular modalities(vision and language),other types of data such as depth,infrared information,and events are also important for multi-modal learning in real-world scenes. 展开更多
关键词 modal utilize incomplete
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部