期刊文献+
共找到21篇文章
< 1 2 >
每页显示 20 50 100
Multi-Stream Temporally Enhanced Network for Video Salient Object Detection
1
作者 Dan Xu Jiale Ru Jinlong Shi 《Computers, Materials & Continua》 SCIE EI 2024年第1期85-104,共20页
Video salient object detection(VSOD)aims at locating the most attractive objects in a video by exploring the spatial and temporal features.VSOD poses a challenging task in computer vision,as it involves processing com... Video salient object detection(VSOD)aims at locating the most attractive objects in a video by exploring the spatial and temporal features.VSOD poses a challenging task in computer vision,as it involves processing complex spatial data that is also influenced by temporal dynamics.Despite the progress made in existing VSOD models,they still struggle in scenes of great background diversity within and between frames.Additionally,they encounter difficulties related to accumulated noise and high time consumption during the extraction of temporal features over a long-term duration.We propose a multi-stream temporal enhanced network(MSTENet)to address these problems.It investigates saliency cues collaboration in the spatial domain with a multi-stream structure to deal with the great background diversity challenge.A straightforward,yet efficient approach for temporal feature extraction is developed to avoid the accumulative noises and reduce time consumption.The distinction between MSTENet and other VSOD methods stems from its incorporation of both foreground supervision and background supervision,facilitating enhanced extraction of collaborative saliency cues.Another notable differentiation is the innovative integration of spatial and temporal features,wherein the temporal module is integrated into the multi-stream structure,enabling comprehensive spatial-temporal interactions within an end-to-end framework.Extensive experimental results demonstrate that the proposed method achieves state-of-the-art performance on five benchmark datasets while maintaining a real-time speed of 27 fps(Titan XP).Our code and models are available at https://github.com/RuJiaLe/MSTENet. 展开更多
关键词 video salient object detection deep learning temporally enhanced foreground-background collaboration
下载PDF
SwinVid:Enhancing Video Object Detection Using Swin Transformer
2
作者 Abdelrahman Maharek Amr Abozeid +1 位作者 Rasha Orban Kamal ElDahshan 《Computer Systems Science & Engineering》 2024年第2期305-320,共16页
What causes object detection in video to be less accurate than it is in still images?Because some video frames have degraded in appearance from fast movement,out-of-focus camera shots,and changes in posture.These reas... What causes object detection in video to be less accurate than it is in still images?Because some video frames have degraded in appearance from fast movement,out-of-focus camera shots,and changes in posture.These reasons have made video object detection(VID)a growing area of research in recent years.Video object detection can be used for various healthcare applications,such as detecting and tracking tumors in medical imaging,monitoring the movement of patients in hospitals and long-term care facilities,and analyzing videos of surgeries to improve technique and training.Additionally,it can be used in telemedicine to help diagnose and monitor patients remotely.Existing VID techniques are based on recurrent neural networks or optical flow for feature aggregation to produce reliable features which can be used for detection.Some of those methods aggregate features on the full-sequence level or from nearby frames.To create feature maps,existing VID techniques frequently use Convolutional Neural Networks(CNNs)as the backbone network.On the other hand,Vision Transformers have outperformed CNNs in various vision tasks,including object detection in still images and image classification.We propose in this research to use Swin-Transformer,a state-of-the-art Vision Transformer,as an alternative to CNN-based backbone networks for object detection in videos.The proposed architecture enhances the accuracy of existing VID methods.The ImageNet VID and EPIC KITCHENS datasets are used to evaluate the suggested methodology.We have demonstrated that our proposed method is efficient by achieving 84.3%mean average precision(mAP)on ImageNet VID using less memory in comparison to other leading VID techniques.The source code is available on the website https://github.com/amaharek/SwinVid. 展开更多
关键词 video object detection vision transformers convolutional neural networks deep learning
下载PDF
Motion connectivity-based initial video object extraction
3
作者 王煜坚 吴镇扬 《Journal of Southeast University(English Edition)》 EI CAS 2007年第4期500-506,共7页
In order to obtain the initial video objects from the video sequences, an improved initial video object extraction algorithm based on motion connectivity is proposed. Moving objects in video sequences are highly conne... In order to obtain the initial video objects from the video sequences, an improved initial video object extraction algorithm based on motion connectivity is proposed. Moving objects in video sequences are highly connected and structured, which makes motion connectivity an advanced feature for segmentation. Accordingly, after sharp noise elimination, the cumulated difference image, which exhibits the coherent motion of the moving object, is adaptively thresholded. Then the maximal connected region is labeled, post-processed and output as the final segmenting mask. Hence the initial video object is effectively extracted. Comparative experimental results show that the proposed algorithm extracts the initial video object automatically, promptly and properly, thereby achieving satisfactory subjective and objective performance. 展开更多
关键词 video object extraction motion connectivity adaptive threshold cumulated difference image
下载PDF
Scribble-Supervised Video Object Segmentation 被引量:3
4
作者 Peiliang Huang Junwei Han +2 位作者 Nian Liu Jun Ren Dingwen Zhang 《IEEE/CAA Journal of Automatica Sinica》 SCIE EI CSCD 2022年第2期339-353,共15页
Recently,video object segmentation has received great attention in the computer vision community.Most of the existing methods heavily rely on the pixel-wise human annotations,which are expensive and time-consuming to ... Recently,video object segmentation has received great attention in the computer vision community.Most of the existing methods heavily rely on the pixel-wise human annotations,which are expensive and time-consuming to obtain.To tackle this problem,we make an early attempt to achieve video object segmentation with scribble-level supervision,which can alleviate large amounts of human labor for collecting the manual annotation.However,using conventional network architectures and learning objective functions under this scenario cannot work well as the supervision information is highly sparse and incomplete.To address this issue,this paper introduces two novel elements to learn the video object segmentation model.The first one is the scribble attention module,which captures more accurate context information and learns an effective attention map to enhance the contrast between foreground and background.The other one is the scribble-supervised loss,which can optimize the unlabeled pixels and dynamically correct inaccurate segmented areas during the training stage.To evaluate the proposed method,we implement experiments on two video object segmentation benchmark datasets,You Tube-video object segmentation(VOS),and densely annotated video segmentation(DAVIS)-2017.We first generate the scribble annotations from the original per-pixel annotations.Then,we train our model and compare its test performance with the baseline models and other existing works.Extensive experiments demonstrate that the proposed method can work effectively and approach to the methods requiring the dense per-pixel annotations. 展开更多
关键词 Convolutional neural networks(CNNs) SCRIBBLE self-attention video object segmentation weakly supervised
下载PDF
Objective Performance Evaluation of Video Segmentation Algorithms with Ground-Truth 被引量:1
5
作者 杨高波 张兆扬 《Journal of Shanghai University(English Edition)》 CAS 2004年第1期70-74,共5页
While the development of particular video segmentation algorithms has attracted considerable research interest, relatively little effort has been devoted to provide a methodology for evaluating their performance. In t... While the development of particular video segmentation algorithms has attracted considerable research interest, relatively little effort has been devoted to provide a methodology for evaluating their performance. In this paper, we propose a methodology to objectively evaluate video segmentation algorithm with ground-truth, which is based on computing the deviation of segmentation results from the reference segmentation. Four different metrics based on classification pixels, edges, relative foreground area and relative position respectively are combined to address the spatial accuracy. Temporal coherency is evaluated by utilizing the difference of spatial accuracy between successive frames. The experimental results show the feasibility of our approach. Moreover, it is computationally more efficient than previous methods. It can be applied to provide an offline ranking among different segmentation algorithms and to optimally set the parameters for a given algorithm. 展开更多
关键词 video object segmentation performance evaluation MPEG-4.
下载PDF
Research on video motion object segmentation for content-based application
6
作者 包红强 ZHANG Zhao- yang +4 位作者 YU Song-yu WANG Suo-zhong WANG Nu-li FANG Yong WANG Zhi-gang 《Journal of Shanghai University(English Edition)》 CAS 2006年第2期142-143,共2页
With the development of the modern information society, more and more multimedia information is available. So the technology of multimedia processing is becoming the important task for the irrelevant area of scientist... With the development of the modern information society, more and more multimedia information is available. So the technology of multimedia processing is becoming the important task for the irrelevant area of scientist. Among of the multimedia, the visual informarion is more attractive due to its direct, vivid characteristic, but at the same rime the huge amount of video data causes many challenges if the video storage, processing and transmission. 展开更多
关键词 image processing video object segmentation spatiotemporal framework MPEG-4.
下载PDF
Initial Object Segmentation for Video Object Plane Generation Using Cellular Neural Networks
7
作者 王慧 杨高波 张兆扬 《Journal of Shanghai University(English Edition)》 CAS 2003年第2期168-172,共5页
MPEG 4 is a basic tool for interactivity and manipulation of video sequences. Video object segmentation is a key issue in defining the content of any video sequence, which is often divided into two steps: initial obj... MPEG 4 is a basic tool for interactivity and manipulation of video sequences. Video object segmentation is a key issue in defining the content of any video sequence, which is often divided into two steps: initial object segmentation and object tracking. In this paper, an initial object segmentation method for video object plane(VOP) generation using color information is proposed. Based on 3 by 3 linear templates, a cellular neural network (CNN) is used to implemented object segmentation. The Experimental results are presented to verify the efficiency and robustness of this approach. 展开更多
关键词 video object plane(VOP) cellular neural networks(CNN) templates.
下载PDF
AUTOMATIC SEGMENTATION OF VIDEO OBJECT PLANES IN MPEG-4 BASED ON SPATIO-TEMPORAL INFORMATION
8
作者 XiaJinxiang HuangShunji 《Journal of Electronics(China)》 2004年第3期206-212,共7页
Segmentation of semantic Video Object Planes (VOP's) from video sequence is a key to the standard MPEG-4 with content-based video coding. In this paper, the approach of automatic Segmentation of VOP's Based on... Segmentation of semantic Video Object Planes (VOP's) from video sequence is a key to the standard MPEG-4 with content-based video coding. In this paper, the approach of automatic Segmentation of VOP's Based on Spatio-Temporal Information (SBSTI) is proposed.The proceeding results demonstrate the good performance of the algorithm. 展开更多
关键词 video sequence segmentation video Object Plane (VOP) Based on spatiotemporal information MPEG-4
下载PDF
VIDEO OBJECT SEGMENTATION BY 2-D MESH-BASED MOTION ANALYSIS
9
作者 Wang Yujian Gao Jianpo Yang Hao Wu Zhenyang 《Journal of Electronics(China)》 2007年第5期668-673,共6页
Video object extraction is a key technology in content-based video coding.A novel video object extracting algorithm by two Dimensional (2-D) mesh-based motion analysis is proposed in this paper.Firstly,a 2-D mesh fitt... Video object extraction is a key technology in content-based video coding.A novel video object extracting algorithm by two Dimensional (2-D) mesh-based motion analysis is proposed in this paper.Firstly,a 2-D mesh fitting the original frame image is obtained via feature detection algorithm. Then,higher order statistics motion analysis is applied on the 2-D mesh representation to get an initial motion detection mask.After post-processing,the final segmenting mask is quickly obtained.And hence the video object is effectively extracted.Experimental results show that the proposed algorithm combines the merits of mesh-based segmenting algorithms and pixel-based segmenting algorithms,and hereby achieves satisfactory subjective and objective performance while dramatically increasing the segmenting speed. 展开更多
关键词 video object extraction 2-D mesh Higher order statistics
下载PDF
A Novel Divide and Conquer Solution for Long-term Video Salient Object Detection
10
作者 Yun-Xiao Li Cheng-Li-Zhao Chen +2 位作者 Shuai Li Ai-Min Hao Hong Qin 《Machine Intelligence Research》 EI CSCD 2024年第4期684-703,共20页
Recently,a new research trend in our video salient object detection(VSOD)research community has focused on enhancing the detection results via model self-fine-tuning using sparsely mined high-quality keyframes from th... Recently,a new research trend in our video salient object detection(VSOD)research community has focused on enhancing the detection results via model self-fine-tuning using sparsely mined high-quality keyframes from the given sequence.Although such a learning scheme is generally effective,it has a critical limitation,i.e.,the model learned on sparse frames only possesses weak generalization ability.This situation could become worse on“long”videos since they tend to have intensive scene variations.Moreover,in such videos,the keyframe information from a longer time span is less relevant to the previous,which could also cause learning conflict and deteriorate the model performance.Thus,the learning scheme is usually incapable of handling complex pattern modeling.To solve this problem,we propose a divide-and-conquer framework,which can convert a complex problem domain into multiple simple ones.First,we devise a novel background consistency analysis(BCA)which effectively divides the mined frames into disjoint groups.Then for each group,we assign an individual deep model on it to capture its key attribute during the fine-tuning phase.During the testing phase,we design a model-matching strategy,which could dynamically select the best-matched model from those fine-tuned ones to handle the given testing frame.Comprehensive experiments show that our method can adapt severe background appearance variation coupling with object movement and obtain robust saliency detection compared with the previous scheme and the state-of-the-art methods. 展开更多
关键词 video salient object detection background consistency analysis weakly supervised learning long-term information background shift.
原文传递
Evaluating quality of motion for unsupervised video object segmentation
11
作者 CHENG Guanjun SONG Huihui 《Optoelectronics Letters》 EI 2024年第6期379-384,共6页
Current mainstream unsupervised video object segmentation(UVOS) approaches typically incorporate optical flow as motion information to locate the primary objects in coherent video frames. However, they fuse appearance... Current mainstream unsupervised video object segmentation(UVOS) approaches typically incorporate optical flow as motion information to locate the primary objects in coherent video frames. However, they fuse appearance and motion information without evaluating the quality of the optical flow. When poor-quality optical flow is used for the interaction with the appearance information, it introduces significant noise and leads to a decline in overall performance. To alleviate this issue, we first employ a quality evaluation module(QEM) to evaluate the optical flow. Then, we select high-quality optical flow as motion cues to fuse with the appearance information, which can prevent poor-quality optical flow from diverting the network's attention. Moreover, we design an appearance-guided fusion module(AGFM) to better integrate appearance and motion information. Extensive experiments on several widely utilized datasets, including DAVIS-16, FBMS-59, and You Tube-Objects, demonstrate that the proposed method outperforms existing methods. 展开更多
关键词 Evaluating quality of motion for unsupervised video object segmentation
原文传递
基于残差密集孪生网络的视频目标跟踪 被引量:1
12
作者 付利华 王路远 +3 位作者 章海涛 闫绍兴 吴会贤 王俊翔 《北京工业大学学报》 CAS CSCD 北大核心 2022年第9期944-951,共8页
针对现有基于孪生网络的视频目标跟踪(video object tracking, VOT)方法存在的特征提取能力不足以及对外观变化过大或平面外旋转等目标跟踪效果不佳的问题,提出一种基于残差密集孪生网络的VOT方法.首先,使用嵌入卷积注意力的残差密集网... 针对现有基于孪生网络的视频目标跟踪(video object tracking, VOT)方法存在的特征提取能力不足以及对外观变化过大或平面外旋转等目标跟踪效果不佳的问题,提出一种基于残差密集孪生网络的VOT方法.首先,使用嵌入卷积注意力的残差密集网络对模板帧图像和检测帧图像分别提取不同层次的特征;然后,将不同层次的特征通过相互独立的区域候选网络进行互相关操作;最后,将多个区域候选网络的输出自适应加权求和,得到最终的跟踪结果.实验结果表明,该方法在应对目标表观变化过大、平面外旋转等挑战时,能够获得较好的跟踪效果. 展开更多
关键词 孪生网络 通道注意力 空间注意力 残差密集网络 视频目标跟踪(video object tracking VOT) 区域候选网络
下载PDF
Full-duplex strategy for video object segmentation 被引量:1
13
作者 Ge-Peng Ji Deng-Ping Fan +3 位作者 Keren Fu Zhe Wu Jianbing Shen Ling Shao 《Computational Visual Media》 SCIE EI CSCD 2023年第1期155-175,共21页
Previous video object segmentation approachesmainly focus on simplex solutions linking appearance and motion,limiting effective feature collaboration between these two cues.In this work,we study a novel and efficient ... Previous video object segmentation approachesmainly focus on simplex solutions linking appearance and motion,limiting effective feature collaboration between these two cues.In this work,we study a novel and efficient full-duplex strategy network(FSNet)to address this issue,by considering a better mutual restraint scheme linking motion and appearance allowing exploitation of cross-modal features from the fusion and decoding stage.Specifically,we introduce a relational cross-attention module(RCAM)to achieve bidirectional message propagation across embedding sub-spaces.To improve the model’s robustness and update inconsistent features from the spatiotemporal embeddings,we adopt a bidirectional purification module after the RCAM.Extensive experiments on five popular benchmarks show that our FSNet is robust to various challenging scenarios(e.g.,motion blur and occlusion),and compares well to leading methods both for video object segmentation and video salient object detection.The project is publicly available at https://github.com/GewelsJI/FSNet. 展开更多
关键词 video object segmentation(VOS) video salient object detection(V-SOD) visual attention
原文传递
An Object-Based Fast Motion Estimation Algorithm in MPEG-4
14
作者 孙雷 张文军 +1 位作者 余松煜 刘勋 《Journal of Shanghai Jiaotong university(Science)》 EI 2006年第3期340-345,共6页
This paper presented an object-based fast motion estimation (ME) algorithm for object-based texture coding in moving picture experts group four (MPEG-4), which takes full advantage of the shape information of video ob... This paper presented an object-based fast motion estimation (ME) algorithm for object-based texture coding in moving picture experts group four (MPEG-4), which takes full advantage of the shape information of video object. Compared with the full search (FS) algorithm, the proposed algorithm can significantly speed the ME process. The speed of ME using the proposed algorithm is faster than that using new three-step search (NTSS), four-step search (4SS), diamond search (DS), and block-based gradient descent search (BBGDS) algorithms with similar motion compensation (MC) errors. The proposed algorithm can be combined with other fast ME algorithm to make the ME process faster. 展开更多
关键词 MPEG 4 motion estimation video object
下载PDF
Recent Advances and Challenges of Visual Signal Quality Assessment 被引量:1
15
作者 马林 邓宸伟 +1 位作者 颜庆义 林维斯 《China Communications》 SCIE CSCD 2013年第5期62-78,共17页
While quality assessment is essential for testing, optimizing, benchmarking, monitoring, and inspecting related systems and services, it also plays an essential role in the design of virtually all visual signal proces... While quality assessment is essential for testing, optimizing, benchmarking, monitoring, and inspecting related systems and services, it also plays an essential role in the design of virtually all visual signal processing and communication algorithms, as well as various related decision-making processes. In this paper, we first provide an overview of recently derived quality assessment approaches for traditional visual signals (i.e., 2D images/videos), with highlights for new trends (such as machine learning approaches). On the other hand, with the ongoing development of devices and multimedia services, newly emerged visual signals (e.g., mobile/3D videos) are becoming more and more popular. This work focuses on recent progresses of quality metrics, which have been reviewed for the newly emerged forms of visual signals, which include scalable and mobile videos, High Dynamic Range (HDR) images, image segmentation results, 3D images/videos, and retargeted images. 展开更多
关键词 objective quality assessment 2D images and videos human perception newly emerged visual signals Human Visual System
下载PDF
A Lossy Contour-Based Representation of Segmentation Maps Using Quadratic Transformations Motion Prediction
16
作者 Ying Zhang Zhao-yang Zhang 《Advances in Manufacturing》 2000年第3期214-219,共6页
In the context of object oriented video coding, the encoding of segmentation maps defined by contour networks is particularly critical. In this paper, we present a lossy contour network encoding algorithm where both t... In the context of object oriented video coding, the encoding of segmentation maps defined by contour networks is particularly critical. In this paper, we present a lossy contour network encoding algorithm where both the rate distortion contour encoding based on maximum operator and the prediction error for the current frame based on quadratic motion model are combined into a optimal polygon contour network compression scheme. The bit rate for the contour network can be further reduced by about 20% in comparison with that in the optimal polygonal boundary encoding scheme using maximum operator in the rate distortion sense. 展开更多
关键词 contour network encoding in the rate distortion sense very low bit rate object oriented video coding
下载PDF
Global video object segmentation with spatial constraint module
17
作者 Yadang Chen Duolin Wang +2 位作者 Zhiguo Chen Zhi-Xin Yang Enhua Wu 《Computational Visual Media》 SCIE EI CSCD 2023年第2期385-400,共16页
We present a lightweight and efficient semisupervised video object segmentation network based on the space-time memory framework.To some extent,our method solves the two difficulties encountered in traditional video o... We present a lightweight and efficient semisupervised video object segmentation network based on the space-time memory framework.To some extent,our method solves the two difficulties encountered in traditional video object segmentation:one is that the single frame calculation time is too long,and the other is that the current frame’s segmentation should use more information from past frames.The algorithm uses a global context(GC)module to achieve highperformance,real-time segmentation.The GC module can effectively integrate multi-frame image information without increased memory and can process each frame in real time.Moreover,the prediction mask of the previous frame is helpful for the segmentation of the current frame,so we input it into a spatial constraint module(SCM),which constrains the areas of segments in the current frame.The SCM effectively alleviates mismatching of similar targets yet consumes few additional resources.We added a refinement module to the decoder to improve boundary segmentation.Our model achieves state-of-the-art results on various datasets,scoring 80.1%on YouTube-VOS 2018 and a J&F score of 78.0%on DAVIS 2017,while taking 0.05 s per frame on the DAVIS 2016 validation dataset. 展开更多
关键词 video object segmentation semantic segmentation global context(GC)module spatial constraint
原文传递
E-GrabCut: an economic method of iterative video object extraction 被引量:1
18
作者 Le DONG Ning FENG +2 位作者 Mengdie MAO Ling HE Jingjing WANG 《Frontiers of Computer Science》 SCIE EI CSCD 2017年第4期649-660,共12页
Efficient, interactive foreground/background seg- mentation in video is of great practical importance in video editing. This paper proposes an interactive and unsupervised video object segmentation algorithm named E-G... Efficient, interactive foreground/background seg- mentation in video is of great practical importance in video editing. This paper proposes an interactive and unsupervised video object segmentation algorithm named E-GrabCut con- centrating on achieving both of the segmentation quality and time efficiency as highly demanded in the related filed. There are three features in the proposed algorithms. Firstly, we have developed a powerful, non-iterative version of the optimiza- tion process for each frame. Secondly, more user interaction in the first frame is used to improve the Gaussian Mixture Model (GMM). Thirdly, a robust algorithm for the follow- ing frame segmentation has been developed by reusing the previous GMM. Extensive experiments demonstrate that our method outperforms the state-of-the-art video segmentation algorithm in terms of integration of time efficiency and seg- mentation quality. 展开更多
关键词 interactive video object extraction video seg-mentation GRABCUT GMM
原文传递
A RNN-based objective video quality measurement 被引量:1
19
作者 黄轩 张荣 庞建新 《Chinese Optics Letters》 SCIE EI CAS CSCD 2009年第11期1004-1006,共3页
Technology used to automatically assess video quality plays a significant role in video processing areas. Because of the complexity of video media, there are great limitations to assess video quality with only one fac... Technology used to automatically assess video quality plays a significant role in video processing areas. Because of the complexity of video media, there are great limitations to assess video quality with only one factor. We propose a new method using artificial random neural networks (RNNs) with motion evaluation as an estimation of perceived visual distortion. The results are obtained through a nonlinear fitting procedure and well correlated with human perception. Compared with other methods, the proposed method performs more adaptable and accurate predictions. 展开更多
关键词 RNN SSIM A RNN-based objective video quality measurement PSNR DMOS
原文传递
Unsupervised object detection with scene-adaptive concept learning 被引量:2
20
作者 Shiliang PU Wei ZHAO +3 位作者 Weijie CHEN Shicai YANG Di XIE Yunhe PAN 《Frontiers of Information Technology & Electronic Engineering》 SCIE EI CSCD 2021年第5期638-651,共14页
Object detection is one of the hottest research directions in computer vision,has already made impressive progress in academia,and has many valuable applications in the industry.However,the mainstream detection method... Object detection is one of the hottest research directions in computer vision,has already made impressive progress in academia,and has many valuable applications in the industry.However,the mainstream detection methods still have two shortcomings:(1)even a model that is well trained using large amounts of data still cannot generally be used across different kinds of scenes;(2)once a model is deployed,it cannot autonomously evolve along with the accumulated unlabeled scene data.To address these problems,and inspired by visual knowledge theory,we propose a novel scene-adaptive evolution unsupervised video object detection algorithm that can decrease the impact of scene changes through the concept of object groups.We first extract a large number of object proposals from unlabeled data through a pre-trained detection model.Second,we build the visual knowledge dictionary of object concepts by clustering the proposals,in which each cluster center represents an object prototype.Third,we look into the relations between different clusters and the object information of different groups,and propose a graph-based group information propagation strategy to determine the category of an object concept,which can effectively distinguish positive and negative proposals.With these pseudo labels,we can easily fine-tune the pretrained model.The effectiveness of the proposed method is verified by performing different experiments,and the significant improvements are achieved. 展开更多
关键词 Visual knowledge Unsupervised video object detection Scene-adaptive learning
原文传递
上一页 1 2 下一页 到第
使用帮助 返回顶部