期刊文献+
共找到2篇文章
< 1 >
每页显示 20 50 100
Dual Frequency Transformer for Efficient SDR-to-HDR Translation
1
作者 Gang Xu Qibin Hou Ming-Ming Cheng 《Machine Intelligence Research》 EI CSCD 2024年第3期538-548,共11页
The SDR-to-HDR translation technique can convert the abundant standard-dynamic-range (SDR) media resources to high-dynamic-range (HDR) ones, which can represent high-contrast scenes, providing more realistic visual ex... The SDR-to-HDR translation technique can convert the abundant standard-dynamic-range (SDR) media resources to high-dynamic-range (HDR) ones, which can represent high-contrast scenes, providing more realistic visual experiences. While recent vision Transformers have achieved promising performance in many low-level vision tasks, there are few works attempting to leverage Transformers for SDR-to-HDR translation. In this paper, we are among the first to investigate the performance of Transformers for SDR-to-HDR translation. We find that directly using the self-attention mechanism may involve artifacts in the results due to the inappropriate way to model long-range dependencies between the low-frequency and high-frequency components. Taking this into account, we advance the self-attention mechanism and present a dual frequency attention (DFA), which leverages the self-attention mechanism to separately encode the low-frequency structural information and high-frequency detail information. Based on the proposed DFA, we further design a multi-scale feature fusion network, named dual frequency Transformer (DFT), for efficient SDR-to-HDR translation. Extensive experiments on the HDRTV1K dataset demonstrate that our DFT can achieve better quantitative and qualitative performance than the recent state-of-the-art methods. The code of our DFT is made publicly available at https://github.com/CS-GangXu/DFT. 展开更多
关键词 Standard-dynamic-range to high-dynamic-range(SDR-to-HDR)translation TRANSFORMER dual frequency attention(DFA) frequency-aware feature decomposition efficient model.
原文传递
Efficient image representation for object recognition via pivots selection 被引量:3
2
作者 Bojun XIE Yi LIU +1 位作者 HuiZHANG Jian YU 《Frontiers of Computer Science》 SCIE EI CSCD 2015年第3期383-391,共9页
Patch-level features are essential for achieving good performance in computer vision tasks. Besides well- known pre-defined patch-level descriptors such as scalein- variant feature transform (SIFT) and histogram of ... Patch-level features are essential for achieving good performance in computer vision tasks. Besides well- known pre-defined patch-level descriptors such as scalein- variant feature transform (SIFT) and histogram of oriented gradient (HOG), the kernel descriptor (KD) method [1] of- fers a new way to "grow-up" features from a match-kernel defined over image patch pairs using kernel principal compo- nent analysis (KPCA) and yields impressive results. In this paper, we present efficient kernel descriptor (EKD) and efficient hierarchical kernel descriptor (EHKD), which are built upon incomplete Cholesky decomposition. EKD au- tomatically selects a small number of pivot features for gener- ating patch-level features to achieve better computational effi- ciency. EHKD recursively applies EKD to form image-level features layer-by-layer. Perhaps due to parsimony, we find surprisingly that the EKD and EHKD approaches achieved competitive results on several public datasets compared with other state-of-the-art methods, at an improved efficiency over KD. 展开更多
关键词 efficient kernel descriptor efficient hierarchi-cal kernel descriptor incomplete Cholesky decomposition patch-level features image-level features
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部