期刊文献+
共找到2篇文章
< 1 >
每页显示 20 50 100
Unsupervised image translation with distributional semantics awareness
1
作者 Zhexi Peng He Wang +2 位作者 Yanlin Weng Yin Yang tianjia shao 《Computational Visual Media》 SCIE EI CSCD 2023年第3期619-631,共13页
Unsupervised image translation(UIT)studies the mapping between two image domains.Since such mappings are under-constrained,existing research has pursued various desirable properties such as distributional matching or ... Unsupervised image translation(UIT)studies the mapping between two image domains.Since such mappings are under-constrained,existing research has pursued various desirable properties such as distributional matching or two-way consistency.In this paper,we re-examine UIT from a new perspective:distributional semantics consistency,based on the observation that data variations contain semantics,e.g.,shoes varying in colors.Further,the semantics can be multi-dimensional,e.g.,shoes also varying in style,functionality,etc.Given two image domains,matching these semantic dimensions during UIT will produce mappings with explicable correspondences,which has not been investigated previously.We propose distributional semantics mapping(DSM),the first UIT method which explicitly matches semantics between two domains.We show that distributional semantics has been rarely considered within and beyond UIT,even though it is a common problem in deep learning.We evaluate DSM on several benchmark datasets,demonstrating its general ability to capture distributional semantics.Extensive comparisons show that DSM not only produces explicable mappings,but also improves image quality in general. 展开更多
关键词 generative adversarial networks(GANs) manifold alignment unsupervised learning image-to-image translation distributional semantics
原文传递
An interactive approach for functional prototype recovery from a single RGBD image
2
作者 Yuliang Rong Youyi Zheng +2 位作者 tianjia shao Yin Yang Kun Zhou 《Computational Visual Media》 2016年第1期87-96,共10页
Inferring the functionality of an object from a single RGBD image is difficult for two reasons:lack of semantic information about the object, and missing data due to occlusion. In this paper, we present an interactive... Inferring the functionality of an object from a single RGBD image is difficult for two reasons:lack of semantic information about the object, and missing data due to occlusion. In this paper, we present an interactive framework to recover a 3D functional prototype from a single RGBD image.Instead of precisely reconstructing the object geometry for the prototype, we mainly focus on recovering the object's functionality along with its geometry.Our system allows users to scribble on the image to create initial rough proxies for the parts. After user annotation of high-level relations between parts, our system automatically jointly optimizes detailed joint parameters(axis and position) and part geometry parameters(size, orientation, and position). Such prototype recovery enables a better understanding of the underlying image geometry and allows for further physically plausible manipulation. We demonstrate our framework on various indoor objects with simple or hybrid functions. 展开更多
关键词 FUNCTIONALITY CUBOID PROXY PROTOTYPE part RELATIONS shape analysis
原文传递
上一页 1 下一页 到第
使用帮助 返回顶部