The topographic information of a closed world is expressed as a graph. The plural mov- ingobjects which go and back in it according to a single moving strategy are supposed.The moving strategy is expressed by numerica...The topographic information of a closed world is expressed as a graph. The plural mov- ingobjects which go and back in it according to a single moving strategy are supposed.The moving strategy is expressed by numerical values as a decision table. Coding is performed with this table as chromosomes, and this is optimized by using genetic algorithm. These environments were realized on a computer, and the simulation was carried out. As the result, the learning of the method to act so that moving objects do not obstruct mutually was recognized, and it was confirmed that these methods are effective for optimizing moving strategy.展开更多
This article introduces a novel variational model for restoring images degraded by Cauchy noise and/or blurring.The model integrates a nonconvex data-fidelity term with two regularization terms,a sparse representation...This article introduces a novel variational model for restoring images degraded by Cauchy noise and/or blurring.The model integrates a nonconvex data-fidelity term with two regularization terms,a sparse representation prior over dictionary learning and total generalized variation(TGV)regularization.The sparse representation prior exploiting patch information enables the preservation of fine features and textural patterns,while adequately denoising in homogeneous regions and contributing natural visual quality.TGV regularization further assists in effectively denoising in smooth regions while retaining edges.By adopting the penalty method and an alternating minimization approach,we present an efficient iterative algorithm to solve the proposed model.Numerical results establish the superiority of the proposed model over other existing models in regard to visual quality and certain image quality assessments.展开更多
It is attractive to formulate problems in computer vision and related fields in term of probabilis- tic estimation where the probability models are defined over graphs, such as grammars. The graphical struc- tures, an...It is attractive to formulate problems in computer vision and related fields in term of probabilis- tic estimation where the probability models are defined over graphs, such as grammars. The graphical struc- tures, and the state variables defined over them, give a rich knowledge representation which can describe the complex structures of objects and images. The proba- bility distributions defined over the graphs capture the statistical variability of these structures. These proba- bility models can be learnt from training data with lim- ited amounts of supervision. But learning these models suffers from the difficulty of evaluating the normaliza- tion constant, or partition function, of the probability distributions which can be extremely computationally demanding. This paper shows that by placing bounds on the normalization constant we can obtain compu- rationally tractable approximations. Surprisingly, for certain choices of loss functions, we obtain many of the standard max-margin criteria used in support vector machines (SVMs) and hence we reduce the learning to standard machine learning methods. We show that many machine learning methods can be obtained in this way as approximations to probabilistic methods including multi-class max-margin, ordinal regression, max-margin Markov networks and parsers, multiple- instance learning, and latent SVM. We illustrate this work by computer vision applications including image labeling, object detection and localization, and motion estimation. We speculate that rained by using better bounds better results can be ob- and approximations.展开更多
文摘The topographic information of a closed world is expressed as a graph. The plural mov- ingobjects which go and back in it according to a single moving strategy are supposed.The moving strategy is expressed by numerical values as a decision table. Coding is performed with this table as chromosomes, and this is optimized by using genetic algorithm. These environments were realized on a computer, and the simulation was carried out. As the result, the learning of the method to act so that moving objects do not obstruct mutually was recognized, and it was confirmed that these methods are effective for optimizing moving strategy.
基金Miyoun Jung was supported by Hankuk University of Foreign Studies Research Fund and the NRF(2017R1A2B1005363)Myungjoo Kang was supported by the NRF(2015R1A15A1009350,2017R1A2A1A17069644).
文摘This article introduces a novel variational model for restoring images degraded by Cauchy noise and/or blurring.The model integrates a nonconvex data-fidelity term with two regularization terms,a sparse representation prior over dictionary learning and total generalized variation(TGV)regularization.The sparse representation prior exploiting patch information enables the preservation of fine features and textural patterns,while adequately denoising in homogeneous regions and contributing natural visual quality.TGV regularization further assists in effectively denoising in smooth regions while retaining edges.By adopting the penalty method and an alternating minimization approach,we present an efficient iterative algorithm to solve the proposed model.Numerical results establish the superiority of the proposed model over other existing models in regard to visual quality and certain image quality assessments.
文摘It is attractive to formulate problems in computer vision and related fields in term of probabilis- tic estimation where the probability models are defined over graphs, such as grammars. The graphical struc- tures, and the state variables defined over them, give a rich knowledge representation which can describe the complex structures of objects and images. The proba- bility distributions defined over the graphs capture the statistical variability of these structures. These proba- bility models can be learnt from training data with lim- ited amounts of supervision. But learning these models suffers from the difficulty of evaluating the normaliza- tion constant, or partition function, of the probability distributions which can be extremely computationally demanding. This paper shows that by placing bounds on the normalization constant we can obtain compu- rationally tractable approximations. Surprisingly, for certain choices of loss functions, we obtain many of the standard max-margin criteria used in support vector machines (SVMs) and hence we reduce the learning to standard machine learning methods. We show that many machine learning methods can be obtained in this way as approximations to probabilistic methods including multi-class max-margin, ordinal regression, max-margin Markov networks and parsers, multiple- instance learning, and latent SVM. We illustrate this work by computer vision applications including image labeling, object detection and localization, and motion estimation. We speculate that rained by using better bounds better results can be ob- and approximations.