Memory deficit,which is often associated with aging and many psychiatric,neurological,and neurodegenerative diseases,has been a challenging issue for treatment.Up till now,all potential drug candidates have failed to ...Memory deficit,which is often associated with aging and many psychiatric,neurological,and neurodegenerative diseases,has been a challenging issue for treatment.Up till now,all potential drug candidates have failed to produce satisfa ctory effects.Therefore,in the search for a solution,we found that a treatment with the gene corresponding to the RGS14414protein in visual area V2,a brain area connected with brain circuits of the ventral stream and the medial temporal lobe,which is crucial for object recognition memory(ORM),can induce enhancement of ORM.In this study,we demonstrated that the same treatment with RGS14414in visual area V2,which is relatively unaffected in neurodegenerative diseases such as Alzheimer s disease,produced longlasting enhancement of ORM in young animals and prevent ORM deficits in rodent models of aging and Alzheimer’s disease.Furthermore,we found that the prevention of memory deficits was mediated through the upregulation of neuronal arbo rization and spine density,as well as an increase in brain-derived neurotrophic factor(BDNF).A knockdown of BDNF gene in RGS14414-treated aging rats and Alzheimer s disease model mice caused complete loss in the upregulation of neuronal structural plasticity and in the prevention of ORM deficits.These findings suggest that BDNF-mediated neuronal structural plasticity in area V2 is crucial in the prevention of memory deficits in RGS14414-treated rodent models of aging and Alzheimer’s disease.Therefore,our findings of RGS14414gene-mediated activation of neuronal circuits in visual area V2 have therapeutic relevance in the treatment of memory deficits.展开更多
For some important object recognition applications such as intelligent robots and unmanned driving, images are collected on a consecutive basis and associated among themselves, besides, the scenes have steady prior fe...For some important object recognition applications such as intelligent robots and unmanned driving, images are collected on a consecutive basis and associated among themselves, besides, the scenes have steady prior features. Yet existing technologies do not take full advantage of this information. In order to take object recognition further than existing algorithms in the above application, an object recognition method that fuses temporal sequence with scene priori information is proposed. This method first employs YOLOv3 as the basic algorithm to recognize objects in single-frame images, then the DeepSort algorithm to establish association among potential objects recognized in images of different moments, and finally the confidence fusion method and temporal boundary processing method designed herein to fuse, at the decision level, temporal sequence information with scene priori information. Experiments using public datasets and self-built industrial scene datasets show that due to the expansion of information sources, the quality of single-frame images has less impact on the recognition results, whereby the object recognition is greatly improved. It is presented herein as a widely applicable framework for the fusion of information under multiple classes. All the object recognition algorithms that output object class, location information and recognition confidence at the same time can be integrated into this information fusion framework to improve performance.展开更多
Objective We aimed to investigate whether antagonism of the cannabinoid CB1 receptor(CB1R)could affect novel object recognition(NOR)memory in chronically rapid eye movement sleep-deprived(RSD)rats.Methods The animals ...Objective We aimed to investigate whether antagonism of the cannabinoid CB1 receptor(CB1R)could affect novel object recognition(NOR)memory in chronically rapid eye movement sleep-deprived(RSD)rats.Methods The animals were examined for recognition memory following a 7-day chronic partial RSD paradigm using the multiple platform technique.The CB1R antagonist rimonabant(1 or 3 mg/kg,i.p.)was administered either at one hour prior to the sample phase for acquisition,or immediately after the sample phase for consolidation,or at one hour before the test phase for retrieval of NOR memory.For the reconsolidation task,rimonabant was administered immediately after the second sample phase.Results The RSD episode impaired acquisition,consolidation,and retrieval,but it did not affect the reconsolidation of NOR memory.Rimonabant administration did not affect acquisition,consolidation,and reconsolidation;however,it attenuated impairment of the retrieval of NOR memory induced by chronic RSD.Conclusions These findings,along with our previous report,would seem to suggest that RSD may affect different phases of recognition memory based on its duration.Importantly,it seems that the CB1R may,at least in part,be involved in the adverse effects of chronic RSD on the retrieval,but not in the acquisition,consolidation,and reconsolidation,of NOR memory.展开更多
In rice production,the prevention and management of pests and diseases have always received special attention.Traditional methods require human experts,which is costly and time-consuming.Due to the complexity of the s...In rice production,the prevention and management of pests and diseases have always received special attention.Traditional methods require human experts,which is costly and time-consuming.Due to the complexity of the structure of rice diseases and pests,quickly and reliably recognizing and locating them is difficult.Recently,deep learning technology has been employed to detect and identify rice diseases and pests.This paper introduces common publicly available datasets;summarizes the applications on rice diseases and pests from the aspects of image recognition,object detection,image segmentation,attention mechanism,and few-shot learning methods according to the network structure differences;and compares the performances of existing studies.Finally,the current issues and challenges are explored fromthe perspective of data acquisition,data processing,and application,providing possible solutions and suggestions.This study aims to review various DL models and provide improved insight into DL techniques and their cutting-edge progress in the prevention and management of rice diseases and pests.展开更多
Recently,there has been a notable surge of interest in scientific research regarding spectral images.The potential of these images to revolutionize the digital photography industry,like aerial photography through Unma...Recently,there has been a notable surge of interest in scientific research regarding spectral images.The potential of these images to revolutionize the digital photography industry,like aerial photography through Unmanned Aerial Vehicles(UAVs),has captured considerable attention.One encouraging aspect is their combination with machine learning and deep learning algorithms,which have demonstrated remarkable outcomes in image classification.As a result of this powerful amalgamation,the adoption of spectral images has experienced exponential growth across various domains,with agriculture being one of the prominent beneficiaries.This paper presents an extensive survey encompassing multispectral and hyperspectral images,focusing on their applications for classification challenges in diverse agricultural areas,including plants,grains,fruits,and vegetables.By meticulously examining primary studies,we delve into the specific agricultural domains where multispectral and hyperspectral images have found practical use.Additionally,our attention is directed towards utilizing machine learning techniques for effectively classifying hyperspectral images within the agricultural context.The findings of our investigation reveal that deep learning and support vector machines have emerged as widely employed methods for hyperspectral image classification in agriculture.Nevertheless,we also shed light on the various issues and limitations of working with spectral images.This comprehensive analysis aims to provide valuable insights into the current state of spectral imaging in agriculture and its potential for future advancements.展开更多
Holograms provide a characteristic manner to display and convey information, and have been improved to provide better user interactions Holographic interactions are important as they improve user interactions with vir...Holograms provide a characteristic manner to display and convey information, and have been improved to provide better user interactions Holographic interactions are important as they improve user interactions with virtual objects. Gesture interaction is a recent research topic, as it allows users to use their bare hands to directly interact with the hologram. However, it remains unclear whether real hand gestures are well suited for hologram applications. Therefore, we discuss the development process and implementation of three-dimensional object manipulation using natural hand gestures in a hologram. We describe the design and development process for hologram applications and its integration with real hand gesture interactions as initial findings. Experimental results from Nasa TLX form are discussed. Based on the findings, we actualize the user interactions in the hologram.展开更多
Object recognition and tracking are two of the most dynamic research sub-areas that belong to the field of Computer Vision.Computer vision is one of the most active research fields that lies at the intersection of dee...Object recognition and tracking are two of the most dynamic research sub-areas that belong to the field of Computer Vision.Computer vision is one of the most active research fields that lies at the intersection of deep learning and machine vision.This paper presents an efficient ensemble algorithm for the recognition and tracking of fixed shapemoving objects while accommodating the shift and scale invariances that the object may encounter.The first part uses the Maximum Average Correlation Height(MACH)filter for object recognition and determines the bounding box coordinates.In case the correlation based MACH filter fails,the algorithms switches to a much reliable but computationally complex feature based object recognition technique i.e.,affine scale invariant feature transform(ASIFT).ASIFT is used to accommodate object shift and scale object variations.ASIFT extracts certain features from the object of interest,providing invariance in up to six affine parameters,namely translation(two parameters),zoom,rotation and two camera axis orientations.However,in this paper,only the shift and scale invariances are used.The second part of the algorithm demonstrates the use of particle filters based Approximate Proximal Gradient(APG)technique to periodically update the coordinates of the object encapsulated in the bounding box.At the end,a comparison of the proposed algorithm with other stateof-the-art tracking algorithms has been presented,which demonstrates the effectiveness of the proposed algorithm with respect to the minimization of tracking errors.展开更多
This paper discusses recognition of three dimensional (3D) moving object from multiple views, which is based on 2D processed frames of a video sequence, view categories (feature aspects) of object, and their transiti...This paper discusses recognition of three dimensional (3D) moving object from multiple views, which is based on 2D processed frames of a video sequence, view categories (feature aspects) of object, and their transitions. Log polar mapping (LPM) and discrete Fourier transformation (DFM) are used to obtain position, scale and rotation invariant feature vectors of 2D characteristic views. ART 2 model is used as memory and classifier of the feature information of the object. ART 2 neural network is improved in experiment with satisfactory results.展开更多
Object recognition and location has always been one of the research hotspots in machine vision.It is of great value and significance to the development and application of current service robots,industrial automation,u...Object recognition and location has always been one of the research hotspots in machine vision.It is of great value and significance to the development and application of current service robots,industrial automation,unmanned driving and other fields.In order to realize the real-time recognition and location of indoor scene objects,this article proposes an improved YOLOv3 neural network model,which combines densely connected networks and residual networks to construct a new YOLOv3 backbone network,which is applied to the detection and recognition of objects in indoor scenes.In this article,RealSense D415 RGB-D camera is used to obtain the RGB map and depth map,the actual distance value is calculated after each pixel in the scene image is mapped to the real scene.Experiment results proved that the detection and recognition accuracy and real-time performance by the new network are obviously improved compared with the previous YOLOV3 neural network model in the same scene.More objects can be detected after the improvement of network which cannot be detected with the YOLOv3 network before the improvement.The running time of objects detection and recognition is reduced to less than half of the original.This improved network has a certain reference value for practical engineering application.展开更多
Ocean underwater exploration is a part of oceanography that investigates the physical and biological conditions for scientific and commercial purposes. And video technology plays an important role and is extensively a...Ocean underwater exploration is a part of oceanography that investigates the physical and biological conditions for scientific and commercial purposes. And video technology plays an important role and is extensively applied for underwater environment observation. Different from the conventional methods, video technology explores the underwater ecosystem continuously and non-invasively. However, due to the scattering and attenuation of light transport in the water, complex noise distribution and lowlight condition cause challenges for underwater video applications including object detection and recognition. In this paper, we propose a new deep encoding-decoding convolutional architecture for underwater object recognition. It uses the deep encoding-decoding network for extracting the discriminative features from the noisy low-light underwater images. To create the deconvolutional layers for classification, we apply the deconvolution kernel with a matched feature map, instead of full connection, to solve the problem of dimension disaster and low accuracy. Moreover, we introduce data augmentation and transfer learning technologies to solve the problem of data starvation. For experiments, we investigated the public datasets with our proposed method and the state-of-the-art methods. The results show that our work achieves significant accuracy. This work provides new underwater technologies applied for ocean exploration.展开更多
Machine learning is an integral technology many people utilize in all areas of human life. It is pervasive in modern living worldwide, and has multiple usages. One application is image classification, embraced across ...Machine learning is an integral technology many people utilize in all areas of human life. It is pervasive in modern living worldwide, and has multiple usages. One application is image classification, embraced across many spheres of influence such as business, finance, medicine, etc. to enhance produces, causes, efficiency, etc. This need for more accurate, detail-oriented classification increases the need for modifications, adaptations, and innovations to Deep Learning Algorithms. This article used Convolutional Neural Networks (CNN) to classify scenes in the CIFAR-10 database, and detect emotions in the KDEF database. The proposed method converted the data to the wavelet domain to attain greater accuracy and comparable efficiency to the spatial domain processing. By dividing image data into subbands, important feature learning occurred over differing low to high frequencies. The combination of the learned low and high frequency features, and processing the fused feature mapping resulted in an advance in the detection accuracy. Comparing the proposed methods to spatial domain CNN and Stacked Denoising Autoencoder (SDA), experimental findings revealed a substantial increase in accuracy.展开更多
Advances in machine vision systems have revolutionized applications such as autonomous driving,robotic navigation,and augmented reality.Despite substantial progress,challenges persist,including dynamic backgrounds,occ...Advances in machine vision systems have revolutionized applications such as autonomous driving,robotic navigation,and augmented reality.Despite substantial progress,challenges persist,including dynamic backgrounds,occlusion,and limited labeled data.To address these challenges,we introduce a comprehensive methodology toenhance image classification and object detection accuracy.The proposed approach involves the integration ofmultiple methods in a complementary way.The process commences with the application of Gaussian filters tomitigate the impact of noise interference.These images are then processed for segmentation using Fuzzy C-Meanssegmentation in parallel with saliency mapping techniques to find the most prominent regions.The Binary RobustIndependent Elementary Features(BRIEF)characteristics are then extracted fromdata derived fromsaliency mapsand segmented images.For precise object separation,Oriented FAST and Rotated BRIEF(ORB)algorithms areemployed.Genetic Algorithms(GAs)are used to optimize Random Forest classifier parameters which lead toimproved performance.Our method stands out due to its comprehensive approach,adeptly addressing challengessuch as changing backdrops,occlusion,and limited labeled data concurrently.A significant enhancement hasbeen achieved by integrating Genetic Algorithms(GAs)to precisely optimize parameters.This minor adjustmentnot only boosts the uniqueness of our system but also amplifies its overall efficacy.The proposed methodologyhas demonstrated notable classification accuracies of 90.9%and 89.0%on the challenging Corel-1k and MSRCdatasets,respectively.Furthermore,detection accuracies of 87.2%and 86.6%have been attained.Although ourmethod performed well in both datasets it may face difficulties in real-world data especially where datasets havehighly complex backgrounds.Despite these limitations,GAintegration for parameter optimization shows a notablestrength in enhancing the overall adaptability and performance of our system.展开更多
Due to the characteristics of high resolution and rich texture information,visible light images are widely used for maritime ship detection.However,these images are suscep-tible to sea fog and ships of different sizes...Due to the characteristics of high resolution and rich texture information,visible light images are widely used for maritime ship detection.However,these images are suscep-tible to sea fog and ships of different sizes,which can result in missed detections and false alarms,ultimately resulting in lower detection accuracy.To address these issues,a novel multi-granularity feature enhancement network,MFENet,which includes a three-way dehazing module(3WDM)and a multi-granularity feature enhancement module(MFEM)is proposed.The 3WDM eliminates sea fog interference by using an image clarity automatic classification algorithm based on three-way decisions and FFA-Net to obtain clear image samples.Additionally,the MFEM improves the accuracy of detecting ships of different sizes by utilising an improved super-resolution reconstruction con-volutional neural network to enhance the resolution and semantic representation capa-bility of the feature maps from YOLOv7.Experimental results demonstrate that MFENet surpasses the other 15 competing models in terms of the mean Average Pre-cision metric on two benchmark datasets,achieving 96.28%on the McShips dataset and 97.71%on the SeaShips dataset.展开更多
The digital twin is the concept of transcending reality,which is the reverse feedback from the real physical space to the virtual digital space.People hold great prospects for this emerging technology.In order to real...The digital twin is the concept of transcending reality,which is the reverse feedback from the real physical space to the virtual digital space.People hold great prospects for this emerging technology.In order to realize the upgrading of the digital twin industrial chain,it is urgent to introduce more modalities,such as vision,haptics,hearing and smell,into the virtual digital space,which assists physical entities and virtual objects in creating a closer connection.Therefore,perceptual understanding and object recognition have become an urgent hot topic in the digital twin.Existing surface material classification schemes often achieve recognition through machine learning or deep learning in a single modality,ignoring the complementarity between multiple modalities.In order to overcome this dilemma,we propose a multimodal fusion network in our article that combines two modalities,visual and haptic,for surface material recognition.On the one hand,the network makes full use of the potential correlations between multiple modalities to deeply mine the modal semantics and complete the data mapping.On the other hand,the network is extensible and can be used as a universal architecture to include more modalities.Experiments show that the constructed multimodal fusion network can achieve 99.42%classification accuracy while reducing complexity.展开更多
Automatically detecting and locating remote occlusion small objects from the images of complex traffic environments is a valuable and challenging research.Since the boundary box location is not sufficiently accurate a...Automatically detecting and locating remote occlusion small objects from the images of complex traffic environments is a valuable and challenging research.Since the boundary box location is not sufficiently accurate and it is difficult to distinguish overlapping and occluded objects,the authors propose a network model with a second-order term attention mechanism and occlusion loss.First,the backbone network is built on CSPDarkNet53.Then a method is designed for the feature extraction network based on an item-wise attention mechanism,which uses the filtered weighted feature vector to replace the original residual fusion and adds a second-order term to reduce the information loss in the process of fusion and accelerate the convergence of the model.Finally,an objected occlusion regression loss function is studied to reduce the problems of missed detections caused by dense objects.Sufficient experimental results demonstrate that the authors’method achieved state-of-the-art performance without reducing the detection speed.The mAP@.5 of the method is 85.8%on the Foggy_cityscapes dataset and the mAP@.5 of the method is 97.8%on the KITTI dataset.展开更多
Humans can perceive our complex world through multi-sensory fusion.Under limited visual conditions,people can sense a variety of tactile signals to identify objects accurately and rapidly.However,replicating this uniq...Humans can perceive our complex world through multi-sensory fusion.Under limited visual conditions,people can sense a variety of tactile signals to identify objects accurately and rapidly.However,replicating this unique capability in robots remains a significant challenge.Here,we present a new form of ultralight multifunctional tactile nano-layered carbon aerogel sensor that provides pressure,temperature,material recognition and 3D location capabilities,which is combined with multimodal supervised learning algorithms for object recognition.The sensor exhibits human-like pressure(0.04–100 kPa)and temperature(21.5–66.2℃)detection,millisecond response times(11 ms),a pressure sensitivity of 92.22 kPa^(−1)and triboelectric durability of over 6000 cycles.The devised algorithm has universality and can accommodate a range of application scenarios.The tactile system can identify common foods in a kitchen scene with 94.63%accuracy and explore the topographic and geomorphic features of a Mars scene with 100%accuracy.This sensing approach empowers robots with versatile tactile perception to advance future society toward heightened sensing,recognition and intelligence.展开更多
In this paper,we first derive two types of transformed Franklin polynomial:substituted and weighted radial Franklin polynomials.Two radial orthogonal moments are proposed based on these two types of polynomials,namely...In this paper,we first derive two types of transformed Franklin polynomial:substituted and weighted radial Franklin polynomials.Two radial orthogonal moments are proposed based on these two types of polynomials,namely substituted Franklin-Fourier moments and weighted Franklin-Fourier moments(SFFMs and WFFMs),which are orthogonal in polar coordinates.The radial kernel functions of SFFMs and WFFMs are transformed Franklin functions and Franklin functions are composed of a class of complete orthogonal splines function system of degree one.Therefore,it provides the possibility of avoiding calculating high order polynomials,and thus the accurate values of SFFMs and WFFMs can be obtained directly with little computational cost.Theoretical and experimental results show that Franklin functions are not well suited for constructing higher-order moments of SFFMs and WFFMs,but compared with traditional orthogonal moments(e.g.,BFMs,OFMs and ZMs)in polar coordinates,the proposed two types of Franklin-Fourier Moments have better performance respectively in lower-order moments.展开更多
In order to accomplish the task of object recognition in natural scenes,a new object recognition algorithm based on an improved convolutional neural network(CNN)is proposed.First,candidate object windows are extracted...In order to accomplish the task of object recognition in natural scenes,a new object recognition algorithm based on an improved convolutional neural network(CNN)is proposed.First,candidate object windows are extracted from the original image.Then,candidate object windows are input into the improved CNN model to obtain deep features.Finally,the deep features are input into the Softmax and the confidence scores of classes are obtained.The candidate object window with the highest confidence score is selected as the object recognition result.Based on AlexNet,Inception V1 is introduced into the improved CNN and the fully connected layer is replaced by the average pooling layer,which widens the network and deepens the network at the same time.Experimental results show that the improved object recognition algorithm can obtain better recognition results in multiple natural scene images,and has a higher degree of accuracy than the classical algorithms in the field of object recognition.展开更多
In this Paper, a classification method based on neural networks is presented for recognition of 3D objects. Indeed, the objective of this paper is to classify an object query against objects in a database, which leads...In this Paper, a classification method based on neural networks is presented for recognition of 3D objects. Indeed, the objective of this paper is to classify an object query against objects in a database, which leads to recognition of the former. 3D objects of this database are transformations of other objects by one element of the overall transformation. The set of transformations considered in this work is the general affine group.展开更多
基金supported by grants from the Ministerio de Economia y Competitividad(BFU2013-43458-R)Junta de Andalucia(P12-CTS-1694 and Proyexcel-00422)to ZUK。
文摘Memory deficit,which is often associated with aging and many psychiatric,neurological,and neurodegenerative diseases,has been a challenging issue for treatment.Up till now,all potential drug candidates have failed to produce satisfa ctory effects.Therefore,in the search for a solution,we found that a treatment with the gene corresponding to the RGS14414protein in visual area V2,a brain area connected with brain circuits of the ventral stream and the medial temporal lobe,which is crucial for object recognition memory(ORM),can induce enhancement of ORM.In this study,we demonstrated that the same treatment with RGS14414in visual area V2,which is relatively unaffected in neurodegenerative diseases such as Alzheimer s disease,produced longlasting enhancement of ORM in young animals and prevent ORM deficits in rodent models of aging and Alzheimer’s disease.Furthermore,we found that the prevention of memory deficits was mediated through the upregulation of neuronal arbo rization and spine density,as well as an increase in brain-derived neurotrophic factor(BDNF).A knockdown of BDNF gene in RGS14414-treated aging rats and Alzheimer s disease model mice caused complete loss in the upregulation of neuronal structural plasticity and in the prevention of ORM deficits.These findings suggest that BDNF-mediated neuronal structural plasticity in area V2 is crucial in the prevention of memory deficits in RGS14414-treated rodent models of aging and Alzheimer’s disease.Therefore,our findings of RGS14414gene-mediated activation of neuronal circuits in visual area V2 have therapeutic relevance in the treatment of memory deficits.
文摘For some important object recognition applications such as intelligent robots and unmanned driving, images are collected on a consecutive basis and associated among themselves, besides, the scenes have steady prior features. Yet existing technologies do not take full advantage of this information. In order to take object recognition further than existing algorithms in the above application, an object recognition method that fuses temporal sequence with scene priori information is proposed. This method first employs YOLOv3 as the basic algorithm to recognize objects in single-frame images, then the DeepSort algorithm to establish association among potential objects recognized in images of different moments, and finally the confidence fusion method and temporal boundary processing method designed herein to fuse, at the decision level, temporal sequence information with scene priori information. Experiments using public datasets and self-built industrial scene datasets show that due to the expansion of information sources, the quality of single-frame images has less impact on the recognition results, whereby the object recognition is greatly improved. It is presented herein as a widely applicable framework for the fusion of information under multiple classes. All the object recognition algorithms that output object class, location information and recognition confidence at the same time can be integrated into this information fusion framework to improve performance.
基金Supported by the Research Council of Kermanshah University of Medical Sciences,Kermanshah,Iran for financial support(grant no.:990812).
文摘Objective We aimed to investigate whether antagonism of the cannabinoid CB1 receptor(CB1R)could affect novel object recognition(NOR)memory in chronically rapid eye movement sleep-deprived(RSD)rats.Methods The animals were examined for recognition memory following a 7-day chronic partial RSD paradigm using the multiple platform technique.The CB1R antagonist rimonabant(1 or 3 mg/kg,i.p.)was administered either at one hour prior to the sample phase for acquisition,or immediately after the sample phase for consolidation,or at one hour before the test phase for retrieval of NOR memory.For the reconsolidation task,rimonabant was administered immediately after the second sample phase.Results The RSD episode impaired acquisition,consolidation,and retrieval,but it did not affect the reconsolidation of NOR memory.Rimonabant administration did not affect acquisition,consolidation,and reconsolidation;however,it attenuated impairment of the retrieval of NOR memory induced by chronic RSD.Conclusions These findings,along with our previous report,would seem to suggest that RSD may affect different phases of recognition memory based on its duration.Importantly,it seems that the CB1R may,at least in part,be involved in the adverse effects of chronic RSD on the retrieval,but not in the acquisition,consolidation,and reconsolidation,of NOR memory.
基金funded by Hunan Provincial Natural Science Foundation of China with Grant Numbers(2022JJ50016,2023JJ50096)Innovation Platform Open Fund of Hengyang Normal University Grant 2021HSKFJJ039Hengyang Science and Technology Plan Guiding Project with Number 202222025902.
文摘In rice production,the prevention and management of pests and diseases have always received special attention.Traditional methods require human experts,which is costly and time-consuming.Due to the complexity of the structure of rice diseases and pests,quickly and reliably recognizing and locating them is difficult.Recently,deep learning technology has been employed to detect and identify rice diseases and pests.This paper introduces common publicly available datasets;summarizes the applications on rice diseases and pests from the aspects of image recognition,object detection,image segmentation,attention mechanism,and few-shot learning methods according to the network structure differences;and compares the performances of existing studies.Finally,the current issues and challenges are explored fromthe perspective of data acquisition,data processing,and application,providing possible solutions and suggestions.This study aims to review various DL models and provide improved insight into DL techniques and their cutting-edge progress in the prevention and management of rice diseases and pests.
文摘Recently,there has been a notable surge of interest in scientific research regarding spectral images.The potential of these images to revolutionize the digital photography industry,like aerial photography through Unmanned Aerial Vehicles(UAVs),has captured considerable attention.One encouraging aspect is their combination with machine learning and deep learning algorithms,which have demonstrated remarkable outcomes in image classification.As a result of this powerful amalgamation,the adoption of spectral images has experienced exponential growth across various domains,with agriculture being one of the prominent beneficiaries.This paper presents an extensive survey encompassing multispectral and hyperspectral images,focusing on their applications for classification challenges in diverse agricultural areas,including plants,grains,fruits,and vegetables.By meticulously examining primary studies,we delve into the specific agricultural domains where multispectral and hyperspectral images have found practical use.Additionally,our attention is directed towards utilizing machine learning techniques for effectively classifying hyperspectral images within the agricultural context.The findings of our investigation reveal that deep learning and support vector machines have emerged as widely employed methods for hyperspectral image classification in agriculture.Nevertheless,we also shed light on the various issues and limitations of working with spectral images.This comprehensive analysis aims to provide valuable insights into the current state of spectral imaging in agriculture and its potential for future advancements.
文摘Holograms provide a characteristic manner to display and convey information, and have been improved to provide better user interactions Holographic interactions are important as they improve user interactions with virtual objects. Gesture interaction is a recent research topic, as it allows users to use their bare hands to directly interact with the hologram. However, it remains unclear whether real hand gestures are well suited for hologram applications. Therefore, we discuss the development process and implementation of three-dimensional object manipulation using natural hand gestures in a hologram. We describe the design and development process for hologram applications and its integration with real hand gesture interactions as initial findings. Experimental results from Nasa TLX form are discussed. Based on the findings, we actualize the user interactions in the hologram.
基金This research was supported by X-mind Corps program of National Research Foundation of Korea(NRF)funded by the Ministry of Science,ICT(No.2019H1D8A1105622)and the Soonchunhyang University Research Fund.
文摘Object recognition and tracking are two of the most dynamic research sub-areas that belong to the field of Computer Vision.Computer vision is one of the most active research fields that lies at the intersection of deep learning and machine vision.This paper presents an efficient ensemble algorithm for the recognition and tracking of fixed shapemoving objects while accommodating the shift and scale invariances that the object may encounter.The first part uses the Maximum Average Correlation Height(MACH)filter for object recognition and determines the bounding box coordinates.In case the correlation based MACH filter fails,the algorithms switches to a much reliable but computationally complex feature based object recognition technique i.e.,affine scale invariant feature transform(ASIFT).ASIFT is used to accommodate object shift and scale object variations.ASIFT extracts certain features from the object of interest,providing invariance in up to six affine parameters,namely translation(two parameters),zoom,rotation and two camera axis orientations.However,in this paper,only the shift and scale invariances are used.The second part of the algorithm demonstrates the use of particle filters based Approximate Proximal Gradient(APG)technique to periodically update the coordinates of the object encapsulated in the bounding box.At the end,a comparison of the proposed algorithm with other stateof-the-art tracking algorithms has been presented,which demonstrates the effectiveness of the proposed algorithm with respect to the minimization of tracking errors.
文摘This paper discusses recognition of three dimensional (3D) moving object from multiple views, which is based on 2D processed frames of a video sequence, view categories (feature aspects) of object, and their transitions. Log polar mapping (LPM) and discrete Fourier transformation (DFM) are used to obtain position, scale and rotation invariant feature vectors of 2D characteristic views. ART 2 model is used as memory and classifier of the feature information of the object. ART 2 neural network is improved in experiment with satisfactory results.
基金supported by Henan Province Science and Technology Project under Grant No.182102210065.
文摘Object recognition and location has always been one of the research hotspots in machine vision.It is of great value and significance to the development and application of current service robots,industrial automation,unmanned driving and other fields.In order to realize the real-time recognition and location of indoor scene objects,this article proposes an improved YOLOv3 neural network model,which combines densely connected networks and residual networks to construct a new YOLOv3 backbone network,which is applied to the detection and recognition of objects in indoor scenes.In this article,RealSense D415 RGB-D camera is used to obtain the RGB map and depth map,the actual distance value is calculated after each pixel in the scene image is mapped to the real scene.Experiment results proved that the detection and recognition accuracy and real-time performance by the new network are obviously improved compared with the previous YOLOV3 neural network model in the same scene.More objects can be detected after the improvement of network which cannot be detected with the YOLOv3 network before the improvement.The running time of objects detection and recognition is reduced to less than half of the original.This improved network has a certain reference value for practical engineering application.
基金supported by the Jilin Science and Technology Development Plan Project (Nos. 20160209006GX, 20170309001GX and 20180201043GX)
文摘Ocean underwater exploration is a part of oceanography that investigates the physical and biological conditions for scientific and commercial purposes. And video technology plays an important role and is extensively applied for underwater environment observation. Different from the conventional methods, video technology explores the underwater ecosystem continuously and non-invasively. However, due to the scattering and attenuation of light transport in the water, complex noise distribution and lowlight condition cause challenges for underwater video applications including object detection and recognition. In this paper, we propose a new deep encoding-decoding convolutional architecture for underwater object recognition. It uses the deep encoding-decoding network for extracting the discriminative features from the noisy low-light underwater images. To create the deconvolutional layers for classification, we apply the deconvolution kernel with a matched feature map, instead of full connection, to solve the problem of dimension disaster and low accuracy. Moreover, we introduce data augmentation and transfer learning technologies to solve the problem of data starvation. For experiments, we investigated the public datasets with our proposed method and the state-of-the-art methods. The results show that our work achieves significant accuracy. This work provides new underwater technologies applied for ocean exploration.
文摘Machine learning is an integral technology many people utilize in all areas of human life. It is pervasive in modern living worldwide, and has multiple usages. One application is image classification, embraced across many spheres of influence such as business, finance, medicine, etc. to enhance produces, causes, efficiency, etc. This need for more accurate, detail-oriented classification increases the need for modifications, adaptations, and innovations to Deep Learning Algorithms. This article used Convolutional Neural Networks (CNN) to classify scenes in the CIFAR-10 database, and detect emotions in the KDEF database. The proposed method converted the data to the wavelet domain to attain greater accuracy and comparable efficiency to the spatial domain processing. By dividing image data into subbands, important feature learning occurred over differing low to high frequencies. The combination of the learned low and high frequency features, and processing the fused feature mapping resulted in an advance in the detection accuracy. Comparing the proposed methods to spatial domain CNN and Stacked Denoising Autoencoder (SDA), experimental findings revealed a substantial increase in accuracy.
基金a grant from the Basic Science Research Program through the National Research Foundation(NRF)(2021R1F1A1063634)funded by the Ministry of Science and ICT(MSIT)Republic of Korea.This research is supported and funded by Princess Nourah bint Abdulrahman University Researchers Supporting Project Number(PNURSP2024R410)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.The authors are thankful to the Deanship of Scientific Research at Najran University for funding this work under the Research Group Funding program Grant Code(NU/RG/SERC/12/6).
文摘Advances in machine vision systems have revolutionized applications such as autonomous driving,robotic navigation,and augmented reality.Despite substantial progress,challenges persist,including dynamic backgrounds,occlusion,and limited labeled data.To address these challenges,we introduce a comprehensive methodology toenhance image classification and object detection accuracy.The proposed approach involves the integration ofmultiple methods in a complementary way.The process commences with the application of Gaussian filters tomitigate the impact of noise interference.These images are then processed for segmentation using Fuzzy C-Meanssegmentation in parallel with saliency mapping techniques to find the most prominent regions.The Binary RobustIndependent Elementary Features(BRIEF)characteristics are then extracted fromdata derived fromsaliency mapsand segmented images.For precise object separation,Oriented FAST and Rotated BRIEF(ORB)algorithms areemployed.Genetic Algorithms(GAs)are used to optimize Random Forest classifier parameters which lead toimproved performance.Our method stands out due to its comprehensive approach,adeptly addressing challengessuch as changing backdrops,occlusion,and limited labeled data concurrently.A significant enhancement hasbeen achieved by integrating Genetic Algorithms(GAs)to precisely optimize parameters.This minor adjustmentnot only boosts the uniqueness of our system but also amplifies its overall efficacy.The proposed methodologyhas demonstrated notable classification accuracies of 90.9%and 89.0%on the challenging Corel-1k and MSRCdatasets,respectively.Furthermore,detection accuracies of 87.2%and 86.6%have been attained.Although ourmethod performed well in both datasets it may face difficulties in real-world data especially where datasets havehighly complex backgrounds.Despite these limitations,GAintegration for parameter optimization shows a notablestrength in enhancing the overall adaptability and performance of our system.
基金National Key Research and Development Program of China,Grant/Award Number:2022YFB3104700National Natural Science Foundation of China,Grant/Award Numbers:62376198,61906137,62076040,62076182,62163016,62006172+1 种基金The China National Scientific Sea‐floor Observatory,The Natural Science Foundation of Shanghai,Grant/Award Number:22ZR1466700The Jiangxi Provincial Natural Science Fund,Grant/Award Number:20212ACB202001。
文摘Due to the characteristics of high resolution and rich texture information,visible light images are widely used for maritime ship detection.However,these images are suscep-tible to sea fog and ships of different sizes,which can result in missed detections and false alarms,ultimately resulting in lower detection accuracy.To address these issues,a novel multi-granularity feature enhancement network,MFENet,which includes a three-way dehazing module(3WDM)and a multi-granularity feature enhancement module(MFEM)is proposed.The 3WDM eliminates sea fog interference by using an image clarity automatic classification algorithm based on three-way decisions and FFA-Net to obtain clear image samples.Additionally,the MFEM improves the accuracy of detecting ships of different sizes by utilising an improved super-resolution reconstruction con-volutional neural network to enhance the resolution and semantic representation capa-bility of the feature maps from YOLOv7.Experimental results demonstrate that MFENet surpasses the other 15 competing models in terms of the mean Average Pre-cision metric on two benchmark datasets,achieving 96.28%on the McShips dataset and 97.71%on the SeaShips dataset.
基金the National Natural Science Foundation of China(62001246,62001248,62171232)Key R&D Program of Jiangsu Province Key project and topics under Grant BE2021095+3 种基金the Natural Science Foundation of Jiangsu Province Higher Education Institutions(20KJB510020)the Future Network Scientific Research Fund Project(FNSRFP-2021-YB-16)the open research fund of Key Lab of Broadband Wireless Communication and Sensor Network Technology(JZNY202110)the NUPTSF under Grant(NY220070).
文摘The digital twin is the concept of transcending reality,which is the reverse feedback from the real physical space to the virtual digital space.People hold great prospects for this emerging technology.In order to realize the upgrading of the digital twin industrial chain,it is urgent to introduce more modalities,such as vision,haptics,hearing and smell,into the virtual digital space,which assists physical entities and virtual objects in creating a closer connection.Therefore,perceptual understanding and object recognition have become an urgent hot topic in the digital twin.Existing surface material classification schemes often achieve recognition through machine learning or deep learning in a single modality,ignoring the complementarity between multiple modalities.In order to overcome this dilemma,we propose a multimodal fusion network in our article that combines two modalities,visual and haptic,for surface material recognition.On the one hand,the network makes full use of the potential correlations between multiple modalities to deeply mine the modal semantics and complete the data mapping.On the other hand,the network is extensible and can be used as a universal architecture to include more modalities.Experiments show that the constructed multimodal fusion network can achieve 99.42%classification accuracy while reducing complexity.
基金Doctoral Talent Training Project of Chongqing University of Posts and Telecommunications,Grant/Award Number:BYJS202007Natural Science Foundation of Chongqing,Grant/Award Number:cstc2021jcyj-msxmX0941+1 种基金National Natural Science Foundation of China,Grant/Award Number:62176034Scientific and Technological Research Program of Chongqing Municipal Education Commission,Grant/Award Number:KJQN202101901。
文摘Automatically detecting and locating remote occlusion small objects from the images of complex traffic environments is a valuable and challenging research.Since the boundary box location is not sufficiently accurate and it is difficult to distinguish overlapping and occluded objects,the authors propose a network model with a second-order term attention mechanism and occlusion loss.First,the backbone network is built on CSPDarkNet53.Then a method is designed for the feature extraction network based on an item-wise attention mechanism,which uses the filtered weighted feature vector to replace the original residual fusion and adds a second-order term to reduce the information loss in the process of fusion and accelerate the convergence of the model.Finally,an objected occlusion regression loss function is studied to reduce the problems of missed detections caused by dense objects.Sufficient experimental results demonstrate that the authors’method achieved state-of-the-art performance without reducing the detection speed.The mAP@.5 of the method is 85.8%on the Foggy_cityscapes dataset and the mAP@.5 of the method is 97.8%on the KITTI dataset.
基金the National Natural Science Foundation of China(Grant No.52072041)the Beijing Natural Science Foundation(Grant No.JQ21007)+2 种基金the University of Chinese Academy of Sciences(Grant No.Y8540XX2D2)the Robotics Rhino-Bird Focused Research Project(No.2020-01-002)the Tencent Robotics X Laboratory.
文摘Humans can perceive our complex world through multi-sensory fusion.Under limited visual conditions,people can sense a variety of tactile signals to identify objects accurately and rapidly.However,replicating this unique capability in robots remains a significant challenge.Here,we present a new form of ultralight multifunctional tactile nano-layered carbon aerogel sensor that provides pressure,temperature,material recognition and 3D location capabilities,which is combined with multimodal supervised learning algorithms for object recognition.The sensor exhibits human-like pressure(0.04–100 kPa)and temperature(21.5–66.2℃)detection,millisecond response times(11 ms),a pressure sensitivity of 92.22 kPa^(−1)and triboelectric durability of over 6000 cycles.The devised algorithm has universality and can accommodate a range of application scenarios.The tactile system can identify common foods in a kitchen scene with 94.63%accuracy and explore the topographic and geomorphic features of a Mars scene with 100%accuracy.This sensing approach empowers robots with versatile tactile perception to advance future society toward heightened sensing,recognition and intelligence.
基金supported by the National Natural Science Foundation of China(61572092,61702403)the Fundamental Research Funds for the Central Universities(JB170308,JBF180301)+2 种基金the Project Funded by China Postdoctoral Science Foundation(2018M633473)the Basic Research Project of Weinan Science and Technology Bureau(ZDYF-JCYJ-17)the Project of Shaanxi Provincial Supports Discipline(Mathematics)
文摘In this paper,we first derive two types of transformed Franklin polynomial:substituted and weighted radial Franklin polynomials.Two radial orthogonal moments are proposed based on these two types of polynomials,namely substituted Franklin-Fourier moments and weighted Franklin-Fourier moments(SFFMs and WFFMs),which are orthogonal in polar coordinates.The radial kernel functions of SFFMs and WFFMs are transformed Franklin functions and Franklin functions are composed of a class of complete orthogonal splines function system of degree one.Therefore,it provides the possibility of avoiding calculating high order polynomials,and thus the accurate values of SFFMs and WFFMs can be obtained directly with little computational cost.Theoretical and experimental results show that Franklin functions are not well suited for constructing higher-order moments of SFFMs and WFFMs,but compared with traditional orthogonal moments(e.g.,BFMs,OFMs and ZMs)in polar coordinates,the proposed two types of Franklin-Fourier Moments have better performance respectively in lower-order moments.
基金Supported by the National Natural Science Foundation of China(61701029)Basic Research Foundation of Beijing Institute of Technology(20170542008)Industry-University Research Innovation Foundation of the Science and Technology Development Center of the Ministry of Education(2018A02012)。
文摘In order to accomplish the task of object recognition in natural scenes,a new object recognition algorithm based on an improved convolutional neural network(CNN)is proposed.First,candidate object windows are extracted from the original image.Then,candidate object windows are input into the improved CNN model to obtain deep features.Finally,the deep features are input into the Softmax and the confidence scores of classes are obtained.The candidate object window with the highest confidence score is selected as the object recognition result.Based on AlexNet,Inception V1 is introduced into the improved CNN and the fully connected layer is replaced by the average pooling layer,which widens the network and deepens the network at the same time.Experimental results show that the improved object recognition algorithm can obtain better recognition results in multiple natural scene images,and has a higher degree of accuracy than the classical algorithms in the field of object recognition.
文摘In this Paper, a classification method based on neural networks is presented for recognition of 3D objects. Indeed, the objective of this paper is to classify an object query against objects in a database, which leads to recognition of the former. 3D objects of this database are transformations of other objects by one element of the overall transformation. The set of transformations considered in this work is the general affine group.