We propose and discuss a novel concept of robust set stabilization by permissible controls; this concept is helpful when dealing with both a priori information of model parameters and different permissible controls in...We propose and discuss a novel concept of robust set stabilization by permissible controls; this concept is helpful when dealing with both a priori information of model parameters and different permissible controls including quantum measurements. Both controllability and stabilization can be regarded as the special case of the novel concept. An instance is presented for a kind of uncertain open quantum systems to further justify this gen- eralized concept. It is underlined that a new type of hybrid control based on periodically perturbed projective measurements can be the permissible control of uncertain open quantum systems when perturbed projective measurements are available. The sufficient conditions are given for the robust set stabilization of uncertain quantum open systems by the hybrid control, and the design of the hybrid control is reduced to selecting the period of measurements.展开更多
Many international brands have a phenomenal Chinese name which,paradoxically,comes from a rather prosaic name.The reason for this may lie in the fact that they need an outstanding translation of their names in order t...Many international brands have a phenomenal Chinese name which,paradoxically,comes from a rather prosaic name.The reason for this may lie in the fact that they need an outstanding translation of their names in order to be successful in international marketing.Hence the translation of brand names is an important part of the advertisement.And a good translation is expected to bridge the differences of cultures,languages,spending habits,thinking patterns,etc.展开更多
Deep Web sources contain a large of high-quality and query-related structured date. One of the challenges in the Deep Web is extracting result schemas of Deep Web sources. To address this challenge, this paper describ...Deep Web sources contain a large of high-quality and query-related structured date. One of the challenges in the Deep Web is extracting result schemas of Deep Web sources. To address this challenge, this paper describes a novel approach that extracts both result data and the result schema of a Web database. The approach first models the query interface of a Deep Web source and fills in it with a specifically query instance. Then the result pages of the Deep Web sources are formatted in the tree structure to retrieve subtrees that contain elements of the query instance, Next, result schema of the Deep Web source is extracted by matching the subtree' nodes with the query instance, in which, a two-phase schema extraction method is adopted for obtaining more accurate result schema. Finally, experiments on real Deep Web sources show the utility of our approach, which provides a high precision and recall.展开更多
Search-based software engineering has mainly dealt with automated test data generation by metaheuristic search techniques. Similarly, we try to generate the test data (i.e., problem instances) which show the worst cas...Search-based software engineering has mainly dealt with automated test data generation by metaheuristic search techniques. Similarly, we try to generate the test data (i.e., problem instances) which show the worst case of algorithms by such a technique. In this paper, in terms of non-functional testing, we re-define the worst case of some algorithms, respectively. By using genetic algorithms (GAs), we illustrate the strategies corresponding to each type of instances. We here adopt three problems for examples;the sorting problem, the 0/1 knapsack problem (0/1KP), and the travelling salesperson problem (TSP). In some algorithms solving these problems, we could find the worst-case instances successfully;the successfulness of the result is based on a statistical approach and comparison to the results by using the random testing. Our tried examples introduce informative guidelines to the use of genetic algorithms in generating the worst-case instance, which is defined in the aspect of algorithm performance.展开更多
This paper proposes a checking method based on mutual instances and discusses three key problems in the method: how to deal with mistakes in the mutual instances and how to deal with too many or too few mutual instan...This paper proposes a checking method based on mutual instances and discusses three key problems in the method: how to deal with mistakes in the mutual instances and how to deal with too many or too few mutual instances. It provides the checking based on the weighted mutual instances considering fault tolerance, gives a way to partition the large-scale mutual instances, and proposes a process greatly reducing the manual annotation work to get more mutual instances. Intension annotation that improves the checking method is also discussed. The method is practical and effective to check subsumption relations between concept queries in different ontologies based on mutual instances.展开更多
HTTP Adaptive Streaming(HAS)of video content is becoming an undivided part of the Internet and accounts for most of today’s network traffic.Video compression technology plays a vital role in efficiently utilizing net...HTTP Adaptive Streaming(HAS)of video content is becoming an undivided part of the Internet and accounts for most of today’s network traffic.Video compression technology plays a vital role in efficiently utilizing network channels,but encoding videos into multiple representations with selected encoding parameters is a significant challenge.However,video encoding is a computationally intensive and time-consuming operation that requires high-performance resources provided by on-premise infrastructures or public clouds.In turn,the public clouds,such as Amazon elastic compute cloud(EC2),provide hundreds of computing instances optimized for different purposes and clients’budgets.Thus,there is a need for algorithms and methods for optimized computing instance selection for specific tasks such as video encoding and transcoding operations.Additionally,the encoding speed directly depends on the selected encoding parameters and the complexity characteristics of video content.In this paper,we first benchmarked the video encoding performance of Amazon EC2 spot instances using multiple×264 codec encoding parameters and video sequences of varying complexity.Then,we proposed a novel fast approach to optimize Amazon EC2 spot instances and minimize video encoding costs.Furthermore,we evaluated how the optimized selection of EC2 spot instances can affect the encoding cost.The results show that our approach,on average,can reduce the encoding costs by at least 15.8%and up to 47.8%when compared to a random selection of EC2 spot instances.展开更多
When deploying workflows in cloud environments,the use of Spot Instances(SIs)is intriguing as they are much cheaper than on-demand ones.However,Sls are volatile and may be revoked at any time,which results in a more c...When deploying workflows in cloud environments,the use of Spot Instances(SIs)is intriguing as they are much cheaper than on-demand ones.However,Sls are volatile and may be revoked at any time,which results in a more challenging scheduling problem involving execution interruption and hence hinders the successful handling of conventional cloud workflow scheduling techniques.Although some scheduling methods for Sls have been proposed,most of them are no more applicable to the latest Sls,as they have evolved by eliminating bidding and simplifying the pricing model.This study focuses on how to minimize the execution cost with a deadline constraint when deploying a workflow on volatile Sls in cloud environments.Based on Monte Carlo simulation and list scheduling,a stochastic scheduling method called MCLS is devised to optimize a utility function introduced for this problem.With the Monte Carlo simulation framework,MCLS employs sampled task execution time to build solutions via deadline distribution and list scheduling,and then returns the most robust solution from all the candidates with a specific evaluation mechanism and selection criteria.Experimental results show that the performance of MCLS is more competitive comparedwithtraditionalalgorithms.展开更多
Infrastructure-as-a-Service(IaaS)cloud platforms offer resources with diverse buying options.Users can run an instance on the on-demand market which is stable but expensive or on the spot market with a significant dis...Infrastructure-as-a-Service(IaaS)cloud platforms offer resources with diverse buying options.Users can run an instance on the on-demand market which is stable but expensive or on the spot market with a significant discount.However,users have to carefully weigh the low cost of spot instances against their poor availability.Spot instances will be revoked when the revocation event occurs.Thus,an important problem that an IaaS user faces now is how to use spot in-stances in a cost-effective and low-risk way.Based on the replication-based fault tolerance mechanism,we propose an on-line termination algorithm that optimizes the cost of using spot instances while ensuring operational stability.We prove that in most cases,the cost of our proposed online algorithm will not exceed twice the minimum cost of the optimal of-fline algorithm that knows the exact future a priori.Through a large number of experiments,we verify that our algorithm in most cases has a competitive ratio of no more than 2,and in other cases it can also reach the guaranteed competitive ratio.展开更多
The real-time detection and instance segmentation of strawberries constitute fundamental components in the development of strawberry harvesting robots.Real-time identification of strawberries in an unstructured envi-r...The real-time detection and instance segmentation of strawberries constitute fundamental components in the development of strawberry harvesting robots.Real-time identification of strawberries in an unstructured envi-ronment is a challenging task.Current instance segmentation algorithms for strawberries suffer from issues such as poor real-time performance and low accuracy.To this end,the present study proposes an Efficient YOLACT(E-YOLACT)algorithm for strawberry detection and segmentation based on the YOLACT framework.The key enhancements of the E-YOLACT encompass the development of a lightweight attention mechanism,pyramid squeeze shuffle attention(PSSA),for efficient feature extraction.Additionally,an attention-guided context-feature pyramid network(AC-FPN)is employed instead of FPN to optimize the architecture’s performance.Furthermore,a feature-enhanced model(FEM)is introduced to enhance the prediction head’s capabilities,while efficient fast non-maximum suppression(EF-NMS)is devised to improve non-maximum suppression.The experimental results demonstrate that the E-YOLACT achieves a Box-mAP and Mask-mAP of 77.9 and 76.6,respectively,on the custom dataset.Moreover,it exhibits an impressive category accuracy of 93.5%.Notably,the E-YOLACT also demonstrates a remarkable real-time detection capability with a speed of 34.8 FPS.The method proposed in this article presents an efficient approach for the vision system of a strawberry-picking robot.展开更多
This study presents a single-class and multi-class instance segmentation approach applied to ancient Palmyrene inscriptions,employing two state-of-the-art deep learning algorithms,namely YOLOv8 and Roboflow 3.0.The go...This study presents a single-class and multi-class instance segmentation approach applied to ancient Palmyrene inscriptions,employing two state-of-the-art deep learning algorithms,namely YOLOv8 and Roboflow 3.0.The goal is to contribute to the preservation and understanding of historical texts,showcasing the potential of modern deep learning methods in archaeological research.Our research culminates in several key findings and scientific contributions.We comprehensively compare the performance of YOLOv8 and Roboflow 3.0 in the context of Palmyrene character segmentation—this comparative analysis mainly focuses on the strengths and weaknesses of each algorithm in this context.We also created and annotated an extensive dataset of Palmyrene inscriptions,a crucial resource for further research in the field.The dataset serves for training and evaluating the segmentation models.We employ comparative evaluation metrics to quantitatively assess the segmentation results,ensuring the reliability and reproducibility of our findings and we present custom visualization tools for predicted segmentation masks.Our study advances the state of the art in semi-automatic reading of Palmyrene inscriptions and establishes a benchmark for future research.The availability of the Palmyrene dataset and the insights into algorithm performance contribute to the broader understanding of historical text analysis.展开更多
Chronic heart failure(CHF)is a clinical syndrome manifested by reduced pumping ability of the heart,increased pressure in heart chambers in both physical activity and at rest.The symptoms of this syndrome are dyspnea,...Chronic heart failure(CHF)is a clinical syndrome manifested by reduced pumping ability of the heart,increased pressure in heart chambers in both physical activity and at rest.The symptoms of this syndrome are dyspnea,undue fatigability,peripheral edema,which follow structural and functional changes of the myocardium.[1]The growing incidence of CHF,especially among elderly people,is an urgent problem for medicine in the vast majority of industrialized countries.For instance,in Russian Federation,CHF is diagnosed in about 7%of cardiovascular patients.At the same time,this indicator varies from 0.3%in young people(20-29 years old)to 70%in the older age group.[2,3].展开更多
The precise detection and segmentation of tumor lesions are very important for lung cancer computer-aided diagnosis.However,in PET/CT(Positron Emission Tomography/Computed Tomography)lung images,the lesion shapes are ...The precise detection and segmentation of tumor lesions are very important for lung cancer computer-aided diagnosis.However,in PET/CT(Positron Emission Tomography/Computed Tomography)lung images,the lesion shapes are complex,the edges are blurred,and the sample numbers are unbalanced.To solve these problems,this paper proposes a Multi-branch Cross-scale Interactive Feature fusion Transformer model(MCIF-Transformer Mask RCNN)for PET/CT lung tumor instance segmentation,The main innovative works of this paper are as follows:Firstly,the ResNet-Transformer backbone network is used to extract global feature and local feature in lung images.The pixel dependence relationship is established in local and non-local fields to improve the model perception ability.Secondly,the Cross-scale Interactive Feature Enhancement auxiliary network is designed to provide the shallow features to the deep features,and the cross-scale interactive feature enhancement module(CIFEM)is used to enhance the attention ability of the fine-grained features.Thirdly,the Cross-scale Interactive Feature fusion FPN network(CIF-FPN)is constructed to realize bidirectional interactive fusion between deep features and shallow features,and the low-level features are enhanced in deep semantic features.Finally,4 ablation experiments,3 comparison experiments of detection,3 comparison experiments of segmentation and 6 comparison experiments with two-stage and single-stage instance segmentation networks are done on PET/CT lung medical image datasets.The results showed that APdet,APseg,ARdet and ARseg indexes are improved by 5.5%,5.15%,3.11%and 6.79%compared with Mask RCNN(resnet50).Based on the above research,the precise detection and segmentation of the lesion region are realized in this paper.This method has positive significance for the detection of lung tumors.展开更多
In actual traffic scenarios,precise recognition of traffic participants,such as vehicles and pedestrians,is crucial for intelligent transportation.This study proposes an improved algorithm built on Mask-RCNN to enhanc...In actual traffic scenarios,precise recognition of traffic participants,such as vehicles and pedestrians,is crucial for intelligent transportation.This study proposes an improved algorithm built on Mask-RCNN to enhance the ability of autonomous driving systems to recognize traffic participants.The algorithmincorporates long and shortterm memory networks and the fused attention module(GSAM,GCT,and Spatial Attention Module)to enhance the algorithm’s capability to process both global and local information.Additionally,to increase the network’s initial operation stability,the original network activation function was replaced with Gaussian error linear unit.Experiments were conducted using the publicly available Cityscapes dataset.Comparing the test results,it was observed that the revised algorithmoutperformed the original algorithmin terms of AP_(50),AP_(75),and othermetrics by 8.7%and 9.6%for target detection and 12.5%and 13.3%for segmentation.展开更多
Dynamic Simultaneous Localization and Mapping(SLAM)in visual scenes is currently a major research area in fields such as robot navigation and autonomous driving.However,in the face of complex real-world envi-ronments,...Dynamic Simultaneous Localization and Mapping(SLAM)in visual scenes is currently a major research area in fields such as robot navigation and autonomous driving.However,in the face of complex real-world envi-ronments,current dynamic SLAM systems struggle to achieve precise localization and map construction.With the advancement of deep learning,there has been increasing interest in the development of deep learning-based dynamic SLAM visual odometry in recent years,and more researchers are turning to deep learning techniques to address the challenges of dynamic SLAM.Compared to dynamic SLAM systems based on deep learning methods such as object detection and semantic segmentation,dynamic SLAM systems based on instance segmentation can not only detect dynamic objects in the scene but also distinguish different instances of the same type of object,thereby reducing the impact of dynamic objects on the SLAM system’s positioning.This article not only introduces traditional dynamic SLAM systems based on mathematical models but also provides a comprehensive analysis of existing instance segmentation algorithms and dynamic SLAM systems based on instance segmentation,comparing and summarizing their advantages and disadvantages.Through comparisons on datasets,it is found that instance segmentation-based methods have significant advantages in accuracy and robustness in dynamic environments.However,the real-time performance of instance segmentation algorithms hinders the widespread application of dynamic SLAM systems.In recent years,the rapid development of single-stage instance segmentationmethods has brought hope for the widespread application of dynamic SLAM systems based on instance segmentation.Finally,possible future research directions and improvementmeasures are discussed for reference by relevant professionals.展开更多
Instance segmentation plays an important role in image processing.The Deep Snake algorithm based on contour iteration deforms an initial bounding box to an instance contour end-to-end,which can improve the performance...Instance segmentation plays an important role in image processing.The Deep Snake algorithm based on contour iteration deforms an initial bounding box to an instance contour end-to-end,which can improve the performance of instance segmentation,but has defects such as slow segmentation speed and sub-optimal initial contour.To solve these problems,a real-time instance segmentation algorithm based on contour learning was proposed.Firstly,ShuffleNet V2 was used as backbone network,and the receptive field of the model was expanded by using a 5×5 convolution kernel.Secondly,a lightweight up-sampling module,multi-stage aggregation(MSA),performs residual fusion of multi-layer features,which not only improves segmentation speed,but also extracts effective features more comprehensively.Thirdly,a contour initialization method for network learning was designed,and a global contour feature aggregation mechanism was used to return a coarse contour,which solves the problem of excessive error between manually initialized contour and real contour.Finally,the Snake deformation module was used to iteratively optimize the coarse contour to obtain the final instance contour.The experimental results showed that the proposed method improved the instance segmentation accuracy on semantic boundaries dataset(SBD),Cityscapes and Kins datasets,and the average precision reached 55.8 on the SBD;Compared with Deep Snake,the model parameters were reduced by 87.2%,calculation amount was reduced by 78.3%,and segmentation speed reached 39.8 frame·s−1 when instance segmentation was performed on an image with a size of 512×512 pixels on a 2080Ti GPU.The proposed method can reduce resource consumption,realize instance segmentation tasks quickly and accurately,and therefore is more suitable for embedded platforms with limited resources.展开更多
Tea leaf picking is a crucial stage in tea production that directly influences the quality and value of the tea.Traditional tea-picking machines may compromise the quality of the tea leaves.High-quality teas are often...Tea leaf picking is a crucial stage in tea production that directly influences the quality and value of the tea.Traditional tea-picking machines may compromise the quality of the tea leaves.High-quality teas are often handpicked and need more delicate operations in intelligent picking machines.Compared with traditional image processing techniques,deep learning models have stronger feature extraction capabilities,and better generalization and are more suitable for practical tea shoot harvesting.However,current research mostly focuses on shoot detection and cannot directly accomplish end-to-end shoot segmentation tasks.We propose a tea shoot instance segmentation model based on multi-scale mixed attention(Mask2FusionNet)using a dataset from the tea garden in Hangzhou.We further analyzed the characteristics of the tea shoot dataset,where the proportion of small to medium-sized targets is 89.9%.Our algorithm is compared with several mainstream object segmentation algorithms,and the results demonstrate that our model achieves an accuracy of 82%in recognizing the tea shoots,showing a better performance compared to other models.Through ablation experiments,we found that ResNet50,PointRend strategy,and the Feature Pyramid Network(FPN)architecture can improve performance by 1.6%,1.4%,and 2.4%,respectively.These experiments demonstrated that our proposed multi-scale and point selection strategy optimizes the feature extraction capability for overlapping small targets.The results indicate that the proposed Mask2FusionNet model can perform the shoot segmentation in unstructured environments,realizing the individual distinction of tea shoots,and complete extraction of the shoot edge contours with a segmentation accuracy of 82.0%.The research results can provide algorithmic support for the segmentation and intelligent harvesting of premium tea shoots at different scales.展开更多
The Third Plenum holds a unique place in the Communist Party of China(CPC)’s political calendar because the Party often introduces significant reforms and outlines strategic priorities during the event.For instance,t...The Third Plenum holds a unique place in the Communist Party of China(CPC)’s political calendar because the Party often introduces significant reforms and outlines strategic priorities during the event.For instance,the Third Plenary Session of the 11th Central Committee of CPC in 1978 marked the beginning of China’s reform and opening-up drive under Deng Xiaoping,after which the country underwent historic economic development.展开更多
Traditional Chinese villages serve as crucial repositories of traditional culture.However,In China,the urgent task of preserving information about traditional village architecture has arisen due to the degradation of ...Traditional Chinese villages serve as crucial repositories of traditional culture.However,In China,the urgent task of preserving information about traditional village architecture has arisen due to the degradation of these villages’appearance caused by rapid urbanization in recent years.This paper proposes a method for preserving information about traditional village rooftops based on high spatial resolution remote sensing imagery.Leveraging an improved Mask R-CNN model,the method conducts target recognition on the rooftops of traditional village buildings and generates vectorized representations of these rooftops.The precision rate,recall rate,and F1-score achieved in the experimental results are 93.26%,86.33%,and 92.02%,respectively.These findings indicate the effectiveness of the proposed method in preserving information about traditional village architecture and providing a viable approach to support the sustainable development of traditional villages in China.展开更多
Correction.The corresponding author of“A Multi-Center Randomized Controlled Study UsingΔP_(CO2)/Ca-v_(O2) as the Target to Guide Early Tissue Hypoperfusion in Sepsis in Plateau Areas,”published in Volume 8 Issue 1(...Correction.The corresponding author of“A Multi-Center Randomized Controlled Study UsingΔP_(CO2)/Ca-v_(O2) as the Target to Guide Early Tissue Hypoperfusion in Sepsis in Plateau Areas,”published in Volume 8 Issue 1(2024)of Journal of Clinical and Nursing Research(https://www.doi.org/10.26689/jcnr.v8i1.5944),wrote to the editors about a typographical error in the article.All instances of“>1.84”(found in the abstract and conclusion)should be corrected to“<1.84.”展开更多
Dear readers,In the field of stroke research and treatment,cutting-edge technologies and studies continue to emerge,providing new perspectives and strategies for exploring the mechanisms and treatment methods of strok...Dear readers,In the field of stroke research and treatment,cutting-edge technologies and studies continue to emerge,providing new perspectives and strategies for exploring the mechanisms and treatment methods of stroke.According to the latest advancements,neuroimaging techniques play a crucial role in stroke research.MRI technology,for instance,is essential for evaluating stroke patients.展开更多
基金Supported by the National Natural Science Foundation of China under Grant Nos 61673389,61273202 and 61134008
文摘We propose and discuss a novel concept of robust set stabilization by permissible controls; this concept is helpful when dealing with both a priori information of model parameters and different permissible controls including quantum measurements. Both controllability and stabilization can be regarded as the special case of the novel concept. An instance is presented for a kind of uncertain open quantum systems to further justify this gen- eralized concept. It is underlined that a new type of hybrid control based on periodically perturbed projective measurements can be the permissible control of uncertain open quantum systems when perturbed projective measurements are available. The sufficient conditions are given for the robust set stabilization of uncertain quantum open systems by the hybrid control, and the design of the hybrid control is reduced to selecting the period of measurements.
文摘Many international brands have a phenomenal Chinese name which,paradoxically,comes from a rather prosaic name.The reason for this may lie in the fact that they need an outstanding translation of their names in order to be successful in international marketing.Hence the translation of brand names is an important part of the advertisement.And a good translation is expected to bridge the differences of cultures,languages,spending habits,thinking patterns,etc.
基金Supported by the National Natural Science Foundation of China (60673139, 60473073, 60573090)
文摘Deep Web sources contain a large of high-quality and query-related structured date. One of the challenges in the Deep Web is extracting result schemas of Deep Web sources. To address this challenge, this paper describes a novel approach that extracts both result data and the result schema of a Web database. The approach first models the query interface of a Deep Web source and fills in it with a specifically query instance. Then the result pages of the Deep Web sources are formatted in the tree structure to retrieve subtrees that contain elements of the query instance, Next, result schema of the Deep Web source is extracted by matching the subtree' nodes with the query instance, in which, a two-phase schema extraction method is adopted for obtaining more accurate result schema. Finally, experiments on real Deep Web sources show the utility of our approach, which provides a high precision and recall.
文摘Search-based software engineering has mainly dealt with automated test data generation by metaheuristic search techniques. Similarly, we try to generate the test data (i.e., problem instances) which show the worst case of algorithms by such a technique. In this paper, in terms of non-functional testing, we re-define the worst case of some algorithms, respectively. By using genetic algorithms (GAs), we illustrate the strategies corresponding to each type of instances. We here adopt three problems for examples;the sorting problem, the 0/1 knapsack problem (0/1KP), and the travelling salesperson problem (TSP). In some algorithms solving these problems, we could find the worst-case instances successfully;the successfulness of the result is based on a statistical approach and comparison to the results by using the random testing. Our tried examples introduce informative guidelines to the use of genetic algorithms in generating the worst-case instance, which is defined in the aspect of algorithm performance.
基金Supported by the National Natural Sciences Foundation of China(60373066 ,60425206 ,90412003) , National Grand Fundamental Research 973 Pro-gramof China(2002CB312000) , National Research Foundation for the Doctoral Pro-gramof Higher Education of China (20020286004)
文摘This paper proposes a checking method based on mutual instances and discusses three key problems in the method: how to deal with mistakes in the mutual instances and how to deal with too many or too few mutual instances. It provides the checking based on the weighted mutual instances considering fault tolerance, gives a way to partition the large-scale mutual instances, and proposes a process greatly reducing the manual annotation work to get more mutual instances. Intension annotation that improves the checking method is also discussed. The method is practical and effective to check subsumption relations between concept queries in different ontologies based on mutual instances.
基金This work has been supported in part by the Austrian Research Promotion Agency(FFG)under the APOLLO and Karnten Fog project.
文摘HTTP Adaptive Streaming(HAS)of video content is becoming an undivided part of the Internet and accounts for most of today’s network traffic.Video compression technology plays a vital role in efficiently utilizing network channels,but encoding videos into multiple representations with selected encoding parameters is a significant challenge.However,video encoding is a computationally intensive and time-consuming operation that requires high-performance resources provided by on-premise infrastructures or public clouds.In turn,the public clouds,such as Amazon elastic compute cloud(EC2),provide hundreds of computing instances optimized for different purposes and clients’budgets.Thus,there is a need for algorithms and methods for optimized computing instance selection for specific tasks such as video encoding and transcoding operations.Additionally,the encoding speed directly depends on the selected encoding parameters and the complexity characteristics of video content.In this paper,we first benchmarked the video encoding performance of Amazon EC2 spot instances using multiple×264 codec encoding parameters and video sequences of varying complexity.Then,we proposed a novel fast approach to optimize Amazon EC2 spot instances and minimize video encoding costs.Furthermore,we evaluated how the optimized selection of EC2 spot instances can affect the encoding cost.The results show that our approach,on average,can reduce the encoding costs by at least 15.8%and up to 47.8%when compared to a random selection of EC2 spot instances.
基金This work was supported by the National Natural Science Foundation of China(Nos.62172065 and 62072060)the Natural Science Foundation of Chongqing(No.cstc2020jcyj-msxmX0137).
文摘When deploying workflows in cloud environments,the use of Spot Instances(SIs)is intriguing as they are much cheaper than on-demand ones.However,Sls are volatile and may be revoked at any time,which results in a more challenging scheduling problem involving execution interruption and hence hinders the successful handling of conventional cloud workflow scheduling techniques.Although some scheduling methods for Sls have been proposed,most of them are no more applicable to the latest Sls,as they have evolved by eliminating bidding and simplifying the pricing model.This study focuses on how to minimize the execution cost with a deadline constraint when deploying a workflow on volatile Sls in cloud environments.Based on Monte Carlo simulation and list scheduling,a stochastic scheduling method called MCLS is devised to optimize a utility function introduced for this problem.With the Monte Carlo simulation framework,MCLS employs sampled task execution time to build solutions via deadline distribution and list scheduling,and then returns the most robust solution from all the candidates with a specific evaluation mechanism and selection criteria.Experimental results show that the performance of MCLS is more competitive comparedwithtraditionalalgorithms.
基金This work was supported by the National Key Research and Development Program of China under Grant No.2018YFB14-04501。
文摘Infrastructure-as-a-Service(IaaS)cloud platforms offer resources with diverse buying options.Users can run an instance on the on-demand market which is stable but expensive or on the spot market with a significant discount.However,users have to carefully weigh the low cost of spot instances against their poor availability.Spot instances will be revoked when the revocation event occurs.Thus,an important problem that an IaaS user faces now is how to use spot in-stances in a cost-effective and low-risk way.Based on the replication-based fault tolerance mechanism,we propose an on-line termination algorithm that optimizes the cost of using spot instances while ensuring operational stability.We prove that in most cases,the cost of our proposed online algorithm will not exceed twice the minimum cost of the optimal of-fline algorithm that knows the exact future a priori.Through a large number of experiments,we verify that our algorithm in most cases has a competitive ratio of no more than 2,and in other cases it can also reach the guaranteed competitive ratio.
基金funded by Anhui Provincial Natural Science Foundation(No.2208085ME128)the Anhui University-Level Special Project of Anhui University of Science and Technology(No.XCZX2021-01)+1 种基金the Research and the Development Fund of the Institute of Environmental Friendly Materials and Occupational Health,Anhui University of Science and Technology(No.ALW2022YF06)Anhui Province New Era Education Quality Project(Graduate Education)(No.2022xscx073).
文摘The real-time detection and instance segmentation of strawberries constitute fundamental components in the development of strawberry harvesting robots.Real-time identification of strawberries in an unstructured envi-ronment is a challenging task.Current instance segmentation algorithms for strawberries suffer from issues such as poor real-time performance and low accuracy.To this end,the present study proposes an Efficient YOLACT(E-YOLACT)algorithm for strawberry detection and segmentation based on the YOLACT framework.The key enhancements of the E-YOLACT encompass the development of a lightweight attention mechanism,pyramid squeeze shuffle attention(PSSA),for efficient feature extraction.Additionally,an attention-guided context-feature pyramid network(AC-FPN)is employed instead of FPN to optimize the architecture’s performance.Furthermore,a feature-enhanced model(FEM)is introduced to enhance the prediction head’s capabilities,while efficient fast non-maximum suppression(EF-NMS)is devised to improve non-maximum suppression.The experimental results demonstrate that the E-YOLACT achieves a Box-mAP and Mask-mAP of 77.9 and 76.6,respectively,on the custom dataset.Moreover,it exhibits an impressive category accuracy of 93.5%.Notably,the E-YOLACT also demonstrates a remarkable real-time detection capability with a speed of 34.8 FPS.The method proposed in this article presents an efficient approach for the vision system of a strawberry-picking robot.
基金The results and knowledge included herein have been obtained owing to support from the following institutional grant.Internal grant agency of the Faculty of Economics and Management,Czech University of Life Sciences Prague,Grant No.2023A0004-“Text Segmentation Methods of Historical Alphabets in OCR Development”.https://iga.pef.czu.cz/.Funds were granted to T.Novák,A.Hamplová,O.Svojše,and A.Veselýfrom the author team.
文摘This study presents a single-class and multi-class instance segmentation approach applied to ancient Palmyrene inscriptions,employing two state-of-the-art deep learning algorithms,namely YOLOv8 and Roboflow 3.0.The goal is to contribute to the preservation and understanding of historical texts,showcasing the potential of modern deep learning methods in archaeological research.Our research culminates in several key findings and scientific contributions.We comprehensively compare the performance of YOLOv8 and Roboflow 3.0 in the context of Palmyrene character segmentation—this comparative analysis mainly focuses on the strengths and weaknesses of each algorithm in this context.We also created and annotated an extensive dataset of Palmyrene inscriptions,a crucial resource for further research in the field.The dataset serves for training and evaluating the segmentation models.We employ comparative evaluation metrics to quantitatively assess the segmentation results,ensuring the reliability and reproducibility of our findings and we present custom visualization tools for predicted segmentation masks.Our study advances the state of the art in semi-automatic reading of Palmyrene inscriptions and establishes a benchmark for future research.The availability of the Palmyrene dataset and the insights into algorithm performance contribute to the broader understanding of historical text analysis.
文摘Chronic heart failure(CHF)is a clinical syndrome manifested by reduced pumping ability of the heart,increased pressure in heart chambers in both physical activity and at rest.The symptoms of this syndrome are dyspnea,undue fatigability,peripheral edema,which follow structural and functional changes of the myocardium.[1]The growing incidence of CHF,especially among elderly people,is an urgent problem for medicine in the vast majority of industrialized countries.For instance,in Russian Federation,CHF is diagnosed in about 7%of cardiovascular patients.At the same time,this indicator varies from 0.3%in young people(20-29 years old)to 70%in the older age group.[2,3].
基金funded by National Natural Science Foundation of China No.62062003Ningxia Natural Science Foundation Project No.2023AAC03293.
文摘The precise detection and segmentation of tumor lesions are very important for lung cancer computer-aided diagnosis.However,in PET/CT(Positron Emission Tomography/Computed Tomography)lung images,the lesion shapes are complex,the edges are blurred,and the sample numbers are unbalanced.To solve these problems,this paper proposes a Multi-branch Cross-scale Interactive Feature fusion Transformer model(MCIF-Transformer Mask RCNN)for PET/CT lung tumor instance segmentation,The main innovative works of this paper are as follows:Firstly,the ResNet-Transformer backbone network is used to extract global feature and local feature in lung images.The pixel dependence relationship is established in local and non-local fields to improve the model perception ability.Secondly,the Cross-scale Interactive Feature Enhancement auxiliary network is designed to provide the shallow features to the deep features,and the cross-scale interactive feature enhancement module(CIFEM)is used to enhance the attention ability of the fine-grained features.Thirdly,the Cross-scale Interactive Feature fusion FPN network(CIF-FPN)is constructed to realize bidirectional interactive fusion between deep features and shallow features,and the low-level features are enhanced in deep semantic features.Finally,4 ablation experiments,3 comparison experiments of detection,3 comparison experiments of segmentation and 6 comparison experiments with two-stage and single-stage instance segmentation networks are done on PET/CT lung medical image datasets.The results showed that APdet,APseg,ARdet and ARseg indexes are improved by 5.5%,5.15%,3.11%and 6.79%compared with Mask RCNN(resnet50).Based on the above research,the precise detection and segmentation of the lesion region are realized in this paper.This method has positive significance for the detection of lung tumors.
基金the National Natural Science Foundation of China(52175236)Qingdao People’s Livelihood Science and Technology Plan(19-6-1-88-nsh).
文摘In actual traffic scenarios,precise recognition of traffic participants,such as vehicles and pedestrians,is crucial for intelligent transportation.This study proposes an improved algorithm built on Mask-RCNN to enhance the ability of autonomous driving systems to recognize traffic participants.The algorithmincorporates long and shortterm memory networks and the fused attention module(GSAM,GCT,and Spatial Attention Module)to enhance the algorithm’s capability to process both global and local information.Additionally,to increase the network’s initial operation stability,the original network activation function was replaced with Gaussian error linear unit.Experiments were conducted using the publicly available Cityscapes dataset.Comparing the test results,it was observed that the revised algorithmoutperformed the original algorithmin terms of AP_(50),AP_(75),and othermetrics by 8.7%and 9.6%for target detection and 12.5%and 13.3%for segmentation.
基金the National Natural Science Foundation of China(No.62063006)the Natural Science Foundation of Guangxi Province(No.2023GXNS-FAA026025)+3 种基金the Innovation Fund of Chinese Universities Industry-University-Research(ID:2021RYC06005)the Research Project for Young andMiddle-Aged Teachers in Guangxi Universi-ties(ID:2020KY15013)the Special Research Project of Hechi University(ID:2021GCC028)financially supported by the Project of Outstanding Thousand Young Teachers’Training in Higher Education Institutions of Guangxi,Guangxi Colleges and Universities Key Laboratory of AI and Information Processing(Hechi University),Education Department of Guangxi Zhuang Autonomous Region.
文摘Dynamic Simultaneous Localization and Mapping(SLAM)in visual scenes is currently a major research area in fields such as robot navigation and autonomous driving.However,in the face of complex real-world envi-ronments,current dynamic SLAM systems struggle to achieve precise localization and map construction.With the advancement of deep learning,there has been increasing interest in the development of deep learning-based dynamic SLAM visual odometry in recent years,and more researchers are turning to deep learning techniques to address the challenges of dynamic SLAM.Compared to dynamic SLAM systems based on deep learning methods such as object detection and semantic segmentation,dynamic SLAM systems based on instance segmentation can not only detect dynamic objects in the scene but also distinguish different instances of the same type of object,thereby reducing the impact of dynamic objects on the SLAM system’s positioning.This article not only introduces traditional dynamic SLAM systems based on mathematical models but also provides a comprehensive analysis of existing instance segmentation algorithms and dynamic SLAM systems based on instance segmentation,comparing and summarizing their advantages and disadvantages.Through comparisons on datasets,it is found that instance segmentation-based methods have significant advantages in accuracy and robustness in dynamic environments.However,the real-time performance of instance segmentation algorithms hinders the widespread application of dynamic SLAM systems.In recent years,the rapid development of single-stage instance segmentationmethods has brought hope for the widespread application of dynamic SLAM systems based on instance segmentation.Finally,possible future research directions and improvementmeasures are discussed for reference by relevant professionals.
基金supported by National Key Research and Development Program(No.2022YFE0112400)National Natural Science Foundation of China(No.21706096)Natural Science Foundation of Jiangsu Province(No.BK20160162).
文摘Instance segmentation plays an important role in image processing.The Deep Snake algorithm based on contour iteration deforms an initial bounding box to an instance contour end-to-end,which can improve the performance of instance segmentation,but has defects such as slow segmentation speed and sub-optimal initial contour.To solve these problems,a real-time instance segmentation algorithm based on contour learning was proposed.Firstly,ShuffleNet V2 was used as backbone network,and the receptive field of the model was expanded by using a 5×5 convolution kernel.Secondly,a lightweight up-sampling module,multi-stage aggregation(MSA),performs residual fusion of multi-layer features,which not only improves segmentation speed,but also extracts effective features more comprehensively.Thirdly,a contour initialization method for network learning was designed,and a global contour feature aggregation mechanism was used to return a coarse contour,which solves the problem of excessive error between manually initialized contour and real contour.Finally,the Snake deformation module was used to iteratively optimize the coarse contour to obtain the final instance contour.The experimental results showed that the proposed method improved the instance segmentation accuracy on semantic boundaries dataset(SBD),Cityscapes and Kins datasets,and the average precision reached 55.8 on the SBD;Compared with Deep Snake,the model parameters were reduced by 87.2%,calculation amount was reduced by 78.3%,and segmentation speed reached 39.8 frame·s−1 when instance segmentation was performed on an image with a size of 512×512 pixels on a 2080Ti GPU.The proposed method can reduce resource consumption,realize instance segmentation tasks quickly and accurately,and therefore is more suitable for embedded platforms with limited resources.
基金This research was supported by the National Natural Science Foundation of China No.62276086the National Key R&D Program of China No.2022YFD2000100Zhejiang Provincial Natural Science Foundation of China under Grant No.LTGN23D010002.
文摘Tea leaf picking is a crucial stage in tea production that directly influences the quality and value of the tea.Traditional tea-picking machines may compromise the quality of the tea leaves.High-quality teas are often handpicked and need more delicate operations in intelligent picking machines.Compared with traditional image processing techniques,deep learning models have stronger feature extraction capabilities,and better generalization and are more suitable for practical tea shoot harvesting.However,current research mostly focuses on shoot detection and cannot directly accomplish end-to-end shoot segmentation tasks.We propose a tea shoot instance segmentation model based on multi-scale mixed attention(Mask2FusionNet)using a dataset from the tea garden in Hangzhou.We further analyzed the characteristics of the tea shoot dataset,where the proportion of small to medium-sized targets is 89.9%.Our algorithm is compared with several mainstream object segmentation algorithms,and the results demonstrate that our model achieves an accuracy of 82%in recognizing the tea shoots,showing a better performance compared to other models.Through ablation experiments,we found that ResNet50,PointRend strategy,and the Feature Pyramid Network(FPN)architecture can improve performance by 1.6%,1.4%,and 2.4%,respectively.These experiments demonstrated that our proposed multi-scale and point selection strategy optimizes the feature extraction capability for overlapping small targets.The results indicate that the proposed Mask2FusionNet model can perform the shoot segmentation in unstructured environments,realizing the individual distinction of tea shoots,and complete extraction of the shoot edge contours with a segmentation accuracy of 82.0%.The research results can provide algorithmic support for the segmentation and intelligent harvesting of premium tea shoots at different scales.
文摘The Third Plenum holds a unique place in the Communist Party of China(CPC)’s political calendar because the Party often introduces significant reforms and outlines strategic priorities during the event.For instance,the Third Plenary Session of the 11th Central Committee of CPC in 1978 marked the beginning of China’s reform and opening-up drive under Deng Xiaoping,after which the country underwent historic economic development.
文摘Traditional Chinese villages serve as crucial repositories of traditional culture.However,In China,the urgent task of preserving information about traditional village architecture has arisen due to the degradation of these villages’appearance caused by rapid urbanization in recent years.This paper proposes a method for preserving information about traditional village rooftops based on high spatial resolution remote sensing imagery.Leveraging an improved Mask R-CNN model,the method conducts target recognition on the rooftops of traditional village buildings and generates vectorized representations of these rooftops.The precision rate,recall rate,and F1-score achieved in the experimental results are 93.26%,86.33%,and 92.02%,respectively.These findings indicate the effectiveness of the proposed method in preserving information about traditional village architecture and providing a viable approach to support the sustainable development of traditional villages in China.
文摘Correction.The corresponding author of“A Multi-Center Randomized Controlled Study UsingΔP_(CO2)/Ca-v_(O2) as the Target to Guide Early Tissue Hypoperfusion in Sepsis in Plateau Areas,”published in Volume 8 Issue 1(2024)of Journal of Clinical and Nursing Research(https://www.doi.org/10.26689/jcnr.v8i1.5944),wrote to the editors about a typographical error in the article.All instances of“>1.84”(found in the abstract and conclusion)should be corrected to“<1.84.”
文摘Dear readers,In the field of stroke research and treatment,cutting-edge technologies and studies continue to emerge,providing new perspectives and strategies for exploring the mechanisms and treatment methods of stroke.According to the latest advancements,neuroimaging techniques play a crucial role in stroke research.MRI technology,for instance,is essential for evaluating stroke patients.