Operation control of power systems has become challenging with an increase in the scale and complexity of power distribution systems and extensive access to renewable energy.Therefore,improvement of the ability of dat...Operation control of power systems has become challenging with an increase in the scale and complexity of power distribution systems and extensive access to renewable energy.Therefore,improvement of the ability of data-driven operation management,intelligent analysis,and mining is urgently required.To investigate and explore similar regularities of the historical operating section of the power distribution system and assist the power grid in obtaining high-value historical operation,maintenance experience,and knowledge by rule and line,a neural information retrieval model with an attention mechanism is proposed based on graph data computing technology.Based on the processing flow of the operating data of the power distribution system,a technical framework of neural information retrieval is established.Combined with the natural graph characteristics of the power distribution system,a unified graph data structure and a data fusion method of data access,data complement,and multi-source data are constructed.Further,a graph node feature-embedding representation learning algorithm and a neural information retrieval algorithm model are constructed.The neural information retrieval algorithm model is trained and tested using the generated graph node feature representation vector set.The model is verified on the operating section of the power distribution system of a provincial grid area.The results show that the proposed method demonstrates high accuracy in the similarity matching of historical operation characteristics and effectively supports intelligent fault diagnosis and elimination in power distribution systems.展开更多
Big data analytics in business intelligence do not provide effective data retrieval methods and job scheduling that will cause execution inefficiency and low system throughput.This paper aims to enhance the capability...Big data analytics in business intelligence do not provide effective data retrieval methods and job scheduling that will cause execution inefficiency and low system throughput.This paper aims to enhance the capability of data retrieval and job scheduling to speed up the operation of big data analytics to overcome inefficiency and low throughput problems.First,integrating stacked sparse autoencoder and Elasticsearch indexing explored fast data searching and distributed indexing,which reduces the search scope of the database and dramatically speeds up data searching.Next,exploiting a deep neural network to predict the approximate execution time of a job gives prioritized job scheduling based on the shortest job first,which reduces the average waiting time of job execution.As a result,the proposed data retrieval approach outperforms the previous method using a deep autoencoder and Solr indexing,significantly improving the speed of data retrieval up to 53%and increasing system throughput by 53%.On the other hand,the proposed job scheduling algorithmdefeats both first-in-first-out andmemory-sensitive heterogeneous early finish time scheduling algorithms,effectively shortening the average waiting time up to 5%and average weighted turnaround time by 19%,respectively.展开更多
In recent years, the rapid decline of Arctic sea ice area (SIA) and sea ice extent (SIE), especially for the multiyear (MY) ice, has led to significant effect on climate change. The accurate retrieval of MY ice ...In recent years, the rapid decline of Arctic sea ice area (SIA) and sea ice extent (SIE), especially for the multiyear (MY) ice, has led to significant effect on climate change. The accurate retrieval of MY ice concentration retrieval is very important and challenging to understand the ongoing changes. Three MY ice concentration retrieval algorithms were systematically evaluated. A similar total ice concentration was yielded by these algorithms, while the retrieved MY sea ice concentrations differs from each other. The MY SIA derived from NASA TEAM algorithm is relatively stable. Other two algorithms created seasonal fluctuations of MY SIA, particularly in autumn and winter. In this paper, we proposed an ice concentration retrieval algorithm, which developed the NASA TEAM algorithm by adding to use AMSR-E 6.9 GHz brightness temperature data and sea ice concentration using 89.0 GHz data. Comparison with the reference MY SIA from reference MY ice, indicates that the mean difference and root mean square (rms) difference of MY SIA derived from the algorithm of this study are 0.65×10^6 km^2 and 0.69×10^6 km^2 during January to March, -0.06×10^6 km^2 and 0.14×10^6 km^2during September to December respectively. Comparison with MY SIE obtained from weekly ice age data provided by University of Colorado show that, the mean difference and rms difference are 0.69×10^6 km^2 and 0.84×10^6 km^2, respectively. The developed algorithm proposed in this study has smaller difference compared with the reference MY ice and MY SIE from ice age data than the Wang's, Lomax' and NASA TEAM algorithms.展开更多
In monitoring systems, multiple sensor nodes can detect a single target of interest simultaneously and the data collected are usually highly correlated and redundant. If each node sends data to the base station, energ...In monitoring systems, multiple sensor nodes can detect a single target of interest simultaneously and the data collected are usually highly correlated and redundant. If each node sends data to the base station, energy will be wasted and thus the network energy will be depleted quickly. Data aggregation is an important paradigm for compressing data so that the energy of the network is spent efficiently. In this paper, a novel data aggregation algorithm called Redundancy Elimination for Accurate Data Aggregation (READA) has been proposed. By exploiting the range of spatial correlations of data in the network, READA applies a grouping and compression mechanism to remove duplicate data in the aggregated set of data to be sent to the base station without largely losing the accuracy of the final aggregated data. One peculiarity of READA is that it uses a prediction model derived from cached values to confirm whether any outlier is actually an event which has occurred. From the various simulations conducted, it was observed that in READA the accuracy of data has been highly preserved taking into consideration the energy dissipated for aggregating the展开更多
Based on the atmospheric horizontal visibility data from forty-seven observational stations along the eastern coast of China near the Taiwan Strait and simultaneous NOAA/AVHRR multichannel satellite data during Januar...Based on the atmospheric horizontal visibility data from forty-seven observational stations along the eastern coast of China near the Taiwan Strait and simultaneous NOAA/AVHRR multichannel satellite data during January 2001 to December 2002, the spectral characters associated with visibility were investigated. Successful retrieval of visibility from multichannel NOAA/AVHRR data was performed using the principal component regression (PCR) method. A sample of retrieved visibility distribution was discussed with a sea fog process. The correlation coefficient between the observed and retrieved visibility was about 0.82, which is far higher than the 99.[KG-*7]9% confidence level by statistical test. The rate of successful retrieval is 94.98% of the 458 cases during 2001-2002. The error distribution showed that high visibilities were usually under-estimated and low visibilities were over-estimated and the relative error between the observed and retrieved visibilities was about 21.4%.展开更多
Based on surfaced-related multiple elimination( SRME),this research has derived the methods on multiples elimination in the inverse data space.Inverse data processing means moving seismic data from forward data space(...Based on surfaced-related multiple elimination( SRME),this research has derived the methods on multiples elimination in the inverse data space.Inverse data processing means moving seismic data from forward data space( FDS) to inverse data space( IDS).The surface-related multiples and primaries can then be separated in the IDS,since surface-related multiples will form a focus region in the IDS.Muting the multiples energy can achieve the purpose of multiples elimination and avoid the damage to primaries energy during the process of adaptive subtraction.Randomized singular value decomposition( RSVD) is used to enhance calculation speed and improve the accuracy in the conversion of FDS to IDS.The synthetic shot record of the salt dome model shows that the relationship between primaries and multiples is simple and clear,and RSVD can easily eliminate multiples and save primaries energy.Compared with conventional multiples elimination methods and ordinary methods of multiples elimination in the inverse data space,this technique has an advantage of high calculation speed and reliable outcomes.展开更多
This paper adopts satellite channel brightness temperature simulation to study M-estimator variational retrieval. This approach combines both the advantages of classical variational inversion and robust M-estimators. ...This paper adopts satellite channel brightness temperature simulation to study M-estimator variational retrieval. This approach combines both the advantages of classical variational inversion and robust M-estimators. Classical variational inversion depends on prior quality control to elim- inate outliers, and its errors follow a Gaussian distribution. We coupled the M-estimators to the framework of classical variational inversion to obtain a M-estimator variational inversion. The cost function contains the M-estimator to guarantee the robustness to outliers and improve the retrieval re- sults. The experimental evaluation adopts Feng Yun-3A (FY-3A) simulated data to add to the Gaussian and Non-Gaussian error. The variational in- version is used to obtain the inversion brightness temperature, and temperature and humidity data are used for validation. The preliminary results demonstrate the potential of M-estimator variational retrieval.展开更多
We develop a data driven method(probability model) to construct a composite shape descriptor by combining a pair of scale-based shape descriptors. The selection of a pair of scale-based shape descriptors is modeled as...We develop a data driven method(probability model) to construct a composite shape descriptor by combining a pair of scale-based shape descriptors. The selection of a pair of scale-based shape descriptors is modeled as the computation of the union of two events, i.e.,retrieving similar shapes by using a single scale-based shape descriptor. The pair of scale-based shape descriptors with the highest probability forms the composite shape descriptor. Given a shape database, the composite shape descriptors for the shapes constitute a planar point set.A VoR-Tree of the planar point set is then used as an indexing structure for efficient query operation. Experiments and comparisons show the effectiveness and efficiency of the proposed composite shape descriptor.展开更多
Direct volume rendering(DVR)is a technique that emphasizes structures of interest(SOIs)within a volume visually,while simultaneously depicting adjacent regional information,e.g.,the spatial location of a structure con...Direct volume rendering(DVR)is a technique that emphasizes structures of interest(SOIs)within a volume visually,while simultaneously depicting adjacent regional information,e.g.,the spatial location of a structure concerning its neighbors.In DVR,transfer function(TF)plays a key role by enabling accurate identification of SOIs interactively as well as ensuring appropriate visibility of them.TF generation typically involves non-intuitive trial-and-error optimization of rendering parameters,which is time-consuming and inefficient.Attempts at mitigating this manual process have led to approaches that make use of a knowledge database consisting of pre-designed TFs by domain experts.In these approaches,a user navigates the knowledge database to find the most suitable pre-designed TF for their input volume to visualize the SOIs.Although these approaches potentially reduce the workload to generate the TFs,they,however,require manual TF navigation of the knowledge database,as well as the likely fine tuning of the selected TF to suit the input.In this work,we propose a TF design approach,CBR-TF,where we introduce a new content-based retrieval(CBR)method to automatically navigate the knowledge database.Instead of pre-designed TFs,our knowledge database contains volumes with SOI labels.Given an input volume,our CBR-TF approach retrieves relevant volumes(with SOI labels)from the knowledge database;the retrieved labels are then used to generate and optimize TFs of the input.This approach largely reduces manual TF navigation and fine tuning.For our CBR-TF approach,we introduce a novel volumetric image feature which includes both a local primitive intensity profile along the SOIs and regional spatial semantics available from the co-planar images to the profile.For the regional spatial semantics,we adopt a convolutional neural network to obtain high-level image feature representations.For the intensity profile,we extend the dynamic time warping technique to address subtle alignment differences between similar profiles(SOIs).Finally,we propose a two-stage CBR scheme to enable the use of these two different feature representations in a complementary manner,thereby improving SOI retrieval performance.We demonstrate the capabilities of our CBR-TF approach with comparison with a conventional approach in visualization,where an intensity profile matching algorithm is used,and also with potential use-cases in medical volume visualization.展开更多
Interference in the data of geochemical hydrocarbon exploration is a large obstacle for anomaly recognition. The multiresolution analysis of wavelet analysis can extract the information at different scales so as to pr...Interference in the data of geochemical hydrocarbon exploration is a large obstacle for anomaly recognition. The multiresolution analysis of wavelet analysis can extract the information at different scales so as to provide a powerful tool for information analysis and processing. Based on the analysis of the geometric nature of hydrocarbon anomalies and background, Mallat wavelet and symmetric border treatment are selected and data pre-processing (logarithm-normalization) is established. This approach provide good results in Shandong and Inner Mongolia, China. It is demonstrated that this approach overcome the disadvantage of backgound variation in the window (interference in window), used in moving average, frame filtering and spatial and scaling modeling methods.展开更多
It is well known that retrieval of parameters is usually ill-posed and highly nonlinear, so parameter retrieval problems are very difficult. There are still many important theoretical issues under research, although g...It is well known that retrieval of parameters is usually ill-posed and highly nonlinear, so parameter retrieval problems are very difficult. There are still many important theoretical issues under research, although great success has been achieved in data assimilation in meteorology and oceanography. This paper reviews the recent research on parameter retrieval, especially that of the authors. First, some concepts and issues of parameter retrieval are introduced and the state-of-the-art parameter retrieval technology in meteorology and oceanography is reviewed briefly, and then atmospheric and oceanic parameters are retrieved using the variational data assimilation method combined with the regularization techniques in four examples: retrieval of the vertical eddy diffusion coefficient; of the turbulivity of the atmospheric boundary layer; of wind from Doppler radar data, and of the physical process parameters. Model parameter retrieval with global and local observations is also introduced.展开更多
A kind of single linked lists named aggregative chain is introduced to the algorithm, thus improving the architecture of FP tree. The new FP tree is a one-way tree and only the pointers that point its parent at each n...A kind of single linked lists named aggregative chain is introduced to the algorithm, thus improving the architecture of FP tree. The new FP tree is a one-way tree and only the pointers that point its parent at each node are kept. Route information of different nodes in a same item are compressed into aggregative chains so that the frequent patterns will be produced in aggregative chains without generating node links and conditional pattern bases. An example of Web key words retrieval is given to analyze and verify the frequent pattern algorithm in this paper.展开更多
The drastic growth of coastal observation sensors results in copious data that provide weather information.The intricacies in sensor-generated big data are heterogeneity and interpretation,driving high-end Information...The drastic growth of coastal observation sensors results in copious data that provide weather information.The intricacies in sensor-generated big data are heterogeneity and interpretation,driving high-end Information Retrieval(IR)systems.The Semantic Web(SW)can solve this issue by integrating data into a single platform for information exchange and knowledge retrieval.This paper focuses on exploiting the SWbase systemto provide interoperability through ontologies by combining the data concepts with ontology classes.This paper presents a 4-phase weather data model:data processing,ontology creation,SW processing,and query engine.The developed Oceanographic Weather Ontology helps to enhance data analysis,discovery,IR,and decision making.In addition to that,it also evaluates the developed ontology with other state-of-the-art ontologies.The proposed ontology’s quality has improved by 39.28%in terms of completeness,and structural complexity has decreased by 45.29%,11%and 37.7%in Precision and Accuracy.Indian Meteorological Satellite INSAT-3D’s ocean data is a typical example of testing the proposed model.The experimental result shows the effectiveness of the proposed data model and its advantages in machine understanding and IR.展开更多
With the increasing popularity of cloud computing,privacy has become one of the key problem in cloud security.When data is outsourced to the cloud,for data owners,they need to ensure the security of their privacy;for ...With the increasing popularity of cloud computing,privacy has become one of the key problem in cloud security.When data is outsourced to the cloud,for data owners,they need to ensure the security of their privacy;for cloud service providers,they need some information of the data to provide high QoS services;and for authorized users,they need to access to the true value of data.The existing privacy-preserving methods can't meet all the needs of the three parties at the same time.To address this issue,we propose a retrievable data perturbation method and use it in the privacy-preserving in data outsourcing in cloud computing.Our scheme comes in four steps.Firstly,an improved random generator is proposed to generate an accurate "noise".Next,a perturbation algorithm is introduced to add noise to the original data.By doing this,the privacy information is hidden,but the mean and covariance of data which the service providers may need remain unchanged.Then,a retrieval algorithm is proposed to get the original data back from the perturbed data.Finally,we combine the retrievable perturbation with the access control process to ensure only the authorized users can retrieve the original data.The experiments show that our scheme perturbs date correctly,efficiently,and securely.展开更多
In this paper we investigate the impact of the Atmospheric Infra-Red Sounder (AIRS) temperature retrievals on data assimilation and the resulting forecasts using the four-dimensional Local Ensemble Transform Kalman Fi...In this paper we investigate the impact of the Atmospheric Infra-Red Sounder (AIRS) temperature retrievals on data assimilation and the resulting forecasts using the four-dimensional Local Ensemble Transform Kalman Filter (LETKF) data assimilation scheme and a reduced resolution version of the NCEP Global Forecast System (GFS).Our results indicate that the AIRS temperature retrievals have a significant and consistent positive impact in the Southern Hemispheric extratropics on both analyses and forecasts,which is found not only in the temperature field but also in other variables.In tropics and the Northern Hemispheric extratropics these impacts are smaller,but are still generally positive or neutral.展开更多
In this paper, we present machine learning algorithms and systems for similar video retrieval. Here, the query is itself a video. For the similarity measurement, exemplars, or representative frames in each video, are ...In this paper, we present machine learning algorithms and systems for similar video retrieval. Here, the query is itself a video. For the similarity measurement, exemplars, or representative frames in each video, are extracted by unsupervised learning. For this learning, we chose the order-aware competitive learning. After obtaining a set of exemplars for each video, the similarity is computed. Because the numbers and positions of the exemplars are different in each video, we use a similarity computing method called M-distance, which generalizes existing global and local alignment methods using followers to the exemplars. To represent each frame in the video, this paper emphasizes the Frame Signature of the ISO/IEC standard so that the total system, along with its graphical user interface, becomes practical. Experiments on the detection of inserted plagiaristic scenes showed excellent precision-recall curves, with precision values very close to 1. Thus, the proposed system can work as a plagiarism detector for videos. In addition, this method can be regarded as the structuring of unstructured data via numerical labeling by exemplars. Finally, further sophistication of this labeling is discussed.展开更多
Mining the content from an information database provides challenging solutions to the industry experts and researchers, due to the overcrowded information in huge data. In web searching, the information retrieved is n...Mining the content from an information database provides challenging solutions to the industry experts and researchers, due to the overcrowded information in huge data. In web searching, the information retrieved is not an appropriate, because it gives ambiguous information for the user query, and the user cannot get relevant information within the stipulated time. To overcome these issues, we propose a new methodology for information retrieval EPCRR by providing the top most exact information to the user, by using the collaborative clustered automated filter which makes use of the collaborative data set and filter works on the prediction by providing the highest ranking for the exact data retrieved. The retrieval works on the basis of recommendation of data which consists of relevant data set with highest priority from the cluster of data which is on high usage. In this work, we make use of the automated wrapper which works similar to the meta crawler functionality and it obtains the content in the semantic usage data format. Obtained information from the user to the agent will be ranked based on the Enabled Pile clustered data with respect to the metadata information from the agent and end-user. The information is given to the end-user with the top most ranking data within the stipulated time and the remaining top information will be moved to the data repository for future use. The data collected will remain stable based on the user preference and works on the intelligence system approach in which the user can choose any information under any instances and can be provided with suitable high range of exact content. In this approach, we find that the proposed algorithm has produced better results than existing work and it costs less online computation time.展开更多
Decision Support Systems(DSS)are man-machine interaction systems,which support the de-cision-makers to solve the unstructured and semi-structured decisions,this paper advances that thefunction of problem-oriented info...Decision Support Systems(DSS)are man-machine interaction systems,which support the de-cision-makers to solve the unstructured and semi-structured decisions,this paper advances that thefunction of problem-oriented information retrieval DSS can meet the needs of enterprise’s topmanagement effectively in comparison with other information retrieval functions,in accordancewith the features of supporting information for decision.An architecture of this system is presented,which dissolves a problem put forward or recognized by the user into the problem recognized by thecomputer,forming retrieval tactics and searching the data the user needs.Designed and developedaccording to the architecture of this system,a prototype system is introduced,which is CF Econom-ic Environment Information Retrieval DSS.展开更多
基金supported by the National Key R&D Program of China(2020YFB0905900).
文摘Operation control of power systems has become challenging with an increase in the scale and complexity of power distribution systems and extensive access to renewable energy.Therefore,improvement of the ability of data-driven operation management,intelligent analysis,and mining is urgently required.To investigate and explore similar regularities of the historical operating section of the power distribution system and assist the power grid in obtaining high-value historical operation,maintenance experience,and knowledge by rule and line,a neural information retrieval model with an attention mechanism is proposed based on graph data computing technology.Based on the processing flow of the operating data of the power distribution system,a technical framework of neural information retrieval is established.Combined with the natural graph characteristics of the power distribution system,a unified graph data structure and a data fusion method of data access,data complement,and multi-source data are constructed.Further,a graph node feature-embedding representation learning algorithm and a neural information retrieval algorithm model are constructed.The neural information retrieval algorithm model is trained and tested using the generated graph node feature representation vector set.The model is verified on the operating section of the power distribution system of a provincial grid area.The results show that the proposed method demonstrates high accuracy in the similarity matching of historical operation characteristics and effectively supports intelligent fault diagnosis and elimination in power distribution systems.
基金supported and granted by the Ministry of Science and Technology,Taiwan(MOST110-2622-E-390-001 and MOST109-2622-E-390-002-CC3).
文摘Big data analytics in business intelligence do not provide effective data retrieval methods and job scheduling that will cause execution inefficiency and low system throughput.This paper aims to enhance the capability of data retrieval and job scheduling to speed up the operation of big data analytics to overcome inefficiency and low throughput problems.First,integrating stacked sparse autoencoder and Elasticsearch indexing explored fast data searching and distributed indexing,which reduces the search scope of the database and dramatically speeds up data searching.Next,exploiting a deep neural network to predict the approximate execution time of a job gives prioritized job scheduling based on the shortest job first,which reduces the average waiting time of job execution.As a result,the proposed data retrieval approach outperforms the previous method using a deep autoencoder and Solr indexing,significantly improving the speed of data retrieval up to 53%and increasing system throughput by 53%.On the other hand,the proposed job scheduling algorithmdefeats both first-in-first-out andmemory-sensitive heterogeneous early finish time scheduling algorithms,effectively shortening the average waiting time up to 5%and average weighted turnaround time by 19%,respectively.
基金The National Natural Science Foundation of China under contract Nos 41330960 and 41276193 and 41206184
文摘In recent years, the rapid decline of Arctic sea ice area (SIA) and sea ice extent (SIE), especially for the multiyear (MY) ice, has led to significant effect on climate change. The accurate retrieval of MY ice concentration retrieval is very important and challenging to understand the ongoing changes. Three MY ice concentration retrieval algorithms were systematically evaluated. A similar total ice concentration was yielded by these algorithms, while the retrieved MY sea ice concentrations differs from each other. The MY SIA derived from NASA TEAM algorithm is relatively stable. Other two algorithms created seasonal fluctuations of MY SIA, particularly in autumn and winter. In this paper, we proposed an ice concentration retrieval algorithm, which developed the NASA TEAM algorithm by adding to use AMSR-E 6.9 GHz brightness temperature data and sea ice concentration using 89.0 GHz data. Comparison with the reference MY SIA from reference MY ice, indicates that the mean difference and root mean square (rms) difference of MY SIA derived from the algorithm of this study are 0.65×10^6 km^2 and 0.69×10^6 km^2 during January to March, -0.06×10^6 km^2 and 0.14×10^6 km^2during September to December respectively. Comparison with MY SIE obtained from weekly ice age data provided by University of Colorado show that, the mean difference and rms difference are 0.69×10^6 km^2 and 0.84×10^6 km^2, respectively. The developed algorithm proposed in this study has smaller difference compared with the reference MY ice and MY SIE from ice age data than the Wang's, Lomax' and NASA TEAM algorithms.
文摘In monitoring systems, multiple sensor nodes can detect a single target of interest simultaneously and the data collected are usually highly correlated and redundant. If each node sends data to the base station, energy will be wasted and thus the network energy will be depleted quickly. Data aggregation is an important paradigm for compressing data so that the energy of the network is spent efficiently. In this paper, a novel data aggregation algorithm called Redundancy Elimination for Accurate Data Aggregation (READA) has been proposed. By exploiting the range of spatial correlations of data in the network, READA applies a grouping and compression mechanism to remove duplicate data in the aggregated set of data to be sent to the base station without largely losing the accuracy of the final aggregated data. One peculiarity of READA is that it uses a prediction model derived from cached values to confirm whether any outlier is actually an event which has occurred. From the various simulations conducted, it was observed that in READA the accuracy of data has been highly preserved taking into consideration the energy dissipated for aggregating the
基金This research is supported by the National High Technology Development Project (863) of China (Grant No. 2002AA639500) the Natural Science Foundation of Guangdong Province (Grant No. 032212)+1 种基金 National Basic Research Program of China (973 Program) (No. 2005CB422301) Program for New Century Excellent Talents in University ( NCET-05-0591 ).
文摘Based on the atmospheric horizontal visibility data from forty-seven observational stations along the eastern coast of China near the Taiwan Strait and simultaneous NOAA/AVHRR multichannel satellite data during January 2001 to December 2002, the spectral characters associated with visibility were investigated. Successful retrieval of visibility from multichannel NOAA/AVHRR data was performed using the principal component regression (PCR) method. A sample of retrieved visibility distribution was discussed with a sea fog process. The correlation coefficient between the observed and retrieved visibility was about 0.82, which is far higher than the 99.[KG-*7]9% confidence level by statistical test. The rate of successful retrieval is 94.98% of the 458 cases during 2001-2002. The error distribution showed that high visibilities were usually under-estimated and low visibilities were over-estimated and the relative error between the observed and retrieved visibilities was about 21.4%.
文摘Based on surfaced-related multiple elimination( SRME),this research has derived the methods on multiples elimination in the inverse data space.Inverse data processing means moving seismic data from forward data space( FDS) to inverse data space( IDS).The surface-related multiples and primaries can then be separated in the IDS,since surface-related multiples will form a focus region in the IDS.Muting the multiples energy can achieve the purpose of multiples elimination and avoid the damage to primaries energy during the process of adaptive subtraction.Randomized singular value decomposition( RSVD) is used to enhance calculation speed and improve the accuracy in the conversion of FDS to IDS.The synthetic shot record of the salt dome model shows that the relationship between primaries and multiples is simple and clear,and RSVD can easily eliminate multiples and save primaries energy.Compared with conventional multiples elimination methods and ordinary methods of multiples elimination in the inverse data space,this technique has an advantage of high calculation speed and reliable outcomes.
基金Supported by Special Scientific Research Fund of Meteorological Public Welfare Profession of China(GYHY201406028)Meteorological Open Research Fund for Huaihe River Basin(HRM201407)Anhui Meteorological Bureau Science and Technology Development Fund(RC201506)
文摘This paper adopts satellite channel brightness temperature simulation to study M-estimator variational retrieval. This approach combines both the advantages of classical variational inversion and robust M-estimators. Classical variational inversion depends on prior quality control to elim- inate outliers, and its errors follow a Gaussian distribution. We coupled the M-estimators to the framework of classical variational inversion to obtain a M-estimator variational inversion. The cost function contains the M-estimator to guarantee the robustness to outliers and improve the retrieval re- sults. The experimental evaluation adopts Feng Yun-3A (FY-3A) simulated data to add to the Gaussian and Non-Gaussian error. The variational in- version is used to obtain the inversion brightness temperature, and temperature and humidity data are used for validation. The preliminary results demonstrate the potential of M-estimator variational retrieval.
基金supported by the National Key R&D Plan of China(2016YFB1001501)
文摘We develop a data driven method(probability model) to construct a composite shape descriptor by combining a pair of scale-based shape descriptors. The selection of a pair of scale-based shape descriptors is modeled as the computation of the union of two events, i.e.,retrieving similar shapes by using a single scale-based shape descriptor. The pair of scale-based shape descriptors with the highest probability forms the composite shape descriptor. Given a shape database, the composite shape descriptors for the shapes constitute a planar point set.A VoR-Tree of the planar point set is then used as an indexing structure for efficient query operation. Experiments and comparisons show the effectiveness and efficiency of the proposed composite shape descriptor.
基金supported by the Korea Health Technology Research and Development Project through the Korea Health Industry Development Institute under Grant No.HI22C1651the National Research Foundation of Korea(NRF)under Grant No.2021R1F1A1059554the Culture,Sports and Tourism Research and Development Program through the Korea Creative Content Agency Grant funded by the Ministry of Culture,Sports and Tourism of Korea under Grant No.RS-2023-00227648.
文摘Direct volume rendering(DVR)is a technique that emphasizes structures of interest(SOIs)within a volume visually,while simultaneously depicting adjacent regional information,e.g.,the spatial location of a structure concerning its neighbors.In DVR,transfer function(TF)plays a key role by enabling accurate identification of SOIs interactively as well as ensuring appropriate visibility of them.TF generation typically involves non-intuitive trial-and-error optimization of rendering parameters,which is time-consuming and inefficient.Attempts at mitigating this manual process have led to approaches that make use of a knowledge database consisting of pre-designed TFs by domain experts.In these approaches,a user navigates the knowledge database to find the most suitable pre-designed TF for their input volume to visualize the SOIs.Although these approaches potentially reduce the workload to generate the TFs,they,however,require manual TF navigation of the knowledge database,as well as the likely fine tuning of the selected TF to suit the input.In this work,we propose a TF design approach,CBR-TF,where we introduce a new content-based retrieval(CBR)method to automatically navigate the knowledge database.Instead of pre-designed TFs,our knowledge database contains volumes with SOI labels.Given an input volume,our CBR-TF approach retrieves relevant volumes(with SOI labels)from the knowledge database;the retrieved labels are then used to generate and optimize TFs of the input.This approach largely reduces manual TF navigation and fine tuning.For our CBR-TF approach,we introduce a novel volumetric image feature which includes both a local primitive intensity profile along the SOIs and regional spatial semantics available from the co-planar images to the profile.For the regional spatial semantics,we adopt a convolutional neural network to obtain high-level image feature representations.For the intensity profile,we extend the dynamic time warping technique to address subtle alignment differences between similar profiles(SOIs).Finally,we propose a two-stage CBR scheme to enable the use of these two different feature representations in a complementary manner,thereby improving SOI retrieval performance.We demonstrate the capabilities of our CBR-TF approach with comparison with a conventional approach in visualization,where an intensity profile matching algorithm is used,and also with potential use-cases in medical volume visualization.
文摘Interference in the data of geochemical hydrocarbon exploration is a large obstacle for anomaly recognition. The multiresolution analysis of wavelet analysis can extract the information at different scales so as to provide a powerful tool for information analysis and processing. Based on the analysis of the geometric nature of hydrocarbon anomalies and background, Mallat wavelet and symmetric border treatment are selected and data pre-processing (logarithm-normalization) is established. This approach provide good results in Shandong and Inner Mongolia, China. It is demonstrated that this approach overcome the disadvantage of backgound variation in the window (interference in window), used in moving average, frame filtering and spatial and scaling modeling methods.
基金This work was supported by the National Natural Science Foundation of China (Grant No. 90411006)by the Shanghai Science and Technology Association (Grant No. 02DJ14032).
文摘It is well known that retrieval of parameters is usually ill-posed and highly nonlinear, so parameter retrieval problems are very difficult. There are still many important theoretical issues under research, although great success has been achieved in data assimilation in meteorology and oceanography. This paper reviews the recent research on parameter retrieval, especially that of the authors. First, some concepts and issues of parameter retrieval are introduced and the state-of-the-art parameter retrieval technology in meteorology and oceanography is reviewed briefly, and then atmospheric and oceanic parameters are retrieved using the variational data assimilation method combined with the regularization techniques in four examples: retrieval of the vertical eddy diffusion coefficient; of the turbulivity of the atmospheric boundary layer; of wind from Doppler radar data, and of the physical process parameters. Model parameter retrieval with global and local observations is also introduced.
基金Supported by the Natural Science Foundation ofLiaoning Province (20042020)
文摘A kind of single linked lists named aggregative chain is introduced to the algorithm, thus improving the architecture of FP tree. The new FP tree is a one-way tree and only the pointers that point its parent at each node are kept. Route information of different nodes in a same item are compressed into aggregative chains so that the frequent patterns will be produced in aggregative chains without generating node links and conditional pattern bases. An example of Web key words retrieval is given to analyze and verify the frequent pattern algorithm in this paper.
基金This work is financially supported by the Ministry of Earth Science(MoES),Government of India,(Grant.No.MoES/36/OOIS/Extra/45/2015),URL:https://www.moes.gov.in。
文摘The drastic growth of coastal observation sensors results in copious data that provide weather information.The intricacies in sensor-generated big data are heterogeneity and interpretation,driving high-end Information Retrieval(IR)systems.The Semantic Web(SW)can solve this issue by integrating data into a single platform for information exchange and knowledge retrieval.This paper focuses on exploiting the SWbase systemto provide interoperability through ontologies by combining the data concepts with ontology classes.This paper presents a 4-phase weather data model:data processing,ontology creation,SW processing,and query engine.The developed Oceanographic Weather Ontology helps to enhance data analysis,discovery,IR,and decision making.In addition to that,it also evaluates the developed ontology with other state-of-the-art ontologies.The proposed ontology’s quality has improved by 39.28%in terms of completeness,and structural complexity has decreased by 45.29%,11%and 37.7%in Precision and Accuracy.Indian Meteorological Satellite INSAT-3D’s ocean data is a typical example of testing the proposed model.The experimental result shows the effectiveness of the proposed data model and its advantages in machine understanding and IR.
基金supported in part by NSFC under Grant No.61172090National Science and Technology Major Project under Grant 2012ZX03002001+3 种基金Research Fund for the Doctoral Program of Higher Education of China under Grant No.20120201110013Scientific and Technological Project in Shaanxi Province under Grant(No.2012K06-30, No.2014JQ8322)Basic Science Research Fund in Xi'an Jiaotong University(No. XJJ2014049,No.XKJC2014008)Shaanxi Science and Technology Innovation Project (2013SZS16-Z01/P01/K01)
文摘With the increasing popularity of cloud computing,privacy has become one of the key problem in cloud security.When data is outsourced to the cloud,for data owners,they need to ensure the security of their privacy;for cloud service providers,they need some information of the data to provide high QoS services;and for authorized users,they need to access to the true value of data.The existing privacy-preserving methods can't meet all the needs of the three parties at the same time.To address this issue,we propose a retrievable data perturbation method and use it in the privacy-preserving in data outsourcing in cloud computing.Our scheme comes in four steps.Firstly,an improved random generator is proposed to generate an accurate "noise".Next,a perturbation algorithm is introduced to add noise to the original data.By doing this,the privacy information is hidden,but the mean and covariance of data which the service providers may need remain unchanged.Then,a retrieval algorithm is proposed to get the original data back from the perturbed data.Finally,we combine the retrievable perturbation with the access control process to ensure only the authorized users can retrieve the original data.The experiments show that our scheme perturbs date correctly,efficiently,and securely.
基金National Natural Science Foundation of China (40975067)973 Program (2009CB421500)+1 种基金CMA Grant GYHY200806029NASA grant NNX07AM97G in U.S.A
文摘In this paper we investigate the impact of the Atmospheric Infra-Red Sounder (AIRS) temperature retrievals on data assimilation and the resulting forecasts using the four-dimensional Local Ensemble Transform Kalman Filter (LETKF) data assimilation scheme and a reduced resolution version of the NCEP Global Forecast System (GFS).Our results indicate that the AIRS temperature retrievals have a significant and consistent positive impact in the Southern Hemispheric extratropics on both analyses and forecasts,which is found not only in the temperature field but also in other variables.In tropics and the Northern Hemispheric extratropics these impacts are smaller,but are still generally positive or neutral.
文摘In this paper, we present machine learning algorithms and systems for similar video retrieval. Here, the query is itself a video. For the similarity measurement, exemplars, or representative frames in each video, are extracted by unsupervised learning. For this learning, we chose the order-aware competitive learning. After obtaining a set of exemplars for each video, the similarity is computed. Because the numbers and positions of the exemplars are different in each video, we use a similarity computing method called M-distance, which generalizes existing global and local alignment methods using followers to the exemplars. To represent each frame in the video, this paper emphasizes the Frame Signature of the ISO/IEC standard so that the total system, along with its graphical user interface, becomes practical. Experiments on the detection of inserted plagiaristic scenes showed excellent precision-recall curves, with precision values very close to 1. Thus, the proposed system can work as a plagiarism detector for videos. In addition, this method can be regarded as the structuring of unstructured data via numerical labeling by exemplars. Finally, further sophistication of this labeling is discussed.
文摘Mining the content from an information database provides challenging solutions to the industry experts and researchers, due to the overcrowded information in huge data. In web searching, the information retrieved is not an appropriate, because it gives ambiguous information for the user query, and the user cannot get relevant information within the stipulated time. To overcome these issues, we propose a new methodology for information retrieval EPCRR by providing the top most exact information to the user, by using the collaborative clustered automated filter which makes use of the collaborative data set and filter works on the prediction by providing the highest ranking for the exact data retrieved. The retrieval works on the basis of recommendation of data which consists of relevant data set with highest priority from the cluster of data which is on high usage. In this work, we make use of the automated wrapper which works similar to the meta crawler functionality and it obtains the content in the semantic usage data format. Obtained information from the user to the agent will be ranked based on the Enabled Pile clustered data with respect to the metadata information from the agent and end-user. The information is given to the end-user with the top most ranking data within the stipulated time and the remaining top information will be moved to the data repository for future use. The data collected will remain stable based on the user preference and works on the intelligence system approach in which the user can choose any information under any instances and can be provided with suitable high range of exact content. In this approach, we find that the proposed algorithm has produced better results than existing work and it costs less online computation time.
文摘Decision Support Systems(DSS)are man-machine interaction systems,which support the de-cision-makers to solve the unstructured and semi-structured decisions,this paper advances that thefunction of problem-oriented information retrieval DSS can meet the needs of enterprise’s topmanagement effectively in comparison with other information retrieval functions,in accordancewith the features of supporting information for decision.An architecture of this system is presented,which dissolves a problem put forward or recognized by the user into the problem recognized by thecomputer,forming retrieval tactics and searching the data the user needs.Designed and developedaccording to the architecture of this system,a prototype system is introduced,which is CF Econom-ic Environment Information Retrieval DSS.