From the beginning,the process of research and its publication is an ever-growing phenomenon and with the emergence of web technologies,its growth rate is overwhelming.On a rough estimate,more than thirty thousand res...From the beginning,the process of research and its publication is an ever-growing phenomenon and with the emergence of web technologies,its growth rate is overwhelming.On a rough estimate,more than thirty thousand research journals have been issuing around four million papers annually on average.Search engines,indexing services,and digital libraries have been searching for such publications over the web.Nevertheless,getting the most relevant articles against the user requests is yet a fantasy.It is mainly because the articles are not appropriately indexed based on the hierarchies of granular subject classification.To overcome this issue,researchers are striving to investigate new techniques for the classification of the research articles especially,when the complete article text is not available(a case of nonopen access articles).The proposed study aims to investigate the multilabel classification over the available metadata in the best possible way and to assess,“to what extent metadata-based features can perform in contrast to content-based approaches.”In this regard,novel techniques for investigating multilabel classification have been proposed,developed,and evaluated on metadata such as the Title and Keywords of the articles.The proposed technique has been assessed for two diverse datasets,namely,from the Journal of universal computer science(J.UCS)and the benchmark dataset comprises of the articles published by the Association for computing machinery(ACM).The proposed technique yields encouraging results in contrast to the state-ofthe-art techniques in the literature.展开更多
Fine-grained silt is widely distributed in the Huanghe River Delta(HRD)in China,and the sedimentary structure is complex,meaning that the clay content in the silt is variable.The piezocone penetration test(CPTu)is the...Fine-grained silt is widely distributed in the Huanghe River Delta(HRD)in China,and the sedimentary structure is complex,meaning that the clay content in the silt is variable.The piezocone penetration test(CPTu)is the most widely approved in situ test method.It can be used to invert soil properties and interpret soil behavior.To analyse the strength properties of surface sediments in the HRD,this paper evaluated the friction angle and its inversion formula through the CPTu penetration test and monotonic simple shear test and other soil unit experiments.The evaluation showed that the empirical formula proposed by Kulhawy and Mayne had better prediction and inversion effect.The HRD silts with clay contents of 9.2%,21.4%and 30.3%were selected as samples for the CPTu variable rate penetration test.The results show as follows.(1)The effects of the clay content on the tip resistance and the pore pressure of silt under different penetration rates were summarized.The tip resistance Q_t is strongly dependent on the clay content of the silt,the B_(q)value of the silt tends to 0 and is not significantly affected by the change of the CPTu penetration rate.(2)Five soil behavior type classification charts and three soil behavior type indexes based on CPTu data were evaluated.The results show that the soil behavior type classification chart based on soil behavior type index ISBT,the Robertson 2010 behavior type classification chart are more suitable for the silty soil in the HRD.展开更多
Many text classifications depend on statistical term measures to implement document representation. Such document representations ignore the lexical semantic contents of terms and the distilled mutual information, lea...Many text classifications depend on statistical term measures to implement document representation. Such document representations ignore the lexical semantic contents of terms and the distilled mutual information, leading to text classification errors.This work proposed a document representation method, Word Net-based lexical semantic VSM, to solve the problem. Using Word Net,this method constructed a data structure of semantic-element information to characterize lexical semantic contents, and adjusted EM modeling to disambiguate word stems. Then, in the lexical-semantic space of corpus, lexical-semantic eigenvector of document representation was built by calculating the weight of each synset, and applied to a widely-recognized algorithm NWKNN. On text corpus Reuter-21578 and its adjusted version of lexical replacement, the experimental results show that the lexical-semantic eigenvector performs F1 measure and scales of dimension better than term-statistic eigenvector based on TF-IDF. Formation of document representation eigenvectors ensures the method a wide prospect of classification applications in text corpus analysis.展开更多
AIM:To support probe-based confocal laser endomi-croscopy (pCLE) diagnosis by designing software for the automated classification of colonic polyps. METHODS:Intravenous fluorescein pCLE imaging of colorectal lesions w...AIM:To support probe-based confocal laser endomi-croscopy (pCLE) diagnosis by designing software for the automated classification of colonic polyps. METHODS:Intravenous fluorescein pCLE imaging of colorectal lesions was performed on patients under-going screening and surveillance colonoscopies, followed by polypectomies. All resected specimens were reviewed by a reference gastrointestinal pathologist blinded to pCLE information. Histopathology was used as the criterion standard for the differentiation between neoplastic and non-neoplastic lesions. The pCLE video sequences, recorded for each polyp, were analyzed off-line by 2 expert endoscopists who were blinded to the endoscopic characteristics and histopathology. These pCLE videos, along with their histopathology diagnosis, were used to train the automated classification software which is a content-based image retrieval technique followed by k-nearest neighbor classification. The performance of the off-line diagnosis of pCLE videos established by the 2 expert endoscopists was compared with that of automated pCLE software classification. All evaluations were performed using leave-one-patient- out cross-validation to avoid bias. RESULTS:Colorectal lesions (135) were imaged in 71 patients. Based on histopathology, 93 of these 135 lesions were neoplastic and 42 were non-neoplastic. The study found no statistical significance for the difference between the performance of automated pCLE software classification (accuracy 89.6%, sensitivity 92.5%, specificity 83.3%, using leave-one-patient-out cross-validation) and the performance of the off-line diagnosis of pCLE videos established by the 2 expert endoscopists (accuracy 89.6%, sensitivity 91.4%, specificity 85.7%). There was very low power (< 6%) to detect the observed differences. The 95% confidence intervals for equivalence testing were:-0.073 to 0.073 for accuracy, -0.068 to 0.089 for sensitivity and -0.18 to 0.13 for specificity. The classification software proposed in this study is not a "black box" but an informative tool based on the query by example model that produces, as intermediate results, visually similar annotated videos that are directly interpretable by the endoscopist. CONCLUSION:The proposed software for automated classification of pCLE videos of colonic polyps achieves high performance, comparable to that of off-line diagnosis of pCLE videos established by expert endoscopists.展开更多
The present study is inscribed within the framework of the geotechnical characterization of the soils of the Santchou plain, their classification for employment as pavement subgrade, various identification tests were ...The present study is inscribed within the framework of the geotechnical characterization of the soils of the Santchou plain, their classification for employment as pavement subgrade, various identification tests were carried out on the samples. The results obtained showed that with a wide range of different grain sizes, the studied soils showed low content in clay grains and dominance of either sand grains or silt grains, this can be explaining how most of these soil are poorly graded. According to the USDA textural classification, the grain size distribution of these soils makes them to be classified as Silty Loam types to Sandy Loam types. Despite of their organic matter content which is less than 10%, according to their respective methylene blue values, the soils studied along the section should be mainly loamy soil of medium plasticity to clayed soil, therefore showing a sensibility of its behavior to variation of water content. That last one is confirmed by the consistency parameters of these soils which show intermediate plasticity to highly plastic. Also, the bearing capacity proposed by these soils at their respective optimum dry densities is relatively small, although most of these experimental CBR values of the studied soils are more important than the ones prescribed by the AASHTO Classification system for A5, A6, and A7 types, and the French Highway Earthworks Manual Classifications system (GTR) for the corresponding A2 and A3 types.展开更多
On the basis of the data obtained from the comprehensive Kuroshio surveys in 1987-1988,this paper analyses the oceanographic characteristics in the area (125°-130° E,27°-31° N) of the continental s...On the basis of the data obtained from the comprehensive Kuroshio surveys in 1987-1988,this paper analyses the oceanographic characteristics in the area (125°-130° E,27°-31° N) of the continental shelf edge of the East China Sea (E. C. S. ) and its adjacent waters and discusses the effects of the Kuroshio front,thermocline and upwelling of the Kuroshio subsurface water on the distribution of standing stock of phytoplankton (chlorophyll-a). The distribution of high content of chlorophylly-a has been detected at 20-50 in depth in the water body on the left side of the Kuroshio front in the continental shelf edge waters of the E. C. S. The high content of chlorophyll-a spreads from the shelf area to the Kuroshio area in the form of a tongue and connects with the maximum layer of subsurface chlorophyll-a of the Kuroshio and pelagic sea. The author considers that the formation of the distribution of high content chlorophyll-a in this area results from the bottom topography and oceanic environment and there are close correlations between the high content of chlorophyll-a and the light-nutrient environment.展开更多
Packet classification (PC) has become the main method to support the quality of service and security of network application. And two-dimensional prefix packet classification (PPC) is the popular one. This paper analyz...Packet classification (PC) has become the main method to support the quality of service and security of network application. And two-dimensional prefix packet classification (PPC) is the popular one. This paper analyzes the problem of ruler conflict, and then presents a TCAMbased two-dimensional PPC algorithm. This algorithm makes use of the parallelism of TCAM to lookup the longest prefix in one instruction cycle. Then it uses a memory image and associated data structures to eliminate the conflicts between rulers, and performs a fast two-dimensional PPC.Compared with other algorithms, this algorithm has the least time complexity and less space complexity.展开更多
The feature of Ternary Content Addressable Memories(TCAMs) makes them particularly attractive for IP address lookup and packet classification applications in a router system. However,the limitations of TCAMs impede th...The feature of Ternary Content Addressable Memories(TCAMs) makes them particularly attractive for IP address lookup and packet classification applications in a router system. However,the limitations of TCAMs impede their utilization. In this paper,the solutions for decreasing the power consumption and avoiding entry expansion in range matching are addressed. Experimental results demonstrate that the proposed techniques can make some big improvements on the performance of TCAMs in IP address lookup and packet classification.展开更多
An embryo classification system for Pinus tabuliformis Carr. was established by time-tracing sam- pling and observation of zygotic embryos. The zygotic embryos were divided into nine stages. Key elements of the zygoti...An embryo classification system for Pinus tabuliformis Carr. was established by time-tracing sam- pling and observation of zygotic embryos. The zygotic embryos were divided into nine stages. Key elements of the zygotic embryo and female gametophyte (FG) tissue of P. tabuliformis were analyzed, using inductively coupled plasma-emission spectroscopy. Several elements--includ- ing aluminum, iron, sodium, and copper--are found in both embryo and FG tissue. Boron, phosphorus, magnesium,zinc, and calcium are also required for zygotic embryo development and therefore accumulated. Manganese is selectively excluded from the embryo. The zygotic embryo development needs a low-sodium and high-potassium nutrition proportion. The results of elemental analysis from zygotic embryos and FGs can provide the mineral targets for optimizing the formulation of culture medium for somatic embryogenesis.展开更多
Engineers often need to look for the right pieces of information by sifting through long engineering documents, It is a very tiring and time-consuming job. To address this issue, researchers are increasingly devoting ...Engineers often need to look for the right pieces of information by sifting through long engineering documents, It is a very tiring and time-consuming job. To address this issue, researchers are increasingly devoting their attention to new ways to help information users, including engineers, to access and retrieve document content. The research reported in this paper explores how to use the key technologies of document decomposition (study of document structure), document mark-up (with EXtensible Mark- up Language (XML), HyperText Mark-up Language (HTML), and Scalable Vector Graphics (SVG)), and a facetted classification mechanism. Document content extraction is implemented via computer programming (with Java). An Engineering Document Content Management System (EDCMS) developed in this research demonstrates that as information providers we can make document content in a more accessible manner for information users including engineers.The main features of the EDCMS system are: 1) EDCMS is a system that enables users, especially engineers, to access and retrieve information at content rather than document level. In other words, it provides the right pieces of information that answer specific questions so that engineers don't need to waste time sifting through the whole document to obtain the required piece of information. 2) Users can use the EDCMS via both the data and metadata of a document to access engineering document content. 3) Users can use the EDCMS to access and retrieve content objects, i.e. text, images and graphics (including engineering drawings) via multiple views and at different granularities based on decomposition schemes. Experiments with the EDCMS have been conducted on semi-structured documents, a textbook of CADCAM, and a set of project posters in the Engineering Design domain. Experimental results show that the system provides information users with a powerful solution to access document content.展开更多
Citations play an important role in the scientific community by assisting in measuring multifarious policies like the impact of journals,researchers,institutions,and countries.Authors cite papers for different reasons...Citations play an important role in the scientific community by assisting in measuring multifarious policies like the impact of journals,researchers,institutions,and countries.Authors cite papers for different reasons,such as extending previous work,comparing their study with the state-of-the-art,providing background of the field,etc.In recent years,researchers have tried to conceptualize all citations into two broad categories,important and incidental.Such a categorization is very important to enhance scientific output in multiple ways,for instance,(1)Helping a researcher in identifying meaningful citations from a list of 100 to 1000 citations(2)Enhancing the impact factor calculation mechanism by more strongly weighting important citations,and(3)Improving researcher,institutional,and university rankings by only considering important citations.All of these uses depend upon correctly identifying the important citations from the list of all citations in a paper.To date,researchers have utilized many features to classify citations into these broad categories:cue phrases,in-text citation counts,and metadata features,etc.However,contemporary approaches are based on identification of in-text citation counts,mapping sections onto the Introduction,Methods,Results,and Discussion(IMRAD)structure,identifying cue phrases,etc.Identifying such features accurately is a challenging task and is normally conducted manually,with the accuracy of citation classification demonstrated in terms of these manually extracted features.This research proposes to examine the content of the cited and citing pair to identify important citing papers for each cited paper.This content similarity approach was adopted from research paper recommendation approaches.Furthermore,a novel section-based content similarity approach is also proposed.The results show that solely using the abstract of the cited and citing papers can achieve similar accuracy as the stateof-the-art approaches.This makes the proposed approach a viable technique that does not depend on manual identification of complex features.展开更多
The growing collection of scientific data in various web repositories is referred to as Scientific Big Data,as it fulfills the four“V’s”of Big Data—volume,variety,velocity,and veracity.This phenomenon has created ...The growing collection of scientific data in various web repositories is referred to as Scientific Big Data,as it fulfills the four“V’s”of Big Data—volume,variety,velocity,and veracity.This phenomenon has created new opportunities for startups;for instance,the extraction of pertinent research papers from enormous knowledge repositories using certain innovative methods has become an important task for researchers and entrepreneurs.Traditionally,the content of the papers are compared to list the relevant papers from a repository.The conventional method results in a long list of papers that is often impossible to interpret productively.Therefore,the need for a novel approach that intelligently utilizes the available data is imminent.Moreover,the primary element of the scientific knowledge base is a research article,which consists of various logical sections such as the Abstract,Introduction,Related Work,Methodology,Results,and Conclusion.Thus,this study utilizes these logical sections of research articles,because they hold significant potential in finding relevant papers.In this study,comprehensive experiments were performed to determine the role of the logical sections-based terms indexing method in improving the quality of results(i.e.,retrieving relevant papers).Therefore,we proposed,implemented,and evaluated the logical sections-based content comparisons method to address the research objective with a standard method of indexing terms.The section-based approach outperformed the standard content-based approach in identifying relevant documents from all classified topics of computer science.Overall,the proposed approach extracted 14%more relevant results from the entire dataset.As the experimental results suggested that employing a finer content similarity technique improved the quality of results,the proposed approach has led the foundation of knowledge-based startups.展开更多
The present study is inscribed within the framework of the amelioration of the soils of the Santchou plain for employment as pavement subgrade. The bearing capacity proposed by these soils at their respective optimum ...The present study is inscribed within the framework of the amelioration of the soils of the Santchou plain for employment as pavement subgrade. The bearing capacity proposed by these soils at their respective optimum dry densities is relatively small, although most of these experimental California Bearing Ratio (CBR) values of the studied soils are more important than the ones prescribed by the American Association of State Highway and Transportation Officials Classification system (AASHTO) for A5, A6, and A7 types. The stabilization of this soils with lime has been chosen to improve the bearing capacity and by association, their resilient modulus. The results of this study show that the increase of lime content is not proportional with the increase of the expected mechanical performances. In fact, the literature explains that when the lime content arrives at an optimum, the mechanical parameters no longer increase, but decrease significantly. After this optimum, the soil stabilization no longer shows advantages in the increase of geo-mechanical properties of soils.展开更多
This paper presents a novel efficient semantic image classification algorithm for high-level feature indexing of high-dimension image database. Experiments show that the algorithm performs well. The size of the train ...This paper presents a novel efficient semantic image classification algorithm for high-level feature indexing of high-dimension image database. Experiments show that the algorithm performs well. The size of the train set and the test set is 7 537 and 5 000 respectively. Based on this theory, another ground is built with 12,000 images, which are divided into three classes: city, landscape and person, the total result of the classifications is 88.92%, meanwhile, some preliminary results are presented for image understanding based on semantic image classification and low level features. The groundtruth for the experiments is built with the images from Corel database, photos and some famous face databases.展开更多
文摘From the beginning,the process of research and its publication is an ever-growing phenomenon and with the emergence of web technologies,its growth rate is overwhelming.On a rough estimate,more than thirty thousand research journals have been issuing around four million papers annually on average.Search engines,indexing services,and digital libraries have been searching for such publications over the web.Nevertheless,getting the most relevant articles against the user requests is yet a fantasy.It is mainly because the articles are not appropriately indexed based on the hierarchies of granular subject classification.To overcome this issue,researchers are striving to investigate new techniques for the classification of the research articles especially,when the complete article text is not available(a case of nonopen access articles).The proposed study aims to investigate the multilabel classification over the available metadata in the best possible way and to assess,“to what extent metadata-based features can perform in contrast to content-based approaches.”In this regard,novel techniques for investigating multilabel classification have been proposed,developed,and evaluated on metadata such as the Title and Keywords of the articles.The proposed technique has been assessed for two diverse datasets,namely,from the Journal of universal computer science(J.UCS)and the benchmark dataset comprises of the articles published by the Association for computing machinery(ACM).The proposed technique yields encouraging results in contrast to the state-ofthe-art techniques in the literature.
基金The National Natural Science Foundation of China under contract No.U2006213。
文摘Fine-grained silt is widely distributed in the Huanghe River Delta(HRD)in China,and the sedimentary structure is complex,meaning that the clay content in the silt is variable.The piezocone penetration test(CPTu)is the most widely approved in situ test method.It can be used to invert soil properties and interpret soil behavior.To analyse the strength properties of surface sediments in the HRD,this paper evaluated the friction angle and its inversion formula through the CPTu penetration test and monotonic simple shear test and other soil unit experiments.The evaluation showed that the empirical formula proposed by Kulhawy and Mayne had better prediction and inversion effect.The HRD silts with clay contents of 9.2%,21.4%and 30.3%were selected as samples for the CPTu variable rate penetration test.The results show as follows.(1)The effects of the clay content on the tip resistance and the pore pressure of silt under different penetration rates were summarized.The tip resistance Q_t is strongly dependent on the clay content of the silt,the B_(q)value of the silt tends to 0 and is not significantly affected by the change of the CPTu penetration rate.(2)Five soil behavior type classification charts and three soil behavior type indexes based on CPTu data were evaluated.The results show that the soil behavior type classification chart based on soil behavior type index ISBT,the Robertson 2010 behavior type classification chart are more suitable for the silty soil in the HRD.
基金Project(2012AA011205)supported by National High-Tech Research and Development Program(863 Program)of ChinaProjects(61272150,61379109,M1321007,61301136,61103034)supported by the National Natural Science Foundation of China+1 种基金Project(20120162110077)supported by Research Fund for the Doctoral Program of Higher Education of ChinaProject(11JJ1012)supported by Excellent Youth Foundation of Hunan Scientific Committee,China
文摘Many text classifications depend on statistical term measures to implement document representation. Such document representations ignore the lexical semantic contents of terms and the distilled mutual information, leading to text classification errors.This work proposed a document representation method, Word Net-based lexical semantic VSM, to solve the problem. Using Word Net,this method constructed a data structure of semantic-element information to characterize lexical semantic contents, and adjusted EM modeling to disambiguate word stems. Then, in the lexical-semantic space of corpus, lexical-semantic eigenvector of document representation was built by calculating the weight of each synset, and applied to a widely-recognized algorithm NWKNN. On text corpus Reuter-21578 and its adjusted version of lexical replacement, the experimental results show that the lexical-semantic eigenvector performs F1 measure and scales of dimension better than term-statistic eigenvector based on TF-IDF. Formation of document representation eigenvectors ensures the method a wide prospect of classification applications in text corpus analysis.
文摘AIM:To support probe-based confocal laser endomi-croscopy (pCLE) diagnosis by designing software for the automated classification of colonic polyps. METHODS:Intravenous fluorescein pCLE imaging of colorectal lesions was performed on patients under-going screening and surveillance colonoscopies, followed by polypectomies. All resected specimens were reviewed by a reference gastrointestinal pathologist blinded to pCLE information. Histopathology was used as the criterion standard for the differentiation between neoplastic and non-neoplastic lesions. The pCLE video sequences, recorded for each polyp, were analyzed off-line by 2 expert endoscopists who were blinded to the endoscopic characteristics and histopathology. These pCLE videos, along with their histopathology diagnosis, were used to train the automated classification software which is a content-based image retrieval technique followed by k-nearest neighbor classification. The performance of the off-line diagnosis of pCLE videos established by the 2 expert endoscopists was compared with that of automated pCLE software classification. All evaluations were performed using leave-one-patient- out cross-validation to avoid bias. RESULTS:Colorectal lesions (135) were imaged in 71 patients. Based on histopathology, 93 of these 135 lesions were neoplastic and 42 were non-neoplastic. The study found no statistical significance for the difference between the performance of automated pCLE software classification (accuracy 89.6%, sensitivity 92.5%, specificity 83.3%, using leave-one-patient-out cross-validation) and the performance of the off-line diagnosis of pCLE videos established by the 2 expert endoscopists (accuracy 89.6%, sensitivity 91.4%, specificity 85.7%). There was very low power (< 6%) to detect the observed differences. The 95% confidence intervals for equivalence testing were:-0.073 to 0.073 for accuracy, -0.068 to 0.089 for sensitivity and -0.18 to 0.13 for specificity. The classification software proposed in this study is not a "black box" but an informative tool based on the query by example model that produces, as intermediate results, visually similar annotated videos that are directly interpretable by the endoscopist. CONCLUSION:The proposed software for automated classification of pCLE videos of colonic polyps achieves high performance, comparable to that of off-line diagnosis of pCLE videos established by expert endoscopists.
文摘The present study is inscribed within the framework of the geotechnical characterization of the soils of the Santchou plain, their classification for employment as pavement subgrade, various identification tests were carried out on the samples. The results obtained showed that with a wide range of different grain sizes, the studied soils showed low content in clay grains and dominance of either sand grains or silt grains, this can be explaining how most of these soil are poorly graded. According to the USDA textural classification, the grain size distribution of these soils makes them to be classified as Silty Loam types to Sandy Loam types. Despite of their organic matter content which is less than 10%, according to their respective methylene blue values, the soils studied along the section should be mainly loamy soil of medium plasticity to clayed soil, therefore showing a sensibility of its behavior to variation of water content. That last one is confirmed by the consistency parameters of these soils which show intermediate plasticity to highly plastic. Also, the bearing capacity proposed by these soils at their respective optimum dry densities is relatively small, although most of these experimental CBR values of the studied soils are more important than the ones prescribed by the AASHTO Classification system for A5, A6, and A7 types, and the French Highway Earthworks Manual Classifications system (GTR) for the corresponding A2 and A3 types.
文摘On the basis of the data obtained from the comprehensive Kuroshio surveys in 1987-1988,this paper analyses the oceanographic characteristics in the area (125°-130° E,27°-31° N) of the continental shelf edge of the East China Sea (E. C. S. ) and its adjacent waters and discusses the effects of the Kuroshio front,thermocline and upwelling of the Kuroshio subsurface water on the distribution of standing stock of phytoplankton (chlorophyll-a). The distribution of high content of chlorophylly-a has been detected at 20-50 in depth in the water body on the left side of the Kuroshio front in the continental shelf edge waters of the E. C. S. The high content of chlorophyll-a spreads from the shelf area to the Kuroshio area in the form of a tongue and connects with the maximum layer of subsurface chlorophyll-a of the Kuroshio and pelagic sea. The author considers that the formation of the distribution of high content chlorophyll-a in this area results from the bottom topography and oceanic environment and there are close correlations between the high content of chlorophyll-a and the light-nutrient environment.
基金Foundation item: supported by Intel Corporation (No. 9078)
文摘Packet classification (PC) has become the main method to support the quality of service and security of network application. And two-dimensional prefix packet classification (PPC) is the popular one. This paper analyzes the problem of ruler conflict, and then presents a TCAMbased two-dimensional PPC algorithm. This algorithm makes use of the parallelism of TCAM to lookup the longest prefix in one instruction cycle. Then it uses a memory image and associated data structures to eliminate the conflicts between rulers, and performs a fast two-dimensional PPC.Compared with other algorithms, this algorithm has the least time complexity and less space complexity.
基金the National Natural Science Foundation of China (No.60532030).
文摘The feature of Ternary Content Addressable Memories(TCAMs) makes them particularly attractive for IP address lookup and packet classification applications in a router system. However,the limitations of TCAMs impede their utilization. In this paper,the solutions for decreasing the power consumption and avoiding entry expansion in range matching are addressed. Experimental results demonstrate that the proposed techniques can make some big improvements on the performance of TCAMs in IP address lookup and packet classification.
基金financially supported by an open fund of the Key Laboratory of Forest Genetics and BiotechnologyMinistry of Education and Jiangsu Province+3 种基金Nanjing Forestry University(FGB200901)the‘‘948’’project of China(2014-4-59)the Project of National Natural Science Foundation of China(No.31370658)‘‘the Mutual Fund of Beijing Government and Central Universities in Beijing(GJ2011-2)’’
文摘An embryo classification system for Pinus tabuliformis Carr. was established by time-tracing sam- pling and observation of zygotic embryos. The zygotic embryos were divided into nine stages. Key elements of the zygotic embryo and female gametophyte (FG) tissue of P. tabuliformis were analyzed, using inductively coupled plasma-emission spectroscopy. Several elements--includ- ing aluminum, iron, sodium, and copper--are found in both embryo and FG tissue. Boron, phosphorus, magnesium,zinc, and calcium are also required for zygotic embryo development and therefore accumulated. Manganese is selectively excluded from the embryo. The zygotic embryo development needs a low-sodium and high-potassium nutrition proportion. The results of elemental analysis from zygotic embryos and FGs can provide the mineral targets for optimizing the formulation of culture medium for somatic embryogenesis.
基金This work was supported by the UK Engineering and Physical Sciences Research Council(EPSRC)(No.GR/R67507/01).
文摘Engineers often need to look for the right pieces of information by sifting through long engineering documents, It is a very tiring and time-consuming job. To address this issue, researchers are increasingly devoting their attention to new ways to help information users, including engineers, to access and retrieve document content. The research reported in this paper explores how to use the key technologies of document decomposition (study of document structure), document mark-up (with EXtensible Mark- up Language (XML), HyperText Mark-up Language (HTML), and Scalable Vector Graphics (SVG)), and a facetted classification mechanism. Document content extraction is implemented via computer programming (with Java). An Engineering Document Content Management System (EDCMS) developed in this research demonstrates that as information providers we can make document content in a more accessible manner for information users including engineers.The main features of the EDCMS system are: 1) EDCMS is a system that enables users, especially engineers, to access and retrieve information at content rather than document level. In other words, it provides the right pieces of information that answer specific questions so that engineers don't need to waste time sifting through the whole document to obtain the required piece of information. 2) Users can use the EDCMS via both the data and metadata of a document to access engineering document content. 3) Users can use the EDCMS to access and retrieve content objects, i.e. text, images and graphics (including engineering drawings) via multiple views and at different granularities based on decomposition schemes. Experiments with the EDCMS have been conducted on semi-structured documents, a textbook of CADCAM, and a set of project posters in the Engineering Design domain. Experimental results show that the system provides information users with a powerful solution to access document content.
文摘Citations play an important role in the scientific community by assisting in measuring multifarious policies like the impact of journals,researchers,institutions,and countries.Authors cite papers for different reasons,such as extending previous work,comparing their study with the state-of-the-art,providing background of the field,etc.In recent years,researchers have tried to conceptualize all citations into two broad categories,important and incidental.Such a categorization is very important to enhance scientific output in multiple ways,for instance,(1)Helping a researcher in identifying meaningful citations from a list of 100 to 1000 citations(2)Enhancing the impact factor calculation mechanism by more strongly weighting important citations,and(3)Improving researcher,institutional,and university rankings by only considering important citations.All of these uses depend upon correctly identifying the important citations from the list of all citations in a paper.To date,researchers have utilized many features to classify citations into these broad categories:cue phrases,in-text citation counts,and metadata features,etc.However,contemporary approaches are based on identification of in-text citation counts,mapping sections onto the Introduction,Methods,Results,and Discussion(IMRAD)structure,identifying cue phrases,etc.Identifying such features accurately is a challenging task and is normally conducted manually,with the accuracy of citation classification demonstrated in terms of these manually extracted features.This research proposes to examine the content of the cited and citing pair to identify important citing papers for each cited paper.This content similarity approach was adopted from research paper recommendation approaches.Furthermore,a novel section-based content similarity approach is also proposed.The results show that solely using the abstract of the cited and citing papers can achieve similar accuracy as the stateof-the-art approaches.This makes the proposed approach a viable technique that does not depend on manual identification of complex features.
基金supported by Institute of Information&communications Technology Planning&Evaluation(IITP)grant funded by the Korea government(MSIT)(2020-0-01592)Basic Science Research Program through the National Research Foundation of Korea(NRF)funded by the Ministry of Education(2019R1F1A1058548).
文摘The growing collection of scientific data in various web repositories is referred to as Scientific Big Data,as it fulfills the four“V’s”of Big Data—volume,variety,velocity,and veracity.This phenomenon has created new opportunities for startups;for instance,the extraction of pertinent research papers from enormous knowledge repositories using certain innovative methods has become an important task for researchers and entrepreneurs.Traditionally,the content of the papers are compared to list the relevant papers from a repository.The conventional method results in a long list of papers that is often impossible to interpret productively.Therefore,the need for a novel approach that intelligently utilizes the available data is imminent.Moreover,the primary element of the scientific knowledge base is a research article,which consists of various logical sections such as the Abstract,Introduction,Related Work,Methodology,Results,and Conclusion.Thus,this study utilizes these logical sections of research articles,because they hold significant potential in finding relevant papers.In this study,comprehensive experiments were performed to determine the role of the logical sections-based terms indexing method in improving the quality of results(i.e.,retrieving relevant papers).Therefore,we proposed,implemented,and evaluated the logical sections-based content comparisons method to address the research objective with a standard method of indexing terms.The section-based approach outperformed the standard content-based approach in identifying relevant documents from all classified topics of computer science.Overall,the proposed approach extracted 14%more relevant results from the entire dataset.As the experimental results suggested that employing a finer content similarity technique improved the quality of results,the proposed approach has led the foundation of knowledge-based startups.
文摘The present study is inscribed within the framework of the amelioration of the soils of the Santchou plain for employment as pavement subgrade. The bearing capacity proposed by these soils at their respective optimum dry densities is relatively small, although most of these experimental California Bearing Ratio (CBR) values of the studied soils are more important than the ones prescribed by the American Association of State Highway and Transportation Officials Classification system (AASHTO) for A5, A6, and A7 types. The stabilization of this soils with lime has been chosen to improve the bearing capacity and by association, their resilient modulus. The results of this study show that the increase of lime content is not proportional with the increase of the expected mechanical performances. In fact, the literature explains that when the lime content arrives at an optimum, the mechanical parameters no longer increase, but decrease significantly. After this optimum, the soil stabilization no longer shows advantages in the increase of geo-mechanical properties of soils.
文摘This paper presents a novel efficient semantic image classification algorithm for high-level feature indexing of high-dimension image database. Experiments show that the algorithm performs well. The size of the train set and the test set is 7 537 and 5 000 respectively. Based on this theory, another ground is built with 12,000 images, which are divided into three classes: city, landscape and person, the total result of the classifications is 88.92%, meanwhile, some preliminary results are presented for image understanding based on semantic image classification and low level features. The groundtruth for the experiments is built with the images from Corel database, photos and some famous face databases.