This paper investigates the bit-interleaved coded generalized spatial modulation(BICGSM) with iterative decoding(BICGSM-ID) for multiple-input multiple-output(MIMO) visible light communications(VLC). In the BICGSM-ID ...This paper investigates the bit-interleaved coded generalized spatial modulation(BICGSM) with iterative decoding(BICGSM-ID) for multiple-input multiple-output(MIMO) visible light communications(VLC). In the BICGSM-ID scheme, the information bits conveyed by the signal-domain(SiD) symbols and the spatial-domain(SpD) light emitting diode(LED)-index patterns are coded by a protograph low-density parity-check(P-LDPC) code. Specifically, we propose a signal-domain symbol expanding and re-allocating(SSER) method for constructing a type of novel generalized spatial modulation(GSM) constellations, referred to as SSERGSM constellations, so as to boost the performance of the BICGSM-ID MIMO-VLC systems.Moreover, by applying a modified PEXIT(MPEXIT) algorithm, we further design a family of rate-compatible P-LDPC codes, referred to as enhanced accumulate-repeat-accumulate(EARA) codes,which possess both excellent decoding thresholds and linear-minimum-distance-growth property. Both analysis and simulation results illustrate that the proposed SSERGSM constellations and P-LDPC codes can remarkably improve the convergence and decoding performance of MIMO-VLC systems. Therefore, the proposed P-LDPC-coded SSERGSM-mapped BICGSMID configuration is envisioned as a promising transmission solution to satisfy the high-throughput requirement of MIMO-VLC applications.展开更多
Tuberculosis treatment course completion is crucial to protect patients against prolonged infectiousness, relapse, lengthened and more expensive therapy due to multidrug resistance TB. Up to 50% of all patients do not...Tuberculosis treatment course completion is crucial to protect patients against prolonged infectiousness, relapse, lengthened and more expensive therapy due to multidrug resistance TB. Up to 50% of all patients do not complete treatment course. To solve this problem, TB treatment with patient supervision and support as an element of the “global plan to stop TB” was considered by the World Health Organization. The plan may require a model to predict the outcome of DOTS therapy;then, this tool may be used to determine how intensive the level of providing services and supports should be. This work applied and compared machine learning techniques initially to predict the outcome of TB therapy. After feature analysis, models by six algorithms including decision tree (DT), artificial neural network (ANN), logistic regression (LR), radial basis function (RBF), Bayesian networks (BN), and support vector machine (SVM) developed and validated. Data of training (N = 4515) and testing (N = 1935) sets were applied and models evaluated by prediction accuracy, F-measure and recall. Seventeen significantly correlated features were identified (P CI = 0.001 - 0.007);DT (C 4.5) was found to be the best algorithm with %74.21 prediction accuracy in comparing with ANN, BN, LR, RBF, and SVM with 62.06%, 57.88%, 57.31%, 53.74%, and 51.36% respectively. Data and distribution may create the opportunity for DT out performance. The predicted class for each TB case might be useful for improving the quality of care through making patients’ supervision and support more case—sensitive in order to enhance the quality of DOTS therapy.展开更多
We have presented an integrated approach based on supervised and unsupervised learning tech- nique to improve the accuracy of six predictive models. They are developed to predict outcome of tuberculosis treatment cour...We have presented an integrated approach based on supervised and unsupervised learning tech- nique to improve the accuracy of six predictive models. They are developed to predict outcome of tuberculosis treatment course and their accuracy needs to be improved as they are not precise as much as necessary. The integrated supervised and unsupervised learning method (ISULM) has been proposed as a new way to improve model accuracy. The dataset of 6450 Iranian TB patients under DOTS therapy was applied to initially select the significant predictors and then develop six predictive models using decision tree, Bayesian network, logistic regression, multilayer perceptron, radial basis function, and support vector machine algorithms. Developed models have integrated with k-mean clustering analysis to calculate more accurate predicted outcome of tuberculosis treatment course. Obtained results, then, have been evaluated to compare prediction accuracy before and after ISULM application. Recall, Precision, F-measure, and ROC area are other criteria used to assess the models validity as well as change percentage to show how different are models before and after ISULM. ISULM led to improve the prediction accuracy for all applied classifiers ranging between 4% and 10%. The most and least improvement for prediction accuracy were shown by logistic regression and support vector machine respectively. Pre-learning by k- mean clustering to relocate the objects and put similar cases in the same group can improve the classification accuracy in the process of integrating supervised and unsupervised learning.展开更多
This paper proposes a high-throughput short reference differential chaos shift keying cooperative communication system with the aid of code index modulation,referred to as CIM-SR-DCSK-CC system.In the proposed CIM-SR-...This paper proposes a high-throughput short reference differential chaos shift keying cooperative communication system with the aid of code index modulation,referred to as CIM-SR-DCSK-CC system.In the proposed CIM-SR-DCSK-CC system,the source transmits information bits to both the relay and destination in the first time slot,while the relay not only forwards the source information bits but also sends new information bits to the destination in the second time slot.To be specific,the relay employs an N-order Walsh code to carry additional log_(2)N information bits,which are superimposed onto the SRDCSK signal carrying the decoded source information bits.Subsequently,the superimposed signal carrying both the source and relay information bits is transmitted to the destination.Moreover,the theoretical bit error rate(BER)expressions of the proposed CIMSR-DCSK-CC system are derived over additive white Gaussian noise(AWGN)and multipath Rayleigh fading channels.Compared with the conventional DCSKCC system and SR-DCSK-CC system,the proposed CIM-SR-DCSK-CC system can significantly improve the throughput without deteriorating any BER performance.As a consequence,the proposed system is very promising for the applications of the 6G-enabled lowpower and high-rate communication.展开更多
In this new information era,the transfer of data and information has become a very important matter.Transferred data must be kept secured from unauthorized persons using cryptography.The science of cryptography depend...In this new information era,the transfer of data and information has become a very important matter.Transferred data must be kept secured from unauthorized persons using cryptography.The science of cryptography depends not only on complex mathematical models but also on encryption keys.Amino acid encryption is a promising model for data security.In this paper,we propose an amino acid encryption model with two encryption keys.The first key is generated randomly using the genetic algorithm.The second key is called the protein key which is generated from converting DNA to a protein message.Then,the protein message and the first key are used in the modified Playfair matrix to generate the cypher message.The experimental results show that the proposed model survives against known attacks such as the Brute-force attack and the Ciphertext-only attack.In addition,the proposed model has been tested over different types of characters including white spaces and special characters,as all the data is encoded to 8-bit binary.The performance of the proposed model is compared with other models using encryption time and decryption time.The model also balances all three principles in the CIA triad.展开更多
Semantic Web(SW)provides new opportunities for the study and application of big data,massive ranges of data sets in varied formats from multiple sources.Related studies focus on potential SW technologies for resolving...Semantic Web(SW)provides new opportunities for the study and application of big data,massive ranges of data sets in varied formats from multiple sources.Related studies focus on potential SW technologies for resolving big data problems,such as structurally and semantically heterogeneous data that result from the variety of data formats(structured,semi-structured,numeric,unstructured text data,email,video,audio,stock ticker).SW offers information semantically both for people and machines to retain the vast volume of data and provide a meaningful output of unstructured data.In the current research,we implement a new semantic Extract Transform Load(ETL)model that uses SW technologies for aggregating,integrating,and representing data as linked data.First,geospatial data resources are aggregated from the internet,and then a semantic ETL model is used to store the aggregated data in a semantic model after converting it to Resource Description Framework(RDF)format for successful integration and representation.The principal contribution of this research is the synthesis,aggregation,and semantic representation of geospatial data to solve problems.A case study of city data is used to illustrate the semantic ETL model’s functionalities.The results show that the proposed model solves the structural and semantic heterogeneity problems in diverse data sources for successful data aggregation,integration,and representation.展开更多
Software systems have been employed in many fields as a means to reduce human efforts;consequently,stakeholders are interested in more updates of their capabilities.Code smells arise as one of the obstacles in the sof...Software systems have been employed in many fields as a means to reduce human efforts;consequently,stakeholders are interested in more updates of their capabilities.Code smells arise as one of the obstacles in the software industry.They are characteristics of software source code that indicate a deeper problem in design.These smells appear not only in the design but also in software implementation.Code smells introduce bugs,affect software maintainability,and lead to higher maintenance costs.Uncovering code smells can be formulated as an optimization problem of finding the best detection rules.Although researchers have recommended different techniques to improve the accuracy of code smell detection,these methods are still unstable and need to be improved.Previous research has sought only to discover a few at a time(three or five types)and did not set rules for detecting their types.Our research improves code smell detection by applying a search-based technique;we use the Whale Optimization Algorithm as a classifier to find ideal detection rules.Applying this algorithm,the Fisher criterion is utilized as a fitness function to maximize the between-class distance over the withinclass variance.The proposed framework adopts if-then detection rules during the software development life cycle.Those rules identify the types for both medium and large projects.Experiments are conducted on five open-source software projects to discover nine smell types that mostly appear in codes.The proposed detection framework has an average of 94.24%precision and 93.4%recall.These accurate values are better than other search-based algorithms of the same field.The proposed framework improves code smell detection,which increases software quality while minimizing maintenance effort,time,and cost.Additionally,the resulting classification rules are analyzed to find the software metrics that differentiate the nine code smells.展开更多
We present a novel quantum algorithm to evaluate the hamming distance between two unknown oracles via measuring the degree of entanglement between two ancillary qubits.In particular,we use the power of the entanglemen...We present a novel quantum algorithm to evaluate the hamming distance between two unknown oracles via measuring the degree of entanglement between two ancillary qubits.In particular,we use the power of the entanglement degree based quantum computing model that preserves at most the locality of interactions within the quantum model structure.This model uses one of two techniques to retrieve the solution of a quantum computing problem at hand.In the first technique,the solution of the problem is obtained based on whether there is an entanglement between the two ancillary qubits or not.In the second,the solution of the quantum computing problem is obtained as a function in the concurrence value,and the number of states that can be generated from the Boolean variables.The proposed algorithm receives two oracles,each oracle represents an unknown Boolean function,then it measures the hamming distance between these two oracles.The hamming distance is evaluated based on the second technique.It is shown that the proposed algorithm provides exponential speedup compared with the classical counterpart for Boolean functions that have large numbers of Boolean variables.The proposed algorithm is explained via a case study.Finally,employing recently developed experimental techniques,the proposed algorithm has been verified using IBM’s quantum computer simulator.展开更多
基金supported in part by the NSF of China under Grant 62322106,62071131the Guangdong Basic and Applied Basic Research Foundation under Grant 2022B1515020086+2 种基金the International Collaborative Research Program of Guangdong Science and Technology Department under Grant 2022A0505050070in part by the Open Research Fund of the State Key Laboratory of Integrated Services Networks under Grant ISN22-23the National Research Foundation,Singapore University of Technology Design under its Future Communications Research&Development Programme“Advanced Error Control Coding for 6G URLLC and mMTC”Grant No.FCP-NTU-RG-2022-020.
文摘This paper investigates the bit-interleaved coded generalized spatial modulation(BICGSM) with iterative decoding(BICGSM-ID) for multiple-input multiple-output(MIMO) visible light communications(VLC). In the BICGSM-ID scheme, the information bits conveyed by the signal-domain(SiD) symbols and the spatial-domain(SpD) light emitting diode(LED)-index patterns are coded by a protograph low-density parity-check(P-LDPC) code. Specifically, we propose a signal-domain symbol expanding and re-allocating(SSER) method for constructing a type of novel generalized spatial modulation(GSM) constellations, referred to as SSERGSM constellations, so as to boost the performance of the BICGSM-ID MIMO-VLC systems.Moreover, by applying a modified PEXIT(MPEXIT) algorithm, we further design a family of rate-compatible P-LDPC codes, referred to as enhanced accumulate-repeat-accumulate(EARA) codes,which possess both excellent decoding thresholds and linear-minimum-distance-growth property. Both analysis and simulation results illustrate that the proposed SSERGSM constellations and P-LDPC codes can remarkably improve the convergence and decoding performance of MIMO-VLC systems. Therefore, the proposed P-LDPC-coded SSERGSM-mapped BICGSMID configuration is envisioned as a promising transmission solution to satisfy the high-throughput requirement of MIMO-VLC applications.
文摘Tuberculosis treatment course completion is crucial to protect patients against prolonged infectiousness, relapse, lengthened and more expensive therapy due to multidrug resistance TB. Up to 50% of all patients do not complete treatment course. To solve this problem, TB treatment with patient supervision and support as an element of the “global plan to stop TB” was considered by the World Health Organization. The plan may require a model to predict the outcome of DOTS therapy;then, this tool may be used to determine how intensive the level of providing services and supports should be. This work applied and compared machine learning techniques initially to predict the outcome of TB therapy. After feature analysis, models by six algorithms including decision tree (DT), artificial neural network (ANN), logistic regression (LR), radial basis function (RBF), Bayesian networks (BN), and support vector machine (SVM) developed and validated. Data of training (N = 4515) and testing (N = 1935) sets were applied and models evaluated by prediction accuracy, F-measure and recall. Seventeen significantly correlated features were identified (P CI = 0.001 - 0.007);DT (C 4.5) was found to be the best algorithm with %74.21 prediction accuracy in comparing with ANN, BN, LR, RBF, and SVM with 62.06%, 57.88%, 57.31%, 53.74%, and 51.36% respectively. Data and distribution may create the opportunity for DT out performance. The predicted class for each TB case might be useful for improving the quality of care through making patients’ supervision and support more case—sensitive in order to enhance the quality of DOTS therapy.
文摘We have presented an integrated approach based on supervised and unsupervised learning tech- nique to improve the accuracy of six predictive models. They are developed to predict outcome of tuberculosis treatment course and their accuracy needs to be improved as they are not precise as much as necessary. The integrated supervised and unsupervised learning method (ISULM) has been proposed as a new way to improve model accuracy. The dataset of 6450 Iranian TB patients under DOTS therapy was applied to initially select the significant predictors and then develop six predictive models using decision tree, Bayesian network, logistic regression, multilayer perceptron, radial basis function, and support vector machine algorithms. Developed models have integrated with k-mean clustering analysis to calculate more accurate predicted outcome of tuberculosis treatment course. Obtained results, then, have been evaluated to compare prediction accuracy before and after ISULM application. Recall, Precision, F-measure, and ROC area are other criteria used to assess the models validity as well as change percentage to show how different are models before and after ISULM. ISULM led to improve the prediction accuracy for all applied classifiers ranging between 4% and 10%. The most and least improvement for prediction accuracy were shown by logistic regression and support vector machine respectively. Pre-learning by k- mean clustering to relocate the objects and put similar cases in the same group can improve the classification accuracy in the process of integrating supervised and unsupervised learning.
基金supported in part by the NSF of China under Grant 62322106,62071131 and 62171135the Guangdong Basic and Applied Basic Research Foundation under Grant 2022B1515020086+2 种基金the NSF of Guangdong Province under Grant 2019A1515011465the International Collaborative Research Program of Guangdong Science and Technology Department under Grant 2022A0505050070the Industrial R&D Project of Haoyang Electronic Co.,Ltd.under Grant 2022440002001494.
文摘This paper proposes a high-throughput short reference differential chaos shift keying cooperative communication system with the aid of code index modulation,referred to as CIM-SR-DCSK-CC system.In the proposed CIM-SR-DCSK-CC system,the source transmits information bits to both the relay and destination in the first time slot,while the relay not only forwards the source information bits but also sends new information bits to the destination in the second time slot.To be specific,the relay employs an N-order Walsh code to carry additional log_(2)N information bits,which are superimposed onto the SRDCSK signal carrying the decoded source information bits.Subsequently,the superimposed signal carrying both the source and relay information bits is transmitted to the destination.Moreover,the theoretical bit error rate(BER)expressions of the proposed CIMSR-DCSK-CC system are derived over additive white Gaussian noise(AWGN)and multipath Rayleigh fading channels.Compared with the conventional DCSKCC system and SR-DCSK-CC system,the proposed CIM-SR-DCSK-CC system can significantly improve the throughput without deteriorating any BER performance.As a consequence,the proposed system is very promising for the applications of the 6G-enabled lowpower and high-rate communication.
文摘In this new information era,the transfer of data and information has become a very important matter.Transferred data must be kept secured from unauthorized persons using cryptography.The science of cryptography depends not only on complex mathematical models but also on encryption keys.Amino acid encryption is a promising model for data security.In this paper,we propose an amino acid encryption model with two encryption keys.The first key is generated randomly using the genetic algorithm.The second key is called the protein key which is generated from converting DNA to a protein message.Then,the protein message and the first key are used in the modified Playfair matrix to generate the cypher message.The experimental results show that the proposed model survives against known attacks such as the Brute-force attack and the Ciphertext-only attack.In addition,the proposed model has been tested over different types of characters including white spaces and special characters,as all the data is encoded to 8-bit binary.The performance of the proposed model is compared with other models using encryption time and decryption time.The model also balances all three principles in the CIA triad.
文摘Semantic Web(SW)provides new opportunities for the study and application of big data,massive ranges of data sets in varied formats from multiple sources.Related studies focus on potential SW technologies for resolving big data problems,such as structurally and semantically heterogeneous data that result from the variety of data formats(structured,semi-structured,numeric,unstructured text data,email,video,audio,stock ticker).SW offers information semantically both for people and machines to retain the vast volume of data and provide a meaningful output of unstructured data.In the current research,we implement a new semantic Extract Transform Load(ETL)model that uses SW technologies for aggregating,integrating,and representing data as linked data.First,geospatial data resources are aggregated from the internet,and then a semantic ETL model is used to store the aggregated data in a semantic model after converting it to Resource Description Framework(RDF)format for successful integration and representation.The principal contribution of this research is the synthesis,aggregation,and semantic representation of geospatial data to solve problems.A case study of city data is used to illustrate the semantic ETL model’s functionalities.The results show that the proposed model solves the structural and semantic heterogeneity problems in diverse data sources for successful data aggregation,integration,and representation.
文摘Software systems have been employed in many fields as a means to reduce human efforts;consequently,stakeholders are interested in more updates of their capabilities.Code smells arise as one of the obstacles in the software industry.They are characteristics of software source code that indicate a deeper problem in design.These smells appear not only in the design but also in software implementation.Code smells introduce bugs,affect software maintainability,and lead to higher maintenance costs.Uncovering code smells can be formulated as an optimization problem of finding the best detection rules.Although researchers have recommended different techniques to improve the accuracy of code smell detection,these methods are still unstable and need to be improved.Previous research has sought only to discover a few at a time(three or five types)and did not set rules for detecting their types.Our research improves code smell detection by applying a search-based technique;we use the Whale Optimization Algorithm as a classifier to find ideal detection rules.Applying this algorithm,the Fisher criterion is utilized as a fitness function to maximize the between-class distance over the withinclass variance.The proposed framework adopts if-then detection rules during the software development life cycle.Those rules identify the types for both medium and large projects.Experiments are conducted on five open-source software projects to discover nine smell types that mostly appear in codes.The proposed detection framework has an average of 94.24%precision and 93.4%recall.These accurate values are better than other search-based algorithms of the same field.The proposed framework improves code smell detection,which increases software quality while minimizing maintenance effort,time,and cost.Additionally,the resulting classification rules are analyzed to find the software metrics that differentiate the nine code smells.
文摘We present a novel quantum algorithm to evaluate the hamming distance between two unknown oracles via measuring the degree of entanglement between two ancillary qubits.In particular,we use the power of the entanglement degree based quantum computing model that preserves at most the locality of interactions within the quantum model structure.This model uses one of two techniques to retrieve the solution of a quantum computing problem at hand.In the first technique,the solution of the problem is obtained based on whether there is an entanglement between the two ancillary qubits or not.In the second,the solution of the quantum computing problem is obtained as a function in the concurrence value,and the number of states that can be generated from the Boolean variables.The proposed algorithm receives two oracles,each oracle represents an unknown Boolean function,then it measures the hamming distance between these two oracles.The hamming distance is evaluated based on the second technique.It is shown that the proposed algorithm provides exponential speedup compared with the classical counterpart for Boolean functions that have large numbers of Boolean variables.The proposed algorithm is explained via a case study.Finally,employing recently developed experimental techniques,the proposed algorithm has been verified using IBM’s quantum computer simulator.