Pattern division multiple access(PDMA),which is a novel non-orthogonal multiple access(NOMA),has been proposed to address the challenges of massive connectivity and higher spectral efficiency for fifth generation(5G) ...Pattern division multiple access(PDMA),which is a novel non-orthogonal multiple access(NOMA),has been proposed to address the challenges of massive connectivity and higher spectral efficiency for fifth generation(5G) mobile network.The performance of PDMA mainly depends on the design of PDMA pattern matrix.In this paper,pattern matrix design of PDMA for 5G uplink(UL) applications in massive machine type communication(mMTC) and enhanced mobile broadband(eMBB) deployment scenarios are studied.The design criteria and examples for application in UL mMTC and UL eMBB are investigated.The performance of the PDMA pattern matrix is analyzed with the discrete constellation constrained capacity(CC-Capacity),and verified by Monte Carlo simulation.The simulation results show that the preferred PDMA pattern matrix can achieve good performance with different overloading factors(OF).展开更多
Let P be a property referring to a real matrix. For a sign pattern A, if there exists a real matrix B in the qualitative class of A such that B has property P, then we say A allows P. Three cases that A allows an M m...Let P be a property referring to a real matrix. For a sign pattern A, if there exists a real matrix B in the qualitative class of A such that B has property P, then we say A allows P. Three cases that A allows an M matrix, an inverse M matrix and a P 0 matrix are considered. The complete characterizations are obtained.展开更多
Matrix principal component analysis (MatPCA), as an effective feature extraction method, can deal with the matrix pattern and the vector pattern. However, like PCA, MatPCA does not use the class information of sampl...Matrix principal component analysis (MatPCA), as an effective feature extraction method, can deal with the matrix pattern and the vector pattern. However, like PCA, MatPCA does not use the class information of samples. As a result, the extracted features cannot provide enough useful information for distinguishing pat- tern from one another, and further resulting in degradation of classification performance. To fullly use class in- formation of samples, a novel method, called the fuzzy within-class MatPCA (F-WMatPCA)is proposed. F-WMatPCA utilizes the fuzzy K-nearest neighbor method(FKNN) to fuzzify the class membership degrees of a training sample and then performs fuzzy MatPCA within these patterns having the same class label. Due to more class information is used in feature extraction, F-WMatPCA can intuitively improve the classification perfor- mance. Experimental results in face databases and some benchmark datasets show that F-WMatPCA is effective and competitive than MatPCA. The experimental analysis on face image databases indicates that F-WMatPCA im- proves the recognition accuracy and is more stable and robust in performing classification than the existing method of fuzzy-based F-Fisherfaces.展开更多
Over the past decade, Graphics Processing Units (GPUs) have revolutionized high-performance computing, playing pivotal roles in advancing fields like IoT, autonomous vehicles, and exascale computing. Despite these adv...Over the past decade, Graphics Processing Units (GPUs) have revolutionized high-performance computing, playing pivotal roles in advancing fields like IoT, autonomous vehicles, and exascale computing. Despite these advancements, efficiently programming GPUs remains a daunting challenge, often relying on trial-and-error optimization methods. This paper introduces an optimization technique for CUDA programs through a novel Data Layout strategy, aimed at restructuring memory data arrangement to significantly enhance data access locality. Focusing on the dynamic programming algorithm for chained matrix multiplication—a critical operation across various domains including artificial intelligence (AI), high-performance computing (HPC), and the Internet of Things (IoT)—this technique facilitates more localized access. We specifically illustrate the importance of efficient matrix multiplication in these areas, underscoring the technique’s broader applicability and its potential to address some of the most pressing computational challenges in GPU-accelerated applications. Our findings reveal a remarkable reduction in memory consumption and a substantial 50% decrease in execution time for CUDA programs utilizing this technique, thereby setting a new benchmark for optimization in GPU computing.展开更多
基金supported by the National High Technology Research and Development Program of China(863 Program,No. 2015AA01A709)
文摘Pattern division multiple access(PDMA),which is a novel non-orthogonal multiple access(NOMA),has been proposed to address the challenges of massive connectivity and higher spectral efficiency for fifth generation(5G) mobile network.The performance of PDMA mainly depends on the design of PDMA pattern matrix.In this paper,pattern matrix design of PDMA for 5G uplink(UL) applications in massive machine type communication(mMTC) and enhanced mobile broadband(eMBB) deployment scenarios are studied.The design criteria and examples for application in UL mMTC and UL eMBB are investigated.The performance of the PDMA pattern matrix is analyzed with the discrete constellation constrained capacity(CC-Capacity),and verified by Monte Carlo simulation.The simulation results show that the preferred PDMA pattern matrix can achieve good performance with different overloading factors(OF).
文摘Let P be a property referring to a real matrix. For a sign pattern A, if there exists a real matrix B in the qualitative class of A such that B has property P, then we say A allows P. Three cases that A allows an M matrix, an inverse M matrix and a P 0 matrix are considered. The complete characterizations are obtained.
文摘Matrix principal component analysis (MatPCA), as an effective feature extraction method, can deal with the matrix pattern and the vector pattern. However, like PCA, MatPCA does not use the class information of samples. As a result, the extracted features cannot provide enough useful information for distinguishing pat- tern from one another, and further resulting in degradation of classification performance. To fullly use class in- formation of samples, a novel method, called the fuzzy within-class MatPCA (F-WMatPCA)is proposed. F-WMatPCA utilizes the fuzzy K-nearest neighbor method(FKNN) to fuzzify the class membership degrees of a training sample and then performs fuzzy MatPCA within these patterns having the same class label. Due to more class information is used in feature extraction, F-WMatPCA can intuitively improve the classification perfor- mance. Experimental results in face databases and some benchmark datasets show that F-WMatPCA is effective and competitive than MatPCA. The experimental analysis on face image databases indicates that F-WMatPCA im- proves the recognition accuracy and is more stable and robust in performing classification than the existing method of fuzzy-based F-Fisherfaces.
文摘Over the past decade, Graphics Processing Units (GPUs) have revolutionized high-performance computing, playing pivotal roles in advancing fields like IoT, autonomous vehicles, and exascale computing. Despite these advancements, efficiently programming GPUs remains a daunting challenge, often relying on trial-and-error optimization methods. This paper introduces an optimization technique for CUDA programs through a novel Data Layout strategy, aimed at restructuring memory data arrangement to significantly enhance data access locality. Focusing on the dynamic programming algorithm for chained matrix multiplication—a critical operation across various domains including artificial intelligence (AI), high-performance computing (HPC), and the Internet of Things (IoT)—this technique facilitates more localized access. We specifically illustrate the importance of efficient matrix multiplication in these areas, underscoring the technique’s broader applicability and its potential to address some of the most pressing computational challenges in GPU-accelerated applications. Our findings reveal a remarkable reduction in memory consumption and a substantial 50% decrease in execution time for CUDA programs utilizing this technique, thereby setting a new benchmark for optimization in GPU computing.