The Internet ofMedical Things(IoMT)is mainly concernedwith the efficient utilisation of wearable devices in the healthcare domain to manage various processes automatically,whereas machine learning approaches enable th...The Internet ofMedical Things(IoMT)is mainly concernedwith the efficient utilisation of wearable devices in the healthcare domain to manage various processes automatically,whereas machine learning approaches enable these smart systems to make informed decisions.Generally,broadcasting is used for the transmission of frames,whereas congestion,energy efficiency,and excessive load are among the common issues associated with existing approaches.In this paper,a machine learning-enabled shortest path identification scheme is presented to ensure reliable transmission of frames,especially with the minimum possible communication overheads in the IoMT network.For this purpose,the proposed scheme utilises a well-known technique,i.e.,Kruskal’s algorithm,to find an optimal path from source to destination wearable devices.Additionally,other evaluation metrics are used to find a reliable and shortest possible communication path between the two interested parties.Apart from that,every device is bound to hold a supplementary path,preferably a second optimised path,for situations where the current communication path is no longer available,either due to device failure or heavy traffic.Furthermore,the machine learning approach helps enable these devices to update their routing tables simultaneously,and an optimal path could be replaced if a better one is available.The proposed mechanism has been tested using a smart environment developed for the healthcare domain using IoMT networks.Simulation results show that the proposed machine learning-oriented approach performs better than existing approaches where the proposed scheme has achieved the minimum possible ratios,i.e.,17%and 23%,in terms of end to end delay and packet losses,respectively.Moreover,the proposed scheme has achieved an approximately 21%improvement in the average throughput compared to the existing schemes.展开更多
Broadcasting gateway equipment generally uses a method of simply switching to a spare input stream when a failure occurs in a main input stream.However,when the transmission environment is unstable,problems such as re...Broadcasting gateway equipment generally uses a method of simply switching to a spare input stream when a failure occurs in a main input stream.However,when the transmission environment is unstable,problems such as reduction in the lifespan of equipment due to frequent switching and interruption,delay,and stoppage of services may occur.Therefore,applying a machine learning(ML)method,which is possible to automatically judge and classify network-related service anomaly,and switch multi-input signals without dropping or changing signals by predicting or quickly determining the time of error occurrence for smooth stream switching when there are problems such as transmission errors,is required.In this paper,we propose an intelligent packet switching method based on the ML method of classification,which is one of the supervised learning methods,that presents the risk level of abnormal multi-stream occurring in broadcasting gateway equipment based on data.Furthermore,we subdivide the risk levels obtained from classification techniques into probabilities and then derive vectorized representative values for each attribute value of the collected input data and continuously update them.The obtained reference vector value is used for switching judgment through the cosine similarity value between input data obtained when a dangerous situation occurs.In the broadcasting gateway equipment to which the proposed method is applied,it is possible to perform more stable and smarter switching than before by solving problems of reliability and broadcasting accidents of the equipment and can maintain stable video streaming as well.展开更多
Extremely large-scale multiple-input multiple-output(XL-MIMO)and terahertz(THz)communications are pivotal candidate technologies for supporting the development of 6G mobile networks.However,these techniques invalidate...Extremely large-scale multiple-input multiple-output(XL-MIMO)and terahertz(THz)communications are pivotal candidate technologies for supporting the development of 6G mobile networks.However,these techniques invalidate the common assumptions of far-field plane waves and introduce many new properties.To accurately understand the performance of these new techniques,spherical wave modeling of near-field communications needs to be applied for future research.Hence,the investigation of near-field communication holds significant importance for the advancement of 6G,which brings many new and open research challenges in contrast to conventional far-field communication.In this paper,we first formulate a general model of the near-field channel and discuss the influence of spatial nonstationary properties on the near-field channel modeling.Subsequently,we discuss the challenges encountered in the near field in terms of beam training,localization,and transmission scheme design,respectively.Finally,we point out some promising research directions for near-field communications.展开更多
As a result of rapid development in electronics and communication technology,large-scale unmanned aerial vehicles(UAVs)are harnessed for various promising applications in a coordinated manner.Although it poses numerou...As a result of rapid development in electronics and communication technology,large-scale unmanned aerial vehicles(UAVs)are harnessed for various promising applications in a coordinated manner.Although it poses numerous advantages,resource management among various domains in large-scale UAV communication networks is the key challenge to be solved urgently.Specifically,due to the inherent requirements and future development trend,distributed resource management is suitable.In this article,we investigate the resource management problem for large-scale UAV communication networks from game-theoretic perspective which are exactly coincident with the distributed and autonomous manner.By exploring the inherent features,the distinctive challenges are discussed.Then,we explore several gametheoretic models that not only combat the challenges but also have broad application prospects.We provide the basics of each game-theoretic model and discuss the potential applications for resource management in large-scale UAV communication networks.Specifically,mean-field game,graphical game,Stackelberg game,coalition game and potential game are included.After that,we propose two innovative case studies to highlight the feasibility of such novel game-theoretic models.Finally,we give some future research directions to shed light on future opportunities and applications.展开更多
Unsourced multiple access(UMA)is a multi-access technology for massive,low-power,uncoordinated,and unsourced Machine Type Communication(MTC)networks.It ensures transmission reliability under the premise of high energy...Unsourced multiple access(UMA)is a multi-access technology for massive,low-power,uncoordinated,and unsourced Machine Type Communication(MTC)networks.It ensures transmission reliability under the premise of high energy efficiency.Based on the analysis of the 6G MTC key performance indicators(KPIs)and scenario characteristics,this paper summarizes its requirements for radio access networks.Following this,the existing multiple access models are analyzed under these standards to determine UMA's advantages for 6G MTC according to its design characteristics.The critical technology of UMA is the design of its multiple-access coding scheme.Therefore,the existing UMA coding schemes from different coding paradigms are further summarized and compared.In particular,this paper comprehensively considers the energy efficiency and computational complexity of these schemes,studies the changes of the above two indexes with the increase of access scale,and considers the trade-off between the two.It is revealed by the above analysis that some guiding rules of UMA coding design.Finally,the open problems and potentials in this field are given for future research.展开更多
Massive machine type communications(mMTC)have been regarded as promising applications in the future.One main feature of mMTC is short packet communication.Different from traditional long packet communication,short pac...Massive machine type communications(mMTC)have been regarded as promising applications in the future.One main feature of mMTC is short packet communication.Different from traditional long packet communication,short packet communication suffers from transmission rate degradation and a significant error rate is introduced.In this case,traditional resource allocation scheme for mMTC is no longer applicable.In this paper,we explore resource allocation for cellular-based mMTC in the finite blocklength regime.First,to mitigate the load of the base station(BS),we establish a framework for cellularbased mMTC,where MTCGs reuse the resources of cellular users(CUs),aggregate the packets generated by MTCDs,and forward them to the BS.Next,we adopt short packet theory to obtain the minimum required blocklength of a packet that transmits a certain amount of information.Then,by modeling the process of MTCGs-assisted communication as a queuing process,we derive the closed-form expression of the average delay of all MTCDs.Guided by this,we propose a joint power allocation and spectrum sharing scheme to minimize the average delay.Finally,the simulation results verify the correctness of the theoretical results and show that the proposed scheme can reduce the average delay efficiently.展开更多
Vehicular ad-hoc networks(VANETs)are mobile networks that use and transfer data with vehicles as the network nodes.Thus,VANETs are essentially mobile ad-hoc networks(MANETs).They allow all the nodes to communicate and...Vehicular ad-hoc networks(VANETs)are mobile networks that use and transfer data with vehicles as the network nodes.Thus,VANETs are essentially mobile ad-hoc networks(MANETs).They allow all the nodes to communicate and connect with one another.One of the main requirements in a VANET is to provide self-decision capability to the vehicles.Cognitive memory,which stores all the previous routes,is used by the vehicles to choose the optimal route.In networks,communication is crucial.In cellular-based vehicle-to-everything(CV2X)communication,vital information is shared using the cooperative awareness message(CAM)that is broadcast by each vehicle.Resources are allocated in a distributed manner,which is known as Mode 4 communication.The support vector machine(SVM)algorithm is used in the SVM-CV2X-M4 system proposed in this study.The k-fold model with different values of k is used to evaluate the accuracy of the SVM-CV2XM4 system.The results show that the proposed system achieves an accuracy of 99.6%.Thus,the proposed system allows vehicles to choose the optimal route and is highly convenient for users.展开更多
The design of large-scale machine system is a very complex problem.These design problems usually have a lot of design variables and constraints so that they are difficult to be solved rapidly and efficiently by using ...The design of large-scale machine system is a very complex problem.These design problems usually have a lot of design variables and constraints so that they are difficult to be solved rapidly and efficiently by using conventional methods.In this paper,a new multilevel design method oriented network environment is proposed,which maps the design problem of large-scale machine system into a hypergraph with degree of linking strength (DLS) between vertices.By decomposition of hypergraph,this method can divide the complex design problem into some small and simple subproblems that can be solved concurrently in a network.展开更多
In recent years,the need for a fast,efficient and a reliable wireless network has increased dramatically.Numerous 5G networks have already been tested while a few are in the early stages of deployment.In noncooperativ...In recent years,the need for a fast,efficient and a reliable wireless network has increased dramatically.Numerous 5G networks have already been tested while a few are in the early stages of deployment.In noncooperative communication scenarios,the recognition of digital signal modulations assists people in identifying the communication targets and ensures an effective management over them.The recent advancements in both Machine Learning(ML)and Deep Learning(DL)models demand the development of effective modulation recognition models with self-learning capability.In this background,the current research article designs aDeep Learning enabled Intelligent Modulation Recognition of Communication Signal(DLIMR-CS)technique for next-generation networks.The aim of the proposed DLIMR-CS technique is to classify different kinds of digitally-modulated signals.In addition,the fractal feature extraction process is appliedwith the help of the Sevcik Fractal Dimension(SFD)approach.Then,the extracted features are fed into the Deep Variational Autoencoder(DVAE)model for the classification of the modulated signals.In order to improve the classification performance of the DVAE model,the Tunicate Swarm Algorithm(TSA)is used to finetune the hyperparameters involved in DVAE model.A wide range of simulations was conducted to establish the enhanced performance of the proposed DLIMR-CS model.The experimental outcomes confirmed the superior recognition rate of the DLIMR-CS model over recent state-of-the-art methods under different evaluation parameters.展开更多
Key challenges for 5G and Beyond networks relate with the requirements for exceptionally low latency, high reliability, and extremely high data rates. The Ultra-Reliable Low Latency Communication (URLLC) use case is t...Key challenges for 5G and Beyond networks relate with the requirements for exceptionally low latency, high reliability, and extremely high data rates. The Ultra-Reliable Low Latency Communication (URLLC) use case is the trickiest to support and current research is focused on physical or MAC layer solutions, while proposals focused on the network layer using Machine Learning (ML) and Artificial Intelligence (AI) algorithms running on base stations and User Equipment (UE) or Internet of Things (IoT) devices are in early stages. In this paper, we describe the operation rationale of the most recent relevant ML algorithms and techniques, and we propose and validate ML algorithms running on both cells (base stations/gNBs) and UEs or IoT devices to handle URLLC service control. One ML algorithm runs on base stations to evaluate latency demands and offload traffic in case of need, while another lightweight algorithm runs on UEs and IoT devices to rank cells with the best URLLC service in real-time to indicate the best one cell for a UE or IoT device to camp. We show that the interplay of these algorithms leads to good service control and eventually optimal load allocation, under slow load mobility. .展开更多
Different from conventional cellular networks, a maritime communication base station(BS) has to cover a much wider area due to the limitation of available BS sites. Accordingly the performance of users far away from t...Different from conventional cellular networks, a maritime communication base station(BS) has to cover a much wider area due to the limitation of available BS sites. Accordingly the performance of users far away from the BS is poor in general. This renders the fairness among users a challenging issue for maritime communications. In this paper, we consider a practical massive MIMO maritime BS with hybrid digital and analog precoding. Only the large-scale channel state information at the transmitter(CSIT) is considered so as to reduce the implementation complexity and overhead of the system. On this basis, we address the problem of fairness-oriented precoding design. A max-min optimization problem is formulated and solved in an iterative way. Simulation results demonstrate that the proposed scheme performs much better than conventional hybrid precoding algorithms in terms of minimum achievable rate of all the users, for the typical three-ray maritime channel model.展开更多
Millimeter-wave(mm Wave) communications will be used in fifth-generation(5G) mobile communication systems, but they experience severe path loss and have high sensitivity to physical objects, leading to smaller cell ra...Millimeter-wave(mm Wave) communications will be used in fifth-generation(5G) mobile communication systems, but they experience severe path loss and have high sensitivity to physical objects, leading to smaller cell radii and complicated network architectures. A coverage extension scheme using large-scale antenna arrays(LSAAs) has been suggested and theoretically proven to be cost-efficient in combination with ultradense small cell networks. To analyze and optimize the LSAA-based network deployments, a comprehensive survey of recent advances in statistical mmWave channel modeling is first presented in terms of channel parameter estimation, large-scale path loss models, and small-scale cluster models. Next, the measurement and modeling results at two 5G candidate mmWave bands(e.g., 28 GHz and 39 GHz) are reviewed and compared in several outdoor scenarios of interest, where the propagation characteristics make crucial contributions to wireless network designs. Finally, the coverage behaviors of systems employing a large number of antenna arrays are discussed, as well as some implications on future mmWave cellular network designs.展开更多
A novel modulation recognition algorithm is proposed by introducing a Chen-Harker-Kanzow-Smale (CHKS) smooth function into the C-support vector machine deformation algorithm. A set of seven characteristic parameters i...A novel modulation recognition algorithm is proposed by introducing a Chen-Harker-Kanzow-Smale (CHKS) smooth function into the C-support vector machine deformation algorithm. A set of seven characteristic parameters is selected from a range of parameters of communication signals including instantaneous amplitude, phase, and frequency. And the Newton-Armijo algorithm is utilized to train the proposed algorithm, namely, smooth CHKS smooth support vector machine (SCHKS-SSVM). Compared with the existing algorithms, the proposed algorithm not only solves the non-differentiable problem of the second order objective function, but also reduces the recognition error. It significantly improves the training speed and also saves a large amount of storage space through large-scale sorting problems. The simulation results show that the recognition rate of the algorithm can batch training. Therefore, the proposed algorithm is suitable for solving the problem of high dimension and its recognition can exceed 95% when the signal-to-noise ratio is no less than 10 dB.展开更多
The lack of communication infrastructure in the ocean inevitably leads to coverage blind zones.In addition to high-throughput marine satellites,unmanned aerial vehicles(UAVs)can be used to provide coverage for these b...The lack of communication infrastructure in the ocean inevitably leads to coverage blind zones.In addition to high-throughput marine satellites,unmanned aerial vehicles(UAVs)can be used to provide coverage for these blind zones along with onshore base stations.In this paper,we consider the use of UAV for maritime coverage enhancement.Particularly,to serve more ships on the vast oceanic area with limited spectrum resources,we employ non-orthogonal multiple access(NOMA).A joint power and transmission duration allocation problem is formulated to maximize the minimum ship throughput,with the constraints on onboard communication energy.Different from previous works,we only assume the slowly time-varying large-scale channel state information(CSI)to reduce the system cost,as the large-scale CSI is locationdependent and can be obtained according to a priori radio map.To solve the non-convex problem,we decompose it into two subproblems and solve them in an iterative way.Simulation results show the effectiveness of the proposed solution.展开更多
With the increasing presence of robots in our daily life,there is a strong need and demand for the strategies to acquire a high quality interaction between robots and users by enabling robots to understand users’mood...With the increasing presence of robots in our daily life,there is a strong need and demand for the strategies to acquire a high quality interaction between robots and users by enabling robots to understand users’mood,intention,and other aspects.During human-human interaction,personality traits have an important influence on human behavior,decision,mood,and many others.Therefore,we propose an efficient computational framework to endow the robot with the capability of understanding the user’s personality traits based on the user’s nonverbal communication cues represented by three visual features including the head motion,gaze,and body motion energy,and three vocal features including voice pitch,voice energy,and mel-frequency cepstral coefficient(MFCC).We used the Pepper robot in this study as a communication robot to interact with each participant by asking questions,and meanwhile,the robot extracts the nonverbal features from each participant’s habitual behavior using its on-board sensors.On the other hand,each participant’s personality traits are evaluated with a questionnaire.We then train the ridge regression and linear support vector machine(SVM)classifiers using the nonverbal features and personality trait labels from a questionnaire and evaluate the performance of the classifiers.We have verified the validity of the proposed models that showed promising binary classification performance on recognizing each of the Big Five personality traits of the participants based on individual differences in nonverbal communication cues.展开更多
Protein-protein interactions are of great significance for human to understand the functional mechanisms of proteins.With the rapid development of high-throughput genomic technologies,massive protein-protein interacti...Protein-protein interactions are of great significance for human to understand the functional mechanisms of proteins.With the rapid development of high-throughput genomic technologies,massive protein-protein interaction(PPI)data have been generated,making it very difficult to analyze them efficiently.To address this problem,this paper presents a distributed framework by reimplementing one of state-of-the-art algorithms,i.e.,CoFex,using MapReduce.To do so,an in-depth analysis of its limitations is conducted from the perspectives of efficiency and memory consumption when applying it for large-scale PPI data analysis and prediction.Respective solutions are then devised to overcome these limitations.In particular,we adopt a novel tree-based data structure to reduce the heavy memory consumption caused by the huge sequence information of proteins.After that,its procedure is modified by following the MapReduce framework to take the prediction task distributively.A series of extensive experiments have been conducted to evaluate the performance of our framework in terms of both efficiency and accuracy.Experimental results well demonstrate that the proposed framework can considerably improve its computational efficiency by more than two orders of magnitude while retaining the same high accuracy.展开更多
BACKGROUND Large-scale functional connectivity(LSFC)patterns in the brain have unique intrinsic characteristics.Abnormal LSFC patterns have been found in patients with dementia,as well as in those with mild cognitive ...BACKGROUND Large-scale functional connectivity(LSFC)patterns in the brain have unique intrinsic characteristics.Abnormal LSFC patterns have been found in patients with dementia,as well as in those with mild cognitive impairment(MCI),and these patterns predicted their cognitive performance.It has been reported that patients with type 2 diabetes mellitus(T2DM)may develop MCI that could progress to dementia.We investigated whether we could adopt LSFC patterns as discriminative features to predict the cognitive function of patients with T2DM,using connectome-based predictive modeling(CPM)and a support vector machine.AIM To investigate the utility of LSFC for predicting cognitive impairment related to T2DM more accurately and reliably.METHODS Resting-state functional magnetic resonance images were derived from 42 patients with T2DM and 24 healthy controls.Cognitive function was assessed using the Montreal Cognitive Assessment(MoCA).Patients with T2DM were divided into two groups,according to the presence(T2DM-C;n=16)or absence(T2DM-NC;n=26)of MCI.Brain regions were marked using Harvard Oxford(HOA-112),automated anatomical labeling(AAL-116),and 264-region functional(Power-264)atlases.LSFC biomarkers for predicting MoCA scores were identified using a new CPM technique.Subsequently,we used a support vector machine based on LSFC patterns for among-group differentiation.The area under the receiver operating characteristic curve determined the appearance of the classification.RESULTS CPM could predict the MoCA scores in patients with T2DM(Pearson’s correlation coefficient between predicted and actual MoCA scores,r=0.32,P=0.0066[HOA-112 atlas];r=0.32,P=0.0078[AAL-116 atlas];r=0.42,P=0.0038[Power-264 atlas]),indicating that LSFC patterns represent cognition-level measures in these patients.Positive(anti-correlated)LSFC networks based on the Power-264 atlas showed the best predictive performance;moreover,we observed new brain regions of interest associated with T2DM-related cognition.The area under the receiver operating characteristic curve values(T2DM-NC group vs.T2DM-C group)were 0.65-0.70,with LSFC matrices based on HOA-112 and Power-264 atlases having the highest value(0.70).Most discriminative and attractive LSFCs were related to the default mode network,limbic system,and basal ganglia.CONCLUSION LSFC provides neuroimaging-based information that may be useful in detecting MCI early and accurately in patients with T2DM.展开更多
The Ethernet and field-bus communications are used in the machine control system (MCS) of HL-2A. The control net, with a programmable logic controller (PLC) as its logic control master, an engineering control mana...The Ethernet and field-bus communications are used in the machine control system (MCS) of HL-2A. The control net, with a programmable logic controller (PLC) as its logic control master, an engineering control management station as its net server, and a timing control PC connected to a number of terminals, flexibly and freely transfers information among the nodes on it with the Ethernet transmission techniques. The PLC masters the field bus, which carries small pieces of information between PLC and the field sites reliably and quickly. The control net is connected into the data net, where Internet access and sharing of more experimental data are enabled. The communication in the MCS guarantees the digitalization, automation and centralization. Also provided are a satisfactory degree of safety, reliability, stability, expandability and flexibility for maintenance.展开更多
Wireless networks are key enablers of ubiquitous communication. With the evolution of networking technologies and the need for these to inter-operate and dynamically adapt to user requirements, intelligent networks ar...Wireless networks are key enablers of ubiquitous communication. With the evolution of networking technologies and the need for these to inter-operate and dynamically adapt to user requirements, intelligent networks are the need of the hour. Use of machine learning techniques allows these networks to adapt to changing environments and enables them to make decisions while continuing to learn about their environment. In this paper, we survey the various problems of wireless networks that have been solved using machine-learning based prediction techniques and identify additional problems to which prediction can be applied. We also look at the gaps in the research done in this area till date.展开更多
文摘The Internet ofMedical Things(IoMT)is mainly concernedwith the efficient utilisation of wearable devices in the healthcare domain to manage various processes automatically,whereas machine learning approaches enable these smart systems to make informed decisions.Generally,broadcasting is used for the transmission of frames,whereas congestion,energy efficiency,and excessive load are among the common issues associated with existing approaches.In this paper,a machine learning-enabled shortest path identification scheme is presented to ensure reliable transmission of frames,especially with the minimum possible communication overheads in the IoMT network.For this purpose,the proposed scheme utilises a well-known technique,i.e.,Kruskal’s algorithm,to find an optimal path from source to destination wearable devices.Additionally,other evaluation metrics are used to find a reliable and shortest possible communication path between the two interested parties.Apart from that,every device is bound to hold a supplementary path,preferably a second optimised path,for situations where the current communication path is no longer available,either due to device failure or heavy traffic.Furthermore,the machine learning approach helps enable these devices to update their routing tables simultaneously,and an optimal path could be replaced if a better one is available.The proposed mechanism has been tested using a smart environment developed for the healthcare domain using IoMT networks.Simulation results show that the proposed machine learning-oriented approach performs better than existing approaches where the proposed scheme has achieved the minimum possible ratios,i.e.,17%and 23%,in terms of end to end delay and packet losses,respectively.Moreover,the proposed scheme has achieved an approximately 21%improvement in the average throughput compared to the existing schemes.
基金This work was supported by a research grant from Seoul Women’s University(2023-0183).
文摘Broadcasting gateway equipment generally uses a method of simply switching to a spare input stream when a failure occurs in a main input stream.However,when the transmission environment is unstable,problems such as reduction in the lifespan of equipment due to frequent switching and interruption,delay,and stoppage of services may occur.Therefore,applying a machine learning(ML)method,which is possible to automatically judge and classify network-related service anomaly,and switch multi-input signals without dropping or changing signals by predicting or quickly determining the time of error occurrence for smooth stream switching when there are problems such as transmission errors,is required.In this paper,we propose an intelligent packet switching method based on the ML method of classification,which is one of the supervised learning methods,that presents the risk level of abnormal multi-stream occurring in broadcasting gateway equipment based on data.Furthermore,we subdivide the risk levels obtained from classification techniques into probabilities and then derive vectorized representative values for each attribute value of the collected input data and continuously update them.The obtained reference vector value is used for switching judgment through the cosine similarity value between input data obtained when a dangerous situation occurs.In the broadcasting gateway equipment to which the proposed method is applied,it is possible to perform more stable and smarter switching than before by solving problems of reliability and broadcasting accidents of the equipment and can maintain stable video streaming as well.
基金supported in part by National Key Research and Develop⁃ment Young Scientist Project 2023YFB2905100the National Natural Sci⁃ence Foundation of China under Grant Nos.62201137 and 62331023+1 种基金the Fundamental Research Funds for the Central Universities under Grant No.2242022k60001the Research Fund of National Mobile Communications Research Laboratory,Southeast University,China under Grant No.2023A03.
文摘Extremely large-scale multiple-input multiple-output(XL-MIMO)and terahertz(THz)communications are pivotal candidate technologies for supporting the development of 6G mobile networks.However,these techniques invalidate the common assumptions of far-field plane waves and introduce many new properties.To accurately understand the performance of these new techniques,spherical wave modeling of near-field communications needs to be applied for future research.Hence,the investigation of near-field communication holds significant importance for the advancement of 6G,which brings many new and open research challenges in contrast to conventional far-field communication.In this paper,we first formulate a general model of the near-field channel and discuss the influence of spatial nonstationary properties on the near-field channel modeling.Subsequently,we discuss the challenges encountered in the near field in terms of beam training,localization,and transmission scheme design,respectively.Finally,we point out some promising research directions for near-field communications.
基金This work was supported by National Key R&D Program of China under Grant 2018YFB1800802in part by the National Natural Science Foundation of China under Grant No.61771488,No.61631020 and No.61827801+1 种基金in part by State Key Laboratory of Air Traffic Management System and Technology under Grant No.SKLATM201808in part by Postgraduate Research and Practice Innovation Program of Jiangsu Province under No.KYCX190188.
文摘As a result of rapid development in electronics and communication technology,large-scale unmanned aerial vehicles(UAVs)are harnessed for various promising applications in a coordinated manner.Although it poses numerous advantages,resource management among various domains in large-scale UAV communication networks is the key challenge to be solved urgently.Specifically,due to the inherent requirements and future development trend,distributed resource management is suitable.In this article,we investigate the resource management problem for large-scale UAV communication networks from game-theoretic perspective which are exactly coincident with the distributed and autonomous manner.By exploring the inherent features,the distinctive challenges are discussed.Then,we explore several gametheoretic models that not only combat the challenges but also have broad application prospects.We provide the basics of each game-theoretic model and discuss the potential applications for resource management in large-scale UAV communication networks.Specifically,mean-field game,graphical game,Stackelberg game,coalition game and potential game are included.After that,we propose two innovative case studies to highlight the feasibility of such novel game-theoretic models.Finally,we give some future research directions to shed light on future opportunities and applications.
基金supported by National Natural Science Foundation of China under Grant 61971062,Grant 92067202,Grant 62071058,Grant 62001049Beijing Natural Science Foundation under Grant 4222012Beijing University of Posts and Telecommunications China Mobile Research Institute Joint Innovation Center。
文摘Unsourced multiple access(UMA)is a multi-access technology for massive,low-power,uncoordinated,and unsourced Machine Type Communication(MTC)networks.It ensures transmission reliability under the premise of high energy efficiency.Based on the analysis of the 6G MTC key performance indicators(KPIs)and scenario characteristics,this paper summarizes its requirements for radio access networks.Following this,the existing multiple access models are analyzed under these standards to determine UMA's advantages for 6G MTC according to its design characteristics.The critical technology of UMA is the design of its multiple-access coding scheme.Therefore,the existing UMA coding schemes from different coding paradigms are further summarized and compared.In particular,this paper comprehensively considers the energy efficiency and computational complexity of these schemes,studies the changes of the above two indexes with the increase of access scale,and considers the trade-off between the two.It is revealed by the above analysis that some guiding rules of UMA coding design.Finally,the open problems and potentials in this field are given for future research.
基金supported by the Jiangsu Provincial Natural Science Fund for Outstanding Young Scholars(No.BK20180028)the Natural Science Foundations of China(No.61671474,No.61771487)the Jiangsu Provincial Natural Science Fund for Excellent Young Scholars(No.BK20170089).
文摘Massive machine type communications(mMTC)have been regarded as promising applications in the future.One main feature of mMTC is short packet communication.Different from traditional long packet communication,short packet communication suffers from transmission rate degradation and a significant error rate is introduced.In this case,traditional resource allocation scheme for mMTC is no longer applicable.In this paper,we explore resource allocation for cellular-based mMTC in the finite blocklength regime.First,to mitigate the load of the base station(BS),we establish a framework for cellularbased mMTC,where MTCGs reuse the resources of cellular users(CUs),aggregate the packets generated by MTCDs,and forward them to the BS.Next,we adopt short packet theory to obtain the minimum required blocklength of a packet that transmits a certain amount of information.Then,by modeling the process of MTCGs-assisted communication as a queuing process,we derive the closed-form expression of the average delay of all MTCDs.Guided by this,we propose a joint power allocation and spectrum sharing scheme to minimize the average delay.Finally,the simulation results verify the correctness of the theoretical results and show that the proposed scheme can reduce the average delay efficiently.
文摘Vehicular ad-hoc networks(VANETs)are mobile networks that use and transfer data with vehicles as the network nodes.Thus,VANETs are essentially mobile ad-hoc networks(MANETs).They allow all the nodes to communicate and connect with one another.One of the main requirements in a VANET is to provide self-decision capability to the vehicles.Cognitive memory,which stores all the previous routes,is used by the vehicles to choose the optimal route.In networks,communication is crucial.In cellular-based vehicle-to-everything(CV2X)communication,vital information is shared using the cooperative awareness message(CAM)that is broadcast by each vehicle.Resources are allocated in a distributed manner,which is known as Mode 4 communication.The support vector machine(SVM)algorithm is used in the SVM-CV2X-M4 system proposed in this study.The k-fold model with different values of k is used to evaluate the accuracy of the SVM-CV2XM4 system.The results show that the proposed system achieves an accuracy of 99.6%.Thus,the proposed system allows vehicles to choose the optimal route and is highly convenient for users.
文摘The design of large-scale machine system is a very complex problem.These design problems usually have a lot of design variables and constraints so that they are difficult to be solved rapidly and efficiently by using conventional methods.In this paper,a new multilevel design method oriented network environment is proposed,which maps the design problem of large-scale machine system into a hypergraph with degree of linking strength (DLS) between vertices.By decomposition of hypergraph,this method can divide the complex design problem into some small and simple subproblems that can be solved concurrently in a network.
文摘In recent years,the need for a fast,efficient and a reliable wireless network has increased dramatically.Numerous 5G networks have already been tested while a few are in the early stages of deployment.In noncooperative communication scenarios,the recognition of digital signal modulations assists people in identifying the communication targets and ensures an effective management over them.The recent advancements in both Machine Learning(ML)and Deep Learning(DL)models demand the development of effective modulation recognition models with self-learning capability.In this background,the current research article designs aDeep Learning enabled Intelligent Modulation Recognition of Communication Signal(DLIMR-CS)technique for next-generation networks.The aim of the proposed DLIMR-CS technique is to classify different kinds of digitally-modulated signals.In addition,the fractal feature extraction process is appliedwith the help of the Sevcik Fractal Dimension(SFD)approach.Then,the extracted features are fed into the Deep Variational Autoencoder(DVAE)model for the classification of the modulated signals.In order to improve the classification performance of the DVAE model,the Tunicate Swarm Algorithm(TSA)is used to finetune the hyperparameters involved in DVAE model.A wide range of simulations was conducted to establish the enhanced performance of the proposed DLIMR-CS model.The experimental outcomes confirmed the superior recognition rate of the DLIMR-CS model over recent state-of-the-art methods under different evaluation parameters.
文摘Key challenges for 5G and Beyond networks relate with the requirements for exceptionally low latency, high reliability, and extremely high data rates. The Ultra-Reliable Low Latency Communication (URLLC) use case is the trickiest to support and current research is focused on physical or MAC layer solutions, while proposals focused on the network layer using Machine Learning (ML) and Artificial Intelligence (AI) algorithms running on base stations and User Equipment (UE) or Internet of Things (IoT) devices are in early stages. In this paper, we describe the operation rationale of the most recent relevant ML algorithms and techniques, and we propose and validate ML algorithms running on both cells (base stations/gNBs) and UEs or IoT devices to handle URLLC service control. One ML algorithm runs on base stations to evaluate latency demands and offload traffic in case of need, while another lightweight algorithm runs on UEs and IoT devices to rank cells with the best URLLC service in real-time to indicate the best one cell for a UE or IoT device to camp. We show that the interplay of these algorithms leads to good service control and eventually optimal load allocation, under slow load mobility. .
基金supported in part by the National Science Foundation of China under grant No. 91638205,grant No. 61771286, and grant No. 61701457, and grant No. 61621091
文摘Different from conventional cellular networks, a maritime communication base station(BS) has to cover a much wider area due to the limitation of available BS sites. Accordingly the performance of users far away from the BS is poor in general. This renders the fairness among users a challenging issue for maritime communications. In this paper, we consider a practical massive MIMO maritime BS with hybrid digital and analog precoding. Only the large-scale channel state information at the transmitter(CSIT) is considered so as to reduce the implementation complexity and overhead of the system. On this basis, we address the problem of fairness-oriented precoding design. A max-min optimization problem is formulated and solved in an iterative way. Simulation results demonstrate that the proposed scheme performs much better than conventional hybrid precoding algorithms in terms of minimum achievable rate of all the users, for the typical three-ray maritime channel model.
基金supported in part by the National Natural Science Foundation of China under Grant No.61671145the Key R&D Program of Jiangsu Province of China under Grant BE2018121
文摘Millimeter-wave(mm Wave) communications will be used in fifth-generation(5G) mobile communication systems, but they experience severe path loss and have high sensitivity to physical objects, leading to smaller cell radii and complicated network architectures. A coverage extension scheme using large-scale antenna arrays(LSAAs) has been suggested and theoretically proven to be cost-efficient in combination with ultradense small cell networks. To analyze and optimize the LSAA-based network deployments, a comprehensive survey of recent advances in statistical mmWave channel modeling is first presented in terms of channel parameter estimation, large-scale path loss models, and small-scale cluster models. Next, the measurement and modeling results at two 5G candidate mmWave bands(e.g., 28 GHz and 39 GHz) are reviewed and compared in several outdoor scenarios of interest, where the propagation characteristics make crucial contributions to wireless network designs. Finally, the coverage behaviors of systems employing a large number of antenna arrays are discussed, as well as some implications on future mmWave cellular network designs.
基金supported by the National Natural Science Foundation of China(61401196)the Jiangsu Provincial Natural Science Foundation of China(BK20140954)+1 种基金the Science and Technology on Information Transmission and Dissemination in Communication Networks Laboratory(KX152600015/ITD-U15006)the Beijing Shengfeifan Electronic System Technology Development Co.,Ltd(KY10800150036)
文摘A novel modulation recognition algorithm is proposed by introducing a Chen-Harker-Kanzow-Smale (CHKS) smooth function into the C-support vector machine deformation algorithm. A set of seven characteristic parameters is selected from a range of parameters of communication signals including instantaneous amplitude, phase, and frequency. And the Newton-Armijo algorithm is utilized to train the proposed algorithm, namely, smooth CHKS smooth support vector machine (SCHKS-SSVM). Compared with the existing algorithms, the proposed algorithm not only solves the non-differentiable problem of the second order objective function, but also reduces the recognition error. It significantly improves the training speed and also saves a large amount of storage space through large-scale sorting problems. The simulation results show that the recognition rate of the algorithm can batch training. Therefore, the proposed algorithm is suitable for solving the problem of high dimension and its recognition can exceed 95% when the signal-to-noise ratio is no less than 10 dB.
基金supported in part by National Natural Science Foundation of China(No.61922049,61771286,61941104)the National Key R&D Program of China(2020YFA0711301)+2 种基金the Beijing National Research Center for Information Science and Technology project(BNR2020RC01016)the Nantong Technology Program(JC2019115)the Beijing Innovation Center for Future Chip。
文摘The lack of communication infrastructure in the ocean inevitably leads to coverage blind zones.In addition to high-throughput marine satellites,unmanned aerial vehicles(UAVs)can be used to provide coverage for these blind zones along with onshore base stations.In this paper,we consider the use of UAV for maritime coverage enhancement.Particularly,to serve more ships on the vast oceanic area with limited spectrum resources,we employ non-orthogonal multiple access(NOMA).A joint power and transmission duration allocation problem is formulated to maximize the minimum ship throughput,with the constraints on onboard communication energy.Different from previous works,we only assume the slowly time-varying large-scale channel state information(CSI)to reduce the system cost,as the large-scale CSI is locationdependent and can be obtained according to a priori radio map.To solve the non-convex problem,we decompose it into two subproblems and solve them in an iterative way.Simulation results show the effectiveness of the proposed solution.
基金supported by the EU-Japan coordinated R&D project on“Culture Aware Robots and Environmental Sensor Systems for Elderly Support,”commissioned by the Ministry of Internal Affairs and Communications of Japan and EC Horizon 2020 Research and Innovation Programme(737858)financial supports from the Air Force Office of Scientific Research(AFOSR-AOARD/FA2386-19-1-4015)。
文摘With the increasing presence of robots in our daily life,there is a strong need and demand for the strategies to acquire a high quality interaction between robots and users by enabling robots to understand users’mood,intention,and other aspects.During human-human interaction,personality traits have an important influence on human behavior,decision,mood,and many others.Therefore,we propose an efficient computational framework to endow the robot with the capability of understanding the user’s personality traits based on the user’s nonverbal communication cues represented by three visual features including the head motion,gaze,and body motion energy,and three vocal features including voice pitch,voice energy,and mel-frequency cepstral coefficient(MFCC).We used the Pepper robot in this study as a communication robot to interact with each participant by asking questions,and meanwhile,the robot extracts the nonverbal features from each participant’s habitual behavior using its on-board sensors.On the other hand,each participant’s personality traits are evaluated with a questionnaire.We then train the ridge regression and linear support vector machine(SVM)classifiers using the nonverbal features and personality trait labels from a questionnaire and evaluate the performance of the classifiers.We have verified the validity of the proposed models that showed promising binary classification performance on recognizing each of the Big Five personality traits of the participants based on individual differences in nonverbal communication cues.
基金This work was supported in part by the National Natural Science Foundation of China(61772493)the CAAI-Huawei MindSpore Open Fund(CAAIXSJLJJ-2020-004B)+4 种基金the Natural Science Foundation of Chongqing(China)(cstc2019jcyjjqX0013)Chongqing Research Program of Technology Innovation and Application(cstc2019jscx-fxydX0024,cstc2019jscx-fxydX0027,cstc2018jszx-cyzdX0041)Guangdong Province Universities and College Pearl River Scholar Funded Scheme(2019)the Pioneer Hundred Talents Program of Chinese Academy of Sciencesthe Deanship of Scientific Research(DSR)at King Abdulaziz University(G-21-135-38).
文摘Protein-protein interactions are of great significance for human to understand the functional mechanisms of proteins.With the rapid development of high-throughput genomic technologies,massive protein-protein interaction(PPI)data have been generated,making it very difficult to analyze them efficiently.To address this problem,this paper presents a distributed framework by reimplementing one of state-of-the-art algorithms,i.e.,CoFex,using MapReduce.To do so,an in-depth analysis of its limitations is conducted from the perspectives of efficiency and memory consumption when applying it for large-scale PPI data analysis and prediction.Respective solutions are then devised to overcome these limitations.In particular,we adopt a novel tree-based data structure to reduce the heavy memory consumption caused by the huge sequence information of proteins.After that,its procedure is modified by following the MapReduce framework to take the prediction task distributively.A series of extensive experiments have been conducted to evaluate the performance of our framework in terms of both efficiency and accuracy.Experimental results well demonstrate that the proposed framework can considerably improve its computational efficiency by more than two orders of magnitude while retaining the same high accuracy.
基金Supported by the National Natural Science Foundation of China,No.81771815.
文摘BACKGROUND Large-scale functional connectivity(LSFC)patterns in the brain have unique intrinsic characteristics.Abnormal LSFC patterns have been found in patients with dementia,as well as in those with mild cognitive impairment(MCI),and these patterns predicted their cognitive performance.It has been reported that patients with type 2 diabetes mellitus(T2DM)may develop MCI that could progress to dementia.We investigated whether we could adopt LSFC patterns as discriminative features to predict the cognitive function of patients with T2DM,using connectome-based predictive modeling(CPM)and a support vector machine.AIM To investigate the utility of LSFC for predicting cognitive impairment related to T2DM more accurately and reliably.METHODS Resting-state functional magnetic resonance images were derived from 42 patients with T2DM and 24 healthy controls.Cognitive function was assessed using the Montreal Cognitive Assessment(MoCA).Patients with T2DM were divided into two groups,according to the presence(T2DM-C;n=16)or absence(T2DM-NC;n=26)of MCI.Brain regions were marked using Harvard Oxford(HOA-112),automated anatomical labeling(AAL-116),and 264-region functional(Power-264)atlases.LSFC biomarkers for predicting MoCA scores were identified using a new CPM technique.Subsequently,we used a support vector machine based on LSFC patterns for among-group differentiation.The area under the receiver operating characteristic curve determined the appearance of the classification.RESULTS CPM could predict the MoCA scores in patients with T2DM(Pearson’s correlation coefficient between predicted and actual MoCA scores,r=0.32,P=0.0066[HOA-112 atlas];r=0.32,P=0.0078[AAL-116 atlas];r=0.42,P=0.0038[Power-264 atlas]),indicating that LSFC patterns represent cognition-level measures in these patients.Positive(anti-correlated)LSFC networks based on the Power-264 atlas showed the best predictive performance;moreover,we observed new brain regions of interest associated with T2DM-related cognition.The area under the receiver operating characteristic curve values(T2DM-NC group vs.T2DM-C group)were 0.65-0.70,with LSFC matrices based on HOA-112 and Power-264 atlases having the highest value(0.70).Most discriminative and attractive LSFCs were related to the default mode network,limbic system,and basal ganglia.CONCLUSION LSFC provides neuroimaging-based information that may be useful in detecting MCI early and accurately in patients with T2DM.
基金The project supported by National Natural Science Foundation of China (No. 10175022) and Sichuan Provincial Youth Foundation
文摘The Ethernet and field-bus communications are used in the machine control system (MCS) of HL-2A. The control net, with a programmable logic controller (PLC) as its logic control master, an engineering control management station as its net server, and a timing control PC connected to a number of terminals, flexibly and freely transfers information among the nodes on it with the Ethernet transmission techniques. The PLC masters the field bus, which carries small pieces of information between PLC and the field sites reliably and quickly. The control net is connected into the data net, where Internet access and sharing of more experimental data are enabled. The communication in the MCS guarantees the digitalization, automation and centralization. Also provided are a satisfactory degree of safety, reliability, stability, expandability and flexibility for maintenance.
文摘Wireless networks are key enablers of ubiquitous communication. With the evolution of networking technologies and the need for these to inter-operate and dynamically adapt to user requirements, intelligent networks are the need of the hour. Use of machine learning techniques allows these networks to adapt to changing environments and enables them to make decisions while continuing to learn about their environment. In this paper, we survey the various problems of wireless networks that have been solved using machine-learning based prediction techniques and identify additional problems to which prediction can be applied. We also look at the gaps in the research done in this area till date.