The forthcoming sixth generation(6G)of mobile communication networks is envisioned to be AInative,supporting intelligent services and pervasive computing at unprecedented scale.Among the key paradigms enabling this vi...The forthcoming sixth generation(6G)of mobile communication networks is envisioned to be AInative,supporting intelligent services and pervasive computing at unprecedented scale.Among the key paradigms enabling this vision,Federated Learning(FL)has gained prominence as a distributed machine learning framework that allows multiple devices to collaboratively train models without sharing raw data,thereby preserving privacy and reducing the need for centralized storage.This capability is particularly attractive for vision-based applications,where image and video data are both sensitive and bandwidth-intensive.However,the integration of FL with 6G networks presents unique challenges,including communication bottlenecks,device heterogeneity,and trade-offs between model accuracy,latency,and energy consumption.In this paper,we developed a simulation-based framework to investigate the performance of FL in representative vision tasks under 6G-like environments.We formalize the system model,incorporating both the federated averaging(FedAvg)training process and a simplified communication costmodel that captures bandwidth constraints,packet loss,and variable latency across edge devices.Using standard image datasets(e.g.,MNIST,CIFAR-10)as benchmarks,we analyze how factors such as the number of participating clients,degree of data heterogeneity,and communication frequency influence convergence speed and model accuracy.Additionally,we evaluate the effectiveness of lightweight communication-efficient strategies,including local update tuning and gradient compression,in mitigating network overhead.The experimental results reveal several key insights:(i)communication limitations can significantly degrade FL convergence in vision tasks if not properly addressed;(ii)judicious tuning of local training epochs and client participation levels enables notable improvements in both efficiency and accuracy;and(iii)communication-efficient FL strategies provide a promising pathway to balance performance with the stringent latency and reliability requirements expected in 6G.These findings highlight the synergistic role of AI and nextgeneration networks in enabling privacy-preserving,real-time vision applications,and they provide concrete design guidelines for researchers and practitioners working at the intersection of FL and 6G.展开更多
In order to support the future digital society,sixth generation(6G)network faces the challenge to work efficiently and flexibly in a wider range of scenarios.The traditional way of system design is to sequentially get...In order to support the future digital society,sixth generation(6G)network faces the challenge to work efficiently and flexibly in a wider range of scenarios.The traditional way of system design is to sequentially get the electromagnetic wave propagation model of typical scenarios firstly and then do the network design by simulation offline,which obviously leads to a 6G network lacking of adaptation to dynamic environments.Recently,with the aid of sensing enhancement,more environment information can be obtained.Based on this,from radio wave propagation perspective,we propose a predictive 6G network with environment sensing enhancement,the electromagnetic wave propagation characteristics prediction enabled network(EWave Net),to further release the potential of 6G.To this end,a prediction plane is created to sense,predict and utilize the physical environment information in EWave Net to realize the electromagnetic wave propagation characteristics prediction timely.A two-level closed feedback workflow is also designed to enhance the sensing and prediction ability for EWave Net.Several promising application cases of EWave Net are analyzed and the open issues to achieve this goal are addressed finally.展开更多
With the evolution of the sixth generation(6G)mobile communication technology,ample attention has gone to the integrated terrestrial-satellite networks.This paper notes that four typical application scenarios of integ...With the evolution of the sixth generation(6G)mobile communication technology,ample attention has gone to the integrated terrestrial-satellite networks.This paper notes that four typical application scenarios of integrated terrestrial-satellite networks are integrated into ultra dense satellite-enabled 6G networks architecture.Then the subchannel and power allocation schemes for the downlink of the ultra dense satellite-enabled 6G heterogeneous networks are introduced.Satellite mobile edge computing(SMEC)with edge caching in three-layer heterogeneous networks serves to reduce the link traffic of networks.Furthermore,a scheme for interference management is presented,involving quality-of-service(QoS)and co-tier/cross-tier interference constraints.The simulation results show that the proposed schemes can significantly increase the total capacity of ultra dense satellite-enabled 6G heterogeneous networks.展开更多
Future 6G communications are envisioned to enable a large catalogue of pioneering applications.These will range from networked Cyber-Physical Systems to edge computing devices,establishing real-time feedback control l...Future 6G communications are envisioned to enable a large catalogue of pioneering applications.These will range from networked Cyber-Physical Systems to edge computing devices,establishing real-time feedback control loops critical for managing Industry 5.0 deployments,digital agriculture systems,and essential infrastructures.The provision of extensive machine-type communications through 6G will render many of these innovative systems autonomous and unsupervised.While full automation will enhance industrial efficiency significantly,it concurrently introduces new cyber risks and vulnerabilities.In particular,unattended systems are highly susceptible to trust issues:malicious nodes and false information can be easily introduced into control loops.Additionally,Denialof-Service attacks can be executed by inundating the network with valueless noise.Current anomaly detection schemes require the entire transformation of the control software to integrate new steps and can only mitigate anomalies that conform to predefined mathematical models.Solutions based on an exhaustive data collection to detect anomalies are precise but extremely slow.Standard models,with their limited understanding of mobile networks,can achieve precision rates no higher than 75%.Therefore,more general and transversal protection mechanisms are needed to detect malicious behaviors transparently.This paper introduces a probabilistic trust model and control algorithm designed to address this gap.The model determines the probability of any node to be trustworthy.Communication channels are pruned for those nodes whose probability is below a given threshold.The trust control algorithmcomprises three primary phases,which feed themodel with three different probabilities,which are weighted and combined.Initially,anomalous nodes are identified using Gaussian mixture models and clustering technologies.Next,traffic patterns are studied using digital Bessel functions and the functional scalar product.Finally,the information coherence and content are analyzed.The noise content and abnormal information sequences are detected using a Volterra filter and a bank of Finite Impulse Response filters.An experimental validation based on simulation tools and environments was carried out.Results show the proposed solution can successfully detect up to 92%of malicious data injection attacks.展开更多
In recent years,the need for a fast,efficient and a reliable wireless network has increased dramatically.Numerous 5G networks have already been tested while a few are in the early stages of deployment.In noncooperativ...In recent years,the need for a fast,efficient and a reliable wireless network has increased dramatically.Numerous 5G networks have already been tested while a few are in the early stages of deployment.In noncooperative communication scenarios,the recognition of digital signal modulations assists people in identifying the communication targets and ensures an effective management over them.The recent advancements in both Machine Learning(ML)and Deep Learning(DL)models demand the development of effective modulation recognition models with self-learning capability.In this background,the current research article designs aDeep Learning enabled Intelligent Modulation Recognition of Communication Signal(DLIMR-CS)technique for next-generation networks.The aim of the proposed DLIMR-CS technique is to classify different kinds of digitally-modulated signals.In addition,the fractal feature extraction process is appliedwith the help of the Sevcik Fractal Dimension(SFD)approach.Then,the extracted features are fed into the Deep Variational Autoencoder(DVAE)model for the classification of the modulated signals.In order to improve the classification performance of the DVAE model,the Tunicate Swarm Algorithm(TSA)is used to finetune the hyperparameters involved in DVAE model.A wide range of simulations was conducted to establish the enhanced performance of the proposed DLIMR-CS model.The experimental outcomes confirmed the superior recognition rate of the DLIMR-CS model over recent state-of-the-art methods under different evaluation parameters.展开更多
Non-orthogonal multiple access (NOMA), multiple-input multiple-output (MIMO) and mobile edge computing (MEC) are prominent technologies to meet high data rate demand in the sixth generation (6G) communication networks...Non-orthogonal multiple access (NOMA), multiple-input multiple-output (MIMO) and mobile edge computing (MEC) are prominent technologies to meet high data rate demand in the sixth generation (6G) communication networks. In this paper, we aim to minimize the transmission delay in the MIMO-MEC in order to improve the spectral efficiency, energy efficiency, and data rate of MEC offloading. Dinkelbach transform and generalized singular value decomposition (GSVD) method are used to solve the delay minimization problem. Analytical results are provided to evaluate the performance of the proposed Hybrid-NOMA-MIMO-MEC system. Simulation results reveal that the H-NOMA-MIMO-MEC system can achieve better delay performance and lower energy consumption compared to OMA.展开更多
Sixth Generation(6G)wireless communication network has been expected to provide global coverage,enhanced spectral efficiency,and AI(Artificial Intelligence)-native intelligence,etc.To meet these requirements,the compu...Sixth Generation(6G)wireless communication network has been expected to provide global coverage,enhanced spectral efficiency,and AI(Artificial Intelligence)-native intelligence,etc.To meet these requirements,the computational concept of Decision-Making of cognition intelligence,its implementation framework adapting to foreseen innovations on networks and services,and its empirical evaluations are key techniques to guarantee the generationagnostic intelligence evolution of wireless communication networks.In this paper,we propose an Intelligent Decision Making(IDM)framework,acting as the role of network brain,based on Reinforcement Learning modelling philosophy to empower autonomous intelligence evolution capability to 6G network.Besides,usage scenarios and simulation demonstrate the generality and efficiency of IDM.We hope that some of the ideas of IDM will assist the research of 6G network in a new or different light.展开更多
In this paper,we propose a new privacy-aware transmission scheduling algorithm for 6G ad hoc networks.This system enables end nodes to select the optimum time and scheme to transmit private data safely.In 6G dynamic h...In this paper,we propose a new privacy-aware transmission scheduling algorithm for 6G ad hoc networks.This system enables end nodes to select the optimum time and scheme to transmit private data safely.In 6G dynamic heterogeneous infrastructures,unstable links and non-uniform hardware capabilities create critical issues regarding security and privacy.Traditional protocols are often too computationally heavy to allow 6G services to achieve their expected Quality-of-Service(QoS).As the transport network is built of ad hoc nodes,there is no guarantee about their trustworthiness or behavior,and transversal functionalities are delegated to the extreme nodes.However,while security can be guaranteed in extreme-to-extreme solutions,privacy cannot,as all intermediate nodes still have to handle the data packets they are transporting.Besides,traditional schemes for private anonymous ad hoc communications are vulnerable against modern intelligent attacks based on learning models.The proposed scheme fulfills this gap.Findings show the probability of a successful intelligent attack reduces by up to 65%compared to ad hoc networks with no privacy protection strategy when used the proposed technology.While congestion probability can remain below 0.001%,as required in 6G services.展开更多
The advent of sixth-generation(6G)networks introduces unprecedented challenges in achieving seamless connectivity,ultra-low latency,and efficient resource management in highly dynamic environments.Although fifth-gener...The advent of sixth-generation(6G)networks introduces unprecedented challenges in achieving seamless connectivity,ultra-low latency,and efficient resource management in highly dynamic environments.Although fifth-generation(5G)networks transformed mobile broadband and machine-type communications at massive scales,their properties of scaling,interference management,and latency remain a limitation in dense high mobility settings.To overcome these limitations,artificial intelligence(AI)and unmanned aerial vehicles(UAVs)have emerged as potential solutions to develop versatile,dynamic,and energy-efficient communication systems.The study proposes an AI-based UAV architecture that utilizes cooperative reinforcement learning(CoRL)to manage an autonomous network.The UAVs collaborate by sharing local observations and real-time state exchanges to optimize user connectivity,movement directions,allocate power,and resource distribution.Unlike conventional centralized or autonomous methods,CoRL involves joint state sharing and conflict-sensitive reward shaping,which ensures fair coverage,less interference,and enhanced adaptability in a dynamic urban environment.Simulations conducted in smart city scenarios with 10 UAVs and 50 ground users demonstrate that the proposed CoRL-based UAV system increases user coverage by up to 10%,achieves convergence 40%faster,and reduces latency and energy consumption by 30%compared with centralized and decentralized baselines.Furthermore,the distributed nature of the algorithm ensures scalability and flexibility,making it well-suited for future large-scale 6G deployments.The results highlighted that AI-enabled UAV systems enhance connectivity,support ultra-reliable low-latency communications(URLLC),and improve 6G network efficiency.Future work will extend the framework with adaptive modulation,beamforming-aware positioning,and real-world testbed deployment.展开更多
Network architectures assisted by Generative Artificial Intelligence(GAI)are envisioned as foundational elements of sixth-generation(6G)communication system.To deliver ubiquitous intelligent services and meet diverse ...Network architectures assisted by Generative Artificial Intelligence(GAI)are envisioned as foundational elements of sixth-generation(6G)communication system.To deliver ubiquitous intelligent services and meet diverse service requirements,6G network architecture should offer personalized services to various mobile devices.Federated learning(FL)with personalized local training,as a privacypreserving machine learning(ML)approach,can be applied to address these challenges.In this paper,we propose a meta-learning-based personalized FL(PFL)method that improves both communication and computation efficiency by utilizing over-the-air computations.Its“pretraining-and-fine-tuning”principle makes it particularly suitable for enabling edge nodes to access personalized GAI services while preserving local privacy.Experiment results demonstrate the outperformance and efficacy of the proposed algorithm,and notably indicate enhanced communication efficiency without compromising accuracy.展开更多
5G/5G-Advanced在持续提升关键性能指标方面被寄予厚望,需要在时延、可靠性、连接数密度与用户体验等方面实现进一步突破。传统以人工操作为主的管理模式在效率、准确性与成本等方面的局限日益凸显。相较于传统优化方法,人工智能技术凭...5G/5G-Advanced在持续提升关键性能指标方面被寄予厚望,需要在时延、可靠性、连接数密度与用户体验等方面实现进一步突破。传统以人工操作为主的管理模式在效率、准确性与成本等方面的局限日益凸显。相较于传统优化方法,人工智能技术凭借其预测性与前瞻性,推动网络管理由被动应对转向主动感知与自优化,实现从“监测-响应”到“预判-编排”的迁移。基于3GPP在无线电接入网(radio access network,RAN)智能化方向的关键技术与标准化路径,结合典型用例场景,分析了AI/ML模型管理、数据采集与交互机制。面向6G智能RAN,进一步提出“意图驱动的协作任务”这一新型架构理念,其关键是通过RAN对应用层信息的感知、任务级别的服务质量(quality of service,QoS)监控、动态组和资源管理等技术实现6G网络人机及碳硅生态系统的无缝交互。展开更多
Although 6G networks combined with artificial intelligence present revolutionary prospects for healthcare delivery,resource management in dense medical device networks stays a basic issue.Reliable communication direct...Although 6G networks combined with artificial intelligence present revolutionary prospects for healthcare delivery,resource management in dense medical device networks stays a basic issue.Reliable communication directly affects patient outcomes in these settings;nonetheless,current resource allocation techniques struggle with complicated interference patterns and different service needs of AI-native healthcare systems.In dense installations where conventional approaches fail,this paper tackles the challenge of combining network efficiency with medical care priority.Thus,we offer a Dueling Deep Q-Network(DDQN)-based resource allocation approach for AI-native healthcare systems in 6G dense networks.First,we create a point-line graph coloringbased interference model to capture the unique characteristics of medical device communications.Building on this foundation,we suggest a DDQN approach to optimal resource allocation over multiple medical services by combining advantage estimate with healthcare-aware state evaluation.Unlike traditional graph-based models,this one correctly depicts the overlapping coverage areas common in hospital environments.Building on this basis,our DDQN design allows the system to prioritize medical needs while distributing resources by separating healthcare state assessment from advantage estimation.Experimental findings show that the suggested DDQN outperforms state-of-the-art techniques in dense healthcare installations by 14.6%greater network throughput and 13.7%better resource use.The solution shows particularly strong in maintaining service quality under vital conditions with 5.5%greater Qo S satisfaction for emergency services and 8.2%quicker recovery from interruptions.展开更多
随着6G的快速发展,天地一体化信息网络成为实现全球无缝覆盖的关键。围绕6G天地一体化网络中的调制技术展开专利分析,重点研究了基于循环前缀(Cyclic Prefix,CP)/离散傅里叶变换(Discrete Fourier Transform,DFT)扩展的正交频分复用(Ort...随着6G的快速发展,天地一体化信息网络成为实现全球无缝覆盖的关键。围绕6G天地一体化网络中的调制技术展开专利分析,重点研究了基于循环前缀(Cyclic Prefix,CP)/离散傅里叶变换(Discrete Fourier Transform,DFT)扩展的正交频分复用(Orthogonal Frequency Division Multiplexing,OFDM)、基于滤波器组/多带滤波的OFDM、正交啁啾分复用(Orthogonal Chirp Division Multiplexing,OCDM)、正交时频空间(Orthogonal Time Frequency and Space,OTFS)和仿射频分复用(Affine Frequency Division Multiplexing,AFDM)这5种调制技术的专利申请趋势、技术热点及竞争格局。研究结果表明,2020年后6G调制技术专利申请量激增,中国在专利数量上占据主导地位,AFDM因其在高动态信道中的优异性能成为未来6G标准的有力候选。揭示了专利领域的技术空白,为后续研发和专利布局提供了战略参考。展开更多
随着第六代移动通信系统(6th generation mobile communication system, 6G)通信技术的发展,空天地一体化网络(Spaceair-ground integrated network, SAGIN)作为6G的重要组成部分,旨在实现卫星、空中平台与地面系统的无缝互联,在应急通...随着第六代移动通信系统(6th generation mobile communication system, 6G)通信技术的发展,空天地一体化网络(Spaceair-ground integrated network, SAGIN)作为6G的重要组成部分,旨在实现卫星、空中平台与地面系统的无缝互联,在应急通信、环境监测、智能交通等领域展现出巨大的潜力.然而,SAGIN具有异构结构、链路动态性高、资源分布广泛等特征,给网络的高效管理与优化带来巨大的挑战.近年来,人工智能(Artificial intelligence, AI)技术凭借强大的感知、学习与自主决策能力应用于通信网络,为SAGIN的智能演进提供了新契机.本文首先系统介绍SAGIN网络架构的基本组成与关键特征,并梳理当前主流AI技术在网络优化中的主要技术体系与适配优势,包括机器学习、图神经网络以及强化学习.其次,本文深入探讨了AI技术在SAGIN中智能资源管理、移动性管理与路由优化、空中平台路径规划、任务卸载与计算协同等典型场景中的应用与最新进展.最后,本文总结了AI技术应用在SAGIN网络中面临的挑战并展望了AI与SAGIN融合发展的未来方向.本文概述了AI技术在SAGIN网络中应用的优势与进展,旨在为AI赋能的SAGIN研究与应用发展提供技术参考.展开更多
The sixth-generation(6G)wireless communication networks are anticipated in integrating aerial,terrestrial,and maritime communication into a robust system to accomplish trustworthy,quick,and low latency needs.It enable...The sixth-generation(6G)wireless communication networks are anticipated in integrating aerial,terrestrial,and maritime communication into a robust system to accomplish trustworthy,quick,and low latency needs.It enables to achieve maximum throughput and delay for several applications.Besides,the evolution of 6G leads to the design of unmanned aerial vehicles(UAVs)in providing inexpensive and effective solutions in various application areas such as healthcare,environment monitoring,and so on.In the UAV network,effective data collection with restricted energy capacity poses a major issue to achieving high quality network communication.It can be addressed by the use of clustering techniques forUAVs in 6G networks.In this aspect,this study develops a novel metaheuristic based energy efficient data gathering scheme for clustered unmanned aerial vehicles(MEEDG-CUAV).The proposed MEEDG-CUAV technique intends in partitioning the UAV networks into various clusters and assign a cluster head(CH)to reduce the overall energy utilization.Besides,the quantum chaotic butterfly optimization algorithm(QCBOA)with a fitness function is derived to choose CHs and construct clusters.The experimental validation of the MEEDG-CUAV technique occurs utilizing benchmark dataset and the experimental results highlighted the better performance over the other state of art techniques interms of different measures.展开更多
The Internet of Things(IoT)is the fourth technological revolution in the global information industry after computers,the Internet,and mobile communication networks.It combines radio-frequency identification devices,in...The Internet of Things(IoT)is the fourth technological revolution in the global information industry after computers,the Internet,and mobile communication networks.It combines radio-frequency identification devices,infrared sensors,global positioning systems,and various other technologies.Information sensing equipment is connected via the Internet,thus forming a vast network.When these physical devices are connected to the Internet,the user terminal can be extended and expanded to exchange information,communicate with anything,and carry out identification,positioning,tracking,monitoring,and triggering of corresponding events on each device in the network.In real life,the IoT has a wide range of applications,covering many fields,such as smart homes,smart logistics,fine agriculture and animal husbandry,national defense,and military.One of the most significant factors in wireless channels is interference,which degrades the system performance.Although the existing QR decomposition-based signal detection method is an emerging topic because of its low complexity,it does not solve the problem of poor detection performance.Therefore,this study proposes a maximumlikelihood-based QR decomposition algorithm.The main idea is to estimate the initial level of detection using the maximum likelihood principle,and then the other layer is detected using a reliable decision.The optimal candidate is selected from the feedback by deploying the candidate points in an unreliable scenario.Simulation results show that the proposed algorithm effectively reduces the interference and propagation error compared with the algorithms reported in the literature.展开更多
5G networks apply adaptive modulation and coding according to the channel condition reported by the user in order to keep the mobile communication quality.However,the delay incurred by the feedback may make the channe...5G networks apply adaptive modulation and coding according to the channel condition reported by the user in order to keep the mobile communication quality.However,the delay incurred by the feedback may make the channel quality indicator(CQI)obsolete.This paper addresses this issue by proposing two approaches,one based on machine learning and another on evolutionary computing,which considers the user context and signal-to-interference-plus-noise ratio(SINR)besides the delay length to estimate the updated SINR to be mapped into a CQI value.Our proposals are designed to run at the user equipment(UE)side,neither requiring any change in the signalling between the base station(gNB)and UE nor overloading the gNB.They are evaluated in terms of mean squared error by adopting 5G network simulation data and the results show their high accuracy and feasibility to be employed in 5G/6G systems.展开更多
文摘The forthcoming sixth generation(6G)of mobile communication networks is envisioned to be AInative,supporting intelligent services and pervasive computing at unprecedented scale.Among the key paradigms enabling this vision,Federated Learning(FL)has gained prominence as a distributed machine learning framework that allows multiple devices to collaboratively train models without sharing raw data,thereby preserving privacy and reducing the need for centralized storage.This capability is particularly attractive for vision-based applications,where image and video data are both sensitive and bandwidth-intensive.However,the integration of FL with 6G networks presents unique challenges,including communication bottlenecks,device heterogeneity,and trade-offs between model accuracy,latency,and energy consumption.In this paper,we developed a simulation-based framework to investigate the performance of FL in representative vision tasks under 6G-like environments.We formalize the system model,incorporating both the federated averaging(FedAvg)training process and a simplified communication costmodel that captures bandwidth constraints,packet loss,and variable latency across edge devices.Using standard image datasets(e.g.,MNIST,CIFAR-10)as benchmarks,we analyze how factors such as the number of participating clients,degree of data heterogeneity,and communication frequency influence convergence speed and model accuracy.Additionally,we evaluate the effectiveness of lightweight communication-efficient strategies,including local update tuning and gradient compression,in mitigating network overhead.The experimental results reveal several key insights:(i)communication limitations can significantly degrade FL convergence in vision tasks if not properly addressed;(ii)judicious tuning of local training epochs and client participation levels enables notable improvements in both efficiency and accuracy;and(iii)communication-efficient FL strategies provide a promising pathway to balance performance with the stringent latency and reliability requirements expected in 6G.These findings highlight the synergistic role of AI and nextgeneration networks in enabling privacy-preserving,real-time vision applications,and they provide concrete design guidelines for researchers and practitioners working at the intersection of FL and 6G.
基金supported by the National Natural Science Foundation of China(No.92167202,61925102,U21B2014,62101069)the National Key R&D Program of China(No.2020YFB1805002)。
文摘In order to support the future digital society,sixth generation(6G)network faces the challenge to work efficiently and flexibly in a wider range of scenarios.The traditional way of system design is to sequentially get the electromagnetic wave propagation model of typical scenarios firstly and then do the network design by simulation offline,which obviously leads to a 6G network lacking of adaptation to dynamic environments.Recently,with the aid of sensing enhancement,more environment information can be obtained.Based on this,from radio wave propagation perspective,we propose a predictive 6G network with environment sensing enhancement,the electromagnetic wave propagation characteristics prediction enabled network(EWave Net),to further release the potential of 6G.To this end,a prediction plane is created to sense,predict and utilize the physical environment information in EWave Net to realize the electromagnetic wave propagation characteristics prediction timely.A two-level closed feedback workflow is also designed to enhance the sensing and prediction ability for EWave Net.Several promising application cases of EWave Net are analyzed and the open issues to achieve this goal are addressed finally.
基金supported in part by the National Key R&D Program of China(2020YFB1806103)the National Natural Science Foundation of China under Grant 62225103 and U22B2003+1 种基金Beijing Natural Science Foundation(L212004)China University Industry-University-Research Collaborative Innovation Fund(2021FNA05001).
文摘With the evolution of the sixth generation(6G)mobile communication technology,ample attention has gone to the integrated terrestrial-satellite networks.This paper notes that four typical application scenarios of integrated terrestrial-satellite networks are integrated into ultra dense satellite-enabled 6G networks architecture.Then the subchannel and power allocation schemes for the downlink of the ultra dense satellite-enabled 6G heterogeneous networks are introduced.Satellite mobile edge computing(SMEC)with edge caching in three-layer heterogeneous networks serves to reduce the link traffic of networks.Furthermore,a scheme for interference management is presented,involving quality-of-service(QoS)and co-tier/cross-tier interference constraints.The simulation results show that the proposed schemes can significantly increase the total capacity of ultra dense satellite-enabled 6G heterogeneous networks.
基金funding by Comunidad de Madrid within the framework of the Multiannual Agreement with Universidad Politécnica de Madrid to encourage research by young doctors(PRINCE project).
文摘Future 6G communications are envisioned to enable a large catalogue of pioneering applications.These will range from networked Cyber-Physical Systems to edge computing devices,establishing real-time feedback control loops critical for managing Industry 5.0 deployments,digital agriculture systems,and essential infrastructures.The provision of extensive machine-type communications through 6G will render many of these innovative systems autonomous and unsupervised.While full automation will enhance industrial efficiency significantly,it concurrently introduces new cyber risks and vulnerabilities.In particular,unattended systems are highly susceptible to trust issues:malicious nodes and false information can be easily introduced into control loops.Additionally,Denialof-Service attacks can be executed by inundating the network with valueless noise.Current anomaly detection schemes require the entire transformation of the control software to integrate new steps and can only mitigate anomalies that conform to predefined mathematical models.Solutions based on an exhaustive data collection to detect anomalies are precise but extremely slow.Standard models,with their limited understanding of mobile networks,can achieve precision rates no higher than 75%.Therefore,more general and transversal protection mechanisms are needed to detect malicious behaviors transparently.This paper introduces a probabilistic trust model and control algorithm designed to address this gap.The model determines the probability of any node to be trustworthy.Communication channels are pruned for those nodes whose probability is below a given threshold.The trust control algorithmcomprises three primary phases,which feed themodel with three different probabilities,which are weighted and combined.Initially,anomalous nodes are identified using Gaussian mixture models and clustering technologies.Next,traffic patterns are studied using digital Bessel functions and the functional scalar product.Finally,the information coherence and content are analyzed.The noise content and abnormal information sequences are detected using a Volterra filter and a bank of Finite Impulse Response filters.An experimental validation based on simulation tools and environments was carried out.Results show the proposed solution can successfully detect up to 92%of malicious data injection attacks.
文摘In recent years,the need for a fast,efficient and a reliable wireless network has increased dramatically.Numerous 5G networks have already been tested while a few are in the early stages of deployment.In noncooperative communication scenarios,the recognition of digital signal modulations assists people in identifying the communication targets and ensures an effective management over them.The recent advancements in both Machine Learning(ML)and Deep Learning(DL)models demand the development of effective modulation recognition models with self-learning capability.In this background,the current research article designs aDeep Learning enabled Intelligent Modulation Recognition of Communication Signal(DLIMR-CS)technique for next-generation networks.The aim of the proposed DLIMR-CS technique is to classify different kinds of digitally-modulated signals.In addition,the fractal feature extraction process is appliedwith the help of the Sevcik Fractal Dimension(SFD)approach.Then,the extracted features are fed into the Deep Variational Autoencoder(DVAE)model for the classification of the modulated signals.In order to improve the classification performance of the DVAE model,the Tunicate Swarm Algorithm(TSA)is used to finetune the hyperparameters involved in DVAE model.A wide range of simulations was conducted to establish the enhanced performance of the proposed DLIMR-CS model.The experimental outcomes confirmed the superior recognition rate of the DLIMR-CS model over recent state-of-the-art methods under different evaluation parameters.
基金supported by Republic of Turkey Ministry of National Education
文摘Non-orthogonal multiple access (NOMA), multiple-input multiple-output (MIMO) and mobile edge computing (MEC) are prominent technologies to meet high data rate demand in the sixth generation (6G) communication networks. In this paper, we aim to minimize the transmission delay in the MIMO-MEC in order to improve the spectral efficiency, energy efficiency, and data rate of MEC offloading. Dinkelbach transform and generalized singular value decomposition (GSVD) method are used to solve the delay minimization problem. Analytical results are provided to evaluate the performance of the proposed Hybrid-NOMA-MIMO-MEC system. Simulation results reveal that the H-NOMA-MIMO-MEC system can achieve better delay performance and lower energy consumption compared to OMA.
基金supported by National Key Research and Development Project 2018YFE0205503Beijing University of Posts and Telecommunications-China Mobile Research Institute Joint Innovation Center。
文摘Sixth Generation(6G)wireless communication network has been expected to provide global coverage,enhanced spectral efficiency,and AI(Artificial Intelligence)-native intelligence,etc.To meet these requirements,the computational concept of Decision-Making of cognition intelligence,its implementation framework adapting to foreseen innovations on networks and services,and its empirical evaluations are key techniques to guarantee the generationagnostic intelligence evolution of wireless communication networks.In this paper,we propose an Intelligent Decision Making(IDM)framework,acting as the role of network brain,based on Reinforcement Learning modelling philosophy to empower autonomous intelligence evolution capability to 6G network.Besides,usage scenarios and simulation demonstrate the generality and efficiency of IDM.We hope that some of the ideas of IDM will assist the research of 6G network in a new or different light.
基金funding from the European Commission by the Ruralities project(grant agreement no.101060876).
文摘In this paper,we propose a new privacy-aware transmission scheduling algorithm for 6G ad hoc networks.This system enables end nodes to select the optimum time and scheme to transmit private data safely.In 6G dynamic heterogeneous infrastructures,unstable links and non-uniform hardware capabilities create critical issues regarding security and privacy.Traditional protocols are often too computationally heavy to allow 6G services to achieve their expected Quality-of-Service(QoS).As the transport network is built of ad hoc nodes,there is no guarantee about their trustworthiness or behavior,and transversal functionalities are delegated to the extreme nodes.However,while security can be guaranteed in extreme-to-extreme solutions,privacy cannot,as all intermediate nodes still have to handle the data packets they are transporting.Besides,traditional schemes for private anonymous ad hoc communications are vulnerable against modern intelligent attacks based on learning models.The proposed scheme fulfills this gap.Findings show the probability of a successful intelligent attack reduces by up to 65%compared to ad hoc networks with no privacy protection strategy when used the proposed technology.While congestion probability can remain below 0.001%,as required in 6G services.
基金supported by the National Research Foundation of Korea(NRF)grant funded by the Korea government(MSIT)(RS-2025-00559546)supported by the IITP(Institute of Information&Coummunications Technology Planning&Evaluation)-ITRC(Information Technology Research Center)grant funded by the Korea government(Ministry of Science and ICT)(IITP-2025-RS-2023-00259004).
文摘The advent of sixth-generation(6G)networks introduces unprecedented challenges in achieving seamless connectivity,ultra-low latency,and efficient resource management in highly dynamic environments.Although fifth-generation(5G)networks transformed mobile broadband and machine-type communications at massive scales,their properties of scaling,interference management,and latency remain a limitation in dense high mobility settings.To overcome these limitations,artificial intelligence(AI)and unmanned aerial vehicles(UAVs)have emerged as potential solutions to develop versatile,dynamic,and energy-efficient communication systems.The study proposes an AI-based UAV architecture that utilizes cooperative reinforcement learning(CoRL)to manage an autonomous network.The UAVs collaborate by sharing local observations and real-time state exchanges to optimize user connectivity,movement directions,allocate power,and resource distribution.Unlike conventional centralized or autonomous methods,CoRL involves joint state sharing and conflict-sensitive reward shaping,which ensures fair coverage,less interference,and enhanced adaptability in a dynamic urban environment.Simulations conducted in smart city scenarios with 10 UAVs and 50 ground users demonstrate that the proposed CoRL-based UAV system increases user coverage by up to 10%,achieves convergence 40%faster,and reduces latency and energy consumption by 30%compared with centralized and decentralized baselines.Furthermore,the distributed nature of the algorithm ensures scalability and flexibility,making it well-suited for future large-scale 6G deployments.The results highlighted that AI-enabled UAV systems enhance connectivity,support ultra-reliable low-latency communications(URLLC),and improve 6G network efficiency.Future work will extend the framework with adaptive modulation,beamforming-aware positioning,and real-world testbed deployment.
基金supported in part by the National Key R&D Program of China under Grant 2024YFE0200700in part by the National Natural Science Foundation of China under Grant 62201504.
文摘Network architectures assisted by Generative Artificial Intelligence(GAI)are envisioned as foundational elements of sixth-generation(6G)communication system.To deliver ubiquitous intelligent services and meet diverse service requirements,6G network architecture should offer personalized services to various mobile devices.Federated learning(FL)with personalized local training,as a privacypreserving machine learning(ML)approach,can be applied to address these challenges.In this paper,we propose a meta-learning-based personalized FL(PFL)method that improves both communication and computation efficiency by utilizing over-the-air computations.Its“pretraining-and-fine-tuning”principle makes it particularly suitable for enabling edge nodes to access personalized GAI services while preserving local privacy.Experiment results demonstrate the outperformance and efficacy of the proposed algorithm,and notably indicate enhanced communication efficiency without compromising accuracy.
文摘5G/5G-Advanced在持续提升关键性能指标方面被寄予厚望,需要在时延、可靠性、连接数密度与用户体验等方面实现进一步突破。传统以人工操作为主的管理模式在效率、准确性与成本等方面的局限日益凸显。相较于传统优化方法,人工智能技术凭借其预测性与前瞻性,推动网络管理由被动应对转向主动感知与自优化,实现从“监测-响应”到“预判-编排”的迁移。基于3GPP在无线电接入网(radio access network,RAN)智能化方向的关键技术与标准化路径,结合典型用例场景,分析了AI/ML模型管理、数据采集与交互机制。面向6G智能RAN,进一步提出“意图驱动的协作任务”这一新型架构理念,其关键是通过RAN对应用层信息的感知、任务级别的服务质量(quality of service,QoS)监控、动态组和资源管理等技术实现6G网络人机及碳硅生态系统的无缝交互。
基金supported by National Natural Science Foundation of China under Granted No.62202247。
文摘Although 6G networks combined with artificial intelligence present revolutionary prospects for healthcare delivery,resource management in dense medical device networks stays a basic issue.Reliable communication directly affects patient outcomes in these settings;nonetheless,current resource allocation techniques struggle with complicated interference patterns and different service needs of AI-native healthcare systems.In dense installations where conventional approaches fail,this paper tackles the challenge of combining network efficiency with medical care priority.Thus,we offer a Dueling Deep Q-Network(DDQN)-based resource allocation approach for AI-native healthcare systems in 6G dense networks.First,we create a point-line graph coloringbased interference model to capture the unique characteristics of medical device communications.Building on this foundation,we suggest a DDQN approach to optimal resource allocation over multiple medical services by combining advantage estimate with healthcare-aware state evaluation.Unlike traditional graph-based models,this one correctly depicts the overlapping coverage areas common in hospital environments.Building on this basis,our DDQN design allows the system to prioritize medical needs while distributing resources by separating healthcare state assessment from advantage estimation.Experimental findings show that the suggested DDQN outperforms state-of-the-art techniques in dense healthcare installations by 14.6%greater network throughput and 13.7%better resource use.The solution shows particularly strong in maintaining service quality under vital conditions with 5.5%greater Qo S satisfaction for emergency services and 8.2%quicker recovery from interruptions.
文摘随着6G的快速发展,天地一体化信息网络成为实现全球无缝覆盖的关键。围绕6G天地一体化网络中的调制技术展开专利分析,重点研究了基于循环前缀(Cyclic Prefix,CP)/离散傅里叶变换(Discrete Fourier Transform,DFT)扩展的正交频分复用(Orthogonal Frequency Division Multiplexing,OFDM)、基于滤波器组/多带滤波的OFDM、正交啁啾分复用(Orthogonal Chirp Division Multiplexing,OCDM)、正交时频空间(Orthogonal Time Frequency and Space,OTFS)和仿射频分复用(Affine Frequency Division Multiplexing,AFDM)这5种调制技术的专利申请趋势、技术热点及竞争格局。研究结果表明,2020年后6G调制技术专利申请量激增,中国在专利数量上占据主导地位,AFDM因其在高动态信道中的优异性能成为未来6G标准的有力候选。揭示了专利领域的技术空白,为后续研发和专利布局提供了战略参考。
基金The authors extend their appreciation to the Deanship of Scientific Research at King Khalid University for funding this work under Grant Number(RGP 1/279/42).www.kku.edu.sa.
文摘The sixth-generation(6G)wireless communication networks are anticipated in integrating aerial,terrestrial,and maritime communication into a robust system to accomplish trustworthy,quick,and low latency needs.It enables to achieve maximum throughput and delay for several applications.Besides,the evolution of 6G leads to the design of unmanned aerial vehicles(UAVs)in providing inexpensive and effective solutions in various application areas such as healthcare,environment monitoring,and so on.In the UAV network,effective data collection with restricted energy capacity poses a major issue to achieving high quality network communication.It can be addressed by the use of clustering techniques forUAVs in 6G networks.In this aspect,this study develops a novel metaheuristic based energy efficient data gathering scheme for clustered unmanned aerial vehicles(MEEDG-CUAV).The proposed MEEDG-CUAV technique intends in partitioning the UAV networks into various clusters and assign a cluster head(CH)to reduce the overall energy utilization.Besides,the quantum chaotic butterfly optimization algorithm(QCBOA)with a fitness function is derived to choose CHs and construct clusters.The experimental validation of the MEEDG-CUAV technique occurs utilizing benchmark dataset and the experimental results highlighted the better performance over the other state of art techniques interms of different measures.
基金This study is supported by Fujitsu-Waseda Digital Annealer FWDA Research Project and Fujitsu Co-Creation Research Laboratory at Waseda University(Joint Research between Waseda University and Fujitsu Lab).The study was also partly supported by the School of Fundamental Science and Engineering,Faculty of Science and Engineering,Waseda University,Japan.This study was supported by the Institute for Information&Communications Technology Planning&Evaluation(IITP)Grant funded by the Korean government(MSIT)(No.2019-0-01343,Training Key Talents in Industrial Convergence Security)and Research Cluster Project,R20143,by the Zayed University Research Office.
文摘The Internet of Things(IoT)is the fourth technological revolution in the global information industry after computers,the Internet,and mobile communication networks.It combines radio-frequency identification devices,infrared sensors,global positioning systems,and various other technologies.Information sensing equipment is connected via the Internet,thus forming a vast network.When these physical devices are connected to the Internet,the user terminal can be extended and expanded to exchange information,communicate with anything,and carry out identification,positioning,tracking,monitoring,and triggering of corresponding events on each device in the network.In real life,the IoT has a wide range of applications,covering many fields,such as smart homes,smart logistics,fine agriculture and animal husbandry,national defense,and military.One of the most significant factors in wireless channels is interference,which degrades the system performance.Although the existing QR decomposition-based signal detection method is an emerging topic because of its low complexity,it does not solve the problem of poor detection performance.Therefore,this study proposes a maximumlikelihood-based QR decomposition algorithm.The main idea is to estimate the initial level of detection using the maximum likelihood principle,and then the other layer is detected using a reliable decision.The optimal candidate is selected from the feedback by deploying the candidate points in an unreliable scenario.Simulation results show that the proposed algorithm effectively reduces the interference and propagation error compared with the algorithms reported in the literature.
基金supported by the Motorola Mobility,the National Council for Scientific and Technological Development(No.433142/2018-9)Research Productivity Fellowship(No.312831/2020-0)the Pernambuco Research Foundation(FACEPE)。
文摘5G networks apply adaptive modulation and coding according to the channel condition reported by the user in order to keep the mobile communication quality.However,the delay incurred by the feedback may make the channel quality indicator(CQI)obsolete.This paper addresses this issue by proposing two approaches,one based on machine learning and another on evolutionary computing,which considers the user context and signal-to-interference-plus-noise ratio(SINR)besides the delay length to estimate the updated SINR to be mapped into a CQI value.Our proposals are designed to run at the user equipment(UE)side,neither requiring any change in the signalling between the base station(gNB)and UE nor overloading the gNB.They are evaluated in terms of mean squared error by adopting 5G network simulation data and the results show their high accuracy and feasibility to be employed in 5G/6G systems.