Aiming at the optimum path excluding characteristics and the full constellation searching characteristics of the K-best detection algorithm, an improved-performance K-best detection algorithm and several reduced-compl...Aiming at the optimum path excluding characteristics and the full constellation searching characteristics of the K-best detection algorithm, an improved-performance K-best detection algorithm and several reduced-complexity K-best detection algorithms are proposed. The improved-performance K-best detection algorithm deploys minimum mean square error (MMSE) filtering of a channel matrix before QR decomposition. This algorithm can decrease the probability of excluding the optimum path and achieve better performance. The reducedcomplexity K-best detection algorithms utilize a sphere decoding method to reduce searching constellation points. Simulation results show that the improved performance K-best detection algorithm obtains a 1 dB performance gain compared to the K- best detection algorithm based on sorted QR decomposition (SQRD). Performance loss occurs when K = 4 in reduced complexity K-best detection algorithms. When K = 8, the reduced complexity K-best detection algorithms require less computational effort compared with traditional K-best detection algorithms and achieve the same performance.展开更多
Existing feature selection methods for intrusion detection systems in the Industrial Internet of Things often suffer from local optimality and high computational complexity.These challenges hinder traditional IDS from...Existing feature selection methods for intrusion detection systems in the Industrial Internet of Things often suffer from local optimality and high computational complexity.These challenges hinder traditional IDS from effectively extracting features while maintaining detection accuracy.This paper proposes an industrial Internet ofThings intrusion detection feature selection algorithm based on an improved whale optimization algorithm(GSLDWOA).The aim is to address the problems that feature selection algorithms under high-dimensional data are prone to,such as local optimality,long detection time,and reduced accuracy.First,the initial population’s diversity is increased using the Gaussian Mutation mechanism.Then,Non-linear Shrinking Factor balances global exploration and local development,avoiding premature convergence.Lastly,Variable-step Levy Flight operator and Dynamic Differential Evolution strategy are introduced to improve the algorithm’s search efficiency and convergence accuracy in highdimensional feature space.Experiments on the NSL-KDD and WUSTL-IIoT-2021 datasets demonstrate that the feature subset selected by GSLDWOA significantly improves detection performance.Compared to the traditional WOA algorithm,the detection rate and F1-score increased by 3.68%and 4.12%.On the WUSTL-IIoT-2021 dataset,accuracy,recall,and F1-score all exceed 99.9%.展开更多
Aiming at the scale adaptation of automatic driving target detection algorithms in low illumination environments and the shortcomings in target occlusion processing,this paper proposes a YOLO-LKSDS automatic driving d...Aiming at the scale adaptation of automatic driving target detection algorithms in low illumination environments and the shortcomings in target occlusion processing,this paper proposes a YOLO-LKSDS automatic driving detection model.Firstly,the Contrast-Limited Adaptive Histogram Equalisation(CLAHE)image enhancement algorithm is improved to increase the image contrast and enhance the detailed features of the target;then,on the basis of the YOLOv5 model,the Kmeans++clustering algorithm is introduced to obtain a suitable anchor frame,and SPPELAN spatial pyramid pooling is improved to enhance the accuracy and robustness of the model for multi-scale target detection.Finally,an improved SEAM(Separated and Enhancement Attention Module)attention mechanism is combined with the DIOU-NMS algorithm to optimize the model’s performance when dealing with occlusion and dense scenes.Compared with the original model,the improved YOLO-LKSDS model achieves a 13.3%improvement in accuracy,a 1.7%improvement in mAP,and 240,000 fewer parameters on the BDD100K dataset.In order to validate the generalization of the improved algorithm,we selected the KITTI dataset for experimentation,which shows that YOLOv5’s accuracy improves by 21.1%,recall by 36.6%,and mAP50 by 29.5%,respectively,on the KITTI dataset.The deployment of this paper’s algorithm is verified by an edge computing platform,where the average speed of detection reaches 24.4 FPS while power consumption remains below 9 W,demonstrating high real-time capability and energy efficiency.展开更多
Community detection is one of the most fundamental applications in understanding the structure of complicated networks.Furthermore,it is an important approach to identifying closely linked clusters of nodes that may r...Community detection is one of the most fundamental applications in understanding the structure of complicated networks.Furthermore,it is an important approach to identifying closely linked clusters of nodes that may represent underlying patterns and relationships.Networking structures are highly sensitive in social networks,requiring advanced techniques to accurately identify the structure of these communities.Most conventional algorithms for detecting communities perform inadequately with complicated networks.In addition,they miss out on accurately identifying clusters.Since single-objective optimization cannot always generate accurate and comprehensive results,as multi-objective optimization can.Therefore,we utilized two objective functions that enable strong connections between communities and weak connections between them.In this study,we utilized the intra function,which has proven effective in state-of-the-art research studies.We proposed a new inter-function that has demonstrated its effectiveness by making the objective of detecting external connections between communities is to make them more distinct and sparse.Furthermore,we proposed a Multi-Objective community strength enhancement algorithm(MOCSE).The proposed algorithm is based on the framework of the Multi-Objective Evolutionary Algorithm with Decomposition(MOEA/D),integrated with a new heuristic mutation strategy,community strength enhancement(CSE).The results demonstrate that the model is effective in accurately identifying community structures while also being computationally efficient.The performance measures used to evaluate the MOEA/D algorithm in our work are normalized mutual information(NMI)and modularity(Q).It was tested using five state-of-the-art algorithms on social networks,comprising real datasets(Zachary,Dolphin,Football,Krebs,SFI,Jazz,and Netscience),as well as twenty synthetic datasets.These results provide the robustness and practical value of the proposed algorithm in multi-objective community identification.展开更多
Previous studies have shown that deep learning is very effective in detecting known attacks.However,when facing unknown attacks,models such as Deep Neural Networks(DNN)combined with Long Short-Term Memory(LSTM),Convol...Previous studies have shown that deep learning is very effective in detecting known attacks.However,when facing unknown attacks,models such as Deep Neural Networks(DNN)combined with Long Short-Term Memory(LSTM),Convolutional Neural Networks(CNN)combined with LSTM,and so on are built by simple stacking,which has the problems of feature loss,low efficiency,and low accuracy.Therefore,this paper proposes an autonomous detectionmodel for Distributed Denial of Service attacks,Multi-Scale Convolutional Neural Network-Bidirectional Gated Recurrent Units-Single Headed Attention(MSCNN-BiGRU-SHA),which is based on a Multistrategy Integrated Zebra Optimization Algorithm(MI-ZOA).The model undergoes training and testing with the CICDDoS2019 dataset,and its performance is evaluated on a new GINKS2023 dataset.The hyperparameters for Conv_filter and GRU_unit are optimized using the Multi-strategy Integrated Zebra Optimization Algorithm(MIZOA).The experimental results show that the test accuracy of the MSCNN-BiGRU-SHA model based on the MIZOA proposed in this paper is as high as 0.9971 in the CICDDoS 2019 dataset.The evaluation accuracy of the new dataset GINKS2023 created in this paper is 0.9386.Compared to the MSCNN-BiGRU-SHA model based on the Zebra Optimization Algorithm(ZOA),the detection accuracy on the GINKS2023 dataset has improved by 5.81%,precisionhas increasedby 1.35%,the recallhas improvedby 9%,and theF1scorehas increasedby 5.55%.Compared to the MSCNN-BiGRU-SHA models developed using Grid Search,Random Search,and Bayesian Optimization,the MSCNN-BiGRU-SHA model optimized with the MI-ZOA exhibits better performance in terms of accuracy,precision,recall,and F1 score.展开更多
Nowadays,abnormal traffic detection for Software-Defined Networking(SDN)faces the challenges of large data volume and high dimensionality.Since traditional machine learning-based detection methods have the problem of ...Nowadays,abnormal traffic detection for Software-Defined Networking(SDN)faces the challenges of large data volume and high dimensionality.Since traditional machine learning-based detection methods have the problem of data redundancy,the Metaheuristic Algorithm(MA)is introduced to select features beforemachine learning to reduce the dimensionality of data.Since a Tyrannosaurus Optimization Algorithm(TROA)has the advantages of few parameters,simple implementation,and fast convergence,and it shows better results in feature selection,TROA can be applied to abnormal traffic detection for SDN.However,TROA suffers frominsufficient global search capability,is easily trapped in local optimums,and has poor search accuracy.Then,this paper tries to improve TROA,namely the Improved Tyrannosaurus Optimization Algorithm(ITROA).It proposes a metaheuristic-driven abnormal traffic detection model for SDN based on ITROA.Finally,the validity of the ITROA is verified by the benchmark function and the UCI dataset,and the feature selection optimization operation is performed on the InSDN dataset by ITROA and other MAs to obtain the optimized feature subset for SDN abnormal traffic detection.The experiment shows that the performance of the proposed ITROA outperforms compared MAs in terms of the metaheuristic-driven model for SDN,achieving an accuracy of 99.37%on binary classification and 96.73%on multiclassification.展开更多
Face recognition has emerged as one of the most prominent applications of image analysis and under-standing,gaining considerable attention in recent years.This growing interest is driven by two key factors:its extensi...Face recognition has emerged as one of the most prominent applications of image analysis and under-standing,gaining considerable attention in recent years.This growing interest is driven by two key factors:its extensive applications in law enforcement and the commercial domain,and the rapid advancement of practical technologies.Despite the significant advancements,modern recognition algorithms still struggle in real-world conditions such as varying lighting conditions,occlusion,and diverse facial postures.In such scenarios,human perception is still well above the capabilities of present technology.Using the systematic mapping study,this paper presents an in-depth review of face detection algorithms and face recognition algorithms,presenting a detailed survey of advancements made between 2015 and 2024.We analyze key methodologies,highlighting their strengths and restrictions in the application context.Additionally,we examine various datasets used for face detection/recognition datasets focusing on the task-specific applications,size,diversity,and complexity.By analyzing these algorithms and datasets,this survey works as a valuable resource for researchers,identifying the research gap in the field of face detection and recognition and outlining potential directions for future research.展开更多
Atrial Fibrillation(AF)is a cardiac disorder characterized by irregular heart rhythms,typically diagnosed using Electrocardiogram(ECG)signals.In remote regions with limited healthcare personnel,automated AF detection ...Atrial Fibrillation(AF)is a cardiac disorder characterized by irregular heart rhythms,typically diagnosed using Electrocardiogram(ECG)signals.In remote regions with limited healthcare personnel,automated AF detection is extremely important.Although recent studies have explored various machine learning and deep learning approaches,challenges such as signal noise and subtle variations between AF and other cardiac rhythms continue to hinder accurate classification.In this study,we propose a novel framework that integrates robust preprocessing,comprehensive feature extraction,and an ensemble classification strategy.In the first step,ECG signals are divided into equal-sized segments using a 5-s sliding window with 50%overlap,followed by bandpass filtering between 0.5 and 45 Hz for noise removal.After preprocessing,both time and frequency-domain features are extracted,and a custom one-dimensional Convolutional Neural Network—Bidirectional Long Short-Term Memory(1D CNN-BiLSTM)architecture is introduced.Handcrafted and automated features are concatenated into a unified feature vector and classified using Support Vector Machine(SVM),Random Forest(RF),and Long Short-Term Memory(LSTM)models.A Quantum Genetic Algorithm(QGA)optimizes weighted averages of the classifier outputs for multi-class classification,distinguishing among AF,noisy,normal,and other rhythms.Evaluated on the PhysioNet 2017 Cardiology Challenge dataset,the proposed method achieved an accuracy of 94.40%and an F1-score of 92.30%,outperforming several state-of-the-art techniques.展开更多
In smart driving for rail transit,a reliable obstacle detection system is an important guarantee for the safety of trains.Therein,the detection of the rail area directly affects the accuracy of the system to identify ...In smart driving for rail transit,a reliable obstacle detection system is an important guarantee for the safety of trains.Therein,the detection of the rail area directly affects the accuracy of the system to identify dangerous targets.Both the rail line and the lane are presented as thin line shapes in the image,but the rail scene is more complex,and the color of the rail line is more difficult to distinguish from the background.By comparison,there are already many deep learning-based lane detection algorithms,but there is a lack of public datasets and targeted deep learning detection algorithms for rail line detection.To address this,this paper constructs a rail image dataset RailwayLine and labels the rail line for the training and testing of models.This dataset contains rich rail images including single-rail,multi-rail,straight rail,curved rail,crossing rails,occlusion,blur,and different lighting conditions.To address the problem of the lack of deep learning-based rail line detection algorithms,we improve the CLRNet algorithm which has an excellent performance in lane detection,and propose the CLRNet-R algorithm for rail line detection.To address the problem of the rail line being thin and occupying fewer pixels in the image,making it difficult to distinguish from complex backgrounds,we introduce an attention mechanism to enhance global feature extraction ability and add a semantic segmentation head to enhance the features of the rail region by the binary probability of rail lines.To address the poor curve recognition performance and unsmooth output lines in the original CLRNet algorithm,we improve the weight allocation for line intersection-over-union calculation in the original framework and propose two loss functions based on local slopes to optimize the model’s local sampling point training constraints,improving the model’s fitting performance on curved rails and obtaining smooth and stable rail line detection results.Through experiments,this paper demonstrates that compared with other mainstream lane detection algorithms,the algorithm proposed in this paper has a better performance for rail line detection.展开更多
The excessive use of artificial intelligence(AI)algorithms has caused the problem of errors in AI algorithms,which has challenged the fairness of decision-making,and has intensified people’s inequality.Therefore,it i...The excessive use of artificial intelligence(AI)algorithms has caused the problem of errors in AI algorithms,which has challenged the fairness of decision-making,and has intensified people’s inequality.Therefore,it is necessary to conduct in-depth research and propose corresponding error detection and error elimination methods.This paper first proposes the root causes and threats of bias in AI algorithms,then summarizes the existing bias detection and error elimination methods,and proposes a bias processing framework in three-level dimensions of data,models,and conclusions,aiming to provide a framework for a comprehensive solution to errors in algorithms.At the same time,it also summarizes the problems and challenges in existing research and makes a prospect for future research trends.It is hoped that it will be helpful for us to build fairer AI.展开更多
Ant colony optimization(ACO)is a random search algorithm based on probability calculation.However,the uninformed search strategy has a slow convergence speed.The Bayesian algorithm uses the historical information of t...Ant colony optimization(ACO)is a random search algorithm based on probability calculation.However,the uninformed search strategy has a slow convergence speed.The Bayesian algorithm uses the historical information of the searched point to determine the next search point during the search process,reducing the uncertainty in the random search process.Due to the ability of the Bayesian algorithm to reduce uncertainty,a Bayesian ACO algorithm is proposed in this paper to increase the convergence speed of the conventional ACO algorithm for image edge detection.In addition,this paper has the following two innovations on the basis of the classical algorithm,one of which is to add random perturbations after completing the pheromone update.The second is the use of adaptive pheromone heuristics.Experimental results illustrate that the proposed Bayesian ACO algorithm has faster convergence and higher precision and recall than the traditional ant colony algorithm,due to the improvement of the pheromone utilization rate.Moreover,Bayesian ACO algorithm outperforms the other comparative methods in edge detection task.展开更多
Aimed at the long and narrow geometric features and poor generalization ability of the damage detection in conveyor belts with steel rope cores using the X-ray image,a detection method of damage X-ray image is propose...Aimed at the long and narrow geometric features and poor generalization ability of the damage detection in conveyor belts with steel rope cores using the X-ray image,a detection method of damage X-ray image is proposed based on the improved fully convolutional one-stage object detection(FCOS)algorithm.The regression performance of bounding boxes was optimized by introducing the complete intersection over union loss function into the improved algorithm.The feature fusion network structure is modified by adding adaptive fusion paths to the feature fusion network structure,which makes full use of the features of accurate localization and semantics of multi-scale feature fusion networks.Finally,the network structure was trained and validated by using the X-ray image dataset of damages in conveyor belts with steel rope cores provided by a flaw detection equipment manufacturer.In addition,the data enhancement methods such as rotating,mirroring,and scaling,were employed to enrich the image dataset so that the model is adequately trained.Experimental results showed that the improved FCOS algorithm promoted the precision rate and the recall rate by 20.9%and 14.8%respectively,compared with the original algorithm.Meanwhile,compared with Fast R-CNN,Faster R-CNN,SSD,and YOLOv3,the improved FCOS algorithm has obvious advantages;detection precision rate and recall rate of the modified network reached 95.8%and 97.0%respectively.Furthermore,it demonstrated a higher detection accuracy without affecting the speed.The results of this work have some reference significance for the automatic identification and detection of steel core conveyor belt damage.展开更多
Domain Generation Algorithms(DGAs)continue to pose a significant threat inmodernmalware infrastructures by enabling resilient and evasive communication with Command and Control(C&C)servers.Traditional detection me...Domain Generation Algorithms(DGAs)continue to pose a significant threat inmodernmalware infrastructures by enabling resilient and evasive communication with Command and Control(C&C)servers.Traditional detection methods-rooted in statistical heuristics,feature engineering,and shallow machine learning-struggle to adapt to the increasing sophistication,linguistic mimicry,and adversarial variability of DGA variants.The emergence of Large Language Models(LLMs)marks a transformative shift in this landscape.Leveraging deep contextual understanding,semantic generalization,and few-shot learning capabilities,LLMs such as BERT,GPT,and T5 have shown promising results in detecting both character-based and dictionary-based DGAs,including previously unseen(zeroday)variants.This paper provides a comprehensive and critical review of LLM-driven DGA detection,introducing a structured taxonomy of LLM architectures,evaluating the linguistic and behavioral properties of benchmark datasets,and comparing recent detection frameworks across accuracy,latency,robustness,and multilingual performance.We also highlight key limitations,including challenges in adversarial resilience,model interpretability,deployment scalability,and privacy risks.To address these gaps,we present a forward-looking research roadmap encompassing adversarial training,model compression,cross-lingual benchmarking,and real-time integration with SIEM/SOAR platforms.This survey aims to serve as a foundational resource for advancing the development of scalable,explainable,and operationally viable LLM-based DGA detection systems.展开更多
The complexity of cloud environments challenges secure resource management,especially for intrusion detection systems(IDS).Existing strategies struggle to balance efficiency,cost fairness,and threat resilience.This pa...The complexity of cloud environments challenges secure resource management,especially for intrusion detection systems(IDS).Existing strategies struggle to balance efficiency,cost fairness,and threat resilience.This paper proposes an innovative approach to managing cloud resources through the integration of a genetic algorithm(GA)with a“double auction”method.This approach seeks to enhance security and efficiency by aligning buyers and sellers within an intelligent market framework.It guarantees equitable pricing while utilizing resources efficiently and optimizing advantages for all stakeholders.The GA functions as an intelligent search mechanism that identifies optimal combinations of bids from users and suppliers,addressing issues arising from the intricacies of cloud systems.Analyses proved that our method surpasses previous strategies,particularly in terms of price accuracy,speed,and the capacity to manage large-scale activities,critical factors for real-time cybersecurity systems,such as IDS.Our research integrates artificial intelligence-inspired evolutionary algorithms with market-driven methods to develop intelligent resource management systems that are secure,scalable,and adaptable to evolving risks,such as process innovation.展开更多
With the birth of Software-Defined Networking(SDN),integration of both SDN and traditional architectures becomes the development trend of computer networks.Network intrusion detection faces challenges in dealing with ...With the birth of Software-Defined Networking(SDN),integration of both SDN and traditional architectures becomes the development trend of computer networks.Network intrusion detection faces challenges in dealing with complex attacks in SDN environments,thus to address the network security issues from the viewpoint of Artificial Intelligence(AI),this paper introduces the Crayfish Optimization Algorithm(COA)to the field of intrusion detection for both SDN and traditional network architectures,and based on the characteristics of the original COA,an Improved Crayfish Optimization Algorithm(ICOA)is proposed by integrating strategies of elite reverse learning,Levy flight,crowding factor and parameter modification.The ICOA is then utilized for AI-integrated feature selection of intrusion detection for both SDN and traditional network architectures,to reduce the dimensionality of the data and improve the performance of network intrusion detection.Finally,the performance evaluation is performed by testing not only the NSL-KDD dataset and the UNSW-NB 15 dataset for traditional networks but also the InSDN dataset for SDN-based networks.Experimental results show that ICOA improves the accuracy by 0.532%and 2.928%respectively compared with GWO and COA in traditional networks.In SDN networks,the accuracy of ICOA is 0.25%and 0.3%higher than COA and PSO.These findings collectively indicate that AI-integrated feature selection based on the proposed ICOA can promote network intrusion detection for both SDN and traditional architectures.展开更多
Multiple Input Multiple Output (MIMO) technology is of great significance in high data rate wireless communication. The K-Best Sphere Decoding (K-Best SD) algorithm was proposed as a powerful method for MIMO detection...Multiple Input Multiple Output (MIMO) technology is of great significance in high data rate wireless communication. The K-Best Sphere Decoding (K-Best SD) algorithm was proposed as a powerful method for MIMO detection that can approach near-optimal performance. However, some extra computational complexity is contained in K-Best SD. In this paper, we propose an improved K-Best SD to reduce the complexity of conventional K-Best SD by assigning K for each level dynamically following some rules. Simulation proves that the performance degradation of the improved K-Best SD is very little and the complexity is significantly reduced.展开更多
In order to solve the problem that the global navigation satellite system(GNSS) receivers can hardly detect the GNSS spoofing when they are deceived by a spoofer,a model-based approach for the identification of the ...In order to solve the problem that the global navigation satellite system(GNSS) receivers can hardly detect the GNSS spoofing when they are deceived by a spoofer,a model-based approach for the identification of the GNSS spoofing is proposed.First,a Hammerstein model is applied to model the spoofer/GNSS transmitter and the wireless channel.Then,a novel method based on the uncultivated wolf pack algorithm(UWPA) is proposed to estimate the model parameters.Taking the estimated model parameters as a feature vector,the identification of the spoofing is realized by comparing the Euclidean distance between the feature vectors.Simulations verify the effectiveness and the robustness of the proposed method.The results show that,compared with the other identification algorithms,such as least square(LS),the iterative method and the bat-inspired algorithm(BA),although the UWPA has a little more time-eomplexity than the LS and the BA algorithm,it has better estimation precision of the model parameters and higher identification rate of the GNSS spoofing,even for relative low signal-to-noise ratios.展开更多
With the development of global position system(GPS),wireless technology and location aware services,it is possible to collect a large quantity of trajectory data.In the field of data mining for moving objects,the pr...With the development of global position system(GPS),wireless technology and location aware services,it is possible to collect a large quantity of trajectory data.In the field of data mining for moving objects,the problem of anomaly detection is a hot topic.Based on the development of anomalous trajectory detection of moving objects,this paper introduces the classical trajectory outlier detection(TRAOD) algorithm,and then proposes a density-based trajectory outlier detection(DBTOD) algorithm,which compensates the disadvantages of the TRAOD algorithm that it is unable to detect anomalous defects when the trajectory is local and dense.The results of employing the proposed algorithm to Elk1993 and Deer1995 datasets are also presented,which show the effectiveness of the algorithm.展开更多
To aim at the multimode character of the data from the airplane detecting system, the paper combines Dempster- Shafer evidence theory and subjective Bayesian algorithm and makes to propose a mixed structure multimode ...To aim at the multimode character of the data from the airplane detecting system, the paper combines Dempster- Shafer evidence theory and subjective Bayesian algorithm and makes to propose a mixed structure multimode data fusion algorithm. The algorithm adopts a prorated algorithm relate to the incertitude evaluation to convert the probability evaluation into the precognition probability in an identity frame, and ensures the adaptability of different data from different source to the mixed system. To guarantee real time fusion, a combination of time domain fusion and space domain fusion is established, this not only assure the fusion of data chain in different time of the same sensor, but also the data fusion from different sensors distributed in different platforms and the data fusion among different modes. The feasibility and practicability are approved through computer simulation.展开更多
Genetic algorithm(GA) has received significant attention for the design and implementation of intrusion detection systems. In this paper, it is proposed to use variable length chromosomes(VLCs) in a GA-based network i...Genetic algorithm(GA) has received significant attention for the design and implementation of intrusion detection systems. In this paper, it is proposed to use variable length chromosomes(VLCs) in a GA-based network intrusion detection system.Fewer chromosomes with relevant features are used for rule generation. An effective fitness function is used to define the fitness of each rule. Each chromosome will have one or more rules in it. As each chromosome is a complete solution to the problem, fewer chromosomes are sufficient for effective intrusion detection. This reduces the computational time. The proposed approach is tested using Defense Advanced Research Project Agency(DARPA) 1998 data. The experimental results show that the proposed approach is efficient in network intrusion detection.展开更多
基金The National High Technology Research and Develop-ment Program of China (863Program)(No.2006AA01Z264)the National Natural Science Foundation of China (No.60572072)
文摘Aiming at the optimum path excluding characteristics and the full constellation searching characteristics of the K-best detection algorithm, an improved-performance K-best detection algorithm and several reduced-complexity K-best detection algorithms are proposed. The improved-performance K-best detection algorithm deploys minimum mean square error (MMSE) filtering of a channel matrix before QR decomposition. This algorithm can decrease the probability of excluding the optimum path and achieve better performance. The reducedcomplexity K-best detection algorithms utilize a sphere decoding method to reduce searching constellation points. Simulation results show that the improved performance K-best detection algorithm obtains a 1 dB performance gain compared to the K- best detection algorithm based on sorted QR decomposition (SQRD). Performance loss occurs when K = 4 in reduced complexity K-best detection algorithms. When K = 8, the reduced complexity K-best detection algorithms require less computational effort compared with traditional K-best detection algorithms and achieve the same performance.
基金supported by the Major Science and Technology Programs in Henan Province(No.241100210100)Henan Provincial Science and Technology Research Project(No.252102211085,No.252102211105)+3 种基金Endogenous Security Cloud Network Convergence R&D Center(No.602431011PQ1)The Special Project for Research and Development in Key Areas of Guangdong Province(No.2021ZDZX1098)The Stabilization Support Program of Science,Technology and Innovation Commission of Shenzhen Municipality(No.20231128083944001)The Key scientific research projects of Henan higher education institutions(No.24A520042).
文摘Existing feature selection methods for intrusion detection systems in the Industrial Internet of Things often suffer from local optimality and high computational complexity.These challenges hinder traditional IDS from effectively extracting features while maintaining detection accuracy.This paper proposes an industrial Internet ofThings intrusion detection feature selection algorithm based on an improved whale optimization algorithm(GSLDWOA).The aim is to address the problems that feature selection algorithms under high-dimensional data are prone to,such as local optimality,long detection time,and reduced accuracy.First,the initial population’s diversity is increased using the Gaussian Mutation mechanism.Then,Non-linear Shrinking Factor balances global exploration and local development,avoiding premature convergence.Lastly,Variable-step Levy Flight operator and Dynamic Differential Evolution strategy are introduced to improve the algorithm’s search efficiency and convergence accuracy in highdimensional feature space.Experiments on the NSL-KDD and WUSTL-IIoT-2021 datasets demonstrate that the feature subset selected by GSLDWOA significantly improves detection performance.Compared to the traditional WOA algorithm,the detection rate and F1-score increased by 3.68%and 4.12%.On the WUSTL-IIoT-2021 dataset,accuracy,recall,and F1-score all exceed 99.9%.
基金supported by the Key R&D Program of Shaanxi Province(No.2025CYYBXM-078).
文摘Aiming at the scale adaptation of automatic driving target detection algorithms in low illumination environments and the shortcomings in target occlusion processing,this paper proposes a YOLO-LKSDS automatic driving detection model.Firstly,the Contrast-Limited Adaptive Histogram Equalisation(CLAHE)image enhancement algorithm is improved to increase the image contrast and enhance the detailed features of the target;then,on the basis of the YOLOv5 model,the Kmeans++clustering algorithm is introduced to obtain a suitable anchor frame,and SPPELAN spatial pyramid pooling is improved to enhance the accuracy and robustness of the model for multi-scale target detection.Finally,an improved SEAM(Separated and Enhancement Attention Module)attention mechanism is combined with the DIOU-NMS algorithm to optimize the model’s performance when dealing with occlusion and dense scenes.Compared with the original model,the improved YOLO-LKSDS model achieves a 13.3%improvement in accuracy,a 1.7%improvement in mAP,and 240,000 fewer parameters on the BDD100K dataset.In order to validate the generalization of the improved algorithm,we selected the KITTI dataset for experimentation,which shows that YOLOv5’s accuracy improves by 21.1%,recall by 36.6%,and mAP50 by 29.5%,respectively,on the KITTI dataset.The deployment of this paper’s algorithm is verified by an edge computing platform,where the average speed of detection reaches 24.4 FPS while power consumption remains below 9 W,demonstrating high real-time capability and energy efficiency.
文摘Community detection is one of the most fundamental applications in understanding the structure of complicated networks.Furthermore,it is an important approach to identifying closely linked clusters of nodes that may represent underlying patterns and relationships.Networking structures are highly sensitive in social networks,requiring advanced techniques to accurately identify the structure of these communities.Most conventional algorithms for detecting communities perform inadequately with complicated networks.In addition,they miss out on accurately identifying clusters.Since single-objective optimization cannot always generate accurate and comprehensive results,as multi-objective optimization can.Therefore,we utilized two objective functions that enable strong connections between communities and weak connections between them.In this study,we utilized the intra function,which has proven effective in state-of-the-art research studies.We proposed a new inter-function that has demonstrated its effectiveness by making the objective of detecting external connections between communities is to make them more distinct and sparse.Furthermore,we proposed a Multi-Objective community strength enhancement algorithm(MOCSE).The proposed algorithm is based on the framework of the Multi-Objective Evolutionary Algorithm with Decomposition(MOEA/D),integrated with a new heuristic mutation strategy,community strength enhancement(CSE).The results demonstrate that the model is effective in accurately identifying community structures while also being computationally efficient.The performance measures used to evaluate the MOEA/D algorithm in our work are normalized mutual information(NMI)and modularity(Q).It was tested using five state-of-the-art algorithms on social networks,comprising real datasets(Zachary,Dolphin,Football,Krebs,SFI,Jazz,and Netscience),as well as twenty synthetic datasets.These results provide the robustness and practical value of the proposed algorithm in multi-objective community identification.
基金supported by Science and Technology Innovation Programfor Postgraduate Students in IDP Subsidized by Fundamental Research Funds for the Central Universities(Project No.ZY20240335)support of the Research Project of the Key Technology of Malicious Code Detection Based on Data Mining in APT Attack(Project No.2022IT173)the Research Project of the Big Data Sensitive Information Supervision Technology Based on Convolutional Neural Network(Project No.2022011033).
文摘Previous studies have shown that deep learning is very effective in detecting known attacks.However,when facing unknown attacks,models such as Deep Neural Networks(DNN)combined with Long Short-Term Memory(LSTM),Convolutional Neural Networks(CNN)combined with LSTM,and so on are built by simple stacking,which has the problems of feature loss,low efficiency,and low accuracy.Therefore,this paper proposes an autonomous detectionmodel for Distributed Denial of Service attacks,Multi-Scale Convolutional Neural Network-Bidirectional Gated Recurrent Units-Single Headed Attention(MSCNN-BiGRU-SHA),which is based on a Multistrategy Integrated Zebra Optimization Algorithm(MI-ZOA).The model undergoes training and testing with the CICDDoS2019 dataset,and its performance is evaluated on a new GINKS2023 dataset.The hyperparameters for Conv_filter and GRU_unit are optimized using the Multi-strategy Integrated Zebra Optimization Algorithm(MIZOA).The experimental results show that the test accuracy of the MSCNN-BiGRU-SHA model based on the MIZOA proposed in this paper is as high as 0.9971 in the CICDDoS 2019 dataset.The evaluation accuracy of the new dataset GINKS2023 created in this paper is 0.9386.Compared to the MSCNN-BiGRU-SHA model based on the Zebra Optimization Algorithm(ZOA),the detection accuracy on the GINKS2023 dataset has improved by 5.81%,precisionhas increasedby 1.35%,the recallhas improvedby 9%,and theF1scorehas increasedby 5.55%.Compared to the MSCNN-BiGRU-SHA models developed using Grid Search,Random Search,and Bayesian Optimization,the MSCNN-BiGRU-SHA model optimized with the MI-ZOA exhibits better performance in terms of accuracy,precision,recall,and F1 score.
基金supported by the National Natural Science Foundation of China under Grant 61602162the Hubei Provincial Science and Technology Plan Project under Grant 2023BCB041.
文摘Nowadays,abnormal traffic detection for Software-Defined Networking(SDN)faces the challenges of large data volume and high dimensionality.Since traditional machine learning-based detection methods have the problem of data redundancy,the Metaheuristic Algorithm(MA)is introduced to select features beforemachine learning to reduce the dimensionality of data.Since a Tyrannosaurus Optimization Algorithm(TROA)has the advantages of few parameters,simple implementation,and fast convergence,and it shows better results in feature selection,TROA can be applied to abnormal traffic detection for SDN.However,TROA suffers frominsufficient global search capability,is easily trapped in local optimums,and has poor search accuracy.Then,this paper tries to improve TROA,namely the Improved Tyrannosaurus Optimization Algorithm(ITROA).It proposes a metaheuristic-driven abnormal traffic detection model for SDN based on ITROA.Finally,the validity of the ITROA is verified by the benchmark function and the UCI dataset,and the feature selection optimization operation is performed on the InSDN dataset by ITROA and other MAs to obtain the optimized feature subset for SDN abnormal traffic detection.The experiment shows that the performance of the proposed ITROA outperforms compared MAs in terms of the metaheuristic-driven model for SDN,achieving an accuracy of 99.37%on binary classification and 96.73%on multiclassification.
文摘Face recognition has emerged as one of the most prominent applications of image analysis and under-standing,gaining considerable attention in recent years.This growing interest is driven by two key factors:its extensive applications in law enforcement and the commercial domain,and the rapid advancement of practical technologies.Despite the significant advancements,modern recognition algorithms still struggle in real-world conditions such as varying lighting conditions,occlusion,and diverse facial postures.In such scenarios,human perception is still well above the capabilities of present technology.Using the systematic mapping study,this paper presents an in-depth review of face detection algorithms and face recognition algorithms,presenting a detailed survey of advancements made between 2015 and 2024.We analyze key methodologies,highlighting their strengths and restrictions in the application context.Additionally,we examine various datasets used for face detection/recognition datasets focusing on the task-specific applications,size,diversity,and complexity.By analyzing these algorithms and datasets,this survey works as a valuable resource for researchers,identifying the research gap in the field of face detection and recognition and outlining potential directions for future research.
基金supported and funded by the Deanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University(IMSIU)(grant number IMSIU-DDRSP2501)。
文摘Atrial Fibrillation(AF)is a cardiac disorder characterized by irregular heart rhythms,typically diagnosed using Electrocardiogram(ECG)signals.In remote regions with limited healthcare personnel,automated AF detection is extremely important.Although recent studies have explored various machine learning and deep learning approaches,challenges such as signal noise and subtle variations between AF and other cardiac rhythms continue to hinder accurate classification.In this study,we propose a novel framework that integrates robust preprocessing,comprehensive feature extraction,and an ensemble classification strategy.In the first step,ECG signals are divided into equal-sized segments using a 5-s sliding window with 50%overlap,followed by bandpass filtering between 0.5 and 45 Hz for noise removal.After preprocessing,both time and frequency-domain features are extracted,and a custom one-dimensional Convolutional Neural Network—Bidirectional Long Short-Term Memory(1D CNN-BiLSTM)architecture is introduced.Handcrafted and automated features are concatenated into a unified feature vector and classified using Support Vector Machine(SVM),Random Forest(RF),and Long Short-Term Memory(LSTM)models.A Quantum Genetic Algorithm(QGA)optimizes weighted averages of the classifier outputs for multi-class classification,distinguishing among AF,noisy,normal,and other rhythms.Evaluated on the PhysioNet 2017 Cardiology Challenge dataset,the proposed method achieved an accuracy of 94.40%and an F1-score of 92.30%,outperforming several state-of-the-art techniques.
基金the Sichuan Science and Technology Program(No.2022YFS0557)the National Natural Science Foundation of China(No.61972271)。
文摘In smart driving for rail transit,a reliable obstacle detection system is an important guarantee for the safety of trains.Therein,the detection of the rail area directly affects the accuracy of the system to identify dangerous targets.Both the rail line and the lane are presented as thin line shapes in the image,but the rail scene is more complex,and the color of the rail line is more difficult to distinguish from the background.By comparison,there are already many deep learning-based lane detection algorithms,but there is a lack of public datasets and targeted deep learning detection algorithms for rail line detection.To address this,this paper constructs a rail image dataset RailwayLine and labels the rail line for the training and testing of models.This dataset contains rich rail images including single-rail,multi-rail,straight rail,curved rail,crossing rails,occlusion,blur,and different lighting conditions.To address the problem of the lack of deep learning-based rail line detection algorithms,we improve the CLRNet algorithm which has an excellent performance in lane detection,and propose the CLRNet-R algorithm for rail line detection.To address the problem of the rail line being thin and occupying fewer pixels in the image,making it difficult to distinguish from complex backgrounds,we introduce an attention mechanism to enhance global feature extraction ability and add a semantic segmentation head to enhance the features of the rail region by the binary probability of rail lines.To address the poor curve recognition performance and unsmooth output lines in the original CLRNet algorithm,we improve the weight allocation for line intersection-over-union calculation in the original framework and propose two loss functions based on local slopes to optimize the model’s local sampling point training constraints,improving the model’s fitting performance on curved rails and obtaining smooth and stable rail line detection results.Through experiments,this paper demonstrates that compared with other mainstream lane detection algorithms,the algorithm proposed in this paper has a better performance for rail line detection.
文摘The excessive use of artificial intelligence(AI)algorithms has caused the problem of errors in AI algorithms,which has challenged the fairness of decision-making,and has intensified people’s inequality.Therefore,it is necessary to conduct in-depth research and propose corresponding error detection and error elimination methods.This paper first proposes the root causes and threats of bias in AI algorithms,then summarizes the existing bias detection and error elimination methods,and proposes a bias processing framework in three-level dimensions of data,models,and conclusions,aiming to provide a framework for a comprehensive solution to errors in algorithms.At the same time,it also summarizes the problems and challenges in existing research and makes a prospect for future research trends.It is hoped that it will be helpful for us to build fairer AI.
基金supported by the National Natural Science Foundation of China(62276055).
文摘Ant colony optimization(ACO)is a random search algorithm based on probability calculation.However,the uninformed search strategy has a slow convergence speed.The Bayesian algorithm uses the historical information of the searched point to determine the next search point during the search process,reducing the uncertainty in the random search process.Due to the ability of the Bayesian algorithm to reduce uncertainty,a Bayesian ACO algorithm is proposed in this paper to increase the convergence speed of the conventional ACO algorithm for image edge detection.In addition,this paper has the following two innovations on the basis of the classical algorithm,one of which is to add random perturbations after completing the pheromone update.The second is the use of adaptive pheromone heuristics.Experimental results illustrate that the proposed Bayesian ACO algorithm has faster convergence and higher precision and recall than the traditional ant colony algorithm,due to the improvement of the pheromone utilization rate.Moreover,Bayesian ACO algorithm outperforms the other comparative methods in edge detection task.
文摘Aimed at the long and narrow geometric features and poor generalization ability of the damage detection in conveyor belts with steel rope cores using the X-ray image,a detection method of damage X-ray image is proposed based on the improved fully convolutional one-stage object detection(FCOS)algorithm.The regression performance of bounding boxes was optimized by introducing the complete intersection over union loss function into the improved algorithm.The feature fusion network structure is modified by adding adaptive fusion paths to the feature fusion network structure,which makes full use of the features of accurate localization and semantics of multi-scale feature fusion networks.Finally,the network structure was trained and validated by using the X-ray image dataset of damages in conveyor belts with steel rope cores provided by a flaw detection equipment manufacturer.In addition,the data enhancement methods such as rotating,mirroring,and scaling,were employed to enrich the image dataset so that the model is adequately trained.Experimental results showed that the improved FCOS algorithm promoted the precision rate and the recall rate by 20.9%and 14.8%respectively,compared with the original algorithm.Meanwhile,compared with Fast R-CNN,Faster R-CNN,SSD,and YOLOv3,the improved FCOS algorithm has obvious advantages;detection precision rate and recall rate of the modified network reached 95.8%and 97.0%respectively.Furthermore,it demonstrated a higher detection accuracy without affecting the speed.The results of this work have some reference significance for the automatic identification and detection of steel core conveyor belt damage.
基金the Deanship of Scientific Research at King Khalid University for funding this work through large group under grant number(GRP.2/663/46).
文摘Domain Generation Algorithms(DGAs)continue to pose a significant threat inmodernmalware infrastructures by enabling resilient and evasive communication with Command and Control(C&C)servers.Traditional detection methods-rooted in statistical heuristics,feature engineering,and shallow machine learning-struggle to adapt to the increasing sophistication,linguistic mimicry,and adversarial variability of DGA variants.The emergence of Large Language Models(LLMs)marks a transformative shift in this landscape.Leveraging deep contextual understanding,semantic generalization,and few-shot learning capabilities,LLMs such as BERT,GPT,and T5 have shown promising results in detecting both character-based and dictionary-based DGAs,including previously unseen(zeroday)variants.This paper provides a comprehensive and critical review of LLM-driven DGA detection,introducing a structured taxonomy of LLM architectures,evaluating the linguistic and behavioral properties of benchmark datasets,and comparing recent detection frameworks across accuracy,latency,robustness,and multilingual performance.We also highlight key limitations,including challenges in adversarial resilience,model interpretability,deployment scalability,and privacy risks.To address these gaps,we present a forward-looking research roadmap encompassing adversarial training,model compression,cross-lingual benchmarking,and real-time integration with SIEM/SOAR platforms.This survey aims to serve as a foundational resource for advancing the development of scalable,explainable,and operationally viable LLM-based DGA detection systems.
文摘The complexity of cloud environments challenges secure resource management,especially for intrusion detection systems(IDS).Existing strategies struggle to balance efficiency,cost fairness,and threat resilience.This paper proposes an innovative approach to managing cloud resources through the integration of a genetic algorithm(GA)with a“double auction”method.This approach seeks to enhance security and efficiency by aligning buyers and sellers within an intelligent market framework.It guarantees equitable pricing while utilizing resources efficiently and optimizing advantages for all stakeholders.The GA functions as an intelligent search mechanism that identifies optimal combinations of bids from users and suppliers,addressing issues arising from the intricacies of cloud systems.Analyses proved that our method surpasses previous strategies,particularly in terms of price accuracy,speed,and the capacity to manage large-scale activities,critical factors for real-time cybersecurity systems,such as IDS.Our research integrates artificial intelligence-inspired evolutionary algorithms with market-driven methods to develop intelligent resource management systems that are secure,scalable,and adaptable to evolving risks,such as process innovation.
基金supported by the National Natural Science Foundation of China under Grant 61602162the Hubei Provincial Science and Technology Plan Project under Grant 2023BCB041.
文摘With the birth of Software-Defined Networking(SDN),integration of both SDN and traditional architectures becomes the development trend of computer networks.Network intrusion detection faces challenges in dealing with complex attacks in SDN environments,thus to address the network security issues from the viewpoint of Artificial Intelligence(AI),this paper introduces the Crayfish Optimization Algorithm(COA)to the field of intrusion detection for both SDN and traditional network architectures,and based on the characteristics of the original COA,an Improved Crayfish Optimization Algorithm(ICOA)is proposed by integrating strategies of elite reverse learning,Levy flight,crowding factor and parameter modification.The ICOA is then utilized for AI-integrated feature selection of intrusion detection for both SDN and traditional network architectures,to reduce the dimensionality of the data and improve the performance of network intrusion detection.Finally,the performance evaluation is performed by testing not only the NSL-KDD dataset and the UNSW-NB 15 dataset for traditional networks but also the InSDN dataset for SDN-based networks.Experimental results show that ICOA improves the accuracy by 0.532%and 2.928%respectively compared with GWO and COA in traditional networks.In SDN networks,the accuracy of ICOA is 0.25%and 0.3%higher than COA and PSO.These findings collectively indicate that AI-integrated feature selection based on the proposed ICOA can promote network intrusion detection for both SDN and traditional architectures.
文摘Multiple Input Multiple Output (MIMO) technology is of great significance in high data rate wireless communication. The K-Best Sphere Decoding (K-Best SD) algorithm was proposed as a powerful method for MIMO detection that can approach near-optimal performance. However, some extra computational complexity is contained in K-Best SD. In this paper, we propose an improved K-Best SD to reduce the complexity of conventional K-Best SD by assigning K for each level dynamically following some rules. Simulation proves that the performance degradation of the improved K-Best SD is very little and the complexity is significantly reduced.
基金The National Natural Science Foundation of China(No.61271214,61471152)the Postdoctoral Science Foundation of Jiangsu Province(No.1402023C)the Natural Science Foundation of Zhejiang Province(No.LZ14F010003)
文摘In order to solve the problem that the global navigation satellite system(GNSS) receivers can hardly detect the GNSS spoofing when they are deceived by a spoofer,a model-based approach for the identification of the GNSS spoofing is proposed.First,a Hammerstein model is applied to model the spoofer/GNSS transmitter and the wireless channel.Then,a novel method based on the uncultivated wolf pack algorithm(UWPA) is proposed to estimate the model parameters.Taking the estimated model parameters as a feature vector,the identification of the spoofing is realized by comparing the Euclidean distance between the feature vectors.Simulations verify the effectiveness and the robustness of the proposed method.The results show that,compared with the other identification algorithms,such as least square(LS),the iterative method and the bat-inspired algorithm(BA),although the UWPA has a little more time-eomplexity than the LS and the BA algorithm,it has better estimation precision of the model parameters and higher identification rate of the GNSS spoofing,even for relative low signal-to-noise ratios.
基金supported by the Aeronautical Science Foundation of China(20111052010)the Jiangsu Graduates Innovation Project (CXZZ120163)+1 种基金the "333" Project of Jiangsu Provincethe Qing Lan Project of Jiangsu Province
文摘With the development of global position system(GPS),wireless technology and location aware services,it is possible to collect a large quantity of trajectory data.In the field of data mining for moving objects,the problem of anomaly detection is a hot topic.Based on the development of anomalous trajectory detection of moving objects,this paper introduces the classical trajectory outlier detection(TRAOD) algorithm,and then proposes a density-based trajectory outlier detection(DBTOD) algorithm,which compensates the disadvantages of the TRAOD algorithm that it is unable to detect anomalous defects when the trajectory is local and dense.The results of employing the proposed algorithm to Elk1993 and Deer1995 datasets are also presented,which show the effectiveness of the algorithm.
文摘To aim at the multimode character of the data from the airplane detecting system, the paper combines Dempster- Shafer evidence theory and subjective Bayesian algorithm and makes to propose a mixed structure multimode data fusion algorithm. The algorithm adopts a prorated algorithm relate to the incertitude evaluation to convert the probability evaluation into the precognition probability in an identity frame, and ensures the adaptability of different data from different source to the mixed system. To guarantee real time fusion, a combination of time domain fusion and space domain fusion is established, this not only assure the fusion of data chain in different time of the same sensor, but also the data fusion from different sensors distributed in different platforms and the data fusion among different modes. The feasibility and practicability are approved through computer simulation.
文摘Genetic algorithm(GA) has received significant attention for the design and implementation of intrusion detection systems. In this paper, it is proposed to use variable length chromosomes(VLCs) in a GA-based network intrusion detection system.Fewer chromosomes with relevant features are used for rule generation. An effective fitness function is used to define the fitness of each rule. Each chromosome will have one or more rules in it. As each chromosome is a complete solution to the problem, fewer chromosomes are sufficient for effective intrusion detection. This reduces the computational time. The proposed approach is tested using Defense Advanced Research Project Agency(DARPA) 1998 data. The experimental results show that the proposed approach is efficient in network intrusion detection.