期刊文献+
共找到21,971篇文章
< 1 2 250 >
每页显示 20 50 100
A Comparative Benchmark of Machine and Deep Learning for Cyberattack Detection in IoT Networks
1
作者 Enzo Hoummady Fehmi Jaafar 《Computers, Materials & Continua》 2026年第4期1070-1092,共23页
With the proliferation of Internet of Things(IoT)devices,securing these interconnected systems against cyberattacks has become a critical challenge.Traditional security paradigms often fail to cope with the scale and ... With the proliferation of Internet of Things(IoT)devices,securing these interconnected systems against cyberattacks has become a critical challenge.Traditional security paradigms often fail to cope with the scale and diversity of IoT network traffic.This paper presents a comparative benchmark of classic machine learning(ML)and state-of-the-art deep learning(DL)algorithms for IoT intrusion detection.Our methodology employs a twophased approach:a preliminary pilot study using a custom-generated dataset to establish baselines,followed by a comprehensive evaluation on the large-scale CICIoTDataset2023.We benchmarked algorithms including Random Forest,XGBoost,CNN,and StackedLSTM.The results indicate that while top-performingmodels frombothcategories achieve over 99%classification accuracy,this metric masks a crucial performance trade-off.We demonstrate that treebased ML ensembles exhibit superior precision(91%)in identifying benign traffic,making them effective at reducing false positives.Conversely,DL models demonstrate superior recall(96%),making them better suited for minimizing the interruption of legitimate traffic.We conclude that the selection of an optimal model is not merely a matter of maximizing accuracy but is a strategic choice dependent on the specific security priority either minimizing false alarms or ensuring service availability.Thiswork provides a practical framework for deploying context-aware security solutions in diverse IoT environments. 展开更多
关键词 Internet of Things deep learning abnormal network traffic cyberattacks machine learning
在线阅读 下载PDF
改进Deep Q Networks的交通信号均衡调度算法
2
作者 贺道坤 《机械设计与制造》 北大核心 2025年第4期135-140,共6页
为进一步缓解城市道路高峰时段十字路口的交通拥堵现象,实现路口各道路车流均衡通过,基于改进Deep Q Networks提出了一种的交通信号均衡调度算法。提取十字路口与交通信号调度最相关的特征,分别建立单向十字路口交通信号模型和线性双向... 为进一步缓解城市道路高峰时段十字路口的交通拥堵现象,实现路口各道路车流均衡通过,基于改进Deep Q Networks提出了一种的交通信号均衡调度算法。提取十字路口与交通信号调度最相关的特征,分别建立单向十字路口交通信号模型和线性双向十字路口交通信号模型,并基于此构建交通信号调度优化模型;针对Deep Q Networks算法在交通信号调度问题应用中所存在的收敛性、过估计等不足,对Deep Q Networks进行竞争网络改进、双网络改进以及梯度更新策略改进,提出相适应的均衡调度算法。通过与经典Deep Q Networks仿真比对,验证论文算法对交通信号调度问题的适用性和优越性。基于城市道路数据,分别针对两种场景进行仿真计算,仿真结果表明该算法能够有效缩减十字路口车辆排队长度,均衡各路口车流通行量,缓解高峰出行方向的道路拥堵现象,有利于十字路口交通信号调度效益的提升。 展开更多
关键词 交通信号调度 十字路口 deep Q networks 深度强化学习 智能交通
在线阅读 下载PDF
DEEP NEURAL NETWORKS COMBINING MULTI-TASK LEARNING FOR SOLVING DELAY INTEGRO-DIFFERENTIAL EQUATIONS 被引量:1
3
作者 WANG Chen-yao SHI Feng 《数学杂志》 2025年第1期13-38,共26页
Deep neural networks(DNNs)are effective in solving both forward and inverse problems for nonlinear partial differential equations(PDEs).However,conventional DNNs are not effective in handling problems such as delay di... Deep neural networks(DNNs)are effective in solving both forward and inverse problems for nonlinear partial differential equations(PDEs).However,conventional DNNs are not effective in handling problems such as delay differential equations(DDEs)and delay integrodifferential equations(DIDEs)with constant delays,primarily due to their low regularity at delayinduced breaking points.In this paper,a DNN method that combines multi-task learning(MTL)which is proposed to solve both the forward and inverse problems of DIDEs.The core idea of this approach is to divide the original equation into multiple tasks based on the delay,using auxiliary outputs to represent the integral terms,followed by the use of MTL to seamlessly incorporate the properties at the breaking points into the loss function.Furthermore,given the increased training dificulty associated with multiple tasks and outputs,we employ a sequential training scheme to reduce training complexity and provide reference solutions for subsequent tasks.This approach significantly enhances the approximation accuracy of solving DIDEs with DNNs,as demonstrated by comparisons with traditional DNN methods.We validate the effectiveness of this method through several numerical experiments,test various parameter sharing structures in MTL and compare the testing results of these structures.Finally,this method is implemented to solve the inverse problem of nonlinear DIDE and the results show that the unknown parameters of DIDE can be discovered with sparse or noisy data. 展开更多
关键词 Delay integro-differential equation Multi-task learning parameter sharing structure deep neural network sequential training scheme
在线阅读 下载PDF
Dynamic Resource Allocation for Multi-Priority Requests Based on Deep Reinforcement Learning in Elastic Optical Network
4
作者 Zhou Yang Yang Xin +1 位作者 Sun Qiang Yang Zhuojia 《China Communications》 2026年第2期312-327,共16页
As the types of traffic requests increase,the elastic optical network(EON)is considered as a promising architecture to carry multiple types of traffic requests simultaneously,including immediate reservation(IR)and adv... As the types of traffic requests increase,the elastic optical network(EON)is considered as a promising architecture to carry multiple types of traffic requests simultaneously,including immediate reservation(IR)and advance reservation(AR).Various resource allocation schemes for IR/AR requests have been designed in EON to reduce bandwidth blocking probability(BBP).However,these schemes do not consider different transmission requirements of IR requests and cannot maintain a low BBP for high-priority requests.In this paper,multi-priority is considered in the hybrid IR/AR request scenario.We modify the asynchronous advantage actor critic(A3C)model and propose an A3C-assisted priority resource allocation(APRA)algorithm.The APRA integrates priority and transmission quality of IR requests to design the A3C reward function,then dynamically allocates dedicated resources for different IR requests according to the time-varying requirements.By maximizing the reward,the transmission quality of IR requests can be matched with the priority,and lower BBP for high-priority IR requests can be ensured.Simulation results show that the APRA reduces the BBP of high-priority IR requests from 0.0341 to0.0138,and the overall network operation gain is improved by 883 compared to the scheme without considering the priority. 展开更多
关键词 deep reinforcement learning dynamic resource allocation elastic optical network multipriority requests
在线阅读 下载PDF
Advancing Breast Cancer Molecular Subtyping:A Comparative Study of Convolutional Neural Networks and Vision Transformers on Mammograms
5
作者 Chee Chin Lim Hui Wen Tiu +2 位作者 Qi Wei Oung Chiew Chea Lau Xiao Jian Tan 《Computers, Materials & Continua》 2026年第3期1287-1308,共22页
critical for guiding treatment and improving patient outcomes.Traditional molecular subtyping via immuno-histochemistry(IHC)test is invasive,time-consuming,and may not fully represent tumor heterogeneity.This study pr... critical for guiding treatment and improving patient outcomes.Traditional molecular subtyping via immuno-histochemistry(IHC)test is invasive,time-consuming,and may not fully represent tumor heterogeneity.This study proposes a non-invasive approach using digital mammography images and deep learning algorithm for classifying breast cancer molecular subtypes.Four pretrained models,including two Convolutional Neural Networks(MobileNet_V3_Large and VGG-16)and two Vision Transformers(ViT_B_16 and ViT_Base_Patch16_Clip_224)were fine-tuned to classify images into HER2-enriched,Luminal,Normal-like,and Triple Negative subtypes.Hyperparameter tuning,including learning rate adjustment and layer freezing strategies,was applied to optimize performance.Among the evaluated models,ViT_Base_Patch16_Clip_224 achieved the highest test accuracy(94.44%),with equally high precision,recall,and F1-score of 0.94,demonstrating excellent generalization.MobileNet_V3_Large achieved the same accuracy but showed less training stability.In contrast,VGG-16 recorded the lowest performance,indicating a limitation in its generalizability for this classification task.The study also highlighted the superior performance of the Vision Transformer models over CNNs,particularly due to their ability to capture global contextual features and the benefit of CLIP-based pretraining in ViT_Base_Patch16_Clip_224.To enhance clinical applicability,a graphical user interface(GUI)named“BCMS Dx”was developed for streamlined subtype prediction.Deep learning applied to mammography has proven effective for accurate and non-invasive molecular subtyping.The proposed Vision Transformer-based model and supporting GUI offer a promising direction for augmenting diagnostic workflows,minimizing the need for invasive procedures,and advancing personalized breast cancer management. 展开更多
关键词 Artificial intelligence breast cancer classification convolutional neural network deep learning hyperparameter tuning MAMMOGRAPHY medical imaging molecular subtypes vision transformer
在线阅读 下载PDF
Deep neural network based on adversarial training for short-term high-resolution precipitation nowcasting from radar echo images
6
作者 Ruikai YANG Shuangjian JIAO Nan YANG 《Journal of Oceanology and Limnology》 2026年第1期85-98,共14页
Precipitation nowcasting is of great importance for disaster prevention and mitigation.However,precipitation is a complex spatio-temporal phenomenon influenced by various underlying physical factors.Even slight change... Precipitation nowcasting is of great importance for disaster prevention and mitigation.However,precipitation is a complex spatio-temporal phenomenon influenced by various underlying physical factors.Even slight changes in the initial precipitation field can have a significant impact on the future precipitation patterns,making the nowcasting of short-term high-resolution precipitation a major challenge.Traditional deep learning methods often have difficulty capturing the long-term spatial dependence of precipitation and are usually at a low resolution.To address these issues,based upon the Simpler yet Better Video Prediction(SimVP)framework,we proposed a deep generative neural network that incorporates the Simple Parameter-Free Attention Module(SimAM)and Generative Adversarial Networks(GANs)for short-term high-resolution precipitation event forecasting.Through an adversarial training strategy,critical precipitation features were extracted from complex radar echo images.During the adversarial learning process,the dynamic competition between the generator and the discriminator could continuously enhance the model in prediction accuracy and resolution for short-term precipitation.Experimental results demonstrate that the proposed method could effectively forecast short-term precipitation events on various scales and showed the best overall performance among existing methods. 展开更多
关键词 precipitation nowcasting deep learning Simple Parameter-Free Attention Module(SimAM) Generative Adversarial networks(GANs)
在线阅读 下载PDF
A Convolutional Neural Network-Based Deep Support Vector Machine for Parkinson’s Disease Detection with Small-Scale and Imbalanced Datasets
7
作者 Kwok Tai Chui Varsha Arya +2 位作者 Brij B.Gupta Miguel Torres-Ruiz Razaz Waheeb Attar 《Computers, Materials & Continua》 2026年第1期1410-1432,共23页
Parkinson’s disease(PD)is a debilitating neurological disorder affecting over 10 million people worldwide.PD classification models using voice signals as input are common in the literature.It is believed that using d... Parkinson’s disease(PD)is a debilitating neurological disorder affecting over 10 million people worldwide.PD classification models using voice signals as input are common in the literature.It is believed that using deep learning algorithms further enhances performance;nevertheless,it is challenging due to the nature of small-scale and imbalanced PD datasets.This paper proposed a convolutional neural network-based deep support vector machine(CNN-DSVM)to automate the feature extraction process using CNN and extend the conventional SVM to a DSVM for better classification performance in small-scale PD datasets.A customized kernel function reduces the impact of biased classification towards the majority class(healthy candidates in our consideration).An improved generative adversarial network(IGAN)was designed to generate additional training data to enhance the model’s performance.For performance evaluation,the proposed algorithm achieves a sensitivity of 97.6%and a specificity of 97.3%.The performance comparison is evaluated from five perspectives,including comparisons with different data generation algorithms,feature extraction techniques,kernel functions,and existing works.Results reveal the effectiveness of the IGAN algorithm,which improves the sensitivity and specificity by 4.05%–4.72%and 4.96%–5.86%,respectively;and the effectiveness of the CNN-DSVM algorithm,which improves the sensitivity by 1.24%–57.4%and specificity by 1.04%–163%and reduces biased detection towards the majority class.The ablation experiments confirm the effectiveness of individual components.Two future research directions have also been suggested. 展开更多
关键词 Convolutional neural network data generation deep support vector machine feature extraction generative artificial intelligence imbalanced dataset medical diagnosis Parkinson’s disease small-scale dataset
在线阅读 下载PDF
A Hybrid Framework Combining Rule-Based and Deep Learning Approaches for Data-Driven Verdict Recommendations
8
作者 Muhammad Hameed Siddiqi Menwa Alshammeri +6 位作者 Jawad Khan Muhammad Faheem Khan Asfandyar Khan Madallah Alruwaili Yousef Alhwaiti Saad Alanazi Irshad Ahmad 《Computers, Materials & Continua》 2025年第6期5345-5371,共27页
As legal cases grow in complexity and volume worldwide,integrating machine learning and artificial intelligence into judicial systems has become a pivotal research focus.This study introduces a comprehensive framework... As legal cases grow in complexity and volume worldwide,integrating machine learning and artificial intelligence into judicial systems has become a pivotal research focus.This study introduces a comprehensive framework for verdict recommendation that synergizes rule-based methods with deep learning techniques specifically tailored to the legal domain.The proposed framework comprises three core modules:legal feature extraction,semantic similarity assessment,and verdict recommendation.For legal feature extraction,a rule-based approach leverages Black’s Law Dictionary and WordNet Synsets to construct feature vectors from judicial texts.Semantic similarity between cases is evaluated using a hybrid method that combines rule-based logic with an LSTM model,analyzing the feature vectors of query cases against a legal knowledge base.Verdicts are then recommended through a rule-based retrieval system,enhanced by predefined legal statutes and regulations.By merging rule-based methodologies with deep learning,this framework addresses the interpretability challenges often associated with contemporary AImodels,thereby enhancing both transparency and generalizability across diverse legal contexts.The system was rigorously tested using a legal corpus of 43,000 case laws across six categories:Criminal,Revenue,Service,Corporate,Constitutional,and Civil law,ensuring its adaptability across a wide range of judicial scenarios.Performance evaluation showed that the feature extraction module achieved an average accuracy of 91.6%with an F-Score of 95%.The semantic similarity module,tested using Manhattan,Euclidean,and Cosine distance metrics,achieved 88%accuracy and a 93%F-Score for short queries(Manhattan),89%accuracy and a 93.7%F-Score for medium-length queries(Euclidean),and 87%accuracy with a 92.5%F-Score for longer queries(Cosine).The verdict recommendation module outperformed existing methods,achieving 90%accuracy and a 93.75%F-Score.This study highlights the potential of hybrid AI frameworks to improve judicial decision-making and streamline legal processes,offering a robust,interpretable,and adaptable solution for the evolving demands of modern legal systems. 展开更多
关键词 Verdict recommendation legal knowledge base judicial text case laws semantic similarity legal domain features rule-based deep learning
在线阅读 下载PDF
Demand Forecasting of a Microgrid-Powered Electric Vehicle Charging Station Enabled by Emerging Technologies and Deep Recurrent Neural Networks
9
作者 Sahbi Boubaker Adel Mellit +3 位作者 Nejib Ghazouani Walid Meskine Mohamed Benghanem Habib Kraiem 《Computer Modeling in Engineering & Sciences》 2025年第5期2237-2259,共23页
Electric vehicles(EVs)are gradually being deployed in the transportation sector.Although they have a high impact on reducing greenhouse gas emissions,their penetration is challenged by their random energy demand and d... Electric vehicles(EVs)are gradually being deployed in the transportation sector.Although they have a high impact on reducing greenhouse gas emissions,their penetration is challenged by their random energy demand and difficult scheduling of their optimal charging.To cope with these problems,this paper presents a novel approach for photovoltaic grid-connected microgrid EV charging station energy demand forecasting.The present study is part of a comprehensive framework involving emerging technologies such as drones and artificial intelligence designed to support the EVs’charging scheduling task.By using predictive algorithms for solar generation and load demand estimation,this approach aimed at ensuring dynamic and efficient energy flow between the solar energy source,the grid and the electric vehicles.The main contribution of this paper lies in developing an intelligent approach based on deep recurrent neural networks to forecast the energy demand using only its previous records.Therefore,various forecasters based on Long Short-term Memory,Gated Recurrent Unit,and their bi-directional and stacked variants were investigated using a real dataset collected from an EV charging station located at Trieste University(Italy).The developed forecasters have been evaluated and compared according to different metrics,including R,RMSE,MAE,and MAPE.We found that the obtained R values for both PV power generation and energy demand ranged between 97%and 98%.These study findings can be used for reliable and efficient decision-making on the management side of the optimal scheduling of the charging operations. 展开更多
关键词 MICROGRID electric vehicles charging station forecasting deep recurrent neural networks energy management system
在线阅读 下载PDF
Application of deep learning-based convolutional neural networks in gastrointestinal disease endoscopic examination
10
作者 Yang-Yang Wang Bin Liu Ji-Han Wang 《World Journal of Gastroenterology》 2025年第36期50-69,共20页
Gastrointestinal(GI)diseases,including gastric and colorectal cancers,signi-ficantly impact global health,necessitating accurate and efficient diagnostic me-thods.Endoscopic examination is the primary diagnostic tool;... Gastrointestinal(GI)diseases,including gastric and colorectal cancers,signi-ficantly impact global health,necessitating accurate and efficient diagnostic me-thods.Endoscopic examination is the primary diagnostic tool;however,its accu-racy is limited by operator dependency and interobserver variability.Advance-ments in deep learning,particularly convolutional neural networks(CNNs),show great potential for enhancing GI disease detection and classification.This review explores the application of CNNs in endoscopic imaging,focusing on polyp and tumor detection,disease classification,endoscopic ultrasound,and capsule endo-scopy analysis.We discuss the performance of CNN models with traditional dia-gnostic methods,highlighting their advantages in accuracy and real-time decision support.Despite promising results,challenges remain,including data availability,model interpretability,and clinical integration.Future directions include impro-ving model generalization,enhancing explainability,and conducting large-scale clinical trials.With continued advancements,CNN-powered artificial intelligence systems could revolutionize GI endoscopy by enhancing early disease detection,reducing diagnostic errors,and improving patient outcomes. 展开更多
关键词 Gastrointestinal diseases Endoscopic examination deep learning Convolutional neural networks Computer-aided diagnosis
在线阅读 下载PDF
A survey of backdoor attacks and defenses:From deep neural networks to large language models
11
作者 Ling-Xin Jin Wei Jiang +5 位作者 Xiang-Yu Wen Mei-Yu Lin Jin-Yu Zhan Xing-Zhi Zhou Maregu Assefa Habtie Naoufel Werghi 《Journal of Electronic Science and Technology》 2025年第3期13-35,共23页
Deep neural networks(DNNs)have found extensive applications in safety-critical artificial intelligence systems,such as autonomous driving and facial recognition systems.However,recent research has revealed their susce... Deep neural networks(DNNs)have found extensive applications in safety-critical artificial intelligence systems,such as autonomous driving and facial recognition systems.However,recent research has revealed their susceptibility to backdoors maliciously injected by adversaries.This vulnerability arises due to the intricate architecture and opacity of DNNs,resulting in numerous redundant neurons embedded within the models.Adversaries exploit these vulnerabilities to conceal malicious backdoor information within DNNs,thereby causing erroneous outputs and posing substantial threats to the efficacy of DNN-based applications.This article presents a comprehensive survey of backdoor attacks against DNNs and the countermeasure methods employed to mitigate them.Initially,we trace the evolution of the concept from traditional backdoor attacks to backdoor attacks against DNNs,highlighting the feasibility and practicality of generating backdoor attacks against DNNs.Subsequently,we provide an overview of notable works encompassing various attack and defense strategies,facilitating a comparative analysis of their approaches.Through these discussions,we offer constructive insights aimed at refining these techniques.Finally,we extend our research perspective to the domain of large language models(LLMs)and synthesize the characteristics and developmental trends of backdoor attacks and defense methods targeting LLMs.Through a systematic review of existing studies on backdoor vulnerabilities in LLMs,we identify critical open challenges in this field and propose actionable directions for future research. 展开更多
关键词 Backdoor Attacks Backdoor defenses deep neural networks Large language model
在线阅读 下载PDF
Improving Fundus Detection Precision in Diabetic Retinopathy Using Derivative-Based Deep Neural Networks
12
作者 Asma Aldrees Hong Min +2 位作者 Ashit Kumar Dutta Yousef Ibrahim Daradkeh Mohd Anjum 《Computer Modeling in Engineering & Sciences》 2025年第3期2487-2511,共25页
Fundoscopic diagnosis involves assessing the proper functioning of the eye’s nerves,blood vessels,retinal health,and the impact of diabetes on the optic nerves.Fundus disorders are a major global health concern,affec... Fundoscopic diagnosis involves assessing the proper functioning of the eye’s nerves,blood vessels,retinal health,and the impact of diabetes on the optic nerves.Fundus disorders are a major global health concern,affecting millions of people worldwide due to their widespread occurrence.Fundus photography generates machine-based eye images that assist in diagnosing and treating ocular diseases such as diabetic retinopathy.As a result,accurate fundus detection is essential for early diagnosis and effective treatment,helping to prevent severe complications and improve patient outcomes.To address this need,this article introduces a Derivative Model for Fundus Detection using Deep NeuralNetworks(DMFD-DNN)to enhance diagnostic precision.Thismethod selects key features for fundus detection using the least derivative,which identifies features correlating with stored fundus images.Feature filtering relies on the minimum derivative,determined by extracting both similar and varying textures.In this research,the DNN model was integrated with the derivative model.Fundus images were segmented,features were extracted,and the DNN was iteratively trained to identify fundus regions reliably.The goal was to improve the precision of fundoscopic diagnosis by training the DNN incrementally,taking into account the least possible derivative across iterations,and using outputs from previous cycles.The hidden layer of the neural network operates on the most significant derivative,which may reduce precision across iterations.These derivatives are treated as inaccurate,and the model is subsequently trained using selective features and their corresponding extractions.The proposed model outperforms previous techniques in detecting fundus regions,achieving 94.98%accuracy and 91.57%sensitivity,with a minimal error rate of 5.43%.It significantly reduces feature extraction time to 1.462 s and minimizes computational overhead,thereby improving operational efficiency and scalability.Ultimately,the proposed model enhances diagnostic precision and reduces errors,leading to more effective fundus dysfunction diagnosis and treatment. 展开更多
关键词 deep neural network feature extraction fundus detection medical image processing
在线阅读 下载PDF
Dynamic Clustering Method for Underwater Wireless Sensor Networks based on Deep Reinforcement Learning
13
作者 Kohyar Bolvary Zadeh Dashtestani Reza Javidan Reza Akbari 《哈尔滨工程大学学报(英文版)》 2025年第4期864-876,共13页
Underwater wireless sensor networks(UWSNs)have emerged as a new paradigm of real-time organized systems,which are utilized in a diverse array of scenarios to manage the underwater environment surrounding them.One of t... Underwater wireless sensor networks(UWSNs)have emerged as a new paradigm of real-time organized systems,which are utilized in a diverse array of scenarios to manage the underwater environment surrounding them.One of the major challenges that these systems confront is topology control via clustering,which reduces the overload of wireless communications within a network and ensures low energy consumption and good scalability.This study aimed to present a clustering technique in which the clustering process and cluster head(CH)selection are performed based on the Markov decision process and deep reinforcement learning(DRL).DRL algorithm selects the CH by maximizing the defined reward function.Subsequently,the sensed data are collected by the CHs and then sent to the autonomous underwater vehicles.In the final phase,the consumed energy by each sensor is calculated,and its residual energy is updated.Then,the autonomous underwater vehicle performs all clustering and CH selection operations.This procedure persists until the point of cessation when the sensor’s power has been reduced to such an extent that no node can become a CH.Through analysis of the findings from this investigation and their comparison with alternative frameworks,the implementation of this method can be used to control the cluster size and the number of CHs,which ultimately augments the energy usage of nodes and prolongs the lifespan of the network.Our simulation results illustrate that the suggested methodology surpasses the conventional low-energy adaptive clustering hierarchy,the distance-and energy-constrained K-means clustering scheme,and the vector-based forward protocol and is viable for deployment in an actual operational environment. 展开更多
关键词 Underwater wireless sensor network CLUSTERING Cluster head selection deep reinforcement learning
暂未订购
Big Texture Dataset Synthesized Based on Gradient and Convolution Kernels Using Pre-Trained Deep Neural Networks
14
作者 Farhan A.Alenizi Faten Khalid Karim +1 位作者 Alaa R.Al-Shamasneh Mohammad Hossein Shakoor 《Computer Modeling in Engineering & Sciences》 2025年第8期1793-1829,共37页
Deep neural networks provide accurate results for most applications.However,they need a big dataset to train properly.Providing a big dataset is a significant challenge in most applications.Image augmentation refers t... Deep neural networks provide accurate results for most applications.However,they need a big dataset to train properly.Providing a big dataset is a significant challenge in most applications.Image augmentation refers to techniques that increase the amount of image data.Common operations for image augmentation include changes in illumination,rotation,contrast,size,viewing angle,and others.Recently,Generative Adversarial Networks(GANs)have been employed for image generation.However,like image augmentation methods,GAN approaches can only generate images that are similar to the original images.Therefore,they also cannot generate new classes of data.Texture images presentmore challenges than general images,and generating textures is more complex than creating other types of images.This study proposes a gradient-based deep neural network method that generates a new class of texture.It is possible to rapidly generate new classes of textures using different kernels from pre-trained deep networks.After generating new textures for each class,the number of textures increases through image augmentation.During this process,several techniques are proposed to automatically remove incomplete and similar textures that are created.The proposed method is faster than some well-known generative networks by around 4 to 10 times.In addition,the quality of the generated textures surpasses that of these networks.The proposed method can generate textures that surpass those of someGANs and parametric models in certain image qualitymetrics.It can provide a big texture dataset to train deep networks.A new big texture dataset is created artificially using the proposed method.This dataset is approximately 2 GB in size and comprises 30,000 textures,each 150×150 pixels in size,organized into 600 classes.It is uploaded to the Kaggle site and Google Drive.This dataset is called BigTex.Compared to other texture datasets,the proposed dataset is the largest and can serve as a comprehensive texture dataset for training more powerful deep neural networks and mitigating overfitting. 展开更多
关键词 Big texture dataset data generation pre-trained deep neural network
在线阅读 下载PDF
Resource Allocation in V2X Networks:A Double Deep Q-Network Approach with Graph Neural Networks
15
作者 Zhengda Huan Jian Sun +3 位作者 Zeyu Chen Ziyi Zhang Xiao Sun Zenghui Xiao 《Computers, Materials & Continua》 2025年第9期5427-5443,共17页
With the advancement of Vehicle-to-Everything(V2X)technology,efficient resource allocation in dynamic vehicular networks has become a critical challenge for achieving optimal performance.Existing methods suffer from h... With the advancement of Vehicle-to-Everything(V2X)technology,efficient resource allocation in dynamic vehicular networks has become a critical challenge for achieving optimal performance.Existing methods suffer from high computational complexity and decision latency under high-density traffic and heterogeneous network conditions.To address these challenges,this study presents an innovative framework that combines Graph Neural Networks(GNNs)with a Double Deep Q-Network(DDQN),utilizing dynamic graph structures and reinforcement learning.An adaptive neighbor sampling mechanism is introduced to dynamically select the most relevant neighbors based on interference levels and network topology,thereby improving decision accuracy and efficiency.Meanwhile,the framework models communication links as nodes and interference relationships as edges,effectively capturing the direct impact of interference on resource allocation while reducing computational complexity and preserving critical interaction information.Employing an aggregation mechanism based on the Graph Attention Network(GAT),it dynamically adjusts the neighbor sampling scope and performs attention-weighted aggregation based on node importance,ensuring more efficient and adaptive resource management.This design ensures reliable Vehicle-to-Vehicle(V2V)communication while maintaining high Vehicle-to-Infrastructure(V2I)throughput.The framework retains the global feature learning capabilities of GNNs and supports distributed network deployment,allowing vehicles to extract low-dimensional graph embeddings from local observations for real-time resource decisions.Experimental results demonstrate that the proposed method significantly reduces computational overhead,mitigates latency,and improves resource utilization efficiency in vehicular networks under complex traffic scenarios.This research not only provides a novel solution to resource allocation challenges in V2X networks but also advances the application of DDQN in intelligent transportation systems,offering substantial theoretical significance and practical value. 展开更多
关键词 Resource allocation V2X double deep Q-network graph neural network
在线阅读 下载PDF
Deep Learning Model for Identifying Internal Flaws Based on Image Quadtree SBFEM and Deep Neural Networks
16
作者 Hanyu Tao Dongye Sun +1 位作者 Tao Fang Wenhu Zhao 《Computer Modeling in Engineering & Sciences》 2025年第10期521-536,共16页
Structural internal flaws often weaken the performance and integral stability,while traditional nondestructive testing or inversion methods face challenges of high cost and low efficiency in quantitative flaw identifi... Structural internal flaws often weaken the performance and integral stability,while traditional nondestructive testing or inversion methods face challenges of high cost and low efficiency in quantitative flaw identification.To quickly identify internal flaws within structures,a deep learning model for flaw detection is proposed based on the image quadtree scaled boundary finite element method(SBFEM)combined with a deep neural network(DNN).The training dataset is generated fromthe numerical simulations using the balanced quadtree algorithmand SBFEM,where the structural domain is discretized based on recursive decomposition principles andmesh refinement is automatically performed in the flaw boundary regions.The model contains only six types of elements and hanging nodes don’t affect the solution accuracy,resulting in a high degree of automation and significantly reducing the cost of the training dataset.The deep artificial neural network for flaw detection is constructed using DNN as the learning framework,effectively mitigating the risk of the objective function converging to local optima during training.Statistical methods are employed to evaluate the accuracy of the inversionmodel,and the influences of flaw size and the number of training samples on the performance are examined.In statistical results of single flaw,the 95%confidence intervals of the relative error for(x,y,r)are[2.16%,2.76%],[1.53%,1.96%]and[1.49%,1.91%],respectively.The 95%confidence interval of the comprehensive relative error for double flaws is[3.06%,3.62%].The results demonstrate that the predicted flaw parameters align closely with the reserved clean data,indicating that themodel can accurately quantify both the location and size of structural flaws. 展开更多
关键词 Flaw detection deep neural network image quadtree scaled boundary finite element method
在线阅读 下载PDF
Resource allocation for AI-native healthcare systems in 6G dense networks using deep reinforcement learning
17
作者 Jianhui Lv Chien-Ming Chen +1 位作者 Saru Kumari Keqin Li 《Digital Communications and Networks》 2025年第6期2016-2029,共14页
Although 6G networks combined with artificial intelligence present revolutionary prospects for healthcare delivery,resource management in dense medical device networks stays a basic issue.Reliable communication direct... Although 6G networks combined with artificial intelligence present revolutionary prospects for healthcare delivery,resource management in dense medical device networks stays a basic issue.Reliable communication directly affects patient outcomes in these settings;nonetheless,current resource allocation techniques struggle with complicated interference patterns and different service needs of AI-native healthcare systems.In dense installations where conventional approaches fail,this paper tackles the challenge of combining network efficiency with medical care priority.Thus,we offer a Dueling Deep Q-Network(DDQN)-based resource allocation approach for AI-native healthcare systems in 6G dense networks.First,we create a point-line graph coloringbased interference model to capture the unique characteristics of medical device communications.Building on this foundation,we suggest a DDQN approach to optimal resource allocation over multiple medical services by combining advantage estimate with healthcare-aware state evaluation.Unlike traditional graph-based models,this one correctly depicts the overlapping coverage areas common in hospital environments.Building on this basis,our DDQN design allows the system to prioritize medical needs while distributing resources by separating healthcare state assessment from advantage estimation.Experimental findings show that the suggested DDQN outperforms state-of-the-art techniques in dense healthcare installations by 14.6%greater network throughput and 13.7%better resource use.The solution shows particularly strong in maintaining service quality under vital conditions with 5.5%greater Qo S satisfaction for emergency services and 8.2%quicker recovery from interruptions. 展开更多
关键词 Resource allocation AI-native healthcare systems 6G dense networks deep reinforcement learning
在线阅读 下载PDF
A Deep Learning Framework for Arabic Cyberbullying Detection in Social Networks
18
作者 Yahya Tashtoush Areen Banysalim +3 位作者 Majdi Maabreh Shorouq Al-Eidi Ola Karajeh Plamen Zahariev 《Computers, Materials & Continua》 2025年第5期3113-3134,共22页
Social media has emerged as one of the most transformative developments on the internet,revolu-tionizing the way people communicate and interact.However,alongside its benefits,social media has also given rise to signi... Social media has emerged as one of the most transformative developments on the internet,revolu-tionizing the way people communicate and interact.However,alongside its benefits,social media has also given rise to significant challenges,one of the most pressing being cyberbullying.This issue has become a major concern in modern society,particularly due to its profound negative impacts on the mental health and well-being of its victims.In the Arab world,where social media usage is exceptionblly high,cyberbullying has become increasingly prevalent,necessitating urgent attention.Early detection of harmful online behavior is critical to fostering safer digital environments and mitigating the adverse efcts of cyberbullying.This underscores the importance of developing advanced tools and systems to identify and address such behavior efectively.This paper investigates the development of a robust cyberbullying detection and classifcation system tailored for Arabic comments on YouTube.The study explores the efectiveness of various deep learning models,including Bi-LSTM(Bidirectional Long Short Term Memory),LSTM(Long Short-Term Memory),CNN(Convolutional Neural Networks),and a hybrid CNN-LSTM,in classifying Arabic comments into binary classes(bullying or not)and multiclass categories.A comprehensive dataset of 20,000 Arabic YouTube comments was collected,preprocessed,and labeled to support these tasks.The results revealed that the CNN and hybrid CNN-LSTM models achieved the highest accuracy in binary classification,reaching an impressive 91.9%.For multiclass dlassification,the LSTM and Bi-LSTM models outperformed others,achieving an accuracy of 89.5%.These findings highlight the efctiveness of deep learning approaches in the mitigation of cyberbullying within Arabic online communities. 展开更多
关键词 Arabic text lassification arabic text mining cyberbullying detection neural networks deep learning CNN LSTM YOUTUBE Bi-LSTM
在线阅读 下载PDF
Electroencephalogram Signal Classification and Artifact Removal with Deep Networks and Adaptive Thresholding
19
作者 MATHE Mariyadasu MIDIDODDI Padmaja BATTULA TIRUMALA Krishna 《Journal of Shanghai Jiaotong university(Science)》 2025年第4期693-701,共9页
Physiological signals such as electroencephalogram(EEG)signals are often corrupted by artifacts during the acquisition and processing.Some of these artifacts may deteriorate the essential properties of the signal that... Physiological signals such as electroencephalogram(EEG)signals are often corrupted by artifacts during the acquisition and processing.Some of these artifacts may deteriorate the essential properties of the signal that pertains to meaningful information.Most of these artifacts occur due to the involuntary movements or actions the human does during the acquisition process.So,it is recommended to eliminate these artifacts with signal processing approaches.This paper presents two mechanisms of classification and elimination of artifacts.In the first step,a customized deep network is employed to classify clean EEG signals and artifact-included signals.The classification is performed at the feature level,where common space pattern features are extracted with convolutional layers,and these features are later classified with a support vector machine classifier.In the second stage of the work,the artifact signals are decomposed with empirical mode decomposition,and they are then eliminated with the proposed adaptive thresholding mechanism where the threshold value changes for every intrinsic mode decomposition in the iterative mechanism. 展开更多
关键词 artifact elimination deep network electroencephalogram(EEG)signal classification empirical mode decomposition
原文传递
System Modeling and Deep Learning-Based Security Analysis of Uplink NOMA Relay Networks with IRS and Fountain Codes
20
作者 Phu Tran Tin Minh-Sang Van Nguyen +2 位作者 Quy-Anh Bui Agbotiname Lucky Imoize Byung-Seo Kim 《Computer Modeling in Engineering & Sciences》 2025年第8期2521-2543,共23页
Digital content such as games,extended reality(XR),and movies has been widely and easily distributed over wireless networks.As a result,unauthorized access,copyright infringement by third parties or eavesdroppers,and ... Digital content such as games,extended reality(XR),and movies has been widely and easily distributed over wireless networks.As a result,unauthorized access,copyright infringement by third parties or eavesdroppers,and cyberattacks over these networks have become pressing concerns.Therefore,protecting copyrighted content and preventing illegal distribution in wireless communications has garnered significant attention.The Intelligent Reflecting Surface(IRS)is regarded as a promising technology for future wireless and mobile networks due to its ability to reconfigure the radio propagation environment.This study investigates the security performance of an uplink Non-Orthogonal Multiple Access(NOMA)system integrated with an IRS and employing Fountain Codes(FCs).Specifically,two users send signals to the base station at separate distances.A relay receives the signal from the nearby user first and then relays it to the base station.The IRS receives the signal from the distant user and reflects it to the relay,which then sends the reflected signal to the base station.Furthermore,a malevolent eavesdropper intercepts both user and relay communications.We construct mathematical equations for Outage Probability(OP),throughput,diversity evaluation,and Interception Probability(IP),offering quantitative insights to assess system security and performance.Additionally,OP and IP are analyzed using a Deep Neural Network(DNN)model.A deeper comprehension of the security performance of the IRS-assisted NOMA systemin signal transmission is provided by Monte Carlo simulations,which are also carried out to confirm the theoretical conclusions. 展开更多
关键词 Copyright management deep neural network fountain codes intelligent reflecting surface non-orthogonal multiple access physical layer security UPLINK
在线阅读 下载PDF
上一页 1 2 250 下一页 到第
使用帮助 返回顶部