This paper presents a new approach to the delineation of local labor markets based on evolutionary computation. The aim of the exercise is the division of a given territory into functional regions based on travel-to-w...This paper presents a new approach to the delineation of local labor markets based on evolutionary computation. The aim of the exercise is the division of a given territory into functional regions based on travel-to-work flows. Such regions are defined so that a high degree of inter-regional separation and of intra-regional integration in both cases in terms of commuting flows is guaranteed. Additional requirements include the absence of overlap between delineated regions and the exhaustive coverage of the whole territory. The procedure is based on the maximization of a fitness function that measures aggregate intra-region interaction under constraints of inter-region separation and minimum size. In the experimentation stage, two variations of the fitness function are used, and the process is also applied as a final stage for the optimization of the results from one of the most successful existing methods, which are used by the British authorities for the delineation of travel-to-work areas (TTWAs). The empirical exercise is conducted using real data for a sufficiently large territory that is considered to be representative given the density and variety of travel-to-work patterns that it embraces. The paper includes the quantitative comparison with alternative traditional methods, the assessment of the performance of the set of operators which has been specifically designed to handle the regionalization problem and the evaluation of the convergence process. The robustness of the solutions, something crucial in a research and policy-making context, is also discussed in the paper.展开更多
Clustering algorithms optimization can minimize topology maintenance overhead in large scale vehicular Ad hoc networks(VANETs)for smart transportation that results from dynamic topology,limited resources and noncentra...Clustering algorithms optimization can minimize topology maintenance overhead in large scale vehicular Ad hoc networks(VANETs)for smart transportation that results from dynamic topology,limited resources and noncentralized architecture.The performance of a clustering algorithm varies with the underlying mobility model to address the topology maintenance overhead issue in VANETs for smart transportation.To design a robust clustering algorithm,careful attention must be paid to components like mobility models and performance objectives.A clustering algorithm may not perform well with every mobility pattern.Therefore,we propose a supervisory protocol(SP)that observes the mobility pattern of vehicles and identies the realistic Mobility model through microscopic features.An analytical model can be used to determine an efcient clustering algorithm for a specic mobility model(MM).SP selects the best clustering scheme according to the mobility model and guarantees a consistent performance throughout VANET operations.The simulation has performed in three parts that is the central part simulation for setting up the clustering environment,In the second part the clustering algorithms are tested for efciency in a constrained atmosphere for some time and the third part represents the proposed scheme.The simulation results show that the proposed scheme outperforms clustering algorithms such as honey bee algorithm-based clustering and memetic clustering in terms of cluster count,re-afliation rate,control overhead and cluster lifetime.展开更多
The world is rapidly changing with the advance of information technology.The expansion of the Internet of Things(IoT)is a huge step in the development of the smart city.The IoT consists of connected devices that trans...The world is rapidly changing with the advance of information technology.The expansion of the Internet of Things(IoT)is a huge step in the development of the smart city.The IoT consists of connected devices that transfer information.The IoT architecture permits on-demand services to a public pool of resources.Cloud computing plays a vital role in developing IoT-enabled smart applications.The integration of cloud computing enhances the offering of distributed resources in the smart city.Improper management of security requirements of cloud-assisted IoT systems can bring about risks to availability,security,performance,condentiality,and privacy.The key reason for cloud-and IoT-enabled smart city application failure is improper security practices at the early stages of development.This article proposes a framework to collect security requirements during the initial development phase of cloud-assisted IoT-enabled smart city applications.Its three-layered architecture includes privacy preserved stakeholder analysis(PPSA),security requirement modeling and validation(SRMV),and secure cloud-assistance(SCA).A case study highlights the applicability and effectiveness of the proposed framework.A hybrid survey enables the identication and evaluation of signicant challenges.展开更多
Since COVID-19 infections are increasing all over the world,there is a need for developing solutions for its early and accurate diagnosis is a must.Detectionmethods for COVID-19 include screeningmethods like Chest X-r...Since COVID-19 infections are increasing all over the world,there is a need for developing solutions for its early and accurate diagnosis is a must.Detectionmethods for COVID-19 include screeningmethods like Chest X-rays and Computed Tomography(CT)scans.More work must be done on preprocessing the datasets,such as eliminating the diaphragm portions,enhancing the image intensity,and minimizing noise.In addition to the detection of COVID-19,the severity of the infection needs to be estimated.The HSDC model is proposed to solve these problems,which will detect and classify the severity of COVID-19 from X-ray and CT-scan images.For CT-scan images,the histogram threshold of the input image is adaptively determined using the ICH Swarm Optimization Segmentation(ICHSeg)algorithm.Based on the Statistical and Shape-based feature vectors(FVs),the extracted regions are classified using a Hybrid model for CT images(HSDCCT)algorithm.When the infections are detected,it’s classified as Normal,Moderate,and Severe.A fused FHI is formed for X-ray images by extracting the features of Histogram-oriented gradient(HOG)and Image profile(IP).The FHI features of X-ray images are classified using Hybrid Support Vector Machine(SVM)and Deep Convolutional Neural Network(DCNN)HSDCX algorithm into COVID-19 or else Pneumonia,or Normal.Experimental results have shown that the accuracy of the HSDC model attains the highest of 94.6 for CT-scan images and 95.6 for X-ray images when compared to SVM and DCNN.This study thus significantly helps medical professionals and doctors diagnose COVID-19 infections quickly,which is the most needed in current years.展开更多
Efficient routing protocols are crucial for enabling secure communication among the highly mobile and self-configurable nodes in Vehicular Ad-Hoc Networks(VANETs). In this work, we present a performance evaluation of ...Efficient routing protocols are crucial for enabling secure communication among the highly mobile and self-configurable nodes in Vehicular Ad-Hoc Networks(VANETs). In this work, we present a performance evaluation of different routing protocols in VANETs based on the currently available research. Our study focuses on analyzing the strength and weaknesses of the routing protocols, namely, Ad-Hoc On-demand Distance Vector(AODV), Dynamic Source Routing(DSR), and Destination-Sequenced Distance-Vector(DSDV), under varying network conditions. We examine the protocols’ performance based on key metrics such as throughput, delay, and energy consumption. We also highlight the advantages and limitations of each protocol in different scenarios, such as varying vehicular densities and mobility patterns. Our results show that AODV outperforms DSR and DSDV in terms of throughput and delay, while DSR consumes the least energy. We also observe that the performance of the routing protocols varies with the density of vehicles and the mobility patterns of the nodes. Our study highlights the importance of conducting real-world experiments to evaluate the performance of routing protocols in VANETs, as they provide more realistic and accurate results than simulation-based studies. Our findings can help in the selection and design of efficient and secure routing protocols for VANETs.展开更多
The introduction of new technologies has increased communication network coverage and the number of associating nodes in dynamic communication networks(DCN).As the network has the characteristics like decentralized an...The introduction of new technologies has increased communication network coverage and the number of associating nodes in dynamic communication networks(DCN).As the network has the characteristics like decentralized and dynamic,few nodes in the network may not associate with other nodes.These uncooperative nodes also known as selfish nodes corrupt the performance of the cooperative nodes.Namely,the nodes cause congestion,high delay,security concerns,and resource depletion.This study presents an effective selfish node detection method to address these problems.The Price of Anarchy(PoA)and the Price of Stability(PoS)in Game Theory with the Presence of Nash Equilibrium(NE)are discussed for the Selfish Node Detection.This is a novel experiment to detect selfish nodes in a network using PoA.Moreover,the least response dynamic-based Capacitated Selfish Resource Allocation(CSRA)game is introduced to improve resource usage among the nodes.The suggested strategy is simulated using the Solar Winds simulator,and the simulation results show that,when compared to earlier methods,the new scheme offers promising performance in terms of delivery rate,delay,and throughput.展开更多
Fog computing is a rapidly growing technology that aids in pipelining the possibility of mitigating breaches between the cloud and edge servers.It facil-itates the benefits of the network edge with the maximized probab...Fog computing is a rapidly growing technology that aids in pipelining the possibility of mitigating breaches between the cloud and edge servers.It facil-itates the benefits of the network edge with the maximized probability of offering interaction with the cloud.However,the fog computing characteristics are suscep-tible to counteract the challenges of security.The issues present with the Physical Layer Security(PLS)aspect in fog computing which included authentication,integrity,and confidentiality has been considered as a reason for the potential issues leading to the security breaches.In this work,the Octonion Algebra-inspired Non-Commutative Ring-based Fully Homomorphic Encryption Scheme(NCR-FHE)was proposed as a secrecy improvement technique to overcome the impersonation attack in cloud computing.The proposed approach was derived through the benefits of Octonion algebra to facilitate the maximum security for big data-based applications.The major issues in the physical layer security which may potentially lead to the possible security issues were identified.The potential issues causing the impersonation attack in the Fog computing environment were identified.The proposed approach was compared with the existing encryption approaches and claimed as a robust approach to identify the impersonation attack for the fog and edge network.The computation cost of the proposed NCR-FHE is identified to be significantly reduced by 7.18%,8.64%,9.42%,and 10.36%in terms of communication overhead for varying packet sizes,when compared to the benchmarked ECDH-DH,LHPPS,BF-PHE and SHE-PABF schemes.展开更多
The creation of the 3D rendering model involves the prediction of an accurate depth map for the input images.A proposed approach of a modified semi-global block matching algorithm with variable window size and the gra...The creation of the 3D rendering model involves the prediction of an accurate depth map for the input images.A proposed approach of a modified semi-global block matching algorithm with variable window size and the gradient assessment of objects predicts the depth map.3D modeling and view synthesis algorithms could effectively handle the obtained disparity maps.This work uses the consistency check method to find an accurate depth map for identifying occluded pixels.The prediction of the disparity map by semi-global block matching has used the benchmark dataset of Middlebury stereo for evaluation.The improved depth map quality within a reasonable process-ing time outperforms the other existing depth map prediction algorithms.The experimental results have shown that the proposed depth map predictioncould identify the inter-object boundaryeven with the presence ofocclusion with less detection error and runtime.We observed that the Middlebury stereo dataset has very few images with occluded objects,which made the attainment of gain cumbersome.Considering this gain,we have created our dataset with occlu-sion using the structured lighting technique.The proposed regularization term as an optimization process in the graph cut algorithm handles occlusion for different smoothing coefficients.The experimented results demonstrated that our dataset had outperformed the Tsukuba dataset regarding the percentage of occluded pixels.展开更多
Background:Pneumonia remains a critical global health challenge,manifesting as a severe respiratory infection caused by viruses,bacteria,and fungi.Early detection is paramount for effective treatment,potentially reduc...Background:Pneumonia remains a critical global health challenge,manifesting as a severe respiratory infection caused by viruses,bacteria,and fungi.Early detection is paramount for effective treatment,potentially reducing mortality rates and optimizing healthcare resource allocation.Despite the importance of chest X-ray diagnosis,image analysis presents significant challenges,particularly in regions with limited medical expertise.This study addresses these challenges by proposing a computer-aided diagnosis system leveraging targeted image preprocessing and optimized deep learning techniques.Methods:We systematically evaluated contrast limited adaptive histogram equalization with varying clip limits for preprocessing chest X-ray images,demonstrating its effectiveness in enhancing feature visibility for diagnostic accuracy.Employing a comprehensive dataset of 5,863 X-ray images(1,583 pneumonia-negative,4,280 pneumonia-positive)collected from multiple healthcare facilities,we conducted a comparative analysis of transfer learning with pre-trained models including ResNet50v2,VGG-19,and MobileNetV2.Statistical validation was performed through 5-fold cross-validation.Results:Our results show that the contrast limited adaptive histogram equalization-enhanced approach with ResNet50v2 achieves 93.40%accuracy,outperforming VGG-19(84.90%)and MobileNetV2(89.70%).Statistical validation confirms the significance of these improvements(P<0.01).The development and optimization resulted in a lightweight mobile application(74 KB)providing rapid diagnostic support(1-2 s response time).Conclusion:The proposed approach demonstrates practical applicability in resource-constrained settings,balancing diagnostic accuracy with deployment efficiency,and offers a viable solution for computer-aided pneumonia diagnosis in areas with limited medical expertise.展开更多
Automated and accurate movie genre classification is crucial for content organization,recommendation systems,and audience targeting in the film industry.Although most existing approaches focus on audiovisual features ...Automated and accurate movie genre classification is crucial for content organization,recommendation systems,and audience targeting in the film industry.Although most existing approaches focus on audiovisual features such as trailers and posters,the text-based classification remains underexplored despite its accessibility and semantic richness.This paper introduces the Genre Attention Model(GAM),a deep learning architecture that integrates transformer models with a hierarchical attention mechanism to extract and leverage contextual information from movie plots formulti-label genre classification.In order to assess its effectiveness,we assessmultiple transformer-based models,including Bidirectional Encoder Representations fromTransformers(BERT),ALite BERT(ALBERT),Distilled BERT(DistilBERT),Robustly Optimized BERT Pretraining Approach(RoBERTa),Efficiently Learning an Encoder that Classifies Token Replacements Accurately(ELECTRA),eXtreme Learning Network(XLNet)and Decodingenhanced BERT with Disentangled Attention(DeBERTa).Experimental results demonstrate the superior performance of DeBERTa-based GAM,which employs a two-tier hierarchical attention mechanism:word-level attention highlights key terms,while sentence-level attention captures critical narrative segments,ensuring a refined and interpretable representation of movie plots.Evaluated on three benchmark datasets Trailers12K,Large Movie Trailer Dataset-9(LMTD-9),and MovieLens37K.GAM achieves micro-average precision scores of 83.63%,83.32%,and 83.34%,respectively,surpassing state-of-the-artmodels.Additionally,GAMis computationally efficient,requiring just 6.10Giga Floating Point Operations Per Second(GFLOPS),making it a scalable and cost-effective solution.These results highlight the growing potential of text-based deep learning models in genre classification and GAM’s effectiveness in improving predictive accuracy while maintaining computational efficiency.With its robust performance,GAM offers a versatile and scalable framework for content recommendation,film indexing,and media analytics,providing an interpretable alternative to traditional audiovisual-based classification techniques.展开更多
Phishing is a type of cybercrime in which cyber-attackers pose themselves as authorized persons or entities and hack the victims’sensitive data.E-mails,instant messages and phone calls are some of the common modes us...Phishing is a type of cybercrime in which cyber-attackers pose themselves as authorized persons or entities and hack the victims’sensitive data.E-mails,instant messages and phone calls are some of the common modes used in cyberattacks.Though the security models are continuously upgraded to prevent cyberattacks,hackers find innovative ways to target the victims.In this background,there is a drastic increase observed in the number of phishing emails sent to potential targets.This scenario necessitates the importance of designing an effective classification model.Though numerous conventional models are available in the literature for proficient classification of phishing emails,the Machine Learning(ML)techniques and the Deep Learning(DL)models have been employed in the literature.The current study presents an Intelligent Cuckoo Search(CS)Optimization Algorithm with a Deep Learning-based Phishing Email Detection and Classification(ICSOA-DLPEC)model.The aim of the proposed ICSOA-DLPEC model is to effectually distinguish the emails as either legitimate or phishing ones.At the initial stage,the pre-processing is performed through three stages such as email cleaning,tokenization and stop-word elimination.Then,the N-gram approach is;moreover,the CS algorithm is applied to extract the useful feature vectors.Moreover,the CS algorithm is employed with the Gated Recurrent Unit(GRU)model to detect and classify phishing emails.Furthermore,the CS algorithm is used to fine-tune the parameters involved in the GRU model.The performance of the proposed ICSOA-DLPEC model was experimentally validated using a benchmark dataset,and the results were assessed under several dimensions.Extensive comparative studies were conducted,and the results confirmed the superior performance of the proposed ICSOA-DLPEC model over other existing approaches.The proposed model achieved a maximum accuracy of 99.72%.展开更多
Bat algorithm(BA)is an eminent meta-heuristic algorithm that has been widely used to solve diverse kinds of optimization problems.BA leverages the echolocation feature of bats produced by imitating the bats’searching...Bat algorithm(BA)is an eminent meta-heuristic algorithm that has been widely used to solve diverse kinds of optimization problems.BA leverages the echolocation feature of bats produced by imitating the bats’searching behavior.BA faces premature convergence due to its local search capability.Instead of using the standard uniform walk,the Torus walk is viewed as a promising alternative to improve the local search capability.In this work,we proposed an improved variation of BA by applying torus walk to improve diversity and convergence.The proposed.Modern Computerized Bat Algorithm(MCBA)approach has been examined for fifteen well-known benchmark test problems.The finding of our technique shows promising performance as compared to the standard PSO and standard BA.The proposed MCBA,BPA,Standard PSO,and Standard BA have been examined for well-known benchmark test problems and training of the artificial neural network(ANN).We have performed experiments using eight benchmark datasets applied from the worldwide famous machine-learning(ML)repository of UCI.Simulation results have shown that the training of an ANN with MCBA-NN algorithm tops the list considering exactness,with more superiority compared to the traditional methodologies.The MCBA-NN algorithm may be used effectively for data classification and statistical problems in the future.展开更多
An intrusion detection system(IDS)becomes an important tool for ensuring security in the network.In recent times,machine learning(ML)and deep learning(DL)models can be applied for the identification of intrusions over...An intrusion detection system(IDS)becomes an important tool for ensuring security in the network.In recent times,machine learning(ML)and deep learning(DL)models can be applied for the identification of intrusions over the network effectively.To resolve the security issues,this paper presents a new Binary Butterfly Optimization algorithm based on Feature Selection with DRL technique,called BBOFS-DRL for intrusion detection.The proposed BBOFSDRL model mainly accomplishes the recognition of intrusions in the network.To attain this,the BBOFS-DRL model initially designs the BBOFS algorithm based on the traditional butterfly optimization algorithm(BOA)to elect feature subsets.Besides,DRL model is employed for the proper identification and classification of intrusions that exist in the network.Furthermore,beetle antenna search(BAS)technique is applied to tune the DRL parameters for enhanced intrusion detection efficiency.For ensuring the superior intrusion detection outcomes of the BBOFS-DRL model,a wide-ranging experimental analysis is performed against benchmark dataset.The simulation results reported the supremacy of the BBOFS-DRL model over its recent state of art approaches.展开更多
The Brain Tumor(BT)is created by an uncontrollable rise of anomalous cells in brain tissue,and it consists of 2 types of cancers they are malignant and benign tumors.The benevolent BT does not affect the neighbouring ...The Brain Tumor(BT)is created by an uncontrollable rise of anomalous cells in brain tissue,and it consists of 2 types of cancers they are malignant and benign tumors.The benevolent BT does not affect the neighbouring healthy and normal tissue;however,the malignant could affect the adjacent brain tissues,which results in death.Initial recognition of BT is highly significant to protecting the patient’s life.Generally,the BT can be identified through the magnetic resonance imaging(MRI)scanning technique.But the radiotherapists are not offering effective tumor segmentation in MRI images because of the position and unequal shape of the tumor in the brain.Recently,ML has prevailed against standard image processing techniques.Several studies denote the superiority of machine learning(ML)techniques over standard techniques.Therefore,this study develops novel brain tumor detection and classification model using met heuristic optimization with machine learning(BTDC-MOML)model.To accomplish the detection of brain tumor effectively,a Computer-Aided Design(CAD)model using Machine Learning(ML)technique is proposed in this research manuscript.Initially,the input image pre-processing is performed using Gaborfiltering(GF)based noise removal,contrast enhancement,and skull stripping.Next,mayfly optimization with the Kapur’s thresholding based segmentation process takes place.For feature extraction proposes,local diagonal extreme patterns(LDEP)are exploited.At last,the Extreme Gradient Boosting(XGBoost)model can be used for the BT classification process.The accuracy analysis is performed in terms of Learning accuracy,and the validation accuracy is performed to determine the efficiency of the proposed research work.The experimental validation of the proposed model demonstrates its promising performance over other existing methods.展开更多
Because of its strong ability to solve problems,evolutionary multitask optimization(EMTO)algorithms have been widely studied recently.Evolutionary algorithms have the advantage of fast searching for the optimal soluti...Because of its strong ability to solve problems,evolutionary multitask optimization(EMTO)algorithms have been widely studied recently.Evolutionary algorithms have the advantage of fast searching for the optimal solution,but it is easy to fall into local optimum and difficult to generalize.Combining evolutionary multitask algorithms with evolutionary optimization algorithms can be an effective method for solving these problems.Through the implicit parallelism of tasks themselves and the knowledge transfer between tasks,more promising individual algorithms can be generated in the evolution process,which can jump out of the local optimum.How to better combine the two has also been studied more and more.This paper explores the existing evolutionary multitasking theory and improvement scheme in detail.Then,it summarizes the application of EMTO in different scenarios.Finally,according to the existing research,the future research trends and potential exploration directions are revealed.展开更多
Internet of Things(IoT)allows several low resources and controlled devices to interconnect,calculate processes and make decisions in the communication network.In the heterogeneous environment for IoT devices,several c...Internet of Things(IoT)allows several low resources and controlled devices to interconnect,calculate processes and make decisions in the communication network.In the heterogeneous environment for IoT devices,several challenging issues such as energy,storage,efficiency,and security.The design of encryption techniques enables the transmission of the data in the IoT environment in a secured way.The proper selection of optimal keys helps to boost the encryption performance.With this motivation,the study presents a signcryption with quantum chaotic krill herd algorithm for secured data transmission(SCQCKH-SDT)in IoT environment.The proposed SCQCKHSDT technique aims to effectively encrypts the data by the use of optimal keys generated by the CQKH algorithm.The proposed SCQCKH-SDT technique initially employs the signcryption technique for the encryption of data.In order to optimize the secrecy,the optimal key generation process is carried out using Chaotic Krill Herd(CQKH)algorithm.The CQKH algorithm incorporates the concept of quantum computing and chaotic theory into the traditional KH algorithm.The performance validation of the SCQCKH-SDT technique is performed using benchmark dataset.An extensive comparative analysis reported the superior performance of the SCQCKH-SDT technique over the recent approaches.展开更多
Numerous methods are analysed in detail to improve task schedulingand data security performance in the cloud environment. The methodsinvolve scheduling according to the factors like makespan, waiting time,cost, deadli...Numerous methods are analysed in detail to improve task schedulingand data security performance in the cloud environment. The methodsinvolve scheduling according to the factors like makespan, waiting time,cost, deadline, and popularity. However, the methods are inappropriate forachieving higher scheduling performance. Regarding data security, existingmethods use various encryption schemes but introduce significant serviceinterruption. This article sketches a practical Real-time Application CentricTRS (Throughput-Resource utilization–Success) Scheduling with Data Security(RATRSDS) model by considering all these issues in task scheduling anddata security. The method identifies the required resource and their claim timeby receiving the service requests. Further, for the list of resources as services,the method computes throughput support (Thrs) according to the number ofstatements executed and the complete statements of the service. Similarly, themethod computes Resource utilization support (Ruts) according to the idletime on any duty cycle and total servicing time. Also, the method computesthe value of Success support (Sus) according to the number of completions forthe number of allocations. The method estimates the TRS score (ThroughputResource utilization Success) for different resources using all these supportmeasures. According to the value of the TRS score, the services are rankedand scheduled. On the other side, based on the requirement of service requests,the method computes Requirement Support (RS). The selection of service isperformed and allocated. Similarly, choosing the route according to the RouteSupport Measure (RSM) enforced route security. Finally, data security hasgets implemented with a service-based encryption technique. The RATRSDSscheme has claimed higher performance in data security and scheduling.展开更多
The impact of a Distributed Denial of Service(DDoS)attack on Soft-ware Defined Networks(SDN)is briefly analyzed.Many approaches to detecting DDoS attacks exist,varying on the feature being considered and the method us...The impact of a Distributed Denial of Service(DDoS)attack on Soft-ware Defined Networks(SDN)is briefly analyzed.Many approaches to detecting DDoS attacks exist,varying on the feature being considered and the method used.Still,the methods have a deficiency in the performance of detecting DDoS attacks and mitigating them.To improve the performance of SDN,an efficient Real-time Multi-Constrained Adaptive Replication and Traffic Approximation Model(RMCARTAM)is sketched in this article.The RMCARTAM considers different parameters or constraints in running different controllers responsible for handling incoming packets.The model is designed with multiple controllers to handle net-work traffic but can turn the controllers according to requirements.The multi-con-straint adaptive replication model monitors different features of network traffic like rate of packet reception,class-based packet reception and target-specific reception.According to these features,the method estimates the Replication Turn-ing Weight(RTW)based on which triggering controllers are performed.Similarly,the method applies Traffic Approximation(TA)in the detection of DDoS attacks.The detection of a DDoS attack is performed by approximating the incoming traf-fic to any service and using various features like hop count,payload,service fre-quency,and malformed frequency to compute various support measures on bandwidth access,data support,frequency support,malformed support,route sup-port,and so on.Using all these support measures,the method computes the value of legitimate weight to conclude the behavior of any source in identifying the mal-icious node.Identified node details are used in the mitigation of DDoS attacks.The method stimulates the network performance by reducing the power factor by switching the controller according to different factors,which also reduces the cost.In the same way,the proposed model improves the accuracy of detecting DDoS attacks by estimating the features of incoming traffic in different corners.展开更多
Reliable transmission is vital to the success of the next generation ofcommunications technologies and Fifth Generation (5G) networks. Many sensitive applications, such as eHealth and medical services, can benefit fro...Reliable transmission is vital to the success of the next generation ofcommunications technologies and Fifth Generation (5G) networks. Many sensitive applications, such as eHealth and medical services, can benefit from a 5G network. The Internet of Medical Things (IoMT) is a new field that fosters themaintenance of trust among various IoMT Device to Device (D2D) modern technologies. In IoMT the medical devices have to be connected through a wirelessnetwork and constantly needs to be self-configured to provide consistent and effi-cient data transmission. The medical devices need to be connected with sophisticated protocols and architecture to handle the synergy of the monitoring devices.Today, one of the commonly used algorithms in D2D communication is the Optimized Link State Routing protocol (OLSR). The OLSR is considerably good ateffectively utilizing the bandwidth and reserving the paths. One of the majorattack against the OLSR is the Node isolation attack, also known as the Gray holedenial of service attack. The Gray hole attack exploits the vulnerabilities presentwith sharing the topological information of the network. The attackers may usethis topological information to maliciously disconnect the target nodes from theexisting network and stops rendering the communication services to the victimnode. Hence, considering the sensitivity and security concerns of the data usedin e-Health applications, these types of attacks must be detected and disabledproactively. In this work, a novel Node Authentication (NA) with OLSR is proposed. The simulation experiments illustrated that the proposed protocol has anexcellent Packet Delivery Ratio, minimal End-End delay, and minimal Packet losswhen compared to the Ad-hoc On-Demand Distance Victor (AODV) protocol andthe proposed authentication scheme was able to protect the OLSR protocol from anode isolation attack.展开更多
Prediction of stock market value is highly risky because it is based on the concept of Time Series forecasting system that can be used for investments in a safe environment with minimized chances of loss.The proposed ...Prediction of stock market value is highly risky because it is based on the concept of Time Series forecasting system that can be used for investments in a safe environment with minimized chances of loss.The proposed model uses a real time dataset offifteen Stocks as input into the system and based on the data,predicts or forecast future stock prices of different companies belonging to different sectors.The dataset includes approximatelyfifteen companies from different sectors and forecasts their results based on which the user can decide whether to invest in the particular company or not;the forecasting is done for the next quarter.Our model uses 3 main concepts for forecasting results.Thefirst one is for stocks that show periodic change throughout the season,the‘Holt-Winters Triple Exponential Smoothing’.3 basic things taken into conclusion by this algorithm are Base Level,Trend Level and Seasoning Factor.The value of all these are calculated by us and then decomposition of all these factors is done by the Holt-Winters Algorithm.The second concept is‘Recurrent Neural Network’.The specific model of recurrent neural network that is being used is Long-Short Term Memory and it’s the same as the Normal Neural Network,the only difference is that each intermediate cell is a memory cell and retails its value till the next feedback loop.The third concept is Recommendation System whichfilters and predict the rating based on the different factors.展开更多
基金This work was supported by Spanish National Plan of R+D+i from Spanish Ministry of Education and Science(Ministerio de Educación y Ciencia)for the project Local Labour Markets:New Methods of Delineation and Analysis(No.8EJ2007-67767-C04-02)the European Social Fund(ESF)and the University of Alicante.
文摘This paper presents a new approach to the delineation of local labor markets based on evolutionary computation. The aim of the exercise is the division of a given territory into functional regions based on travel-to-work flows. Such regions are defined so that a high degree of inter-regional separation and of intra-regional integration in both cases in terms of commuting flows is guaranteed. Additional requirements include the absence of overlap between delineated regions and the exhaustive coverage of the whole territory. The procedure is based on the maximization of a fitness function that measures aggregate intra-region interaction under constraints of inter-region separation and minimum size. In the experimentation stage, two variations of the fitness function are used, and the process is also applied as a final stage for the optimization of the results from one of the most successful existing methods, which are used by the British authorities for the delineation of travel-to-work areas (TTWAs). The empirical exercise is conducted using real data for a sufficiently large territory that is considered to be representative given the density and variety of travel-to-work patterns that it embraces. The paper includes the quantitative comparison with alternative traditional methods, the assessment of the performance of the set of operators which has been specifically designed to handle the regionalization problem and the evaluation of the convergence process. The robustness of the solutions, something crucial in a research and policy-making context, is also discussed in the paper.
基金The authors extend their appreciation to King Saud University for funding this work through Researchers supporting project number(RSP-2020/133),King Saud University,Riyadh,Saudi Arabia.
文摘Clustering algorithms optimization can minimize topology maintenance overhead in large scale vehicular Ad hoc networks(VANETs)for smart transportation that results from dynamic topology,limited resources and noncentralized architecture.The performance of a clustering algorithm varies with the underlying mobility model to address the topology maintenance overhead issue in VANETs for smart transportation.To design a robust clustering algorithm,careful attention must be paid to components like mobility models and performance objectives.A clustering algorithm may not perform well with every mobility pattern.Therefore,we propose a supervisory protocol(SP)that observes the mobility pattern of vehicles and identies the realistic Mobility model through microscopic features.An analytical model can be used to determine an efcient clustering algorithm for a specic mobility model(MM).SP selects the best clustering scheme according to the mobility model and guarantees a consistent performance throughout VANET operations.The simulation has performed in three parts that is the central part simulation for setting up the clustering environment,In the second part the clustering algorithms are tested for efciency in a constrained atmosphere for some time and the third part represents the proposed scheme.The simulation results show that the proposed scheme outperforms clustering algorithms such as honey bee algorithm-based clustering and memetic clustering in terms of cluster count,re-afliation rate,control overhead and cluster lifetime.
基金Taif University Researchers Supporting Project No.(TURSP-2020/126),Taif University,Taif,Saudi Arabia。
文摘The world is rapidly changing with the advance of information technology.The expansion of the Internet of Things(IoT)is a huge step in the development of the smart city.The IoT consists of connected devices that transfer information.The IoT architecture permits on-demand services to a public pool of resources.Cloud computing plays a vital role in developing IoT-enabled smart applications.The integration of cloud computing enhances the offering of distributed resources in the smart city.Improper management of security requirements of cloud-assisted IoT systems can bring about risks to availability,security,performance,condentiality,and privacy.The key reason for cloud-and IoT-enabled smart city application failure is improper security practices at the early stages of development.This article proposes a framework to collect security requirements during the initial development phase of cloud-assisted IoT-enabled smart city applications.Its three-layered architecture includes privacy preserved stakeholder analysis(PPSA),security requirement modeling and validation(SRMV),and secure cloud-assistance(SCA).A case study highlights the applicability and effectiveness of the proposed framework.A hybrid survey enables the identication and evaluation of signicant challenges.
文摘Since COVID-19 infections are increasing all over the world,there is a need for developing solutions for its early and accurate diagnosis is a must.Detectionmethods for COVID-19 include screeningmethods like Chest X-rays and Computed Tomography(CT)scans.More work must be done on preprocessing the datasets,such as eliminating the diaphragm portions,enhancing the image intensity,and minimizing noise.In addition to the detection of COVID-19,the severity of the infection needs to be estimated.The HSDC model is proposed to solve these problems,which will detect and classify the severity of COVID-19 from X-ray and CT-scan images.For CT-scan images,the histogram threshold of the input image is adaptively determined using the ICH Swarm Optimization Segmentation(ICHSeg)algorithm.Based on the Statistical and Shape-based feature vectors(FVs),the extracted regions are classified using a Hybrid model for CT images(HSDCCT)algorithm.When the infections are detected,it’s classified as Normal,Moderate,and Severe.A fused FHI is formed for X-ray images by extracting the features of Histogram-oriented gradient(HOG)and Image profile(IP).The FHI features of X-ray images are classified using Hybrid Support Vector Machine(SVM)and Deep Convolutional Neural Network(DCNN)HSDCX algorithm into COVID-19 or else Pneumonia,or Normal.Experimental results have shown that the accuracy of the HSDC model attains the highest of 94.6 for CT-scan images and 95.6 for X-ray images when compared to SVM and DCNN.This study thus significantly helps medical professionals and doctors diagnose COVID-19 infections quickly,which is the most needed in current years.
文摘Efficient routing protocols are crucial for enabling secure communication among the highly mobile and self-configurable nodes in Vehicular Ad-Hoc Networks(VANETs). In this work, we present a performance evaluation of different routing protocols in VANETs based on the currently available research. Our study focuses on analyzing the strength and weaknesses of the routing protocols, namely, Ad-Hoc On-demand Distance Vector(AODV), Dynamic Source Routing(DSR), and Destination-Sequenced Distance-Vector(DSDV), under varying network conditions. We examine the protocols’ performance based on key metrics such as throughput, delay, and energy consumption. We also highlight the advantages and limitations of each protocol in different scenarios, such as varying vehicular densities and mobility patterns. Our results show that AODV outperforms DSR and DSDV in terms of throughput and delay, while DSR consumes the least energy. We also observe that the performance of the routing protocols varies with the density of vehicles and the mobility patterns of the nodes. Our study highlights the importance of conducting real-world experiments to evaluate the performance of routing protocols in VANETs, as they provide more realistic and accurate results than simulation-based studies. Our findings can help in the selection and design of efficient and secure routing protocols for VANETs.
文摘The introduction of new technologies has increased communication network coverage and the number of associating nodes in dynamic communication networks(DCN).As the network has the characteristics like decentralized and dynamic,few nodes in the network may not associate with other nodes.These uncooperative nodes also known as selfish nodes corrupt the performance of the cooperative nodes.Namely,the nodes cause congestion,high delay,security concerns,and resource depletion.This study presents an effective selfish node detection method to address these problems.The Price of Anarchy(PoA)and the Price of Stability(PoS)in Game Theory with the Presence of Nash Equilibrium(NE)are discussed for the Selfish Node Detection.This is a novel experiment to detect selfish nodes in a network using PoA.Moreover,the least response dynamic-based Capacitated Selfish Resource Allocation(CSRA)game is introduced to improve resource usage among the nodes.The suggested strategy is simulated using the Solar Winds simulator,and the simulation results show that,when compared to earlier methods,the new scheme offers promising performance in terms of delivery rate,delay,and throughput.
文摘Fog computing is a rapidly growing technology that aids in pipelining the possibility of mitigating breaches between the cloud and edge servers.It facil-itates the benefits of the network edge with the maximized probability of offering interaction with the cloud.However,the fog computing characteristics are suscep-tible to counteract the challenges of security.The issues present with the Physical Layer Security(PLS)aspect in fog computing which included authentication,integrity,and confidentiality has been considered as a reason for the potential issues leading to the security breaches.In this work,the Octonion Algebra-inspired Non-Commutative Ring-based Fully Homomorphic Encryption Scheme(NCR-FHE)was proposed as a secrecy improvement technique to overcome the impersonation attack in cloud computing.The proposed approach was derived through the benefits of Octonion algebra to facilitate the maximum security for big data-based applications.The major issues in the physical layer security which may potentially lead to the possible security issues were identified.The potential issues causing the impersonation attack in the Fog computing environment were identified.The proposed approach was compared with the existing encryption approaches and claimed as a robust approach to identify the impersonation attack for the fog and edge network.The computation cost of the proposed NCR-FHE is identified to be significantly reduced by 7.18%,8.64%,9.42%,and 10.36%in terms of communication overhead for varying packet sizes,when compared to the benchmarked ECDH-DH,LHPPS,BF-PHE and SHE-PABF schemes.
文摘The creation of the 3D rendering model involves the prediction of an accurate depth map for the input images.A proposed approach of a modified semi-global block matching algorithm with variable window size and the gradient assessment of objects predicts the depth map.3D modeling and view synthesis algorithms could effectively handle the obtained disparity maps.This work uses the consistency check method to find an accurate depth map for identifying occluded pixels.The prediction of the disparity map by semi-global block matching has used the benchmark dataset of Middlebury stereo for evaluation.The improved depth map quality within a reasonable process-ing time outperforms the other existing depth map prediction algorithms.The experimental results have shown that the proposed depth map predictioncould identify the inter-object boundaryeven with the presence ofocclusion with less detection error and runtime.We observed that the Middlebury stereo dataset has very few images with occluded objects,which made the attainment of gain cumbersome.Considering this gain,we have created our dataset with occlu-sion using the structured lighting technique.The proposed regularization term as an optimization process in the graph cut algorithm handles occlusion for different smoothing coefficients.The experimented results demonstrated that our dataset had outperformed the Tsukuba dataset regarding the percentage of occluded pixels.
文摘Background:Pneumonia remains a critical global health challenge,manifesting as a severe respiratory infection caused by viruses,bacteria,and fungi.Early detection is paramount for effective treatment,potentially reducing mortality rates and optimizing healthcare resource allocation.Despite the importance of chest X-ray diagnosis,image analysis presents significant challenges,particularly in regions with limited medical expertise.This study addresses these challenges by proposing a computer-aided diagnosis system leveraging targeted image preprocessing and optimized deep learning techniques.Methods:We systematically evaluated contrast limited adaptive histogram equalization with varying clip limits for preprocessing chest X-ray images,demonstrating its effectiveness in enhancing feature visibility for diagnostic accuracy.Employing a comprehensive dataset of 5,863 X-ray images(1,583 pneumonia-negative,4,280 pneumonia-positive)collected from multiple healthcare facilities,we conducted a comparative analysis of transfer learning with pre-trained models including ResNet50v2,VGG-19,and MobileNetV2.Statistical validation was performed through 5-fold cross-validation.Results:Our results show that the contrast limited adaptive histogram equalization-enhanced approach with ResNet50v2 achieves 93.40%accuracy,outperforming VGG-19(84.90%)and MobileNetV2(89.70%).Statistical validation confirms the significance of these improvements(P<0.01).The development and optimization resulted in a lightweight mobile application(74 KB)providing rapid diagnostic support(1-2 s response time).Conclusion:The proposed approach demonstrates practical applicability in resource-constrained settings,balancing diagnostic accuracy with deployment efficiency,and offers a viable solution for computer-aided pneumonia diagnosis in areas with limited medical expertise.
基金would like to thank the Deanship of Graduate Studies and Scientific Research at Qassim University for financial support(QU-APC-2025).
文摘Automated and accurate movie genre classification is crucial for content organization,recommendation systems,and audience targeting in the film industry.Although most existing approaches focus on audiovisual features such as trailers and posters,the text-based classification remains underexplored despite its accessibility and semantic richness.This paper introduces the Genre Attention Model(GAM),a deep learning architecture that integrates transformer models with a hierarchical attention mechanism to extract and leverage contextual information from movie plots formulti-label genre classification.In order to assess its effectiveness,we assessmultiple transformer-based models,including Bidirectional Encoder Representations fromTransformers(BERT),ALite BERT(ALBERT),Distilled BERT(DistilBERT),Robustly Optimized BERT Pretraining Approach(RoBERTa),Efficiently Learning an Encoder that Classifies Token Replacements Accurately(ELECTRA),eXtreme Learning Network(XLNet)and Decodingenhanced BERT with Disentangled Attention(DeBERTa).Experimental results demonstrate the superior performance of DeBERTa-based GAM,which employs a two-tier hierarchical attention mechanism:word-level attention highlights key terms,while sentence-level attention captures critical narrative segments,ensuring a refined and interpretable representation of movie plots.Evaluated on three benchmark datasets Trailers12K,Large Movie Trailer Dataset-9(LMTD-9),and MovieLens37K.GAM achieves micro-average precision scores of 83.63%,83.32%,and 83.34%,respectively,surpassing state-of-the-artmodels.Additionally,GAMis computationally efficient,requiring just 6.10Giga Floating Point Operations Per Second(GFLOPS),making it a scalable and cost-effective solution.These results highlight the growing potential of text-based deep learning models in genre classification and GAM’s effectiveness in improving predictive accuracy while maintaining computational efficiency.With its robust performance,GAM offers a versatile and scalable framework for content recommendation,film indexing,and media analytics,providing an interpretable alternative to traditional audiovisual-based classification techniques.
基金This research was supported in part by Basic Science Research Program through the National Research Foundation of Korea(NRF),funded by the Ministry of Education(NRF-2021R1A6A1A03039493)in part by the NRF grant funded by the Korea government(MSIT)(NRF-2022R1A2C1004401).
文摘Phishing is a type of cybercrime in which cyber-attackers pose themselves as authorized persons or entities and hack the victims’sensitive data.E-mails,instant messages and phone calls are some of the common modes used in cyberattacks.Though the security models are continuously upgraded to prevent cyberattacks,hackers find innovative ways to target the victims.In this background,there is a drastic increase observed in the number of phishing emails sent to potential targets.This scenario necessitates the importance of designing an effective classification model.Though numerous conventional models are available in the literature for proficient classification of phishing emails,the Machine Learning(ML)techniques and the Deep Learning(DL)models have been employed in the literature.The current study presents an Intelligent Cuckoo Search(CS)Optimization Algorithm with a Deep Learning-based Phishing Email Detection and Classification(ICSOA-DLPEC)model.The aim of the proposed ICSOA-DLPEC model is to effectually distinguish the emails as either legitimate or phishing ones.At the initial stage,the pre-processing is performed through three stages such as email cleaning,tokenization and stop-word elimination.Then,the N-gram approach is;moreover,the CS algorithm is applied to extract the useful feature vectors.Moreover,the CS algorithm is employed with the Gated Recurrent Unit(GRU)model to detect and classify phishing emails.Furthermore,the CS algorithm is used to fine-tune the parameters involved in the GRU model.The performance of the proposed ICSOA-DLPEC model was experimentally validated using a benchmark dataset,and the results were assessed under several dimensions.Extensive comparative studies were conducted,and the results confirmed the superior performance of the proposed ICSOA-DLPEC model over other existing approaches.The proposed model achieved a maximum accuracy of 99.72%.
基金The APC was funded by PPPI,University Malaysia Sabah,KK,Sabah,Malaysia,https://www.ums.edu.my.
文摘Bat algorithm(BA)is an eminent meta-heuristic algorithm that has been widely used to solve diverse kinds of optimization problems.BA leverages the echolocation feature of bats produced by imitating the bats’searching behavior.BA faces premature convergence due to its local search capability.Instead of using the standard uniform walk,the Torus walk is viewed as a promising alternative to improve the local search capability.In this work,we proposed an improved variation of BA by applying torus walk to improve diversity and convergence.The proposed.Modern Computerized Bat Algorithm(MCBA)approach has been examined for fifteen well-known benchmark test problems.The finding of our technique shows promising performance as compared to the standard PSO and standard BA.The proposed MCBA,BPA,Standard PSO,and Standard BA have been examined for well-known benchmark test problems and training of the artificial neural network(ANN).We have performed experiments using eight benchmark datasets applied from the worldwide famous machine-learning(ML)repository of UCI.Simulation results have shown that the training of an ANN with MCBA-NN algorithm tops the list considering exactness,with more superiority compared to the traditional methodologies.The MCBA-NN algorithm may be used effectively for data classification and statistical problems in the future.
文摘An intrusion detection system(IDS)becomes an important tool for ensuring security in the network.In recent times,machine learning(ML)and deep learning(DL)models can be applied for the identification of intrusions over the network effectively.To resolve the security issues,this paper presents a new Binary Butterfly Optimization algorithm based on Feature Selection with DRL technique,called BBOFS-DRL for intrusion detection.The proposed BBOFSDRL model mainly accomplishes the recognition of intrusions in the network.To attain this,the BBOFS-DRL model initially designs the BBOFS algorithm based on the traditional butterfly optimization algorithm(BOA)to elect feature subsets.Besides,DRL model is employed for the proper identification and classification of intrusions that exist in the network.Furthermore,beetle antenna search(BAS)technique is applied to tune the DRL parameters for enhanced intrusion detection efficiency.For ensuring the superior intrusion detection outcomes of the BBOFS-DRL model,a wide-ranging experimental analysis is performed against benchmark dataset.The simulation results reported the supremacy of the BBOFS-DRL model over its recent state of art approaches.
文摘The Brain Tumor(BT)is created by an uncontrollable rise of anomalous cells in brain tissue,and it consists of 2 types of cancers they are malignant and benign tumors.The benevolent BT does not affect the neighbouring healthy and normal tissue;however,the malignant could affect the adjacent brain tissues,which results in death.Initial recognition of BT is highly significant to protecting the patient’s life.Generally,the BT can be identified through the magnetic resonance imaging(MRI)scanning technique.But the radiotherapists are not offering effective tumor segmentation in MRI images because of the position and unequal shape of the tumor in the brain.Recently,ML has prevailed against standard image processing techniques.Several studies denote the superiority of machine learning(ML)techniques over standard techniques.Therefore,this study develops novel brain tumor detection and classification model using met heuristic optimization with machine learning(BTDC-MOML)model.To accomplish the detection of brain tumor effectively,a Computer-Aided Design(CAD)model using Machine Learning(ML)technique is proposed in this research manuscript.Initially,the input image pre-processing is performed using Gaborfiltering(GF)based noise removal,contrast enhancement,and skull stripping.Next,mayfly optimization with the Kapur’s thresholding based segmentation process takes place.For feature extraction proposes,local diagonal extreme patterns(LDEP)are exploited.At last,the Extreme Gradient Boosting(XGBoost)model can be used for the BT classification process.The accuracy analysis is performed in terms of Learning accuracy,and the validation accuracy is performed to determine the efficiency of the proposed research work.The experimental validation of the proposed model demonstrates its promising performance over other existing methods.
基金Natural Science Basic Research Plan in Shaanxi Province of China under Grant 2022JM-327 and in part by the CAAI-Huawei MindSpore Academic Open Fund.
文摘Because of its strong ability to solve problems,evolutionary multitask optimization(EMTO)algorithms have been widely studied recently.Evolutionary algorithms have the advantage of fast searching for the optimal solution,but it is easy to fall into local optimum and difficult to generalize.Combining evolutionary multitask algorithms with evolutionary optimization algorithms can be an effective method for solving these problems.Through the implicit parallelism of tasks themselves and the knowledge transfer between tasks,more promising individual algorithms can be generated in the evolution process,which can jump out of the local optimum.How to better combine the two has also been studied more and more.This paper explores the existing evolutionary multitasking theory and improvement scheme in detail.Then,it summarizes the application of EMTO in different scenarios.Finally,according to the existing research,the future research trends and potential exploration directions are revealed.
文摘Internet of Things(IoT)allows several low resources and controlled devices to interconnect,calculate processes and make decisions in the communication network.In the heterogeneous environment for IoT devices,several challenging issues such as energy,storage,efficiency,and security.The design of encryption techniques enables the transmission of the data in the IoT environment in a secured way.The proper selection of optimal keys helps to boost the encryption performance.With this motivation,the study presents a signcryption with quantum chaotic krill herd algorithm for secured data transmission(SCQCKH-SDT)in IoT environment.The proposed SCQCKHSDT technique aims to effectively encrypts the data by the use of optimal keys generated by the CQKH algorithm.The proposed SCQCKH-SDT technique initially employs the signcryption technique for the encryption of data.In order to optimize the secrecy,the optimal key generation process is carried out using Chaotic Krill Herd(CQKH)algorithm.The CQKH algorithm incorporates the concept of quantum computing and chaotic theory into the traditional KH algorithm.The performance validation of the SCQCKH-SDT technique is performed using benchmark dataset.An extensive comparative analysis reported the superior performance of the SCQCKH-SDT technique over the recent approaches.
文摘Numerous methods are analysed in detail to improve task schedulingand data security performance in the cloud environment. The methodsinvolve scheduling according to the factors like makespan, waiting time,cost, deadline, and popularity. However, the methods are inappropriate forachieving higher scheduling performance. Regarding data security, existingmethods use various encryption schemes but introduce significant serviceinterruption. This article sketches a practical Real-time Application CentricTRS (Throughput-Resource utilization–Success) Scheduling with Data Security(RATRSDS) model by considering all these issues in task scheduling anddata security. The method identifies the required resource and their claim timeby receiving the service requests. Further, for the list of resources as services,the method computes throughput support (Thrs) according to the number ofstatements executed and the complete statements of the service. Similarly, themethod computes Resource utilization support (Ruts) according to the idletime on any duty cycle and total servicing time. Also, the method computesthe value of Success support (Sus) according to the number of completions forthe number of allocations. The method estimates the TRS score (ThroughputResource utilization Success) for different resources using all these supportmeasures. According to the value of the TRS score, the services are rankedand scheduled. On the other side, based on the requirement of service requests,the method computes Requirement Support (RS). The selection of service isperformed and allocated. Similarly, choosing the route according to the RouteSupport Measure (RSM) enforced route security. Finally, data security hasgets implemented with a service-based encryption technique. The RATRSDSscheme has claimed higher performance in data security and scheduling.
文摘The impact of a Distributed Denial of Service(DDoS)attack on Soft-ware Defined Networks(SDN)is briefly analyzed.Many approaches to detecting DDoS attacks exist,varying on the feature being considered and the method used.Still,the methods have a deficiency in the performance of detecting DDoS attacks and mitigating them.To improve the performance of SDN,an efficient Real-time Multi-Constrained Adaptive Replication and Traffic Approximation Model(RMCARTAM)is sketched in this article.The RMCARTAM considers different parameters or constraints in running different controllers responsible for handling incoming packets.The model is designed with multiple controllers to handle net-work traffic but can turn the controllers according to requirements.The multi-con-straint adaptive replication model monitors different features of network traffic like rate of packet reception,class-based packet reception and target-specific reception.According to these features,the method estimates the Replication Turn-ing Weight(RTW)based on which triggering controllers are performed.Similarly,the method applies Traffic Approximation(TA)in the detection of DDoS attacks.The detection of a DDoS attack is performed by approximating the incoming traf-fic to any service and using various features like hop count,payload,service fre-quency,and malformed frequency to compute various support measures on bandwidth access,data support,frequency support,malformed support,route sup-port,and so on.Using all these support measures,the method computes the value of legitimate weight to conclude the behavior of any source in identifying the mal-icious node.Identified node details are used in the mitigation of DDoS attacks.The method stimulates the network performance by reducing the power factor by switching the controller according to different factors,which also reduces the cost.In the same way,the proposed model improves the accuracy of detecting DDoS attacks by estimating the features of incoming traffic in different corners.
文摘Reliable transmission is vital to the success of the next generation ofcommunications technologies and Fifth Generation (5G) networks. Many sensitive applications, such as eHealth and medical services, can benefit from a 5G network. The Internet of Medical Things (IoMT) is a new field that fosters themaintenance of trust among various IoMT Device to Device (D2D) modern technologies. In IoMT the medical devices have to be connected through a wirelessnetwork and constantly needs to be self-configured to provide consistent and effi-cient data transmission. The medical devices need to be connected with sophisticated protocols and architecture to handle the synergy of the monitoring devices.Today, one of the commonly used algorithms in D2D communication is the Optimized Link State Routing protocol (OLSR). The OLSR is considerably good ateffectively utilizing the bandwidth and reserving the paths. One of the majorattack against the OLSR is the Node isolation attack, also known as the Gray holedenial of service attack. The Gray hole attack exploits the vulnerabilities presentwith sharing the topological information of the network. The attackers may usethis topological information to maliciously disconnect the target nodes from theexisting network and stops rendering the communication services to the victimnode. Hence, considering the sensitivity and security concerns of the data usedin e-Health applications, these types of attacks must be detected and disabledproactively. In this work, a novel Node Authentication (NA) with OLSR is proposed. The simulation experiments illustrated that the proposed protocol has anexcellent Packet Delivery Ratio, minimal End-End delay, and minimal Packet losswhen compared to the Ad-hoc On-Demand Distance Victor (AODV) protocol andthe proposed authentication scheme was able to protect the OLSR protocol from anode isolation attack.
文摘Prediction of stock market value is highly risky because it is based on the concept of Time Series forecasting system that can be used for investments in a safe environment with minimized chances of loss.The proposed model uses a real time dataset offifteen Stocks as input into the system and based on the data,predicts or forecast future stock prices of different companies belonging to different sectors.The dataset includes approximatelyfifteen companies from different sectors and forecasts their results based on which the user can decide whether to invest in the particular company or not;the forecasting is done for the next quarter.Our model uses 3 main concepts for forecasting results.Thefirst one is for stocks that show periodic change throughout the season,the‘Holt-Winters Triple Exponential Smoothing’.3 basic things taken into conclusion by this algorithm are Base Level,Trend Level and Seasoning Factor.The value of all these are calculated by us and then decomposition of all these factors is done by the Holt-Winters Algorithm.The second concept is‘Recurrent Neural Network’.The specific model of recurrent neural network that is being used is Long-Short Term Memory and it’s the same as the Normal Neural Network,the only difference is that each intermediate cell is a memory cell and retails its value till the next feedback loop.The third concept is Recommendation System whichfilters and predict the rating based on the different factors.