The rapid and increasing growth in the volume and number of cyber threats from malware is not a real danger;the real threat lies in the obfuscation of these cyberattacks,as they constantly change their behavior,making...The rapid and increasing growth in the volume and number of cyber threats from malware is not a real danger;the real threat lies in the obfuscation of these cyberattacks,as they constantly change their behavior,making detection more difficult.Numerous researchers and developers have devoted considerable attention to this topic;however,the research field has not yet been fully saturated with high-quality studies that address these problems.For this reason,this paper presents a novel multi-objective Markov-enhanced adaptive whale optimization(MOMEAWO)cybersecurity model to improve the classification of binary and multi-class malware threats through the proposed MOMEAWO approach.The proposed MOMEAWO cybersecurity model aims to provide an innovative solution for analyzing,detecting,and classifying the behavior of obfuscated malware within their respective families.The proposed model includes three classification types:Binary classification and multi-class classification(e.g.,four families and 16 malware families).To evaluate the performance of this model,we used a recently published dataset called the Canadian Institute for Cybersecurity Malware Memory Analysis(CIC-MalMem-2022)that contains balanced data.The results show near-perfect accuracy in binary classification and high accuracy in multi-class classification compared with related work using the same dataset.展开更多
In recent decades,the proliferation of email communication has markedly escalated,resulting in a concomitant surge in spam emails that congest networks and presenting security risks.This study introduces an innovative...In recent decades,the proliferation of email communication has markedly escalated,resulting in a concomitant surge in spam emails that congest networks and presenting security risks.This study introduces an innovative spam detection method utilizing the Horse Herd Optimization Algorithm(HHOA),designed for binary classification within multi⁃objective framework.The method proficiently identifies essential features,minimizing redundancy and improving classification precision.The suggested HHOA attained an impressive accuracy of 97.21%on the Kaggle email dataset,with precision of 94.30%,recall of 90.50%,and F1⁃score of 92.80%.Compared to conventional techniques,such as Support Vector Machine(93.89%accuracy),Random Forest(96.14%accuracy),and K⁃Nearest Neighbours(92.08%accuracy),HHOA exhibited enhanced performance with reduced computing complexity.The suggested method demonstrated enhanced feature selection efficiency,decreasing the number of selected features while maintaining high classification accuracy.The results underscore the efficacy of HHOA in spam identification and indicate its potential for further applications in practical email filtering systems.展开更多
Visual diagnosis of skin cancer is challenging due to subtle inter-class similarities,variations in skin texture,the presence of hair,and inconsistent illumination.Deep learning models have shown promise in assisting ...Visual diagnosis of skin cancer is challenging due to subtle inter-class similarities,variations in skin texture,the presence of hair,and inconsistent illumination.Deep learning models have shown promise in assisting early detection,yet their performance is often limited by the severe class imbalance present in dermoscopic datasets.This paper proposes CANNSkin,a skin cancer classification framework that integrates a convolutional autoencoder with latent-space oversampling to address this imbalance.The autoencoder is trained to reconstruct lesion images,and its latent embeddings are used as features for classification.To enhance minority-class representation,the Synthetic Minority Oversampling Technique(SMOTE)is applied directly to the latent vectors before classifier training.The encoder and classifier are first trained independently and later fine-tuned end-to-end.On the HAM10000 dataset,CANNSkin achieves an accuracy of 93.01%,a macro-F1 of 88.54%,and an ROC–AUC of 98.44%,demonstrating strong robustness across ten test subsets.Evaluation on the more complex ISIC 2019 dataset further confirms the model’s effectiveness,where CANNSkin achieves 94.27%accuracy,93.95%precision,94.09%recall,and 99.02%F1-score,supported by high reconstruction fidelity(PSNR 35.03 dB,SSIM 0.86).These results demonstrate the effectiveness of our proposed latent-space balancing and fine-tuned representation learning as a new benchmark method for robust and accurate skin cancer classification across heterogeneous datasets.展开更多
Autonomous connected vehicles(ACV)involve advanced control strategies to effectively balance safety,efficiency,energy consumption,and passenger comfort.This research introduces a deep reinforcement learning(DRL)-based...Autonomous connected vehicles(ACV)involve advanced control strategies to effectively balance safety,efficiency,energy consumption,and passenger comfort.This research introduces a deep reinforcement learning(DRL)-based car-following(CF)framework employing the Deep Deterministic Policy Gradient(DDPG)algorithm,which integrates a multi-objective reward function that balances the four goals while maintaining safe policy learning.Utilizing real-world driving data from the highD dataset,the proposed model learns adaptive speed control policies suitable for dynamic traffic scenarios.The performance of the DRL-based model is evaluated against a traditional model predictive control-adaptive cruise control(MPC-ACC)controller.Results show that theDRLmodel significantly enhances safety,achieving zero collisions and a higher average time-to-collision(TTC)of 8.45 s,compared to 5.67 s for MPC and 6.12 s for human drivers.For efficiency,the model demonstrates 89.2% headway compliance and maintains speed tracking errors below 1.2 m/s in 90% of cases.In terms of energy optimization,the proposed approach reduces fuel consumption by 5.4% relative to MPC.Additionally,it enhances passenger comfort by lowering jerk values by 65%,achieving 0.12 m/s3 vs.0.34 m/s3 for human drivers.A multi-objective reward function is integrated to ensure stable policy convergence while simultaneously balancing the four key performance metrics.Moreover,the findings underscore the potential of DRL in advancing autonomous vehicle control,offering a robust and sustainable solution for safer,more efficient,and more comfortable transportation systems.展开更多
Task scheduling in cloud computing is a multi-objective optimization problem,often involving conflicting objectives such as minimizing execution time,reducing operational cost,and maximizing resource utilization.Howev...Task scheduling in cloud computing is a multi-objective optimization problem,often involving conflicting objectives such as minimizing execution time,reducing operational cost,and maximizing resource utilization.However,traditional approaches frequently rely on single-objective optimization methods which are insufficient for capturing the complexity of such problems.To address this limitation,we introduce MDMOSA(Multi-objective Dwarf Mongoose Optimization with Simulated Annealing),a hybrid that integrates multi-objective optimization for efficient task scheduling in Infrastructure-as-a-Service(IaaS)cloud environments.MDMOSA harmonizes the exploration capabilities of the biologically inspired Dwarf Mongoose Optimization(DMO)with the exploitation strengths of Simulated Annealing(SA),achieving a balanced search process.The algorithm aims to optimize task allocation by reducing makespan and financial cost while improving system resource utilization.We evaluate MDMOSA through extensive simulations using the real-world Google Cloud Jobs(GoCJ)dataset within the CloudSim environment.Comparative analysis against benchmarked algorithms such as SMOACO,MOTSGWO,and MFPAGWO reveals that MDMOSA consistently achieves superior performance in terms of scheduling efficiency,cost-effectiveness,and scalability.These results confirm the potential of MDMOSA as a robust and adaptable solution for resource scheduling in dynamic and heterogeneous cloud computing infrastructures.展开更多
Deployable Composite Thin-Walled Structures(DCTWS)are widely used in space applications due to their ability to compactly fold and self-deploy in orbit,enabled by cutouts.Cutout design is crucial for balancing structu...Deployable Composite Thin-Walled Structures(DCTWS)are widely used in space applications due to their ability to compactly fold and self-deploy in orbit,enabled by cutouts.Cutout design is crucial for balancing structural rigidity and flexibility,ensuring material integrity during large deformations,and providing adequate load-bearing capacity and stability once deployed.Most research has focused on optimizing cutout size and shape,while topology optimization offers a broader design space.However,the anisotropic properties of woven composite laminates,complex failure criteria,and multi-performance optimization needs have limited the exploration of topology optimization in this field.This work derives the sensitivities of bending stiffness,critical buckling load,and the failure index of woven composite materials with respect to element density,and formulates both single-objective and multi-objective topology optimization models using a linear weighted aggregation approach.The developed method was integrated with the commercial finite element software ABAQUS via a Python script,allowing efficient application to cutout design in various DCTWS configurations to maximize bending stiffness and critical buckling load under material failure constraints.Optimization of a classical tubular hinge resulted in improvements of 107.7%in bending stiffness and 420.5%in critical buckling load compared to level-set topology optimization results reported in the literature,validating the effectiveness of the approach.To facilitate future research and encourage the broader adoption of topology optimization techniques in DCTWS design,the source code for this work is made publicly available via a Git Hub link:https://github.com/jinhao-ok1/Topo-for-DCTWS.git.展开更多
In a wide range of engineering applications,complex constrained multi-objective optimization problems(CMOPs)present significant challenges,as the complexity of constraints often hampers algorithmic convergence and red...In a wide range of engineering applications,complex constrained multi-objective optimization problems(CMOPs)present significant challenges,as the complexity of constraints often hampers algorithmic convergence and reduces population diversity.To address these challenges,we propose a novel algorithm named Constraint IntensityDriven Evolutionary Multitasking(CIDEMT),which employs a two-stage,tri-task framework to dynamically integrates problem structure and knowledge transfer.In the first stage,three cooperative tasks are designed to explore the Constrained Pareto Front(CPF),the Unconstrained Pareto Front(UPF),and theε-relaxed constraint boundary,respectively.A CPF-UPF relationship classifier is employed to construct a problem-type-aware evolutionary strategy pool.At the end of the first stage,each task selects strategies from this strategy pool based on the specific type of problem,thereby guiding the subsequent evolutionary process.In the second stage,while each task continues to evolve,aτ-driven knowledge transfer mechanism is introduced to selectively incorporate effective solutions across tasks.enhancing the convergence and feasibility of the main task.Extensive experiments conducted on 32 benchmark problems from three test suites(LIRCMOP,DASCMOP,and DOC)demonstrate that CIDEMT achieves the best Inverted Generational Distance(IGD)values on 24 problems and the best Hypervolume values(HV)on 22 problems.Furthermore,CIDEMT significantly outperforms six state-of-the-art constrained multi-objective evolutionary algorithms(CMOEAs).These results confirm CIDEMT’s superiority in promoting convergence,diversity,and robustness in solving complex CMOPs.展开更多
Community detection is one of the most fundamental applications in understanding the structure of complicated networks.Furthermore,it is an important approach to identifying closely linked clusters of nodes that may r...Community detection is one of the most fundamental applications in understanding the structure of complicated networks.Furthermore,it is an important approach to identifying closely linked clusters of nodes that may represent underlying patterns and relationships.Networking structures are highly sensitive in social networks,requiring advanced techniques to accurately identify the structure of these communities.Most conventional algorithms for detecting communities perform inadequately with complicated networks.In addition,they miss out on accurately identifying clusters.Since single-objective optimization cannot always generate accurate and comprehensive results,as multi-objective optimization can.Therefore,we utilized two objective functions that enable strong connections between communities and weak connections between them.In this study,we utilized the intra function,which has proven effective in state-of-the-art research studies.We proposed a new inter-function that has demonstrated its effectiveness by making the objective of detecting external connections between communities is to make them more distinct and sparse.Furthermore,we proposed a Multi-Objective community strength enhancement algorithm(MOCSE).The proposed algorithm is based on the framework of the Multi-Objective Evolutionary Algorithm with Decomposition(MOEA/D),integrated with a new heuristic mutation strategy,community strength enhancement(CSE).The results demonstrate that the model is effective in accurately identifying community structures while also being computationally efficient.The performance measures used to evaluate the MOEA/D algorithm in our work are normalized mutual information(NMI)and modularity(Q).It was tested using five state-of-the-art algorithms on social networks,comprising real datasets(Zachary,Dolphin,Football,Krebs,SFI,Jazz,and Netscience),as well as twenty synthetic datasets.These results provide the robustness and practical value of the proposed algorithm in multi-objective community identification.展开更多
Arrhythmias are a frequently occurring phenomenon in clinical practice,but how to accurately dis-tinguish subtle rhythm abnormalities remains an ongoing difficulty faced by the entire research community when conductin...Arrhythmias are a frequently occurring phenomenon in clinical practice,but how to accurately dis-tinguish subtle rhythm abnormalities remains an ongoing difficulty faced by the entire research community when conducting ECG-based studies.From a review of existing studies,two main factors appear to contribute to this problem:the uneven distribution of arrhythmia classes and the limited expressiveness of features learned by current models.To overcome these limitations,this study proposes a dual-path multimodal framework,termed DM-EHC(Dual-Path Multimodal ECG Heartbeat Classifier),for ECG-based heartbeat classification.The proposed framework links 1D ECG temporal features with 2D time–frequency features.By setting up the dual paths described above,the model can process more dimensions of feature information.The MIT-BIH arrhythmia database was selected as the baseline dataset for the experiments.Experimental results show that the proposed method outperforms single modalities and performs better for certain specific types of arrhythmias.The model achieved mean precision,recall,and F1 score of 95.14%,92.26%,and 93.65%,respectively.These results indicate that the framework is robust and has potential value in automated arrhythmia classification.展开更多
Near-Earth objects are important not only in studying the early formation of the Solar System,but also because they pose a serious hazard to humanity when they make close approaches to the Earth.Study of their physica...Near-Earth objects are important not only in studying the early formation of the Solar System,but also because they pose a serious hazard to humanity when they make close approaches to the Earth.Study of their physical properties can provide useful information on their origin,evolution,and hazard to human beings.However,it remains challenging to investigate small,newly discovered,near-Earth objects because of our limited observational window.This investigation seeks to determine the visible colors of near-Earth asteroids(NEAs),perform an initial taxonomic classification based on visible colors and analyze possible correlations between the distribution of taxonomic classification and asteroid size or orbital parameters.Observations were performed in the broadband BVRI Johnson−Cousins photometric system,applied to images from the Yaoan High Precision Telescope and the 1.88 m telescope at the Kottamia Astronomical Observatory.We present new photometric observations of 84 near-Earth asteroids,and classify 80 of them taxonomically,based on their photometric colors.We find that nearly half(46.3%)of the objects in our sample can be classified as S-complex,26.3%as C-complex,6%as D-complex,and 15.0%as X-complex;the remaining belong to the A-or V-types.Additionally,we identify three P-type NEAs in our sample,according to the Tholen scheme.The fractional abundances of the C/X-complex members with absolute magnitude H≥17.0 were more than twice as large as those with H<17.0.However,the fractions of C-and S-complex members with diameters≤1 km and>1 km are nearly equal,while X-complex members tend to have sub-kilometer diameters.In our sample,the C/D-complex objects are predominant among those with a Jovian Tisserand parameter of T_(J)<3.1.These bodies could have a cometary origin.C-and S-complex members account for a considerable proportion of the asteroids that are potentially hazardous.展开更多
Graph Neural Networks(GNNs)have proven highly effective for graph classification across diverse fields such as social networks,bioinformatics,and finance,due to their capability to learn complex graph structures.Howev...Graph Neural Networks(GNNs)have proven highly effective for graph classification across diverse fields such as social networks,bioinformatics,and finance,due to their capability to learn complex graph structures.However,despite their success,GNNs remain vulnerable to adversarial attacks that can significantly degrade their classification accuracy.Existing adversarial attack strategies primarily rely on label information to guide the attacks,which limits their applicability in scenarios where such information is scarce or unavailable.This paper introduces an innovative unsupervised attack method for graph classification,which operates without relying on label information,thereby enhancing its applicability in a broad range of scenarios.Specifically,our method first leverages a graph contrastive learning loss to learn high-quality graph embeddings by comparing different stochastic augmented views of the graphs.To effectively perturb the graphs,we then introduce an implicit estimator that measures the impact of various modifications on graph structures.The proposed strategy identifies and flips edges with the top-K highest scores,determined by the estimator,to maximize the degradation of the model’s performance.In addition,to defend against such attack,we propose a lightweight regularization-based defense mechanism that is specifically tailored to mitigate the structural perturbations introduced by our attack strategy.It enhances model robustness by enforcing embedding consistency and edge-level smoothness during training.We conduct experiments on six public TU graph classification datasets:NCI1,NCI109,Mutagenicity,ENZYMES,COLLAB,and DBLP_v1,to evaluate the effectiveness of our attack and defense strategies.Under an attack budget of 3,the maximum reduction in model accuracy reaches 6.67%on the Graph Convolutional Network(GCN)and 11.67%on the Graph Attention Network(GAT)across different datasets,indicating that our unsupervised method induces degradation comparable to state-of-the-art supervised attacks.Meanwhile,our defense achieves the highest accuracy recovery of 3.89%(GCN)and 5.00%(GAT),demonstrating improved robustness against structural perturbations.展开更多
Vehicle Edge Computing(VEC)and Cloud Computing(CC)significantly enhance the processing efficiency of delay-sensitive and computation-intensive applications by offloading compute-intensive tasks from resource-constrain...Vehicle Edge Computing(VEC)and Cloud Computing(CC)significantly enhance the processing efficiency of delay-sensitive and computation-intensive applications by offloading compute-intensive tasks from resource-constrained onboard devices to nearby Roadside Unit(RSU),thereby achieving lower delay and energy consumption.However,due to the limited storage capacity and energy budget of RSUs,it is challenging to meet the demands of the highly dynamic Internet of Vehicles(IoV)environment.Therefore,determining reasonable service caching and computation offloading strategies is crucial.To address this,this paper proposes a joint service caching scheme for cloud-edge collaborative IoV computation offloading.By modeling the dynamic optimization problem using Markov Decision Processes(MDP),the scheme jointly optimizes task delay,energy consumption,load balancing,and privacy entropy to achieve better quality of service.Additionally,a dynamic adaptive multi-objective deep reinforcement learning algorithm is proposed.Each Double Deep Q-Network(DDQN)agent obtains rewards for different objectives based on distinct reward functions and dynamically updates the objective weights by learning the value changes between objectives using Radial Basis Function Networks(RBFN),thereby efficiently approximating the Pareto-optimal decisions for multiple objectives.Extensive experiments demonstrate that the proposed algorithm can better coordinate the three-tier computing resources of cloud,edge,and vehicles.Compared to existing algorithms,the proposed method reduces task delay and energy consumption by 10.64%and 5.1%,respectively.展开更多
Spaceborne antennas are essential for remote sensing,deep-space communication,and Earth observation,yet their trajectory planning is complicated by nonlinear base-manipulator coupling and antenna flexibility.To addres...Spaceborne antennas are essential for remote sensing,deep-space communication,and Earth observation,yet their trajectory planning is complicated by nonlinear base-manipulator coupling and antenna flexibility.To address these challenges,this paper proposes a multi-objective trajectory optimization framework.The system dynamics capture both nonlinear rigid-flexible coupling and antenna deformation through a reduced-order formulation.To enhance discretization efficiency,a predictive-terminal hp-adaptive pseudospectral method is employed,assigning collocation density based on task-phase characteristics:finer resolution is applied to dynamic segments requiring higher accuracy,especially near the terminal phase.This enables efficient transcription of the continuous-time problem into a Nonlinear Programming Problem(NLP).The resulting NLP is then solved using a multi-objective optimization strategy based on the nondominated sorting genetic algorithm II,which explores trade-offs among antenna pointing accuracy,energy consumption,and structural vibration.Numerical results demonstrate that the proposed method achieves a reduction of approximately 14.0% in control energy and 41.8%in peak actuation compared to a GPOPS-II baseline,while significantly enhancing vibration suppression.The resulting Pareto front reveals structured trade-offs and clustered solutions,offering robust and diverse options for precision,low-disturbance mission planning.展开更多
Skin diseases affect millions worldwide.Early detection is key to preventing disfigurement,lifelong disability,or death.Dermoscopic images acquired in primary-care settings show high intra-class visual similarity and ...Skin diseases affect millions worldwide.Early detection is key to preventing disfigurement,lifelong disability,or death.Dermoscopic images acquired in primary-care settings show high intra-class visual similarity and severe class imbalance,and occasional imaging artifacts can create ambiguity for state-of-the-art convolutional neural networks(CNNs).We frame skin lesion recognition as graph-based reasoning and,to ensure fair evaluation and avoid data leakage,adopt a strict lesion-level partitioning strategy.Each image is first over-segmented using SLIC(Simple Linear Iterative Clustering)to produce perceptually homogeneous superpixels.These superpixels form the nodes of a region-adjacency graph whose edges encode spatial continuity.Node attributes are 1280-dimensional embeddings extracted with a lightweight yet expressive EfficientNet-B0 backbone,providing strong representational power at modest computational cost.The resulting graphs are processed by a five-layer Graph Attention Network(GAT)that learns to weight inter-node relationships dynamically and aggregates multi-hop context before classifying lesions into seven classes with a log-softmax output.Extensive experiments on the DermaMNIST benchmark show the proposed pipeline achieves 88.35%accuracy and 98.04%AUC,outperforming contemporary CNNs,AutoML approaches,and alternative graph neural networks.An ablation study indicates EfficientNet-B0 produces superior node descriptors compared with ResNet-18 and DenseNet,and that roughly five GAT layers strike a good balance between being too shallow and over-deep while avoiding oversmoothing.The method requires no data augmentation or external metadata,making it a drop-in upgrade for clinical computer-aided diagnosis systems.展开更多
With the recent increase in data volume and diversity,traditional text representation techniques are struggling to capture context,particularly in environments with sparse data.To address these challenges,this study p...With the recent increase in data volume and diversity,traditional text representation techniques are struggling to capture context,particularly in environments with sparse data.To address these challenges,this study proposes a new model,the Masked Joint Representation Model(MJRM).MJRM approximates the original hypothesis by leveraging multiple elements in a limited context.It dynamically adapts to changes in characteristics based on data distribution through three main components.First,masking-based representation learning,termed selective dynamic masking,integrates topic modeling and sentiment clustering to generate and train multiple instances across different data subsets,whose predictions are then aggregated with optimized weights.This design alleviates sparsity,suppresses noise,and preserves contextual structures.Second,regularization-based improvements are applied.Third,techniques for addressing sparse data are used to perform final inference.As a result,MJRM improves performance by up to 4%compared to existing AI techniques.In our experiments,we analyzed the contribution of each factor,demonstrating that masking,dynamic learning,and aggregating multiple instances complement each other to improve performance.This demonstrates that a masking-based multi-learning strategy is effective for context-aware sparse text classification,and can be useful even in challenging situations such as data shortage or data distribution variations.We expect that the approach can be extended to diverse fields such as sentiment analysis,spam filtering,and domain-specific document classification.展开更多
Classifying job offers into occupational categories is a fundamental task in human resource information systems,as it improves and streamlines indexing,search,and matching between openings and job seekers.Comprehensiv...Classifying job offers into occupational categories is a fundamental task in human resource information systems,as it improves and streamlines indexing,search,and matching between openings and job seekers.Comprehensive occupational databases such as O∗NET or ESCO provide detailed taxonomies of interrelated positions that can be leveraged to align the textual content of postings with occupational categories,thereby facilitating standardization,cross-system interoperability,and access to metadata for each occupation(e.g.,tasks,knowledge,skills,and abilities).In this work,we explore the effectiveness of fine-tuning existing language models(LMs)to classify job offers with occupational descriptors from O∗NET.This enables a more precise assessment of candidate suitability by identifying the specific knowledge and skills required for each position,and helps automate recruitment processes by mitigating human bias and subjectivity in candidate selection.We evaluate three representative BERT-like models:BERT,RoBERTa,and DeBERTa.BERT serves as the baseline encoder-only architecture;RoBERTa incorporates advances in pretraining objectives and data scale;and DeBERTa introduces architectural improvements through disentangled attention mechanisms.The best performance was achieved with the DeBERTa model,although the other models also produced strong results,and no statistically significant differences were observed acrossmodels.We also find that these models typically reach optimal performance after only a few training epochs,and that training with smaller,balanced datasets is effective.Consequently,comparable results can be obtained with models that require fewer computational resources and less training time,facilitating deployment and practical use.展开更多
Container transportation is pivotal in global trade due to its efficiency,safety,and cost-effectiveness.However,structural defects—particularly in grapple slots—can result in cargo damage,financial loss,and elevated...Container transportation is pivotal in global trade due to its efficiency,safety,and cost-effectiveness.However,structural defects—particularly in grapple slots—can result in cargo damage,financial loss,and elevated safety risks,including container drops during lifting operations.Timely and accurate inspection before and after transit is therefore essential.Traditional inspection methods rely heavily on manual observation of internal and external surfaces,which are time-consuming,resource-intensive,and prone to subjective errors.Container roofs pose additional challenges due to limited visibility,while grapple slots are especially vulnerable to wear from frequent use.This study proposes a two-stage automated detection framework targeting defects in container roof grapple slots.In the first stage,YOLOv7 is employed to localize grapple slot regions with high precision.In the second stage,ResNet50 classifies the extracted slots as either intact or defective.The results from both stages are integrated into a human-machine interface for real-time visualization and user verification.Experimental evaluations demonstrate that YOLOv7 achieves a 99%detection rate at 100 frames per second(FPS),while ResNet50 attains 87%classification accuracy at 34 FPS.Compared to some state of the arts,the proposed system offers significant speed,reliability,and usability improvements,enabling efficient defect identification and visual reconfirmation via the interface.展开更多
Recently,large-scale deep learning models have been increasingly adopted for point cloud classification.However,thesemethods typically require collecting extensive datasets frommultiple clients,which may lead to priva...Recently,large-scale deep learning models have been increasingly adopted for point cloud classification.However,thesemethods typically require collecting extensive datasets frommultiple clients,which may lead to privacy leaks.Federated learning provides an effective solution to data leakage by eliminating the need for data transmission,relying instead on the exchange of model parameters.However,the uneven distribution of client data can still affect the model’s ability to generalize effectively.To address these challenges,we propose a new framework for point cloud classification called Federated Dynamic Aggregation Selection Strategy-based Multi-Receptive Field Fusion Classification Framework(FDASS-MRFCF).Specifically,we tackle these challenges with two key innovations:(1)During the client local training phase,we propose a Multi-Receptive Field Fusion Classification Model(MRFCM),which captures local and global structures in point cloud data through dynamic convolution and multi-scale feature fusion,enhancing the robustness of point cloud classification.(2)In the server aggregation phase,we introduce a Federated Dynamic Aggregation Selection Strategy(FDASS),which employs a hybrid strategy to average client model parameters,skip aggregation,or reallocate local models to different clients,thereby balancing global consistency and local diversity.We evaluate our framework using the ModelNet40 and ShapeNetPart benchmarks,demonstrating its effectiveness.The proposed method is expected to significantly advance the field of point cloud classification in a secure environment.展开更多
Evaluating the adversarial robustness of classification algorithms in machine learning is a crucial domain.However,current methods lack measurable and interpretable metrics.To address this issue,this paper introduces ...Evaluating the adversarial robustness of classification algorithms in machine learning is a crucial domain.However,current methods lack measurable and interpretable metrics.To address this issue,this paper introduces a visual evaluation index named confidence centroid skewing quadrilateral,which is based on a classification confidence-based confusion matrix,offering a quantitative and visual comparison of the adversarial robustness among different classification algorithms,and enhances intuitiveness and interpretability of attack impacts.We first conduct a validity test and sensitive analysis of the method.Then,prove its effectiveness through the experiments of five classification algorithms including artificial neural network(ANN),logistic regression(LR),support vector machine(SVM),convolutional neural network(CNN)and transformer against three adversarial attacks such as fast gradient sign method(FGSM),DeepFool,and projected gradient descent(PGD)attack.展开更多
Legal case classification involves the categorization of legal documents into predefined categories,which facilitates legal information retrieval and case management.However,real-world legal datasets often suffer from...Legal case classification involves the categorization of legal documents into predefined categories,which facilitates legal information retrieval and case management.However,real-world legal datasets often suffer from class imbalances due to the uneven distribution of case types across legal domains.This leads to biased model performance,in the form of high accuracy for overrepresented categories and underperformance for minority classes.To address this issue,in this study,we propose a data augmentation method that masks unimportant terms within a document selectively while preserving key terms fromthe perspective of the legal domain.This approach enhances data diversity and improves the generalization capability of conventional models.Our experiments demonstrate consistent improvements achieved by the proposed augmentation strategy in terms of accuracy and F1 score across all models,validating the effectiveness of the proposed method in legal case classification.展开更多
文摘The rapid and increasing growth in the volume and number of cyber threats from malware is not a real danger;the real threat lies in the obfuscation of these cyberattacks,as they constantly change their behavior,making detection more difficult.Numerous researchers and developers have devoted considerable attention to this topic;however,the research field has not yet been fully saturated with high-quality studies that address these problems.For this reason,this paper presents a novel multi-objective Markov-enhanced adaptive whale optimization(MOMEAWO)cybersecurity model to improve the classification of binary and multi-class malware threats through the proposed MOMEAWO approach.The proposed MOMEAWO cybersecurity model aims to provide an innovative solution for analyzing,detecting,and classifying the behavior of obfuscated malware within their respective families.The proposed model includes three classification types:Binary classification and multi-class classification(e.g.,four families and 16 malware families).To evaluate the performance of this model,we used a recently published dataset called the Canadian Institute for Cybersecurity Malware Memory Analysis(CIC-MalMem-2022)that contains balanced data.The results show near-perfect accuracy in binary classification and high accuracy in multi-class classification compared with related work using the same dataset.
文摘In recent decades,the proliferation of email communication has markedly escalated,resulting in a concomitant surge in spam emails that congest networks and presenting security risks.This study introduces an innovative spam detection method utilizing the Horse Herd Optimization Algorithm(HHOA),designed for binary classification within multi⁃objective framework.The method proficiently identifies essential features,minimizing redundancy and improving classification precision.The suggested HHOA attained an impressive accuracy of 97.21%on the Kaggle email dataset,with precision of 94.30%,recall of 90.50%,and F1⁃score of 92.80%.Compared to conventional techniques,such as Support Vector Machine(93.89%accuracy),Random Forest(96.14%accuracy),and K⁃Nearest Neighbours(92.08%accuracy),HHOA exhibited enhanced performance with reduced computing complexity.The suggested method demonstrated enhanced feature selection efficiency,decreasing the number of selected features while maintaining high classification accuracy.The results underscore the efficacy of HHOA in spam identification and indicate its potential for further applications in practical email filtering systems.
基金supported and funded by the Deanship of Scientific Research at Imam Mohammad Ibn Saud Islamic University(IMSIU)(grant number IMSIU-DDRSP2601).
文摘Visual diagnosis of skin cancer is challenging due to subtle inter-class similarities,variations in skin texture,the presence of hair,and inconsistent illumination.Deep learning models have shown promise in assisting early detection,yet their performance is often limited by the severe class imbalance present in dermoscopic datasets.This paper proposes CANNSkin,a skin cancer classification framework that integrates a convolutional autoencoder with latent-space oversampling to address this imbalance.The autoencoder is trained to reconstruct lesion images,and its latent embeddings are used as features for classification.To enhance minority-class representation,the Synthetic Minority Oversampling Technique(SMOTE)is applied directly to the latent vectors before classifier training.The encoder and classifier are first trained independently and later fine-tuned end-to-end.On the HAM10000 dataset,CANNSkin achieves an accuracy of 93.01%,a macro-F1 of 88.54%,and an ROC–AUC of 98.44%,demonstrating strong robustness across ten test subsets.Evaluation on the more complex ISIC 2019 dataset further confirms the model’s effectiveness,where CANNSkin achieves 94.27%accuracy,93.95%precision,94.09%recall,and 99.02%F1-score,supported by high reconstruction fidelity(PSNR 35.03 dB,SSIM 0.86).These results demonstrate the effectiveness of our proposed latent-space balancing and fine-tuned representation learning as a new benchmark method for robust and accurate skin cancer classification across heterogeneous datasets.
基金the Hebei Province Science and Technology Plan Project(19221909D)rincess Nourah bint Abdulrahman University Researchers Supporting Project number(PNURSP2025R308),Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.
文摘Autonomous connected vehicles(ACV)involve advanced control strategies to effectively balance safety,efficiency,energy consumption,and passenger comfort.This research introduces a deep reinforcement learning(DRL)-based car-following(CF)framework employing the Deep Deterministic Policy Gradient(DDPG)algorithm,which integrates a multi-objective reward function that balances the four goals while maintaining safe policy learning.Utilizing real-world driving data from the highD dataset,the proposed model learns adaptive speed control policies suitable for dynamic traffic scenarios.The performance of the DRL-based model is evaluated against a traditional model predictive control-adaptive cruise control(MPC-ACC)controller.Results show that theDRLmodel significantly enhances safety,achieving zero collisions and a higher average time-to-collision(TTC)of 8.45 s,compared to 5.67 s for MPC and 6.12 s for human drivers.For efficiency,the model demonstrates 89.2% headway compliance and maintains speed tracking errors below 1.2 m/s in 90% of cases.In terms of energy optimization,the proposed approach reduces fuel consumption by 5.4% relative to MPC.Additionally,it enhances passenger comfort by lowering jerk values by 65%,achieving 0.12 m/s3 vs.0.34 m/s3 for human drivers.A multi-objective reward function is integrated to ensure stable policy convergence while simultaneously balancing the four key performance metrics.Moreover,the findings underscore the potential of DRL in advancing autonomous vehicle control,offering a robust and sustainable solution for safer,more efficient,and more comfortable transportation systems.
文摘Task scheduling in cloud computing is a multi-objective optimization problem,often involving conflicting objectives such as minimizing execution time,reducing operational cost,and maximizing resource utilization.However,traditional approaches frequently rely on single-objective optimization methods which are insufficient for capturing the complexity of such problems.To address this limitation,we introduce MDMOSA(Multi-objective Dwarf Mongoose Optimization with Simulated Annealing),a hybrid that integrates multi-objective optimization for efficient task scheduling in Infrastructure-as-a-Service(IaaS)cloud environments.MDMOSA harmonizes the exploration capabilities of the biologically inspired Dwarf Mongoose Optimization(DMO)with the exploitation strengths of Simulated Annealing(SA),achieving a balanced search process.The algorithm aims to optimize task allocation by reducing makespan and financial cost while improving system resource utilization.We evaluate MDMOSA through extensive simulations using the real-world Google Cloud Jobs(GoCJ)dataset within the CloudSim environment.Comparative analysis against benchmarked algorithms such as SMOACO,MOTSGWO,and MFPAGWO reveals that MDMOSA consistently achieves superior performance in terms of scheduling efficiency,cost-effectiveness,and scalability.These results confirm the potential of MDMOSA as a robust and adaptable solution for resource scheduling in dynamic and heterogeneous cloud computing infrastructures.
基金supported by the National Natural Science Foundation of China(No.12202295)the International(Regional)Cooperation and Exchange Projects of the National Natural Science Foundation of China(No.W2421002)+2 种基金the Sichuan Science and Technology Program(No.2025ZNSFSC0845)Zhejiang Provincial Natural Science Foundation of China(No.ZCLZ24A0201)the Fundamental Research Funds for the Provincial Universities of Zhejiang(No.GK249909299001-004)。
文摘Deployable Composite Thin-Walled Structures(DCTWS)are widely used in space applications due to their ability to compactly fold and self-deploy in orbit,enabled by cutouts.Cutout design is crucial for balancing structural rigidity and flexibility,ensuring material integrity during large deformations,and providing adequate load-bearing capacity and stability once deployed.Most research has focused on optimizing cutout size and shape,while topology optimization offers a broader design space.However,the anisotropic properties of woven composite laminates,complex failure criteria,and multi-performance optimization needs have limited the exploration of topology optimization in this field.This work derives the sensitivities of bending stiffness,critical buckling load,and the failure index of woven composite materials with respect to element density,and formulates both single-objective and multi-objective topology optimization models using a linear weighted aggregation approach.The developed method was integrated with the commercial finite element software ABAQUS via a Python script,allowing efficient application to cutout design in various DCTWS configurations to maximize bending stiffness and critical buckling load under material failure constraints.Optimization of a classical tubular hinge resulted in improvements of 107.7%in bending stiffness and 420.5%in critical buckling load compared to level-set topology optimization results reported in the literature,validating the effectiveness of the approach.To facilitate future research and encourage the broader adoption of topology optimization techniques in DCTWS design,the source code for this work is made publicly available via a Git Hub link:https://github.com/jinhao-ok1/Topo-for-DCTWS.git.
基金supported by the National Natural Science Foundation of China under Grant No.61972040the Science and Technology Research and Development Project funded by China Railway Material Trade Group Luban Company.
文摘In a wide range of engineering applications,complex constrained multi-objective optimization problems(CMOPs)present significant challenges,as the complexity of constraints often hampers algorithmic convergence and reduces population diversity.To address these challenges,we propose a novel algorithm named Constraint IntensityDriven Evolutionary Multitasking(CIDEMT),which employs a two-stage,tri-task framework to dynamically integrates problem structure and knowledge transfer.In the first stage,three cooperative tasks are designed to explore the Constrained Pareto Front(CPF),the Unconstrained Pareto Front(UPF),and theε-relaxed constraint boundary,respectively.A CPF-UPF relationship classifier is employed to construct a problem-type-aware evolutionary strategy pool.At the end of the first stage,each task selects strategies from this strategy pool based on the specific type of problem,thereby guiding the subsequent evolutionary process.In the second stage,while each task continues to evolve,aτ-driven knowledge transfer mechanism is introduced to selectively incorporate effective solutions across tasks.enhancing the convergence and feasibility of the main task.Extensive experiments conducted on 32 benchmark problems from three test suites(LIRCMOP,DASCMOP,and DOC)demonstrate that CIDEMT achieves the best Inverted Generational Distance(IGD)values on 24 problems and the best Hypervolume values(HV)on 22 problems.Furthermore,CIDEMT significantly outperforms six state-of-the-art constrained multi-objective evolutionary algorithms(CMOEAs).These results confirm CIDEMT’s superiority in promoting convergence,diversity,and robustness in solving complex CMOPs.
文摘Community detection is one of the most fundamental applications in understanding the structure of complicated networks.Furthermore,it is an important approach to identifying closely linked clusters of nodes that may represent underlying patterns and relationships.Networking structures are highly sensitive in social networks,requiring advanced techniques to accurately identify the structure of these communities.Most conventional algorithms for detecting communities perform inadequately with complicated networks.In addition,they miss out on accurately identifying clusters.Since single-objective optimization cannot always generate accurate and comprehensive results,as multi-objective optimization can.Therefore,we utilized two objective functions that enable strong connections between communities and weak connections between them.In this study,we utilized the intra function,which has proven effective in state-of-the-art research studies.We proposed a new inter-function that has demonstrated its effectiveness by making the objective of detecting external connections between communities is to make them more distinct and sparse.Furthermore,we proposed a Multi-Objective community strength enhancement algorithm(MOCSE).The proposed algorithm is based on the framework of the Multi-Objective Evolutionary Algorithm with Decomposition(MOEA/D),integrated with a new heuristic mutation strategy,community strength enhancement(CSE).The results demonstrate that the model is effective in accurately identifying community structures while also being computationally efficient.The performance measures used to evaluate the MOEA/D algorithm in our work are normalized mutual information(NMI)and modularity(Q).It was tested using five state-of-the-art algorithms on social networks,comprising real datasets(Zachary,Dolphin,Football,Krebs,SFI,Jazz,and Netscience),as well as twenty synthetic datasets.These results provide the robustness and practical value of the proposed algorithm in multi-objective community identification.
基金supported by the Innovative Human Resource Development for Local Intel-lectualization program through the Institute of Information&Communications Technology Planning&Evaluation(IITP)grant funded by the Korea government(MSIT)(No.IITP-2026-2020-0-01741)the research fund of Hanyang University(HY-2025-1110).
文摘Arrhythmias are a frequently occurring phenomenon in clinical practice,but how to accurately dis-tinguish subtle rhythm abnormalities remains an ongoing difficulty faced by the entire research community when conducting ECG-based studies.From a review of existing studies,two main factors appear to contribute to this problem:the uneven distribution of arrhythmia classes and the limited expressiveness of features learned by current models.To overcome these limitations,this study proposes a dual-path multimodal framework,termed DM-EHC(Dual-Path Multimodal ECG Heartbeat Classifier),for ECG-based heartbeat classification.The proposed framework links 1D ECG temporal features with 2D time–frequency features.By setting up the dual paths described above,the model can process more dimensions of feature information.The MIT-BIH arrhythmia database was selected as the baseline dataset for the experiments.Experimental results show that the proposed method outperforms single modalities and performs better for certain specific types of arrhythmias.The model achieved mean precision,recall,and F1 score of 95.14%,92.26%,and 93.65%,respectively.These results indicate that the framework is robust and has potential value in automated arrhythmia classification.
基金funded by the China National Space Administration(KJSP2023020105)supported by the National Key R&D Program of China(Grant No.2023YFA1608100)+2 种基金the NSFC(Grant No.62227901)the Minor Planet Foundationsupported by the Egyptian Science,Technology&Innovation Funding Authority(STDF)under Grant No.48102.
文摘Near-Earth objects are important not only in studying the early formation of the Solar System,but also because they pose a serious hazard to humanity when they make close approaches to the Earth.Study of their physical properties can provide useful information on their origin,evolution,and hazard to human beings.However,it remains challenging to investigate small,newly discovered,near-Earth objects because of our limited observational window.This investigation seeks to determine the visible colors of near-Earth asteroids(NEAs),perform an initial taxonomic classification based on visible colors and analyze possible correlations between the distribution of taxonomic classification and asteroid size or orbital parameters.Observations were performed in the broadband BVRI Johnson−Cousins photometric system,applied to images from the Yaoan High Precision Telescope and the 1.88 m telescope at the Kottamia Astronomical Observatory.We present new photometric observations of 84 near-Earth asteroids,and classify 80 of them taxonomically,based on their photometric colors.We find that nearly half(46.3%)of the objects in our sample can be classified as S-complex,26.3%as C-complex,6%as D-complex,and 15.0%as X-complex;the remaining belong to the A-or V-types.Additionally,we identify three P-type NEAs in our sample,according to the Tholen scheme.The fractional abundances of the C/X-complex members with absolute magnitude H≥17.0 were more than twice as large as those with H<17.0.However,the fractions of C-and S-complex members with diameters≤1 km and>1 km are nearly equal,while X-complex members tend to have sub-kilometer diameters.In our sample,the C/D-complex objects are predominant among those with a Jovian Tisserand parameter of T_(J)<3.1.These bodies could have a cometary origin.C-and S-complex members account for a considerable proportion of the asteroids that are potentially hazardous.
基金funded by the National Key Research and Development Program of China(Grant No.2024YFE0209000)the NSFC(Grant No.U23B2019).
文摘Graph Neural Networks(GNNs)have proven highly effective for graph classification across diverse fields such as social networks,bioinformatics,and finance,due to their capability to learn complex graph structures.However,despite their success,GNNs remain vulnerable to adversarial attacks that can significantly degrade their classification accuracy.Existing adversarial attack strategies primarily rely on label information to guide the attacks,which limits their applicability in scenarios where such information is scarce or unavailable.This paper introduces an innovative unsupervised attack method for graph classification,which operates without relying on label information,thereby enhancing its applicability in a broad range of scenarios.Specifically,our method first leverages a graph contrastive learning loss to learn high-quality graph embeddings by comparing different stochastic augmented views of the graphs.To effectively perturb the graphs,we then introduce an implicit estimator that measures the impact of various modifications on graph structures.The proposed strategy identifies and flips edges with the top-K highest scores,determined by the estimator,to maximize the degradation of the model’s performance.In addition,to defend against such attack,we propose a lightweight regularization-based defense mechanism that is specifically tailored to mitigate the structural perturbations introduced by our attack strategy.It enhances model robustness by enforcing embedding consistency and edge-level smoothness during training.We conduct experiments on six public TU graph classification datasets:NCI1,NCI109,Mutagenicity,ENZYMES,COLLAB,and DBLP_v1,to evaluate the effectiveness of our attack and defense strategies.Under an attack budget of 3,the maximum reduction in model accuracy reaches 6.67%on the Graph Convolutional Network(GCN)and 11.67%on the Graph Attention Network(GAT)across different datasets,indicating that our unsupervised method induces degradation comparable to state-of-the-art supervised attacks.Meanwhile,our defense achieves the highest accuracy recovery of 3.89%(GCN)and 5.00%(GAT),demonstrating improved robustness against structural perturbations.
基金supported by Key Science and Technology Program of Henan Province,China(Grant Nos.242102210147,242102210027)Fujian Province Young and Middle aged Teacher Education Research Project(Science and Technology Category)(No.JZ240101)(Corresponding author:Dong Yuan).
文摘Vehicle Edge Computing(VEC)and Cloud Computing(CC)significantly enhance the processing efficiency of delay-sensitive and computation-intensive applications by offloading compute-intensive tasks from resource-constrained onboard devices to nearby Roadside Unit(RSU),thereby achieving lower delay and energy consumption.However,due to the limited storage capacity and energy budget of RSUs,it is challenging to meet the demands of the highly dynamic Internet of Vehicles(IoV)environment.Therefore,determining reasonable service caching and computation offloading strategies is crucial.To address this,this paper proposes a joint service caching scheme for cloud-edge collaborative IoV computation offloading.By modeling the dynamic optimization problem using Markov Decision Processes(MDP),the scheme jointly optimizes task delay,energy consumption,load balancing,and privacy entropy to achieve better quality of service.Additionally,a dynamic adaptive multi-objective deep reinforcement learning algorithm is proposed.Each Double Deep Q-Network(DDQN)agent obtains rewards for different objectives based on distinct reward functions and dynamically updates the objective weights by learning the value changes between objectives using Radial Basis Function Networks(RBFN),thereby efficiently approximating the Pareto-optimal decisions for multiple objectives.Extensive experiments demonstrate that the proposed algorithm can better coordinate the three-tier computing resources of cloud,edge,and vehicles.Compared to existing algorithms,the proposed method reduces task delay and energy consumption by 10.64%and 5.1%,respectively.
基金supported by the National Natural Science Foundation of China(No.62173107).
文摘Spaceborne antennas are essential for remote sensing,deep-space communication,and Earth observation,yet their trajectory planning is complicated by nonlinear base-manipulator coupling and antenna flexibility.To address these challenges,this paper proposes a multi-objective trajectory optimization framework.The system dynamics capture both nonlinear rigid-flexible coupling and antenna deformation through a reduced-order formulation.To enhance discretization efficiency,a predictive-terminal hp-adaptive pseudospectral method is employed,assigning collocation density based on task-phase characteristics:finer resolution is applied to dynamic segments requiring higher accuracy,especially near the terminal phase.This enables efficient transcription of the continuous-time problem into a Nonlinear Programming Problem(NLP).The resulting NLP is then solved using a multi-objective optimization strategy based on the nondominated sorting genetic algorithm II,which explores trade-offs among antenna pointing accuracy,energy consumption,and structural vibration.Numerical results demonstrate that the proposed method achieves a reduction of approximately 14.0% in control energy and 41.8%in peak actuation compared to a GPOPS-II baseline,while significantly enhancing vibration suppression.The resulting Pareto front reveals structured trade-offs and clustered solutions,offering robust and diverse options for precision,low-disturbance mission planning.
基金funded by the Deanship of Graduate Studies and Scientific Research at Jouf University under grant No.(DGSSR-2025-02-01296).
文摘Skin diseases affect millions worldwide.Early detection is key to preventing disfigurement,lifelong disability,or death.Dermoscopic images acquired in primary-care settings show high intra-class visual similarity and severe class imbalance,and occasional imaging artifacts can create ambiguity for state-of-the-art convolutional neural networks(CNNs).We frame skin lesion recognition as graph-based reasoning and,to ensure fair evaluation and avoid data leakage,adopt a strict lesion-level partitioning strategy.Each image is first over-segmented using SLIC(Simple Linear Iterative Clustering)to produce perceptually homogeneous superpixels.These superpixels form the nodes of a region-adjacency graph whose edges encode spatial continuity.Node attributes are 1280-dimensional embeddings extracted with a lightweight yet expressive EfficientNet-B0 backbone,providing strong representational power at modest computational cost.The resulting graphs are processed by a five-layer Graph Attention Network(GAT)that learns to weight inter-node relationships dynamically and aggregates multi-hop context before classifying lesions into seven classes with a log-softmax output.Extensive experiments on the DermaMNIST benchmark show the proposed pipeline achieves 88.35%accuracy and 98.04%AUC,outperforming contemporary CNNs,AutoML approaches,and alternative graph neural networks.An ablation study indicates EfficientNet-B0 produces superior node descriptors compared with ResNet-18 and DenseNet,and that roughly five GAT layers strike a good balance between being too shallow and over-deep while avoiding oversmoothing.The method requires no data augmentation or external metadata,making it a drop-in upgrade for clinical computer-aided diagnosis systems.
基金supported by the SungKyunKwan University and the BK21 FOUR(Graduate School Innovation)funded by the Ministry of Education(MOE,Korea)and National Research Foundation of Korea(NRF).
文摘With the recent increase in data volume and diversity,traditional text representation techniques are struggling to capture context,particularly in environments with sparse data.To address these challenges,this study proposes a new model,the Masked Joint Representation Model(MJRM).MJRM approximates the original hypothesis by leveraging multiple elements in a limited context.It dynamically adapts to changes in characteristics based on data distribution through three main components.First,masking-based representation learning,termed selective dynamic masking,integrates topic modeling and sentiment clustering to generate and train multiple instances across different data subsets,whose predictions are then aggregated with optimized weights.This design alleviates sparsity,suppresses noise,and preserves contextual structures.Second,regularization-based improvements are applied.Third,techniques for addressing sparse data are used to perform final inference.As a result,MJRM improves performance by up to 4%compared to existing AI techniques.In our experiments,we analyzed the contribution of each factor,demonstrating that masking,dynamic learning,and aggregating multiple instances complement each other to improve performance.This demonstrates that a masking-based multi-learning strategy is effective for context-aware sparse text classification,and can be useful even in challenging situations such as data shortage or data distribution variations.We expect that the approach can be extended to diverse fields such as sentiment analysis,spam filtering,and domain-specific document classification.
文摘Classifying job offers into occupational categories is a fundamental task in human resource information systems,as it improves and streamlines indexing,search,and matching between openings and job seekers.Comprehensive occupational databases such as O∗NET or ESCO provide detailed taxonomies of interrelated positions that can be leveraged to align the textual content of postings with occupational categories,thereby facilitating standardization,cross-system interoperability,and access to metadata for each occupation(e.g.,tasks,knowledge,skills,and abilities).In this work,we explore the effectiveness of fine-tuning existing language models(LMs)to classify job offers with occupational descriptors from O∗NET.This enables a more precise assessment of candidate suitability by identifying the specific knowledge and skills required for each position,and helps automate recruitment processes by mitigating human bias and subjectivity in candidate selection.We evaluate three representative BERT-like models:BERT,RoBERTa,and DeBERTa.BERT serves as the baseline encoder-only architecture;RoBERTa incorporates advances in pretraining objectives and data scale;and DeBERTa introduces architectural improvements through disentangled attention mechanisms.The best performance was achieved with the DeBERTa model,although the other models also produced strong results,and no statistically significant differences were observed acrossmodels.We also find that these models typically reach optimal performance after only a few training epochs,and that training with smaller,balanced datasets is effective.Consequently,comparable results can be obtained with models that require fewer computational resources and less training time,facilitating deployment and practical use.
文摘Container transportation is pivotal in global trade due to its efficiency,safety,and cost-effectiveness.However,structural defects—particularly in grapple slots—can result in cargo damage,financial loss,and elevated safety risks,including container drops during lifting operations.Timely and accurate inspection before and after transit is therefore essential.Traditional inspection methods rely heavily on manual observation of internal and external surfaces,which are time-consuming,resource-intensive,and prone to subjective errors.Container roofs pose additional challenges due to limited visibility,while grapple slots are especially vulnerable to wear from frequent use.This study proposes a two-stage automated detection framework targeting defects in container roof grapple slots.In the first stage,YOLOv7 is employed to localize grapple slot regions with high precision.In the second stage,ResNet50 classifies the extracted slots as either intact or defective.The results from both stages are integrated into a human-machine interface for real-time visualization and user verification.Experimental evaluations demonstrate that YOLOv7 achieves a 99%detection rate at 100 frames per second(FPS),while ResNet50 attains 87%classification accuracy at 34 FPS.Compared to some state of the arts,the proposed system offers significant speed,reliability,and usability improvements,enabling efficient defect identification and visual reconfirmation via the interface.
基金supported in part by the National Key Research and Development Program of Chinaunder(Grant 2021YFB3101100)in part by the National Natural Science Foundation of Chinaunder(Grant 42461057),(Grant 62272123),and(Grant 42371470)+1 种基金in part by the Fundamental Research Program of Shanxi Province under(Grant 202303021212164)in part by the Postgraduate Education Innovation Program of Shanxi Province under(Grant 2024KY474).
文摘Recently,large-scale deep learning models have been increasingly adopted for point cloud classification.However,thesemethods typically require collecting extensive datasets frommultiple clients,which may lead to privacy leaks.Federated learning provides an effective solution to data leakage by eliminating the need for data transmission,relying instead on the exchange of model parameters.However,the uneven distribution of client data can still affect the model’s ability to generalize effectively.To address these challenges,we propose a new framework for point cloud classification called Federated Dynamic Aggregation Selection Strategy-based Multi-Receptive Field Fusion Classification Framework(FDASS-MRFCF).Specifically,we tackle these challenges with two key innovations:(1)During the client local training phase,we propose a Multi-Receptive Field Fusion Classification Model(MRFCM),which captures local and global structures in point cloud data through dynamic convolution and multi-scale feature fusion,enhancing the robustness of point cloud classification.(2)In the server aggregation phase,we introduce a Federated Dynamic Aggregation Selection Strategy(FDASS),which employs a hybrid strategy to average client model parameters,skip aggregation,or reallocate local models to different clients,thereby balancing global consistency and local diversity.We evaluate our framework using the ModelNet40 and ShapeNetPart benchmarks,demonstrating its effectiveness.The proposed method is expected to significantly advance the field of point cloud classification in a secure environment.
文摘Evaluating the adversarial robustness of classification algorithms in machine learning is a crucial domain.However,current methods lack measurable and interpretable metrics.To address this issue,this paper introduces a visual evaluation index named confidence centroid skewing quadrilateral,which is based on a classification confidence-based confusion matrix,offering a quantitative and visual comparison of the adversarial robustness among different classification algorithms,and enhances intuitiveness and interpretability of attack impacts.We first conduct a validity test and sensitive analysis of the method.Then,prove its effectiveness through the experiments of five classification algorithms including artificial neural network(ANN),logistic regression(LR),support vector machine(SVM),convolutional neural network(CNN)and transformer against three adversarial attacks such as fast gradient sign method(FGSM),DeepFool,and projected gradient descent(PGD)attack.
基金supported by the Institute of Information&Communications Technology Planning&Evaluation(IITP)grant funded by the Korea government(MSIT)[RS-2021-II211341,Artificial Intelligence Graduate School Program(Chung-Ang University)],and by the Chung-Ang University Graduate Research Scholarship in 2024.
文摘Legal case classification involves the categorization of legal documents into predefined categories,which facilitates legal information retrieval and case management.However,real-world legal datasets often suffer from class imbalances due to the uneven distribution of case types across legal domains.This leads to biased model performance,in the form of high accuracy for overrepresented categories and underperformance for minority classes.To address this issue,in this study,we propose a data augmentation method that masks unimportant terms within a document selectively while preserving key terms fromthe perspective of the legal domain.This approach enhances data diversity and improves the generalization capability of conventional models.Our experiments demonstrate consistent improvements achieved by the proposed augmentation strategy in terms of accuracy and F1 score across all models,validating the effectiveness of the proposed method in legal case classification.