Regularization inversion uses constraints and a regularization factor to solve ill- posed inversion problems in geophysics. The choice of the regularization factor and of the initial model is critical in regularizatio...Regularization inversion uses constraints and a regularization factor to solve ill- posed inversion problems in geophysics. The choice of the regularization factor and of the initial model is critical in regularization inversion. To deal with these problems, we propose a multiobjective particle swarm inversion (MOPSOI) algorithm to simultaneously minimize the data misfit and model constraints, and obtain a multiobjective inversion solution set without the gradient information of the objective function and the regularization factor. We then choose the optimum solution from the solution set based on the trade-off between data misfit and constraints that substitute for the regularization factor. The inversion of synthetic two-dimensional magnetic data suggests that the MOPSOI algorithm can obtain as many feasible solutions as possible; thus, deeper insights of the inversion process can be gained and more reasonable solutions can be obtained by balancing the data misfit and constraints. The proposed MOPSOI algorithm can deal with the problems of choosing the right regularization factor and the initial model.展开更多
Automatic modulation recognition(AMR)of radiation source signals is a research focus in the field of cognitive radio.However,the AMR of radiation source signals at low SNRs still faces a great challenge.Therefore,the ...Automatic modulation recognition(AMR)of radiation source signals is a research focus in the field of cognitive radio.However,the AMR of radiation source signals at low SNRs still faces a great challenge.Therefore,the AMR method of radiation source signals based on two-dimensional data matrix and improved residual neural network is proposed in this paper.First,the time series of the radiation source signals are reconstructed into two-dimensional data matrix,which greatly simplifies the signal preprocessing process.Second,the depthwise convolution and large-size convolutional kernels based residual neural network(DLRNet)is proposed to improve the feature extraction capability of the AMR model.Finally,the model performs feature extraction and classification on the two-dimensional data matrix to obtain the recognition vector that represents the signal modulation type.Theoretical analysis and simulation results show that the AMR method based on two-dimensional data matrix and improved residual network can significantly improve the accuracy of the AMR method.The recognition accuracy of the proposed method maintains a high level greater than 90% even at -14 dB SNR.展开更多
The explicit solution to Cauchy problem for linearized system of two-dimensional isentropic flow with axisymmetrical initial data in gas dynamics is given.
In this paper,an Adaptive-Weighted Time-Dimensional and Space-Dimensional(AWTDSD) data aggregation algorithm for a clustered sensor network is proposed for prolonging the lifetime of the network as well as improving t...In this paper,an Adaptive-Weighted Time-Dimensional and Space-Dimensional(AWTDSD) data aggregation algorithm for a clustered sensor network is proposed for prolonging the lifetime of the network as well as improving the accuracy of the data gathered in the network.AWTDSD contains three phases:(1) the time-dimensional aggregation phase for eliminating the data redundancy;(2) the adaptive-weighted aggregation phase for further aggregating the data as well as improving the accuracy of the aggregated data; and(3) the space-dimensional aggregation phase for reducing the size and the amount of the data transmission to the base station.AWTDSD utilizes the correlations between the sensed data for reducing the data transmission and increasing the data accuracy as well.Experimental result shows that AWTDSD can not only save almost a half of the total energy consumption but also greatly increase the accuracy of the data monitored by the sensors in the clustered network.展开更多
Mechanically cleaved two-dimensional materials are random in size and thickness.Recognizing atomically thin flakes by human experts is inefficient and unsuitable for scalable production.Deep learning algorithms have b...Mechanically cleaved two-dimensional materials are random in size and thickness.Recognizing atomically thin flakes by human experts is inefficient and unsuitable for scalable production.Deep learning algorithms have been adopted as an alternative,nevertheless a major challenge is a lack of sufficient actual training images.Here we report the generation of synthetic two-dimensional materials images using StyleGAN3 to complement the dataset.DeepLabv3Plus network is trained with the synthetic images which reduces overfitting and improves recognition accuracy to over 90%.A semi-supervisory technique for labeling images is introduced to reduce manual efforts.The sharper edges recognized by this method facilitate material stacking with precise edge alignment,which benefits exploring novel properties of layered-material devices that crucially depend on the interlayer twist-angle.This feasible and efficient method allows for the rapid and high-quality manufacturing of atomically thin materials and devices.展开更多
Missing data presents a crucial challenge in data analysis,especially in high-dimensional datasets,where missing data often leads to biased conclusions and degraded model performance.In this study,we present a novel a...Missing data presents a crucial challenge in data analysis,especially in high-dimensional datasets,where missing data often leads to biased conclusions and degraded model performance.In this study,we present a novel autoencoder-based imputation framework that integrates a composite loss function to enhance robustness and precision.The proposed loss combines(i)a guided,masked mean squared error focusing on missing entries;(ii)a noise-aware regularization term to improve resilience against data corruption;and(iii)a variance penalty to encourage expressive yet stable reconstructions.We evaluate the proposed model across four missingness mechanisms,such as Missing Completely at Random,Missing at Random,Missing Not at Random,and Missing Not at Random with quantile censorship,under systematically varied feature counts,sample sizes,and missingness ratios ranging from 5%to 60%.Four publicly available real-world datasets(Stroke Prediction,Pima Indians Diabetes,Cardiovascular Disease,and Framingham Heart Study)were used,and the obtained results show that our proposed model consistently outperforms baseline methods,including traditional and deep learning-based techniques.An ablation study reveals the additive value of each component in the loss function.Additionally,we assessed the downstream utility of imputed data through classification tasks,where datasets imputed by the proposed method yielded the highest receiver operating characteristic area under the curve scores across all scenarios.The model demonstrates strong scalability and robustness,improving performance with larger datasets and higher feature counts.These results underscore the capacity of the proposed method to produce not only numerically accurate but also semantically useful imputations,making it a promising solution for robust data recovery in clinical applications.展开更多
Modern intrusion detection systems(MIDS)face persistent challenges in coping with the rapid evolution of cyber threats,high-volume network traffic,and imbalanced datasets.Traditional models often lack the robustness a...Modern intrusion detection systems(MIDS)face persistent challenges in coping with the rapid evolution of cyber threats,high-volume network traffic,and imbalanced datasets.Traditional models often lack the robustness and explainability required to detect novel and sophisticated attacks effectively.This study introduces an advanced,explainable machine learning framework for multi-class IDS using the KDD99 and IDS datasets,which reflects real-world network behavior through a blend of normal and diverse attack classes.The methodology begins with sophisticated data preprocessing,incorporating both RobustScaler and QuantileTransformer to address outliers and skewed feature distributions,ensuring standardized and model-ready inputs.Critical dimensionality reduction is achieved via the Harris Hawks Optimization(HHO)algorithm—a nature-inspired metaheuristic modeled on hawks’hunting strategies.HHO efficiently identifies the most informative features by optimizing a fitness function based on classification performance.Following feature selection,the SMOTE is applied to the training data to resolve class imbalance by synthetically augmenting underrepresented attack types.The stacked architecture is then employed,combining the strengths of XGBoost,SVM,and RF as base learners.This layered approach improves prediction robustness and generalization by balancing bias and variance across diverse classifiers.The model was evaluated using standard classification metrics:precision,recall,F1-score,and overall accuracy.The best overall performance was recorded with an accuracy of 99.44%for UNSW-NB15,demonstrating the model’s effectiveness.After balancing,the model demonstrated a clear improvement in detecting the attacks.We tested the model on four datasets to show the effectiveness of the proposed approach and performed the ablation study to check the effect of each parameter.Also,the proposed model is computationaly efficient.To support transparency and trust in decision-making,explainable AI(XAI)techniques are incorporated that provides both global and local insight into feature contributions,and offers intuitive visualizations for individual predictions.This makes it suitable for practical deployment in cybersecurity environments that demand both precision and accountability.展开更多
Reversible data hiding(RDH)enables secret data embedding while preserving complete cover image recovery,making it crucial for applications requiring image integrity.The pixel value ordering(PVO)technique used in multi...Reversible data hiding(RDH)enables secret data embedding while preserving complete cover image recovery,making it crucial for applications requiring image integrity.The pixel value ordering(PVO)technique used in multi-stego images provides good image quality but often results in low embedding capability.To address these challenges,this paper proposes a high-capacity RDH scheme based on PVO that generates three stego images from a single cover image.The cover image is partitioned into non-overlapping blocks with pixels sorted in ascending order.Four secret bits are embedded into each block’s maximum pixel value,while three additional bits are embedded into the second-largest value when the pixel difference exceeds a predefined threshold.A similar embedding strategy is also applied to the minimum side of the block,including the second-smallest pixel value.This design enables each block to embed up to 14 bits of secret data.Experimental results demonstrate that the proposed method achieves significantly higher embedding capacity and improved visual quality compared to existing triple-stego RDH approaches,advancing the field of reversible steganography.展开更多
With the increasing emphasis on personal information protection,encryption through security protocols has emerged as a critical requirement in data transmission and reception processes.Nevertheless,IoT ecosystems comp...With the increasing emphasis on personal information protection,encryption through security protocols has emerged as a critical requirement in data transmission and reception processes.Nevertheless,IoT ecosystems comprise heterogeneous networks where outdated systems coexist with the latest devices,spanning a range of devices from non-encrypted ones to fully encrypted ones.Given the limited visibility into payloads in this context,this study investigates AI-based attack detection methods that leverage encrypted traffic metadata,eliminating the need for decryption and minimizing system performance degradation—especially in light of these heterogeneous devices.Using the UNSW-NB15 and CICIoT-2023 dataset,encrypted and unencrypted traffic were categorized according to security protocol,and AI-based intrusion detection experiments were conducted for each traffic type based on metadata.To mitigate the problem of class imbalance,eight different data sampling techniques were applied.The effectiveness of these sampling techniques was then comparatively analyzed using two ensemble models and three Deep Learning(DL)models from various perspectives.The experimental results confirmed that metadata-based attack detection is feasible using only encrypted traffic.In the UNSW-NB15 dataset,the f1-score of encrypted traffic was approximately 0.98,which is 4.3%higher than that of unencrypted traffic(approximately 0.94).In addition,analysis of the encrypted traffic in the CICIoT-2023 dataset using the same method showed a significantly lower f1-score of roughly 0.43,indicating that the quality of the dataset and the preprocessing approach have a substantial impact on detection performance.Furthermore,when data sampling techniques were applied to encrypted traffic,the recall in the UNSWNB15(Encrypted)dataset improved by up to 23.0%,and in the CICIoT-2023(Encrypted)dataset by 20.26%,showing a similar level of improvement.Notably,in CICIoT-2023,f1-score and Receiver Operation Characteristic-Area Under the Curve(ROC-AUC)increased by 59.0%and 55.94%,respectively.These results suggest that data sampling can have a positive effect even in encrypted environments.However,the extent of the improvement may vary depending on data quality,model architecture,and sampling strategy.展开更多
Automated essay scoring(AES)systems have gained significant importance in educational settings,offering a scalable,efficient,and objective method for evaluating student essays.However,developing AES systems for Arabic...Automated essay scoring(AES)systems have gained significant importance in educational settings,offering a scalable,efficient,and objective method for evaluating student essays.However,developing AES systems for Arabic poses distinct challenges due to the language’s complex morphology,diglossia,and the scarcity of annotated datasets.This paper presents a hybrid approach to Arabic AES by combining text-based,vector-based,and embeddingbased similarity measures to improve essay scoring accuracy while minimizing the training data required.Using a large Arabic essay dataset categorized into thematic groups,the study conducted four experiments to evaluate the impact of feature selection,data size,and model performance.Experiment 1 established a baseline using a non-machine learning approach,selecting top-N correlated features to predict essay scores.The subsequent experiments employed 5-fold cross-validation.Experiment 2 showed that combining embedding-based,text-based,and vector-based features in a Random Forest(RF)model achieved an R2 of 88.92%and an accuracy of 83.3%within a 0.5-point tolerance.Experiment 3 further refined the feature selection process,demonstrating that 19 correlated features yielded optimal results,improving R2 to 88.95%.In Experiment 4,an optimal data efficiency training approach was introduced,where training data portions increased from 5%to 50%.The study found that using just 10%of the data achieved near-peak performance,with an R2 of 85.49%,emphasizing an effective trade-off between performance and computational costs.These findings highlight the potential of the hybrid approach for developing scalable Arabic AES systems,especially in low-resource environments,addressing linguistic challenges while ensuring efficient data usage.展开更多
Objective expertise evaluation of individuals,as a prerequisite stage for team formation,has been a long-term desideratum in large software development companies.With the rapid advancements in machine learning methods...Objective expertise evaluation of individuals,as a prerequisite stage for team formation,has been a long-term desideratum in large software development companies.With the rapid advancements in machine learning methods,based on reliable existing data stored in project management tools’datasets,automating this evaluation process becomes a natural step forward.In this context,our approach focuses on quantifying software developer expertise by using metadata from the task-tracking systems.For this,we mathematically formalize two categories of expertise:technology-specific expertise,which denotes the skills required for a particular technology,and general expertise,which encapsulates overall knowledge in the software industry.Afterward,we automatically classify the zones of expertise associated with each task a developer has worked on using Bidirectional Encoder Representations from Transformers(BERT)-like transformers to handle the unique characteristics of project tool datasets effectively.Finally,our method evaluates the proficiency of each software specialist across already completed projects from both technology-specific and general perspectives.The method was experimentally validated,yielding promising results.展开更多
Objective: to study and try to apply Internet medical technology to clinical nursing work, so as to improve clinical nursing service. Methods: a total of 100 patients in Department of general surgery and Department of...Objective: to study and try to apply Internet medical technology to clinical nursing work, so as to improve clinical nursing service. Methods: a total of 100 patients in Department of general surgery and Department of orthopedics were collected. The data of temperature, blood pressure, pulse and respiration were collected and uploaded to the cloud database. Make the website QR code as the only identification mark of the patient. The QR code corresponds to the home page of the patient's vital signs browsing, and automatically refresh the home page to search the real-time vital signs data in the cloud. The QR code is posted on the bedside card for scanning or forwarding. Then investigate the satisfaction of family members and medical staff. Results: to realize the integration of Internet technology and clinical nursing service, and make the patient's condition information public to the family members, so as to explore a new model of clinical nursing based on Internet technology.展开更多
Parkinson’s disease(PD)is a debilitating neurological disorder affecting over 10 million people worldwide.PD classification models using voice signals as input are common in the literature.It is believed that using d...Parkinson’s disease(PD)is a debilitating neurological disorder affecting over 10 million people worldwide.PD classification models using voice signals as input are common in the literature.It is believed that using deep learning algorithms further enhances performance;nevertheless,it is challenging due to the nature of small-scale and imbalanced PD datasets.This paper proposed a convolutional neural network-based deep support vector machine(CNN-DSVM)to automate the feature extraction process using CNN and extend the conventional SVM to a DSVM for better classification performance in small-scale PD datasets.A customized kernel function reduces the impact of biased classification towards the majority class(healthy candidates in our consideration).An improved generative adversarial network(IGAN)was designed to generate additional training data to enhance the model’s performance.For performance evaluation,the proposed algorithm achieves a sensitivity of 97.6%and a specificity of 97.3%.The performance comparison is evaluated from five perspectives,including comparisons with different data generation algorithms,feature extraction techniques,kernel functions,and existing works.Results reveal the effectiveness of the IGAN algorithm,which improves the sensitivity and specificity by 4.05%–4.72%and 4.96%–5.86%,respectively;and the effectiveness of the CNN-DSVM algorithm,which improves the sensitivity by 1.24%–57.4%and specificity by 1.04%–163%and reduces biased detection towards the majority class.The ablation experiments confirm the effectiveness of individual components.Two future research directions have also been suggested.展开更多
Some theorems of compactly supported non-tensor product form two-dimension Daubechies wavelet were analysed carefully. Compactly supported non-tensor product form two-dimension wavelet was constructed, then non-tensor...Some theorems of compactly supported non-tensor product form two-dimension Daubechies wavelet were analysed carefully. Compactly supported non-tensor product form two-dimension wavelet was constructed, then non-tensor product form two dimension wavelet finite element was used to solve the deflection problem of elastic thin plate. The error order was researched. A numerical example was given at last.展开更多
The adaptive mesh refinement (AMR) method is applied in the 2-D Euler multi-component elasticplastic hydrodynamics code (MEPH2Y). It is applied on detonation. Firstly, the AMR method is described, including a cons...The adaptive mesh refinement (AMR) method is applied in the 2-D Euler multi-component elasticplastic hydrodynamics code (MEPH2Y). It is applied on detonation. Firstly, the AMR method is described, including a conservative spatial interpolation, the time integration methodology with the adapitve time increment and an adaptive computational region method. The advantage of AMR technique is exhibited by numerical examples, including the 1-D C-J detonation and the 2-D implosion ignited from a single point. Results show that AMR can promote the computational efficiency, keeping the accuracy in interesting regions.展开更多
D-T_(2)two-dimensional nuclear magnetic resonance(2D NMR)logging technology can distinguish pore fluid types intuitively,and it is widely used in oil and gas exploration.Many 2D NMR inversion methods(e.g.,truncated si...D-T_(2)two-dimensional nuclear magnetic resonance(2D NMR)logging technology can distinguish pore fluid types intuitively,and it is widely used in oil and gas exploration.Many 2D NMR inversion methods(e.g.,truncated singular value decomposition(TSVD),Butler-Reds-Dawson(BRD),LM-norm smoothing,and TIST-L1 regularization methods)have been proposed successively,but most are limited to numerical simulations.This study focused on the applicability of different inversion methods for NMR logging data of various acquisition sequences,from which the optimal inversion method was selected based on the comparative analysis.First,the two-dimensional NMR logging principle was studied.Then,these inversion methods were studied in detail,and the precision and computational efficiency of CPMG and diffusion editing(DE)sequences obtained from oil-water and gas-water models were compared,respectively.The inversion results and calculation time of truncated singular value decomposition(TSVD),Butler-Reds-Dawson(BRD),LM-norm smoothing,and TIST-L1 regularization were compared and analyzed through numerical simulations.The inversion method was optimized to process SP mode logging data from the MR Scanner instrument.The results showed that the TIST-regularization and LM-norm smoothing methods were more accurate for the CPMG and DE sequence echo trains of the oil-water and gas-water models.However,the LM-norm smoothing method was less time-consuming,making it more suitable for logging data processing.A case study in well A25 showed that the processing results by the LM-norm smoothing method were consistent with GEOLOG software.This demonstrates that the LM-norm smoothing method is applicable in practical NMR logging processing.展开更多
Reviewing the empirical and theoretical parameter relationships between various parameters is a good way to understand more about contact binary systems.In this investigation,two-dimensional(2D)relationships for P–MV...Reviewing the empirical and theoretical parameter relationships between various parameters is a good way to understand more about contact binary systems.In this investigation,two-dimensional(2D)relationships for P–MV(system),P–L1,2,M1,2–L1,2,and q–Lratiowere revisited.The sample used is related to 118 contact binary systems with an orbital period shorter than 0.6 days whose absolute parameters were estimated based on the Gaia Data Release 3 parallax.We reviewed previous studies on 2D relationships and updated six parameter relationships.Therefore,Markov chain Monte Carlo and Machine Learning methods were used,and the outcomes were compared.We selected 22 contact binary systems from eight previous studies for comparison,which had light curve solutions using spectroscopic data.The results show that the systems are in good agreement with the results of this study.展开更多
In the two-dimensional positioning method of pulsars, the grid method is used to provide non-sensitive direction and positional estimates. However, the grid method has a high computational load and low accuracy due to...In the two-dimensional positioning method of pulsars, the grid method is used to provide non-sensitive direction and positional estimates. However, the grid method has a high computational load and low accuracy due to the interval of the grid. To improve estimation accuracy and reduce the computational load, we propose a fast twodimensional positioning method for the crab pulsar based on multiple optimization algorithms(FTPCO). The FTPCO uses the Levenberg–Marquardt(LM) algorithm, three-point orientation(TPO) method, particle swarm optimization(PSO) and Newton–Raphson-based optimizer(NRBO) to substitute the grid method. First, to avoid the influence of the non-sensitive direction on positioning, we take an orbital error and the distortion of the pulsar profile as optimization objectives and combine the grid method with the LM algorithm or PSO to search for the non-sensitive direction. Then, on the sensitive plane perpendicular to the non-sensitive direction, the TPO method is proposed to fast search the sensitive direction and sub-sensitive direction. Finally, the NRBO is employed on the sensitive and sub-sensitive directions to achieve two-dimensional positioning of the Crab pulsar. The simulation results show that the computational load of the FTPCO is reduced by 89.4% and the positioning accuracy of the FTPCO is improved by approximately 38% compared with the grid method. The FTPCO has the advantage of high real-time accuracy and does not fall into the local optimum.展开更多
By considering the effect of suspended solid particles in the ordinary equations for two-dimension inviscid incompressible mixing layer, the Rayleigh equation and the modified Rayleigh equation are obtained. And then,...By considering the effect of suspended solid particles in the ordinary equations for two-dimension inviscid incompressible mixing layer, the Rayleigh equation and the modified Rayleigh equation are obtained. And then, by solving the corresponding eigen-value equations with numerical computational method, the relation curves between perturbation frequency and spacial growth rate of the mixing layer for the varying particle loading, ratio of particle velocity to fluid velocity and Stokes number are got. Sever al important conclusions on the effect of suspended solid particles on unstability of the mixing layer are presented in the end by analyzing all the relation curves.展开更多
基金supported by the Natural Science Foundation of China(No.61273179)Department of Education,Science and Technology Research Project of Hubei Province of China(No.D20131206,No.20141304)
文摘Regularization inversion uses constraints and a regularization factor to solve ill- posed inversion problems in geophysics. The choice of the regularization factor and of the initial model is critical in regularization inversion. To deal with these problems, we propose a multiobjective particle swarm inversion (MOPSOI) algorithm to simultaneously minimize the data misfit and model constraints, and obtain a multiobjective inversion solution set without the gradient information of the objective function and the regularization factor. We then choose the optimum solution from the solution set based on the trade-off between data misfit and constraints that substitute for the regularization factor. The inversion of synthetic two-dimensional magnetic data suggests that the MOPSOI algorithm can obtain as many feasible solutions as possible; thus, deeper insights of the inversion process can be gained and more reasonable solutions can be obtained by balancing the data misfit and constraints. The proposed MOPSOI algorithm can deal with the problems of choosing the right regularization factor and the initial model.
基金National Natural Science Foundation of China under Grant No.61973037China Postdoctoral Science Foundation under Grant No.2022M720419。
文摘Automatic modulation recognition(AMR)of radiation source signals is a research focus in the field of cognitive radio.However,the AMR of radiation source signals at low SNRs still faces a great challenge.Therefore,the AMR method of radiation source signals based on two-dimensional data matrix and improved residual neural network is proposed in this paper.First,the time series of the radiation source signals are reconstructed into two-dimensional data matrix,which greatly simplifies the signal preprocessing process.Second,the depthwise convolution and large-size convolutional kernels based residual neural network(DLRNet)is proposed to improve the feature extraction capability of the AMR model.Finally,the model performs feature extraction and classification on the two-dimensional data matrix to obtain the recognition vector that represents the signal modulation type.Theoretical analysis and simulation results show that the AMR method based on two-dimensional data matrix and improved residual network can significantly improve the accuracy of the AMR method.The recognition accuracy of the proposed method maintains a high level greater than 90% even at -14 dB SNR.
文摘The explicit solution to Cauchy problem for linearized system of two-dimensional isentropic flow with axisymmetrical initial data in gas dynamics is given.
基金Supported by the Promotive Research Fund for Excellent Young and Middle-aged Scientists of Shandong Province(No.BS2010DX010)the Project of Higher Educational Science and Technology Program of Shandong Province(No.J12LN36)
文摘In this paper,an Adaptive-Weighted Time-Dimensional and Space-Dimensional(AWTDSD) data aggregation algorithm for a clustered sensor network is proposed for prolonging the lifetime of the network as well as improving the accuracy of the data gathered in the network.AWTDSD contains three phases:(1) the time-dimensional aggregation phase for eliminating the data redundancy;(2) the adaptive-weighted aggregation phase for further aggregating the data as well as improving the accuracy of the aggregated data; and(3) the space-dimensional aggregation phase for reducing the size and the amount of the data transmission to the base station.AWTDSD utilizes the correlations between the sensed data for reducing the data transmission and increasing the data accuracy as well.Experimental result shows that AWTDSD can not only save almost a half of the total energy consumption but also greatly increase the accuracy of the data monitored by the sensors in the clustered network.
基金Project supported by the National Key Research and Development Program of China(Grant No.2022YFB2803900)the National Natural Science Foundation of China(Grant Nos.61974075 and 61704121)+2 种基金the Natural Science Foundation of Tianjin Municipality(Grant Nos.22JCZDJC00460 and 19JCQNJC00700)Tianjin Municipal Education Commission(Grant No.2019KJ028)Fundamental Research Funds for the Central Universities(Grant No.22JCZDJC00460).
文摘Mechanically cleaved two-dimensional materials are random in size and thickness.Recognizing atomically thin flakes by human experts is inefficient and unsuitable for scalable production.Deep learning algorithms have been adopted as an alternative,nevertheless a major challenge is a lack of sufficient actual training images.Here we report the generation of synthetic two-dimensional materials images using StyleGAN3 to complement the dataset.DeepLabv3Plus network is trained with the synthetic images which reduces overfitting and improves recognition accuracy to over 90%.A semi-supervisory technique for labeling images is introduced to reduce manual efforts.The sharper edges recognized by this method facilitate material stacking with precise edge alignment,which benefits exploring novel properties of layered-material devices that crucially depend on the interlayer twist-angle.This feasible and efficient method allows for the rapid and high-quality manufacturing of atomically thin materials and devices.
文摘Missing data presents a crucial challenge in data analysis,especially in high-dimensional datasets,where missing data often leads to biased conclusions and degraded model performance.In this study,we present a novel autoencoder-based imputation framework that integrates a composite loss function to enhance robustness and precision.The proposed loss combines(i)a guided,masked mean squared error focusing on missing entries;(ii)a noise-aware regularization term to improve resilience against data corruption;and(iii)a variance penalty to encourage expressive yet stable reconstructions.We evaluate the proposed model across four missingness mechanisms,such as Missing Completely at Random,Missing at Random,Missing Not at Random,and Missing Not at Random with quantile censorship,under systematically varied feature counts,sample sizes,and missingness ratios ranging from 5%to 60%.Four publicly available real-world datasets(Stroke Prediction,Pima Indians Diabetes,Cardiovascular Disease,and Framingham Heart Study)were used,and the obtained results show that our proposed model consistently outperforms baseline methods,including traditional and deep learning-based techniques.An ablation study reveals the additive value of each component in the loss function.Additionally,we assessed the downstream utility of imputed data through classification tasks,where datasets imputed by the proposed method yielded the highest receiver operating characteristic area under the curve scores across all scenarios.The model demonstrates strong scalability and robustness,improving performance with larger datasets and higher feature counts.These results underscore the capacity of the proposed method to produce not only numerically accurate but also semantically useful imputations,making it a promising solution for robust data recovery in clinical applications.
基金funded by Princess Nourah bint Abdulrahman University Researchers Supporting Project number(PNURSP2025R104)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.
文摘Modern intrusion detection systems(MIDS)face persistent challenges in coping with the rapid evolution of cyber threats,high-volume network traffic,and imbalanced datasets.Traditional models often lack the robustness and explainability required to detect novel and sophisticated attacks effectively.This study introduces an advanced,explainable machine learning framework for multi-class IDS using the KDD99 and IDS datasets,which reflects real-world network behavior through a blend of normal and diverse attack classes.The methodology begins with sophisticated data preprocessing,incorporating both RobustScaler and QuantileTransformer to address outliers and skewed feature distributions,ensuring standardized and model-ready inputs.Critical dimensionality reduction is achieved via the Harris Hawks Optimization(HHO)algorithm—a nature-inspired metaheuristic modeled on hawks’hunting strategies.HHO efficiently identifies the most informative features by optimizing a fitness function based on classification performance.Following feature selection,the SMOTE is applied to the training data to resolve class imbalance by synthetically augmenting underrepresented attack types.The stacked architecture is then employed,combining the strengths of XGBoost,SVM,and RF as base learners.This layered approach improves prediction robustness and generalization by balancing bias and variance across diverse classifiers.The model was evaluated using standard classification metrics:precision,recall,F1-score,and overall accuracy.The best overall performance was recorded with an accuracy of 99.44%for UNSW-NB15,demonstrating the model’s effectiveness.After balancing,the model demonstrated a clear improvement in detecting the attacks.We tested the model on four datasets to show the effectiveness of the proposed approach and performed the ablation study to check the effect of each parameter.Also,the proposed model is computationaly efficient.To support transparency and trust in decision-making,explainable AI(XAI)techniques are incorporated that provides both global and local insight into feature contributions,and offers intuitive visualizations for individual predictions.This makes it suitable for practical deployment in cybersecurity environments that demand both precision and accountability.
基金funded by University of Transport and Communications(UTC)under grant number T2025-CN-004.
文摘Reversible data hiding(RDH)enables secret data embedding while preserving complete cover image recovery,making it crucial for applications requiring image integrity.The pixel value ordering(PVO)technique used in multi-stego images provides good image quality but often results in low embedding capability.To address these challenges,this paper proposes a high-capacity RDH scheme based on PVO that generates three stego images from a single cover image.The cover image is partitioned into non-overlapping blocks with pixels sorted in ascending order.Four secret bits are embedded into each block’s maximum pixel value,while three additional bits are embedded into the second-largest value when the pixel difference exceeds a predefined threshold.A similar embedding strategy is also applied to the minimum side of the block,including the second-smallest pixel value.This design enables each block to embed up to 14 bits of secret data.Experimental results demonstrate that the proposed method achieves significantly higher embedding capacity and improved visual quality compared to existing triple-stego RDH approaches,advancing the field of reversible steganography.
基金supported by the Institute of Information&Communications Technology Planning&Evaluation(IITP)grant funded by the Korea government(MSIT)(No.RS-2023-00235509Development of security monitoring technology based network behavior against encrypted cyber threats in ICT convergence environment).
文摘With the increasing emphasis on personal information protection,encryption through security protocols has emerged as a critical requirement in data transmission and reception processes.Nevertheless,IoT ecosystems comprise heterogeneous networks where outdated systems coexist with the latest devices,spanning a range of devices from non-encrypted ones to fully encrypted ones.Given the limited visibility into payloads in this context,this study investigates AI-based attack detection methods that leverage encrypted traffic metadata,eliminating the need for decryption and minimizing system performance degradation—especially in light of these heterogeneous devices.Using the UNSW-NB15 and CICIoT-2023 dataset,encrypted and unencrypted traffic were categorized according to security protocol,and AI-based intrusion detection experiments were conducted for each traffic type based on metadata.To mitigate the problem of class imbalance,eight different data sampling techniques were applied.The effectiveness of these sampling techniques was then comparatively analyzed using two ensemble models and three Deep Learning(DL)models from various perspectives.The experimental results confirmed that metadata-based attack detection is feasible using only encrypted traffic.In the UNSW-NB15 dataset,the f1-score of encrypted traffic was approximately 0.98,which is 4.3%higher than that of unencrypted traffic(approximately 0.94).In addition,analysis of the encrypted traffic in the CICIoT-2023 dataset using the same method showed a significantly lower f1-score of roughly 0.43,indicating that the quality of the dataset and the preprocessing approach have a substantial impact on detection performance.Furthermore,when data sampling techniques were applied to encrypted traffic,the recall in the UNSWNB15(Encrypted)dataset improved by up to 23.0%,and in the CICIoT-2023(Encrypted)dataset by 20.26%,showing a similar level of improvement.Notably,in CICIoT-2023,f1-score and Receiver Operation Characteristic-Area Under the Curve(ROC-AUC)increased by 59.0%and 55.94%,respectively.These results suggest that data sampling can have a positive effect even in encrypted environments.However,the extent of the improvement may vary depending on data quality,model architecture,and sampling strategy.
基金funded by Deanship of Graduate studies and Scientific Research at Jouf University under grant No.(DGSSR-2024-02-01264).
文摘Automated essay scoring(AES)systems have gained significant importance in educational settings,offering a scalable,efficient,and objective method for evaluating student essays.However,developing AES systems for Arabic poses distinct challenges due to the language’s complex morphology,diglossia,and the scarcity of annotated datasets.This paper presents a hybrid approach to Arabic AES by combining text-based,vector-based,and embeddingbased similarity measures to improve essay scoring accuracy while minimizing the training data required.Using a large Arabic essay dataset categorized into thematic groups,the study conducted four experiments to evaluate the impact of feature selection,data size,and model performance.Experiment 1 established a baseline using a non-machine learning approach,selecting top-N correlated features to predict essay scores.The subsequent experiments employed 5-fold cross-validation.Experiment 2 showed that combining embedding-based,text-based,and vector-based features in a Random Forest(RF)model achieved an R2 of 88.92%and an accuracy of 83.3%within a 0.5-point tolerance.Experiment 3 further refined the feature selection process,demonstrating that 19 correlated features yielded optimal results,improving R2 to 88.95%.In Experiment 4,an optimal data efficiency training approach was introduced,where training data portions increased from 5%to 50%.The study found that using just 10%of the data achieved near-peak performance,with an R2 of 85.49%,emphasizing an effective trade-off between performance and computational costs.These findings highlight the potential of the hybrid approach for developing scalable Arabic AES systems,especially in low-resource environments,addressing linguistic challenges while ensuring efficient data usage.
基金supported by the project“Romanian Hub for Artificial Intelligence-HRIA”,Smart Growth,Digitization and Financial Instruments Program,2021–2027,MySMIS No.334906.
文摘Objective expertise evaluation of individuals,as a prerequisite stage for team formation,has been a long-term desideratum in large software development companies.With the rapid advancements in machine learning methods,based on reliable existing data stored in project management tools’datasets,automating this evaluation process becomes a natural step forward.In this context,our approach focuses on quantifying software developer expertise by using metadata from the task-tracking systems.For this,we mathematically formalize two categories of expertise:technology-specific expertise,which denotes the skills required for a particular technology,and general expertise,which encapsulates overall knowledge in the software industry.Afterward,we automatically classify the zones of expertise associated with each task a developer has worked on using Bidirectional Encoder Representations from Transformers(BERT)-like transformers to handle the unique characteristics of project tool datasets effectively.Finally,our method evaluates the proficiency of each software specialist across already completed projects from both technology-specific and general perspectives.The method was experimentally validated,yielding promising results.
文摘Objective: to study and try to apply Internet medical technology to clinical nursing work, so as to improve clinical nursing service. Methods: a total of 100 patients in Department of general surgery and Department of orthopedics were collected. The data of temperature, blood pressure, pulse and respiration were collected and uploaded to the cloud database. Make the website QR code as the only identification mark of the patient. The QR code corresponds to the home page of the patient's vital signs browsing, and automatically refresh the home page to search the real-time vital signs data in the cloud. The QR code is posted on the bedside card for scanning or forwarding. Then investigate the satisfaction of family members and medical staff. Results: to realize the integration of Internet technology and clinical nursing service, and make the patient's condition information public to the family members, so as to explore a new model of clinical nursing based on Internet technology.
基金The work described in this paper was fully supported by a grant from Hong Kong Metropolitan University(RIF/2021/05).
文摘Parkinson’s disease(PD)is a debilitating neurological disorder affecting over 10 million people worldwide.PD classification models using voice signals as input are common in the literature.It is believed that using deep learning algorithms further enhances performance;nevertheless,it is challenging due to the nature of small-scale and imbalanced PD datasets.This paper proposed a convolutional neural network-based deep support vector machine(CNN-DSVM)to automate the feature extraction process using CNN and extend the conventional SVM to a DSVM for better classification performance in small-scale PD datasets.A customized kernel function reduces the impact of biased classification towards the majority class(healthy candidates in our consideration).An improved generative adversarial network(IGAN)was designed to generate additional training data to enhance the model’s performance.For performance evaluation,the proposed algorithm achieves a sensitivity of 97.6%and a specificity of 97.3%.The performance comparison is evaluated from five perspectives,including comparisons with different data generation algorithms,feature extraction techniques,kernel functions,and existing works.Results reveal the effectiveness of the IGAN algorithm,which improves the sensitivity and specificity by 4.05%–4.72%and 4.96%–5.86%,respectively;and the effectiveness of the CNN-DSVM algorithm,which improves the sensitivity by 1.24%–57.4%and specificity by 1.04%–163%and reduces biased detection towards the majority class.The ablation experiments confirm the effectiveness of individual components.Two future research directions have also been suggested.
文摘Some theorems of compactly supported non-tensor product form two-dimension Daubechies wavelet were analysed carefully. Compactly supported non-tensor product form two-dimension wavelet was constructed, then non-tensor product form two dimension wavelet finite element was used to solve the deflection problem of elastic thin plate. The error order was researched. A numerical example was given at last.
基金Sponsored by the National Natural Science Foundation of China(10676120)Laboratory of Computational Physics Foundation(9140C690101070C69)
文摘The adaptive mesh refinement (AMR) method is applied in the 2-D Euler multi-component elasticplastic hydrodynamics code (MEPH2Y). It is applied on detonation. Firstly, the AMR method is described, including a conservative spatial interpolation, the time integration methodology with the adapitve time increment and an adaptive computational region method. The advantage of AMR technique is exhibited by numerical examples, including the 1-D C-J detonation and the 2-D implosion ignited from a single point. Results show that AMR can promote the computational efficiency, keeping the accuracy in interesting regions.
基金sponsored by the National Natural Science Foundation of China(Nos.42174149,41774144)the National Major Projects(No.2016ZX05014-001).
文摘D-T_(2)two-dimensional nuclear magnetic resonance(2D NMR)logging technology can distinguish pore fluid types intuitively,and it is widely used in oil and gas exploration.Many 2D NMR inversion methods(e.g.,truncated singular value decomposition(TSVD),Butler-Reds-Dawson(BRD),LM-norm smoothing,and TIST-L1 regularization methods)have been proposed successively,but most are limited to numerical simulations.This study focused on the applicability of different inversion methods for NMR logging data of various acquisition sequences,from which the optimal inversion method was selected based on the comparative analysis.First,the two-dimensional NMR logging principle was studied.Then,these inversion methods were studied in detail,and the precision and computational efficiency of CPMG and diffusion editing(DE)sequences obtained from oil-water and gas-water models were compared,respectively.The inversion results and calculation time of truncated singular value decomposition(TSVD),Butler-Reds-Dawson(BRD),LM-norm smoothing,and TIST-L1 regularization were compared and analyzed through numerical simulations.The inversion method was optimized to process SP mode logging data from the MR Scanner instrument.The results showed that the TIST-regularization and LM-norm smoothing methods were more accurate for the CPMG and DE sequence echo trains of the oil-water and gas-water models.However,the LM-norm smoothing method was less time-consuming,making it more suitable for logging data processing.A case study in well A25 showed that the processing results by the LM-norm smoothing method were consistent with GEOLOG software.This demonstrates that the LM-norm smoothing method is applicable in practical NMR logging processing.
基金The Binary Systems of South and North(BSN)project(https://bsnp.info/)。
文摘Reviewing the empirical and theoretical parameter relationships between various parameters is a good way to understand more about contact binary systems.In this investigation,two-dimensional(2D)relationships for P–MV(system),P–L1,2,M1,2–L1,2,and q–Lratiowere revisited.The sample used is related to 118 contact binary systems with an orbital period shorter than 0.6 days whose absolute parameters were estimated based on the Gaia Data Release 3 parallax.We reviewed previous studies on 2D relationships and updated six parameter relationships.Therefore,Markov chain Monte Carlo and Machine Learning methods were used,and the outcomes were compared.We selected 22 contact binary systems from eight previous studies for comparison,which had light curve solutions using spectroscopic data.The results show that the systems are in good agreement with the results of this study.
基金supported by the National Natural Science Foundation of China (Nos. 61873196 and 62373030)the Innovation Program for Quantum Science and Technology(No. 2021ZD0303400)。
文摘In the two-dimensional positioning method of pulsars, the grid method is used to provide non-sensitive direction and positional estimates. However, the grid method has a high computational load and low accuracy due to the interval of the grid. To improve estimation accuracy and reduce the computational load, we propose a fast twodimensional positioning method for the crab pulsar based on multiple optimization algorithms(FTPCO). The FTPCO uses the Levenberg–Marquardt(LM) algorithm, three-point orientation(TPO) method, particle swarm optimization(PSO) and Newton–Raphson-based optimizer(NRBO) to substitute the grid method. First, to avoid the influence of the non-sensitive direction on positioning, we take an orbital error and the distortion of the pulsar profile as optimization objectives and combine the grid method with the LM algorithm or PSO to search for the non-sensitive direction. Then, on the sensitive plane perpendicular to the non-sensitive direction, the TPO method is proposed to fast search the sensitive direction and sub-sensitive direction. Finally, the NRBO is employed on the sensitive and sub-sensitive directions to achieve two-dimensional positioning of the Crab pulsar. The simulation results show that the computational load of the FTPCO is reduced by 89.4% and the positioning accuracy of the FTPCO is improved by approximately 38% compared with the grid method. The FTPCO has the advantage of high real-time accuracy and does not fall into the local optimum.
文摘By considering the effect of suspended solid particles in the ordinary equations for two-dimension inviscid incompressible mixing layer, the Rayleigh equation and the modified Rayleigh equation are obtained. And then, by solving the corresponding eigen-value equations with numerical computational method, the relation curves between perturbation frequency and spacial growth rate of the mixing layer for the varying particle loading, ratio of particle velocity to fluid velocity and Stokes number are got. Sever al important conclusions on the effect of suspended solid particles on unstability of the mixing layer are presented in the end by analyzing all the relation curves.