As the crypto-asset ecosystem matures,the use of high-frequency data has become increasingly common in decentralized finance literature.Using bibliometric analysis,we characterize the existing cryptocurrency literatur...As the crypto-asset ecosystem matures,the use of high-frequency data has become increasingly common in decentralized finance literature.Using bibliometric analysis,we characterize the existing cryptocurrency literature that employs high-frequency data.We highlighted the most influential authors,articles,and journals based on 189 articles from the Scopus database from 2015 to 2022.This approach enables us to identify emerging trends and research hotspots with the aid of co-citation and cartographic analyses.It shows knowledge expansion through authors’collaboration in cryptocurrency research with co-authorship analysis.We identify four major streams of research:(i)return prediction and measurement of cryptocurrency volatility,(ii)(in)efficiency of cryptocurrencies,(iii)price dynamics and bubbles in cryptocurrencies,and(iv)the diversification,safe haven,and hedging properties of Bitcoin.We conclude that highly traded cryptocurrencies’investment features and economic outcomes are analyzed predominantly on a tick-by-tick basis.This study also provides recommendations for future studies.展开更多
The modeling of volatility and correlation is important in order to calculate hedge ratios, value at risk estimates, CAPM (Capital Asset Pricing Model betas), derivate pricing and risk management in general. Recent ...The modeling of volatility and correlation is important in order to calculate hedge ratios, value at risk estimates, CAPM (Capital Asset Pricing Model betas), derivate pricing and risk management in general. Recent access to intra-daily high-frequency data for two of the most liquid contracts at the Nord Pool exchange has made it possible to apply new and promising methods for analyzing volatility and correlation. The concepts of realized volatility and realized correlation are applied, and this study statistically describes the distribution (both distributional properties and temporal dependencies) of electricity forward data from 2005 to 2009. The main findings show that the logarithmic realized volatility is approximately normally distributed, while realized correlation seems not to be. Further, realized volatility and realized correlation have a long-memory feature. There also seems to be a high correlation between realized correlation and volatilities and positive relations between trading volume and realized volatility and between trading volume and realized correlation. These results are to a large extent consistent with earlier studies of stylized facts of other financial and commodity markets.展开更多
This paper studies an M-estimator of a proxy periodic GARCH (p, q) scaling model and establishes its consistency and asymptotic normality. Simulation studies are carried out to assess the performance of the estimato...This paper studies an M-estimator of a proxy periodic GARCH (p, q) scaling model and establishes its consistency and asymptotic normality. Simulation studies are carried out to assess the performance of the estimator. The numerical results show that our M-estimator is more efficient and robust than other estimators without the use of high-frequency data.展开更多
The low-pass fi ltering eff ect of the Earth results in the absorption and attenuation of the high-frequency components of seismic signals by the stratum during propagation.Hence,seismic data have low resolution.Consi...The low-pass fi ltering eff ect of the Earth results in the absorption and attenuation of the high-frequency components of seismic signals by the stratum during propagation.Hence,seismic data have low resolution.Considering the limitations of traditional high-frequency compensation methods,this paper presents a new method based on adaptive generalized S transform.This method is based on the study of frequency spectrum attenuation law of seismic signals,and the Gauss window function of adaptive generalized S transform is used to fi t the attenuation trend of seismic signals to seek the optimal Gauss window function.The amplitude spectrum compensation function constructed using the optimal Gauss window function is used to modify the time-frequency spectrum of the adaptive generalized S transform of seismic signals and reconstruct seismic signals to compensate for high-frequency attenuation.Practical data processing results show that the method can compensate for the high-frequency components that are absorbed and attenuated by the stratum,thereby eff ectively improving the resolution and quality of seismic data.展开更多
The oil-based mud(OBM) borehole measurement environment presents significant limitations on the application of existing electrical logging instruments in high-resistance formations. In this paper, we propose a novel l...The oil-based mud(OBM) borehole measurement environment presents significant limitations on the application of existing electrical logging instruments in high-resistance formations. In this paper, we propose a novel logging method for detection of high-resistance formations in OBM using highfrequency electrodes. The method addresses the issue of shallow depth of investigation(DOI) in existing electrical logging instruments, while simultaneously ensuring the vertical resolution. Based on the principle of current continuity, the total impedance of the loop is obtained by equating the measurement loop to the series form of a capacitively coupled circuit. and its validity is verified in a homogeneous formation model and a radial two-layer formation model with a mud standoff. Then, the instrument operating frequency and electrode system parameters were preferentially determined by numerical simulation, and the effect of mud gap on impedance measurement was investigated. Subsequently, the DOI of the instrument was investigated utilizing the pseudo-geometric factor defined by the real part of impedance. It was determined that the detection depth of the instrument is 8.74 cm, while the effective vertical resolution was not less than 2 cm. Finally, a focused high-frequency electrode-type instrument was designed by introducing a pair of focused electrodes, which effectively enhanced the DOI of the instrument and was successfully deployed in the Oklahoma formation model. The simulation results demonstrate that the novel method can achieve a detection depth of 17.40 cm in highly-resistive formations drilling with OBM, which is approximately twice the depth of detection of the existing oil-based mud microimager instruments. Furthermore, its effective vertical resolution remains at or above 2 cm,which is comparable to the resolution of the existing OBM electrical logging instrument.展开更多
Objective:To analyze the significance of high-frequency ultrasound in differentiating benign and malignant breast micronodules.Methods:Eighty-five patients with breast micronodules admitted for diagnosis between Octob...Objective:To analyze the significance of high-frequency ultrasound in differentiating benign and malignant breast micronodules.Methods:Eighty-five patients with breast micronodules admitted for diagnosis between October 2022 and October 2024 were selected for high-frequency ultrasound diagnosis.The diagnostic efficacy of high-frequency ultrasound was evaluated by comparing it with the results of surgical pathology.Results:High-frequency ultrasound detected 50 benign nodules,primarily breast fibroadenomas,and 35 malignant nodules,mainly breast ductal carcinoma in situ.Based on surgical pathology results,the diagnostic accuracy of high-frequency ultrasound was 96.47%,specificity was 97.96%,and sensitivity was 94.44%.In high-frequency ultrasound diagnosis,the proportion of grade III and IV blood flow in malignant nodules was higher than that in benign nodules,while the proportion of regular shape and clear margins was lower.The proportion of microcalcifications and posterior echo attenuation was higher in malignant nodules,and the resistance index(RI)and peak blood flow velocity were lower than those in benign nodules(P<0.05).Conclusion:High-frequency ultrasound can effectively differentiate benign and malignant breast micronodules,determine specific nodule types,and exhibits high diagnostic accuracy and sensitivity.Additionally,benign and malignant nodules can be differentiated based on the grading of blood flow signals,sonographic features,and blood flow velocity,providing reasonable guidance for subsequent treatment plans.展开更多
New electric power systems characterized by a high proportion of renewable energy and power electronics equipment face significant challenges due to high-frequency(HF)electromagnetic interference from the high-speed s...New electric power systems characterized by a high proportion of renewable energy and power electronics equipment face significant challenges due to high-frequency(HF)electromagnetic interference from the high-speed switching of power converters.To address this situation,this paper offers an in-depth review of HF interference problems and challenges originating from power electronic devices.First,the root cause of HF electromagnetic interference,i.e.,the resonant response of the parasitic parameters of the system to high-speed switching transients,is analyzed,and various scenarios of HF interference in power systems are highlighted.Next,the types of HF interference are summarized,with a focus on common-mode interference in grounding systems.This paper thoroughly reviews and compares various suppression methods for conducted HF interference.Finally,the challenges involved and suggestions for addressing emerging HF interference problems from the perspective of both power electronics equipment and power systems are discussed.This review aims to offer a structured understanding of HF interference problems and their suppression techniques for researchers and practitioners.展开更多
China has a long history of coal mining,among which open-pit coal mines have a large number of small coal mine goafs underground.The distribution,shape,structure and other characteristics of goafs are isolated and dis...China has a long history of coal mining,among which open-pit coal mines have a large number of small coal mine goafs underground.The distribution,shape,structure and other characteristics of goafs are isolated and discontinuous,and there is no definite geological law to follow,which seriously threatens the safety of coal mine production and personnel life.Conventional ground geophysical methods have low accuracy in detecting goaf areas affected by mechanical interference from open-pit mines,especially for waterless goaf areas,which cannot be detected by existing methods.This article proposes the use of high-frequency electromagnetic waves for goaf detection.The feasibility of using drilling radar to detect goaf was theoretically analyzed,and a goaf detection model was established.The response characteristics of different fillers in the goaf under different frequencies of high-frequency electromagnetic waves were simulated and analyzed.In a certain open-pit mine in Inner Mongolia,100MHz high-frequency electromagnetic waves were used to detect the goaf through directional drilling on the ground.After detection,excavation verification was carried out,and the location of one goaf detected was verified.The results of engineering practice show that the application of high-frequency electromagnetic waves in goaf detection expands the detection radius of boreholes,has the advantages of high efficiency and accuracy,and has important theoretical and practical significance.展开更多
Single-shot ultrafast compressed imaging(UCI)is an effective tool for studying ultrafast dynamics in physics,chemistry,or material science because of its excellent high frame rate and large frame number.However,the ra...Single-shot ultrafast compressed imaging(UCI)is an effective tool for studying ultrafast dynamics in physics,chemistry,or material science because of its excellent high frame rate and large frame number.However,the random code(Rcode)used in traditional UCI will lead to low-frequency noise covering high-frequency information due to its uneven sampling interval,which is a great challenge in the fidelity of large-frame reconstruction.Here,a high-frequency enhanced compressed active photography(H-CAP)is proposed.By uniformizing the sampling interval of R-code,H-CAP capture the ultrafast process with a random uniform sampling mode.This sampling mode makes the high-frequency sampling energy dominant,which greatly suppresses the low-frequency noise blurring caused by R-code and achieves high-frequency information of image enhanced.The superior dynamic performance and large-frame reconstruction ability of H-CAP are verified by imaging optical self-focusing effect and static object,respectively.We applied H-CAP to the spatial-temporal characterization of double-pulse induced silicon surface ablation dynamics,which is performed within 220 frames in a single-shot of 300 ps.H-CAP provides a high-fidelity imaging method for observing ultrafast unrepeatable dynamic processes with large frames.展开更多
Objective:To analyze the therapeutic effect of high-frequency electrosurgical knife surgery guided by painless digestive endoscopy(PDE)in elderly patients with gastrointestinal polyps(GP).Methods:A total of 100 elderl...Objective:To analyze the therapeutic effect of high-frequency electrosurgical knife surgery guided by painless digestive endoscopy(PDE)in elderly patients with gastrointestinal polyps(GP).Methods:A total of 100 elderly GP patients admitted between June 2021 and December 2022 were selected.Patients were randomly divided into two groups:the painless group(50 cases)underwent high-frequency electrosurgical knife surgery guided by PDE,while the conventional group(50 cases)underwent the same surgery guided by traditional digestive endoscopy(DE).The total treatment efficacy,perioperative indicators,gastrointestinal hormone levels,oxidative stress(OS)markers,and complication rates were compared between the two groups.Results:The total treatment efficacy in the painless group was higher than that in the conventional group,and perioperative indicators were superior in the painless group(P<0.05).One week after treatment,the gastrointestinal hormone levels and OS-related markers in the painless group were better than those in the conventional group(P<0.05).The complication rate in the painless group was lower than in the conventional group(P<0.05).Conclusion:High-frequency electrosurgical knife surgery guided by PDE improves the effectiveness of polyp removal in elderly GP patients and accelerates postoperative recovery.It also protects gastrointestinal function,reduces postoperative OS,and ensures higher surgical safety.展开更多
Soft magnetic composites made from metallic magnetic particles with an easy magnetization plane(referred to as easy-plane metallic soft magnetic composites(SMC))are considered ideal materials for the next generation o...Soft magnetic composites made from metallic magnetic particles with an easy magnetization plane(referred to as easy-plane metallic soft magnetic composites(SMC))are considered ideal materials for the next generation of power electronic devices.This advantage is attributed to their ability to maintain high permeability at elevated frequencies.Despite these advantages,a definitive mathematical model that connects the high-frequency magnetic properties(e.g.,effective permeability)of easy-plane metallic SMCs to the intrinsic properties of the particles is still lacking.In this work,a theoretical calculation model for the effective permeability of easy-plane metallic SMCs was formulated.This model was derived from a skin effect-corrected Landau-Lifshitz-Gilbert(LLG)equation and integrated with effective medium theory incorporating inter-particle interaction.To validate the model,we prepared samples of easy-plane Y_(2)Co_(17)particle/PU SMCs with varying particle sizes and volume fractions.The experimental results showed a strong agreement with the calculated values.This research offers critical theoretical backing for the design and optimization of soft magnetic materials intended for high-frequency applications.展开更多
With the evolution of DC distribution networks from traditional radial topologies to more complex multi-branch structures,the number of measurement points supporting synchronous communication remains relatively limite...With the evolution of DC distribution networks from traditional radial topologies to more complex multi-branch structures,the number of measurement points supporting synchronous communication remains relatively limited.This poses challenges for conventional fault distance estimation methods,which are often tailored to simple topologies and are thus difficult to apply to large-scale,multi-node DC networks.To address this,a fault distance estimation method based on sparse measurement of high-frequency electrical quantities is proposed in this paper.First,a preliminary fault line identification model based on compressed sensing is constructed to effectively narrow the fault search range and improve localization efficiency.Then,leveraging the high-frequency impedance characteristics and the voltage-current relationship of electrical quantities,a fault distance estimation approach based on high-frequency measurements from both ends of a line is designed.This enables accurate distance estimation even when the measurement devices are not directly placed at both ends of the faulted line,overcoming the dependence on specific sensor placement inherent in traditional methods.Finally,to further enhance accuracy,an optimization model based on minimizing the high-frequency voltage error at the fault point is introduced to reduce estimation error.Simulation results demonstrate that the proposed method achieves a fault distance estimation error of less than 1%under normal conditions,and maintains good performance even under adverse scenarios.展开更多
The internal hotspot temperature rise prediction in nanocrystalline high-frequency transformers(nanoHFTs) is essential to ensure reliable operation. This paper presents a three-dimensional thermal network(3DTN) model ...The internal hotspot temperature rise prediction in nanocrystalline high-frequency transformers(nanoHFTs) is essential to ensure reliable operation. This paper presents a three-dimensional thermal network(3DTN) model for epoxy resin encapsulated nano HFTs, which aims to precisely predict the temperature distribution inside the transformer in combination with the finite element method(FEM). A magnetothermal bidirectional coupling 3DTN model is established by analyzing the thermal conduction between the core, windings, and epoxy resin, while also considering the convection and radiation heat transfer mechanisms on the surface of the epoxy resin. The model considers the impact of loss distribution in the core and windings on the temperature field and adopts a simplified 1/2 thermal network model to reduce computational complexity. Furthermore, the results of FEM are compared with experimental results to verify the accuracy of the 3DTN model in predicting the temperature rise of nano HFT. The results show that the 3DTN model reduces errors by an average of 5.25% over the traditional two-dimensional thermal network(2DTN) model, particularly for temperature distributions in the windings and core. This paper provides a temperature rise prediction method for the thermal design and offers a theoretical basis and engineering guidance for the optimization of their thermal management systems.展开更多
Missing data presents a crucial challenge in data analysis,especially in high-dimensional datasets,where missing data often leads to biased conclusions and degraded model performance.In this study,we present a novel a...Missing data presents a crucial challenge in data analysis,especially in high-dimensional datasets,where missing data often leads to biased conclusions and degraded model performance.In this study,we present a novel autoencoder-based imputation framework that integrates a composite loss function to enhance robustness and precision.The proposed loss combines(i)a guided,masked mean squared error focusing on missing entries;(ii)a noise-aware regularization term to improve resilience against data corruption;and(iii)a variance penalty to encourage expressive yet stable reconstructions.We evaluate the proposed model across four missingness mechanisms,such as Missing Completely at Random,Missing at Random,Missing Not at Random,and Missing Not at Random with quantile censorship,under systematically varied feature counts,sample sizes,and missingness ratios ranging from 5%to 60%.Four publicly available real-world datasets(Stroke Prediction,Pima Indians Diabetes,Cardiovascular Disease,and Framingham Heart Study)were used,and the obtained results show that our proposed model consistently outperforms baseline methods,including traditional and deep learning-based techniques.An ablation study reveals the additive value of each component in the loss function.Additionally,we assessed the downstream utility of imputed data through classification tasks,where datasets imputed by the proposed method yielded the highest receiver operating characteristic area under the curve scores across all scenarios.The model demonstrates strong scalability and robustness,improving performance with larger datasets and higher feature counts.These results underscore the capacity of the proposed method to produce not only numerically accurate but also semantically useful imputations,making it a promising solution for robust data recovery in clinical applications.展开更多
Modern intrusion detection systems(MIDS)face persistent challenges in coping with the rapid evolution of cyber threats,high-volume network traffic,and imbalanced datasets.Traditional models often lack the robustness a...Modern intrusion detection systems(MIDS)face persistent challenges in coping with the rapid evolution of cyber threats,high-volume network traffic,and imbalanced datasets.Traditional models often lack the robustness and explainability required to detect novel and sophisticated attacks effectively.This study introduces an advanced,explainable machine learning framework for multi-class IDS using the KDD99 and IDS datasets,which reflects real-world network behavior through a blend of normal and diverse attack classes.The methodology begins with sophisticated data preprocessing,incorporating both RobustScaler and QuantileTransformer to address outliers and skewed feature distributions,ensuring standardized and model-ready inputs.Critical dimensionality reduction is achieved via the Harris Hawks Optimization(HHO)algorithm—a nature-inspired metaheuristic modeled on hawks’hunting strategies.HHO efficiently identifies the most informative features by optimizing a fitness function based on classification performance.Following feature selection,the SMOTE is applied to the training data to resolve class imbalance by synthetically augmenting underrepresented attack types.The stacked architecture is then employed,combining the strengths of XGBoost,SVM,and RF as base learners.This layered approach improves prediction robustness and generalization by balancing bias and variance across diverse classifiers.The model was evaluated using standard classification metrics:precision,recall,F1-score,and overall accuracy.The best overall performance was recorded with an accuracy of 99.44%for UNSW-NB15,demonstrating the model’s effectiveness.After balancing,the model demonstrated a clear improvement in detecting the attacks.We tested the model on four datasets to show the effectiveness of the proposed approach and performed the ablation study to check the effect of each parameter.Also,the proposed model is computationaly efficient.To support transparency and trust in decision-making,explainable AI(XAI)techniques are incorporated that provides both global and local insight into feature contributions,and offers intuitive visualizations for individual predictions.This makes it suitable for practical deployment in cybersecurity environments that demand both precision and accountability.展开更多
Reversible data hiding(RDH)enables secret data embedding while preserving complete cover image recovery,making it crucial for applications requiring image integrity.The pixel value ordering(PVO)technique used in multi...Reversible data hiding(RDH)enables secret data embedding while preserving complete cover image recovery,making it crucial for applications requiring image integrity.The pixel value ordering(PVO)technique used in multi-stego images provides good image quality but often results in low embedding capability.To address these challenges,this paper proposes a high-capacity RDH scheme based on PVO that generates three stego images from a single cover image.The cover image is partitioned into non-overlapping blocks with pixels sorted in ascending order.Four secret bits are embedded into each block’s maximum pixel value,while three additional bits are embedded into the second-largest value when the pixel difference exceeds a predefined threshold.A similar embedding strategy is also applied to the minimum side of the block,including the second-smallest pixel value.This design enables each block to embed up to 14 bits of secret data.Experimental results demonstrate that the proposed method achieves significantly higher embedding capacity and improved visual quality compared to existing triple-stego RDH approaches,advancing the field of reversible steganography.展开更多
The increasing complexity of China’s electricity market creates substantial challenges for settlement automation,data consistency,and operational scalability.Existing provincial settlement systems are fragmented,lack...The increasing complexity of China’s electricity market creates substantial challenges for settlement automation,data consistency,and operational scalability.Existing provincial settlement systems are fragmented,lack a unified data structure,and depend heavily on manual intervention to process high-frequency and retroactive transactions.To address these limitations,a graph-based unified settlement framework is proposed to enhance automation,flexibility,and adaptability in electricity market settlements.A flexible attribute-graph model is employed to represent heterogeneousmulti-market data,enabling standardized integration,rapid querying,and seamless adaptation to evolving business requirements.An extensible operator library is designed to support configurable settlement rules,and a suite of modular tools—including dataset generation,formula configuration,billing templates,and task scheduling—facilitates end-to-end automated settlement processing.A robust refund-clearing mechanism is further incorporated,utilizing sandbox execution,data-version snapshots,dynamic lineage tracing,and real-time changecapture technologies to enable rapid and accurate recalculations under dynamic policy and data revisions.Case studies based on real-world data from regional Chinese markets validate the effectiveness of the proposed approach,demonstrating marked improvements in computational efficiency,system robustness,and automation.Moreover,enhanced settlement accuracy and high temporal granularity improve price-signal fidelity,promote cost-reflective tariffs,and incentivize energy-efficient and demand-responsive behavior among market participants.The method not only supports equitable and transparent market operations but also provides a generalizable,scalable foundation for modern electricity settlement platforms in increasingly complex and dynamic market environments.展开更多
With the increasing emphasis on personal information protection,encryption through security protocols has emerged as a critical requirement in data transmission and reception processes.Nevertheless,IoT ecosystems comp...With the increasing emphasis on personal information protection,encryption through security protocols has emerged as a critical requirement in data transmission and reception processes.Nevertheless,IoT ecosystems comprise heterogeneous networks where outdated systems coexist with the latest devices,spanning a range of devices from non-encrypted ones to fully encrypted ones.Given the limited visibility into payloads in this context,this study investigates AI-based attack detection methods that leverage encrypted traffic metadata,eliminating the need for decryption and minimizing system performance degradation—especially in light of these heterogeneous devices.Using the UNSW-NB15 and CICIoT-2023 dataset,encrypted and unencrypted traffic were categorized according to security protocol,and AI-based intrusion detection experiments were conducted for each traffic type based on metadata.To mitigate the problem of class imbalance,eight different data sampling techniques were applied.The effectiveness of these sampling techniques was then comparatively analyzed using two ensemble models and three Deep Learning(DL)models from various perspectives.The experimental results confirmed that metadata-based attack detection is feasible using only encrypted traffic.In the UNSW-NB15 dataset,the f1-score of encrypted traffic was approximately 0.98,which is 4.3%higher than that of unencrypted traffic(approximately 0.94).In addition,analysis of the encrypted traffic in the CICIoT-2023 dataset using the same method showed a significantly lower f1-score of roughly 0.43,indicating that the quality of the dataset and the preprocessing approach have a substantial impact on detection performance.Furthermore,when data sampling techniques were applied to encrypted traffic,the recall in the UNSWNB15(Encrypted)dataset improved by up to 23.0%,and in the CICIoT-2023(Encrypted)dataset by 20.26%,showing a similar level of improvement.Notably,in CICIoT-2023,f1-score and Receiver Operation Characteristic-Area Under the Curve(ROC-AUC)increased by 59.0%and 55.94%,respectively.These results suggest that data sampling can have a positive effect even in encrypted environments.However,the extent of the improvement may vary depending on data quality,model architecture,and sampling strategy.展开更多
Automated essay scoring(AES)systems have gained significant importance in educational settings,offering a scalable,efficient,and objective method for evaluating student essays.However,developing AES systems for Arabic...Automated essay scoring(AES)systems have gained significant importance in educational settings,offering a scalable,efficient,and objective method for evaluating student essays.However,developing AES systems for Arabic poses distinct challenges due to the language’s complex morphology,diglossia,and the scarcity of annotated datasets.This paper presents a hybrid approach to Arabic AES by combining text-based,vector-based,and embeddingbased similarity measures to improve essay scoring accuracy while minimizing the training data required.Using a large Arabic essay dataset categorized into thematic groups,the study conducted four experiments to evaluate the impact of feature selection,data size,and model performance.Experiment 1 established a baseline using a non-machine learning approach,selecting top-N correlated features to predict essay scores.The subsequent experiments employed 5-fold cross-validation.Experiment 2 showed that combining embedding-based,text-based,and vector-based features in a Random Forest(RF)model achieved an R2 of 88.92%and an accuracy of 83.3%within a 0.5-point tolerance.Experiment 3 further refined the feature selection process,demonstrating that 19 correlated features yielded optimal results,improving R2 to 88.95%.In Experiment 4,an optimal data efficiency training approach was introduced,where training data portions increased from 5%to 50%.The study found that using just 10%of the data achieved near-peak performance,with an R2 of 85.49%,emphasizing an effective trade-off between performance and computational costs.These findings highlight the potential of the hybrid approach for developing scalable Arabic AES systems,especially in low-resource environments,addressing linguistic challenges while ensuring efficient data usage.展开更多
Objective expertise evaluation of individuals,as a prerequisite stage for team formation,has been a long-term desideratum in large software development companies.With the rapid advancements in machine learning methods...Objective expertise evaluation of individuals,as a prerequisite stage for team formation,has been a long-term desideratum in large software development companies.With the rapid advancements in machine learning methods,based on reliable existing data stored in project management tools’datasets,automating this evaluation process becomes a natural step forward.In this context,our approach focuses on quantifying software developer expertise by using metadata from the task-tracking systems.For this,we mathematically formalize two categories of expertise:technology-specific expertise,which denotes the skills required for a particular technology,and general expertise,which encapsulates overall knowledge in the software industry.Afterward,we automatically classify the zones of expertise associated with each task a developer has worked on using Bidirectional Encoder Representations from Transformers(BERT)-like transformers to handle the unique characteristics of project tool datasets effectively.Finally,our method evaluates the proficiency of each software specialist across already completed projects from both technology-specific and general perspectives.The method was experimentally validated,yielding promising results.展开更多
文摘As the crypto-asset ecosystem matures,the use of high-frequency data has become increasingly common in decentralized finance literature.Using bibliometric analysis,we characterize the existing cryptocurrency literature that employs high-frequency data.We highlighted the most influential authors,articles,and journals based on 189 articles from the Scopus database from 2015 to 2022.This approach enables us to identify emerging trends and research hotspots with the aid of co-citation and cartographic analyses.It shows knowledge expansion through authors’collaboration in cryptocurrency research with co-authorship analysis.We identify four major streams of research:(i)return prediction and measurement of cryptocurrency volatility,(ii)(in)efficiency of cryptocurrencies,(iii)price dynamics and bubbles in cryptocurrencies,and(iv)the diversification,safe haven,and hedging properties of Bitcoin.We conclude that highly traded cryptocurrencies’investment features and economic outcomes are analyzed predominantly on a tick-by-tick basis.This study also provides recommendations for future studies.
文摘The modeling of volatility and correlation is important in order to calculate hedge ratios, value at risk estimates, CAPM (Capital Asset Pricing Model betas), derivate pricing and risk management in general. Recent access to intra-daily high-frequency data for two of the most liquid contracts at the Nord Pool exchange has made it possible to apply new and promising methods for analyzing volatility and correlation. The concepts of realized volatility and realized correlation are applied, and this study statistically describes the distribution (both distributional properties and temporal dependencies) of electricity forward data from 2005 to 2009. The main findings show that the logarithmic realized volatility is approximately normally distributed, while realized correlation seems not to be. Further, realized volatility and realized correlation have a long-memory feature. There also seems to be a high correlation between realized correlation and volatilities and positive relations between trading volume and realized volatility and between trading volume and realized correlation. These results are to a large extent consistent with earlier studies of stylized facts of other financial and commodity markets.
基金Supported by the National Natural Science Foundation of China(No.71673315)Foundation of Beijing Technology and Business University(LKJJ2016-03)Capital Circulation Research Base(JD-YB-2017-016)
文摘This paper studies an M-estimator of a proxy periodic GARCH (p, q) scaling model and establishes its consistency and asymptotic normality. Simulation studies are carried out to assess the performance of the estimator. The numerical results show that our M-estimator is more efficient and robust than other estimators without the use of high-frequency data.
基金This research is supported by the National Science and Technology Major Project of China(No.2011ZX05024-001-03)the Natural Science Basic Research Plan in Shaanxi Province of China(No.2021JQ-588)Innovation Fund for graduate students of Xi’an Shiyou University(No.YCS17111017).
文摘The low-pass fi ltering eff ect of the Earth results in the absorption and attenuation of the high-frequency components of seismic signals by the stratum during propagation.Hence,seismic data have low resolution.Considering the limitations of traditional high-frequency compensation methods,this paper presents a new method based on adaptive generalized S transform.This method is based on the study of frequency spectrum attenuation law of seismic signals,and the Gauss window function of adaptive generalized S transform is used to fi t the attenuation trend of seismic signals to seek the optimal Gauss window function.The amplitude spectrum compensation function constructed using the optimal Gauss window function is used to modify the time-frequency spectrum of the adaptive generalized S transform of seismic signals and reconstruct seismic signals to compensate for high-frequency attenuation.Practical data processing results show that the method can compensate for the high-frequency components that are absorbed and attenuated by the stratum,thereby eff ectively improving the resolution and quality of seismic data.
基金the National Natural Science Foundation of China(42074134,42474152,42374150)CNPC Innovation Found(2024DQ02-0152).
文摘The oil-based mud(OBM) borehole measurement environment presents significant limitations on the application of existing electrical logging instruments in high-resistance formations. In this paper, we propose a novel logging method for detection of high-resistance formations in OBM using highfrequency electrodes. The method addresses the issue of shallow depth of investigation(DOI) in existing electrical logging instruments, while simultaneously ensuring the vertical resolution. Based on the principle of current continuity, the total impedance of the loop is obtained by equating the measurement loop to the series form of a capacitively coupled circuit. and its validity is verified in a homogeneous formation model and a radial two-layer formation model with a mud standoff. Then, the instrument operating frequency and electrode system parameters were preferentially determined by numerical simulation, and the effect of mud gap on impedance measurement was investigated. Subsequently, the DOI of the instrument was investigated utilizing the pseudo-geometric factor defined by the real part of impedance. It was determined that the detection depth of the instrument is 8.74 cm, while the effective vertical resolution was not less than 2 cm. Finally, a focused high-frequency electrode-type instrument was designed by introducing a pair of focused electrodes, which effectively enhanced the DOI of the instrument and was successfully deployed in the Oklahoma formation model. The simulation results demonstrate that the novel method can achieve a detection depth of 17.40 cm in highly-resistive formations drilling with OBM, which is approximately twice the depth of detection of the existing oil-based mud microimager instruments. Furthermore, its effective vertical resolution remains at or above 2 cm,which is comparable to the resolution of the existing OBM electrical logging instrument.
文摘Objective:To analyze the significance of high-frequency ultrasound in differentiating benign and malignant breast micronodules.Methods:Eighty-five patients with breast micronodules admitted for diagnosis between October 2022 and October 2024 were selected for high-frequency ultrasound diagnosis.The diagnostic efficacy of high-frequency ultrasound was evaluated by comparing it with the results of surgical pathology.Results:High-frequency ultrasound detected 50 benign nodules,primarily breast fibroadenomas,and 35 malignant nodules,mainly breast ductal carcinoma in situ.Based on surgical pathology results,the diagnostic accuracy of high-frequency ultrasound was 96.47%,specificity was 97.96%,and sensitivity was 94.44%.In high-frequency ultrasound diagnosis,the proportion of grade III and IV blood flow in malignant nodules was higher than that in benign nodules,while the proportion of regular shape and clear margins was lower.The proportion of microcalcifications and posterior echo attenuation was higher in malignant nodules,and the resistance index(RI)and peak blood flow velocity were lower than those in benign nodules(P<0.05).Conclusion:High-frequency ultrasound can effectively differentiate benign and malignant breast micronodules,determine specific nodule types,and exhibits high diagnostic accuracy and sensitivity.Additionally,benign and malignant nodules can be differentiated based on the grading of blood flow signals,sonographic features,and blood flow velocity,providing reasonable guidance for subsequent treatment plans.
基金supported by the science and technology project of State Grid Shanghai Municipal Electric Power Company(No.52094023003L).
文摘New electric power systems characterized by a high proportion of renewable energy and power electronics equipment face significant challenges due to high-frequency(HF)electromagnetic interference from the high-speed switching of power converters.To address this situation,this paper offers an in-depth review of HF interference problems and challenges originating from power electronic devices.First,the root cause of HF electromagnetic interference,i.e.,the resonant response of the parasitic parameters of the system to high-speed switching transients,is analyzed,and various scenarios of HF interference in power systems are highlighted.Next,the types of HF interference are summarized,with a focus on common-mode interference in grounding systems.This paper thoroughly reviews and compares various suppression methods for conducted HF interference.Finally,the challenges involved and suggestions for addressing emerging HF interference problems from the perspective of both power electronics equipment and power systems are discussed.This review aims to offer a structured understanding of HF interference problems and their suppression techniques for researchers and practitioners.
文摘China has a long history of coal mining,among which open-pit coal mines have a large number of small coal mine goafs underground.The distribution,shape,structure and other characteristics of goafs are isolated and discontinuous,and there is no definite geological law to follow,which seriously threatens the safety of coal mine production and personnel life.Conventional ground geophysical methods have low accuracy in detecting goaf areas affected by mechanical interference from open-pit mines,especially for waterless goaf areas,which cannot be detected by existing methods.This article proposes the use of high-frequency electromagnetic waves for goaf detection.The feasibility of using drilling radar to detect goaf was theoretically analyzed,and a goaf detection model was established.The response characteristics of different fillers in the goaf under different frequencies of high-frequency electromagnetic waves were simulated and analyzed.In a certain open-pit mine in Inner Mongolia,100MHz high-frequency electromagnetic waves were used to detect the goaf through directional drilling on the ground.After detection,excavation verification was carried out,and the location of one goaf detected was verified.The results of engineering practice show that the application of high-frequency electromagnetic waves in goaf detection expands the detection radius of boreholes,has the advantages of high efficiency and accuracy,and has important theoretical and practical significance.
基金supported by the National Science Foundation of China(No.12127806,No.62175195 and No.12304382)the International Joint Research Laboratory for Micro/Nano Manufacturing and Measurement Technologies.
文摘Single-shot ultrafast compressed imaging(UCI)is an effective tool for studying ultrafast dynamics in physics,chemistry,or material science because of its excellent high frame rate and large frame number.However,the random code(Rcode)used in traditional UCI will lead to low-frequency noise covering high-frequency information due to its uneven sampling interval,which is a great challenge in the fidelity of large-frame reconstruction.Here,a high-frequency enhanced compressed active photography(H-CAP)is proposed.By uniformizing the sampling interval of R-code,H-CAP capture the ultrafast process with a random uniform sampling mode.This sampling mode makes the high-frequency sampling energy dominant,which greatly suppresses the low-frequency noise blurring caused by R-code and achieves high-frequency information of image enhanced.The superior dynamic performance and large-frame reconstruction ability of H-CAP are verified by imaging optical self-focusing effect and static object,respectively.We applied H-CAP to the spatial-temporal characterization of double-pulse induced silicon surface ablation dynamics,which is performed within 220 frames in a single-shot of 300 ps.H-CAP provides a high-fidelity imaging method for observing ultrafast unrepeatable dynamic processes with large frames.
文摘Objective:To analyze the therapeutic effect of high-frequency electrosurgical knife surgery guided by painless digestive endoscopy(PDE)in elderly patients with gastrointestinal polyps(GP).Methods:A total of 100 elderly GP patients admitted between June 2021 and December 2022 were selected.Patients were randomly divided into two groups:the painless group(50 cases)underwent high-frequency electrosurgical knife surgery guided by PDE,while the conventional group(50 cases)underwent the same surgery guided by traditional digestive endoscopy(DE).The total treatment efficacy,perioperative indicators,gastrointestinal hormone levels,oxidative stress(OS)markers,and complication rates were compared between the two groups.Results:The total treatment efficacy in the painless group was higher than that in the conventional group,and perioperative indicators were superior in the painless group(P<0.05).One week after treatment,the gastrointestinal hormone levels and OS-related markers in the painless group were better than those in the conventional group(P<0.05).The complication rate in the painless group was lower than in the conventional group(P<0.05).Conclusion:High-frequency electrosurgical knife surgery guided by PDE improves the effectiveness of polyp removal in elderly GP patients and accelerates postoperative recovery.It also protects gastrointestinal function,reduces postoperative OS,and ensures higher surgical safety.
基金supported by the National Key R&D Program of China(Grant No.2021YFB3501300)the 9th Research Institute of China Electronics Technology Group Corporation’s open projects(Grant No.2024SK-002-01)the Science and Technology Project of Gansu Province(Grant No.22YF7GA001).
文摘Soft magnetic composites made from metallic magnetic particles with an easy magnetization plane(referred to as easy-plane metallic soft magnetic composites(SMC))are considered ideal materials for the next generation of power electronic devices.This advantage is attributed to their ability to maintain high permeability at elevated frequencies.Despite these advantages,a definitive mathematical model that connects the high-frequency magnetic properties(e.g.,effective permeability)of easy-plane metallic SMCs to the intrinsic properties of the particles is still lacking.In this work,a theoretical calculation model for the effective permeability of easy-plane metallic SMCs was formulated.This model was derived from a skin effect-corrected Landau-Lifshitz-Gilbert(LLG)equation and integrated with effective medium theory incorporating inter-particle interaction.To validate the model,we prepared samples of easy-plane Y_(2)Co_(17)particle/PU SMCs with varying particle sizes and volume fractions.The experimental results showed a strong agreement with the calculated values.This research offers critical theoretical backing for the design and optimization of soft magnetic materials intended for high-frequency applications.
基金National Natural Science Foundation of China, grant number 52177074.
文摘With the evolution of DC distribution networks from traditional radial topologies to more complex multi-branch structures,the number of measurement points supporting synchronous communication remains relatively limited.This poses challenges for conventional fault distance estimation methods,which are often tailored to simple topologies and are thus difficult to apply to large-scale,multi-node DC networks.To address this,a fault distance estimation method based on sparse measurement of high-frequency electrical quantities is proposed in this paper.First,a preliminary fault line identification model based on compressed sensing is constructed to effectively narrow the fault search range and improve localization efficiency.Then,leveraging the high-frequency impedance characteristics and the voltage-current relationship of electrical quantities,a fault distance estimation approach based on high-frequency measurements from both ends of a line is designed.This enables accurate distance estimation even when the measurement devices are not directly placed at both ends of the faulted line,overcoming the dependence on specific sensor placement inherent in traditional methods.Finally,to further enhance accuracy,an optimization model based on minimizing the high-frequency voltage error at the fault point is introduced to reduce estimation error.Simulation results demonstrate that the proposed method achieves a fault distance estimation error of less than 1%under normal conditions,and maintains good performance even under adverse scenarios.
基金supported by the Project of the National Key Research and Development Program of China under Grant 2022YFB2404100。
文摘The internal hotspot temperature rise prediction in nanocrystalline high-frequency transformers(nanoHFTs) is essential to ensure reliable operation. This paper presents a three-dimensional thermal network(3DTN) model for epoxy resin encapsulated nano HFTs, which aims to precisely predict the temperature distribution inside the transformer in combination with the finite element method(FEM). A magnetothermal bidirectional coupling 3DTN model is established by analyzing the thermal conduction between the core, windings, and epoxy resin, while also considering the convection and radiation heat transfer mechanisms on the surface of the epoxy resin. The model considers the impact of loss distribution in the core and windings on the temperature field and adopts a simplified 1/2 thermal network model to reduce computational complexity. Furthermore, the results of FEM are compared with experimental results to verify the accuracy of the 3DTN model in predicting the temperature rise of nano HFT. The results show that the 3DTN model reduces errors by an average of 5.25% over the traditional two-dimensional thermal network(2DTN) model, particularly for temperature distributions in the windings and core. This paper provides a temperature rise prediction method for the thermal design and offers a theoretical basis and engineering guidance for the optimization of their thermal management systems.
文摘Missing data presents a crucial challenge in data analysis,especially in high-dimensional datasets,where missing data often leads to biased conclusions and degraded model performance.In this study,we present a novel autoencoder-based imputation framework that integrates a composite loss function to enhance robustness and precision.The proposed loss combines(i)a guided,masked mean squared error focusing on missing entries;(ii)a noise-aware regularization term to improve resilience against data corruption;and(iii)a variance penalty to encourage expressive yet stable reconstructions.We evaluate the proposed model across four missingness mechanisms,such as Missing Completely at Random,Missing at Random,Missing Not at Random,and Missing Not at Random with quantile censorship,under systematically varied feature counts,sample sizes,and missingness ratios ranging from 5%to 60%.Four publicly available real-world datasets(Stroke Prediction,Pima Indians Diabetes,Cardiovascular Disease,and Framingham Heart Study)were used,and the obtained results show that our proposed model consistently outperforms baseline methods,including traditional and deep learning-based techniques.An ablation study reveals the additive value of each component in the loss function.Additionally,we assessed the downstream utility of imputed data through classification tasks,where datasets imputed by the proposed method yielded the highest receiver operating characteristic area under the curve scores across all scenarios.The model demonstrates strong scalability and robustness,improving performance with larger datasets and higher feature counts.These results underscore the capacity of the proposed method to produce not only numerically accurate but also semantically useful imputations,making it a promising solution for robust data recovery in clinical applications.
基金funded by Princess Nourah bint Abdulrahman University Researchers Supporting Project number(PNURSP2025R104)Princess Nourah bint Abdulrahman University,Riyadh,Saudi Arabia.
文摘Modern intrusion detection systems(MIDS)face persistent challenges in coping with the rapid evolution of cyber threats,high-volume network traffic,and imbalanced datasets.Traditional models often lack the robustness and explainability required to detect novel and sophisticated attacks effectively.This study introduces an advanced,explainable machine learning framework for multi-class IDS using the KDD99 and IDS datasets,which reflects real-world network behavior through a blend of normal and diverse attack classes.The methodology begins with sophisticated data preprocessing,incorporating both RobustScaler and QuantileTransformer to address outliers and skewed feature distributions,ensuring standardized and model-ready inputs.Critical dimensionality reduction is achieved via the Harris Hawks Optimization(HHO)algorithm—a nature-inspired metaheuristic modeled on hawks’hunting strategies.HHO efficiently identifies the most informative features by optimizing a fitness function based on classification performance.Following feature selection,the SMOTE is applied to the training data to resolve class imbalance by synthetically augmenting underrepresented attack types.The stacked architecture is then employed,combining the strengths of XGBoost,SVM,and RF as base learners.This layered approach improves prediction robustness and generalization by balancing bias and variance across diverse classifiers.The model was evaluated using standard classification metrics:precision,recall,F1-score,and overall accuracy.The best overall performance was recorded with an accuracy of 99.44%for UNSW-NB15,demonstrating the model’s effectiveness.After balancing,the model demonstrated a clear improvement in detecting the attacks.We tested the model on four datasets to show the effectiveness of the proposed approach and performed the ablation study to check the effect of each parameter.Also,the proposed model is computationaly efficient.To support transparency and trust in decision-making,explainable AI(XAI)techniques are incorporated that provides both global and local insight into feature contributions,and offers intuitive visualizations for individual predictions.This makes it suitable for practical deployment in cybersecurity environments that demand both precision and accountability.
基金funded by University of Transport and Communications(UTC)under grant number T2025-CN-004.
文摘Reversible data hiding(RDH)enables secret data embedding while preserving complete cover image recovery,making it crucial for applications requiring image integrity.The pixel value ordering(PVO)technique used in multi-stego images provides good image quality but often results in low embedding capability.To address these challenges,this paper proposes a high-capacity RDH scheme based on PVO that generates three stego images from a single cover image.The cover image is partitioned into non-overlapping blocks with pixels sorted in ascending order.Four secret bits are embedded into each block’s maximum pixel value,while three additional bits are embedded into the second-largest value when the pixel difference exceeds a predefined threshold.A similar embedding strategy is also applied to the minimum side of the block,including the second-smallest pixel value.This design enables each block to embed up to 14 bits of secret data.Experimental results demonstrate that the proposed method achieves significantly higher embedding capacity and improved visual quality compared to existing triple-stego RDH approaches,advancing the field of reversible steganography.
基金funded by the Science and Technology Project of State Grid Corporation of China(5108-202355437A-3-2-ZN).
文摘The increasing complexity of China’s electricity market creates substantial challenges for settlement automation,data consistency,and operational scalability.Existing provincial settlement systems are fragmented,lack a unified data structure,and depend heavily on manual intervention to process high-frequency and retroactive transactions.To address these limitations,a graph-based unified settlement framework is proposed to enhance automation,flexibility,and adaptability in electricity market settlements.A flexible attribute-graph model is employed to represent heterogeneousmulti-market data,enabling standardized integration,rapid querying,and seamless adaptation to evolving business requirements.An extensible operator library is designed to support configurable settlement rules,and a suite of modular tools—including dataset generation,formula configuration,billing templates,and task scheduling—facilitates end-to-end automated settlement processing.A robust refund-clearing mechanism is further incorporated,utilizing sandbox execution,data-version snapshots,dynamic lineage tracing,and real-time changecapture technologies to enable rapid and accurate recalculations under dynamic policy and data revisions.Case studies based on real-world data from regional Chinese markets validate the effectiveness of the proposed approach,demonstrating marked improvements in computational efficiency,system robustness,and automation.Moreover,enhanced settlement accuracy and high temporal granularity improve price-signal fidelity,promote cost-reflective tariffs,and incentivize energy-efficient and demand-responsive behavior among market participants.The method not only supports equitable and transparent market operations but also provides a generalizable,scalable foundation for modern electricity settlement platforms in increasingly complex and dynamic market environments.
基金supported by the Institute of Information&Communications Technology Planning&Evaluation(IITP)grant funded by the Korea government(MSIT)(No.RS-2023-00235509Development of security monitoring technology based network behavior against encrypted cyber threats in ICT convergence environment).
文摘With the increasing emphasis on personal information protection,encryption through security protocols has emerged as a critical requirement in data transmission and reception processes.Nevertheless,IoT ecosystems comprise heterogeneous networks where outdated systems coexist with the latest devices,spanning a range of devices from non-encrypted ones to fully encrypted ones.Given the limited visibility into payloads in this context,this study investigates AI-based attack detection methods that leverage encrypted traffic metadata,eliminating the need for decryption and minimizing system performance degradation—especially in light of these heterogeneous devices.Using the UNSW-NB15 and CICIoT-2023 dataset,encrypted and unencrypted traffic were categorized according to security protocol,and AI-based intrusion detection experiments were conducted for each traffic type based on metadata.To mitigate the problem of class imbalance,eight different data sampling techniques were applied.The effectiveness of these sampling techniques was then comparatively analyzed using two ensemble models and three Deep Learning(DL)models from various perspectives.The experimental results confirmed that metadata-based attack detection is feasible using only encrypted traffic.In the UNSW-NB15 dataset,the f1-score of encrypted traffic was approximately 0.98,which is 4.3%higher than that of unencrypted traffic(approximately 0.94).In addition,analysis of the encrypted traffic in the CICIoT-2023 dataset using the same method showed a significantly lower f1-score of roughly 0.43,indicating that the quality of the dataset and the preprocessing approach have a substantial impact on detection performance.Furthermore,when data sampling techniques were applied to encrypted traffic,the recall in the UNSWNB15(Encrypted)dataset improved by up to 23.0%,and in the CICIoT-2023(Encrypted)dataset by 20.26%,showing a similar level of improvement.Notably,in CICIoT-2023,f1-score and Receiver Operation Characteristic-Area Under the Curve(ROC-AUC)increased by 59.0%and 55.94%,respectively.These results suggest that data sampling can have a positive effect even in encrypted environments.However,the extent of the improvement may vary depending on data quality,model architecture,and sampling strategy.
基金funded by Deanship of Graduate studies and Scientific Research at Jouf University under grant No.(DGSSR-2024-02-01264).
文摘Automated essay scoring(AES)systems have gained significant importance in educational settings,offering a scalable,efficient,and objective method for evaluating student essays.However,developing AES systems for Arabic poses distinct challenges due to the language’s complex morphology,diglossia,and the scarcity of annotated datasets.This paper presents a hybrid approach to Arabic AES by combining text-based,vector-based,and embeddingbased similarity measures to improve essay scoring accuracy while minimizing the training data required.Using a large Arabic essay dataset categorized into thematic groups,the study conducted four experiments to evaluate the impact of feature selection,data size,and model performance.Experiment 1 established a baseline using a non-machine learning approach,selecting top-N correlated features to predict essay scores.The subsequent experiments employed 5-fold cross-validation.Experiment 2 showed that combining embedding-based,text-based,and vector-based features in a Random Forest(RF)model achieved an R2 of 88.92%and an accuracy of 83.3%within a 0.5-point tolerance.Experiment 3 further refined the feature selection process,demonstrating that 19 correlated features yielded optimal results,improving R2 to 88.95%.In Experiment 4,an optimal data efficiency training approach was introduced,where training data portions increased from 5%to 50%.The study found that using just 10%of the data achieved near-peak performance,with an R2 of 85.49%,emphasizing an effective trade-off between performance and computational costs.These findings highlight the potential of the hybrid approach for developing scalable Arabic AES systems,especially in low-resource environments,addressing linguistic challenges while ensuring efficient data usage.
基金supported by the project“Romanian Hub for Artificial Intelligence-HRIA”,Smart Growth,Digitization and Financial Instruments Program,2021–2027,MySMIS No.334906.
文摘Objective expertise evaluation of individuals,as a prerequisite stage for team formation,has been a long-term desideratum in large software development companies.With the rapid advancements in machine learning methods,based on reliable existing data stored in project management tools’datasets,automating this evaluation process becomes a natural step forward.In this context,our approach focuses on quantifying software developer expertise by using metadata from the task-tracking systems.For this,we mathematically formalize two categories of expertise:technology-specific expertise,which denotes the skills required for a particular technology,and general expertise,which encapsulates overall knowledge in the software industry.Afterward,we automatically classify the zones of expertise associated with each task a developer has worked on using Bidirectional Encoder Representations from Transformers(BERT)-like transformers to handle the unique characteristics of project tool datasets effectively.Finally,our method evaluates the proficiency of each software specialist across already completed projects from both technology-specific and general perspectives.The method was experimentally validated,yielding promising results.