To study the uncertainty quantification of resonant states in open quantum systems,we developed a Bayesian framework by integrating a reduced basis method(RBM)emulator with the Gamow coupled-channel(GCC)approach.The R...To study the uncertainty quantification of resonant states in open quantum systems,we developed a Bayesian framework by integrating a reduced basis method(RBM)emulator with the Gamow coupled-channel(GCC)approach.The RBM,constructed via eigenvector continuation and trained on both bound and resonant configurations,enables the fast and accurate emulation of resonance properties across the parameter space.To identify the physical resonant states from the emulator’s output,we introduce an overlap-based selection technique that effectively isolates true solutions from background artifacts.By applying this framework to unbound nucleus ^(6)Be,we quantified the model uncertainty in the predicted complex energies.The results demonstrate relative errors of 17.48%in the real part and 8.24%in the imaginary part,while achieving a speedup of four orders of magnitude compared with the full GCC calculations.To further investigate the asymptotic behavior of the resonant-state wavefunctions within the RBM framework,we employed a Lippmann–Schwinger(L–S)-based correction scheme.This approach not only improves the consistency between eigenvalues and wavefunctions but also enables a seamless extension from real-space training data to the complex energy plane.By bridging the gap between bound-state and continuum regimes,the L–S correction significantly enhances the emulator’s capability to accurately capture continuum structures in open quantum systems.展开更多
The incremental capacity analysis(ICA)technique is notably limited by its sensitivity to variations in charging conditions,which constrains its practical applicability in real-world scenarios.This paper introduces an ...The incremental capacity analysis(ICA)technique is notably limited by its sensitivity to variations in charging conditions,which constrains its practical applicability in real-world scenarios.This paper introduces an ICA-compensation technique to address this limitation and propose a generalized framework for assessing the state of health(SOH)of batteries based on ICA that is applicable under differing charging conditions.This novel approach calculates the voltage profile under quasi-static conditions by subtracting the voltage increase attributable to the additional polarization effects at high currents from the measured voltage profile.This approach's efficacy is contingent upon precisely acquiring the equivalent impedance.To obtain the equivalent impedance throughout the batteries'lifespan while minimizing testing costs,this study employs a current interrupt technique in conjunction with a long short-term memory(LSTM)network to develop a predictive model for equivalent impedance.Following the derivation of ICA curves using voltage profiles under quasi-static conditions,the research explores two scenarios for SOH estimation:one utilizing only incremental capacity(IC)features and the other incorporating both IC features and IC sampling.A genetic algorithm-optimized backpropagation neural network(GABPNN)is employed for the SOH estimation.The proposed generalized framework is validated using independent training and test datasets.Variable test conditions are applied for the test set to rigorously evaluate the methodology under challenging conditions.These evaluation results demonstrate that the proposed framework achieves an estimation accuracy of 1.04%for RMSE and 0.90%for MAPE across a spectrum of charging rates ranging from 0.1 C to 1 C and starting SOCs between 0%and 70%,which constitutes a major advancement compared to established ICA methods.It also significantly enhances the applicability of conventional ICA techniques in varying charging conditions and negates the necessity for separate testing protocols for each charging scenario.展开更多
This paper concentrates on methods for comparing activity units found relatively efficient by data envelopment analysis (DEA). The use of the basic DEA models does not provide direct information regarding the performa...This paper concentrates on methods for comparing activity units found relatively efficient by data envelopment analysis (DEA). The use of the basic DEA models does not provide direct information regarding the performance of such units. The paper provides a systematic framework of alternative ways for ranking DEA-efficient units. The framework contains criteria derived as by-products of the basic DEA models and also criteria derived from complementary DEA analysis that needs to be carried out. The proposed framework is applied to rank a set of relatively efficient restaurants on the basis of their market efficiency.展开更多
Electrocardiogram(ECG)is a low-cost,simple,fast,and non-invasive test.It can reflect the heart’s electrical activity and provide valuable diagnostic clues about the health of the entire body.Therefore,ECG has been wi...Electrocardiogram(ECG)is a low-cost,simple,fast,and non-invasive test.It can reflect the heart’s electrical activity and provide valuable diagnostic clues about the health of the entire body.Therefore,ECG has been widely used in various biomedical applications such as arrhythmia detection,disease-specific detection,mortality prediction,and biometric recognition.In recent years,ECG-related studies have been carried out using a variety of publicly available datasets,with many differences in the datasets used,data preprocessing methods,targeted challenges,and modeling and analysis techniques.Here we systematically summarize and analyze the ECGbased automatic analysis methods and applications.Specifically,we first reviewed 22 commonly used ECG public datasets and provided an overview of data preprocessing processes.Then we described some of the most widely used applications of ECG signals and analyzed the advanced methods involved in these applications.Finally,we elucidated some of the challenges in ECG analysis and provided suggestions for further research.展开更多
The development of adaptation measures to climate change relies on data from climate models or impact models. In order to analyze these large data sets or an ensemble of these data sets, the use of statistical methods...The development of adaptation measures to climate change relies on data from climate models or impact models. In order to analyze these large data sets or an ensemble of these data sets, the use of statistical methods is required. In this paper, the methodological approach to collecting, structuring and publishing the methods, which have been used or developed by former or present adaptation initiatives, is described. The intention is to communicate achieved knowledge and thus support future users. A key component is the participation of users in the development process. Main elements of the approach are standardized, template-based descriptions of the methods including the specific applications, references, and method assessment. All contributions have been quality checked, sorted, and placed in a larger context. The result is a report on statistical methods which is freely available as printed or online version. Examples of how to use the methods are presented in this paper and are also included in the brochure.展开更多
In mineral exploration, the apparent resistivity and apparent frequency (or apparent polarizability) parameters of induced polarization method are commonly utilized to describe the induced polarization anomaly. When...In mineral exploration, the apparent resistivity and apparent frequency (or apparent polarizability) parameters of induced polarization method are commonly utilized to describe the induced polarization anomaly. When the target geology structure is significantly complicated, these parameters would fail to reflect the nature of the anomaly source, and wrong conclusions may be obtained. A wavelet approach and a metal factor method were used to comprehensively interpret the induced polarization anomaly of complex geologic bodies in the Adi Bladia mine. Db5 wavelet basis was used to conduct two-scale decomposition and reconstruction, which effectively suppress the noise interference of greenschist facies regional metamorphism and magma intrusion, making energy concentrated and boundary problem unobservable. On the basis of that, the ore-induced anomaly was effectively extracted by the metal factor method.展开更多
We develop various statistical methods important for multidimensional genetic data analysis. Theorems justifying application of these methods are established. We concentrate on the multifactor dimensionality reduction...We develop various statistical methods important for multidimensional genetic data analysis. Theorems justifying application of these methods are established. We concentrate on the multifactor dimensionality reduction, logic regression, random forests, stochastic gradient boosting along with their new modifications. We use complementary approaches to study the risk of complex diseases such as cardiovascular ones. The roles of certain combinations of single nucleotide polymorphisms and non-genetic risk factors are examined. To perform the data analysis concerning the coronary heart disease and myocardial infarction the Lomonosov Moscow State University supercomputer “Chebyshev” was employed.展开更多
Mitigating increasing cyberattack incidents may require strategies such as reinforcing organizations’ networks with Honeypots and effectively analyzing attack traffic for detection of zero-day attacks and vulnerabili...Mitigating increasing cyberattack incidents may require strategies such as reinforcing organizations’ networks with Honeypots and effectively analyzing attack traffic for detection of zero-day attacks and vulnerabilities. To effectively detect and mitigate cyberattacks, both computerized and visual analyses are typically required. However, most security analysts are not adequately trained in visualization principles and/or methods, which is required for effective visual perception of useful attack information hidden in attack data. Additionally, Honeypot has proven useful in cyberattack research, but no studies have comprehensively investigated visualization practices in the field. In this paper, we reviewed visualization practices and methods commonly used in the discovery and communication of attack patterns based on Honeypot network traffic data. Using the PRISMA methodology, we identified and screened 218 papers and evaluated only 37 papers having a high impact. Most Honeypot papers conducted summary statistics of Honeypot data based on static data metrics such as IP address, port, and packet size. They visually analyzed Honeypot attack data using simple graphical methods (such as line, bar, and pie charts) that tend to hide useful attack information. Furthermore, only a few papers conducted extended attack analysis, and commonly visualized attack data using scatter and linear plots. Papers rarely included simple yet sophisticated graphical methods, such as box plots and histograms, which allow for critical evaluation of analysis results. While a significant number of automated visualization tools have incorporated visualization standards by default, the construction of effective and expressive graphical methods for easy pattern discovery and explainable insights still requires applied knowledge and skill of visualization principles and tools, and occasionally, an interdisciplinary collaboration with peers. We, therefore, suggest the need, going forward, for non-classical graphical methods for visualizing attack patterns and communicating analysis results. We also recommend training investigators in visualization principles and standards for effective visual perception and presentation.展开更多
11% of Irish electricity was consumed by data centres in 2020. The Irish data centre industry and the cooling methods utilised require reformative actions in the coming years to meet EU Energy policies. The resell of ...11% of Irish electricity was consumed by data centres in 2020. The Irish data centre industry and the cooling methods utilised require reformative actions in the coming years to meet EU Energy policies. The resell of heat, alternative cooling methods or carbon reduction methods are all possibilities to conform to these policies. This study aims to determine the viability of the resell of waste heat from data centres both technically and economically. This was determined using a novel application of thermodynamics to determine waste heat recovery potential in Irish data centres, and the current methods of heat generation for economical comparison. This paper also explores policy surrounding waste heat recovery within the industry. The Recoverable Carnot Equivalent Power (RCEP) is theoretically calculated for the three potential cooling methods for Irish data centres. These are air, hybrid, and immersion cooling techniques. This is the maximum useable heat that can be recovered from a data centre rack. This study is established under current operating conditions which are optimised for cooling performance, that air cooling has the highest potential RCEP of 0.39 kW/rack. This is approximately 8% of the input electrical power that can be captured as useable heat. Indicating that Irish data centres have the energy potential to be heat providers in the Irish economy. This study highlighted the technical and economic aspects of prevalent cooling techniques and determined air cooling heat recovery cost can be reduced to 0.01 €/kWhth using offsetting. This is financially competitive with current heating solutions in Ireland.展开更多
Gravitational wave detection is one of the most cutting-edge research areas in modern physics, with its success relying on advanced data analysis and signal processing techniques. This study provides a comprehensive r...Gravitational wave detection is one of the most cutting-edge research areas in modern physics, with its success relying on advanced data analysis and signal processing techniques. This study provides a comprehensive review of data analysis methods and signal processing techniques in gravitational wave detection. The research begins by introducing the characteristics of gravitational wave signals and the challenges faced in their detection, such as extremely low signal-to-noise ratios and complex noise backgrounds. It then systematically analyzes the application of time-frequency analysis methods in extracting transient gravitational wave signals, including wavelet transforms and Hilbert-Huang transforms. The study focuses on discussing the crucial role of matched filtering techniques in improving signal detection sensitivity and explores strategies for template bank optimization. Additionally, the research evaluates the potential of machine learning algorithms, especially deep learning networks, in rapidly identifying and classifying gravitational wave events. The study also analyzes the application of Bayesian inference methods in parameter estimation and model selection, as well as their advantages in handling uncertainties. However, the research also points out the challenges faced by current technologies, such as dealing with non-Gaussian noise and improving computational efficiency. To address these issues, the study proposes a hybrid analysis framework combining physical models and data-driven methods. Finally, the research looks ahead to the potential applications of quantum computing in future gravitational wave data analysis. This study provides a comprehensive theoretical foundation for the optimization and innovation of gravitational wave data analysis methods, contributing to the advancement of gravitational wave astronomy.展开更多
This article presents a micro-structure tensor enhanced elasto-plastic finite element(FE)method to address strength anisotropy in three-dimensional(3D)soil slope stability analysis.The gravity increase method(GIM)is e...This article presents a micro-structure tensor enhanced elasto-plastic finite element(FE)method to address strength anisotropy in three-dimensional(3D)soil slope stability analysis.The gravity increase method(GIM)is employed to analyze the stability of 3D anisotropic soil slopes.The accuracy of the proposed method is first verified against the data in the literature.We then simulate the 3D soil slope with a straight slope surface and the convex and concave slope surfaces with a 90turning corner to study the 3D effect on slope stability and the failure mechanism under anisotropy conditions.Based on our numerical results,the end effect significantly impacts the failure mechanism and safety factor.Anisotropy degree notably affects the safety factor,with higher degrees leading to deeper landslides.For concave slopes,they can be approximated by straight slopes with suitable boundary conditions to assess their stability.Furthermore,a case study of the Saint-Alban test embankment A in Quebec,Canada,is provided to demonstrate the applicability of the proposed FE model.展开更多
DNA microarray technology is an extremely effective technique for studying gene expression patterns in cells, and the main challenge currently faced by this technology is how to analyze the large amount of gene expres...DNA microarray technology is an extremely effective technique for studying gene expression patterns in cells, and the main challenge currently faced by this technology is how to analyze the large amount of gene expression data generated. To address this, this paper employs a mixed-effects model to analyze gene expression data. In terms of data selection, 1176 genes from the white mouse gene expression dataset under two experimental conditions were chosen, setting up two conditions: pneumococcal infection and no infection, and constructing a mixed-effects model. After preprocessing the gene chip information, the data were imported into the model, preliminary results were calculated, and permutation tests were performed to biologically validate the preliminary results using GSEA. The final dataset consists of 20 groups of gene expression data from pneumococcal infection, which categorizes functionally related genes based on the similarity of their expression profiles, facilitating the study of genes with unknown functions.展开更多
The issue of strong noise has increasingly become a bottleneck restricting the precision and application space of electromagnetic exploration methods.Noise suppression and extraction of effective electromagnetic respo...The issue of strong noise has increasingly become a bottleneck restricting the precision and application space of electromagnetic exploration methods.Noise suppression and extraction of effective electromagnetic response information under a strong noise background is a crucial scientific task to be addressed.To solve the noise suppression problem of the controlled-source electromagnetic method in strong interference areas,we propose an approach based on complex-plane 2D k-means clustering for data processing.Based on the stability of the controlled-source signal response,clustering analysis is applied to classify the spectra of different sources and noises in multiple time segments.By identifying the power spectra with controlled-source characteristics,it helps to improve the quality of the controlled-source response extraction.This paper presents the principle and workflow of the proposed algorithm,and demonstrates feasibility and effectiveness of the new algorithm through synthetic and real data examples.The results show that,compared with the conventional Robust denoising method,the clustering algorithm has a stronger suppression effect on common noise,can identify high-quality signals,and improve the preprocessing data quality of the controlledsource electromagnetic method.展开更多
With the advent of the big data era,real-time data analysis and decision-support systems have been recognized as essential tools for enhancing enterprise competitiveness and optimizing the decision-making process.This...With the advent of the big data era,real-time data analysis and decision-support systems have been recognized as essential tools for enhancing enterprise competitiveness and optimizing the decision-making process.This study aims to explore the development strategies of real-time data analysis and decision-support systems,and analyze their application status and future development trends in various industries.The article first reviews the basic concepts and importance of real-time data analysis and decision-support systems,and then discusses in detail the key technical aspects such as system architecture,data collection and processing,analysis methods,and visualization techniques.展开更多
This paper analyzes the advantages of legal digital currencies and explores their impact on bank big data practices.By combining bank big data collection and processing,it clarifies that legal digital currencies can e...This paper analyzes the advantages of legal digital currencies and explores their impact on bank big data practices.By combining bank big data collection and processing,it clarifies that legal digital currencies can enhance the efficiency of bank data processing,enrich data types,and strengthen data analysis and application capabilities.In response to future development needs,it is necessary to strengthen data collection management,enhance data processing capabilities,innovate big data application models,and provide references for bank big data practices,promoting the transformation and upgrading of the banking industry in the context of legal digital currencies.展开更多
With the acceleration of intelligent transformation of energy system,the monitoring of equipment operation status and optimization of production process in thermal power plants face the challenge of multi-source heter...With the acceleration of intelligent transformation of energy system,the monitoring of equipment operation status and optimization of production process in thermal power plants face the challenge of multi-source heterogeneous data integration.In view of the heterogeneous characteristics of physical sensor data,including temperature,vibration and pressure that generated by boilers,steam turbines and other key equipment and real-time working condition data of SCADA system,this paper proposes a multi-source heterogeneous data fusion and analysis platform for thermal power plants based on edge computing and deep learning.By constructing a multi-level fusion architecture,the platform adopts dynamic weight allocation strategy and 5D digital twin model to realize the collaborative analysis of physical sensor data,simulation calculation results and expert knowledge.The data fusion module combines Kalman filter,wavelet transform and Bayesian estimation method to solve the problem of data time series alignment and dimension difference.Simulation results show that the data fusion accuracy can be improved to more than 98%,and the calculation delay can be controlled within 500 ms.The data analysis module integrates Dymola simulation model and AERMOD pollutant diffusion model,supports the cascade analysis of boiler combustion efficiency prediction and flue gas emission monitoring,system response time is less than 2 seconds,and data consistency verification accuracy reaches 99.5%.展开更多
With Beijing Huilongguan Sports and Cultural Park as the research object,this study was conducted to investigate public service satisfaction in the park by the Importance-Performance Analysis(IPA)method.A questionnair...With Beijing Huilongguan Sports and Cultural Park as the research object,this study was conducted to investigate public service satisfaction in the park by the Importance-Performance Analysis(IPA)method.A questionnaire covering six dimensions,including public transportation,sanitation and environment,and supporting facility construction,was designed.A total of 208 valid samples were collected,and SPSS was employed for reliability and validity tests as well as IPA analysis.The findings were as follows:①Visitors were generally quite satisfied with the overall public services in Huilongguan Sports and Cultural Park.②The highest satisfaction levels were observed in sanitation and environment services and the sports and cultural atmosphere,while lower satisfaction was noted for supporting facility construction and public information services.③The advantage enhancement zone includes sanitation and environment services and sports and cultural atmosphere;and the continuous maintenance zone includes public transportation services and security management amd maintenance;the subsequent opportunity zone includes supporting facility construction and public information services;and there are no dimensions in the urgent improvement zone.The study recommends strengthening the service connotations from three aspects:enhancing facilities with sports as the core,optimizing services with a people-centered approach,and upgrading the information platform through technological efficiency.Additionally,a multi-stakeholder collaborative mechanism involving the government in coordinating policy resources,the operator in improving implementation efficiency,and the public participating in supervision and evaluation is proposed to drive the enhancement of public service quality at Huilongguan Sports and Cultural Park.展开更多
In the contemporary era,characterized by the Internet and digitalization as fundamental features,the operation and application of digital currency have gradually developed into a comprehensive structural system.This s...In the contemporary era,characterized by the Internet and digitalization as fundamental features,the operation and application of digital currency have gradually developed into a comprehensive structural system.This system restores the essential characteristics of currency while providing auxiliary services related to the formation,circulation,storage,application,and promotion of digital currency.Compared to traditional currency management technologies,big data analysis technology,which is primarily embedded in digital currency systems,enables the rapid acquisition of information.This facilitates the identification of standard associations within currency data and provides technical support for the operational framework of digital currency.展开更多
The fracture volume is gradually changed with the depletion of fracture pressure during the production process.However,there are few flowback models available so far that can estimate the fracture volume loss using pr...The fracture volume is gradually changed with the depletion of fracture pressure during the production process.However,there are few flowback models available so far that can estimate the fracture volume loss using pressure transient and rate transient data.The initial flowback involves producing back the fracturing fuid after hydraulic fracturing,while the second flowback involves producing back the preloading fluid injected into the parent wells before fracturing of child wells.The main objective of this research is to compare the initial and second flowback data to capture the changes in fracture volume after production and preload processes.Such a comparison is useful for evaluating well performance and optimizing frac-turing operations.We construct rate-normalized pressure(RNP)versus material balance time(MBT)diagnostic plots using both initial and second flowback data(FB;and FBs,respectively)of six multi-fractured horizontal wells completed in Niobrara and Codell formations in DJ Basin.In general,the slope of RNP plot during the FB,period is higher than that during the FB;period,indicating a potential loss of fracture volume from the FB;to the FB,period.We estimate the changes in effective fracture volume(Ver)by analyzing the changes in the RNP slope and total compressibility between these two flowback periods.Ver during FB,is in general 3%-45%lower than that during FB:.We also compare the drive mechanisms for the two flowback periods by calculating the compaction-drive index(CDI),hydrocarbon-drive index(HDI),and water-drive index(WDI).The dominant drive mechanism during both flowback periods is CDI,but its contribution is reduced by 16%in the FB,period.This drop is generally compensated by a relatively higher HDI during this period.The loss of effective fracture volume might be attributed to the pressure depletion in fractures,which occurs during the production period and can extend 800 days.展开更多
In the industrial process situation, principal component analysis (PCA) is ageneral method in data reconciliation. However, PCA sometime is unfeasible to nonlinear featureanalysis and limited in application to nonline...In the industrial process situation, principal component analysis (PCA) is ageneral method in data reconciliation. However, PCA sometime is unfeasible to nonlinear featureanalysis and limited in application to nonlinear industrial process. Kernel PCA (KPCA) is extensionof PCA and can be used for nonlinear feature analysis. A nonlinear data reconciliation method basedon KPCA is proposed. The basic idea of this method is that firstly original data are mapped to highdimensional feature space by nonlinear function, and PCA is implemented in the feature space. Thennonlinear feature analysis is implemented and data are reconstructed by using the kernel. The datareconciliation method based on KPCA is applied to ternary distillation column. Simulation resultsshow that this method can filter the noise in measurements of nonlinear process and reconciliateddata can represent the true information of nonlinear process.展开更多
基金supported by the National Key Research and Development Program(MOST 2023YFA1606404 and MOST 2022YFA1602303)the National Natural Science Foundation of China(Nos.12347106,12147101,and 12447122)the China Postdoctoral Science Foundation(No.2024M760489).
文摘To study the uncertainty quantification of resonant states in open quantum systems,we developed a Bayesian framework by integrating a reduced basis method(RBM)emulator with the Gamow coupled-channel(GCC)approach.The RBM,constructed via eigenvector continuation and trained on both bound and resonant configurations,enables the fast and accurate emulation of resonance properties across the parameter space.To identify the physical resonant states from the emulator’s output,we introduce an overlap-based selection technique that effectively isolates true solutions from background artifacts.By applying this framework to unbound nucleus ^(6)Be,we quantified the model uncertainty in the predicted complex energies.The results demonstrate relative errors of 17.48%in the real part and 8.24%in the imaginary part,while achieving a speedup of four orders of magnitude compared with the full GCC calculations.To further investigate the asymptotic behavior of the resonant-state wavefunctions within the RBM framework,we employed a Lippmann–Schwinger(L–S)-based correction scheme.This approach not only improves the consistency between eigenvalues and wavefunctions but also enables a seamless extension from real-space training data to the complex energy plane.By bridging the gap between bound-state and continuum regimes,the L–S correction significantly enhances the emulator’s capability to accurately capture continuum structures in open quantum systems.
基金funded by the Bavarian State Ministry of ScienceResearch and Art(Grant number:H.2-F1116.WE/52/2)。
文摘The incremental capacity analysis(ICA)technique is notably limited by its sensitivity to variations in charging conditions,which constrains its practical applicability in real-world scenarios.This paper introduces an ICA-compensation technique to address this limitation and propose a generalized framework for assessing the state of health(SOH)of batteries based on ICA that is applicable under differing charging conditions.This novel approach calculates the voltage profile under quasi-static conditions by subtracting the voltage increase attributable to the additional polarization effects at high currents from the measured voltage profile.This approach's efficacy is contingent upon precisely acquiring the equivalent impedance.To obtain the equivalent impedance throughout the batteries'lifespan while minimizing testing costs,this study employs a current interrupt technique in conjunction with a long short-term memory(LSTM)network to develop a predictive model for equivalent impedance.Following the derivation of ICA curves using voltage profiles under quasi-static conditions,the research explores two scenarios for SOH estimation:one utilizing only incremental capacity(IC)features and the other incorporating both IC features and IC sampling.A genetic algorithm-optimized backpropagation neural network(GABPNN)is employed for the SOH estimation.The proposed generalized framework is validated using independent training and test datasets.Variable test conditions are applied for the test set to rigorously evaluate the methodology under challenging conditions.These evaluation results demonstrate that the proposed framework achieves an estimation accuracy of 1.04%for RMSE and 0.90%for MAPE across a spectrum of charging rates ranging from 0.1 C to 1 C and starting SOCs between 0%and 70%,which constitutes a major advancement compared to established ICA methods.It also significantly enhances the applicability of conventional ICA techniques in varying charging conditions and negates the necessity for separate testing protocols for each charging scenario.
文摘This paper concentrates on methods for comparing activity units found relatively efficient by data envelopment analysis (DEA). The use of the basic DEA models does not provide direct information regarding the performance of such units. The paper provides a systematic framework of alternative ways for ranking DEA-efficient units. The framework contains criteria derived as by-products of the basic DEA models and also criteria derived from complementary DEA analysis that needs to be carried out. The proposed framework is applied to rank a set of relatively efficient restaurants on the basis of their market efficiency.
基金Supported by the NSFC-Zhejiang Joint Fund for the Integration of Industrialization and Informatization(U1909208)the Science and Technology Major Project of Changsha(kh2202004)the Changsha Municipal Natural Science Foundation(kq2202106).
文摘Electrocardiogram(ECG)is a low-cost,simple,fast,and non-invasive test.It can reflect the heart’s electrical activity and provide valuable diagnostic clues about the health of the entire body.Therefore,ECG has been widely used in various biomedical applications such as arrhythmia detection,disease-specific detection,mortality prediction,and biometric recognition.In recent years,ECG-related studies have been carried out using a variety of publicly available datasets,with many differences in the datasets used,data preprocessing methods,targeted challenges,and modeling and analysis techniques.Here we systematically summarize and analyze the ECGbased automatic analysis methods and applications.Specifically,we first reviewed 22 commonly used ECG public datasets and provided an overview of data preprocessing processes.Then we described some of the most widely used applications of ECG signals and analyzed the advanced methods involved in these applications.Finally,we elucidated some of the challenges in ECG analysis and provided suggestions for further research.
文摘The development of adaptation measures to climate change relies on data from climate models or impact models. In order to analyze these large data sets or an ensemble of these data sets, the use of statistical methods is required. In this paper, the methodological approach to collecting, structuring and publishing the methods, which have been used or developed by former or present adaptation initiatives, is described. The intention is to communicate achieved knowledge and thus support future users. A key component is the participation of users in the development process. Main elements of the approach are standardized, template-based descriptions of the methods including the specific applications, references, and method assessment. All contributions have been quality checked, sorted, and placed in a larger context. The result is a report on statistical methods which is freely available as printed or online version. Examples of how to use the methods are presented in this paper and are also included in the brochure.
基金Project(41174103)supported by the National Natural Science Foundation of ChinaProject(2010-211)supported by the Foreign Mineral Resources Venture Exploration Special Fund of China
文摘In mineral exploration, the apparent resistivity and apparent frequency (or apparent polarizability) parameters of induced polarization method are commonly utilized to describe the induced polarization anomaly. When the target geology structure is significantly complicated, these parameters would fail to reflect the nature of the anomaly source, and wrong conclusions may be obtained. A wavelet approach and a metal factor method were used to comprehensively interpret the induced polarization anomaly of complex geologic bodies in the Adi Bladia mine. Db5 wavelet basis was used to conduct two-scale decomposition and reconstruction, which effectively suppress the noise interference of greenschist facies regional metamorphism and magma intrusion, making energy concentrated and boundary problem unobservable. On the basis of that, the ore-induced anomaly was effectively extracted by the metal factor method.
文摘We develop various statistical methods important for multidimensional genetic data analysis. Theorems justifying application of these methods are established. We concentrate on the multifactor dimensionality reduction, logic regression, random forests, stochastic gradient boosting along with their new modifications. We use complementary approaches to study the risk of complex diseases such as cardiovascular ones. The roles of certain combinations of single nucleotide polymorphisms and non-genetic risk factors are examined. To perform the data analysis concerning the coronary heart disease and myocardial infarction the Lomonosov Moscow State University supercomputer “Chebyshev” was employed.
文摘Mitigating increasing cyberattack incidents may require strategies such as reinforcing organizations’ networks with Honeypots and effectively analyzing attack traffic for detection of zero-day attacks and vulnerabilities. To effectively detect and mitigate cyberattacks, both computerized and visual analyses are typically required. However, most security analysts are not adequately trained in visualization principles and/or methods, which is required for effective visual perception of useful attack information hidden in attack data. Additionally, Honeypot has proven useful in cyberattack research, but no studies have comprehensively investigated visualization practices in the field. In this paper, we reviewed visualization practices and methods commonly used in the discovery and communication of attack patterns based on Honeypot network traffic data. Using the PRISMA methodology, we identified and screened 218 papers and evaluated only 37 papers having a high impact. Most Honeypot papers conducted summary statistics of Honeypot data based on static data metrics such as IP address, port, and packet size. They visually analyzed Honeypot attack data using simple graphical methods (such as line, bar, and pie charts) that tend to hide useful attack information. Furthermore, only a few papers conducted extended attack analysis, and commonly visualized attack data using scatter and linear plots. Papers rarely included simple yet sophisticated graphical methods, such as box plots and histograms, which allow for critical evaluation of analysis results. While a significant number of automated visualization tools have incorporated visualization standards by default, the construction of effective and expressive graphical methods for easy pattern discovery and explainable insights still requires applied knowledge and skill of visualization principles and tools, and occasionally, an interdisciplinary collaboration with peers. We, therefore, suggest the need, going forward, for non-classical graphical methods for visualizing attack patterns and communicating analysis results. We also recommend training investigators in visualization principles and standards for effective visual perception and presentation.
文摘11% of Irish electricity was consumed by data centres in 2020. The Irish data centre industry and the cooling methods utilised require reformative actions in the coming years to meet EU Energy policies. The resell of heat, alternative cooling methods or carbon reduction methods are all possibilities to conform to these policies. This study aims to determine the viability of the resell of waste heat from data centres both technically and economically. This was determined using a novel application of thermodynamics to determine waste heat recovery potential in Irish data centres, and the current methods of heat generation for economical comparison. This paper also explores policy surrounding waste heat recovery within the industry. The Recoverable Carnot Equivalent Power (RCEP) is theoretically calculated for the three potential cooling methods for Irish data centres. These are air, hybrid, and immersion cooling techniques. This is the maximum useable heat that can be recovered from a data centre rack. This study is established under current operating conditions which are optimised for cooling performance, that air cooling has the highest potential RCEP of 0.39 kW/rack. This is approximately 8% of the input electrical power that can be captured as useable heat. Indicating that Irish data centres have the energy potential to be heat providers in the Irish economy. This study highlighted the technical and economic aspects of prevalent cooling techniques and determined air cooling heat recovery cost can be reduced to 0.01 €/kWhth using offsetting. This is financially competitive with current heating solutions in Ireland.
文摘Gravitational wave detection is one of the most cutting-edge research areas in modern physics, with its success relying on advanced data analysis and signal processing techniques. This study provides a comprehensive review of data analysis methods and signal processing techniques in gravitational wave detection. The research begins by introducing the characteristics of gravitational wave signals and the challenges faced in their detection, such as extremely low signal-to-noise ratios and complex noise backgrounds. It then systematically analyzes the application of time-frequency analysis methods in extracting transient gravitational wave signals, including wavelet transforms and Hilbert-Huang transforms. The study focuses on discussing the crucial role of matched filtering techniques in improving signal detection sensitivity and explores strategies for template bank optimization. Additionally, the research evaluates the potential of machine learning algorithms, especially deep learning networks, in rapidly identifying and classifying gravitational wave events. The study also analyzes the application of Bayesian inference methods in parameter estimation and model selection, as well as their advantages in handling uncertainties. However, the research also points out the challenges faced by current technologies, such as dealing with non-Gaussian noise and improving computational efficiency. To address these issues, the study proposes a hybrid analysis framework combining physical models and data-driven methods. Finally, the research looks ahead to the potential applications of quantum computing in future gravitational wave data analysis. This study provides a comprehensive theoretical foundation for the optimization and innovation of gravitational wave data analysis methods, contributing to the advancement of gravitational wave astronomy.
基金supported by the National Natural Science Foundation of China(Grant Nos.51890912,51979025 and 52011530189).
文摘This article presents a micro-structure tensor enhanced elasto-plastic finite element(FE)method to address strength anisotropy in three-dimensional(3D)soil slope stability analysis.The gravity increase method(GIM)is employed to analyze the stability of 3D anisotropic soil slopes.The accuracy of the proposed method is first verified against the data in the literature.We then simulate the 3D soil slope with a straight slope surface and the convex and concave slope surfaces with a 90turning corner to study the 3D effect on slope stability and the failure mechanism under anisotropy conditions.Based on our numerical results,the end effect significantly impacts the failure mechanism and safety factor.Anisotropy degree notably affects the safety factor,with higher degrees leading to deeper landslides.For concave slopes,they can be approximated by straight slopes with suitable boundary conditions to assess their stability.Furthermore,a case study of the Saint-Alban test embankment A in Quebec,Canada,is provided to demonstrate the applicability of the proposed FE model.
文摘DNA microarray technology is an extremely effective technique for studying gene expression patterns in cells, and the main challenge currently faced by this technology is how to analyze the large amount of gene expression data generated. To address this, this paper employs a mixed-effects model to analyze gene expression data. In terms of data selection, 1176 genes from the white mouse gene expression dataset under two experimental conditions were chosen, setting up two conditions: pneumococcal infection and no infection, and constructing a mixed-effects model. After preprocessing the gene chip information, the data were imported into the model, preliminary results were calculated, and permutation tests were performed to biologically validate the preliminary results using GSEA. The final dataset consists of 20 groups of gene expression data from pneumococcal infection, which categorizes functionally related genes based on the similarity of their expression profiles, facilitating the study of genes with unknown functions.
基金supported by the National Key Research and Development Program Project of China(Grant No.2023YFF0718003)the key research and development plan project of Yunnan Province(Grant No.202303AA080006).
文摘The issue of strong noise has increasingly become a bottleneck restricting the precision and application space of electromagnetic exploration methods.Noise suppression and extraction of effective electromagnetic response information under a strong noise background is a crucial scientific task to be addressed.To solve the noise suppression problem of the controlled-source electromagnetic method in strong interference areas,we propose an approach based on complex-plane 2D k-means clustering for data processing.Based on the stability of the controlled-source signal response,clustering analysis is applied to classify the spectra of different sources and noises in multiple time segments.By identifying the power spectra with controlled-source characteristics,it helps to improve the quality of the controlled-source response extraction.This paper presents the principle and workflow of the proposed algorithm,and demonstrates feasibility and effectiveness of the new algorithm through synthetic and real data examples.The results show that,compared with the conventional Robust denoising method,the clustering algorithm has a stronger suppression effect on common noise,can identify high-quality signals,and improve the preprocessing data quality of the controlledsource electromagnetic method.
文摘With the advent of the big data era,real-time data analysis and decision-support systems have been recognized as essential tools for enhancing enterprise competitiveness and optimizing the decision-making process.This study aims to explore the development strategies of real-time data analysis and decision-support systems,and analyze their application status and future development trends in various industries.The article first reviews the basic concepts and importance of real-time data analysis and decision-support systems,and then discusses in detail the key technical aspects such as system architecture,data collection and processing,analysis methods,and visualization techniques.
文摘This paper analyzes the advantages of legal digital currencies and explores their impact on bank big data practices.By combining bank big data collection and processing,it clarifies that legal digital currencies can enhance the efficiency of bank data processing,enrich data types,and strengthen data analysis and application capabilities.In response to future development needs,it is necessary to strengthen data collection management,enhance data processing capabilities,innovate big data application models,and provide references for bank big data practices,promoting the transformation and upgrading of the banking industry in the context of legal digital currencies.
文摘With the acceleration of intelligent transformation of energy system,the monitoring of equipment operation status and optimization of production process in thermal power plants face the challenge of multi-source heterogeneous data integration.In view of the heterogeneous characteristics of physical sensor data,including temperature,vibration and pressure that generated by boilers,steam turbines and other key equipment and real-time working condition data of SCADA system,this paper proposes a multi-source heterogeneous data fusion and analysis platform for thermal power plants based on edge computing and deep learning.By constructing a multi-level fusion architecture,the platform adopts dynamic weight allocation strategy and 5D digital twin model to realize the collaborative analysis of physical sensor data,simulation calculation results and expert knowledge.The data fusion module combines Kalman filter,wavelet transform and Bayesian estimation method to solve the problem of data time series alignment and dimension difference.Simulation results show that the data fusion accuracy can be improved to more than 98%,and the calculation delay can be controlled within 500 ms.The data analysis module integrates Dymola simulation model and AERMOD pollutant diffusion model,supports the cascade analysis of boiler combustion efficiency prediction and flue gas emission monitoring,system response time is less than 2 seconds,and data consistency verification accuracy reaches 99.5%.
基金Sponsored by The Youth Project of National Social Science Foundation of China(21CTY007)Special Fund for Basic Scientific Research Business Expenses of Central Universities(2024DAWH008).
文摘With Beijing Huilongguan Sports and Cultural Park as the research object,this study was conducted to investigate public service satisfaction in the park by the Importance-Performance Analysis(IPA)method.A questionnaire covering six dimensions,including public transportation,sanitation and environment,and supporting facility construction,was designed.A total of 208 valid samples were collected,and SPSS was employed for reliability and validity tests as well as IPA analysis.The findings were as follows:①Visitors were generally quite satisfied with the overall public services in Huilongguan Sports and Cultural Park.②The highest satisfaction levels were observed in sanitation and environment services and the sports and cultural atmosphere,while lower satisfaction was noted for supporting facility construction and public information services.③The advantage enhancement zone includes sanitation and environment services and sports and cultural atmosphere;and the continuous maintenance zone includes public transportation services and security management amd maintenance;the subsequent opportunity zone includes supporting facility construction and public information services;and there are no dimensions in the urgent improvement zone.The study recommends strengthening the service connotations from three aspects:enhancing facilities with sports as the core,optimizing services with a people-centered approach,and upgrading the information platform through technological efficiency.Additionally,a multi-stakeholder collaborative mechanism involving the government in coordinating policy resources,the operator in improving implementation efficiency,and the public participating in supervision and evaluation is proposed to drive the enhancement of public service quality at Huilongguan Sports and Cultural Park.
文摘In the contemporary era,characterized by the Internet and digitalization as fundamental features,the operation and application of digital currency have gradually developed into a comprehensive structural system.This system restores the essential characteristics of currency while providing auxiliary services related to the formation,circulation,storage,application,and promotion of digital currency.Compared to traditional currency management technologies,big data analysis technology,which is primarily embedded in digital currency systems,enables the rapid acquisition of information.This facilitates the identification of standard associations within currency data and provides technical support for the operational framework of digital currency.
文摘The fracture volume is gradually changed with the depletion of fracture pressure during the production process.However,there are few flowback models available so far that can estimate the fracture volume loss using pressure transient and rate transient data.The initial flowback involves producing back the fracturing fuid after hydraulic fracturing,while the second flowback involves producing back the preloading fluid injected into the parent wells before fracturing of child wells.The main objective of this research is to compare the initial and second flowback data to capture the changes in fracture volume after production and preload processes.Such a comparison is useful for evaluating well performance and optimizing frac-turing operations.We construct rate-normalized pressure(RNP)versus material balance time(MBT)diagnostic plots using both initial and second flowback data(FB;and FBs,respectively)of six multi-fractured horizontal wells completed in Niobrara and Codell formations in DJ Basin.In general,the slope of RNP plot during the FB,period is higher than that during the FB;period,indicating a potential loss of fracture volume from the FB;to the FB,period.We estimate the changes in effective fracture volume(Ver)by analyzing the changes in the RNP slope and total compressibility between these two flowback periods.Ver during FB,is in general 3%-45%lower than that during FB:.We also compare the drive mechanisms for the two flowback periods by calculating the compaction-drive index(CDI),hydrocarbon-drive index(HDI),and water-drive index(WDI).The dominant drive mechanism during both flowback periods is CDI,but its contribution is reduced by 16%in the FB,period.This drop is generally compensated by a relatively higher HDI during this period.The loss of effective fracture volume might be attributed to the pressure depletion in fractures,which occurs during the production period and can extend 800 days.
基金This project is supported by Special Foundation for Major State Basic Research of China (Project 973, No.G1998030415)
文摘In the industrial process situation, principal component analysis (PCA) is ageneral method in data reconciliation. However, PCA sometime is unfeasible to nonlinear featureanalysis and limited in application to nonlinear industrial process. Kernel PCA (KPCA) is extensionof PCA and can be used for nonlinear feature analysis. A nonlinear data reconciliation method basedon KPCA is proposed. The basic idea of this method is that firstly original data are mapped to highdimensional feature space by nonlinear function, and PCA is implemented in the feature space. Thennonlinear feature analysis is implemented and data are reconstructed by using the kernel. The datareconciliation method based on KPCA is applied to ternary distillation column. Simulation resultsshow that this method can filter the noise in measurements of nonlinear process and reconciliateddata can represent the true information of nonlinear process.