Accurately evaluating the lifespan of the Printed Circuit Board(PCB)in airborne equipment is an essential issue for aircraft design and operation in the marine atmospheric environment.This paper presents a novel evalu...Accurately evaluating the lifespan of the Printed Circuit Board(PCB)in airborne equipment is an essential issue for aircraft design and operation in the marine atmospheric environment.This paper presents a novel evaluation method by fusing Accelerated Degradation Testing(ADT)data,degradation data,and life data of small samples based on the uncertainty degradation process.An uncertain life model of PCB in airborne equipment is constructed by employing the uncertain distribution that considers the accelerated factor of multiple environmental conditions such as temperature,humidity,and salinity.In addition,a degradation process model of PCB in airborne equipment is constructed by employing the uncertain process of fusing ADT data and field data,in which the performance characteristics of dynamic cumulative change are included.Based on minimizing the pth sample moments,an integrated method for parameter estimation of the PCB in airborne equipment is proposed by fusing the multi-source data of life,degradation,and ADT.An engineering case illustrates the effectiveness and advantage of the proposed method.展开更多
Data compression plays a vital role in datamanagement and information theory by reducing redundancy.However,it lacks built-in security features such as secret keys or password-based access control,leaving sensitive da...Data compression plays a vital role in datamanagement and information theory by reducing redundancy.However,it lacks built-in security features such as secret keys or password-based access control,leaving sensitive data vulnerable to unauthorized access and misuse.With the exponential growth of digital data,robust security measures are essential.Data encryption,a widely used approach,ensures data confidentiality by making it unreadable and unalterable through secret key control.Despite their individual benefits,both require significant computational resources.Additionally,performing them separately for the same data increases complexity and processing time.Recognizing the need for integrated approaches that balance compression ratios and security levels,this research proposes an integrated data compression and encryption algorithm,named IDCE,for enhanced security and efficiency.Thealgorithmoperates on 128-bit block sizes and a 256-bit secret key length.It combines Huffman coding for compression and a Tent map for encryption.Additionally,an iterative Arnold cat map further enhances cryptographic confusion properties.Experimental analysis validates the effectiveness of the proposed algorithm,showcasing competitive performance in terms of compression ratio,security,and overall efficiency when compared to prior algorithms in the field.展开更多
The study of plant diversity is often hindered by the challenge of integrating data from different sources and different data types.A standardized data system would facilitate detailed exploration of plant distributio...The study of plant diversity is often hindered by the challenge of integrating data from different sources and different data types.A standardized data system would facilitate detailed exploration of plant distribution patterns and dynamics for botanists,ecologists,conservation biologists,and biogeographers.This study proposes a gridded vector data integration method,combining grid-based techniques with vectorization to integrate diverse data types from multiple sources into grids of the same scale.Here we demonstrate the methodology by creating a comprehensive 1°×1°database of western China that includes plant distribution information and environmental factor data.This approach addresses the need for a standardized data system to facilitate exploration of plant distribution patterns and dynamic changes in the region.展开更多
Digital twin is a novel technology that has achieved significant progress in industrial manufactur-ing systems in recent years.In the digital twin envi-ronment,entities in the virtual space collect data from devices i...Digital twin is a novel technology that has achieved significant progress in industrial manufactur-ing systems in recent years.In the digital twin envi-ronment,entities in the virtual space collect data from devices in the physical space to analyze their states.However,since a lot of devices exist in the physical space,the digital twin system needs to aggregate data from multiple devices at the edge gateway.Homomor-phic integrity and confidentiality protections are two important requirements for this data aggregation pro-cess.Unfortunately,existing homomorphic encryp-tion algorithms do not support integrity protection,and existing homomorphic signing algorithms require all signers to use the same signing key,which is not feasible in the digital twin environment.Moreover,for both integrity and confidentiality protections,the homomorphic signing algorithm must be compatible with the aggregation manner of the homomorphic en-cryption algorithm.To address these issues,this paper designs a novel homomorphic aggregation scheme,which allows multiple devices in the physical space to sign different data using different keys and support in-tegrity and confidentiality protections.Finally,the security of the newly designed scheme is analyzed,and its efficiency is evaluated.Experimental results show that our scheme is feasible for real world applications.展开更多
The distillation process is an important chemical process,and the application of data-driven modelling approach has the potential to reduce model complexity compared to mechanistic modelling,thus improving the efficie...The distillation process is an important chemical process,and the application of data-driven modelling approach has the potential to reduce model complexity compared to mechanistic modelling,thus improving the efficiency of process optimization or monitoring studies.However,the distillation process is highly nonlinear and has multiple uncertainty perturbation intervals,which brings challenges to accurate data-driven modelling of distillation processes.This paper proposes a systematic data-driven modelling framework to solve these problems.Firstly,data segment variance was introduced into the K-means algorithm to form K-means data interval(KMDI)clustering in order to cluster the data into perturbed and steady state intervals for steady-state data extraction.Secondly,maximal information coefficient(MIC)was employed to calculate the nonlinear correlation between variables for removing redundant features.Finally,extreme gradient boosting(XGBoost)was integrated as the basic learner into adaptive boosting(AdaBoost)with the error threshold(ET)set to improve weights update strategy to construct the new integrated learning algorithm,XGBoost-AdaBoost-ET.The superiority of the proposed framework is verified by applying this data-driven modelling framework to a real industrial process of propylene distillation.展开更多
Lead(Pb)plays a significant role in the nuclear industry and is extensively used in radiation shielding,radiation protection,neutron moderation,radiation measurements,and various other critical functions.Consequently,...Lead(Pb)plays a significant role in the nuclear industry and is extensively used in radiation shielding,radiation protection,neutron moderation,radiation measurements,and various other critical functions.Consequently,the measurement and evaluation of Pb nuclear data are highly regarded in nuclear scientific research,emphasizing its crucial role in the field.Using the time-of-flight(ToF)method,the neutron leakage spectra from three^(nat)Pb samples were measured at 60°and 120°based on the neutronics integral experimental facility at the China Institute of Atomic Energy(CIAE).The^(nat)Pb sample sizes were30 cm×30 cm×5 cm,30 cm×30 cm×10 cm,and 30 cm×30 cm×15 cm.Neutron sources were generated by the Cockcroft-Walton accelerator,producing approximately 14.5 MeV and 3.5 MeV neutrons through the T(d,n)^(4)He and D(d,n)^(3)He reactions,respectively.Leakage neutron spectra were also calculated by employing the Monte Carlo code of MCNP-4C,and the nuclear data of Pb isotopes from four libraries:CENDL-3.2,JEFF-3.3,JENDL-5,and ENDF/B-Ⅷ.0 were used individually.By comparing the simulation and experimental results,improvements and deficiencies in the evaluated nuclear data of the Pb isotopes were analyzed.Most of the calculated results were consistent with the experimental results;however,a few areas did not fit well.In the(n,el)energy range,the simulated results from CENDL-3.2 were significantly overestimated;in the(n,inl)D and the(n,inl)C energy regions,the results from CENDL-3.2 and ENDF/B-Ⅷ.0 were significantly overestimated at 120°,and the results from JENDL-5 and JEFF-3.3 are underestimated at 60°in the(n,inl)D energy region.The calculated spectra were analyzed by comparing them with the experimental spectra in terms of the neutron spectrum shape and C/E values.The results indicate that the theoretical simulations,using different data libraries,overestimated or underestimated the measured values in certain energy ranges.Secondary neutron energies and angular distributions in the data files have been presented to explain these discrepancies.展开更多
Plant morphogenesis relies on precise gene expression programs at the proper time and position which is orchestrated by transcription factors(TFs)in intricate regulatory networks in a cell-type specific manner.Here we...Plant morphogenesis relies on precise gene expression programs at the proper time and position which is orchestrated by transcription factors(TFs)in intricate regulatory networks in a cell-type specific manner.Here we introduced a comprehensive single-cell transcriptomic atlas of Arabidopsis seedlings.This atlas is the result of meticulous integration of 63 previously published scRNA-seq datasets,addressing batch effects and conserving biological variance.This integration spans a broad spectrum of tissues,including both below-and above-ground parts.Utilizing a rigorous approach for cell type annotation,we identified 47 distinct cell types or states,largely expanding our current view of plant cell compositions.We systematically constructed cell-type specific gene regulatory networks and uncovered key regulators that act in a coordinated manner to control cell-type specific gene expression.Taken together,our study not only offers extensive plant cell atlas exploration that serves as a valuable resource,but also provides molecular insights into gene-regulatory programs that varies from different cell types.展开更多
Bioinformatic analysis of large and complex omics datasets has become increasingly useful in modern day biology by providing a great depth of information,with its application to neuroscience termed neuroinformatics.Da...Bioinformatic analysis of large and complex omics datasets has become increasingly useful in modern day biology by providing a great depth of information,with its application to neuroscience termed neuroinformatics.Data mining of omics datasets has enabled the generation of new hypotheses based on differentially regulated biological molecules associated with disease mechanisms,which can be tested experimentally for improved diagnostic and therapeutic targeting of neurodegenerative diseases.Importantly,integrating multi-omics data using a systems bioinformatics approach will advance the understanding of the layered and interactive network of biological regulation that exchanges systemic knowledge to facilitate the development of a comprehensive human brain profile.In this review,we first summarize data mining studies utilizing datasets from the individual type of omics analysis,including epigenetics/epigenomics,transcriptomics,proteomics,metabolomics,lipidomics,and spatial omics,pertaining to Alzheimer's disease,Parkinson's disease,and multiple sclerosis.We then discuss multi-omics integration approaches,including independent biological integration and unsupervised integration methods,for more intuitive and informative interpretation of the biological data obtained across different omics layers.We further assess studies that integrate multi-omics in data mining which provide convoluted biological insights and offer proof-of-concept proposition towards systems bioinformatics in the reconstruction of brain networks.Finally,we recommend a combination of high dimensional bioinformatics analysis with experimental validation to achieve translational neuroscience applications including biomarker discovery,therapeutic development,and elucidation of disease mechanisms.We conclude by providing future perspectives and opportunities in applying integrative multi-omics and systems bioinformatics to achieve precision phenotyping of neurodegenerative diseases and towards personalized medicine.展开更多
To solve the query processing correctness problem for semantic-based relational data integration,the semantics of SAPRQL(simple protocol and RDF query language) queries is defined.In the course of query rewriting,al...To solve the query processing correctness problem for semantic-based relational data integration,the semantics of SAPRQL(simple protocol and RDF query language) queries is defined.In the course of query rewriting,all relative tables are found and decomposed into minimal connectable units.Minimal connectable units are joined according to semantic queries to produce the semantically correct query plans.Algorithms for query rewriting and transforming are presented.Computational complexity of the algorithms is discussed.Under the worst case,the query decomposing algorithm can be finished in O(n2) time and the query rewriting algorithm requires O(nm) time.And the performance of the algorithms is verified by experiments,and experimental results show that when the length of query is less than 8,the query processing algorithms can provide satisfactory performance.展开更多
Background Mastitis caused by multiple factors remains one of the most common and costly disease of the dairy industry.Multi-omics approaches enable the comprehensive investigation of the complex interactions between ...Background Mastitis caused by multiple factors remains one of the most common and costly disease of the dairy industry.Multi-omics approaches enable the comprehensive investigation of the complex interactions between mul-tiple layers of information to provide a more holistic view of disease pathogenesis.Therefore,this study investigated the genomic and epigenomic signatures and the possible regulatory mechanisms underlying subclinical mastitis by integrating RNA sequencing data(mRNA and lncRNA),small RNA sequencing data(miRNA)and DNA methylation sequencing data of milk somatic cells from 10 healthy cows and 20 cows with naturally occurring subclinical mastitis caused by Staphylococcus aureus or Staphylococcus chromogenes.Results Functional investigation of the data sets through gene set analysis uncovered 3458 biological process GO terms and 170 KEGG pathways with altered activities during subclinical mastitis,provided further insights into subclin-ical mastitis and revealed the involvement of multi-omics signatures in the altered immune responses and impaired mammary gland productivity during subclinical mastitis.The abundant genomic and epigenomic signatures with sig-nificant alterations related to subclinical mastitis were observed,including 30,846,2552,1276 and 57 differential methylation haplotype blocks(dMHBs),differentially expressed genes(DEGs),lncRNAs(DELs)and miRNAs(DEMs),respectively.Next,5 factors presenting the principal variation of differential multi-omics signatures were identified.The important roles of Factor 1(DEG,DEM and DEL)and Factor 2(dMHB and DEM),in the regulation of immune defense and impaired mammary gland functions during subclinical mastitis were revealed.Each of the omics within Factors 1 and 2 explained about 20%of the source of variation in subclinical mastitis.Also,networks of impor-tant functional gene sets with the involvement of multi-omics signatures were demonstrated,which contributed to a comprehensive view of the possible regulatory mechanisms underlying subclinical mastitis.Furthermore,multi-omics integration enabled the association of the epigenomic regulatory factors(dMHBs,DELs and DEMs)of altered genes in important pathways,such as‘Staphylococcus aureus infection pathway’and‘natural killer cell mediated cyto-toxicity pathway’,etc.,which provides further insights into mastitis regulatory mechanisms.Moreover,few multi-omics signatures(14 dMHBs,25 DEGs,18 DELs and 5 DEMs)were identified as candidate discriminant signatures with capac-ity of distinguishing subclinical mastitis cows from healthy cows.Conclusion The integration of genomic and epigenomic data by multi-omics approaches in this study provided a better understanding of the molecular mechanisms underlying subclinical mastitis and identified multi-omics candidate discriminant signatures for subclinical mastitis,which may ultimately lead to the development of more effective mastitis control and management strategies.展开更多
Integrated data and energy transfer(IDET)enables the electromagnetic waves to transmit wireless energy at the same time of data delivery for lowpower devices.In this paper,an energy harvesting modulation(EHM)assisted ...Integrated data and energy transfer(IDET)enables the electromagnetic waves to transmit wireless energy at the same time of data delivery for lowpower devices.In this paper,an energy harvesting modulation(EHM)assisted multi-user IDET system is studied,where all the received signals at the users are exploited for energy harvesting without the degradation of wireless data transfer(WDT)performance.The joint IDET performance is then analysed theoretically by conceiving a practical time-dependent wireless channel.With the aid of the AO based algorithm,the average effective data rate among users are maximized by ensuring the BER and the wireless energy transfer(WET)performance.Simulation results validate and evaluate the IDET performance of the EHM assisted system,which also demonstrates that the optimal number of user clusters and IDET time slots should be allocated,in order to improve the WET and WDT performance.展开更多
Industrial big data integration and sharing(IBDIS)is of great significance in managing and providing data for big data analysis in manufacturing systems.A novel fog-computing-based IBDIS approach called Fog-IBDIS is p...Industrial big data integration and sharing(IBDIS)is of great significance in managing and providing data for big data analysis in manufacturing systems.A novel fog-computing-based IBDIS approach called Fog-IBDIS is proposed in order to integrate and share industrial big data with high raw data security and low network traffic loads by moving the integration task from the cloud to the edge of networks.First,a task flow graph(TFG)is designed to model the data analysis process.The TFG is composed of several tasks,which are executed by the data owners through the Fog-IBDIS platform in order to protect raw data privacy.Second,the function of Fog-IBDIS to enable data integration and sharing is presented in five modules:TFG management,compilation and running control,the data integration model,the basic algorithm library,and the management component.Finally,a case study is presented to illustrate the implementation of Fog-IBDIS,which ensures raw data security by deploying the analysis tasks executed by the data generators,and eases the network traffic load by greatly reducing the volume of transmitted data.展开更多
With the rapid development of Web, there are more and more Web databases available for users to access. At the same time, job searchers often have difficulties in first finding the right sources and then querying over...With the rapid development of Web, there are more and more Web databases available for users to access. At the same time, job searchers often have difficulties in first finding the right sources and then querying over them, providing such an integrated job search system over Web databases has become a Web application in high demand. Based on such consideration, we build a deep Web data integration system that supports unified access for users to multiple job Web sites as a job meta-search engine. In this paper, the architecture of the system is given first, and the key components in the system are introduced.展开更多
This paper analyzes the status of existing resources through extensive research and international cooperation on the basis of four typical global monthly surface temperature datasets including the climate research dat...This paper analyzes the status of existing resources through extensive research and international cooperation on the basis of four typical global monthly surface temperature datasets including the climate research dataset of the University of East Anglia(CRUTEM3), the dataset of the U.S. National Climatic Data Center(GHCN-V3), the dataset of the U.S. National Aeronautics and Space Administration(GISSTMP), and the Berkeley Earth surface temperature dataset(Berkeley). China's first global monthly temperature dataset over land was developed by integrating the four aforementioned global temperature datasets and several regional datasets from major countries or regions. This dataset contains information from 9,519 stations worldwide of at least 20 years for monthly mean temperature, 7,073 for maximum temperature, and 6,587 for minimum temperature. Compared with CRUTEM3 and GHCN-V3, the station density is much higher particularly for South America, Africa,and Asia. Moreover, data from significantly more stations were available after the year 1990 which dramatically reduced the uncertainty of the estimated global temperature trend during 1990e2011. The integrated dataset can serve as a reliable data source for global climate change research.展开更多
Cloud computing has emerged as a viable alternative to traditional computing infrastructures,offering various benefits.However,the adoption of cloud storage poses significant risks to data secrecy and integrity.This a...Cloud computing has emerged as a viable alternative to traditional computing infrastructures,offering various benefits.However,the adoption of cloud storage poses significant risks to data secrecy and integrity.This article presents an effective mechanism to preserve the secrecy and integrity of data stored on the public cloud by leveraging blockchain technology,smart contracts,and cryptographic primitives.The proposed approach utilizes a Solidity-based smart contract as an auditor for maintaining and verifying the integrity of outsourced data.To preserve data secrecy,symmetric encryption systems are employed to encrypt user data before outsourcing it.An extensive performance analysis is conducted to illustrate the efficiency of the proposed mechanism.Additionally,a rigorous assessment is conducted to ensure that the developed smart contract is free from vulnerabilities and to measure its associated running costs.The security analysis of the proposed system confirms that our approach can securely maintain the confidentiality and integrity of cloud storage,even in the presence of malicious entities.The proposed mechanism contributes to enhancing data security in cloud computing environments and can be used as a foundation for developing more secure cloud storage systems.展开更多
Currently,there is a growing trend among users to store their data in the cloud.However,the cloud is vulnerable to persistent data corruption risks arising from equipment failures and hacker attacks.Additionally,when ...Currently,there is a growing trend among users to store their data in the cloud.However,the cloud is vulnerable to persistent data corruption risks arising from equipment failures and hacker attacks.Additionally,when users perform file operations,the semantic integrity of the data can be compromised.Ensuring both data integrity and semantic correctness has become a critical issue that requires attention.We introduce a pioneering solution called Sec-Auditor,the first of its kind with the ability to verify data integrity and semantic correctness simultaneously,while maintaining a constant communication cost independent of the audited data volume.Sec-Auditor also supports public auditing,enabling anyone with access to public information to conduct data audits.This feature makes Sec-Auditor highly adaptable to open data environments,such as the cloud.In Sec-Auditor,users are assigned specific rules that are utilized to verify the accuracy of data semantic.Furthermore,users are given the flexibility to update their own rules as needed.We conduct in-depth analyses of the correctness and security of Sec-Auditor.We also compare several important security attributes with existing schemes,demonstrating the superior properties of Sec-Auditor.Evaluation results demonstrate that even for time-consuming file upload operations,our solution is more efficient than the comparison one.展开更多
Background:Presently,multi-omics data(e.g.,genomics,transcriptomics,proteomics,and metabolomics)are available to improve genomic predictors.Omics data not only offers new data layers for genomic prediction but also pr...Background:Presently,multi-omics data(e.g.,genomics,transcriptomics,proteomics,and metabolomics)are available to improve genomic predictors.Omics data not only offers new data layers for genomic prediction but also provides a bridge between organismal phenotypes and genome variation that cannot be readily captured at the genome sequence level.Therefore,using multi-omics data to select feature markers is a feasible strategy to improve the accuracy of genomic prediction.In this study,simultaneously using whole-genome sequencing(WGS)and gene expression level data,four strategies for single-nucleotide polymorphism(SNP)preselection were investigated for genomic predictions in the Drosophila Genetic Reference Panel.Results:Using genomic best linear unbiased prediction(GBLUP)with complete WGS data,the prediction accuracies were 0.208±0.020(0.181±0.022)for the startle response and 0.272±0.017(0.307±0.015)for starvation resistance in the female(male)lines.Compared with GBLUP using complete WGS data,both GBLUP and the genomic feature BLUP(GFBLUP)did not improve the prediction accuracy using SNPs preselected from complete WGS data based on the results of genome-wide association studies(GWASs)or transcriptome-wide association studies(TWASs).Furthermore,by using SNPs preselected from the WGS data based on the results of the expression quantitative trait locus(eQTL)mapping of all genes,only the startle response had greater accuracy than GBLUP with the complete WGS data.The best accuracy values in the female and male lines were 0.243±0.020 and 0.220±0.022,respectively.Importantly,by using SNPs preselected based on the results of the eQTL mapping of significant genes from TWAS,both GBLUP and GFBLUP resulted in great accuracy and small bias of genomic prediction.Compared with the GBLUP using complete WGS data,the best accuracy values represented increases of 60.66%and 39.09%for the starvation resistance and 27.40%and 35.36%for startle response in the female and male lines,respectively.Conclusions:Overall,multi-omics data can assist genomic feature preselection and improve the performance of genomic prediction.The new knowledge gained from this study will enrich the use of multi-omics in genomic prediction.展开更多
With the deepening informationization of Resources & Environment Remote Sensing geological survey conducted,some potential problems and deficiency are:(1) shortage of unified-planed running environment;(2) inconsi...With the deepening informationization of Resources & Environment Remote Sensing geological survey conducted,some potential problems and deficiency are:(1) shortage of unified-planed running environment;(2) inconsistent methods of data integration;and(3) disadvantages of different performing ways of data integration.This paper solves the above problems through overall planning and design,constructs unified running environment, consistent methods of data integration and system structure in order to advance the informationization展开更多
Technical and accessibility issues in hospitals often prevent patients from receiving optimal mental and physical health care,which is essential for independent living,especially as societies age and chronic diseases ...Technical and accessibility issues in hospitals often prevent patients from receiving optimal mental and physical health care,which is essential for independent living,especially as societies age and chronic diseases like diabetes and cardiovascular disease become more common.Recent advances in the Internet of Things(IoT)-enabled wearable devices offer potential solutions for remote health monitoring and everyday activity recognition,gaining significant attention in personalized healthcare.This paper comprehensively reviews wearable healthcare technology integrated with the IoT for continuous vital sign monitoring.Relevant papers were extracted and analyzed using a systematic numerical review method,covering various aspects such as sports monitoring,disease detection,patient monitoring,and medical diagnosis.The review highlights the transformative impact of IoTenabled wearable devices in healthcare,facilitating real-time monitoring of vital signs,including blood pressure,temperature,oxygen levels,and heart rate.Results from the reviewed papers demonstrate high accuracy and efficiency in predicting health conditions,improving sports performance,enhancing patient care,and diagnosing diseases.The integration of IoT in wearable healthcare devices enables remote patient monitoring,personalized care,and efficient data transmission,ultimately transcending traditional boundaries of healthcare and leading to better patient outcomes.展开更多
Microsoft Excel is essential for the End-User Approach (EUA), offering versatility in data organization, analysis, and visualization, as well as widespread accessibility. It fosters collaboration and informed decision...Microsoft Excel is essential for the End-User Approach (EUA), offering versatility in data organization, analysis, and visualization, as well as widespread accessibility. It fosters collaboration and informed decision-making across diverse domains. Conversely, Python is indispensable for professional programming due to its versatility, readability, extensive libraries, and robust community support. It enables efficient development, advanced data analysis, data mining, and automation, catering to diverse industries and applications. However, one primary issue when using Microsoft Excel with Python libraries is compatibility and interoperability. While Excel is a widely used tool for data storage and analysis, it may not seamlessly integrate with Python libraries, leading to challenges in reading and writing data, especially in complex or large datasets. Additionally, manipulating Excel files with Python may not always preserve formatting or formulas accurately, potentially affecting data integrity. Moreover, dependency on Excel’s graphical user interface (GUI) for automation can limit scalability and reproducibility compared to Python’s scripting capabilities. This paper covers the integration solution of empowering non-programmers to leverage Python’s capabilities within the familiar Excel environment. This enables users to perform advanced data analysis and automation tasks without requiring extensive programming knowledge. Based on Soliciting feedback from non-programmers who have tested the integration solution, the case study shows how the solution evaluates the ease of implementation, performance, and compatibility of Python with Excel versions.展开更多
基金supported by the National Natural Science Foundation of China(No.62073009).
文摘Accurately evaluating the lifespan of the Printed Circuit Board(PCB)in airborne equipment is an essential issue for aircraft design and operation in the marine atmospheric environment.This paper presents a novel evaluation method by fusing Accelerated Degradation Testing(ADT)data,degradation data,and life data of small samples based on the uncertainty degradation process.An uncertain life model of PCB in airborne equipment is constructed by employing the uncertain distribution that considers the accelerated factor of multiple environmental conditions such as temperature,humidity,and salinity.In addition,a degradation process model of PCB in airborne equipment is constructed by employing the uncertain process of fusing ADT data and field data,in which the performance characteristics of dynamic cumulative change are included.Based on minimizing the pth sample moments,an integrated method for parameter estimation of the PCB in airborne equipment is proposed by fusing the multi-source data of life,degradation,and ADT.An engineering case illustrates the effectiveness and advantage of the proposed method.
基金the Deanship of Graduate Studies and Scientific Research at Qassim University for financial support(QU-APC-2025).
文摘Data compression plays a vital role in datamanagement and information theory by reducing redundancy.However,it lacks built-in security features such as secret keys or password-based access control,leaving sensitive data vulnerable to unauthorized access and misuse.With the exponential growth of digital data,robust security measures are essential.Data encryption,a widely used approach,ensures data confidentiality by making it unreadable and unalterable through secret key control.Despite their individual benefits,both require significant computational resources.Additionally,performing them separately for the same data increases complexity and processing time.Recognizing the need for integrated approaches that balance compression ratios and security levels,this research proposes an integrated data compression and encryption algorithm,named IDCE,for enhanced security and efficiency.Thealgorithmoperates on 128-bit block sizes and a 256-bit secret key length.It combines Huffman coding for compression and a Tent map for encryption.Additionally,an iterative Arnold cat map further enhances cryptographic confusion properties.Experimental analysis validates the effectiveness of the proposed algorithm,showcasing competitive performance in terms of compression ratio,security,and overall efficiency when compared to prior algorithms in the field.
基金supported by the Second Tibetan Plateau Scientific Expedition and Research(STEP)program(2019QZKK0502)the National Natural Science Foundation of China(32322006)+1 种基金the Major Program for Basic Research Project of Yunnan Province(202103AF140005 and 202101BC070002)the Practice Innovation Fund for Professional Degree Graduates of Yunnan University(ZC-22222401).
文摘The study of plant diversity is often hindered by the challenge of integrating data from different sources and different data types.A standardized data system would facilitate detailed exploration of plant distribution patterns and dynamics for botanists,ecologists,conservation biologists,and biogeographers.This study proposes a gridded vector data integration method,combining grid-based techniques with vectorization to integrate diverse data types from multiple sources into grids of the same scale.Here we demonstrate the methodology by creating a comprehensive 1°×1°database of western China that includes plant distribution information and environmental factor data.This approach addresses the need for a standardized data system to facilitate exploration of plant distribution patterns and dynamic changes in the region.
基金supported by ZTE Industry-University-Institute Cooperation Funds under Grant No.IA20230628015the State Key Laboratory of Particle Detection and Electronics under Grant No.SKLPDE-KF-202314.
文摘Digital twin is a novel technology that has achieved significant progress in industrial manufactur-ing systems in recent years.In the digital twin envi-ronment,entities in the virtual space collect data from devices in the physical space to analyze their states.However,since a lot of devices exist in the physical space,the digital twin system needs to aggregate data from multiple devices at the edge gateway.Homomor-phic integrity and confidentiality protections are two important requirements for this data aggregation pro-cess.Unfortunately,existing homomorphic encryp-tion algorithms do not support integrity protection,and existing homomorphic signing algorithms require all signers to use the same signing key,which is not feasible in the digital twin environment.Moreover,for both integrity and confidentiality protections,the homomorphic signing algorithm must be compatible with the aggregation manner of the homomorphic en-cryption algorithm.To address these issues,this paper designs a novel homomorphic aggregation scheme,which allows multiple devices in the physical space to sign different data using different keys and support in-tegrity and confidentiality protections.Finally,the security of the newly designed scheme is analyzed,and its efficiency is evaluated.Experimental results show that our scheme is feasible for real world applications.
基金supported by the National Key Research and Development Program of China(2023YFB3307801)the National Natural Science Foundation of China(62394343,62373155,62073142)+3 种基金Major Science and Technology Project of Xinjiang(No.2022A01006-4)the Programme of Introducing Talents of Discipline to Universities(the 111 Project)under Grant B17017the Fundamental Research Funds for the Central Universities,Science Foundation of China University of Petroleum,Beijing(No.2462024YJRC011)the Open Research Project of the State Key Laboratory of Industrial Control Technology,China(Grant No.ICT2024B70).
文摘The distillation process is an important chemical process,and the application of data-driven modelling approach has the potential to reduce model complexity compared to mechanistic modelling,thus improving the efficiency of process optimization or monitoring studies.However,the distillation process is highly nonlinear and has multiple uncertainty perturbation intervals,which brings challenges to accurate data-driven modelling of distillation processes.This paper proposes a systematic data-driven modelling framework to solve these problems.Firstly,data segment variance was introduced into the K-means algorithm to form K-means data interval(KMDI)clustering in order to cluster the data into perturbed and steady state intervals for steady-state data extraction.Secondly,maximal information coefficient(MIC)was employed to calculate the nonlinear correlation between variables for removing redundant features.Finally,extreme gradient boosting(XGBoost)was integrated as the basic learner into adaptive boosting(AdaBoost)with the error threshold(ET)set to improve weights update strategy to construct the new integrated learning algorithm,XGBoost-AdaBoost-ET.The superiority of the proposed framework is verified by applying this data-driven modelling framework to a real industrial process of propylene distillation.
基金supported by the National Natural Science Foundation of China(Nos.11775311 and U2067205)the Stable Support Basic Research Program Grant(BJ010261223282)the Research and Development Project of China National Nuclear Corporation。
文摘Lead(Pb)plays a significant role in the nuclear industry and is extensively used in radiation shielding,radiation protection,neutron moderation,radiation measurements,and various other critical functions.Consequently,the measurement and evaluation of Pb nuclear data are highly regarded in nuclear scientific research,emphasizing its crucial role in the field.Using the time-of-flight(ToF)method,the neutron leakage spectra from three^(nat)Pb samples were measured at 60°and 120°based on the neutronics integral experimental facility at the China Institute of Atomic Energy(CIAE).The^(nat)Pb sample sizes were30 cm×30 cm×5 cm,30 cm×30 cm×10 cm,and 30 cm×30 cm×15 cm.Neutron sources were generated by the Cockcroft-Walton accelerator,producing approximately 14.5 MeV and 3.5 MeV neutrons through the T(d,n)^(4)He and D(d,n)^(3)He reactions,respectively.Leakage neutron spectra were also calculated by employing the Monte Carlo code of MCNP-4C,and the nuclear data of Pb isotopes from four libraries:CENDL-3.2,JEFF-3.3,JENDL-5,and ENDF/B-Ⅷ.0 were used individually.By comparing the simulation and experimental results,improvements and deficiencies in the evaluated nuclear data of the Pb isotopes were analyzed.Most of the calculated results were consistent with the experimental results;however,a few areas did not fit well.In the(n,el)energy range,the simulated results from CENDL-3.2 were significantly overestimated;in the(n,inl)D and the(n,inl)C energy regions,the results from CENDL-3.2 and ENDF/B-Ⅷ.0 were significantly overestimated at 120°,and the results from JENDL-5 and JEFF-3.3 are underestimated at 60°in the(n,inl)D energy region.The calculated spectra were analyzed by comparing them with the experimental spectra in terms of the neutron spectrum shape and C/E values.The results indicate that the theoretical simulations,using different data libraries,overestimated or underestimated the measured values in certain energy ranges.Secondary neutron energies and angular distributions in the data files have been presented to explain these discrepancies.
基金supported by the National Natural Science Foundation of China (No.32070656)the Nanjing University Deng Feng Scholars Program+1 种基金the Priority Academic Program Development (PAPD) of Jiangsu Higher Education Institutions,China Postdoctoral Science Foundation funded project (No.2022M711563)Jiangsu Funding Program for Excellent Postdoctoral Talent (No.2022ZB50)
文摘Plant morphogenesis relies on precise gene expression programs at the proper time and position which is orchestrated by transcription factors(TFs)in intricate regulatory networks in a cell-type specific manner.Here we introduced a comprehensive single-cell transcriptomic atlas of Arabidopsis seedlings.This atlas is the result of meticulous integration of 63 previously published scRNA-seq datasets,addressing batch effects and conserving biological variance.This integration spans a broad spectrum of tissues,including both below-and above-ground parts.Utilizing a rigorous approach for cell type annotation,we identified 47 distinct cell types or states,largely expanding our current view of plant cell compositions.We systematically constructed cell-type specific gene regulatory networks and uncovered key regulators that act in a coordinated manner to control cell-type specific gene expression.Taken together,our study not only offers extensive plant cell atlas exploration that serves as a valuable resource,but also provides molecular insights into gene-regulatory programs that varies from different cell types.
基金supported by a Lee Kong Chian School of Medicine Dean’s Postdoctoral Fellowship(021207-00001)from Nanyang Technological University(NTU)Singapore and a Mistletoe Research Fellowship(022522-00001)from the Momental Foundation USA.Jialiu Zeng is supported by a Presidential Postdoctoral Fellowship(021229-00001)from NTU Singapore and an Open Fund Young Investigator Research Grant(OF-YIRG)(MOH-001147)from the National Medical Research Council(NMRC)SingaporeSu Bin Lim is supported by the National Research Foundation(NRF)of Korea(Grant Nos.:2020R1A6A1A03043539,2020M3A9D8037604,2022R1C1C1004756)a grant of the Korea Health Technology R&D Project through the Korea Health Industry Development Institute(KHIDI),funded by the Ministry of Health&Welfare,Republic of Korea(Grant No.:HR22C1734).
文摘Bioinformatic analysis of large and complex omics datasets has become increasingly useful in modern day biology by providing a great depth of information,with its application to neuroscience termed neuroinformatics.Data mining of omics datasets has enabled the generation of new hypotheses based on differentially regulated biological molecules associated with disease mechanisms,which can be tested experimentally for improved diagnostic and therapeutic targeting of neurodegenerative diseases.Importantly,integrating multi-omics data using a systems bioinformatics approach will advance the understanding of the layered and interactive network of biological regulation that exchanges systemic knowledge to facilitate the development of a comprehensive human brain profile.In this review,we first summarize data mining studies utilizing datasets from the individual type of omics analysis,including epigenetics/epigenomics,transcriptomics,proteomics,metabolomics,lipidomics,and spatial omics,pertaining to Alzheimer's disease,Parkinson's disease,and multiple sclerosis.We then discuss multi-omics integration approaches,including independent biological integration and unsupervised integration methods,for more intuitive and informative interpretation of the biological data obtained across different omics layers.We further assess studies that integrate multi-omics in data mining which provide convoluted biological insights and offer proof-of-concept proposition towards systems bioinformatics in the reconstruction of brain networks.Finally,we recommend a combination of high dimensional bioinformatics analysis with experimental validation to achieve translational neuroscience applications including biomarker discovery,therapeutic development,and elucidation of disease mechanisms.We conclude by providing future perspectives and opportunities in applying integrative multi-omics and systems bioinformatics to achieve precision phenotyping of neurodegenerative diseases and towards personalized medicine.
基金Weaponry Equipment Pre-Research Foundation of PLA Equipment Ministry (No. 9140A06050409JB8102)Pre-Research Foundation of PLA University of Science and Technology (No. 2009JSJ11)
文摘To solve the query processing correctness problem for semantic-based relational data integration,the semantics of SAPRQL(simple protocol and RDF query language) queries is defined.In the course of query rewriting,all relative tables are found and decomposed into minimal connectable units.Minimal connectable units are joined according to semantic queries to produce the semantically correct query plans.Algorithms for query rewriting and transforming are presented.Computational complexity of the algorithms is discussed.Under the worst case,the query decomposing algorithm can be finished in O(n2) time and the query rewriting algorithm requires O(nm) time.And the performance of the algorithms is verified by experiments,and experimental results show that when the length of query is less than 8,the query processing algorithms can provide satisfactory performance.
基金The help and support of owners of the dairy farms enrolled in this study is gratefully acknowledged.The financial support from the program of China Scholarship Council during the PhD study of Mengqi Wang in Canada is acknowledged(No.202008880009).
文摘Background Mastitis caused by multiple factors remains one of the most common and costly disease of the dairy industry.Multi-omics approaches enable the comprehensive investigation of the complex interactions between mul-tiple layers of information to provide a more holistic view of disease pathogenesis.Therefore,this study investigated the genomic and epigenomic signatures and the possible regulatory mechanisms underlying subclinical mastitis by integrating RNA sequencing data(mRNA and lncRNA),small RNA sequencing data(miRNA)and DNA methylation sequencing data of milk somatic cells from 10 healthy cows and 20 cows with naturally occurring subclinical mastitis caused by Staphylococcus aureus or Staphylococcus chromogenes.Results Functional investigation of the data sets through gene set analysis uncovered 3458 biological process GO terms and 170 KEGG pathways with altered activities during subclinical mastitis,provided further insights into subclin-ical mastitis and revealed the involvement of multi-omics signatures in the altered immune responses and impaired mammary gland productivity during subclinical mastitis.The abundant genomic and epigenomic signatures with sig-nificant alterations related to subclinical mastitis were observed,including 30,846,2552,1276 and 57 differential methylation haplotype blocks(dMHBs),differentially expressed genes(DEGs),lncRNAs(DELs)and miRNAs(DEMs),respectively.Next,5 factors presenting the principal variation of differential multi-omics signatures were identified.The important roles of Factor 1(DEG,DEM and DEL)and Factor 2(dMHB and DEM),in the regulation of immune defense and impaired mammary gland functions during subclinical mastitis were revealed.Each of the omics within Factors 1 and 2 explained about 20%of the source of variation in subclinical mastitis.Also,networks of impor-tant functional gene sets with the involvement of multi-omics signatures were demonstrated,which contributed to a comprehensive view of the possible regulatory mechanisms underlying subclinical mastitis.Furthermore,multi-omics integration enabled the association of the epigenomic regulatory factors(dMHBs,DELs and DEMs)of altered genes in important pathways,such as‘Staphylococcus aureus infection pathway’and‘natural killer cell mediated cyto-toxicity pathway’,etc.,which provides further insights into mastitis regulatory mechanisms.Moreover,few multi-omics signatures(14 dMHBs,25 DEGs,18 DELs and 5 DEMs)were identified as candidate discriminant signatures with capac-ity of distinguishing subclinical mastitis cows from healthy cows.Conclusion The integration of genomic and epigenomic data by multi-omics approaches in this study provided a better understanding of the molecular mechanisms underlying subclinical mastitis and identified multi-omics candidate discriminant signatures for subclinical mastitis,which may ultimately lead to the development of more effective mastitis control and management strategies.
基金supported in part by the MOST Major Research and Development Project(Grant No.2021YFB2900204)the National Natural Science Foundation of China(NSFC)(Grant No.62201123,No.62132004,No.61971102)+3 种基金China Postdoctoral Science Foundation(Grant No.2022TQ0056)in part by the financial support of the Sichuan Science and Technology Program(Grant No.2022YFH0022)Sichuan Major R&D Project(Grant No.22QYCX0168)the Municipal Government of Quzhou(Grant No.2022D031)。
文摘Integrated data and energy transfer(IDET)enables the electromagnetic waves to transmit wireless energy at the same time of data delivery for lowpower devices.In this paper,an energy harvesting modulation(EHM)assisted multi-user IDET system is studied,where all the received signals at the users are exploited for energy harvesting without the degradation of wireless data transfer(WDT)performance.The joint IDET performance is then analysed theoretically by conceiving a practical time-dependent wireless channel.With the aid of the AO based algorithm,the average effective data rate among users are maximized by ensuring the BER and the wireless energy transfer(WET)performance.Simulation results validate and evaluate the IDET performance of the EHM assisted system,which also demonstrates that the optimal number of user clusters and IDET time slots should be allocated,in order to improve the WET and WDT performance.
基金This work was supported in part by the National Natural Science Foundation of China(51435009)Shanghai Sailing Program(19YF1401500)the Fundamental Research Funds for the Central Universities(2232019D3-34).
文摘Industrial big data integration and sharing(IBDIS)is of great significance in managing and providing data for big data analysis in manufacturing systems.A novel fog-computing-based IBDIS approach called Fog-IBDIS is proposed in order to integrate and share industrial big data with high raw data security and low network traffic loads by moving the integration task from the cloud to the edge of networks.First,a task flow graph(TFG)is designed to model the data analysis process.The TFG is composed of several tasks,which are executed by the data owners through the Fog-IBDIS platform in order to protect raw data privacy.Second,the function of Fog-IBDIS to enable data integration and sharing is presented in five modules:TFG management,compilation and running control,the data integration model,the basic algorithm library,and the management component.Finally,a case study is presented to illustrate the implementation of Fog-IBDIS,which ensures raw data security by deploying the analysis tasks executed by the data generators,and eases the network traffic load by greatly reducing the volume of transmitted data.
基金Supportted by the Natural Science Foundation ofChina (60573091 ,60273018) National Basic Research and Develop-ment Programof China (2003CB317000) the Key Project of Minis-try of Education of China (03044) .
文摘With the rapid development of Web, there are more and more Web databases available for users to access. At the same time, job searchers often have difficulties in first finding the right sources and then querying over them, providing such an integrated job search system over Web databases has become a Web application in high demand. Based on such consideration, we build a deep Web data integration system that supports unified access for users to multiple job Web sites as a job meta-search engine. In this paper, the architecture of the system is given first, and the key components in the system are introduced.
基金supported by the China Meteorological Administration Special Public Welfare Research Fund (GYHY201206012, GYHY201406016)the Climate Change Foundation of the China Meteorological Administration (CCSF201338)
文摘This paper analyzes the status of existing resources through extensive research and international cooperation on the basis of four typical global monthly surface temperature datasets including the climate research dataset of the University of East Anglia(CRUTEM3), the dataset of the U.S. National Climatic Data Center(GHCN-V3), the dataset of the U.S. National Aeronautics and Space Administration(GISSTMP), and the Berkeley Earth surface temperature dataset(Berkeley). China's first global monthly temperature dataset over land was developed by integrating the four aforementioned global temperature datasets and several regional datasets from major countries or regions. This dataset contains information from 9,519 stations worldwide of at least 20 years for monthly mean temperature, 7,073 for maximum temperature, and 6,587 for minimum temperature. Compared with CRUTEM3 and GHCN-V3, the station density is much higher particularly for South America, Africa,and Asia. Moreover, data from significantly more stations were available after the year 1990 which dramatically reduced the uncertainty of the estimated global temperature trend during 1990e2011. The integrated dataset can serve as a reliable data source for global climate change research.
文摘Cloud computing has emerged as a viable alternative to traditional computing infrastructures,offering various benefits.However,the adoption of cloud storage poses significant risks to data secrecy and integrity.This article presents an effective mechanism to preserve the secrecy and integrity of data stored on the public cloud by leveraging blockchain technology,smart contracts,and cryptographic primitives.The proposed approach utilizes a Solidity-based smart contract as an auditor for maintaining and verifying the integrity of outsourced data.To preserve data secrecy,symmetric encryption systems are employed to encrypt user data before outsourcing it.An extensive performance analysis is conducted to illustrate the efficiency of the proposed mechanism.Additionally,a rigorous assessment is conducted to ensure that the developed smart contract is free from vulnerabilities and to measure its associated running costs.The security analysis of the proposed system confirms that our approach can securely maintain the confidentiality and integrity of cloud storage,even in the presence of malicious entities.The proposed mechanism contributes to enhancing data security in cloud computing environments and can be used as a foundation for developing more secure cloud storage systems.
基金This research was supported by the Qinghai Provincial High-End Innovative and Entrepreneurial Talents Project.
文摘Currently,there is a growing trend among users to store their data in the cloud.However,the cloud is vulnerable to persistent data corruption risks arising from equipment failures and hacker attacks.Additionally,when users perform file operations,the semantic integrity of the data can be compromised.Ensuring both data integrity and semantic correctness has become a critical issue that requires attention.We introduce a pioneering solution called Sec-Auditor,the first of its kind with the ability to verify data integrity and semantic correctness simultaneously,while maintaining a constant communication cost independent of the audited data volume.Sec-Auditor also supports public auditing,enabling anyone with access to public information to conduct data audits.This feature makes Sec-Auditor highly adaptable to open data environments,such as the cloud.In Sec-Auditor,users are assigned specific rules that are utilized to verify the accuracy of data semantic.Furthermore,users are given the flexibility to update their own rules as needed.We conduct in-depth analyses of the correctness and security of Sec-Auditor.We also compare several important security attributes with existing schemes,demonstrating the superior properties of Sec-Auditor.Evaluation results demonstrate that even for time-consuming file upload operations,our solution is more efficient than the comparison one.
基金supported by the National Natural Science Foundation of China(31772556)the Local Innovative and Research Teams Project of Guangdong Province(2019BT02N630)+1 种基金the grants from the earmarked fund for China Agriculture Research System(CARS-35)the Science and Technology Innovation Strategy projects of Guangdong Province(Grant No.2018B020203002).
文摘Background:Presently,multi-omics data(e.g.,genomics,transcriptomics,proteomics,and metabolomics)are available to improve genomic predictors.Omics data not only offers new data layers for genomic prediction but also provides a bridge between organismal phenotypes and genome variation that cannot be readily captured at the genome sequence level.Therefore,using multi-omics data to select feature markers is a feasible strategy to improve the accuracy of genomic prediction.In this study,simultaneously using whole-genome sequencing(WGS)and gene expression level data,four strategies for single-nucleotide polymorphism(SNP)preselection were investigated for genomic predictions in the Drosophila Genetic Reference Panel.Results:Using genomic best linear unbiased prediction(GBLUP)with complete WGS data,the prediction accuracies were 0.208±0.020(0.181±0.022)for the startle response and 0.272±0.017(0.307±0.015)for starvation resistance in the female(male)lines.Compared with GBLUP using complete WGS data,both GBLUP and the genomic feature BLUP(GFBLUP)did not improve the prediction accuracy using SNPs preselected from complete WGS data based on the results of genome-wide association studies(GWASs)or transcriptome-wide association studies(TWASs).Furthermore,by using SNPs preselected from the WGS data based on the results of the expression quantitative trait locus(eQTL)mapping of all genes,only the startle response had greater accuracy than GBLUP with the complete WGS data.The best accuracy values in the female and male lines were 0.243±0.020 and 0.220±0.022,respectively.Importantly,by using SNPs preselected based on the results of the eQTL mapping of significant genes from TWAS,both GBLUP and GFBLUP resulted in great accuracy and small bias of genomic prediction.Compared with the GBLUP using complete WGS data,the best accuracy values represented increases of 60.66%and 39.09%for the starvation resistance and 27.40%and 35.36%for startle response in the female and male lines,respectively.Conclusions:Overall,multi-omics data can assist genomic feature preselection and improve the performance of genomic prediction.The new knowledge gained from this study will enrich the use of multi-omics in genomic prediction.
文摘With the deepening informationization of Resources & Environment Remote Sensing geological survey conducted,some potential problems and deficiency are:(1) shortage of unified-planed running environment;(2) inconsistent methods of data integration;and(3) disadvantages of different performing ways of data integration.This paper solves the above problems through overall planning and design,constructs unified running environment, consistent methods of data integration and system structure in order to advance the informationization
文摘Technical and accessibility issues in hospitals often prevent patients from receiving optimal mental and physical health care,which is essential for independent living,especially as societies age and chronic diseases like diabetes and cardiovascular disease become more common.Recent advances in the Internet of Things(IoT)-enabled wearable devices offer potential solutions for remote health monitoring and everyday activity recognition,gaining significant attention in personalized healthcare.This paper comprehensively reviews wearable healthcare technology integrated with the IoT for continuous vital sign monitoring.Relevant papers were extracted and analyzed using a systematic numerical review method,covering various aspects such as sports monitoring,disease detection,patient monitoring,and medical diagnosis.The review highlights the transformative impact of IoTenabled wearable devices in healthcare,facilitating real-time monitoring of vital signs,including blood pressure,temperature,oxygen levels,and heart rate.Results from the reviewed papers demonstrate high accuracy and efficiency in predicting health conditions,improving sports performance,enhancing patient care,and diagnosing diseases.The integration of IoT in wearable healthcare devices enables remote patient monitoring,personalized care,and efficient data transmission,ultimately transcending traditional boundaries of healthcare and leading to better patient outcomes.
文摘Microsoft Excel is essential for the End-User Approach (EUA), offering versatility in data organization, analysis, and visualization, as well as widespread accessibility. It fosters collaboration and informed decision-making across diverse domains. Conversely, Python is indispensable for professional programming due to its versatility, readability, extensive libraries, and robust community support. It enables efficient development, advanced data analysis, data mining, and automation, catering to diverse industries and applications. However, one primary issue when using Microsoft Excel with Python libraries is compatibility and interoperability. While Excel is a widely used tool for data storage and analysis, it may not seamlessly integrate with Python libraries, leading to challenges in reading and writing data, especially in complex or large datasets. Additionally, manipulating Excel files with Python may not always preserve formatting or formulas accurately, potentially affecting data integrity. Moreover, dependency on Excel’s graphical user interface (GUI) for automation can limit scalability and reproducibility compared to Python’s scripting capabilities. This paper covers the integration solution of empowering non-programmers to leverage Python’s capabilities within the familiar Excel environment. This enables users to perform advanced data analysis and automation tasks without requiring extensive programming knowledge. Based on Soliciting feedback from non-programmers who have tested the integration solution, the case study shows how the solution evaluates the ease of implementation, performance, and compatibility of Python with Excel versions.