Small angle x-ray scattering(SAXS)is an advanced technique for characterizing the particle size distribution(PSD)of nanoparticles.However,the ill-posed nature of inverse problems in SAXS data analysis often reduces th...Small angle x-ray scattering(SAXS)is an advanced technique for characterizing the particle size distribution(PSD)of nanoparticles.However,the ill-posed nature of inverse problems in SAXS data analysis often reduces the accuracy of conventional methods.This article proposes a user-friendly software for PSD analysis,GranuSAS,which employs an algorithm that integrates truncated singular value decomposition(TSVD)with the Chahine method.This approach employs TSVD for data preprocessing,generating a set of initial solutions with noise suppression.A high-quality initial solution is subsequently selected via the L-curve method.This selected candidate solution is then iteratively refined by the Chahine algorithm,enforcing constraints such as non-negativity and improving physical interpretability.Most importantly,GranuSAS employs a parallel architecture that simultaneously yields inversion results from multiple shape models and,by evaluating the accuracy of each model's reconstructed scattering curve,offers a suggestion for model selection in material systems.To systematically validate the accuracy and efficiency of the software,verification was performed using both simulated and experimental datasets.The results demonstrate that the proposed software delivers both satisfactory accuracy and reliable computational efficiency.It provides an easy-to-use and reliable tool for researchers in materials science,helping them fully exploit the potential of SAXS in nanoparticle characterization.展开更多
Aiming at the problems such as low throughput and unbalanced load of data center network caused by traditional multipath routing strategy,a dynamic load balancing strategy for flow classification oriented to Fat-Tree ...Aiming at the problems such as low throughput and unbalanced load of data center network caused by traditional multipath routing strategy,a dynamic load balancing strategy for flow classification oriented to Fat-Tree topology based on the software defined network(SDN)architecture is proposed,named DLB-FC.Multi-index evaluation methods such as link state information and network traffic characteristics are considered.DLB-FC mechanism can dynamically adjust the flow classification threshold to differentiate between large and small flows.The scheme selects different forwarding paths to meet the transmission performance requirements of different flow characteristics.On this basis,an SDN simulation platform is built for performance testing.The simulation results show that DLB-FC algorithm can dynamically distinguish large flows from small flows and achieve load balancing effectively.Compared with equal-cost multi-path(ECMP),global first fit(GFF)and minmum total delay load routing(MTDLR)algorithms,DLB-FC scheme improves the network throughput and link utilization of the data center network effectively.The transmission delay is also reduced with better load balance.展开更多
This paper proposes a multivariate data fusion based quality evaluation model for software talent cultivation.The model constructs a comprehensive ability and quality evaluation index system for college students from ...This paper proposes a multivariate data fusion based quality evaluation model for software talent cultivation.The model constructs a comprehensive ability and quality evaluation index system for college students from a perspective of engineering course,especially of software engineering.As for evaluation method,relying on the behavioral data of students during their school years,we aim to construct the evaluation model as objective as possible,effectively weakening the negative impact of personal subjective assumptions on the evaluation results.展开更多
E-learning produces the data on the learners’utilization of the software,which helps the teacher to perceive the learners’mental status and learning efficiency,so it is of great value to make full use of the data.Wi...E-learning produces the data on the learners’utilization of the software,which helps the teacher to perceive the learners’mental status and learning efficiency,so it is of great value to make full use of the data.With Speexx foreign language learning system being the case,this thesis introduces the function of such data and the modes of how to use them to facilitate the blendedteaching and learning.展开更多
To comprehensively understand the Arctic and Antarctic upper atmosphere, it is often crucial to analyze various data that are obtained from many regions. Infrastructure that promotes such interdisciplinary studies on ...To comprehensively understand the Arctic and Antarctic upper atmosphere, it is often crucial to analyze various data that are obtained from many regions. Infrastructure that promotes such interdisciplinary studies on the upper atmosphere has been developed by a Japanese inter-university project called the Inter-university Upper atmosphere Global Observation Network (1UGONET). The objective of this paper is to describe the infrastructure and tools developed by IUGONET. We focus on the data analysis software. It is written in Interactive Data Language (IDL) and is a plug-in for the THEMIS Data Analysis Software suite (TDAS), which is a set of IDL libraries used to visualize and analyze satellite- and ground-based data. We present plots of upper atmospheric data provided by IUGONET as examples of applications, and verify the usefulness of the software in the study of polar science. We discuss IUGONET's new and unique developments, i.e., an executable file of TDAS that can run on the IDL Virtual Machine, IDL routines to retrieve metadata from the IUGONET database, and an archive of 3-D simulation data that uses the Common Data Format so that it can easily be used with TDAS.展开更多
Masked data are the system failure data when exact component causing system failure might be unknown.In this paper,the mathematical description of general masked data was presented in software reliability engineering....Masked data are the system failure data when exact component causing system failure might be unknown.In this paper,the mathematical description of general masked data was presented in software reliability engineering.Furthermore,a general maskedbased additive non-homogeneous Poisson process(NHPP) model was considered to analyze component reliability.However,the problem of masked-based additive model lies in the difficulty of estimating parameters.The maximum likelihood estimation procedure was derived to estimate parameters.Finally,a numerical example was given to illustrate the applicability of proposed model,and the immune particle swarm optimization(IPSO) algorithm was used in maximize log-likelihood function.展开更多
When castings become complicated and the demands for precision of numerical simulation become higher,the numerical data of casting numerical simulation become more massive.On a general personal computer,these massive ...When castings become complicated and the demands for precision of numerical simulation become higher,the numerical data of casting numerical simulation become more massive.On a general personal computer,these massive numerical data may probably exceed the capacity of available memory,resulting in failure of rendering.Based on the out-of-core technique,this paper proposes a method to effectively utilize external storage and reduce memory usage dramatically,so as to solve the problem of insufficient memory for massive data rendering on general personal computers.Based on this method,a new postprocessor is developed.It is capable to illustrate filling and solidification processes of casting,as well as thermal stess.The new post-processor also provides fast interaction to simulation results.Theoretical analysis as well as several practical examples prove that the memory usage and loading time of the post-processor are independent of the size of the relevant files,but the proportion of the number of cells on surface.Meanwhile,the speed of rendering and fetching of value from the mouse is appreciable,and the demands of real-time and interaction are satisfied.展开更多
Software defect prediction is a critical component in maintaining software quality,enabling early identification and resolution of issues that could lead to system failures and significant financial losses.With the in...Software defect prediction is a critical component in maintaining software quality,enabling early identification and resolution of issues that could lead to system failures and significant financial losses.With the increasing reliance on user-generated content,social media reviews have emerged as a valuable source of real-time feedback,offering insights into potential software defects that traditional testing methods may overlook.However,existing models face challenges like handling imbalanced data,high computational complexity,and insufficient inte-gration of contextual information from these reviews.To overcome these limitations,this paper introduces the SESDP(Sentiment Analysis-Based Early Software Defect Prediction)model.SESDP employs a Transformer-Based Multi-Task Learning approach using Robustly Optimized Bidirectional Encoder Representations from Transformers Approach(RoBERTa)to simultaneously perform sentiment analysis and defect prediction.By integrating text embedding extraction,sentiment score computation,and feature fusion,the model effectively captures both the contextual nuances and sentiment expressed in user reviews.Experimental results show that SESDP achieves superior performance with an accuracy of 96.37%,precision of 94.7%,and recall of 95.4%,particularly excelling in handling imbalanced datasets compared to baseline models.This approach offers a scalable and efficient solution for early software defect detection,enhancing proactive software quality assurance.展开更多
The magnetic resonance spectroscopy(MRS)results are greatly influenced by reconstruction of the spectrum and quantitative analysis.Because of this requirement a number of programs dedicated to MRS data analysis were d...The magnetic resonance spectroscopy(MRS)results are greatly influenced by reconstruction of the spectrum and quantitative analysis.Because of this requirement a number of programs dedicated to MRS data analysis were developed.The selection and use of appropriate software is crucial not only in clinical procedures,but also while carrying out scientific research.The choice of the software to suit the user's needs should be based on the analysis of the functionality of the program.It is particularly important from the user's viewpoint to identify what data can be loaded and processed in the program.The specific programs allow the user different degree of control over analysis parameters.Moreover,the programs for MRS data analysis differ in terms of the applied signal processing algorithms.The aim of this work,therefore,is to review available packages designed for MRS data analysis,taking into account their capabilities and limitations.展开更多
Software defect prediction plays a critical role in software development and quality assurance processes. Effective defect prediction enables testers to accurately prioritize testing efforts and enhance defect detecti...Software defect prediction plays a critical role in software development and quality assurance processes. Effective defect prediction enables testers to accurately prioritize testing efforts and enhance defect detection efficiency. Additionally, this technology provides developers with a means to quickly identify errors, thereby improving software robustness and overall quality. However, current research in software defect prediction often faces challenges, such as relying on a single data source or failing to adequately account for the characteristics of multiple coexisting data sources. This approach may overlook the differences and potential value of various data sources, affecting the accuracy and generalization performance of prediction results. To address this issue, this study proposes a multivariate heterogeneous hybrid deep learning algorithm for defect prediction (DP-MHHDL). Initially, Abstract Syntax Tree (AST), Code Dependency Network (CDN), and code static quality metrics are extracted from source code files and used as inputs to ensure data diversity. Subsequently, for the three types of heterogeneous data, the study employs a graph convolutional network optimization model based on adjacency and spatial topologies, a Convolutional Neural Network-Bidirectional Long Short-Term Memory (CNN-BiLSTM) hybrid neural network model, and a TabNet model to extract data features. These features are then concatenated and processed through a fully connected neural network for defect prediction. Finally, the proposed framework is evaluated using ten promise defect repository projects, and performance is assessed with three metrics: F1, Area under the curve (AUC), and Matthews correlation coefficient (MCC). The experimental results demonstrate that the proposed algorithm outperforms existing methods, offering a novel solution for software defect prediction.展开更多
In light of the escalating demand and intricacy of services in contemporary terrestrial,maritime,and aerial combat operations,there is a compelling need for enhanced service quality and efficiency in airborne cluster ...In light of the escalating demand and intricacy of services in contemporary terrestrial,maritime,and aerial combat operations,there is a compelling need for enhanced service quality and efficiency in airborne cluster communication networks.Software-Defined Networking(SDN)proffers a viable solution for the multifaceted task of cooperative communication transmission and management across different operational domains within complex combat contexts,due to its intrinsic ability to flexibly allocate and centrally administer network resources.This study pivots around the optimization of SDN controller deployment within airborne data link clusters.A collaborative multi-controller architecture predicated on airborne data link clusters is thus proposed.Within this architectural framework,the controller deployment issue is reframed as a two-fold problem:subdomain partition-ing and central interaction node selection.We advocate a subdomain segmentation approach grounded in node value ranking(NDVR)and a central interaction node selection methodology predicated on an enhanced Artificial Fish Swarm Algorithm(AFSA).The advanced NDVR-AFSA(Node value ranking-Improved artificial fish swarm algorithm)algorithm makes use of a chaos algorithm for population initialization,boosting population diversity and circumventing premature algorithm convergence.By the integration of adaptive strategies and incorporation of the genetic algorithm’s crossover and mutation operations,the algorithm’s search range adaptability is enhanced,thereby increasing the possibility of obtaining globally optimal solutions,while concurrently augmenting cluster reliability.The simulation results verify the advantages of the NDVR-IAFSA algorithm,achieve a better load balancing effect,improve the reliability of aviation data link cluster,and significantly reduce the average propagation delay and disconnection rate,respectively,by 12.8%and 11.7%.This shows that the optimization scheme has important significance in practical application,and can meet the high requirements of modern sea,land,and air operations to aviation airborne communication networks.展开更多
Increasing global competition forces manufacturers of products from alltechnical fields to guarantee a high product quality for a long period of time. At thesame time it is necessary to minimize production costs. In o...Increasing global competition forces manufacturers of products from alltechnical fields to guarantee a high product quality for a long period of time. At thesame time it is necessary to minimize production costs. In order to meet all theserequirements, on-line data acquisition and processing are of increasing importancein distributed automation systems. A software bus operating on industrial Ethernethas an ability to minimize operating costs by offering easy installation, scalability,high degree of reliability and remote monitoring and control.展开更多
The </span></span><span><span><span style="font-family:"">software reliability model is the stochastic model to measure the software <span>reliability quantitatively....The </span></span><span><span><span style="font-family:"">software reliability model is the stochastic model to measure the software <span>reliability quantitatively. A Hazard-Rate Model is </span></span></span></span><span><span><span style="font-family:"">the </span></span></span><span><span><span style="font-family:"">well</span></span></span><span><span><span style="font-family:"">-</span></span></span><span><span><span style="font-family:"">known one as the</span></span></span><span><span><span style="font-family:""> typical software reliability model. We propose Hazard-Rate Models Consider<span>ing Fault Severity Levels (CFSL) for Open Source Software (OSS). The purpose of </span><span>this research is to </span></span></span></span><span><span><span style="font-family:"">make </span></span></span><span><span><span style="font-family:"">the Hazard-Rate Model considering CFSL adapt to</span></span></span><span><span><span style="font-family:""> </span></span></span><span><span><span style="font-family:"">baseline hazard function and 2 kinds of faults data in Bug Tracking System <span>(BTS)</span></span></span></span><span><span><span style="font-family:"">,</span></span></span><span><span><span style="font-family:""> <i>i.e.</i>, we use the covariate vectors in Cox proportional Hazard-Rate</span></span></span><span><span><span style="font-family:""> Model. Also, <span>we show the numerical examples by evaluating the performance of our pro</span><span>posed model. As the result, we compare the performance of our model with the</span> Hazard-Rate Model CFSL.展开更多
This paper presents a methodology driven by database constraints for designing and developing(database)software applications.Much needed and with excellent results,this paradigm guarantees the highest possible quality...This paper presents a methodology driven by database constraints for designing and developing(database)software applications.Much needed and with excellent results,this paradigm guarantees the highest possible quality of the managed data.The proposed methodology is illustrated with an easy to understand,yet complex medium-sized genealogy software application driven by more than 200 database constraints,which fully meets such expectations.展开更多
Large latency of applications will bring revenue loss to cloud infrastructure providers in the cloud data center. The existing controllers of software-defined networking architecture can fetch and process traffic info...Large latency of applications will bring revenue loss to cloud infrastructure providers in the cloud data center. The existing controllers of software-defined networking architecture can fetch and process traffic information in the network. Therefore, the controllers can only optimize the network latency of applications. However, the serving latency of applications is also an important factor in delivered user-experience for arrival requests. Unintelligent request routing will cause large serving latency if arrival requests are allocated to overloaded virtual machines. To deal with the request routing problem, this paper proposes the workload-aware software-defined networking controller architecture. Then, request routing algorithms are proposed to minimize the total round trip time for every type of request by considering the congestion in the network and the workload in virtual machines(VMs). This paper finally provides the evaluation of the proposed algorithms in a simulated prototype. The simulation results show that the proposed methodology is efficient compared with the existing approaches.展开更多
A software package to be used in high-speed oscilloscope-basedthree-dimensionalbunch-by-bunch charge and position measurement is presented.The software package takes the pick-up electrode signal waveform recorded by t...A software package to be used in high-speed oscilloscope-basedthree-dimensionalbunch-by-bunch charge and position measurement is presented.The software package takes the pick-up electrode signal waveform recorded by the high-speed oscilloscope as input,and it calculates and outputs the bunch-by-bunch charge and position.In addition to enabling a three-dimensional observation of the motion of each passing bunch on all beam position monitor pick-up electrodes,it offers many additional features such as injection analysis,bunch response function reconstruction,and turn-by-turn beam analysis.The software package has an easy-to-understand graphical user interface and convenient interactive operation,which has been verified on the Windows 10 system.展开更多
In order to effectively detect and analyze the backdoors this paper introduces a method named Backdoor Analysis based on Sensitive flow tracking and Concolic Execution(BASEC).BASEC uses sensitive flow tracking to ef...In order to effectively detect and analyze the backdoors this paper introduces a method named Backdoor Analysis based on Sensitive flow tracking and Concolic Execution(BASEC).BASEC uses sensitive flow tracking to effectively discover backdoor behaviors, such as stealing secret information and injecting evil data into system, with less false negatives. With concolic execution on predetermined path, the backdoor trigger condition can be extracted and analyzed to achieve high accuracy. BASEC has been implemented and experimented on several software backdoor samples widespread on the Internet, and over 90% of them can be detected. Compared with behavior-based and system-call-based detection methods, BASEC relies less on the historical sample collections, and is more effective in detecting software backdoors, especially those injected into software by modifying and recompiling source codes.展开更多
Driven by market requirements, software services organizations have adopted various software engineering process models (such as capability maturity model (CMM), capability maturity model integration (CMMI), ISO ...Driven by market requirements, software services organizations have adopted various software engineering process models (such as capability maturity model (CMM), capability maturity model integration (CMMI), ISO 9001:2000, etc.) and practice of the project management concepts defined in the project management body of knowledge. While this has definitely helped organizations to bring some methods into the software development madness, there always exists a demand for comparing various groups within the organization in terms of the practice of these defined process models. Even though there exist many metrics for comparison, considering the variety of projects in terms of technology, life cycle, etc., finding a single metric that caters to this is a difficult task. This paper proposes a model for arriving at a rating on group maturity within the organization. Considering the linguistic or imprecise and uncertain nature of software measurements, fuzzy logic approach is used for the proposed model. Without the barriers like technology or life cycle difference, the proposed model helps the organization to compare different groups within it with reasonable precision.展开更多
Entity recognition and extraction are the foundations of knowledge graph construction.Entity data in the field of software engineering come from different platforms and communities,and have different formats.This pape...Entity recognition and extraction are the foundations of knowledge graph construction.Entity data in the field of software engineering come from different platforms and communities,and have different formats.This paper divides multi-source software knowledge entities into unstructured data,semi-structured data and code data.For these different types of data,Bi-directional Long Short-Term Memory(Bi-LSTM)with Conditional Random Field(CRF),template matching,and abstract syntax tree are used and integrated into a multi-source software knowledge entity extraction integration model(MEIM)to extract software entities.The model can be updated continuously based on user’s feedbacks to improve the accuracy.To deal with the shortage of entity annotation datasets,keyword extraction methods based on Term Frequency–Inverse Document Frequency(TF-IDF),TextRank,and K-Means are applied to annotate tasks.The proposed MEIM model is applied to the Spring Boot framework,which demonstrates good adaptability.The extracted entities are used to construct a knowledge graph,which is applied to association retrieval and association visualization.展开更多
基金Project supported by the Project of the Anhui Provincial Natural Science Foundation(Grant No.2308085MA19)Strategic Priority Research Program of the Chinese Academy of Sciences(Grant No.XDA0410401)+2 种基金the National Natural Science Foundation of China(Grant No.52202120)the National Key Research and Development Program of China(Grant No.2023YFA1609800)USTC Research Funds of the Double First-Class Initiative(Grant No.YD2310002013)。
文摘Small angle x-ray scattering(SAXS)is an advanced technique for characterizing the particle size distribution(PSD)of nanoparticles.However,the ill-posed nature of inverse problems in SAXS data analysis often reduces the accuracy of conventional methods.This article proposes a user-friendly software for PSD analysis,GranuSAS,which employs an algorithm that integrates truncated singular value decomposition(TSVD)with the Chahine method.This approach employs TSVD for data preprocessing,generating a set of initial solutions with noise suppression.A high-quality initial solution is subsequently selected via the L-curve method.This selected candidate solution is then iteratively refined by the Chahine algorithm,enforcing constraints such as non-negativity and improving physical interpretability.Most importantly,GranuSAS employs a parallel architecture that simultaneously yields inversion results from multiple shape models and,by evaluating the accuracy of each model's reconstructed scattering curve,offers a suggestion for model selection in material systems.To systematically validate the accuracy and efficiency of the software,verification was performed using both simulated and experimental datasets.The results demonstrate that the proposed software delivers both satisfactory accuracy and reliable computational efficiency.It provides an easy-to-use and reliable tool for researchers in materials science,helping them fully exploit the potential of SAXS in nanoparticle characterization.
基金Supported by the National Natural Science Foundation of China(No.61672270)Jiangsu Provionce Teaching Reform Project for Cloud Computing Technology and Application Talent Training(No.201802130049).
文摘Aiming at the problems such as low throughput and unbalanced load of data center network caused by traditional multipath routing strategy,a dynamic load balancing strategy for flow classification oriented to Fat-Tree topology based on the software defined network(SDN)architecture is proposed,named DLB-FC.Multi-index evaluation methods such as link state information and network traffic characteristics are considered.DLB-FC mechanism can dynamically adjust the flow classification threshold to differentiate between large and small flows.The scheme selects different forwarding paths to meet the transmission performance requirements of different flow characteristics.On this basis,an SDN simulation platform is built for performance testing.The simulation results show that DLB-FC algorithm can dynamically distinguish large flows from small flows and achieve load balancing effectively.Compared with equal-cost multi-path(ECMP),global first fit(GFF)and minmum total delay load routing(MTDLR)algorithms,DLB-FC scheme improves the network throughput and link utilization of the data center network effectively.The transmission delay is also reduced with better load balance.
基金supported in part by the Education Reform Key Projects of Heilongjiang Province(Grant No.SJGZ20220011,SJGZ20220012)the Excellent Project of Ministry of Education and China Higher Education Association on Digital Ideological and Political Education in Universities(Grant No.GXSZSZJPXM001)。
文摘This paper proposes a multivariate data fusion based quality evaluation model for software talent cultivation.The model constructs a comprehensive ability and quality evaluation index system for college students from a perspective of engineering course,especially of software engineering.As for evaluation method,relying on the behavioral data of students during their school years,we aim to construct the evaluation model as objective as possible,effectively weakening the negative impact of personal subjective assumptions on the evaluation results.
文摘E-learning produces the data on the learners’utilization of the software,which helps the teacher to perceive the learners’mental status and learning efficiency,so it is of great value to make full use of the data.With Speexx foreign language learning system being the case,this thesis introduces the function of such data and the modes of how to use them to facilitate the blendedteaching and learning.
基金supported by the Special Edu-cational Research Budget(Research Promotion)[FY2009]the Special Budget(Project)[FY2010 and later years]from the Ministry of Education,Culture,Sports,Science and Technology(MEXT),Japansupported by the GRENE Arctic Climate Change Research Project,Japan
文摘To comprehensively understand the Arctic and Antarctic upper atmosphere, it is often crucial to analyze various data that are obtained from many regions. Infrastructure that promotes such interdisciplinary studies on the upper atmosphere has been developed by a Japanese inter-university project called the Inter-university Upper atmosphere Global Observation Network (1UGONET). The objective of this paper is to describe the infrastructure and tools developed by IUGONET. We focus on the data analysis software. It is written in Interactive Data Language (IDL) and is a plug-in for the THEMIS Data Analysis Software suite (TDAS), which is a set of IDL libraries used to visualize and analyze satellite- and ground-based data. We present plots of upper atmospheric data provided by IUGONET as examples of applications, and verify the usefulness of the software in the study of polar science. We discuss IUGONET's new and unique developments, i.e., an executable file of TDAS that can run on the IDL Virtual Machine, IDL routines to retrieve metadata from the IUGONET database, and an archive of 3-D simulation data that uses the Common Data Format so that it can easily be used with TDAS.
基金Technology Foundation of Guizhou Province,China(No.QianKeHeJZi[2015]2064)Scientific Research Foundation for Advanced Talents in Guizhou Institue of Technology and Science,China(No.XJGC20150106)Joint Foundation of Guizhou Province,China(No.QianKeHeLHZi[2015]7105)
文摘Masked data are the system failure data when exact component causing system failure might be unknown.In this paper,the mathematical description of general masked data was presented in software reliability engineering.Furthermore,a general maskedbased additive non-homogeneous Poisson process(NHPP) model was considered to analyze component reliability.However,the problem of masked-based additive model lies in the difficulty of estimating parameters.The maximum likelihood estimation procedure was derived to estimate parameters.Finally,a numerical example was given to illustrate the applicability of proposed model,and the immune particle swarm optimization(IPSO) algorithm was used in maximize log-likelihood function.
基金supported by the New Century Excellent Talents in University(NCET-09-0396)the National Science&Technology Key Projects of Numerical Control(2012ZX04014-031)+1 种基金the Natural Science Foundation of Hubei Province(2011CDB279)the Foundation for Innovative Research Groups of the Natural Science Foundation of Hubei Province,China(2010CDA067)
文摘When castings become complicated and the demands for precision of numerical simulation become higher,the numerical data of casting numerical simulation become more massive.On a general personal computer,these massive numerical data may probably exceed the capacity of available memory,resulting in failure of rendering.Based on the out-of-core technique,this paper proposes a method to effectively utilize external storage and reduce memory usage dramatically,so as to solve the problem of insufficient memory for massive data rendering on general personal computers.Based on this method,a new postprocessor is developed.It is capable to illustrate filling and solidification processes of casting,as well as thermal stess.The new post-processor also provides fast interaction to simulation results.Theoretical analysis as well as several practical examples prove that the memory usage and loading time of the post-processor are independent of the size of the relevant files,but the proportion of the number of cells on surface.Meanwhile,the speed of rendering and fetching of value from the mouse is appreciable,and the demands of real-time and interaction are satisfied.
基金funded by a grant from the Center of Excellence in Information Assurance(CoEIA),King Saud University(KSU).
文摘Software defect prediction is a critical component in maintaining software quality,enabling early identification and resolution of issues that could lead to system failures and significant financial losses.With the increasing reliance on user-generated content,social media reviews have emerged as a valuable source of real-time feedback,offering insights into potential software defects that traditional testing methods may overlook.However,existing models face challenges like handling imbalanced data,high computational complexity,and insufficient inte-gration of contextual information from these reviews.To overcome these limitations,this paper introduces the SESDP(Sentiment Analysis-Based Early Software Defect Prediction)model.SESDP employs a Transformer-Based Multi-Task Learning approach using Robustly Optimized Bidirectional Encoder Representations from Transformers Approach(RoBERTa)to simultaneously perform sentiment analysis and defect prediction.By integrating text embedding extraction,sentiment score computation,and feature fusion,the model effectively captures both the contextual nuances and sentiment expressed in user reviews.Experimental results show that SESDP achieves superior performance with an accuracy of 96.37%,precision of 94.7%,and recall of 95.4%,particularly excelling in handling imbalanced datasets compared to baseline models.This approach offers a scalable and efficient solution for early software defect detection,enhancing proactive software quality assurance.
文摘The magnetic resonance spectroscopy(MRS)results are greatly influenced by reconstruction of the spectrum and quantitative analysis.Because of this requirement a number of programs dedicated to MRS data analysis were developed.The selection and use of appropriate software is crucial not only in clinical procedures,but also while carrying out scientific research.The choice of the software to suit the user's needs should be based on the analysis of the functionality of the program.It is particularly important from the user's viewpoint to identify what data can be loaded and processed in the program.The specific programs allow the user different degree of control over analysis parameters.Moreover,the programs for MRS data analysis differ in terms of the applied signal processing algorithms.The aim of this work,therefore,is to review available packages designed for MRS data analysis,taking into account their capabilities and limitations.
文摘Software defect prediction plays a critical role in software development and quality assurance processes. Effective defect prediction enables testers to accurately prioritize testing efforts and enhance defect detection efficiency. Additionally, this technology provides developers with a means to quickly identify errors, thereby improving software robustness and overall quality. However, current research in software defect prediction often faces challenges, such as relying on a single data source or failing to adequately account for the characteristics of multiple coexisting data sources. This approach may overlook the differences and potential value of various data sources, affecting the accuracy and generalization performance of prediction results. To address this issue, this study proposes a multivariate heterogeneous hybrid deep learning algorithm for defect prediction (DP-MHHDL). Initially, Abstract Syntax Tree (AST), Code Dependency Network (CDN), and code static quality metrics are extracted from source code files and used as inputs to ensure data diversity. Subsequently, for the three types of heterogeneous data, the study employs a graph convolutional network optimization model based on adjacency and spatial topologies, a Convolutional Neural Network-Bidirectional Long Short-Term Memory (CNN-BiLSTM) hybrid neural network model, and a TabNet model to extract data features. These features are then concatenated and processed through a fully connected neural network for defect prediction. Finally, the proposed framework is evaluated using ten promise defect repository projects, and performance is assessed with three metrics: F1, Area under the curve (AUC), and Matthews correlation coefficient (MCC). The experimental results demonstrate that the proposed algorithm outperforms existing methods, offering a novel solution for software defect prediction.
基金supported by the following funds:Defense Industrial Technology Development Program Grant:G20210513Shaanxi Provincal Department of Science and Technology Grant:2021KW-07Shaanxi Provincal Department of Science and Technology Grant:2022 QFY01-14.
文摘In light of the escalating demand and intricacy of services in contemporary terrestrial,maritime,and aerial combat operations,there is a compelling need for enhanced service quality and efficiency in airborne cluster communication networks.Software-Defined Networking(SDN)proffers a viable solution for the multifaceted task of cooperative communication transmission and management across different operational domains within complex combat contexts,due to its intrinsic ability to flexibly allocate and centrally administer network resources.This study pivots around the optimization of SDN controller deployment within airborne data link clusters.A collaborative multi-controller architecture predicated on airborne data link clusters is thus proposed.Within this architectural framework,the controller deployment issue is reframed as a two-fold problem:subdomain partition-ing and central interaction node selection.We advocate a subdomain segmentation approach grounded in node value ranking(NDVR)and a central interaction node selection methodology predicated on an enhanced Artificial Fish Swarm Algorithm(AFSA).The advanced NDVR-AFSA(Node value ranking-Improved artificial fish swarm algorithm)algorithm makes use of a chaos algorithm for population initialization,boosting population diversity and circumventing premature algorithm convergence.By the integration of adaptive strategies and incorporation of the genetic algorithm’s crossover and mutation operations,the algorithm’s search range adaptability is enhanced,thereby increasing the possibility of obtaining globally optimal solutions,while concurrently augmenting cluster reliability.The simulation results verify the advantages of the NDVR-IAFSA algorithm,achieve a better load balancing effect,improve the reliability of aviation data link cluster,and significantly reduce the average propagation delay and disconnection rate,respectively,by 12.8%and 11.7%.This shows that the optimization scheme has important significance in practical application,and can meet the high requirements of modern sea,land,and air operations to aviation airborne communication networks.
文摘Increasing global competition forces manufacturers of products from alltechnical fields to guarantee a high product quality for a long period of time. At thesame time it is necessary to minimize production costs. In order to meet all theserequirements, on-line data acquisition and processing are of increasing importancein distributed automation systems. A software bus operating on industrial Ethernethas an ability to minimize operating costs by offering easy installation, scalability,high degree of reliability and remote monitoring and control.
文摘The </span></span><span><span><span style="font-family:"">software reliability model is the stochastic model to measure the software <span>reliability quantitatively. A Hazard-Rate Model is </span></span></span></span><span><span><span style="font-family:"">the </span></span></span><span><span><span style="font-family:"">well</span></span></span><span><span><span style="font-family:"">-</span></span></span><span><span><span style="font-family:"">known one as the</span></span></span><span><span><span style="font-family:""> typical software reliability model. We propose Hazard-Rate Models Consider<span>ing Fault Severity Levels (CFSL) for Open Source Software (OSS). The purpose of </span><span>this research is to </span></span></span></span><span><span><span style="font-family:"">make </span></span></span><span><span><span style="font-family:"">the Hazard-Rate Model considering CFSL adapt to</span></span></span><span><span><span style="font-family:""> </span></span></span><span><span><span style="font-family:"">baseline hazard function and 2 kinds of faults data in Bug Tracking System <span>(BTS)</span></span></span></span><span><span><span style="font-family:"">,</span></span></span><span><span><span style="font-family:""> <i>i.e.</i>, we use the covariate vectors in Cox proportional Hazard-Rate</span></span></span><span><span><span style="font-family:""> Model. Also, <span>we show the numerical examples by evaluating the performance of our pro</span><span>posed model. As the result, we compare the performance of our model with the</span> Hazard-Rate Model CFSL.
文摘This paper presents a methodology driven by database constraints for designing and developing(database)software applications.Much needed and with excellent results,this paradigm guarantees the highest possible quality of the managed data.The proposed methodology is illustrated with an easy to understand,yet complex medium-sized genealogy software application driven by more than 200 database constraints,which fully meets such expectations.
基金supported by the National Postdoctoral Science Foundation of China(2014M550068)
文摘Large latency of applications will bring revenue loss to cloud infrastructure providers in the cloud data center. The existing controllers of software-defined networking architecture can fetch and process traffic information in the network. Therefore, the controllers can only optimize the network latency of applications. However, the serving latency of applications is also an important factor in delivered user-experience for arrival requests. Unintelligent request routing will cause large serving latency if arrival requests are allocated to overloaded virtual machines. To deal with the request routing problem, this paper proposes the workload-aware software-defined networking controller architecture. Then, request routing algorithms are proposed to minimize the total round trip time for every type of request by considering the congestion in the network and the workload in virtual machines(VMs). This paper finally provides the evaluation of the proposed algorithms in a simulated prototype. The simulation results show that the proposed methodology is efficient compared with the existing approaches.
基金supported by the Ten Thousand Talent Program and National Natural Science Foundation of China(No.11575282)the Ten Thousand Talent Program and Chinese Academy of Sciences Key Technology Talent Program。
文摘A software package to be used in high-speed oscilloscope-basedthree-dimensionalbunch-by-bunch charge and position measurement is presented.The software package takes the pick-up electrode signal waveform recorded by the high-speed oscilloscope as input,and it calculates and outputs the bunch-by-bunch charge and position.In addition to enabling a three-dimensional observation of the motion of each passing bunch on all beam position monitor pick-up electrodes,it offers many additional features such as injection analysis,bunch response function reconstruction,and turn-by-turn beam analysis.The software package has an easy-to-understand graphical user interface and convenient interactive operation,which has been verified on the Windows 10 system.
基金Supported in part by the National Natural Science Foundation of China(61272493)the Specialized Research Fund for the Doctoral Program of Higher Education of China(20113402120026)Oversea Academic Training Funds of University of Science and Technology of China
文摘In order to effectively detect and analyze the backdoors this paper introduces a method named Backdoor Analysis based on Sensitive flow tracking and Concolic Execution(BASEC).BASEC uses sensitive flow tracking to effectively discover backdoor behaviors, such as stealing secret information and injecting evil data into system, with less false negatives. With concolic execution on predetermined path, the backdoor trigger condition can be extracted and analyzed to achieve high accuracy. BASEC has been implemented and experimented on several software backdoor samples widespread on the Internet, and over 90% of them can be detected. Compared with behavior-based and system-call-based detection methods, BASEC relies less on the historical sample collections, and is more effective in detecting software backdoors, especially those injected into software by modifying and recompiling source codes.
文摘Driven by market requirements, software services organizations have adopted various software engineering process models (such as capability maturity model (CMM), capability maturity model integration (CMMI), ISO 9001:2000, etc.) and practice of the project management concepts defined in the project management body of knowledge. While this has definitely helped organizations to bring some methods into the software development madness, there always exists a demand for comparing various groups within the organization in terms of the practice of these defined process models. Even though there exist many metrics for comparison, considering the variety of projects in terms of technology, life cycle, etc., finding a single metric that caters to this is a difficult task. This paper proposes a model for arriving at a rating on group maturity within the organization. Considering the linguistic or imprecise and uncertain nature of software measurements, fuzzy logic approach is used for the proposed model. Without the barriers like technology or life cycle difference, the proposed model helps the organization to compare different groups within it with reasonable precision.
基金Zhifang Liao:Ministry of Science and Technology:Key Research and Development Project(2018YFB003800),Hunan Provincial Key Laboratory of Finance&Economics Big Data Scienceand Technology(Hunan University of Finance and Economics)2017TP1025,HNNSF 2018JJ2535Shengzong Liu:NSF61802120.
文摘Entity recognition and extraction are the foundations of knowledge graph construction.Entity data in the field of software engineering come from different platforms and communities,and have different formats.This paper divides multi-source software knowledge entities into unstructured data,semi-structured data and code data.For these different types of data,Bi-directional Long Short-Term Memory(Bi-LSTM)with Conditional Random Field(CRF),template matching,and abstract syntax tree are used and integrated into a multi-source software knowledge entity extraction integration model(MEIM)to extract software entities.The model can be updated continuously based on user’s feedbacks to improve the accuracy.To deal with the shortage of entity annotation datasets,keyword extraction methods based on Term Frequency–Inverse Document Frequency(TF-IDF),TextRank,and K-Means are applied to annotate tasks.The proposed MEIM model is applied to the Spring Boot framework,which demonstrates good adaptability.The extracted entities are used to construct a knowledge graph,which is applied to association retrieval and association visualization.