Construction project is not a standalone engineering maneuver.It is closely linked to the well-being of local communities in concern.The city renovation in Beijing down center for Olympic 2008 transformed many antique...Construction project is not a standalone engineering maneuver.It is closely linked to the well-being of local communities in concern.The city renovation in Beijing down center for Olympic 2008 transformed many antique architecture and regional landscape.It gave a world-recognized achievement in China s modem development and manifested a major milestone in China's economic development.In the course of metro construction projects,there are substantial interwoven municipal structures influencing the success of the projects,which including,but the least,all underground cables and ducts,sewage system,the power consumption of construction works,traffic diversion,air pollution,expatriate business activities and social security.There are many US and UK project insurance companies moving into Asia Pacific.They are doing re-insurance business on major construction guarantee,such as machinery damage,project on-time,power consumption,claims from contractors and communities.Environmental information,such as water quality,indoor and outdoor air quality,people inflow and lift waiting time play deterministic roles in construction's fit-touse.Big Data is a contemporary buzzword since 2013,and the key competence is to provide real time response to heuristic syndrome in order to make short-term prediction.This paper attempts to develop a conceptual model in big data for construction展开更多
The bug tracking system is well known as the project support tool of open source software. There are many categorical data sets recorded on the bug tracking system. In the past, many reliability assessment methods hav...The bug tracking system is well known as the project support tool of open source software. There are many categorical data sets recorded on the bug tracking system. In the past, many reliability assessment methods have been proposed in the research area of software reliability. Also, there are several software project analyses based on the software effort data such as the earned value management. In particular, the software reliability growth models can </span><span style="font-family:Verdana;">apply to the system testing phase of software development. On the other</span><span style="font-family:Verdana;"> hand, the software effort analysis can apply to all development phase, because the fault data is only recorded on the testing phase. We focus on the big fault data and effort data of open source software. Then, it is difficult to assess by using the typical statistical assessment method, because the data recorded on the bug tracking system is large scale. Also, we discuss the jump diffusion process model based on the estimation method of jump parameters by using the discriminant analysis. Moreover, we analyze actual big fault data to show numerical examples of software effort assessment considering many categorical data set.展开更多
The importance of the project selection phase in any six sigma initiative cannot be emphasized enough. The successfulness of the six sigma initiative is affected by successful project selection. Recently, Data Envelop...The importance of the project selection phase in any six sigma initiative cannot be emphasized enough. The successfulness of the six sigma initiative is affected by successful project selection. Recently, Data Envelopment Analysis (DEA) has been proposed as a six sigma project selection tool. However, there exist a number of different DEA formulations which may affect the selection process and the wining project being selected. This work initially applies nine different DEA formulations to several case studies and concludes that different DEA formulations select different wining projects. Also in this work, a Multi-DEA Unified Scoring Framework is proposed to overcome this problem. This framework is applied to several case studies and proved to successfully select the six sigma project with the best performance. The framework is also successful in filtering out some of the projects that have “selective” excellent performance, i.e. projects with excellent performance in some of the DEA formulations and worse performance in others. It is also successful in selecting stable projects;these are projects that perform well in the majority of the DEA formulations, even if it has not been selected as a wining project by any of the DEA formulations.展开更多
Dimensionality reduction and data visualization are useful and important processes in pattern recognition. Many techniques have been developed in the recent years. The self-organizing map (SOM) can be an efficient m...Dimensionality reduction and data visualization are useful and important processes in pattern recognition. Many techniques have been developed in the recent years. The self-organizing map (SOM) can be an efficient method for this purpose. This paper reviews recent advances in this area and related approaches such as multidimensional scaling (MDS), nonlinear PC A, principal manifolds, as well as the connections of the SOM and its recent variant, the visualization induced SOM (ViSOM), with these approaches. The SOM is shown to produce a quantized, qualitative scaling and while the ViSOM a quantitative or metric scaling and approximates principal curve/surface. The SOM can also be regarded as a generalized MDS to relate two metric spaces by forming a topological mapping between them. The relationships among various recently proposed techniques such as ViSOM, Isomap, LLE, and eigenmap are discussed and compared.展开更多
The marking scheme method removes the low scores of the contractor's attributes given by experts when the overall score is calculated, which may result in that a contractor with some latent risks will win the proj...The marking scheme method removes the low scores of the contractor's attributes given by experts when the overall score is calculated, which may result in that a contractor with some latent risks will win the project. In order to remedy the above defect of the marking scheme method, an outlier detection model, which is one mission of knowledge discovery in data, is established on the basis of the sum of similar coefficients. Then, the model is applied to the historical score data of tender evaluation for civil projects in Tianjin, China, according to which the outliers of the scores of the contractor's attributes can be detected and analyzed. Consequently, risk pre-warning can be carried out, and some advice to employers can be given to prevent some latent risks and help them improve the success rate of bidding projects.展开更多
It is common in industrial construction projects for data to be collected and discarded without being analyzed to extract useful knowledge. A proposed integrated methodology based on a five-step Knowledge Discovery in...It is common in industrial construction projects for data to be collected and discarded without being analyzed to extract useful knowledge. A proposed integrated methodology based on a five-step Knowledge Discovery in Data (KDD) model was developed to address this issue. The framework transfers existing multidimensional historical data from completed projects into useful knowledge for future projects. The model starts by understanding the problem domain, industrial construction projects. The second step is analyzing the problem data and its multiple dimensions. The target dataset is the labour resources data generated while managing industrial construction projects. The next step is developing the data collection model and prototype data ware-house. The data warehouse stores collected data in a ready-for-mining format and produces dynamic On Line Analytical Processing (OLAP) reports and graphs. Data was collected from a large western-Canadian structural steel fabricator to prove the applicability of the developed methodology. The proposed framework was applied to three different case studies to validate the applicability of the developed framework to real projects data.展开更多
Spatial data, including geometrical data, attribute data, image data and DEM data, are huge in volume and relations among them are complex. How to effectively organize and manage those data is an important problem in ...Spatial data, including geometrical data, attribute data, image data and DEM data, are huge in volume and relations among them are complex. How to effectively organize and manage those data is an important problem in GIS. Several problems about space data organization and management in GeoStar which is a basic GIS software made in China are discussed in this paper. The paper emphasizes on object model of spatial vector, data organization, data management and how to realize the goal, and the like.展开更多
Risk management is relevant for every project that which seeks to avoid and suppress unanticipated costs, basically calling for pre-emptive action. The current work proposes a new approach for handling risks based on ...Risk management is relevant for every project that which seeks to avoid and suppress unanticipated costs, basically calling for pre-emptive action. The current work proposes a new approach for handling risks based on predictive analytics and machine learning (ML) that can work in real-time to help avoid risks and increase project adaptability. The main research aim of the study is to ascertain risk presence in projects by using historical data from previous projects, focusing on important aspects such as time, task time, resources and project results. t-SNE technique applies feature engineering in the reduction of the dimensionality while preserving important structural properties. This process is analysed using measures including recall, F1-score, accuracy and precision measurements. The results demonstrate that the Gradient Boosting Machine (GBM) achieves an impressive 85% accuracy, 82% precision, 85% recall, and 80% F1-score, surpassing previous models. Additionally, predictive analytics achieves a resource utilisation efficiency of 85%, compared to 70% for traditional allocation methods, and a project cost reduction of 10%, double the 5% achieved by traditional approaches. Furthermore, the study indicates that while GBM excels in overall accuracy, Logistic Regression (LR) offers more favourable precision-recall trade-offs, highlighting the importance of model selection in project risk management.展开更多
文摘Construction project is not a standalone engineering maneuver.It is closely linked to the well-being of local communities in concern.The city renovation in Beijing down center for Olympic 2008 transformed many antique architecture and regional landscape.It gave a world-recognized achievement in China s modem development and manifested a major milestone in China's economic development.In the course of metro construction projects,there are substantial interwoven municipal structures influencing the success of the projects,which including,but the least,all underground cables and ducts,sewage system,the power consumption of construction works,traffic diversion,air pollution,expatriate business activities and social security.There are many US and UK project insurance companies moving into Asia Pacific.They are doing re-insurance business on major construction guarantee,such as machinery damage,project on-time,power consumption,claims from contractors and communities.Environmental information,such as water quality,indoor and outdoor air quality,people inflow and lift waiting time play deterministic roles in construction's fit-touse.Big Data is a contemporary buzzword since 2013,and the key competence is to provide real time response to heuristic syndrome in order to make short-term prediction.This paper attempts to develop a conceptual model in big data for construction
文摘The bug tracking system is well known as the project support tool of open source software. There are many categorical data sets recorded on the bug tracking system. In the past, many reliability assessment methods have been proposed in the research area of software reliability. Also, there are several software project analyses based on the software effort data such as the earned value management. In particular, the software reliability growth models can </span><span style="font-family:Verdana;">apply to the system testing phase of software development. On the other</span><span style="font-family:Verdana;"> hand, the software effort analysis can apply to all development phase, because the fault data is only recorded on the testing phase. We focus on the big fault data and effort data of open source software. Then, it is difficult to assess by using the typical statistical assessment method, because the data recorded on the bug tracking system is large scale. Also, we discuss the jump diffusion process model based on the estimation method of jump parameters by using the discriminant analysis. Moreover, we analyze actual big fault data to show numerical examples of software effort assessment considering many categorical data set.
文摘The importance of the project selection phase in any six sigma initiative cannot be emphasized enough. The successfulness of the six sigma initiative is affected by successful project selection. Recently, Data Envelopment Analysis (DEA) has been proposed as a six sigma project selection tool. However, there exist a number of different DEA formulations which may affect the selection process and the wining project being selected. This work initially applies nine different DEA formulations to several case studies and concludes that different DEA formulations select different wining projects. Also in this work, a Multi-DEA Unified Scoring Framework is proposed to overcome this problem. This framework is applied to several case studies and proved to successfully select the six sigma project with the best performance. The framework is also successful in filtering out some of the projects that have “selective” excellent performance, i.e. projects with excellent performance in some of the DEA formulations and worse performance in others. It is also successful in selecting stable projects;these are projects that perform well in the majority of the DEA formulations, even if it has not been selected as a wining project by any of the DEA formulations.
文摘Dimensionality reduction and data visualization are useful and important processes in pattern recognition. Many techniques have been developed in the recent years. The self-organizing map (SOM) can be an efficient method for this purpose. This paper reviews recent advances in this area and related approaches such as multidimensional scaling (MDS), nonlinear PC A, principal manifolds, as well as the connections of the SOM and its recent variant, the visualization induced SOM (ViSOM), with these approaches. The SOM is shown to produce a quantized, qualitative scaling and while the ViSOM a quantitative or metric scaling and approximates principal curve/surface. The SOM can also be regarded as a generalized MDS to relate two metric spaces by forming a topological mapping between them. The relationships among various recently proposed techniques such as ViSOM, Isomap, LLE, and eigenmap are discussed and compared.
基金Project of Tianjin Water Resources Bureau(No.KY2007-09)
文摘The marking scheme method removes the low scores of the contractor's attributes given by experts when the overall score is calculated, which may result in that a contractor with some latent risks will win the project. In order to remedy the above defect of the marking scheme method, an outlier detection model, which is one mission of knowledge discovery in data, is established on the basis of the sum of similar coefficients. Then, the model is applied to the historical score data of tender evaluation for civil projects in Tianjin, China, according to which the outliers of the scores of the contractor's attributes can be detected and analyzed. Consequently, risk pre-warning can be carried out, and some advice to employers can be given to prevent some latent risks and help them improve the success rate of bidding projects.
文摘It is common in industrial construction projects for data to be collected and discarded without being analyzed to extract useful knowledge. A proposed integrated methodology based on a five-step Knowledge Discovery in Data (KDD) model was developed to address this issue. The framework transfers existing multidimensional historical data from completed projects into useful knowledge for future projects. The model starts by understanding the problem domain, industrial construction projects. The second step is analyzing the problem data and its multiple dimensions. The target dataset is the labour resources data generated while managing industrial construction projects. The next step is developing the data collection model and prototype data ware-house. The data warehouse stores collected data in a ready-for-mining format and produces dynamic On Line Analytical Processing (OLAP) reports and graphs. Data was collected from a large western-Canadian structural steel fabricator to prove the applicability of the developed methodology. The proposed framework was applied to three different case studies to validate the applicability of the developed framework to real projects data.
文摘Spatial data, including geometrical data, attribute data, image data and DEM data, are huge in volume and relations among them are complex. How to effectively organize and manage those data is an important problem in GIS. Several problems about space data organization and management in GeoStar which is a basic GIS software made in China are discussed in this paper. The paper emphasizes on object model of spatial vector, data organization, data management and how to realize the goal, and the like.
文摘Risk management is relevant for every project that which seeks to avoid and suppress unanticipated costs, basically calling for pre-emptive action. The current work proposes a new approach for handling risks based on predictive analytics and machine learning (ML) that can work in real-time to help avoid risks and increase project adaptability. The main research aim of the study is to ascertain risk presence in projects by using historical data from previous projects, focusing on important aspects such as time, task time, resources and project results. t-SNE technique applies feature engineering in the reduction of the dimensionality while preserving important structural properties. This process is analysed using measures including recall, F1-score, accuracy and precision measurements. The results demonstrate that the Gradient Boosting Machine (GBM) achieves an impressive 85% accuracy, 82% precision, 85% recall, and 80% F1-score, surpassing previous models. Additionally, predictive analytics achieves a resource utilisation efficiency of 85%, compared to 70% for traditional allocation methods, and a project cost reduction of 10%, double the 5% achieved by traditional approaches. Furthermore, the study indicates that while GBM excels in overall accuracy, Logistic Regression (LR) offers more favourable precision-recall trade-offs, highlighting the importance of model selection in project risk management.