The paper presents a set of techniques of digital watermarking by which copyright and user rights messages are hidden into geo-spatial graphics data,as well as techniques of compressing and encrypting the watermarked ...The paper presents a set of techniques of digital watermarking by which copyright and user rights messages are hidden into geo-spatial graphics data,as well as techniques of compressing and encrypting the watermarked geo-spatial graphics data.The technology aims at tracing and resisting the illegal distribution and duplication of the geo-spatial graphics data product,so as to effectively protect the data producer's rights as well as to facilitate the secure sharing of geo-spatial graphics data.So far in the CIS field throughout the world,few researches have been made on digital watermarking.The research is a novel exploration both in the field of security management of geo-spatial graphics data and in the applications of digital watermarking technique.An application software employing the proposed technology has been developed.A number of experimental tests on the 1:500,000 digital bathymetric chart of the South China Sea and 1:10,000 digital topographic map of Jiangsu Province have been conducted to verify the feasibility of the proposed technology.展开更多
With the rise of data-intensive research,data literacy has become a critical capability for improving scientific data quality and achieving artificial intelligence(AI)readiness.In the biomedical domain,data are charac...With the rise of data-intensive research,data literacy has become a critical capability for improving scientific data quality and achieving artificial intelligence(AI)readiness.In the biomedical domain,data are characterized by high complexity and privacy sensitivity,calling for robust and systematic data management skills.This paper reviews current trends in scientific data governance and the evolving policy landscape,highlighting persistent challenges such as inconsistent standards,semantic misalignment,and limited awareness of compliance.These issues are largely rooted in the lack of structured training and practical support for researchers.In response,this study builds on existing data literacy frameworks and integrates the specific demands of biomedical research to propose a comprehensive,lifecycle-oriented data literacy competency model with an emphasis on ethics and regulatory awareness.Furthermore,it outlines a tiered training strategy tailored to different research stages—undergraduate,graduate,and professional,offering theoretical foundations and practical pathways for universities and research institutions to advance data literacy education.展开更多
With the explosive growth of data available, there is an urgent need to develop continuous data mining which reduces manual interaction evidently. A novel model for data mining is proposed in evolving environment. Fir...With the explosive growth of data available, there is an urgent need to develop continuous data mining which reduces manual interaction evidently. A novel model for data mining is proposed in evolving environment. First, some valid mining task schedules are generated, and then au tonomous and local mining are executed periodically, finally, previous results are merged and refined. The framework based on the model creates a communication mechanism to in corporate domain knowledge into continuous process through ontology service. The local and merge mining are transparent to the end user and heterogeneous data ,source by ontology. Experiments suggest that the framework should be useful in guiding the continuous mining process.展开更多
Reliability and remaining useful life(RUL)estimation for a satellite rechargeable lithium battery(RLB)are significant for prognostic and health management(PHM).A novel Bayesian framework is proposed to do reliability ...Reliability and remaining useful life(RUL)estimation for a satellite rechargeable lithium battery(RLB)are significant for prognostic and health management(PHM).A novel Bayesian framework is proposed to do reliability analysis by synthesizing multisource data,including bivariate degradation data and lifetime data.Bivariate degradation means that there are two degraded performance characteristics leading to the failure of the system.First,linear Wiener process and Frank Copula function are used to model the dependent degradation processes of the RLB's temperature and discharge voltage.Next,the Bayesian method,in combination with Markov Chain Monte Carlo(MCMC)simulations,is provided to integrate limited bivariate degradation data with other congeneric RLBs'lifetime data.Then reliability evaluation and RUL prediction are carried out for PHM.A simulation study demonstrates that due to the data fusion,parameter estimations and predicted RUL obtained from our model are more precise than models only using degradation data or ignoring the dependency of different degradation processes.Finally,a practical case study of a satellite RLB verifies the usability of the model.展开更多
This paper investigates the problem of data scarcity in spectrum prediction.A cognitive radio equipment may frequently switch the target frequency as the electromagnetic environment changes.The previously trained mode...This paper investigates the problem of data scarcity in spectrum prediction.A cognitive radio equipment may frequently switch the target frequency as the electromagnetic environment changes.The previously trained model for prediction often cannot maintain a good performance when facing small amount of historical data of the new target frequency.Moreover,the cognitive radio equipment usually implements the dynamic spectrum access in real time which means the time to recollect the data of the new task frequency band and retrain the model is very limited.To address the above issues,we develop a crossband data augmentation framework for spectrum prediction by leveraging the recent advances of generative adversarial network(GAN)and deep transfer learning.Firstly,through the similarity measurement,we pre-train a GAN model using the historical data of the frequency band that is the most similar to the target frequency band.Then,through the data augmentation by feeding the small amount of the target data into the pre-trained GAN,temporal-spectral residual network is further trained using deep transfer learning and the generated data with high similarity from GAN.Finally,experiment results demonstrate the effectiveness of the proposed framework.展开更多
The tremendous growth of the cloud computing environments requires new architecture for security services. Cloud computing is the utilization of many servers/data centers or cloud data storages (CDSs) housed in many d...The tremendous growth of the cloud computing environments requires new architecture for security services. Cloud computing is the utilization of many servers/data centers or cloud data storages (CDSs) housed in many different locations and interconnected by high speed networks. CDS, like any other emerging technology, is experiencing growing pains. It is immature, it is fragmented and it lacks standardization. Although security issues are delaying its fast adoption, cloud computing is an unstoppable force and we need to provide security mechanisms to ensure its secure adoption. In this paper a comprehensive security framework based on Multi-Agent System (MAS) architecture for CDS to facilitate confidentiality, correctness assurance, availability and integrity of users' data in the cloud is proposed. Our security framework consists of two main layers as agent layer and CDS layer. Our propose MAS architecture includes main five types of agents: Cloud Service Provider Agent (CSPA), Cloud Data Confidentiality Agent (CDConA), Cloud Data Correctness Agent (CDCorA), Cloud Data Availability Agent (CDAA) and Cloud Data Integrity Agent (CDIA). In order to verify our proposed security framework based on MAS architecture, pilot study is conducted using a questionnaire survey. Rasch Methodology is used to analyze the pilot data. Item reliability is found to be poor and a few respondents and items are identified as misfits with distorted measurements. As a result, some problematic questions are revised and some predictably easy questions are excluded from the questionnaire. A prototype of the system is implemented using Java. To simulate the agents, oracle database packages and triggers are used to implement agent functions and oracle jobs are utilized to create agents.展开更多
Cloud accounting is based on the traditional financial work process, the context of big data, and the necessary trend of future corporate accounting development. Its emergence and rapid development will have a fundame...Cloud accounting is based on the traditional financial work process, the context of big data, and the necessary trend of future corporate accounting development. Its emergence and rapid development will have a fundamental impact on corporate environmental information disclosure. In the big data era of information sharing, companies will have a new understanding of the emergence, balance, and final consideration of social responsibility, and will have new changes in their overall decision-making and information disclosure methods. "Knowing" and "behavior" will be combined on the basis of rational judgment, so that corporate environmental information disclosure is more in line with the overall social development requirements. Based on the background of big data, this article starts with the disclosure of impact factors, footholds, and path choices. It describes the evolution of corporate environmental information disclosure and provides reference suggestions for enterprises to disclose environmental information truthfully and perform social responsibilities.展开更多
In several fields like financial dealing,industry,business,medicine,et cetera,Big Data(BD)has been utilized extensively,which is nothing but a collection of a huge amount of data.However,it is highly complicated along...In several fields like financial dealing,industry,business,medicine,et cetera,Big Data(BD)has been utilized extensively,which is nothing but a collection of a huge amount of data.However,it is highly complicated along with time-consuming to process a massive amount of data.Thus,to design the Distribution Preserving Framework for BD,a novel methodology has been proposed utilizing Manhattan Distance(MD)-centered Partition Around Medoid(MD–PAM)along with Conjugate Gradient Artificial Neural Network(CG-ANN),which undergoes various steps to reduce the complications of BD.Firstly,the data are processed in the pre-processing phase by mitigating the data repetition utilizing the map-reduce function;subsequently,the missing data are handled by substituting or by ignoring the missed values.After that,the data are transmuted into a normalized form.Next,to enhance the classification performance,the data’s dimensionalities are minimized by employing Gaussian Kernel(GK)-Fisher Discriminant Analysis(GK-FDA).Afterwards,the processed data is submitted to the partitioning phase after transmuting it into a structured format.In the partition phase,by utilizing the MD-PAM,the data are partitioned along with grouped into a cluster.Lastly,by employing CG-ANN,the data are classified in the classification phase so that the needed data can be effortlessly retrieved by the user.To analogize the outcomes of the CG-ANN with the prevailing methodologies,the NSL-KDD openly accessible datasets are utilized.The experiential outcomes displayed that an efficient result along with a reduced computation cost was shown by the proposed CG-ANN.The proposed work outperforms well in terms of accuracy,sensitivity and specificity than the existing systems.展开更多
An idea is presented about the development of a data processing and analysis system for ICF experiments, which is based on an object oriented framework. The design and preliminary implementation of the data processing...An idea is presented about the development of a data processing and analysis system for ICF experiments, which is based on an object oriented framework. The design and preliminary implementation of the data processing and analysis framework based on the ROOT system have been completed. Software for unfolding soft X-ray spectra has been developed to test the functions of this framework.展开更多
The main aim of an educational institute is to offer high-quality education to students. The system to achieve better quality in the educational system is to find the knowledge from educational data and to discover th...The main aim of an educational institute is to offer high-quality education to students. The system to achieve better quality in the educational system is to find the knowledge from educational data and to discover the attributes that manipulate the performance of students. Student performance prediction is a major issue in education and training, specifically in the educational data mining system. This research presents the student performance prediction approach with the MapReduce framework based on the proposed fractional competitive multi-verse optimization-based deep neuro-fuzzy network. The proposed fractional competitive multi-verse optimization-based deep neuro-fuzzy network is derived by integrating fractional calculus with competitive multi-verse optimization. The MapReduce framework is designed with the mapper and the reducer phase to perform the student performance prediction mechanism with the deep learning classifier. The input data is partitioned at the mapper phase to perform the data transformation process, and thereby the features are selected using the distance measure. The selected unique features are employed for the data segmentation process, and thereafter the prediction strategy is accomplished at the reducer phase by the deep neuro-fuzzy network classifier. The proposed method obtained the performance in terms of mean square error, root mean square error and mean absolute error with the values of 0.338 3, 0.581 7, and 0.391 5, respectively.展开更多
Antarctic data management is the research focus, which the international Antarctic organizations, e.g. Antarctic Treaty Consultative Meeting(ATCM) , Scientific Committee on Antarctic Research(SCAR), and Council of Man...Antarctic data management is the research focus, which the international Antarctic organizations, e.g. Antarctic Treaty Consultative Meeting(ATCM) , Scientific Committee on Antarctic Research(SCAR), and Council of Managers of National Antarctic Programmes(COMNAP) have been paying close attention to and promoting actively. Through the co effort of international Antarctic organizations and member countries concerned in recent years, Antarctic Data Directory Syatem(ADDS) is established as the most important basic programme for development of the international Antarctic data management system. At present, Joint Committee on Antarctic Data Management(JCADM) is responsible for organizing and coordinating the international Antarctic data management, and implementing the project ADDS.In this paper the background on Antarctic data management in time sequence and the structure of international framework are introduced, meanwhile, it is necessary to develop ADDS first of all. The ADDS mainly consists of the two principal parts: National Antarctic Data Center(NADCs) of all the party members and Antarctic Main Directory(AMD), the best available technology for creating ADDS is to make full use of International Directory Network(IDN) and adopt its Directory Interchange Formats(DIF). In the light of the above requirements, combined with Chinese specific situation, the contents, technical and administrative methods on Chinese Antarctic data management are discussed to promote our related work.展开更多
Crop phenomics has rapidly progressed in recent years due to the growing need for crop functional geno-mics,digital breeding,and smart cultivation.Despite this advancement,the lack of standards for the cre-ation and u...Crop phenomics has rapidly progressed in recent years due to the growing need for crop functional geno-mics,digital breeding,and smart cultivation.Despite this advancement,the lack of standards for the cre-ation and usage of crop phenomics technology and equipment has become a bottleneck,limiting the industry’s high-quality development.This paper begins with an overview of the crop phenotyping indus-try and presents an industrial mapping of technology and equipment for big data in crop phenomics.It analyzes the necessity and current state of constructing a standard framework for crop phenotyping.Furthermore,this paper proposes the intended organizational structure and goals of the standard frame-work.It details the essentials of the standard framework in the research and development of hardware and equipment,data acquisition,and the storage and management of crop phenotyping data.Finally,it discusses promoting the construction and evaluation of the standard framework,aiming to provide ideas for developing a high-quality standard framework for crop phenotyping.展开更多
With an increase in population and economic development,water withdrawals are close to or even exceed the amount of water available in many regions of the world.Modelling water withdrawals could help water planners im...With an increase in population and economic development,water withdrawals are close to or even exceed the amount of water available in many regions of the world.Modelling water withdrawals could help water planners improve the efficiency of water use,water resources allocation,and management in order to alleviate water crises.However,minimal information has been obtained on how water withdrawals have changed over space and time,especially on a regional or local scale.This research proposes a data-driven framework to help estimate county-level distribution of water withdrawals.Using this framework,spatial statistical methods are used to estimate water withdrawals for agricultural,industrial,and domestic purposes in the Huaihe River watershed in China for the period 1978–2018.Total water withdrawals were found to have more than doubled,from 292.55×10^(8)m^(3) in 1978 to 642.93×10^(8)m^(3) in 2009,and decreased to 602.63×10^(8)m^(3) in 2018.Agricultural water increased from 208.17×10^(8)m^(3) in 1978 to 435.80×10^(8)m^(3) in 2009 and decreased to 360.84×10^(8)m^(3) in 2018.Industrial and domestic water usage constantly increased throughout the 1978–2018 period.In 1978,industrial and domestic demands were 20.35×10^(8)m^(3) and 60.04×10^(8)m^(3),respectively,and up until 2018,the figures were 105.58×10^(8)m^(3) and 136.20×10^(8)m^(3).From a spatial distribution perspective,Moran’s I statistical results show that the total water withdrawal has significant spatial autocorrelation during 1978–2018.The overall trend was a gradual increase in 1978–2010 with withdrawal beginning to decline in 2010–2018.The results of Getis-Ord G_(i)^(*)statistical calculations showed spatially contiguous clusters of total water withdrawal in the Huaihe River watershed during1978–2010,and the spatial agglomeration weakened from 2010 to 2018.This study provides a data-driven framework for assessing water withdrawals to enable a deeper understanding of competing water use among economic sectors as well as water withdrawal modelled with proper data resource and method.展开更多
The paper discusses the need of a high-level query language to allow analysts,geographers and,in general,non-programmers to easily cross-analyze multi-source VGI created by means of apps,crowd-sourced data from social...The paper discusses the need of a high-level query language to allow analysts,geographers and,in general,non-programmers to easily cross-analyze multi-source VGI created by means of apps,crowd-sourced data from social networks and authoritative geo-referenced data,usually represented as JSON data sets(nowadays,the de facto standard for data exported by social networks).Since an easy to use high-level language for querying and manipulating collections of possibly geo-tagged JSON objects is still unavailable,we propose a truly declarative language,named J-CO-QL,that is based on a well-defined execution model.A plug-in for a GIS permits to visualize geo-tagged data sets stored in a NoSQL database such as MongoDB;furthermore,the same plug-in can be used to write and execute J-CO-QL queries on those databases.The paper introduces the language by exemplifying its operators within a real study case,the aim of which is to understand the mobility of people in the neighborhood of Bergamo city.Cross-analysis of data about transportation networks and VGI from travelers is performed,by means of J-CO-QL language,capable to manipulate and transform,combine and join possibly geo-tagged JSON objects,in order to produce new possibly geo-tagged JSON objects satisfying users’needs.展开更多
This paper proposes a two-step general framework for reversible data hiding(RDH)schemes with controllable contrast enhancement.The first step aims at preserving visual perception as much as possible on the basis of ac...This paper proposes a two-step general framework for reversible data hiding(RDH)schemes with controllable contrast enhancement.The first step aims at preserving visual perception as much as possible on the basis of achieving high embedding capacity(EC),while the second step is used for increasing image contrast.In the second step,some peak-pairs are utilized so that the histogram of pixel values is modified to perform histogram equalization(HE),which would lead to the image contrast enhancement.However,for HE,the utilization of some peak-pairs easily leads to over-enhanced image contrast when a large number of bits are embedded.Therefore,in our proposed framework,contrast over-enhancement is avoided by controlling the degree of contrast enhancement.Since the second step can only provide a small amount of data due to controlled contrast enhancement,the first one helps to achieve a large amount of data without degrading visual quality.Any RDH method which can achieve high EC while preserve good visual quality,can be selected for the first step.In fact,Gao et al.’s method is a special case of our proposed framework.In addition,two simple and commonly-used RDH methods are also introduced to further demonstrate the generalization of our framework.展开更多
Point of Care (PoC) devices and systems can be categorized into three broad classes (CAT 1, CAT 2, and CAT 3) based on the context of operation and usage. In this paper, the categories are defined to address certain u...Point of Care (PoC) devices and systems can be categorized into three broad classes (CAT 1, CAT 2, and CAT 3) based on the context of operation and usage. In this paper, the categories are defined to address certain usage models of the PoC device. PoC devices that are used for PoC testing and diagnostic applications are defined CAT 1 devices;PoC devices that are used for patient monitoring are defined as CAT 2 devices (PoCM);PoC devices that are used for as interfacing with other devices are defined as CAT 3 devices (PoCI). The PoCI devices provide an interface gateway for collecting and aggregating data from other medical devices. In all categories, data security is an important aspect. This paper presents a security framework concept, which is applicable for all of the classes of PoC operation. It outlines the concepts and security framework for preventing security challenges in unauthorized access to data, unintended data flow, and data tampering during communication between system entities, the user, and the PoC system. The security framework includes secure layering of basic PoC system architecture, protection of PoC devices in the context of application and network. Developing the security framework is taken into account of a thread model of the PoC system. A proposal for a low-level protocol is discussed. This protocol is independent of communications technologies, and it is elaborated in relation to providing security. An algorithm that can be used to overcome the threat challenges has been shown using the elements in the protocol. The paper further discusses the vulnerability scanning process for the PoC system interconnected network. The paper also presents a four-step process of authentication and authorization framework for providing the security for the PoC system. Finally, the paper concludes with the machine to machine (M2M) security viewpoint and discusses the key stakeholders within an actual deployment of the PoC system and its security challenges.展开更多
The importance of the project selection phase in any six sigma initiative cannot be emphasized enough. The successfulness of the six sigma initiative is affected by successful project selection. Recently, Data Envelop...The importance of the project selection phase in any six sigma initiative cannot be emphasized enough. The successfulness of the six sigma initiative is affected by successful project selection. Recently, Data Envelopment Analysis (DEA) has been proposed as a six sigma project selection tool. However, there exist a number of different DEA formulations which may affect the selection process and the wining project being selected. This work initially applies nine different DEA formulations to several case studies and concludes that different DEA formulations select different wining projects. Also in this work, a Multi-DEA Unified Scoring Framework is proposed to overcome this problem. This framework is applied to several case studies and proved to successfully select the six sigma project with the best performance. The framework is also successful in filtering out some of the projects that have “selective” excellent performance, i.e. projects with excellent performance in some of the DEA formulations and worse performance in others. It is also successful in selecting stable projects;these are projects that perform well in the majority of the DEA formulations, even if it has not been selected as a wining project by any of the DEA formulations.展开更多
Digital educational content is gaining importance as an incubator of pedagogical methodologies in formal and informal online educational settings. Its educational efficiency is directly dependent on its quality, howev...Digital educational content is gaining importance as an incubator of pedagogical methodologies in formal and informal online educational settings. Its educational efficiency is directly dependent on its quality, however educational content is more than information and data. This paper presents a new data quality framework for assessing digital educational content used for teaching in distance learning environments. The model relies on the ISO2500 series quality standard and beside providing the mechanisms for multi-facet quality assessment it also supports organizations that design, create, manage and use educational content with the quality tools (expressed as quality metrics and measurement methods) to provide a more efficient distance education experience. The model describes the quality characteristics of the educational material content using data and software quality characteristics.展开更多
基金Under the auspices of Jiangsu Provincial Science and Technology Fundation of Surveying and Mapping (No. 200416 )
文摘The paper presents a set of techniques of digital watermarking by which copyright and user rights messages are hidden into geo-spatial graphics data,as well as techniques of compressing and encrypting the watermarked geo-spatial graphics data.The technology aims at tracing and resisting the illegal distribution and duplication of the geo-spatial graphics data product,so as to effectively protect the data producer's rights as well as to facilitate the secure sharing of geo-spatial graphics data.So far in the CIS field throughout the world,few researches have been made on digital watermarking.The research is a novel exploration both in the field of security management of geo-spatial graphics data and in the applications of digital watermarking technique.An application software employing the proposed technology has been developed.A number of experimental tests on the 1:500,000 digital bathymetric chart of the South China Sea and 1:10,000 digital topographic map of Jiangsu Province have been conducted to verify the feasibility of the proposed technology.
文摘With the rise of data-intensive research,data literacy has become a critical capability for improving scientific data quality and achieving artificial intelligence(AI)readiness.In the biomedical domain,data are characterized by high complexity and privacy sensitivity,calling for robust and systematic data management skills.This paper reviews current trends in scientific data governance and the evolving policy landscape,highlighting persistent challenges such as inconsistent standards,semantic misalignment,and limited awareness of compliance.These issues are largely rooted in the lack of structured training and practical support for researchers.In response,this study builds on existing data literacy frameworks and integrates the specific demands of biomedical research to propose a comprehensive,lifecycle-oriented data literacy competency model with an emphasis on ethics and regulatory awareness.Furthermore,it outlines a tiered training strategy tailored to different research stages—undergraduate,graduate,and professional,offering theoretical foundations and practical pathways for universities and research institutions to advance data literacy education.
基金Supported by the National Natural Science Foun-dation of China (60173058 ,70372024)
文摘With the explosive growth of data available, there is an urgent need to develop continuous data mining which reduces manual interaction evidently. A novel model for data mining is proposed in evolving environment. First, some valid mining task schedules are generated, and then au tonomous and local mining are executed periodically, finally, previous results are merged and refined. The framework based on the model creates a communication mechanism to in corporate domain knowledge into continuous process through ontology service. The local and merge mining are transparent to the end user and heterogeneous data ,source by ontology. Experiments suggest that the framework should be useful in guiding the continuous mining process.
基金Project(71371182) supported by the National Natural Science Foundation of China
文摘Reliability and remaining useful life(RUL)estimation for a satellite rechargeable lithium battery(RLB)are significant for prognostic and health management(PHM).A novel Bayesian framework is proposed to do reliability analysis by synthesizing multisource data,including bivariate degradation data and lifetime data.Bivariate degradation means that there are two degraded performance characteristics leading to the failure of the system.First,linear Wiener process and Frank Copula function are used to model the dependent degradation processes of the RLB's temperature and discharge voltage.Next,the Bayesian method,in combination with Markov Chain Monte Carlo(MCMC)simulations,is provided to integrate limited bivariate degradation data with other congeneric RLBs'lifetime data.Then reliability evaluation and RUL prediction are carried out for PHM.A simulation study demonstrates that due to the data fusion,parameter estimations and predicted RUL obtained from our model are more precise than models only using degradation data or ignoring the dependency of different degradation processes.Finally,a practical case study of a satellite RLB verifies the usability of the model.
基金This work was supported by the Science and Technology Innovation 2030-Key Project of“New Generation Artificial Intelligence”of China under Grant 2018AAA0102303the Natural Science Foundation for Distinguished Young Scholars of Jiangsu Province(No.BK20190030)the National Natural Science Foundation of China(No.61631020,No.61871398,No.61931011 and No.U20B2038).
文摘This paper investigates the problem of data scarcity in spectrum prediction.A cognitive radio equipment may frequently switch the target frequency as the electromagnetic environment changes.The previously trained model for prediction often cannot maintain a good performance when facing small amount of historical data of the new target frequency.Moreover,the cognitive radio equipment usually implements the dynamic spectrum access in real time which means the time to recollect the data of the new task frequency band and retrain the model is very limited.To address the above issues,we develop a crossband data augmentation framework for spectrum prediction by leveraging the recent advances of generative adversarial network(GAN)and deep transfer learning.Firstly,through the similarity measurement,we pre-train a GAN model using the historical data of the frequency band that is the most similar to the target frequency band.Then,through the data augmentation by feeding the small amount of the target data into the pre-trained GAN,temporal-spectral residual network is further trained using deep transfer learning and the generated data with high similarity from GAN.Finally,experiment results demonstrate the effectiveness of the proposed framework.
文摘The tremendous growth of the cloud computing environments requires new architecture for security services. Cloud computing is the utilization of many servers/data centers or cloud data storages (CDSs) housed in many different locations and interconnected by high speed networks. CDS, like any other emerging technology, is experiencing growing pains. It is immature, it is fragmented and it lacks standardization. Although security issues are delaying its fast adoption, cloud computing is an unstoppable force and we need to provide security mechanisms to ensure its secure adoption. In this paper a comprehensive security framework based on Multi-Agent System (MAS) architecture for CDS to facilitate confidentiality, correctness assurance, availability and integrity of users' data in the cloud is proposed. Our security framework consists of two main layers as agent layer and CDS layer. Our propose MAS architecture includes main five types of agents: Cloud Service Provider Agent (CSPA), Cloud Data Confidentiality Agent (CDConA), Cloud Data Correctness Agent (CDCorA), Cloud Data Availability Agent (CDAA) and Cloud Data Integrity Agent (CDIA). In order to verify our proposed security framework based on MAS architecture, pilot study is conducted using a questionnaire survey. Rasch Methodology is used to analyze the pilot data. Item reliability is found to be poor and a few respondents and items are identified as misfits with distorted measurements. As a result, some problematic questions are revised and some predictably easy questions are excluded from the questionnaire. A prototype of the system is implemented using Java. To simulate the agents, oracle database packages and triggers are used to implement agent functions and oracle jobs are utilized to create agents.
基金supported by the Philosophy and Social Science Research Project of Daqing City (Grant No. DSGB2017112)the Postgraduate Innovation Research Project of Heilongjiang Bayi Agricultural University (Grant No. YJSCX2017Y79)
文摘Cloud accounting is based on the traditional financial work process, the context of big data, and the necessary trend of future corporate accounting development. Its emergence and rapid development will have a fundamental impact on corporate environmental information disclosure. In the big data era of information sharing, companies will have a new understanding of the emergence, balance, and final consideration of social responsibility, and will have new changes in their overall decision-making and information disclosure methods. "Knowing" and "behavior" will be combined on the basis of rational judgment, so that corporate environmental information disclosure is more in line with the overall social development requirements. Based on the background of big data, this article starts with the disclosure of impact factors, footholds, and path choices. It describes the evolution of corporate environmental information disclosure and provides reference suggestions for enterprises to disclose environmental information truthfully and perform social responsibilities.
文摘In several fields like financial dealing,industry,business,medicine,et cetera,Big Data(BD)has been utilized extensively,which is nothing but a collection of a huge amount of data.However,it is highly complicated along with time-consuming to process a massive amount of data.Thus,to design the Distribution Preserving Framework for BD,a novel methodology has been proposed utilizing Manhattan Distance(MD)-centered Partition Around Medoid(MD–PAM)along with Conjugate Gradient Artificial Neural Network(CG-ANN),which undergoes various steps to reduce the complications of BD.Firstly,the data are processed in the pre-processing phase by mitigating the data repetition utilizing the map-reduce function;subsequently,the missing data are handled by substituting or by ignoring the missed values.After that,the data are transmuted into a normalized form.Next,to enhance the classification performance,the data’s dimensionalities are minimized by employing Gaussian Kernel(GK)-Fisher Discriminant Analysis(GK-FDA).Afterwards,the processed data is submitted to the partitioning phase after transmuting it into a structured format.In the partition phase,by utilizing the MD-PAM,the data are partitioned along with grouped into a cluster.Lastly,by employing CG-ANN,the data are classified in the classification phase so that the needed data can be effortlessly retrieved by the user.To analogize the outcomes of the CG-ANN with the prevailing methodologies,the NSL-KDD openly accessible datasets are utilized.The experiential outcomes displayed that an efficient result along with a reduced computation cost was shown by the proposed CG-ANN.The proposed work outperforms well in terms of accuracy,sensitivity and specificity than the existing systems.
基金the National Natural Science Foundation of China(No.61303094)the Program of Science and Technology Commission of Shanghai Municipality(No.16511102400)the Innovation Program of Shanghai Municipal Education Commission(No.14YZ024)
基金This project supported by the National High-Tech Research and Development Plan (863-804-3)
文摘An idea is presented about the development of a data processing and analysis system for ICF experiments, which is based on an object oriented framework. The design and preliminary implementation of the data processing and analysis framework based on the ROOT system have been completed. Software for unfolding soft X-ray spectra has been developed to test the functions of this framework.
文摘The main aim of an educational institute is to offer high-quality education to students. The system to achieve better quality in the educational system is to find the knowledge from educational data and to discover the attributes that manipulate the performance of students. Student performance prediction is a major issue in education and training, specifically in the educational data mining system. This research presents the student performance prediction approach with the MapReduce framework based on the proposed fractional competitive multi-verse optimization-based deep neuro-fuzzy network. The proposed fractional competitive multi-verse optimization-based deep neuro-fuzzy network is derived by integrating fractional calculus with competitive multi-verse optimization. The MapReduce framework is designed with the mapper and the reducer phase to perform the student performance prediction mechanism with the deep learning classifier. The input data is partitioned at the mapper phase to perform the data transformation process, and thereby the features are selected using the distance measure. The selected unique features are employed for the data segmentation process, and thereafter the prediction strategy is accomplished at the reducer phase by the deep neuro-fuzzy network classifier. The proposed method obtained the performance in terms of mean square error, root mean square error and mean absolute error with the values of 0.338 3, 0.581 7, and 0.391 5, respectively.
文摘Antarctic data management is the research focus, which the international Antarctic organizations, e.g. Antarctic Treaty Consultative Meeting(ATCM) , Scientific Committee on Antarctic Research(SCAR), and Council of Managers of National Antarctic Programmes(COMNAP) have been paying close attention to and promoting actively. Through the co effort of international Antarctic organizations and member countries concerned in recent years, Antarctic Data Directory Syatem(ADDS) is established as the most important basic programme for development of the international Antarctic data management system. At present, Joint Committee on Antarctic Data Management(JCADM) is responsible for organizing and coordinating the international Antarctic data management, and implementing the project ADDS.In this paper the background on Antarctic data management in time sequence and the structure of international framework are introduced, meanwhile, it is necessary to develop ADDS first of all. The ADDS mainly consists of the two principal parts: National Antarctic Data Center(NADCs) of all the party members and Antarctic Main Directory(AMD), the best available technology for creating ADDS is to make full use of International Directory Network(IDN) and adopt its Directory Interchange Formats(DIF). In the light of the above requirements, combined with Chinese specific situation, the contents, technical and administrative methods on Chinese Antarctic data management are discussed to promote our related work.
基金supported by the National Key R&D Program of China(2022YFD2002300)the Construction of Collaborative Innovation Center of Beijing Academy of Agricultural and Forestry Sciences(KJCX20240406)+1 种基金the National Natural Science Foundation of China(32071891)the earmarked fund(CARS-02 and CARS-054).
文摘Crop phenomics has rapidly progressed in recent years due to the growing need for crop functional geno-mics,digital breeding,and smart cultivation.Despite this advancement,the lack of standards for the cre-ation and usage of crop phenomics technology and equipment has become a bottleneck,limiting the industry’s high-quality development.This paper begins with an overview of the crop phenotyping indus-try and presents an industrial mapping of technology and equipment for big data in crop phenomics.It analyzes the necessity and current state of constructing a standard framework for crop phenotyping.Furthermore,this paper proposes the intended organizational structure and goals of the standard frame-work.It details the essentials of the standard framework in the research and development of hardware and equipment,data acquisition,and the storage and management of crop phenotyping data.Finally,it discusses promoting the construction and evaluation of the standard framework,aiming to provide ideas for developing a high-quality standard framework for crop phenotyping.
基金Under the auspices of the National Natural Science Foundation of China(No.71203200)the National Social Science Fund Project(No.20&ZD138)+1 种基金the National Science and Technology Platform Construction Project(No.2005DKA32300)the Major Research Projects of the Ministry of Education(No.16JJD770019)。
文摘With an increase in population and economic development,water withdrawals are close to or even exceed the amount of water available in many regions of the world.Modelling water withdrawals could help water planners improve the efficiency of water use,water resources allocation,and management in order to alleviate water crises.However,minimal information has been obtained on how water withdrawals have changed over space and time,especially on a regional or local scale.This research proposes a data-driven framework to help estimate county-level distribution of water withdrawals.Using this framework,spatial statistical methods are used to estimate water withdrawals for agricultural,industrial,and domestic purposes in the Huaihe River watershed in China for the period 1978–2018.Total water withdrawals were found to have more than doubled,from 292.55×10^(8)m^(3) in 1978 to 642.93×10^(8)m^(3) in 2009,and decreased to 602.63×10^(8)m^(3) in 2018.Agricultural water increased from 208.17×10^(8)m^(3) in 1978 to 435.80×10^(8)m^(3) in 2009 and decreased to 360.84×10^(8)m^(3) in 2018.Industrial and domestic water usage constantly increased throughout the 1978–2018 period.In 1978,industrial and domestic demands were 20.35×10^(8)m^(3) and 60.04×10^(8)m^(3),respectively,and up until 2018,the figures were 105.58×10^(8)m^(3) and 136.20×10^(8)m^(3).From a spatial distribution perspective,Moran’s I statistical results show that the total water withdrawal has significant spatial autocorrelation during 1978–2018.The overall trend was a gradual increase in 1978–2010 with withdrawal beginning to decline in 2010–2018.The results of Getis-Ord G_(i)^(*)statistical calculations showed spatially contiguous clusters of total water withdrawal in the Huaihe River watershed during1978–2010,and the spatial agglomeration weakened from 2010 to 2018.This study provides a data-driven framework for assessing water withdrawals to enable a deeper understanding of competing water use among economic sectors as well as water withdrawal modelled with proper data resource and method.
文摘The paper discusses the need of a high-level query language to allow analysts,geographers and,in general,non-programmers to easily cross-analyze multi-source VGI created by means of apps,crowd-sourced data from social networks and authoritative geo-referenced data,usually represented as JSON data sets(nowadays,the de facto standard for data exported by social networks).Since an easy to use high-level language for querying and manipulating collections of possibly geo-tagged JSON objects is still unavailable,we propose a truly declarative language,named J-CO-QL,that is based on a well-defined execution model.A plug-in for a GIS permits to visualize geo-tagged data sets stored in a NoSQL database such as MongoDB;furthermore,the same plug-in can be used to write and execute J-CO-QL queries on those databases.The paper introduces the language by exemplifying its operators within a real study case,the aim of which is to understand the mobility of people in the neighborhood of Bergamo city.Cross-analysis of data about transportation networks and VGI from travelers is performed,by means of J-CO-QL language,capable to manipulate and transform,combine and join possibly geo-tagged JSON objects,in order to produce new possibly geo-tagged JSON objects satisfying users’needs.
基金This work was supported in part by National NSF of China(Nos.61872095,61872128,61571139 and 61201393)New Star of Pearl River on Science and Technology of Guangzhou(No.2014J2200085)+2 种基金the Open Project Program of Shenzhen Key Laboratory of Media Security(Grant No.ML-2018-03)the Opening Project of Guang Dong Province Key Laboratory of Information Security Technology(Grant No.2017B030314131-15)Natural Science Foundation of Xizang(No.2016ZR-MZ-01).
文摘This paper proposes a two-step general framework for reversible data hiding(RDH)schemes with controllable contrast enhancement.The first step aims at preserving visual perception as much as possible on the basis of achieving high embedding capacity(EC),while the second step is used for increasing image contrast.In the second step,some peak-pairs are utilized so that the histogram of pixel values is modified to perform histogram equalization(HE),which would lead to the image contrast enhancement.However,for HE,the utilization of some peak-pairs easily leads to over-enhanced image contrast when a large number of bits are embedded.Therefore,in our proposed framework,contrast over-enhancement is avoided by controlling the degree of contrast enhancement.Since the second step can only provide a small amount of data due to controlled contrast enhancement,the first one helps to achieve a large amount of data without degrading visual quality.Any RDH method which can achieve high EC while preserve good visual quality,can be selected for the first step.In fact,Gao et al.’s method is a special case of our proposed framework.In addition,two simple and commonly-used RDH methods are also introduced to further demonstrate the generalization of our framework.
文摘Point of Care (PoC) devices and systems can be categorized into three broad classes (CAT 1, CAT 2, and CAT 3) based on the context of operation and usage. In this paper, the categories are defined to address certain usage models of the PoC device. PoC devices that are used for PoC testing and diagnostic applications are defined CAT 1 devices;PoC devices that are used for patient monitoring are defined as CAT 2 devices (PoCM);PoC devices that are used for as interfacing with other devices are defined as CAT 3 devices (PoCI). The PoCI devices provide an interface gateway for collecting and aggregating data from other medical devices. In all categories, data security is an important aspect. This paper presents a security framework concept, which is applicable for all of the classes of PoC operation. It outlines the concepts and security framework for preventing security challenges in unauthorized access to data, unintended data flow, and data tampering during communication between system entities, the user, and the PoC system. The security framework includes secure layering of basic PoC system architecture, protection of PoC devices in the context of application and network. Developing the security framework is taken into account of a thread model of the PoC system. A proposal for a low-level protocol is discussed. This protocol is independent of communications technologies, and it is elaborated in relation to providing security. An algorithm that can be used to overcome the threat challenges has been shown using the elements in the protocol. The paper further discusses the vulnerability scanning process for the PoC system interconnected network. The paper also presents a four-step process of authentication and authorization framework for providing the security for the PoC system. Finally, the paper concludes with the machine to machine (M2M) security viewpoint and discusses the key stakeholders within an actual deployment of the PoC system and its security challenges.
文摘The importance of the project selection phase in any six sigma initiative cannot be emphasized enough. The successfulness of the six sigma initiative is affected by successful project selection. Recently, Data Envelopment Analysis (DEA) has been proposed as a six sigma project selection tool. However, there exist a number of different DEA formulations which may affect the selection process and the wining project being selected. This work initially applies nine different DEA formulations to several case studies and concludes that different DEA formulations select different wining projects. Also in this work, a Multi-DEA Unified Scoring Framework is proposed to overcome this problem. This framework is applied to several case studies and proved to successfully select the six sigma project with the best performance. The framework is also successful in filtering out some of the projects that have “selective” excellent performance, i.e. projects with excellent performance in some of the DEA formulations and worse performance in others. It is also successful in selecting stable projects;these are projects that perform well in the majority of the DEA formulations, even if it has not been selected as a wining project by any of the DEA formulations.
文摘Digital educational content is gaining importance as an incubator of pedagogical methodologies in formal and informal online educational settings. Its educational efficiency is directly dependent on its quality, however educational content is more than information and data. This paper presents a new data quality framework for assessing digital educational content used for teaching in distance learning environments. The model relies on the ISO2500 series quality standard and beside providing the mechanisms for multi-facet quality assessment it also supports organizations that design, create, manage and use educational content with the quality tools (expressed as quality metrics and measurement methods) to provide a more efficient distance education experience. The model describes the quality characteristics of the educational material content using data and software quality characteristics.