Well logging technology has accumulated a large amount of historical data through four generations of technological development,which forms the basis of well logging big data and digital assets.However,the value of th...Well logging technology has accumulated a large amount of historical data through four generations of technological development,which forms the basis of well logging big data and digital assets.However,the value of these data has not been well stored,managed and mined.With the development of cloud computing technology,it provides a rare development opportunity for logging big data private cloud.The traditional petrophysical evaluation and interpretation model has encountered great challenges in the face of new evaluation objects.The solution research of logging big data distributed storage,processing and learning functions integrated in logging big data private cloud has not been carried out yet.To establish a distributed logging big-data private cloud platform centered on a unifi ed learning model,which achieves the distributed storage and processing of logging big data and facilitates the learning of novel knowledge patterns via the unifi ed logging learning model integrating physical simulation and data models in a large-scale functional space,thus resolving the geo-engineering evaluation problem of geothermal fi elds.Based on the research idea of“logging big data cloud platform-unifi ed logging learning model-large function space-knowledge learning&discovery-application”,the theoretical foundation of unified learning model,cloud platform architecture,data storage and learning algorithm,arithmetic power allocation and platform monitoring,platform stability,data security,etc.have been carried on analysis.The designed logging big data cloud platform realizes parallel distributed storage and processing of data and learning algorithms.The feasibility of constructing a well logging big data cloud platform based on a unifi ed learning model of physics and data is analyzed in terms of the structure,ecology,management and security of the cloud platform.The case study shows that the logging big data cloud platform has obvious technical advantages over traditional logging evaluation methods in terms of knowledge discovery method,data software and results sharing,accuracy,speed and complexity.展开更多
This paper addresses urban sustainability challenges amid global urbanization, emphasizing the need for innova tive approaches aligned with the Sustainable Development Goals. While traditional tools and linear models ...This paper addresses urban sustainability challenges amid global urbanization, emphasizing the need for innova tive approaches aligned with the Sustainable Development Goals. While traditional tools and linear models offer insights, they fall short in presenting a holistic view of complex urban challenges. System dynamics (SD) models that are often utilized to provide holistic, systematic understanding of a research subject, like the urban system, emerge as valuable tools, but data scarcity and theoretical inadequacy pose challenges. The research reviews relevant papers on recent SD model applications in urban sustainability since 2018, categorizing them based on nine key indicators. Among the reviewed papers, data limitations and model assumptions were identified as ma jor challenges in applying SD models to urban sustainability. This led to exploring the transformative potential of big data analytics, a rare approach in this field as identified by this study, to enhance SD models’ empirical foundation. Integrating big data could provide data-driven calibration, potentially improving predictive accuracy and reducing reliance on simplified assumptions. The paper concludes by advocating for new approaches that reduce assumptions and promote real-time applicable models, contributing to a comprehensive understanding of urban sustainability through the synergy of big data and SD models.展开更多
Recently,internet stimulates the explosive progress of knowledge discovery in big volume data resource,to dig the valuable and hidden rules by computing.Simultaneously,the wireless channel measurement data reveals big...Recently,internet stimulates the explosive progress of knowledge discovery in big volume data resource,to dig the valuable and hidden rules by computing.Simultaneously,the wireless channel measurement data reveals big volume feature,considering the massive antennas,huge bandwidth and versatile application scenarios.This article firstly presents a comprehensive survey of channel measurement and modeling research for mobile communication,especially for 5th Generation(5G) and beyond.Considering the big data research progress,then a cluster-nuclei based model is proposed,which takes advantages of both the stochastical model and deterministic model.The novel model has low complexity with the limited number of cluster-nuclei while the cluster-nuclei has the physical mapping to real propagation objects.Combining the channel properties variation principles with antenna size,frequency,mobility and scenario dug from the channel data,the proposed model can be expanded in versatile application to support future mobile research.展开更多
We examine the role of big data and machine learning in cancer research.We describe an example in cancer research where gene-level data from The Cancer Genome Atlas(TCGA) consortium is interpreted using a pathway-leve...We examine the role of big data and machine learning in cancer research.We describe an example in cancer research where gene-level data from The Cancer Genome Atlas(TCGA) consortium is interpreted using a pathway-level model.As the complexity of computational models increases,their sample requirements grow exponentially.This growth stems from the fact that the number of combinations of variables grows exponentially as the number of variables increases.Thus,a large sample size is needed.The number of variables in a computational model can be reduced by incorporating biological knowledge.One particularly successful way of doing this is by using available gene regulatory,signaling,metabolic,or context-specific pathway information.We conclude that the incorporation of existing biological knowledge is essential for the progress in using big data for cancer research.展开更多
This paper provides a new obstacle avoidance control method for cars based on big data and just-in-time modeling. Just-in-time modeling is a new kind of data-driven control technique in the age of big data and is used...This paper provides a new obstacle avoidance control method for cars based on big data and just-in-time modeling. Just-in-time modeling is a new kind of data-driven control technique in the age of big data and is used in various real systems. The main property of the proposed method is that a gain and a control time which are parameters in the control input to avoid an encountered obstacle are computed from a database which includes a lot of driving data in various situations. Especially, the important advantage of the method is small computation time, and hence it realizes real-time obstacle avoidance control for cars. From some numerical simulations, it is showed that the new control method can make the car avoid various obstacles efficiently in comparison with the previous method.展开更多
These last years we have been witnessing a tremendous growth in the volume and availability of data. This fact results primarily from the emergence of a multitude of sources (e.g. computers, mobile devices, sensors or...These last years we have been witnessing a tremendous growth in the volume and availability of data. This fact results primarily from the emergence of a multitude of sources (e.g. computers, mobile devices, sensors or social networks) that are continuously producing either structured, semi-structured or unstructured data. Database Management Systems and Data Warehouses are no longer the only technologies used to store and analyze datasets, namely due to the volume and complex structure of nowadays data that degrade their performance and scalability. Big Data is one of the recent challenges, since it implies new requirements in terms of data storage, processing and visualization. Despite that, analyzing properly Big Data can constitute great advantages because it allows discovering patterns and correlations in datasets. Users can use this processed information to gain deeper insights and to get business advantages. Thus, data modeling and data analytics are evolved in a way that we are able to process huge amounts of data without compromising performance and availability, but instead by “relaxing” the usual ACID properties. This paper provides a broad view and discussion of the current state of this subject with a particular focus on data modeling and data analytics, describing and clarifying the main differences between the three main approaches in what concerns these aspects, namely: operational databases, decision support databases and Big Data technologies.展开更多
The 19th National Congress of the Communist Party of China has put forward higher requirements for Chinese government governance. The government governance has developed to a higher stage. Meanwhile, it faces more cha...The 19th National Congress of the Communist Party of China has put forward higher requirements for Chinese government governance. The government governance has developed to a higher stage. Meanwhile, it faces more challenges, like lack of top-level design and information sharing. To develop a government governance decision-making innovation model, we should make good use of big data to mine in the grassroots government data management network. Both the characteristics of the times and the experience of the practice have proven that big data can empower government governance and promote the construction of a service-oriented government.展开更多
In recent years,more and more researches focus on the self characteristics and spatial location of housing,and explore the influencing factors of urban housing price from the micro perspective.As representative of big...In recent years,more and more researches focus on the self characteristics and spatial location of housing,and explore the influencing factors of urban housing price from the micro perspective.As representative of big cities,spatial distribution pattern of housing price in national central cities has attracted much attention.In order to return the spatial distribution pattern of housing price to the research on influencing factors of housing price,the reasons behind the spatial distribution pattern of housing price in three national central cities:Beijing,Wuhan and Chongqing are explored.The results show that①urban housing price is affected by many factors.Due to different social and economic conditions in each city,there are differences in the influence direction of the proximity to expressways,city squares,universities and living facilities,characteristics of companies and enterprises on Beijing,Wuhan and Chongqing.②Various factors have different value-added effects on housing price in different cities.The location of ring line in Beijing and Wuhan has the greatest increase effect on housing price,while metro station of Chongqing has the greatest increase effect on housing price.展开更多
Facing the development of future 5 G, the emerging technologies such as Internet of things, big data, cloud computing, and artificial intelligence is enhancing an explosive growth in data traffic. Radical changes in c...Facing the development of future 5 G, the emerging technologies such as Internet of things, big data, cloud computing, and artificial intelligence is enhancing an explosive growth in data traffic. Radical changes in communication theory and implement technologies, the wireless communications and wireless networks have entered a new era. Among them, wireless big data(WBD) has tremendous value, and artificial intelligence(AI) gives unthinkable possibilities. However, in the big data development and artificial intelligence application groups, the lack of a sound theoretical foundation and mathematical methods is regarded as a real challenge that needs to be solved. From the basic problem of wireless communication, the interrelationship of demand, environment and ability, this paper intends to investigate the concept and data model of WBD, the wireless data mining, the wireless knowledge and wireless knowledge learning(WKL), and typical practices examples, to facilitate and open up more opportunities of WBD research and developments. Such research is beneficial for creating new theoretical foundation and emerging technologies of future wireless communications.展开更多
The current global challenges that threaten biodiversity are immense and rapidly growing.These biodiversity challenges demand approaches that meld bioinformatics,large-scale phylogeny reconstruction,use of digitized s...The current global challenges that threaten biodiversity are immense and rapidly growing.These biodiversity challenges demand approaches that meld bioinformatics,large-scale phylogeny reconstruction,use of digitized specimen data,and complex post-tree analyses(e.g.niche modeling,niche diversification,and other ecological analyses).Recent developments in phylogenetics coupled with emerging cyberinfrastructure and new data sources provide unparalleled opportunities for mobilizing and integrating massive amounts of biological data,driving the discovery of complex patterns and new hypotheses for further study.These developments are not trivial in that biodiversity data on the global scale now being collected and analyzed are inherently complex.The ongoing integration and maturation of biodiversity tools discussed here is transforming biodiversity science,enabling what we broadly term"next-generation"investigations in systematics,ecology,and evolution(i.e.,"biodiversity science").New training that integrates domain knowledge in biodiversity and data science skills is also needed to accelerate research in these areas.Integrative biodiversity science is crucial to the future of global biodiversity.We cannot simply react to continued threats to biodiversity,but via the use of an integrative,multifaceted,big data approach,researchers can now make biodiversity projections to provide crucial data not only for scientists,but also for the public,land managers,policy makers,urban planners,and agriculture.展开更多
Eighty percent of big data are associated with spatial information,and thus are Big Spatial Data(BSD).BSD provides new and great opportunities to rework problems in urban and environmental sustainability with advanced...Eighty percent of big data are associated with spatial information,and thus are Big Spatial Data(BSD).BSD provides new and great opportunities to rework problems in urban and environmental sustainability with advanced BSD analytics.To fully leverage the advantages of BSD,it is integrated with conventional data(e.g.remote sensing images)and improved methods are developed.This paper introduces four case studies:(1)Detection of polycentric urban structures;(2)Evaluation of urban vibrancy;(3)Estimation of population exposure to PM2.5;and(4)Urban land-use classification via deep learning.The results provide evidence that integrated methods can harness the advantages of both traditional data and BSD.Meanwhile,they can also improve the effectiveness of big data itself.Finally,this study makes three key recommendations for the development of BSD with regards to data fusion,data and predicting analytics,and theoretical modeling.展开更多
In big data of business service or transaction,it is impossible to provide entire information to both of services from cyber system,so some service providers made use of maliciously services to get more interests.Trus...In big data of business service or transaction,it is impossible to provide entire information to both of services from cyber system,so some service providers made use of maliciously services to get more interests.Trust management is an effective solution to deal with these malicious actions.This paper gave a trust computing model based on service-recommendation in big data.This model takes into account difference of recommendation trust between familiar node and stranger node.Thus,to ensure accuracy of recommending trust computing,paper proposed a fine-granularity similarity computing method based on the similarity of service concept domain ontology.This model is more accurate in computing trust value of cyber service nodes and prevents better cheating and attacking of malicious service nodes.Experiment results illustrated our model is effective.展开更多
Multi-level searching is called Drill down search.Right now,no drill down search feature is available in the existing search engines like Google,Yahoo,Bing and Baidu.Drill down search is very much useful for the end u...Multi-level searching is called Drill down search.Right now,no drill down search feature is available in the existing search engines like Google,Yahoo,Bing and Baidu.Drill down search is very much useful for the end user tofind the exact search results among the huge paginated search results.Higher level of drill down search with category based search feature leads to get the most accurate search results but it increases the number and size of thefile system.The purpose of this manuscript is to implement a big data storage reduction binaryfile system model for category based drill down search engine that offers fast multi-levelfiltering capability.The basic methodology of the proposed model stores the search engine data in the binaryfile system model.To verify the effectiveness of the proposedfile system model,5 million unique keyword data are stored into a binaryfile,thereby analysing the proposedfile system with efficiency.Some experimental results are also provided based on real data that show our storage model speed and superiority.Experiments demonstrated that ourfile system expansion ratio is constant and it reduces the disk storage space up to 30%with conventional database/file system and it also increases the search performance for any levels of search.To discuss deeply,the paper starts with the short introduction of drill down search followed by the discussion of important technologies used to implement big data storage reduction system in detail.展开更多
Damage caused by people and organizations unconnected with the pipeline management is a major risk faced by pipelines,and its consequences can have a huge impact.However,the present measures to monitor this have major...Damage caused by people and organizations unconnected with the pipeline management is a major risk faced by pipelines,and its consequences can have a huge impact.However,the present measures to monitor this have major problems such as time delays,overlooking threats,and false alarms.To overcome the disadvantages of these methods,analysis of big location data from mobile phone systems was applied to prevent third-party damage to pipelines,and a third-party damage prevention system was developed for pipelines including encryption mobile phone data,data preprocessing,and extraction of characteristic patterns.By applying this to natural gas pipelines,a large amount of location data was collected for data feature recognition and model analysis.Third-party illegal construction and occupation activities were discovered in a timely manner.This is important for preventing third-party damage to pipelines.展开更多
This study focuses on meeting the challenges of big data visualization by using of data reduction methods based the feature selection methods.To reduce the volume of big data and minimize model training time(Tt)while ...This study focuses on meeting the challenges of big data visualization by using of data reduction methods based the feature selection methods.To reduce the volume of big data and minimize model training time(Tt)while maintaining data quality.We contributed to meeting the challenges of big data visualization using the embedded method based“Select from model(SFM)”method by using“Random forest Importance algorithm(RFI)”and comparing it with the filter method by using“Select percentile(SP)”method based chi square“Chi2”tool for selecting the most important features,which are then fed into a classification process using the logistic regression(LR)algorithm and the k-nearest neighbor(KNN)algorithm.Thus,the classification accuracy(AC)performance of LRis also compared to theKNN approach in python on eight data sets to see which method produces the best rating when feature selection methods are applied.Consequently,the study concluded that the feature selection methods have a significant impact on the analysis and visualization of the data after removing the repetitive data and the data that do not affect the goal.After making several comparisons,the study suggests(SFMLR)using SFM based on RFI algorithm for feature selection,with LR algorithm for data classify.The proposal proved its efficacy by comparing its results with recent literature.展开更多
With the rapid development of technology,geological big data is increasing explosively,and plays an increasingly important position in the national economy(Zhang and Zhou,2017;Zhou et al.,2018).Governments and agencie...With the rapid development of technology,geological big data is increasing explosively,and plays an increasingly important position in the national economy(Zhang and Zhou,2017;Zhou et al.,2018).Governments and agencies attach great importance to the open internet service of geological big data and information at home,and abroad(Yan et al.,2013;Guo et al.,2014).The basic norms of western countries’geological data information services are rich and varied products.展开更多
With market competition becoming fiercer,enterprises must update their products by constantly assimilating new big data knowledge and private knowledge to maintain their market shares at different time points in the b...With market competition becoming fiercer,enterprises must update their products by constantly assimilating new big data knowledge and private knowledge to maintain their market shares at different time points in the big data environment.Typically,there is mutual influence between each knowledge transfer if the time interval is not too long.It is necessary to study the problem of continuous knowledge transfer in the big data environment.Based on research on one-time knowledge transfer,a model of continuous knowledge transfer is presented,which can consider the interaction between knowledge transfer and determine the optimal knowledge transfer time at different time points in the big data environment.Simulation experiments were performed by adjusting several parameters.The experimental results verified the model’s validity and facilitated conclusions regarding their practical application values.The experimental results can provide more effective decisions for enterprises that must carry out continuous knowledge transfer in the big data environment.展开更多
Global warming has caused the Arctic Ocean ice cover to shrink.This endangers the environment but has made traversing the Arctic channel possible.Therefore,the strategic position of the Arctic has been significantly i...Global warming has caused the Arctic Ocean ice cover to shrink.This endangers the environment but has made traversing the Arctic channel possible.Therefore,the strategic position of the Arctic has been significantly improved.As a near-Arctic country,China has formulated relevant policies that will be directly impacted by changes in the international relations between the eight Arctic countries(regions).A comprehensive and real-time analysis of the various characteristics of the Arctic geographical relationship is required in China,which helps formulate political,economic,and diplomatic countermeasures.Massive global real-time open databases provide news data from major media in various countries.This makes it possible to monitor geographical relationships in real-time.This paper explores key elements of the social development of eight Arctic countries(regions)over 2013-2019 based on the GDELT database and the method of labeled latent Dirichlet allocation.This paper also constructs the national interaction network and identifies the evolution pattern for the relationships between Arctic countries(regions).The following conclusions are drawn.(1)Arctic news hotspot is now focusing on climate change/ice cap melting which is becoming the main driving factor for changes in geographical relationships in the Arctic.(2)There is a strong correlation between the number of news pieces about ice cap melting and the sea ice area.(3)With the melting of the ice caps,the social,economic,and military activities in the Arctic have been booming,and the competition for dominance is becoming increasingly fierce.In general,there is a pattern of domination by Russia and Canada.展开更多
Although the Internet of Things has been widely applied,the problems of cloud computing in the application of digital smart medical Big Data collection,processing,analysis,and storage remain,especially the low efficie...Although the Internet of Things has been widely applied,the problems of cloud computing in the application of digital smart medical Big Data collection,processing,analysis,and storage remain,especially the low efficiency of medical diagnosis.And with the wide application of the Internet of Things and Big Data in the medical field,medical Big Data is increasing in geometric magnitude resulting in cloud service overload,insufficient storage,communication delay,and network congestion.In order to solve these medical and network problems,a medical big-data-oriented fog computing architec-ture and BP algorithm application are proposed,and its structural advantages and characteristics are studied.This architecture enables the medical Big Data generated by medical edge devices and the existing data in the cloud service center to calculate,compare and analyze the fog node through the Internet of Things.The diagnosis results are designed to reduce the business processing delay and improve the diagnosis effect.Considering the weak computing of each edge device,the artificial intelligence BP neural network algorithm is used in the core computing model of the medical diagnosis system to improve the system computing power,enhance the medical intelligence-aided decision-making,and improve the clinical diagnosis and treatment efficiency.In the application process,combined with the characteristics of medical Big Data technology,through fog architecture design and Big Data technology integration,we could research the processing and analysis of heterogeneous data of the medical diagnosis system in the context of the Internet of Things.The results are promising:The medical platform network is smooth,the data storage space is sufficient,the data processing and analysis speed is fast,the diagnosis effect is remarkable,and it is a good assistant to doctors’treatment effect.It not only effectively solves the problem of low clinical diagnosis,treatment efficiency and quality,but also reduces the waiting time of patients,effectively solves the contradiction between doctors and patients,and improves the medical service quality and management level.展开更多
Today,we are living in the era of“big data”where massive amounts of data are used for quantitative decisions and communication management.With the continuous penetration of big data-based intelligent technology in a...Today,we are living in the era of“big data”where massive amounts of data are used for quantitative decisions and communication management.With the continuous penetration of big data-based intelligent technology in all fields of human life,the enormous commercial value inherent in the data industry has become a crucial force that drives the aggregation of new industries.For the publishing industry,the introduction of big data and relevant intelligent technologies,such as data intelligence analysis and scenario services,into the structure and value system of the publishing industry,has become an effective path to expanding and reshaping the demand space of publishing products,content decisions,workflow chain,and marketing direction.In the integration and reconstruction of big data,cloud computing,artificial intelligence,and other related technologies,it is expected that a generalized publishing industry pattern dominated by virtual interaction will be formed in the future.展开更多
基金supported By Grant (PLN2022-14) of State Key Laboratory of Oil and Gas Reservoir Geology and Exploitation (Southwest Petroleum University)。
文摘Well logging technology has accumulated a large amount of historical data through four generations of technological development,which forms the basis of well logging big data and digital assets.However,the value of these data has not been well stored,managed and mined.With the development of cloud computing technology,it provides a rare development opportunity for logging big data private cloud.The traditional petrophysical evaluation and interpretation model has encountered great challenges in the face of new evaluation objects.The solution research of logging big data distributed storage,processing and learning functions integrated in logging big data private cloud has not been carried out yet.To establish a distributed logging big-data private cloud platform centered on a unifi ed learning model,which achieves the distributed storage and processing of logging big data and facilitates the learning of novel knowledge patterns via the unifi ed logging learning model integrating physical simulation and data models in a large-scale functional space,thus resolving the geo-engineering evaluation problem of geothermal fi elds.Based on the research idea of“logging big data cloud platform-unifi ed logging learning model-large function space-knowledge learning&discovery-application”,the theoretical foundation of unified learning model,cloud platform architecture,data storage and learning algorithm,arithmetic power allocation and platform monitoring,platform stability,data security,etc.have been carried on analysis.The designed logging big data cloud platform realizes parallel distributed storage and processing of data and learning algorithms.The feasibility of constructing a well logging big data cloud platform based on a unifi ed learning model of physics and data is analyzed in terms of the structure,ecology,management and security of the cloud platform.The case study shows that the logging big data cloud platform has obvious technical advantages over traditional logging evaluation methods in terms of knowledge discovery method,data software and results sharing,accuracy,speed and complexity.
基金sponsored by the U.S.Department of Housing and Urban Development(Grant No.NJLTS0027-22)The opinions expressed in this study are the authors alone,and do not represent the U.S.Depart-ment of HUD’s opinions.
文摘This paper addresses urban sustainability challenges amid global urbanization, emphasizing the need for innova tive approaches aligned with the Sustainable Development Goals. While traditional tools and linear models offer insights, they fall short in presenting a holistic view of complex urban challenges. System dynamics (SD) models that are often utilized to provide holistic, systematic understanding of a research subject, like the urban system, emerge as valuable tools, but data scarcity and theoretical inadequacy pose challenges. The research reviews relevant papers on recent SD model applications in urban sustainability since 2018, categorizing them based on nine key indicators. Among the reviewed papers, data limitations and model assumptions were identified as ma jor challenges in applying SD models to urban sustainability. This led to exploring the transformative potential of big data analytics, a rare approach in this field as identified by this study, to enhance SD models’ empirical foundation. Integrating big data could provide data-driven calibration, potentially improving predictive accuracy and reducing reliance on simplified assumptions. The paper concludes by advocating for new approaches that reduce assumptions and promote real-time applicable models, contributing to a comprehensive understanding of urban sustainability through the synergy of big data and SD models.
基金supported in part by National Natural Science Foundation of China (61322110, 6141101115)Doctoral Fund of Ministry of Education (201300051100013)
文摘Recently,internet stimulates the explosive progress of knowledge discovery in big volume data resource,to dig the valuable and hidden rules by computing.Simultaneously,the wireless channel measurement data reveals big volume feature,considering the massive antennas,huge bandwidth and versatile application scenarios.This article firstly presents a comprehensive survey of channel measurement and modeling research for mobile communication,especially for 5th Generation(5G) and beyond.Considering the big data research progress,then a cluster-nuclei based model is proposed,which takes advantages of both the stochastical model and deterministic model.The novel model has low complexity with the limited number of cluster-nuclei while the cluster-nuclei has the physical mapping to real propagation objects.Combining the channel properties variation principles with antenna size,frequency,mobility and scenario dug from the channel data,the proposed model can be expanded in versatile application to support future mobile research.
文摘We examine the role of big data and machine learning in cancer research.We describe an example in cancer research where gene-level data from The Cancer Genome Atlas(TCGA) consortium is interpreted using a pathway-level model.As the complexity of computational models increases,their sample requirements grow exponentially.This growth stems from the fact that the number of combinations of variables grows exponentially as the number of variables increases.Thus,a large sample size is needed.The number of variables in a computational model can be reduced by incorporating biological knowledge.One particularly successful way of doing this is by using available gene regulatory,signaling,metabolic,or context-specific pathway information.We conclude that the incorporation of existing biological knowledge is essential for the progress in using big data for cancer research.
文摘This paper provides a new obstacle avoidance control method for cars based on big data and just-in-time modeling. Just-in-time modeling is a new kind of data-driven control technique in the age of big data and is used in various real systems. The main property of the proposed method is that a gain and a control time which are parameters in the control input to avoid an encountered obstacle are computed from a database which includes a lot of driving data in various situations. Especially, the important advantage of the method is small computation time, and hence it realizes real-time obstacle avoidance control for cars. From some numerical simulations, it is showed that the new control method can make the car avoid various obstacles efficiently in comparison with the previous method.
文摘These last years we have been witnessing a tremendous growth in the volume and availability of data. This fact results primarily from the emergence of a multitude of sources (e.g. computers, mobile devices, sensors or social networks) that are continuously producing either structured, semi-structured or unstructured data. Database Management Systems and Data Warehouses are no longer the only technologies used to store and analyze datasets, namely due to the volume and complex structure of nowadays data that degrade their performance and scalability. Big Data is one of the recent challenges, since it implies new requirements in terms of data storage, processing and visualization. Despite that, analyzing properly Big Data can constitute great advantages because it allows discovering patterns and correlations in datasets. Users can use this processed information to gain deeper insights and to get business advantages. Thus, data modeling and data analytics are evolved in a way that we are able to process huge amounts of data without compromising performance and availability, but instead by “relaxing” the usual ACID properties. This paper provides a broad view and discussion of the current state of this subject with a particular focus on data modeling and data analytics, describing and clarifying the main differences between the three main approaches in what concerns these aspects, namely: operational databases, decision support databases and Big Data technologies.
文摘The 19th National Congress of the Communist Party of China has put forward higher requirements for Chinese government governance. The government governance has developed to a higher stage. Meanwhile, it faces more challenges, like lack of top-level design and information sharing. To develop a government governance decision-making innovation model, we should make good use of big data to mine in the grassroots government data management network. Both the characteristics of the times and the experience of the practice have proven that big data can empower government governance and promote the construction of a service-oriented government.
基金Sponsored by National Natural Science Foundation of China (51808413)General Project of Hubei Social Science Fund (2018193)+1 种基金Innovation and Entrepreneurship Training Program for College Students in Hubei Province (S201910490024)University-level Graduate Innovation Fund of Wuhan Institute of Technology (CX2019036)。
文摘In recent years,more and more researches focus on the self characteristics and spatial location of housing,and explore the influencing factors of urban housing price from the micro perspective.As representative of big cities,spatial distribution pattern of housing price in national central cities has attracted much attention.In order to return the spatial distribution pattern of housing price to the research on influencing factors of housing price,the reasons behind the spatial distribution pattern of housing price in three national central cities:Beijing,Wuhan and Chongqing are explored.The results show that①urban housing price is affected by many factors.Due to different social and economic conditions in each city,there are differences in the influence direction of the proximity to expressways,city squares,universities and living facilities,characteristics of companies and enterprises on Beijing,Wuhan and Chongqing.②Various factors have different value-added effects on housing price in different cities.The location of ring line in Beijing and Wuhan has the greatest increase effect on housing price,while metro station of Chongqing has the greatest increase effect on housing price.
文摘Facing the development of future 5 G, the emerging technologies such as Internet of things, big data, cloud computing, and artificial intelligence is enhancing an explosive growth in data traffic. Radical changes in communication theory and implement technologies, the wireless communications and wireless networks have entered a new era. Among them, wireless big data(WBD) has tremendous value, and artificial intelligence(AI) gives unthinkable possibilities. However, in the big data development and artificial intelligence application groups, the lack of a sound theoretical foundation and mathematical methods is regarded as a real challenge that needs to be solved. From the basic problem of wireless communication, the interrelationship of demand, environment and ability, this paper intends to investigate the concept and data model of WBD, the wireless data mining, the wireless knowledge and wireless knowledge learning(WKL), and typical practices examples, to facilitate and open up more opportunities of WBD research and developments. Such research is beneficial for creating new theoretical foundation and emerging technologies of future wireless communications.
基金supported in part by US NSF grants EF-1115210,DBI-1547229,DBI-1458640,DEB-1442280,and DEB-1208809
文摘The current global challenges that threaten biodiversity are immense and rapidly growing.These biodiversity challenges demand approaches that meld bioinformatics,large-scale phylogeny reconstruction,use of digitized specimen data,and complex post-tree analyses(e.g.niche modeling,niche diversification,and other ecological analyses).Recent developments in phylogenetics coupled with emerging cyberinfrastructure and new data sources provide unparalleled opportunities for mobilizing and integrating massive amounts of biological data,driving the discovery of complex patterns and new hypotheses for further study.These developments are not trivial in that biodiversity data on the global scale now being collected and analyzed are inherently complex.The ongoing integration and maturation of biodiversity tools discussed here is transforming biodiversity science,enabling what we broadly term"next-generation"investigations in systematics,ecology,and evolution(i.e.,"biodiversity science").New training that integrates domain knowledge in biodiversity and data science skills is also needed to accelerate research in these areas.Integrative biodiversity science is crucial to the future of global biodiversity.We cannot simply react to continued threats to biodiversity,but via the use of an integrative,multifaceted,big data approach,researchers can now make biodiversity projections to provide crucial data not only for scientists,but also for the public,land managers,policy makers,urban planners,and agriculture.
文摘Eighty percent of big data are associated with spatial information,and thus are Big Spatial Data(BSD).BSD provides new and great opportunities to rework problems in urban and environmental sustainability with advanced BSD analytics.To fully leverage the advantages of BSD,it is integrated with conventional data(e.g.remote sensing images)and improved methods are developed.This paper introduces four case studies:(1)Detection of polycentric urban structures;(2)Evaluation of urban vibrancy;(3)Estimation of population exposure to PM2.5;and(4)Urban land-use classification via deep learning.The results provide evidence that integrated methods can harness the advantages of both traditional data and BSD.Meanwhile,they can also improve the effectiveness of big data itself.Finally,this study makes three key recommendations for the development of BSD with regards to data fusion,data and predicting analytics,and theoretical modeling.
文摘In big data of business service or transaction,it is impossible to provide entire information to both of services from cyber system,so some service providers made use of maliciously services to get more interests.Trust management is an effective solution to deal with these malicious actions.This paper gave a trust computing model based on service-recommendation in big data.This model takes into account difference of recommendation trust between familiar node and stranger node.Thus,to ensure accuracy of recommending trust computing,paper proposed a fine-granularity similarity computing method based on the similarity of service concept domain ontology.This model is more accurate in computing trust value of cyber service nodes and prevents better cheating and attacking of malicious service nodes.Experiment results illustrated our model is effective.
文摘Multi-level searching is called Drill down search.Right now,no drill down search feature is available in the existing search engines like Google,Yahoo,Bing and Baidu.Drill down search is very much useful for the end user tofind the exact search results among the huge paginated search results.Higher level of drill down search with category based search feature leads to get the most accurate search results but it increases the number and size of thefile system.The purpose of this manuscript is to implement a big data storage reduction binaryfile system model for category based drill down search engine that offers fast multi-levelfiltering capability.The basic methodology of the proposed model stores the search engine data in the binaryfile system model.To verify the effectiveness of the proposedfile system model,5 million unique keyword data are stored into a binaryfile,thereby analysing the proposedfile system with efficiency.Some experimental results are also provided based on real data that show our storage model speed and superiority.Experiments demonstrated that ourfile system expansion ratio is constant and it reduces the disk storage space up to 30%with conventional database/file system and it also increases the search performance for any levels of search.To discuss deeply,the paper starts with the short introduction of drill down search followed by the discussion of important technologies used to implement big data storage reduction system in detail.
基金supported by Pipeline Management Data Analysis and Typical Model Research [Grant Number 2016B-3105-0501]CNPC (China National Petroleum Corporation) project, Research on Oil and Gas Pipeline Safety and Reliability Operating [Grant Number 2015-B025-0628]
文摘Damage caused by people and organizations unconnected with the pipeline management is a major risk faced by pipelines,and its consequences can have a huge impact.However,the present measures to monitor this have major problems such as time delays,overlooking threats,and false alarms.To overcome the disadvantages of these methods,analysis of big location data from mobile phone systems was applied to prevent third-party damage to pipelines,and a third-party damage prevention system was developed for pipelines including encryption mobile phone data,data preprocessing,and extraction of characteristic patterns.By applying this to natural gas pipelines,a large amount of location data was collected for data feature recognition and model analysis.Third-party illegal construction and occupation activities were discovered in a timely manner.This is important for preventing third-party damage to pipelines.
文摘This study focuses on meeting the challenges of big data visualization by using of data reduction methods based the feature selection methods.To reduce the volume of big data and minimize model training time(Tt)while maintaining data quality.We contributed to meeting the challenges of big data visualization using the embedded method based“Select from model(SFM)”method by using“Random forest Importance algorithm(RFI)”and comparing it with the filter method by using“Select percentile(SP)”method based chi square“Chi2”tool for selecting the most important features,which are then fed into a classification process using the logistic regression(LR)algorithm and the k-nearest neighbor(KNN)algorithm.Thus,the classification accuracy(AC)performance of LRis also compared to theKNN approach in python on eight data sets to see which method produces the best rating when feature selection methods are applied.Consequently,the study concluded that the feature selection methods have a significant impact on the analysis and visualization of the data after removing the repetitive data and the data that do not affect the goal.After making several comparisons,the study suggests(SFMLR)using SFM based on RFI algorithm for feature selection,with LR algorithm for data classify.The proposal proved its efficacy by comparing its results with recent literature.
基金granted by the National Key R&D Program of China(Grant No.2016YFC0600510)the Ministry of Land and Resources"Twelfth Five-Year Plan"Key Projects(Grant No.1212011220352).
文摘With the rapid development of technology,geological big data is increasing explosively,and plays an increasingly important position in the national economy(Zhang and Zhou,2017;Zhou et al.,2018).Governments and agencies attach great importance to the open internet service of geological big data and information at home,and abroad(Yan et al.,2013;Guo et al.,2014).The basic norms of western countries’geological data information services are rich and varied products.
基金supported by the National Natural Science Foundation of China(Grant No.71704016,71331008)the Natural Science Foundation of Hunan Province(Grant No.2017JJ2267)+1 种基金Key Projects of Chinese Ministry of Education(17JZD022)the Project of China Scholarship Council for Overseas Studies(201208430233,201508430121),which are acknowledged.
文摘With market competition becoming fiercer,enterprises must update their products by constantly assimilating new big data knowledge and private knowledge to maintain their market shares at different time points in the big data environment.Typically,there is mutual influence between each knowledge transfer if the time interval is not too long.It is necessary to study the problem of continuous knowledge transfer in the big data environment.Based on research on one-time knowledge transfer,a model of continuous knowledge transfer is presented,which can consider the interaction between knowledge transfer and determine the optimal knowledge transfer time at different time points in the big data environment.Simulation experiments were performed by adjusting several parameters.The experimental results verified the model’s validity and facilitated conclusions regarding their practical application values.The experimental results can provide more effective decisions for enterprises that must carry out continuous knowledge transfer in the big data environment.
基金National Natural Science Foundation of China(42071153)The Strategic Priority Research Program of Chinese Academy of Sciences(XDA19040401)The Strategic Priority Research Program of Chinese Academy of Sciences(XDA20080100)。
文摘Global warming has caused the Arctic Ocean ice cover to shrink.This endangers the environment but has made traversing the Arctic channel possible.Therefore,the strategic position of the Arctic has been significantly improved.As a near-Arctic country,China has formulated relevant policies that will be directly impacted by changes in the international relations between the eight Arctic countries(regions).A comprehensive and real-time analysis of the various characteristics of the Arctic geographical relationship is required in China,which helps formulate political,economic,and diplomatic countermeasures.Massive global real-time open databases provide news data from major media in various countries.This makes it possible to monitor geographical relationships in real-time.This paper explores key elements of the social development of eight Arctic countries(regions)over 2013-2019 based on the GDELT database and the method of labeled latent Dirichlet allocation.This paper also constructs the national interaction network and identifies the evolution pattern for the relationships between Arctic countries(regions).The following conclusions are drawn.(1)Arctic news hotspot is now focusing on climate change/ice cap melting which is becoming the main driving factor for changes in geographical relationships in the Arctic.(2)There is a strong correlation between the number of news pieces about ice cap melting and the sea ice area.(3)With the melting of the ice caps,the social,economic,and military activities in the Arctic have been booming,and the competition for dominance is becoming increasingly fierce.In general,there is a pattern of domination by Russia and Canada.
基金supported by 2020 Foshan Science and Technology Project(Numbering:2020001005356),Baoling Qin received the grant.
文摘Although the Internet of Things has been widely applied,the problems of cloud computing in the application of digital smart medical Big Data collection,processing,analysis,and storage remain,especially the low efficiency of medical diagnosis.And with the wide application of the Internet of Things and Big Data in the medical field,medical Big Data is increasing in geometric magnitude resulting in cloud service overload,insufficient storage,communication delay,and network congestion.In order to solve these medical and network problems,a medical big-data-oriented fog computing architec-ture and BP algorithm application are proposed,and its structural advantages and characteristics are studied.This architecture enables the medical Big Data generated by medical edge devices and the existing data in the cloud service center to calculate,compare and analyze the fog node through the Internet of Things.The diagnosis results are designed to reduce the business processing delay and improve the diagnosis effect.Considering the weak computing of each edge device,the artificial intelligence BP neural network algorithm is used in the core computing model of the medical diagnosis system to improve the system computing power,enhance the medical intelligence-aided decision-making,and improve the clinical diagnosis and treatment efficiency.In the application process,combined with the characteristics of medical Big Data technology,through fog architecture design and Big Data technology integration,we could research the processing and analysis of heterogeneous data of the medical diagnosis system in the context of the Internet of Things.The results are promising:The medical platform network is smooth,the data storage space is sufficient,the data processing and analysis speed is fast,the diagnosis effect is remarkable,and it is a good assistant to doctors’treatment effect.It not only effectively solves the problem of low clinical diagnosis,treatment efficiency and quality,but also reduces the waiting time of patients,effectively solves the contradiction between doctors and patients,and improves the medical service quality and management level.
文摘Today,we are living in the era of“big data”where massive amounts of data are used for quantitative decisions and communication management.With the continuous penetration of big data-based intelligent technology in all fields of human life,the enormous commercial value inherent in the data industry has become a crucial force that drives the aggregation of new industries.For the publishing industry,the introduction of big data and relevant intelligent technologies,such as data intelligence analysis and scenario services,into the structure and value system of the publishing industry,has become an effective path to expanding and reshaping the demand space of publishing products,content decisions,workflow chain,and marketing direction.In the integration and reconstruction of big data,cloud computing,artificial intelligence,and other related technologies,it is expected that a generalized publishing industry pattern dominated by virtual interaction will be formed in the future.