Open networks and heterogeneous services in the Internet of Vehicles(IoV)can lead to security and privacy challenges.One key requirement for such systems is the preservation of user privacy,ensuring a seamless experie...Open networks and heterogeneous services in the Internet of Vehicles(IoV)can lead to security and privacy challenges.One key requirement for such systems is the preservation of user privacy,ensuring a seamless experience in driving,navigation,and communication.These privacy needs are influenced by various factors,such as data collected at different intervals,trip durations,and user interactions.To address this,the paper proposes a Support Vector Machine(SVM)model designed to process large amounts of aggregated data and recommend privacy preserving measures.The model analyzes data based on user demands and interactions with service providers or neighboring infrastructure.It aims to minimize privacy risks while ensuring service continuity and sustainability.The SVMmodel helps validate the system’s reliability by creating a hyperplane that distinguishes between maximum and minimum privacy recommendations.The results demonstrate the effectiveness of the proposed SVM model in enhancing both privacy and service performance.展开更多
Miniature air quality sensors are widely used in urban grid-based monitoring due to their flexibility in deployment and low cost.However,the raw data collected by these devices often suffer from low accuracy caused by...Miniature air quality sensors are widely used in urban grid-based monitoring due to their flexibility in deployment and low cost.However,the raw data collected by these devices often suffer from low accuracy caused by environmental interference and sensor drift,highlighting the need for effective calibration methods to improve data reliability.This study proposes a data correction method based on Bayesian Optimization Support Vector Regression(BO-SVR),which combines the nonlinear modeling capability of Support Vector Regression(SVR)with the efficient global hyperparameter search of Bayesian Optimization.By introducing cross-validation loss as the optimization objective and using Gaussian process modeling with an Expected Improvement acquisition strategy,the approach automatically determines optimal hyperparameters for accurate pollutant concentration prediction.Experiments on real-world micro-sensor datasets demonstrate that BO-SVR outperforms traditional SVR,grid search SVR,and random forest(RF)models across multiple pollutants,including PM_(2.5),PM_(10),CO,NO_(2),SO_(2),and O_(3).The proposed method achieves lower prediction residuals,higher fitting accuracy,and better generalization,offering an efficient and practical solution for enhancing the quality of micro-sensor air monitoring data.展开更多
In wireless communication,the problem of authenticating the transmitter’s identity is challeng-ing,especially for those terminal devices in which the security schemes based on cryptography are approxi-mately unfeasib...In wireless communication,the problem of authenticating the transmitter’s identity is challeng-ing,especially for those terminal devices in which the security schemes based on cryptography are approxi-mately unfeasible owing to limited resources.In this paper,a physical layer authentication scheme is pro-posed to detect whether there is anomalous access by the attackers disguised as legitimate users.Explicitly,channel state information(CSI)is used as a form of fingerprint to exploit spatial discrimination among de-vices in the wireless network and machine learning(ML)technology is employed to promote the improve-ment of authentication accuracy.Considering that the falsified messages are not accessible for authenticator during the training phase,deep support vector data de-scription(Deep SVDD)is selected to solve the one-class classification(OCC)problem.Simulation results show that Deep SVDD based scheme can tackle the challenges of physical layer authentication in wireless communication environments.展开更多
Deformation monitoring is a critical measure for intuitively reflecting the operational behavior of a dam.However,the deformation monitoring data are often incomplete due to environmental changes,monitoring instrument...Deformation monitoring is a critical measure for intuitively reflecting the operational behavior of a dam.However,the deformation monitoring data are often incomplete due to environmental changes,monitoring instrument faults,and human operational errors,thereby often hindering the accurate assessment of actual deformation patterns.This study proposed a method for quantifying deformation similarity between measurement points by recognizing the spatiotemporal characteristics of concrete dam deformation monitoring data.It introduces a spatiotemporal clustering analysis of the concrete dam deformation behavior and employs the support vector machine model to address the missing data in concrete dam deformation monitoring.The proposed method was validated in a concrete dam project,with the model error maintaining within 5%,demonstrating its effectiveness in processing missing deformation data.This approach enhances the capability of early-warning systems and contributes to enhanced dam safety management.展开更多
The paper presents a set of techniques of digital watermarking by which copyright and user rights messages are hidden into geo-spatial graphics data,as well as techniques of compressing and encrypting the watermarked ...The paper presents a set of techniques of digital watermarking by which copyright and user rights messages are hidden into geo-spatial graphics data,as well as techniques of compressing and encrypting the watermarked geo-spatial graphics data.The technology aims at tracing and resisting the illegal distribution and duplication of the geo-spatial graphics data product,so as to effectively protect the data producer's rights as well as to facilitate the secure sharing of geo-spatial graphics data.So far in the CIS field throughout the world,few researches have been made on digital watermarking.The research is a novel exploration both in the field of security management of geo-spatial graphics data and in the applications of digital watermarking technique.An application software employing the proposed technology has been developed.A number of experimental tests on the 1:500,000 digital bathymetric chart of the South China Sea and 1:10,000 digital topographic map of Jiangsu Province have been conducted to verify the feasibility of the proposed technology.展开更多
There are multiple operating modes in the real industrial process, and the collected data follow the complex multimodal distribution, so most traditional process monitoring methods are no longer applicable because the...There are multiple operating modes in the real industrial process, and the collected data follow the complex multimodal distribution, so most traditional process monitoring methods are no longer applicable because their presumptions are that sampled-data should obey the single Gaussian distribution or non-Gaussian distribution. In order to solve these problems, a novel weighted local standardization(WLS) strategy is proposed to standardize the multimodal data, which can eliminate the multi-mode characteristics of the collected data, and normalize them into unimodal data distribution. After detailed analysis of the raised data preprocessing strategy, a new algorithm using WLS strategy with support vector data description(SVDD) is put forward to apply for multi-mode monitoring process. Unlike the strategy of building multiple local models, the developed method only contains a model without the prior knowledge of multi-mode process. To demonstrate the proposed method's validity, it is applied to a numerical example and a Tennessee Eastman(TE) process. Finally, the simulation results show that the WLS strategy is very effective to standardize multimodal data, and the WLS-SVDD monitoring method has great advantages over the traditional SVDD and PCA combined with a local standardization strategy(LNS-PCA) in multi-mode process monitoring.展开更多
According to the chaotic and non-linear characters of power load data,the time series matrix is established with the theory of phase-space reconstruction,and then Lyapunov exponents with chaotic time series are comput...According to the chaotic and non-linear characters of power load data,the time series matrix is established with the theory of phase-space reconstruction,and then Lyapunov exponents with chaotic time series are computed to determine the time delay and the embedding dimension.Due to different features of the data,data mining algorithm is conducted to classify the data into different groups.Redundant information is eliminated by the advantage of data mining technology,and the historical loads that have highly similar features with the forecasting day are searched by the system.As a result,the training data can be decreased and the computing speed can also be improved when constructing support vector machine(SVM) model.Then,SVM algorithm is used to predict power load with parameters that get in pretreatment.In order to prove the effectiveness of the new model,the calculation with data mining SVM algorithm is compared with that of single SVM and back propagation network.It can be seen that the new DSVM algorithm effectively improves the forecast accuracy by 0.75%,1.10% and 1.73% compared with SVM for two random dimensions of 11-dimension,14-dimension and BP network,respectively.This indicates that the DSVM gains perfect improvement effect in the short-term power load forecasting.展开更多
High compression ratio,high decoding performance,and progressive data transmission are the most important require-ments of vector data compression algorithms for WebGIS.To meet these requirements,we present a new comp...High compression ratio,high decoding performance,and progressive data transmission are the most important require-ments of vector data compression algorithms for WebGIS.To meet these requirements,we present a new compression approach.This paper begins with the generation of multiscale data by converting float coordinates to integer coordinates.It is proved that the distance between the converted point and the original point on screen is within 2 pixels,and therefore,our approach is suitable for the visualization of vector data on the client side.Integer coordinates are passed to an Integer Wavelet Transformer,and the high-frequency coefficients produced by the transformer are encoded by Canonical Huffman codes.The experimental results on river data and road data demonstrate the effectiveness of the proposed approach:compression ratio can reach 10% for river data and 20% for road data,respectively.We conclude that more attention needs be paid to correlation between curves that contain a few points.展开更多
The high observation efficiency,scanning speed and observation frequency of the Fengyun-4A(FY-4A)satellite indicates the progress of Chinese geostationary meteorological satellites.The characteristics of FY-4A atmosph...The high observation efficiency,scanning speed and observation frequency of the Fengyun-4A(FY-4A)satellite indicates the progress of Chinese geostationary meteorological satellites.The characteristics of FY-4A atmospheric motion vectors(AMVs)derived from the high-level water vapor(WV-High)channel,mid-level water vapor(WV-Mid)channel,and infrared(IR)channel of FY-4A are analyzed,and their corresponding observation errors estimated.Then,the impacts of single-channel and multi-channel FY-4A AMVs on RMAPS-ST(the Rapid-refresh Multi-scale Analysis and Prediction System-Short Term)are evaluated based on one-month data assimilation cycling and forecasting experiments.Results show that the observation errors of FY-4A AMVs from the three channels have an explicit vertical structure.Results from the cycling experiments indicate that the assimilation of AMVs from WV-High produces more apparent improvement of the wind in the upper layer,while a more positive effect in the lower layer is achieved by the assimilation of AMVs from IR.Furthermore,the assimilation of AMVs from IR is more skillful for medium and moderate precipitation than from other channels owing to the good quality of data in the lower layer in the AMVs from IR.Assimilation of FY-4A AMVs from the three channels could combine the advantages of assimilation from each individual channel to improve the wind in the upper,middle and lower layers simultaneously.展开更多
Workers’exposure to excessive noise is a big universal work-related challenges.One of the major consequences of exposure to noise is permanent or transient hearing loss.The current study sought to utilize audiometric...Workers’exposure to excessive noise is a big universal work-related challenges.One of the major consequences of exposure to noise is permanent or transient hearing loss.The current study sought to utilize audiometric data to weigh and prioritize the factors affecting workers’hearing loss based using the Support Vector Machine(SVM)algorithm.This cross sectional-descriptive study was conducted in 2017 in a mining industry in southeast Iran.The participating workers(n=150)were divided into three groups of 50 based on the sound pressure level to which they were exposed(two experimental groups and one control group).Audiometric tests were carried out for all members of each group.The study generally entailed the following steps:(1)selecting predicting variables to weigh and prioritize factors affecting hearing loss;(2)conducting audiometric tests and assessing permanent hearing loss in each ear and then evaluating total hearing loss;(3)categorizing different types of hearing loss;(4)weighing and prioritizing factors that affect hearing loss based on the SVM algorithm;and(5)assessing the error rate and accuracy of the models.The collected data were fed into SPSS 18,followed by conducting linear regression and paired samples t-test.It was revealed that,in the first model(SPL<70 dBA),the frequency of 8 KHz had the greatest impact(with a weight of 33%),while noise had the smallest influence(with a weight of 5%).The accuracy of this model was 100%.In the second model(70<SPL<80 dBA),the frequency of 4 KHz had the most profound effect(with a weight of 21%),whereas the frequency of 250 Hz had the lowest impact(with a weight of 6%).The accuracy of this model was 100%too.In the third model(SPL>85 dBA),the frequency of 4 KHz had the highest impact(with a weight of 22%),while the frequency of 250 Hz had the smallest influence(with a weight of 3%).The accuracy of this model was 100%too.In the fourth model,the frequency of 4 KHz had the greatest effect(with a weight of 24%),while the frequency of 500 Hz had the smallest effect(with a weight of 4%).The accuracy of this model was found to be 94%.According to the modeling conducted using the SVM algorithm,the frequency of 4 KHz has the most profound effect on predicting changes in hearing loss.Given the high accuracy of the obtained model,this algorithm is an appropriate and powerful tool to predict and model hearing loss.展开更多
Hybrid data assimilation (DA) is a method seeing more use in recent hydrology and water resources research. In this study, a DA method coupled with the support vector machines (SVMs) and the ensemble Kalman filter...Hybrid data assimilation (DA) is a method seeing more use in recent hydrology and water resources research. In this study, a DA method coupled with the support vector machines (SVMs) and the ensemble Kalman filter (EnKF) technology was used for the prediction of soil moisture in different soil layers: 0-5 cm, 30 cm, 50 cm, 100 cm, 200 cm, and 300 cm. The SVM methodology was first used to train the ground measurements of soil moisture and meteorological parameters from the Meilin study area, in East China, to construct soil moisture statistical prediction models. Subsequent observations and their statistics were used for predictions, with two approaches: the SVM predictor and the SVM-EnKF model made by coupling the SVM model with the EnKF technique using the DA method. Validation results showed that the proposed SVM-EnKF model can improve the prediction results of soil moisture in different layers, from the surface to the root zone.展开更多
Due to the conflict between huge amount of map data and limited network bandwidth, rapid trans- mission of vector map data over the Internet has become a bottleneck of spatial data delivery in web-based environment. T...Due to the conflict between huge amount of map data and limited network bandwidth, rapid trans- mission of vector map data over the Internet has become a bottleneck of spatial data delivery in web-based environment. This paper proposed an approach to organizing and transmitting multi-scale vector river network data via the Internet progressively. This approach takes account of two levels of importance, i.e. the importance of river branches and the importance of the points belonging to each river branch, and forms data packages ac- cording to these. Our experiments have shown that the proposed approach can reduce 90% of original data while preserving the river structure well.展开更多
Radiometric normalization,as an essential step for multi-source and multi-temporal data processing,has received critical attention.Relative Radiometric Normalization(RRN)method has been primarily used for eliminating ...Radiometric normalization,as an essential step for multi-source and multi-temporal data processing,has received critical attention.Relative Radiometric Normalization(RRN)method has been primarily used for eliminating the radiometric inconsistency.The radiometric trans-forming relation between the subject image and the reference image is an essential aspect of RRN.Aimed at accurate radiometric transforming relation modeling,the learning-based nonlinear regression method,Support Vector machine Regression(SVR)is used for fitting the complicated radiometric transforming relation for the coarse-resolution data-referenced RRN.To evaluate the effectiveness of the proposed method,a series of experiments are performed,including two synthetic data experiments and one real data experiment.And the proposed method is compared with other methods that use linear regression,Artificial Neural Network(ANN)or Random Forest(RF)for radiometric transforming relation modeling.The results show that the proposed method performs well on fitting the radiometric transforming relation and could enhance the RRN performance.展开更多
Multi-source multi-class classification methods based on multi-class Support Vector Machines and data fusion strategies are proposed in this paper. The centralized and distributed fusion schemes are applied to combine...Multi-source multi-class classification methods based on multi-class Support Vector Machines and data fusion strategies are proposed in this paper. The centralized and distributed fusion schemes are applied to combine information from several data sources. In the centralized scheme, all information from several data sources is centralized to construct an input space. Then a multi-class Support Vector Machine classifier is trained. In the distributed schemes, the individual data sources are proc-essed separately and modelled by using the multi-class Support Vector Machine. Then new data fusion strategies are proposed to combine the information from the individual multi-class Support Vector Machine models. Our proposed fusion strategies take into account that an Support Vector Machine (SVM) classifier achieves classification by finding the optimal classification hyperplane with maximal margin. The proposed methods are applied for fault diagnosis of a diesel engine. The experimental results showed that almost all the proposed approaches can largely improve the diagnostic accuracy. The robustness of diagnosis is also improved because of the implementation of data fusion strategies. The proposed methods can also be applied in other fields.展开更多
Parallel vector buffer analysis approaches can be classified into 2 types:algorithm-oriented parallel strategy and the data-oriented parallel strategy.These methods do not take its applicability on the existing geogra...Parallel vector buffer analysis approaches can be classified into 2 types:algorithm-oriented parallel strategy and the data-oriented parallel strategy.These methods do not take its applicability on the existing geographic information systems(GIS)platforms into consideration.In order to address the problem,a spatial decomposition approach for accelerating buffer analysis of vector data is proposed.The relationship between the number of vertices of each feature and the buffer analysis computing time is analyzed to generate computational intensity transformation functions(CITFs).Then,computational intensity grids(CIGs)of polyline and polygon are constructed based on the relative CITFs.Using the corresponding CIGs,a spatial decomposition method for parallel buffer analysis is developed.Based on the computational intensity of the features and the sub-domains generated in the decomposition,the features are averagely assigned within the sub-domains into parallel buffer analysis tasks for load balance.Compared with typical regular domain decomposition methods,the new approach accomplishes greater balanced decomposition of computational intensity for parallel buffer analysis and achieves near-linear speedups.展开更多
Geophysical data sets are growing at an ever-increasing rate,requiring computationally efficient data selection (thinning) methods to preserve essential information.Satellites,such as WindSat,provide large data sets...Geophysical data sets are growing at an ever-increasing rate,requiring computationally efficient data selection (thinning) methods to preserve essential information.Satellites,such as WindSat,provide large data sets for assessing the accuracy and computational efficiency of data selection techniques.A new data thinning technique,based on support vector regression (SVR),is developed and tested.To manage large on-line satellite data streams,observations from WindSat are formed into subsets by Voronoi tessellation and then each is thinned by SVR (TSVR).Three experiments are performed.The first confirms the viability of TSVR for a relatively small sample,comparing it to several commonly used data thinning methods (random selection,averaging and Barnes filtering),producing a 10% thinning rate (90% data reduction),low mean absolute errors (MAE) and large correlations with the original data.A second experiment,using a larger dataset,shows TSVR retrievals with MAE < 1 m s-1 and correlations ≥ 0.98.TSVR was an order of magnitude faster than the commonly used thinning methods.A third experiment applies a two-stage pipeline to TSVR,to accommodate online data.The pipeline subsets reconstruct the wind field with the same accuracy as the second experiment,is an order of magnitude faster than the nonpipeline TSVR.Therefore,pipeline TSVR is two orders of magnitude faster than commonly used thinning methods that ingest the entire data set.This study demonstrates that TSVR pipeline thinning is an accurate and computationally efficient alternative to commonly used data selection techniques.展开更多
The geological data are constructed in vector format in geographical information system (GIS) while other data such as remote sensing images, geographical data and geochemical data are saved in raster ones. This paper...The geological data are constructed in vector format in geographical information system (GIS) while other data such as remote sensing images, geographical data and geochemical data are saved in raster ones. This paper converts the vector data into 8 bit images according to their importance to mineralization each by programming. We can communicate the geological meaning with the raster images by this method. The paper also fuses geographical data and geochemical data with the programmed strata data. The result shows that image fusion can express different intensities effectively and visualize the structure characters in 2 dimensions. Furthermore, it also can produce optimized information from multi-source data and express them more directly.展开更多
Multistage Vector Quantization(MSVQ) can achieve very low encoding and storage complexity in comparison to unstructured vector quantization. However, the conventional MSVQ is suboptimal with respect to the overall per...Multistage Vector Quantization(MSVQ) can achieve very low encoding and storage complexity in comparison to unstructured vector quantization. However, the conventional MSVQ is suboptimal with respect to the overall performance measure. This paper proposes a new technology to design the decoder codebook, which is different from the encoder codebook to optimise the overall performance. The performance improvement is achieved with no effect on encoding complexity, both storage and time consuming, but a modest increase in storage complexity of decoder.展开更多
With the rapid growth of the Internet,the copyright protection problem occurs frequently,and unauthorized copying and distributing of geospatial data threaten the investments of data producers.Digital watermarking is ...With the rapid growth of the Internet,the copyright protection problem occurs frequently,and unauthorized copying and distributing of geospatial data threaten the investments of data producers.Digital watermarking is a possible solution to solve this issue.However,watermarking causes modifications in the original data resulting in distortion and affects accuracy,which is very important to geospatial vector data.This article provides distortion assessment of watermarked geospatial data using wavelet-based invisible watermarking.Eight wavelets at different wavelet decomposition levels are used for accuracy evaluation with the help of error measures such as maximum error and mean square error.Normalized correlation is used as a similarity index between original and extracted watermark.It is observed that the increase in the strength of embedding increases visual degradation.Haar wavelet outperforms the other wavelets,and the third wavelet decomposition level is proved to be optimal level for watermarking.展开更多
基金supported by the Deanship of Graduate Studies and Scientific Research at University of Bisha for funding this research through the promising program under grant number(UB-Promising-33-1445).
文摘Open networks and heterogeneous services in the Internet of Vehicles(IoV)can lead to security and privacy challenges.One key requirement for such systems is the preservation of user privacy,ensuring a seamless experience in driving,navigation,and communication.These privacy needs are influenced by various factors,such as data collected at different intervals,trip durations,and user interactions.To address this,the paper proposes a Support Vector Machine(SVM)model designed to process large amounts of aggregated data and recommend privacy preserving measures.The model analyzes data based on user demands and interactions with service providers or neighboring infrastructure.It aims to minimize privacy risks while ensuring service continuity and sustainability.The SVMmodel helps validate the system’s reliability by creating a hyperplane that distinguishes between maximum and minimum privacy recommendations.The results demonstrate the effectiveness of the proposed SVM model in enhancing both privacy and service performance.
文摘Miniature air quality sensors are widely used in urban grid-based monitoring due to their flexibility in deployment and low cost.However,the raw data collected by these devices often suffer from low accuracy caused by environmental interference and sensor drift,highlighting the need for effective calibration methods to improve data reliability.This study proposes a data correction method based on Bayesian Optimization Support Vector Regression(BO-SVR),which combines the nonlinear modeling capability of Support Vector Regression(SVR)with the efficient global hyperparameter search of Bayesian Optimization.By introducing cross-validation loss as the optimization objective and using Gaussian process modeling with an Expected Improvement acquisition strategy,the approach automatically determines optimal hyperparameters for accurate pollutant concentration prediction.Experiments on real-world micro-sensor datasets demonstrate that BO-SVR outperforms traditional SVR,grid search SVR,and random forest(RF)models across multiple pollutants,including PM_(2.5),PM_(10),CO,NO_(2),SO_(2),and O_(3).The proposed method achieves lower prediction residuals,higher fitting accuracy,and better generalization,offering an efficient and practical solution for enhancing the quality of micro-sensor air monitoring data.
基金partially supported by the National Key Research and Development Project under Grant2020YFB1806805Social Development Projects of Jiangsu Science and Technology Department under Grant No.BE2018704
文摘In wireless communication,the problem of authenticating the transmitter’s identity is challeng-ing,especially for those terminal devices in which the security schemes based on cryptography are approxi-mately unfeasible owing to limited resources.In this paper,a physical layer authentication scheme is pro-posed to detect whether there is anomalous access by the attackers disguised as legitimate users.Explicitly,channel state information(CSI)is used as a form of fingerprint to exploit spatial discrimination among de-vices in the wireless network and machine learning(ML)technology is employed to promote the improve-ment of authentication accuracy.Considering that the falsified messages are not accessible for authenticator during the training phase,deep support vector data de-scription(Deep SVDD)is selected to solve the one-class classification(OCC)problem.Simulation results show that Deep SVDD based scheme can tackle the challenges of physical layer authentication in wireless communication environments.
基金supported by the National Key R&D Program of China(Grant No.2022YFC3005401)the Fundamental Research Funds for the Central Universities(Grant No.B230201013)+2 种基金the National Natural Science Foundation of China(Grants No.52309152,U2243223,and U23B20150)the Natural Science Foundation of Jiangsu Province(Grant No.BK20220978)the Open Fund of National Dam Safety Research Center(Grant No.CX2023B03).
文摘Deformation monitoring is a critical measure for intuitively reflecting the operational behavior of a dam.However,the deformation monitoring data are often incomplete due to environmental changes,monitoring instrument faults,and human operational errors,thereby often hindering the accurate assessment of actual deformation patterns.This study proposed a method for quantifying deformation similarity between measurement points by recognizing the spatiotemporal characteristics of concrete dam deformation monitoring data.It introduces a spatiotemporal clustering analysis of the concrete dam deformation behavior and employs the support vector machine model to address the missing data in concrete dam deformation monitoring.The proposed method was validated in a concrete dam project,with the model error maintaining within 5%,demonstrating its effectiveness in processing missing deformation data.This approach enhances the capability of early-warning systems and contributes to enhanced dam safety management.
基金Under the auspices of Jiangsu Provincial Science and Technology Fundation of Surveying and Mapping (No. 200416 )
文摘The paper presents a set of techniques of digital watermarking by which copyright and user rights messages are hidden into geo-spatial graphics data,as well as techniques of compressing and encrypting the watermarked geo-spatial graphics data.The technology aims at tracing and resisting the illegal distribution and duplication of the geo-spatial graphics data product,so as to effectively protect the data producer's rights as well as to facilitate the secure sharing of geo-spatial graphics data.So far in the CIS field throughout the world,few researches have been made on digital watermarking.The research is a novel exploration both in the field of security management of geo-spatial graphics data and in the applications of digital watermarking technique.An application software employing the proposed technology has been developed.A number of experimental tests on the 1:500,000 digital bathymetric chart of the South China Sea and 1:10,000 digital topographic map of Jiangsu Province have been conducted to verify the feasibility of the proposed technology.
基金Project(61374140)supported by the National Natural Science Foundation of China
文摘There are multiple operating modes in the real industrial process, and the collected data follow the complex multimodal distribution, so most traditional process monitoring methods are no longer applicable because their presumptions are that sampled-data should obey the single Gaussian distribution or non-Gaussian distribution. In order to solve these problems, a novel weighted local standardization(WLS) strategy is proposed to standardize the multimodal data, which can eliminate the multi-mode characteristics of the collected data, and normalize them into unimodal data distribution. After detailed analysis of the raised data preprocessing strategy, a new algorithm using WLS strategy with support vector data description(SVDD) is put forward to apply for multi-mode monitoring process. Unlike the strategy of building multiple local models, the developed method only contains a model without the prior knowledge of multi-mode process. To demonstrate the proposed method's validity, it is applied to a numerical example and a Tennessee Eastman(TE) process. Finally, the simulation results show that the WLS strategy is very effective to standardize multimodal data, and the WLS-SVDD monitoring method has great advantages over the traditional SVDD and PCA combined with a local standardization strategy(LNS-PCA) in multi-mode process monitoring.
基金Project(70671039) supported by the National Natural Science Foundation of China
文摘According to the chaotic and non-linear characters of power load data,the time series matrix is established with the theory of phase-space reconstruction,and then Lyapunov exponents with chaotic time series are computed to determine the time delay and the embedding dimension.Due to different features of the data,data mining algorithm is conducted to classify the data into different groups.Redundant information is eliminated by the advantage of data mining technology,and the historical loads that have highly similar features with the forecasting day are searched by the system.As a result,the training data can be decreased and the computing speed can also be improved when constructing support vector machine(SVM) model.Then,SVM algorithm is used to predict power load with parameters that get in pretreatment.In order to prove the effectiveness of the new model,the calculation with data mining SVM algorithm is compared with that of single SVM and back propagation network.It can be seen that the new DSVM algorithm effectively improves the forecast accuracy by 0.75%,1.10% and 1.73% compared with SVM for two random dimensions of 11-dimension,14-dimension and BP network,respectively.This indicates that the DSVM gains perfect improvement effect in the short-term power load forecasting.
基金Supported by the National High-tech R&D Program of China(NO.2007AA120501)
文摘High compression ratio,high decoding performance,and progressive data transmission are the most important require-ments of vector data compression algorithms for WebGIS.To meet these requirements,we present a new compression approach.This paper begins with the generation of multiscale data by converting float coordinates to integer coordinates.It is proved that the distance between the converted point and the original point on screen is within 2 pixels,and therefore,our approach is suitable for the visualization of vector data on the client side.Integer coordinates are passed to an Integer Wavelet Transformer,and the high-frequency coefficients produced by the transformer are encoded by Canonical Huffman codes.The experimental results on river data and road data demonstrate the effectiveness of the proposed approach:compression ratio can reach 10% for river data and 20% for road data,respectively.We conclude that more attention needs be paid to correlation between curves that contain a few points.
基金the National Key Research and Development Plan(Grant No.2018YFC1507105).
文摘The high observation efficiency,scanning speed and observation frequency of the Fengyun-4A(FY-4A)satellite indicates the progress of Chinese geostationary meteorological satellites.The characteristics of FY-4A atmospheric motion vectors(AMVs)derived from the high-level water vapor(WV-High)channel,mid-level water vapor(WV-Mid)channel,and infrared(IR)channel of FY-4A are analyzed,and their corresponding observation errors estimated.Then,the impacts of single-channel and multi-channel FY-4A AMVs on RMAPS-ST(the Rapid-refresh Multi-scale Analysis and Prediction System-Short Term)are evaluated based on one-month data assimilation cycling and forecasting experiments.Results show that the observation errors of FY-4A AMVs from the three channels have an explicit vertical structure.Results from the cycling experiments indicate that the assimilation of AMVs from WV-High produces more apparent improvement of the wind in the upper layer,while a more positive effect in the lower layer is achieved by the assimilation of AMVs from IR.Furthermore,the assimilation of AMVs from IR is more skillful for medium and moderate precipitation than from other channels owing to the good quality of data in the lower layer in the AMVs from IR.Assimilation of FY-4A AMVs from the three channels could combine the advantages of assimilation from each individual channel to improve the wind in the upper,middle and lower layers simultaneously.
基金This study stemmed from a research project(code number:96000838)which was sponsored by the Institute for Futures Studies in Health at Kerman University of Medical Sciences.
文摘Workers’exposure to excessive noise is a big universal work-related challenges.One of the major consequences of exposure to noise is permanent or transient hearing loss.The current study sought to utilize audiometric data to weigh and prioritize the factors affecting workers’hearing loss based using the Support Vector Machine(SVM)algorithm.This cross sectional-descriptive study was conducted in 2017 in a mining industry in southeast Iran.The participating workers(n=150)were divided into three groups of 50 based on the sound pressure level to which they were exposed(two experimental groups and one control group).Audiometric tests were carried out for all members of each group.The study generally entailed the following steps:(1)selecting predicting variables to weigh and prioritize factors affecting hearing loss;(2)conducting audiometric tests and assessing permanent hearing loss in each ear and then evaluating total hearing loss;(3)categorizing different types of hearing loss;(4)weighing and prioritizing factors that affect hearing loss based on the SVM algorithm;and(5)assessing the error rate and accuracy of the models.The collected data were fed into SPSS 18,followed by conducting linear regression and paired samples t-test.It was revealed that,in the first model(SPL<70 dBA),the frequency of 8 KHz had the greatest impact(with a weight of 33%),while noise had the smallest influence(with a weight of 5%).The accuracy of this model was 100%.In the second model(70<SPL<80 dBA),the frequency of 4 KHz had the most profound effect(with a weight of 21%),whereas the frequency of 250 Hz had the lowest impact(with a weight of 6%).The accuracy of this model was 100%too.In the third model(SPL>85 dBA),the frequency of 4 KHz had the highest impact(with a weight of 22%),while the frequency of 250 Hz had the smallest influence(with a weight of 3%).The accuracy of this model was 100%too.In the fourth model,the frequency of 4 KHz had the greatest effect(with a weight of 24%),while the frequency of 500 Hz had the smallest effect(with a weight of 4%).The accuracy of this model was found to be 94%.According to the modeling conducted using the SVM algorithm,the frequency of 4 KHz has the most profound effect on predicting changes in hearing loss.Given the high accuracy of the obtained model,this algorithm is an appropriate and powerful tool to predict and model hearing loss.
基金supported by the National Basic Research Program of China (the 973 Program,Grant No.2010CB951101)the Program for Changjiang Scholars and Innovative Research Teams in Universities,the Ministry of Education,China (Grant No. IRT0717)
文摘Hybrid data assimilation (DA) is a method seeing more use in recent hydrology and water resources research. In this study, a DA method coupled with the support vector machines (SVMs) and the ensemble Kalman filter (EnKF) technology was used for the prediction of soil moisture in different soil layers: 0-5 cm, 30 cm, 50 cm, 100 cm, 200 cm, and 300 cm. The SVM methodology was first used to train the ground measurements of soil moisture and meteorological parameters from the Meilin study area, in East China, to construct soil moisture statistical prediction models. Subsequent observations and their statistics were used for predictions, with two approaches: the SVM predictor and the SVM-EnKF model made by coupling the SVM model with the EnKF technique using the DA method. Validation results showed that the proposed SVM-EnKF model can improve the prediction results of soil moisture in different layers, from the surface to the root zone.
文摘Due to the conflict between huge amount of map data and limited network bandwidth, rapid trans- mission of vector map data over the Internet has become a bottleneck of spatial data delivery in web-based environment. This paper proposed an approach to organizing and transmitting multi-scale vector river network data via the Internet progressively. This approach takes account of two levels of importance, i.e. the importance of river branches and the importance of the points belonging to each river branch, and forms data packages ac- cording to these. Our experiments have shown that the proposed approach can reduce 90% of original data while preserving the river structure well.
基金This research was funded by the National Natural Science Fund of China[grant number 41701415]Science fund project of Wuhan Institute of Technology[grant number K201724]Science and Technology Development Funds Project of Department of Transportation of Hubei Province[grant number 201900001].
文摘Radiometric normalization,as an essential step for multi-source and multi-temporal data processing,has received critical attention.Relative Radiometric Normalization(RRN)method has been primarily used for eliminating the radiometric inconsistency.The radiometric trans-forming relation between the subject image and the reference image is an essential aspect of RRN.Aimed at accurate radiometric transforming relation modeling,the learning-based nonlinear regression method,Support Vector machine Regression(SVR)is used for fitting the complicated radiometric transforming relation for the coarse-resolution data-referenced RRN.To evaluate the effectiveness of the proposed method,a series of experiments are performed,including two synthetic data experiments and one real data experiment.And the proposed method is compared with other methods that use linear regression,Artificial Neural Network(ANN)or Random Forest(RF)for radiometric transforming relation modeling.The results show that the proposed method performs well on fitting the radiometric transforming relation and could enhance the RRN performance.
文摘Multi-source multi-class classification methods based on multi-class Support Vector Machines and data fusion strategies are proposed in this paper. The centralized and distributed fusion schemes are applied to combine information from several data sources. In the centralized scheme, all information from several data sources is centralized to construct an input space. Then a multi-class Support Vector Machine classifier is trained. In the distributed schemes, the individual data sources are proc-essed separately and modelled by using the multi-class Support Vector Machine. Then new data fusion strategies are proposed to combine the information from the individual multi-class Support Vector Machine models. Our proposed fusion strategies take into account that an Support Vector Machine (SVM) classifier achieves classification by finding the optimal classification hyperplane with maximal margin. The proposed methods are applied for fault diagnosis of a diesel engine. The experimental results showed that almost all the proposed approaches can largely improve the diagnostic accuracy. The robustness of diagnosis is also improved because of the implementation of data fusion strategies. The proposed methods can also be applied in other fields.
基金the National Natural Science Foundation of China(No.41971356,41701446)National Key Research and Development Program of China(No.2017YFB0503600,2018YFB0505500,2017YFC0602204).
文摘Parallel vector buffer analysis approaches can be classified into 2 types:algorithm-oriented parallel strategy and the data-oriented parallel strategy.These methods do not take its applicability on the existing geographic information systems(GIS)platforms into consideration.In order to address the problem,a spatial decomposition approach for accelerating buffer analysis of vector data is proposed.The relationship between the number of vertices of each feature and the buffer analysis computing time is analyzed to generate computational intensity transformation functions(CITFs).Then,computational intensity grids(CIGs)of polyline and polygon are constructed based on the relative CITFs.Using the corresponding CIGs,a spatial decomposition method for parallel buffer analysis is developed.Based on the computational intensity of the features and the sub-domains generated in the decomposition,the features are averagely assigned within the sub-domains into parallel buffer analysis tasks for load balance.Compared with typical regular domain decomposition methods,the new approach accomplishes greater balanced decomposition of computational intensity for parallel buffer analysis and achieves near-linear speedups.
基金Supported by National High Technology Research and Development Program of China (863 Program) (2006AA040308), National Natural Science Foundation of China (60736021), and the National Creative Research Groups Science Foundation of China (60721062)
基金NOAA Grant NA17RJ1227 and NSF Grant EIA-0205628 for providing financial support for this worksupported by RSF Grant 14-41-00039
文摘Geophysical data sets are growing at an ever-increasing rate,requiring computationally efficient data selection (thinning) methods to preserve essential information.Satellites,such as WindSat,provide large data sets for assessing the accuracy and computational efficiency of data selection techniques.A new data thinning technique,based on support vector regression (SVR),is developed and tested.To manage large on-line satellite data streams,observations from WindSat are formed into subsets by Voronoi tessellation and then each is thinned by SVR (TSVR).Three experiments are performed.The first confirms the viability of TSVR for a relatively small sample,comparing it to several commonly used data thinning methods (random selection,averaging and Barnes filtering),producing a 10% thinning rate (90% data reduction),low mean absolute errors (MAE) and large correlations with the original data.A second experiment,using a larger dataset,shows TSVR retrievals with MAE < 1 m s-1 and correlations ≥ 0.98.TSVR was an order of magnitude faster than the commonly used thinning methods.A third experiment applies a two-stage pipeline to TSVR,to accommodate online data.The pipeline subsets reconstruct the wind field with the same accuracy as the second experiment,is an order of magnitude faster than the nonpipeline TSVR.Therefore,pipeline TSVR is two orders of magnitude faster than commonly used thinning methods that ingest the entire data set.This study demonstrates that TSVR pipeline thinning is an accurate and computationally efficient alternative to commonly used data selection techniques.
文摘The geological data are constructed in vector format in geographical information system (GIS) while other data such as remote sensing images, geographical data and geochemical data are saved in raster ones. This paper converts the vector data into 8 bit images according to their importance to mineralization each by programming. We can communicate the geological meaning with the raster images by this method. The paper also fuses geographical data and geochemical data with the programmed strata data. The result shows that image fusion can express different intensities effectively and visualize the structure characters in 2 dimensions. Furthermore, it also can produce optimized information from multi-source data and express them more directly.
文摘Multistage Vector Quantization(MSVQ) can achieve very low encoding and storage complexity in comparison to unstructured vector quantization. However, the conventional MSVQ is suboptimal with respect to the overall performance measure. This paper proposes a new technology to design the decoder codebook, which is different from the encoder codebook to optimise the overall performance. The performance improvement is achieved with no effect on encoding complexity, both storage and time consuming, but a modest increase in storage complexity of decoder.
文摘With the rapid growth of the Internet,the copyright protection problem occurs frequently,and unauthorized copying and distributing of geospatial data threaten the investments of data producers.Digital watermarking is a possible solution to solve this issue.However,watermarking causes modifications in the original data resulting in distortion and affects accuracy,which is very important to geospatial vector data.This article provides distortion assessment of watermarked geospatial data using wavelet-based invisible watermarking.Eight wavelets at different wavelet decomposition levels are used for accuracy evaluation with the help of error measures such as maximum error and mean square error.Normalized correlation is used as a similarity index between original and extracted watermark.It is observed that the increase in the strength of embedding increases visual degradation.Haar wavelet outperforms the other wavelets,and the third wavelet decomposition level is proved to be optimal level for watermarking.