In wireless communication,the problem of authenticating the transmitter’s identity is challeng-ing,especially for those terminal devices in which the security schemes based on cryptography are approxi-mately unfeasib...In wireless communication,the problem of authenticating the transmitter’s identity is challeng-ing,especially for those terminal devices in which the security schemes based on cryptography are approxi-mately unfeasible owing to limited resources.In this paper,a physical layer authentication scheme is pro-posed to detect whether there is anomalous access by the attackers disguised as legitimate users.Explicitly,channel state information(CSI)is used as a form of fingerprint to exploit spatial discrimination among de-vices in the wireless network and machine learning(ML)technology is employed to promote the improve-ment of authentication accuracy.Considering that the falsified messages are not accessible for authenticator during the training phase,deep support vector data de-scription(Deep SVDD)is selected to solve the one-class classification(OCC)problem.Simulation results show that Deep SVDD based scheme can tackle the challenges of physical layer authentication in wireless communication environments.展开更多
There are multiple operating modes in the real industrial process, and the collected data follow the complex multimodal distribution, so most traditional process monitoring methods are no longer applicable because the...There are multiple operating modes in the real industrial process, and the collected data follow the complex multimodal distribution, so most traditional process monitoring methods are no longer applicable because their presumptions are that sampled-data should obey the single Gaussian distribution or non-Gaussian distribution. In order to solve these problems, a novel weighted local standardization(WLS) strategy is proposed to standardize the multimodal data, which can eliminate the multi-mode characteristics of the collected data, and normalize them into unimodal data distribution. After detailed analysis of the raised data preprocessing strategy, a new algorithm using WLS strategy with support vector data description(SVDD) is put forward to apply for multi-mode monitoring process. Unlike the strategy of building multiple local models, the developed method only contains a model without the prior knowledge of multi-mode process. To demonstrate the proposed method's validity, it is applied to a numerical example and a Tennessee Eastman(TE) process. Finally, the simulation results show that the WLS strategy is very effective to standardize multimodal data, and the WLS-SVDD monitoring method has great advantages over the traditional SVDD and PCA combined with a local standardization strategy(LNS-PCA) in multi-mode process monitoring.展开更多
Hybrid data assimilation (DA) is a method seeing more use in recent hydrology and water resources research. In this study, a DA method coupled with the support vector machines (SVMs) and the ensemble Kalman filter...Hybrid data assimilation (DA) is a method seeing more use in recent hydrology and water resources research. In this study, a DA method coupled with the support vector machines (SVMs) and the ensemble Kalman filter (EnKF) technology was used for the prediction of soil moisture in different soil layers: 0-5 cm, 30 cm, 50 cm, 100 cm, 200 cm, and 300 cm. The SVM methodology was first used to train the ground measurements of soil moisture and meteorological parameters from the Meilin study area, in East China, to construct soil moisture statistical prediction models. Subsequent observations and their statistics were used for predictions, with two approaches: the SVM predictor and the SVM-EnKF model made by coupling the SVM model with the EnKF technique using the DA method. Validation results showed that the proposed SVM-EnKF model can improve the prediction results of soil moisture in different layers, from the surface to the root zone.展开更多
An efficient trust-aware secure routing and network strategy-based data collection scheme is presented in this paper to enhance the performance and security of wireless sensor networks during data collection.The metho...An efficient trust-aware secure routing and network strategy-based data collection scheme is presented in this paper to enhance the performance and security of wireless sensor networks during data collection.The method first discovers the routes between the data sensors and the sink node.Several factors are considered for each sensor node along the route,including energy,number of neighbours,previous transmissions,and energy depletion ratio.Considering all these variables,the Sink Reachable Support Measure and the Secure Communication Support Measure,the method evaluates two distinct measures.The method calculates the data carrier support value using these two metrics.A single route is chosen to collect data based on the value of data carrier support.It has contributed to the design of Secure Communication Support(SCS)Estimation.This has been measured according to the strategy of each hop of the route.The suggested method improves the security and efficacy of data collection in wireless sensor networks.The second stage uses the two-fish approach to build a trust model for secure data transfer.A sim-ulation exercise was conducted to evaluate the effectiveness of the suggested framework.Metrics,including PDR,end-to-end latency,and average residual energy,were assessed for the proposed model.The efficiency of the suggested route design serves as evidence for the average residual energy for the proposed framework.展开更多
In recent years,deep learning-based signal recognition technology has gained attention and emerged as an important approach for safeguarding the electromagnetic environment.However,training deep learning-based classif...In recent years,deep learning-based signal recognition technology has gained attention and emerged as an important approach for safeguarding the electromagnetic environment.However,training deep learning-based classifiers on large signal datasets with redundant samples requires significant memory and high costs.This paper proposes a support databased core-set selection method(SD)for signal recognition,aiming to screen a representative subset that approximates the large signal dataset.Specifically,this subset can be identified by employing the labeled information during the early stages of model training,as some training samples are labeled as supporting data frequently.This support data is crucial for model training and can be found using a border sample selector.Simulation results demonstrate that the SD method minimizes the impact on model recognition performance while reducing the dataset size,and outperforms five other state-of-the-art core-set selection methods when the fraction of training sample kept is less than or equal to 0.3 on the RML2016.04C dataset or 0.5 on the RML22 dataset.The SD method is particularly helpful for signal recognition tasks with limited memory and computing resources.展开更多
In order to compete in the global manufacturing mar ke t, agility is the only possible solution to response to the fragmented market se gments and frequently changed customer requirements. However, manufacturing agil ...In order to compete in the global manufacturing mar ke t, agility is the only possible solution to response to the fragmented market se gments and frequently changed customer requirements. However, manufacturing agil ity can only be attained through the deployment of knowledge. To embed knowledge into a CAD system to form a knowledge intensive CAD (KIC) system is one of way to enhance the design compatibility of a manufacturing company. The most difficu lt phase to develop a KIC system is to capitalize a huge amount of legacy data t o form a knowledge database. In the past, such capitalization process could only be done solely manually or semi-automatic. In this paper, a five step model fo r automatic design knowledge capitalization through the use of data mining is pr oposed whilst details of how to select, verify and performance benchmarking an a ppropriate data mining algorithm for a specific design task will also be discuss ed. A case study concerning the design of a plastic toaster casing was used as an illustration for the proposed methodology and it was found that the avera ge absolute error of the predictions for the most appropriate algorithm is withi n 17%.展开更多
Geophysical data sets are growing at an ever-increasing rate,requiring computationally efficient data selection (thinning) methods to preserve essential information.Satellites,such as WindSat,provide large data sets...Geophysical data sets are growing at an ever-increasing rate,requiring computationally efficient data selection (thinning) methods to preserve essential information.Satellites,such as WindSat,provide large data sets for assessing the accuracy and computational efficiency of data selection techniques.A new data thinning technique,based on support vector regression (SVR),is developed and tested.To manage large on-line satellite data streams,observations from WindSat are formed into subsets by Voronoi tessellation and then each is thinned by SVR (TSVR).Three experiments are performed.The first confirms the viability of TSVR for a relatively small sample,comparing it to several commonly used data thinning methods (random selection,averaging and Barnes filtering),producing a 10% thinning rate (90% data reduction),low mean absolute errors (MAE) and large correlations with the original data.A second experiment,using a larger dataset,shows TSVR retrievals with MAE < 1 m s-1 and correlations ≥ 0.98.TSVR was an order of magnitude faster than the commonly used thinning methods.A third experiment applies a two-stage pipeline to TSVR,to accommodate online data.The pipeline subsets reconstruct the wind field with the same accuracy as the second experiment,is an order of magnitude faster than the nonpipeline TSVR.Therefore,pipeline TSVR is two orders of magnitude faster than commonly used thinning methods that ingest the entire data set.This study demonstrates that TSVR pipeline thinning is an accurate and computationally efficient alternative to commonly used data selection techniques.展开更多
Complex industry processes often need multiple operation modes to meet the change of production conditions. In the same mode,there are discrete samples belonging to this mode. Therefore,it is important to consider the...Complex industry processes often need multiple operation modes to meet the change of production conditions. In the same mode,there are discrete samples belonging to this mode. Therefore,it is important to consider the samples which are sparse in the mode.To solve this issue,a new approach called density-based support vector data description( DBSVDD) is proposed. In this article,an algorithm using Gaussian mixture model( GMM) with the DBSVDD technique is proposed for process monitoring. The GMM method is used to obtain the center of each mode and determine the number of the modes. Considering the complexity of the data distribution and discrete samples in monitoring process,the DBSVDD is utilized for process monitoring. Finally,the validity and effectiveness of the DBSVDD method are illustrated through the Tennessee Eastman( TE) process.展开更多
Switzerland is one of the most desirable European destinations for Chinese tourists;therefore, a better understanding of Chinese tourists is essential for successful business practices. In China, the largest and leadi...Switzerland is one of the most desirable European destinations for Chinese tourists;therefore, a better understanding of Chinese tourists is essential for successful business practices. In China, the largest and leading social media platform—Sina Weibo, a hybrid of Twitter and Facebook—has more than 600 million users. Weibo’s great market penetration suggests that tourism operators and markets need to understand how to build effective and sustainable communications on Chinese social media platforms. In order to offer a better decision support platform to tourism destination managers as well as Chinese tourists, we proposed a framework using linked data on Sina Weibo. Linked Data is a term referring to using the Internet to connect related data. We will show how it can be used and how ontology can be designed to include the users’ context (e.g., GPS locations). Our framework will provide a good theoretical foundation for further understand Chinese tourists’ expectation, experiences, behaviors and new trends in Switzerland.展开更多
Data refinement refers to the processes by which a dataset’s resolution,in particular,the spatial one,is refined,and is thus synonymous to spatial downscaling.Spatial resolution indicates measurement scale and can be...Data refinement refers to the processes by which a dataset’s resolution,in particular,the spatial one,is refined,and is thus synonymous to spatial downscaling.Spatial resolution indicates measurement scale and can be seen as an index for regular data support.As a type of change of scale,data refinement is useful for many scenarios where spatial scales of existing data,desired analyses,or specific applications need to be made commensurate and refined.As spatial data are related to certain data support,they can be conceived of as support-specific realizations of random fields,suggesting that multivariate geostatistics should be explored for refining datasets from their coarser-resolution versions to the finerresolution ones.In this paper,geostatistical methods for downscaling are described,and were implemented using GTOPO30 data and sampled Shuttle Radar Topography Mission data at a site in northwest China,with the latter’s majority grid cells used as surrogate reference data.It was found that proper structural modeling is important for achieving increased accuracy in data refinement;here,structural modeling can be done through proper decomposition of elevation fields into trends and residuals and thereafter.It was confirmed that effects of semantic differences on data refinement can be reduced through properly estimating and incorporating biases in local means.展开更多
基金partially supported by the National Key Research and Development Project under Grant2020YFB1806805Social Development Projects of Jiangsu Science and Technology Department under Grant No.BE2018704
文摘In wireless communication,the problem of authenticating the transmitter’s identity is challeng-ing,especially for those terminal devices in which the security schemes based on cryptography are approxi-mately unfeasible owing to limited resources.In this paper,a physical layer authentication scheme is pro-posed to detect whether there is anomalous access by the attackers disguised as legitimate users.Explicitly,channel state information(CSI)is used as a form of fingerprint to exploit spatial discrimination among de-vices in the wireless network and machine learning(ML)technology is employed to promote the improve-ment of authentication accuracy.Considering that the falsified messages are not accessible for authenticator during the training phase,deep support vector data de-scription(Deep SVDD)is selected to solve the one-class classification(OCC)problem.Simulation results show that Deep SVDD based scheme can tackle the challenges of physical layer authentication in wireless communication environments.
基金Project(61374140)supported by the National Natural Science Foundation of China
文摘There are multiple operating modes in the real industrial process, and the collected data follow the complex multimodal distribution, so most traditional process monitoring methods are no longer applicable because their presumptions are that sampled-data should obey the single Gaussian distribution or non-Gaussian distribution. In order to solve these problems, a novel weighted local standardization(WLS) strategy is proposed to standardize the multimodal data, which can eliminate the multi-mode characteristics of the collected data, and normalize them into unimodal data distribution. After detailed analysis of the raised data preprocessing strategy, a new algorithm using WLS strategy with support vector data description(SVDD) is put forward to apply for multi-mode monitoring process. Unlike the strategy of building multiple local models, the developed method only contains a model without the prior knowledge of multi-mode process. To demonstrate the proposed method's validity, it is applied to a numerical example and a Tennessee Eastman(TE) process. Finally, the simulation results show that the WLS strategy is very effective to standardize multimodal data, and the WLS-SVDD monitoring method has great advantages over the traditional SVDD and PCA combined with a local standardization strategy(LNS-PCA) in multi-mode process monitoring.
基金supported by the National Basic Research Program of China (the 973 Program,Grant No.2010CB951101)the Program for Changjiang Scholars and Innovative Research Teams in Universities,the Ministry of Education,China (Grant No. IRT0717)
文摘Hybrid data assimilation (DA) is a method seeing more use in recent hydrology and water resources research. In this study, a DA method coupled with the support vector machines (SVMs) and the ensemble Kalman filter (EnKF) technology was used for the prediction of soil moisture in different soil layers: 0-5 cm, 30 cm, 50 cm, 100 cm, 200 cm, and 300 cm. The SVM methodology was first used to train the ground measurements of soil moisture and meteorological parameters from the Meilin study area, in East China, to construct soil moisture statistical prediction models. Subsequent observations and their statistics were used for predictions, with two approaches: the SVM predictor and the SVM-EnKF model made by coupling the SVM model with the EnKF technique using the DA method. Validation results showed that the proposed SVM-EnKF model can improve the prediction results of soil moisture in different layers, from the surface to the root zone.
文摘An efficient trust-aware secure routing and network strategy-based data collection scheme is presented in this paper to enhance the performance and security of wireless sensor networks during data collection.The method first discovers the routes between the data sensors and the sink node.Several factors are considered for each sensor node along the route,including energy,number of neighbours,previous transmissions,and energy depletion ratio.Considering all these variables,the Sink Reachable Support Measure and the Secure Communication Support Measure,the method evaluates two distinct measures.The method calculates the data carrier support value using these two metrics.A single route is chosen to collect data based on the value of data carrier support.It has contributed to the design of Secure Communication Support(SCS)Estimation.This has been measured according to the strategy of each hop of the route.The suggested method improves the security and efficacy of data collection in wireless sensor networks.The second stage uses the two-fish approach to build a trust model for secure data transfer.A sim-ulation exercise was conducted to evaluate the effectiveness of the suggested framework.Metrics,including PDR,end-to-end latency,and average residual energy,were assessed for the proposed model.The efficiency of the suggested route design serves as evidence for the average residual energy for the proposed framework.
基金supported by National Natural Science Foundation of China(62371098)Natural Science Foundation of Sichuan Province(2023NSFSC1422)+1 种基金National Key Research and Development Program of China(2021YFB2900404)Central Universities of South west Minzu University(ZYN2022032).
文摘In recent years,deep learning-based signal recognition technology has gained attention and emerged as an important approach for safeguarding the electromagnetic environment.However,training deep learning-based classifiers on large signal datasets with redundant samples requires significant memory and high costs.This paper proposes a support databased core-set selection method(SD)for signal recognition,aiming to screen a representative subset that approximates the large signal dataset.Specifically,this subset can be identified by employing the labeled information during the early stages of model training,as some training samples are labeled as supporting data frequently.This support data is crucial for model training and can be found using a border sample selector.Simulation results demonstrate that the SD method minimizes the impact on model recognition performance while reducing the dataset size,and outperforms five other state-of-the-art core-set selection methods when the fraction of training sample kept is less than or equal to 0.3 on the RML2016.04C dataset or 0.5 on the RML22 dataset.The SD method is particularly helpful for signal recognition tasks with limited memory and computing resources.
基金Supported by National High Technology Research and Development Program of China(863 Program)(2006AA040308)National Natural Science Foundation of China(60736021)the National Creative Research Groups Science Foundation of China(60721062)
文摘In order to compete in the global manufacturing mar ke t, agility is the only possible solution to response to the fragmented market se gments and frequently changed customer requirements. However, manufacturing agil ity can only be attained through the deployment of knowledge. To embed knowledge into a CAD system to form a knowledge intensive CAD (KIC) system is one of way to enhance the design compatibility of a manufacturing company. The most difficu lt phase to develop a KIC system is to capitalize a huge amount of legacy data t o form a knowledge database. In the past, such capitalization process could only be done solely manually or semi-automatic. In this paper, a five step model fo r automatic design knowledge capitalization through the use of data mining is pr oposed whilst details of how to select, verify and performance benchmarking an a ppropriate data mining algorithm for a specific design task will also be discuss ed. A case study concerning the design of a plastic toaster casing was used as an illustration for the proposed methodology and it was found that the avera ge absolute error of the predictions for the most appropriate algorithm is withi n 17%.
基金NOAA Grant NA17RJ1227 and NSF Grant EIA-0205628 for providing financial support for this worksupported by RSF Grant 14-41-00039
文摘Geophysical data sets are growing at an ever-increasing rate,requiring computationally efficient data selection (thinning) methods to preserve essential information.Satellites,such as WindSat,provide large data sets for assessing the accuracy and computational efficiency of data selection techniques.A new data thinning technique,based on support vector regression (SVR),is developed and tested.To manage large on-line satellite data streams,observations from WindSat are formed into subsets by Voronoi tessellation and then each is thinned by SVR (TSVR).Three experiments are performed.The first confirms the viability of TSVR for a relatively small sample,comparing it to several commonly used data thinning methods (random selection,averaging and Barnes filtering),producing a 10% thinning rate (90% data reduction),low mean absolute errors (MAE) and large correlations with the original data.A second experiment,using a larger dataset,shows TSVR retrievals with MAE < 1 m s-1 and correlations ≥ 0.98.TSVR was an order of magnitude faster than the commonly used thinning methods.A third experiment applies a two-stage pipeline to TSVR,to accommodate online data.The pipeline subsets reconstruct the wind field with the same accuracy as the second experiment,is an order of magnitude faster than the nonpipeline TSVR.Therefore,pipeline TSVR is two orders of magnitude faster than commonly used thinning methods that ingest the entire data set.This study demonstrates that TSVR pipeline thinning is an accurate and computationally efficient alternative to commonly used data selection techniques.
基金National Natural Science Foundation of China(No.61374140)the Youth Foundation of National Natural Science Foundation of China(No.61403072)
文摘Complex industry processes often need multiple operation modes to meet the change of production conditions. In the same mode,there are discrete samples belonging to this mode. Therefore,it is important to consider the samples which are sparse in the mode.To solve this issue,a new approach called density-based support vector data description( DBSVDD) is proposed. In this article,an algorithm using Gaussian mixture model( GMM) with the DBSVDD technique is proposed for process monitoring. The GMM method is used to obtain the center of each mode and determine the number of the modes. Considering the complexity of the data distribution and discrete samples in monitoring process,the DBSVDD is utilized for process monitoring. Finally,the validity and effectiveness of the DBSVDD method are illustrated through the Tennessee Eastman( TE) process.
文摘Switzerland is one of the most desirable European destinations for Chinese tourists;therefore, a better understanding of Chinese tourists is essential for successful business practices. In China, the largest and leading social media platform—Sina Weibo, a hybrid of Twitter and Facebook—has more than 600 million users. Weibo’s great market penetration suggests that tourism operators and markets need to understand how to build effective and sustainable communications on Chinese social media platforms. In order to offer a better decision support platform to tourism destination managers as well as Chinese tourists, we proposed a framework using linked data on Sina Weibo. Linked Data is a term referring to using the Internet to connect related data. We will show how it can be used and how ontology can be designed to include the users’ context (e.g., GPS locations). Our framework will provide a good theoretical foundation for further understand Chinese tourists’ expectation, experiences, behaviors and new trends in Switzerland.
基金Research reported in this paper is supported by the National Natural Science Foundation of China(grant numbers 41171346,41471375).
文摘Data refinement refers to the processes by which a dataset’s resolution,in particular,the spatial one,is refined,and is thus synonymous to spatial downscaling.Spatial resolution indicates measurement scale and can be seen as an index for regular data support.As a type of change of scale,data refinement is useful for many scenarios where spatial scales of existing data,desired analyses,or specific applications need to be made commensurate and refined.As spatial data are related to certain data support,they can be conceived of as support-specific realizations of random fields,suggesting that multivariate geostatistics should be explored for refining datasets from their coarser-resolution versions to the finerresolution ones.In this paper,geostatistical methods for downscaling are described,and were implemented using GTOPO30 data and sampled Shuttle Radar Topography Mission data at a site in northwest China,with the latter’s majority grid cells used as surrogate reference data.It was found that proper structural modeling is important for achieving increased accuracy in data refinement;here,structural modeling can be done through proper decomposition of elevation fields into trends and residuals and thereafter.It was confirmed that effects of semantic differences on data refinement can be reduced through properly estimating and incorporating biases in local means.