Piezo actuators are widely used in ultra-precision fields because of their high response and nano-scale step length.However,their hysteresis characteristics seriously affect the accuracy and stability of piezo actuato...Piezo actuators are widely used in ultra-precision fields because of their high response and nano-scale step length.However,their hysteresis characteristics seriously affect the accuracy and stability of piezo actuators.Existing methods for fitting hysteresis loops include operator class,differential equation class,and machine learning class.The modeling cost of operator class and differential equation class methods is high,the model complexity is high,and the process of machine learning,such as neural network calculation,is opaque.The physical model framework cannot be directly extracted.Therefore,the sparse identification of nonlinear dynamics(SINDy)algorithm is proposed to fit hysteresis loops.Furthermore,the SINDy algorithm is improved.While the SINDy algorithm builds an orthogonal candidate database for modeling,the sparse regression model is simplified,and the Relay operator is introduced for piecewise fitting to solve the distortion problem of the SINDy algorithm fitting singularities.The Relay-SINDy algorithm proposed in this paper is applied to fitting hysteresis loops.Good performance is obtained with the experimental results of open and closed loops.Compared with the existing methods,the modeling cost and model complexity are reduced,and the modeling accuracy of the hysteresis loop is improved.展开更多
The pricing of moving window Asian option with an early exercise feature is considered a challenging problem in option pricing. The computational challenge lies in the unknown optimal exercise strategy and in the high...The pricing of moving window Asian option with an early exercise feature is considered a challenging problem in option pricing. The computational challenge lies in the unknown optimal exercise strategy and in the high dimensionality required for approximating the early exercise boundary. We use sparse grid basis functions in the Least Squares Monte Carlo approach to solve this “curse of dimensionality” problem. The resulting algorithm provides a general and convergent method for pricing moving window Asian options. The sparse grid technique presented in this paper can be generalized to pricing other high-dimensional, early-exercisable derivatives.展开更多
Latent factor(LF)models are highly effective in extracting useful knowledge from High-Dimensional and Sparse(HiDS)matrices which are commonly seen in various industrial applications.An LF model usually adopts iterativ...Latent factor(LF)models are highly effective in extracting useful knowledge from High-Dimensional and Sparse(HiDS)matrices which are commonly seen in various industrial applications.An LF model usually adopts iterative optimizers,which may consume many iterations to achieve a local optima,resulting in considerable time cost.Hence,determining how to accelerate the training process for LF models has become a significant issue.To address this,this work proposes a randomized latent factor(RLF)model.It incorporates the principle of randomized learning techniques from neural networks into the LF analysis of HiDS matrices,thereby greatly alleviating computational burden.It also extends a standard learning process for randomized neural networks in context of LF analysis to make the resulting model represent an HiDS matrix correctly.Experimental results on three HiDS matrices from industrial applications demonstrate that compared with state-of-the-art LF models,RLF is able to achieve significantly higher computational efficiency and comparable prediction accuracy for missing data.I provides an important alternative approach to LF analysis of HiDS matrices,which is especially desired for industrial applications demanding highly efficient models.展开更多
Modern agricultural mechanization has put forward higher requirements for the intelligent defect diagnosis.However,the fault features are usually learned and classified under all speeds without considering the effects...Modern agricultural mechanization has put forward higher requirements for the intelligent defect diagnosis.However,the fault features are usually learned and classified under all speeds without considering the effects of speed fluctuation.To overcome this deficiency,a novel intelligent defect detection framework based on time-frequency transformation is presented in this work.In the framework,the samples under one speed are employed for training sparse filtering model,and the remaining samples under different speeds are adopted for testing the effectiveness.Our proposed approach contains two stages:1)the time-frequency domain signals are acquired from the mechanical raw vibration data by the short time Fourier transform algorithm,and then the defect features are extracted from time-frequency domain signals by sparse filtering algorithm;2)different defect types are classified by the softmax regression using the defect features.The proposed approach can be employed to mine available fault characteristics adaptively and is an effective intelligent method for fault detection of agricultural equipment.The fault detection performances confirm that our approach not only owns strong ability for fault classification under different speeds,but also obtains higher identification accuracy than the other methods.展开更多
High-dimensional and sparse(HiDS)matrices commonly arise in various industrial applications,e.g.,recommender systems(RSs),social networks,and wireless sensor networks.Since they contain rich information,how to accurat...High-dimensional and sparse(HiDS)matrices commonly arise in various industrial applications,e.g.,recommender systems(RSs),social networks,and wireless sensor networks.Since they contain rich information,how to accurately represent them is of great significance.A latent factor(LF)model is one of the most popular and successful ways to address this issue.Current LF models mostly adopt L2-norm-oriented Loss to represent an HiDS matrix,i.e.,they sum the errors between observed data and predicted ones with L2-norm.Yet L2-norm is sensitive to outlier data.Unfortunately,outlier data usually exist in such matrices.For example,an HiDS matrix from RSs commonly contains many outlier ratings due to some heedless/malicious users.To address this issue,this work proposes a smooth L1-norm-oriented latent factor(SL-LF)model.Its main idea is to adopt smooth L1-norm rather than L2-norm to form its Loss,making it have both strong robustness and high accuracy in predicting the missing data of an HiDS matrix.Experimental results on eight HiDS matrices generated by industrial applications verify that the proposed SL-LF model not only is robust to the outlier data but also has significantly higher prediction accuracy than state-of-the-art models when they are used to predict the missing data of HiDS matrices.展开更多
For high-dimensional models with a focus on classification performance,the?1-penalized logistic regression is becoming important and popular.However,the Lasso estimates could be problematic when penalties of different...For high-dimensional models with a focus on classification performance,the?1-penalized logistic regression is becoming important and popular.However,the Lasso estimates could be problematic when penalties of different coefficients are all the same and not related to the data.We propose two types of weighted Lasso estimates,depending upon covariates determined by the Mc Diarmid inequality.Given sample size n and a dimension of covariates p,the finite sample behavior of our proposed method with a diverging number of predictors is illustrated by non-asymptotic oracle inequalities such as the?1-estimation error and the squared prediction error of the unknown parameters.We compare the performance of our method with that of former weighted estimates on simulated data,then apply it to do real data analysis.展开更多
The traditional estimation of Gaussian mixture model is sensitive to heavy-tailed errors;thus we propose a robust mixture regression model by assuming that the error terms follow a Laplace distribution in this article...The traditional estimation of Gaussian mixture model is sensitive to heavy-tailed errors;thus we propose a robust mixture regression model by assuming that the error terms follow a Laplace distribution in this article. And for the variable selection problem in our new robust mixture regression model, we introduce the adaptive sparse group Lasso penalty to achieve sparsity at both the group-level and within-group-level. As numerical experiments show, compared with other alternative methods, our method has better performances in variable selection and parameter estimation. Finally, we apply our proposed method to analyze NBA salary data during the period from 2018 to 2019.展开更多
Sparse coding is a prevalent method for image inpainting and feature extraction,which can repair corrupted images or improve data processing efficiency,and has numerous applications in computer vision and signal proce...Sparse coding is a prevalent method for image inpainting and feature extraction,which can repair corrupted images or improve data processing efficiency,and has numerous applications in computer vision and signal processing.Recently,sev-eral memristor-based in-memory computing systems have been proposed to enhance the efficiency of sparse coding remark-ably.However,the variations and low precision of the devices will deteriorate the dictionary,causing inevitable degradation in the accuracy and reliability of the application.In this work,a digital-analog hybrid memristive sparse coding system is pro-posed utilizing a multilevel Pt/Al_(2)O_(3)/AlO_(x)/W memristor,which employs the forward stagewise regression algorithm:The approxi-mate cosine distance calculation is conducted in the analog part to speed up the computation,followed by high-precision coeffi-cient updates performed in the digital portion.We determine that four states of the aforementioned memristor are sufficient for the processing of natural images.Furthermore,through dynamic adjustment of the mapping ratio,the precision require-ment for the digit-to-analog converters can be reduced to 4 bits.Compared to the previous system,our system achieves higher image reconstruction quality of the 38 dB peak-signal-to-noise ratio.Moreover,in the context of image inpainting,images containing 50%missing pixels can be restored with a reconstruction error of 0.0424 root-mean-squared error.展开更多
Due to the large scale and high dimension of teaching data,the using of traditional clustering algorithms has problems such as high computational complexity and low accuracy.Therefore,this paper proposes a weighted bl...Due to the large scale and high dimension of teaching data,the using of traditional clustering algorithms has problems such as high computational complexity and low accuracy.Therefore,this paper proposes a weighted block sparse subspace clustering algorithm based on information entropy.The introduction of information entropy weight and block diagonal constraints can obtain the prior probability that two pixels belong to the same category before the simulation experiment,thereby positively intervening that the solutions solved by the model tend to be the optimal approximate solutions of the block diagonal structure.It can enable the model to obtain the performance against noise and outliers,and thereby improving the discriminative ability of the model classification.The experimental results show that the average probability Rand index of the proposed method is 0.86,which is higher than that of other algorithms.The average information change index of the proposed method is 1.55,which is lower than that of other algorithms,proving its strong robustness.On different datasets,the misclassification rates of the design method are 1.2%and 0.9%respectively,which proves that its classification accuracy is relatively high.The proposed method has high reliability in processing highdimensional teaching data.It can play an important role in the field of educational data analysis and provide strong support for intelligent teaching.展开更多
Current high-dimensional feature screening methods still face significant challenges in handling mixed linear and nonlinear relationships,controlling redundant information,and improving model robustness.In this study,...Current high-dimensional feature screening methods still face significant challenges in handling mixed linear and nonlinear relationships,controlling redundant information,and improving model robustness.In this study,we propose a Dynamic Conditional Feature Screening(DCFS)method tailored for high-dimensional economic forecasting tasks.Our goal is to accurately identify key variables,enhance predictive performance,and provide both theoretical foundations and practical tools for macroeconomic modeling.The DCFS method constructs a comprehensive test statistic by integrating conditional mutual information with conditional regression error differences.By introducing a dynamic weighting mechanism,DCFS adaptively balances the linear and nonlinear contributions of features during the screening process.In addition,a dynamic thresholding mechanism is designed to effectively control the false discovery rate(FDR),thereby improving the stability and reliability of the screening results.On the theoretical front,we rigorously prove that the proposed method satisfies the sure screening property and rank consistency,ensuring accurate identification of the truly important feature set in high-dimensional settings.Simulation results demonstrate that under purely linear,purely nonlinear,and mixed dependency structures,DCFS consistently outperforms classical screening methods such as SIS,CSIS,and IG-SIS in terms of true positive rate(TPR),false discovery rate(FDR),and rank correlation.These results highlight the superior accuracy,robustness,and stability of our method.Furthermore,an empirical analysis based on the U.S.FRED-MD macroeconomic dataset confirms the practical value of DCFS in real-world forecasting tasks.The experimental results show that DCFS achieves lower prediction errors(RMSE and MAE)and higher R2 values in forecasting GDP growth.The selected key variables-including the Industrial Production Index(IP),Federal Funds Rate,Consumer Price Index(CPI),and Money Supply(M2)-possess clear economic interpretability,offering reliable support for economic forecasting and policy formulation.展开更多
In order to solve the problem of the reliability of slope engineering due to complex uncertainties, the Monte Carlo simulation method is adopted. Based on the characteristics of sparse grid, an interpolation algorithm...In order to solve the problem of the reliability of slope engineering due to complex uncertainties, the Monte Carlo simulation method is adopted. Based on the characteristics of sparse grid, an interpolation algorithm, which can be applied to high dimensional problems, is introduced. A surrogate model of high dimensional implicit function is established, which makes Monte Carlo method more adaptable. Finally, a reliability analysis method is proposed to evaluate the reliability of the slope engineering, and is applied in the Sau Mau Ping slope project in Hong Kong. The reliability analysis method has great theoretical and practical significance for engineering quality evaluation and natural disaster assessment.展开更多
The solution of normal least squares support vector regression(LSSVR)is lack of sparseness,which limits the real-time and hampers the wide applications to a certain degree.To overcome this obstacle,a scheme,named I2FS...The solution of normal least squares support vector regression(LSSVR)is lack of sparseness,which limits the real-time and hampers the wide applications to a certain degree.To overcome this obstacle,a scheme,named I2FSA-LSSVR,is proposed.Compared with the previously approximate algorithms,it not only adopts the partial reduction strategy but considers the influence between the previously selected support vectors and the willselected support vector during the process of computing the supporting weights.As a result,I2FSA-LSSVR reduces the number of support vectors and enhances the real-time.To confirm the feasibility and effectiveness of the proposed algorithm,experiments on benchmark data sets are conducted,whose results support the presented I2FSA-LSSVR.展开更多
Moderate resolution imaging spectroradiometer(MODIS)imaging has various applications in the field of ground monitoring,cloud classification and meteorological research.However,the limitations of the sensors and extern...Moderate resolution imaging spectroradiometer(MODIS)imaging has various applications in the field of ground monitoring,cloud classification and meteorological research.However,the limitations of the sensors and external disturbance make the resolution of image still limited in a certain level.The goal of this paper is to use a single image super-resolution(SISR)method to predict a high-resolution(HR)MODIS image from a single low-resolution(LR)input.Recently,although the method based on sparse representation has tackled the ill-posed problem effectively,two fatal issues have been ignored.First,many methods ignore the relationships among patches,resulting in some unfaithful output.Second,the high computational complexity of sparse coding using l_1 norm is needed in reconstruction stage.In this work,we discover the semantic relationships among LR patches and the corresponding HR patches and group the documents with similar semantic into topics by probabilistic Latent Semantic Analysis(p LSA).Then,we can learn dual dictionaries for each topic in the low-resolution(LR)patch space and high-resolution(HR)patch space and also pre-compute corresponding regression matrices for dictionary pairs.Finally,for the test image,we infer locally which topic it corresponds to and adaptive to select the regression matrix to reconstruct HR image by semantic relationships.Our method discovered the relationships among patches and pre-computed the regression matrices for topics.Therefore,our method can greatly reduce the artifacts and get some speed-up in the reconstruction phase.Experiment manifests that our method performs MODIS image super-resolution effectively,results in higher PSNR,reconstructs faster,and gets better visual quality than some current state-of-art methods.展开更多
Model averaging has attracted increasing attention in recent years for the analysis of high-dimensional data. By weighting several competing statistical models suitably, model averaging attempts to achieve stable and ...Model averaging has attracted increasing attention in recent years for the analysis of high-dimensional data. By weighting several competing statistical models suitably, model averaging attempts to achieve stable and improved prediction. To obtain a better understanding of the available model averaging methods, their properties and the relationships between them, this paper is devoted to make a review on some recent progresses in high-dimensional model averaging from the frequentist perspective. Some future research topics are also discussed.展开更多
The sparse phase retrieval aims to recover the sparse signal from quadratic measurements. However, the measurements are often affected by outliers and asymmetric distribution noise. This paper introduces a novel metho...The sparse phase retrieval aims to recover the sparse signal from quadratic measurements. However, the measurements are often affected by outliers and asymmetric distribution noise. This paper introduces a novel method that combines the quantile regression and the L<sub>1/2</sub>-regularizer. It is a non-convex, non-smooth, non-Lipschitz optimization problem. We propose an efficient algorithm based on the Alternating Direction Methods of Multiplier (ADMM) to solve the corresponding optimization problem. Numerous numerical experiments show that this method can recover sparse signals with fewer measurements and is robust to dense bounded noise and Laplace noise.展开更多
A Single Image Super-Resolution (SISR) reconstruction method that uses clustered sparse representation and adaptive patch aggregation is proposed. First, we randomly extract image patch pairs from the training images,...A Single Image Super-Resolution (SISR) reconstruction method that uses clustered sparse representation and adaptive patch aggregation is proposed. First, we randomly extract image patch pairs from the training images, and divide these patch pairs into different groups by K-means clustering. Then, we learn an over-complete sub-dictionary pair offline from corresponding group patch pairs. For a given low-resolution patch, we adaptively select one sub-dictionary to reconstruct the high resolution patch online. In addition, non-local self-similarity and steering kernel regression constraints are integrated into patch aggregation to improve the quality of the recovered images. Experiments show that the proposed method is able to realize state-of-the-art performance in terms of both objective evaluation and visual perception.展开更多
This paper studies variable selection using the penalized likelihood method for dis-tributed sparse regression with large sample size n under a limited memory constraint.This is a much needed research problem to be so...This paper studies variable selection using the penalized likelihood method for dis-tributed sparse regression with large sample size n under a limited memory constraint.This is a much needed research problem to be solved in the big data era.A naive divide-and-conquer method solving this problem is to split the whole data into N parts and run each part on one of N machines,aggregate the results from all machines via averaging,andfinally obtain the selected variables.However,it tends to select more noise variables,and the false discovery rate may not be well controlled.We improve it by a special designed weighted average in aggregation.Although the alternating direction method of multiplier can be used to deal with massive data in the literature,our proposed method reduces the computational burden a lot and performs better by mean square error in most cases.Theoretically,we establish asymptotic properties of the resulting estimators for the likelihood models with a diverging number of parame-ters.Under some regularity conditions,we establish oracle properties in the sense that our distributed estimator shares the same asymptotic efficiency as the estimator based on the full sample.Computationally,a distributed penalized likelihood algorithm is proposed to refine the results in the context of general likelihoods.Furthermore,the proposed method is evaluated by simulations and a real example.展开更多
针对过程复杂且结构未知的对象,在保证模型有效性的前提下,根据数据信息构建简单模型来简化控制器的求解是亟待解决的问题。以受控自回归模型为例,提出一种基于修正最小角回归算法的稀疏辨识方法。首先将系统模型转化为过参数化的高维...针对过程复杂且结构未知的对象,在保证模型有效性的前提下,根据数据信息构建简单模型来简化控制器的求解是亟待解决的问题。以受控自回归模型为例,提出一种基于修正最小角回归算法的稀疏辨识方法。首先将系统模型转化为过参数化的高维稀疏模型,然后将最小角回归算法用于稀疏系统辨识,并提出绝对角度停止准则,使算法经过少量的迭代即可获得模型的稀疏参数估计,并同时获得有效的时滞和阶次估计。结合辨识得到的受控自回归模型,引入一种基于指定相位点频率和增益的比例-积分-微分(proportional integral derivative,PID)控制器。数值仿真和平衡机器人的姿态控制仿真表明,该稀疏辨识算法在低数据量下具有较高的辨识精度,建立的模型具有较好的泛化性能,控制器具有良好的控制效果。展开更多
基金National Natural Science Foundation of China(62203118)。
文摘Piezo actuators are widely used in ultra-precision fields because of their high response and nano-scale step length.However,their hysteresis characteristics seriously affect the accuracy and stability of piezo actuators.Existing methods for fitting hysteresis loops include operator class,differential equation class,and machine learning class.The modeling cost of operator class and differential equation class methods is high,the model complexity is high,and the process of machine learning,such as neural network calculation,is opaque.The physical model framework cannot be directly extracted.Therefore,the sparse identification of nonlinear dynamics(SINDy)algorithm is proposed to fit hysteresis loops.Furthermore,the SINDy algorithm is improved.While the SINDy algorithm builds an orthogonal candidate database for modeling,the sparse regression model is simplified,and the Relay operator is introduced for piecewise fitting to solve the distortion problem of the SINDy algorithm fitting singularities.The Relay-SINDy algorithm proposed in this paper is applied to fitting hysteresis loops.Good performance is obtained with the experimental results of open and closed loops.Compared with the existing methods,the modeling cost and model complexity are reduced,and the modeling accuracy of the hysteresis loop is improved.
文摘The pricing of moving window Asian option with an early exercise feature is considered a challenging problem in option pricing. The computational challenge lies in the unknown optimal exercise strategy and in the high dimensionality required for approximating the early exercise boundary. We use sparse grid basis functions in the Least Squares Monte Carlo approach to solve this “curse of dimensionality” problem. The resulting algorithm provides a general and convergent method for pricing moving window Asian options. The sparse grid technique presented in this paper can be generalized to pricing other high-dimensional, early-exercisable derivatives.
基金supported in part by the National Natural Science Foundation of China (6177249391646114)+1 种基金Chongqing research program of technology innovation and application (cstc2017rgzn-zdyfX0020)in part by the Pioneer Hundred Talents Program of Chinese Academy of Sciences
文摘Latent factor(LF)models are highly effective in extracting useful knowledge from High-Dimensional and Sparse(HiDS)matrices which are commonly seen in various industrial applications.An LF model usually adopts iterative optimizers,which may consume many iterations to achieve a local optima,resulting in considerable time cost.Hence,determining how to accelerate the training process for LF models has become a significant issue.To address this,this work proposes a randomized latent factor(RLF)model.It incorporates the principle of randomized learning techniques from neural networks into the LF analysis of HiDS matrices,thereby greatly alleviating computational burden.It also extends a standard learning process for randomized neural networks in context of LF analysis to make the resulting model represent an HiDS matrix correctly.Experimental results on three HiDS matrices from industrial applications demonstrate that compared with state-of-the-art LF models,RLF is able to achieve significantly higher computational efficiency and comparable prediction accuracy for missing data.I provides an important alternative approach to LF analysis of HiDS matrices,which is especially desired for industrial applications demanding highly efficient models.
基金Project(51675262)supported by the National Natural Science Foundation of ChinaProject(2016YFD0700800)supported by the National Key Research and Development Program of China+2 种基金Project(6140210020102)supported by the Advance Research Field Fund Project of ChinaProject(NP2018304)supported by the Fundamental Research Funds for the Central Universities,ChinaProject(2017-IV-0008-0045)supported by the National Science and Technology Major Project
文摘Modern agricultural mechanization has put forward higher requirements for the intelligent defect diagnosis.However,the fault features are usually learned and classified under all speeds without considering the effects of speed fluctuation.To overcome this deficiency,a novel intelligent defect detection framework based on time-frequency transformation is presented in this work.In the framework,the samples under one speed are employed for training sparse filtering model,and the remaining samples under different speeds are adopted for testing the effectiveness.Our proposed approach contains two stages:1)the time-frequency domain signals are acquired from the mechanical raw vibration data by the short time Fourier transform algorithm,and then the defect features are extracted from time-frequency domain signals by sparse filtering algorithm;2)different defect types are classified by the softmax regression using the defect features.The proposed approach can be employed to mine available fault characteristics adaptively and is an effective intelligent method for fault detection of agricultural equipment.The fault detection performances confirm that our approach not only owns strong ability for fault classification under different speeds,but also obtains higher identification accuracy than the other methods.
基金supported in part by the National Natural Science Foundation of China(61702475,61772493,61902370,62002337)in part by the Natural Science Foundation of Chongqing,China(cstc2019jcyj-msxmX0578,cstc2019jcyjjqX0013)+1 种基金in part by the Chinese Academy of Sciences“Light of West China”Program,in part by the Pioneer Hundred Talents Program of Chinese Academy of Sciencesby Technology Innovation and Application Development Project of Chongqing,China(cstc2019jscx-fxydX0027)。
文摘High-dimensional and sparse(HiDS)matrices commonly arise in various industrial applications,e.g.,recommender systems(RSs),social networks,and wireless sensor networks.Since they contain rich information,how to accurately represent them is of great significance.A latent factor(LF)model is one of the most popular and successful ways to address this issue.Current LF models mostly adopt L2-norm-oriented Loss to represent an HiDS matrix,i.e.,they sum the errors between observed data and predicted ones with L2-norm.Yet L2-norm is sensitive to outlier data.Unfortunately,outlier data usually exist in such matrices.For example,an HiDS matrix from RSs commonly contains many outlier ratings due to some heedless/malicious users.To address this issue,this work proposes a smooth L1-norm-oriented latent factor(SL-LF)model.Its main idea is to adopt smooth L1-norm rather than L2-norm to form its Loss,making it have both strong robustness and high accuracy in predicting the missing data of an HiDS matrix.Experimental results on eight HiDS matrices generated by industrial applications verify that the proposed SL-LF model not only is robust to the outlier data but also has significantly higher prediction accuracy than state-of-the-art models when they are used to predict the missing data of HiDS matrices.
基金Supported by the National Natural Science Foundation of China(61877023)the Fundamental Research Funds for the Central Universities(CCNU19TD009)。
文摘For high-dimensional models with a focus on classification performance,the?1-penalized logistic regression is becoming important and popular.However,the Lasso estimates could be problematic when penalties of different coefficients are all the same and not related to the data.We propose two types of weighted Lasso estimates,depending upon covariates determined by the Mc Diarmid inequality.Given sample size n and a dimension of covariates p,the finite sample behavior of our proposed method with a diverging number of predictors is illustrated by non-asymptotic oracle inequalities such as the?1-estimation error and the squared prediction error of the unknown parameters.We compare the performance of our method with that of former weighted estimates on simulated data,then apply it to do real data analysis.
文摘The traditional estimation of Gaussian mixture model is sensitive to heavy-tailed errors;thus we propose a robust mixture regression model by assuming that the error terms follow a Laplace distribution in this article. And for the variable selection problem in our new robust mixture regression model, we introduce the adaptive sparse group Lasso penalty to achieve sparsity at both the group-level and within-group-level. As numerical experiments show, compared with other alternative methods, our method has better performances in variable selection and parameter estimation. Finally, we apply our proposed method to analyze NBA salary data during the period from 2018 to 2019.
基金This work was supported by the National Key R&D Program of China(Grant No.2019YFB2205100)in part by Hubei Key Laboratory of Advanced Memories.
文摘Sparse coding is a prevalent method for image inpainting and feature extraction,which can repair corrupted images or improve data processing efficiency,and has numerous applications in computer vision and signal processing.Recently,sev-eral memristor-based in-memory computing systems have been proposed to enhance the efficiency of sparse coding remark-ably.However,the variations and low precision of the devices will deteriorate the dictionary,causing inevitable degradation in the accuracy and reliability of the application.In this work,a digital-analog hybrid memristive sparse coding system is pro-posed utilizing a multilevel Pt/Al_(2)O_(3)/AlO_(x)/W memristor,which employs the forward stagewise regression algorithm:The approxi-mate cosine distance calculation is conducted in the analog part to speed up the computation,followed by high-precision coeffi-cient updates performed in the digital portion.We determine that four states of the aforementioned memristor are sufficient for the processing of natural images.Furthermore,through dynamic adjustment of the mapping ratio,the precision require-ment for the digit-to-analog converters can be reduced to 4 bits.Compared to the previous system,our system achieves higher image reconstruction quality of the 38 dB peak-signal-to-noise ratio.Moreover,in the context of image inpainting,images containing 50%missing pixels can be restored with a reconstruction error of 0.0424 root-mean-squared error.
文摘Due to the large scale and high dimension of teaching data,the using of traditional clustering algorithms has problems such as high computational complexity and low accuracy.Therefore,this paper proposes a weighted block sparse subspace clustering algorithm based on information entropy.The introduction of information entropy weight and block diagonal constraints can obtain the prior probability that two pixels belong to the same category before the simulation experiment,thereby positively intervening that the solutions solved by the model tend to be the optimal approximate solutions of the block diagonal structure.It can enable the model to obtain the performance against noise and outliers,and thereby improving the discriminative ability of the model classification.The experimental results show that the average probability Rand index of the proposed method is 0.86,which is higher than that of other algorithms.The average information change index of the proposed method is 1.55,which is lower than that of other algorithms,proving its strong robustness.On different datasets,the misclassification rates of the design method are 1.2%and 0.9%respectively,which proves that its classification accuracy is relatively high.The proposed method has high reliability in processing highdimensional teaching data.It can play an important role in the field of educational data analysis and provide strong support for intelligent teaching.
文摘Current high-dimensional feature screening methods still face significant challenges in handling mixed linear and nonlinear relationships,controlling redundant information,and improving model robustness.In this study,we propose a Dynamic Conditional Feature Screening(DCFS)method tailored for high-dimensional economic forecasting tasks.Our goal is to accurately identify key variables,enhance predictive performance,and provide both theoretical foundations and practical tools for macroeconomic modeling.The DCFS method constructs a comprehensive test statistic by integrating conditional mutual information with conditional regression error differences.By introducing a dynamic weighting mechanism,DCFS adaptively balances the linear and nonlinear contributions of features during the screening process.In addition,a dynamic thresholding mechanism is designed to effectively control the false discovery rate(FDR),thereby improving the stability and reliability of the screening results.On the theoretical front,we rigorously prove that the proposed method satisfies the sure screening property and rank consistency,ensuring accurate identification of the truly important feature set in high-dimensional settings.Simulation results demonstrate that under purely linear,purely nonlinear,and mixed dependency structures,DCFS consistently outperforms classical screening methods such as SIS,CSIS,and IG-SIS in terms of true positive rate(TPR),false discovery rate(FDR),and rank correlation.These results highlight the superior accuracy,robustness,and stability of our method.Furthermore,an empirical analysis based on the U.S.FRED-MD macroeconomic dataset confirms the practical value of DCFS in real-world forecasting tasks.The experimental results show that DCFS achieves lower prediction errors(RMSE and MAE)and higher R2 values in forecasting GDP growth.The selected key variables-including the Industrial Production Index(IP),Federal Funds Rate,Consumer Price Index(CPI),and Money Supply(M2)-possess clear economic interpretability,offering reliable support for economic forecasting and policy formulation.
基金Supported by projects of China Ocean Research Mineral Resources R&D Association(COMRA)Special Foundation(DY135-R2-1-01,DY135-46)the Province/Jilin University Co-Construction Project-Funds for New Materials(SXGJSF2017-3)
文摘In order to solve the problem of the reliability of slope engineering due to complex uncertainties, the Monte Carlo simulation method is adopted. Based on the characteristics of sparse grid, an interpolation algorithm, which can be applied to high dimensional problems, is introduced. A surrogate model of high dimensional implicit function is established, which makes Monte Carlo method more adaptable. Finally, a reliability analysis method is proposed to evaluate the reliability of the slope engineering, and is applied in the Sau Mau Ping slope project in Hong Kong. The reliability analysis method has great theoretical and practical significance for engineering quality evaluation and natural disaster assessment.
基金Supported by the National Natural Science Foundation of China(51006052)
文摘The solution of normal least squares support vector regression(LSSVR)is lack of sparseness,which limits the real-time and hampers the wide applications to a certain degree.To overcome this obstacle,a scheme,named I2FSA-LSSVR,is proposed.Compared with the previously approximate algorithms,it not only adopts the partial reduction strategy but considers the influence between the previously selected support vectors and the willselected support vector during the process of computing the supporting weights.As a result,I2FSA-LSSVR reduces the number of support vectors and enhances the real-time.To confirm the feasibility and effectiveness of the proposed algorithm,experiments on benchmark data sets are conducted,whose results support the presented I2FSA-LSSVR.
基金partially supported by the National Natural Science Foundation of China(61471212)Natural Science Foundation of Zhejiang Province(LY16F010001)Natural Science Foundation of Ningbo(2016A610091,2017A610297)
文摘Moderate resolution imaging spectroradiometer(MODIS)imaging has various applications in the field of ground monitoring,cloud classification and meteorological research.However,the limitations of the sensors and external disturbance make the resolution of image still limited in a certain level.The goal of this paper is to use a single image super-resolution(SISR)method to predict a high-resolution(HR)MODIS image from a single low-resolution(LR)input.Recently,although the method based on sparse representation has tackled the ill-posed problem effectively,two fatal issues have been ignored.First,many methods ignore the relationships among patches,resulting in some unfaithful output.Second,the high computational complexity of sparse coding using l_1 norm is needed in reconstruction stage.In this work,we discover the semantic relationships among LR patches and the corresponding HR patches and group the documents with similar semantic into topics by probabilistic Latent Semantic Analysis(p LSA).Then,we can learn dual dictionaries for each topic in the low-resolution(LR)patch space and high-resolution(HR)patch space and also pre-compute corresponding regression matrices for dictionary pairs.Finally,for the test image,we infer locally which topic it corresponds to and adaptive to select the regression matrix to reconstruct HR image by semantic relationships.Our method discovered the relationships among patches and pre-computed the regression matrices for topics.Therefore,our method can greatly reduce the artifacts and get some speed-up in the reconstruction phase.Experiment manifests that our method performs MODIS image super-resolution effectively,results in higher PSNR,reconstructs faster,and gets better visual quality than some current state-of-art methods.
文摘Model averaging has attracted increasing attention in recent years for the analysis of high-dimensional data. By weighting several competing statistical models suitably, model averaging attempts to achieve stable and improved prediction. To obtain a better understanding of the available model averaging methods, their properties and the relationships between them, this paper is devoted to make a review on some recent progresses in high-dimensional model averaging from the frequentist perspective. Some future research topics are also discussed.
文摘The sparse phase retrieval aims to recover the sparse signal from quadratic measurements. However, the measurements are often affected by outliers and asymmetric distribution noise. This paper introduces a novel method that combines the quantile regression and the L<sub>1/2</sub>-regularizer. It is a non-convex, non-smooth, non-Lipschitz optimization problem. We propose an efficient algorithm based on the Alternating Direction Methods of Multiplier (ADMM) to solve the corresponding optimization problem. Numerous numerical experiments show that this method can recover sparse signals with fewer measurements and is robust to dense bounded noise and Laplace noise.
基金partially supported by the National Natural Science Foundation of China under Grants No. 61071146, No. 61171165the Natural Science Foundation of Jiangsu Province under Grant No. BK2010488+1 种基金sponsored by Qing Lan Project, Project 333 "The Six Top Talents" of Jiangsu Province
文摘A Single Image Super-Resolution (SISR) reconstruction method that uses clustered sparse representation and adaptive patch aggregation is proposed. First, we randomly extract image patch pairs from the training images, and divide these patch pairs into different groups by K-means clustering. Then, we learn an over-complete sub-dictionary pair offline from corresponding group patch pairs. For a given low-resolution patch, we adaptively select one sub-dictionary to reconstruct the high resolution patch online. In addition, non-local self-similarity and steering kernel regression constraints are integrated into patch aggregation to improve the quality of the recovered images. Experiments show that the proposed method is able to realize state-of-the-art performance in terms of both objective evaluation and visual perception.
基金supported by NSFC(11871263)NSF grant of Guangdong Province of China(No.2017A030313012).
文摘This paper studies variable selection using the penalized likelihood method for dis-tributed sparse regression with large sample size n under a limited memory constraint.This is a much needed research problem to be solved in the big data era.A naive divide-and-conquer method solving this problem is to split the whole data into N parts and run each part on one of N machines,aggregate the results from all machines via averaging,andfinally obtain the selected variables.However,it tends to select more noise variables,and the false discovery rate may not be well controlled.We improve it by a special designed weighted average in aggregation.Although the alternating direction method of multiplier can be used to deal with massive data in the literature,our proposed method reduces the computational burden a lot and performs better by mean square error in most cases.Theoretically,we establish asymptotic properties of the resulting estimators for the likelihood models with a diverging number of parame-ters.Under some regularity conditions,we establish oracle properties in the sense that our distributed estimator shares the same asymptotic efficiency as the estimator based on the full sample.Computationally,a distributed penalized likelihood algorithm is proposed to refine the results in the context of general likelihoods.Furthermore,the proposed method is evaluated by simulations and a real example.
文摘针对过程复杂且结构未知的对象,在保证模型有效性的前提下,根据数据信息构建简单模型来简化控制器的求解是亟待解决的问题。以受控自回归模型为例,提出一种基于修正最小角回归算法的稀疏辨识方法。首先将系统模型转化为过参数化的高维稀疏模型,然后将最小角回归算法用于稀疏系统辨识,并提出绝对角度停止准则,使算法经过少量的迭代即可获得模型的稀疏参数估计,并同时获得有效的时滞和阶次估计。结合辨识得到的受控自回归模型,引入一种基于指定相位点频率和增益的比例-积分-微分(proportional integral derivative,PID)控制器。数值仿真和平衡机器人的姿态控制仿真表明,该稀疏辨识算法在低数据量下具有较高的辨识精度,建立的模型具有较好的泛化性能,控制器具有良好的控制效果。