In this paper,we consider the knot placement problem in B-spline curve approximation.A novel two-stage framework is proposed for addressing this problem.In the first step,the l_(∞,1)-norm model is introduced for the ...In this paper,we consider the knot placement problem in B-spline curve approximation.A novel two-stage framework is proposed for addressing this problem.In the first step,the l_(∞,1)-norm model is introduced for the sparse selection of candidate knots from an initial knot vector.By this step,the knot number is determined.In the second step,knot positions are formulated into a nonlinear optimization problem and optimized by a global optimization algorithm—the differential evolution algorithm(DE).The candidate knots selected in the first step are served for initial values of the DE algorithm.Since the candidate knots provide a good guess of knot positions,the DE algorithm can quickly converge.One advantage of the proposed algorithm is that the knot number and knot positions are determined automatically.Compared with the current existing algorithms,the proposed algorithm finds approximations with smaller fitting error when the knot number is fixed in advance.Furthermore,the proposed algorithm is robust to noisy data and can handle with few data points.We illustrate with some examples and applications.展开更多
Seismic data regularization is an important preprocessing step in seismic signal processing. Traditional seismic acquisition methods follow the Shannon–Nyquist sampling theorem, whereas compressive sensing(CS) prov...Seismic data regularization is an important preprocessing step in seismic signal processing. Traditional seismic acquisition methods follow the Shannon–Nyquist sampling theorem, whereas compressive sensing(CS) provides a fundamentally new paradigm to overcome limitations in data acquisition. Besides the sparse representation of seismic signal in some transform domain and the 1-norm reconstruction algorithm, the seismic data regularization quality of CS-based techniques strongly depends on random undersampling schemes. For 2D seismic data, discrete uniform-based methods have been investigated, where some seismic traces are randomly sampled with an equal probability. However, in theory and practice, some seismic traces with different probability are required to be sampled for satisfying the assumptions in CS. Therefore, designing new undersampling schemes is imperative. We propose a Bernoulli-based random undersampling scheme and its jittered version to determine the regular traces that are randomly sampled with different probability, while both schemes comply with the Bernoulli process distribution. We performed experiments using the Fourier and curvelet transforms and the spectral projected gradient reconstruction algorithm for 1-norm(SPGL1), and ten different random seeds. According to the signal-to-noise ratio(SNR) between the original and reconstructed seismic data, the detailed experimental results from 2D numerical and physical simulation data show that the proposed novel schemes perform overall better than the discrete uniform schemes.展开更多
Principal component analysis(PCA) is fundamental in many pattern recognition applications.Much research has been performed to minimize the reconstruction error in L1-norm based reconstruction error minimization(L1-PCA...Principal component analysis(PCA) is fundamental in many pattern recognition applications.Much research has been performed to minimize the reconstruction error in L1-norm based reconstruction error minimization(L1-PCA-REM) since conventional L2-norm based PCA(L2-PCA) is sensitive to outliers.Recently,the variance maximization formulation of PCA with L1-norm(L1-PCA-VM) has been proposed,where new greedy and nongreedy solutions are developed.Armed with the gradient ascent perspective for optimization,we show that the L1-PCA-VM formulation is problematic in learning principal components and that only a greedy solution can achieve robustness motivation,which are verified by experiments on synthetic and real-world datasets.展开更多
基金supported by the National Natural Science Foundation of China(Nos.11871447,11801393)the Natural Science Foundation of Jiangsu Province(No.BK20180831).
文摘In this paper,we consider the knot placement problem in B-spline curve approximation.A novel two-stage framework is proposed for addressing this problem.In the first step,the l_(∞,1)-norm model is introduced for the sparse selection of candidate knots from an initial knot vector.By this step,the knot number is determined.In the second step,knot positions are formulated into a nonlinear optimization problem and optimized by a global optimization algorithm—the differential evolution algorithm(DE).The candidate knots selected in the first step are served for initial values of the DE algorithm.Since the candidate knots provide a good guess of knot positions,the DE algorithm can quickly converge.One advantage of the proposed algorithm is that the knot number and knot positions are determined automatically.Compared with the current existing algorithms,the proposed algorithm finds approximations with smaller fitting error when the knot number is fixed in advance.Furthermore,the proposed algorithm is robust to noisy data and can handle with few data points.We illustrate with some examples and applications.
基金financially supported by The 2011 Prospective Research Project of SINOPEC(P11096)
文摘Seismic data regularization is an important preprocessing step in seismic signal processing. Traditional seismic acquisition methods follow the Shannon–Nyquist sampling theorem, whereas compressive sensing(CS) provides a fundamentally new paradigm to overcome limitations in data acquisition. Besides the sparse representation of seismic signal in some transform domain and the 1-norm reconstruction algorithm, the seismic data regularization quality of CS-based techniques strongly depends on random undersampling schemes. For 2D seismic data, discrete uniform-based methods have been investigated, where some seismic traces are randomly sampled with an equal probability. However, in theory and practice, some seismic traces with different probability are required to be sampled for satisfying the assumptions in CS. Therefore, designing new undersampling schemes is imperative. We propose a Bernoulli-based random undersampling scheme and its jittered version to determine the regular traces that are randomly sampled with different probability, while both schemes comply with the Bernoulli process distribution. We performed experiments using the Fourier and curvelet transforms and the spectral projected gradient reconstruction algorithm for 1-norm(SPGL1), and ten different random seeds. According to the signal-to-noise ratio(SNR) between the original and reconstructed seismic data, the detailed experimental results from 2D numerical and physical simulation data show that the proposed novel schemes perform overall better than the discrete uniform schemes.
基金Project supported by the National Natural Science Foundation of China (Nos. 61071131 and 61271388)the Beijing Natural Science Foundation (No. 4122040)+1 种基金the Research Project of Tsinghua University (No. 2012Z01011)the United Technologies Research Center (UTRC)
文摘Principal component analysis(PCA) is fundamental in many pattern recognition applications.Much research has been performed to minimize the reconstruction error in L1-norm based reconstruction error minimization(L1-PCA-REM) since conventional L2-norm based PCA(L2-PCA) is sensitive to outliers.Recently,the variance maximization formulation of PCA with L1-norm(L1-PCA-VM) has been proposed,where new greedy and nongreedy solutions are developed.Armed with the gradient ascent perspective for optimization,we show that the L1-PCA-VM formulation is problematic in learning principal components and that only a greedy solution can achieve robustness motivation,which are verified by experiments on synthetic and real-world datasets.