The El Niño-Southern Oscillation(ENSO)is a naturally recurring interannual climate fluctuation that affects the global climate system.The advent of deep learning-based approaches has led to transformative changes...The El Niño-Southern Oscillation(ENSO)is a naturally recurring interannual climate fluctuation that affects the global climate system.The advent of deep learning-based approaches has led to transformative changes in ENSO forecasts,resulting in significant progress.Most deep learning-based ENSO prediction models which primarily rely solely on reanalysis data may lead to challenges in intensity underestimation in long-term forecasts,reducing the forecasting skills.To this end,we propose a deep residual-coupled model prediction(Res-CMP)model,which integrates historical reanalysis data and coupled model forecast data for multiyear ENSO prediction.The Res-CMP model is designed as a lightweight model that leverages only short-term reanalysis data and nudging assimilation prediction results of the Community Earth System Model(CESM)for effective prediction of the Niño 3.4 index.We also developed a transfer learning strategy for this model to overcome the limitations of inadequate forecast data.After determining the optimal configuration,which included selecting a suitable transfer learning rate during training,along with input variables and CESM forecast lengths,Res-CMP demonstrated a high correlation ability for 19-month lead time predictions(correlation coefficients exceeding 0.5).The Res-CMP model also alleviated the spring predictability barrier(SPB).When validated against actual ENSO events,Res-CMP successfully captured the temporal evolution of the Niño 3.4 index during La Niña events(1998/99 and 2020/21)and El Niño events(2009/10 and 2015/16).Our proposed model has the potential to further enhance ENSO prediction performance by using coupled models to assist deep learning methods.展开更多
Arctic sea ice is an important component of the global climate system and has experienced rapid changes during in the past few decades,the prediction of which is a significant application for climate models.In this st...Arctic sea ice is an important component of the global climate system and has experienced rapid changes during in the past few decades,the prediction of which is a significant application for climate models.In this study,a Localized Error Subspace Transform Kalman Filter is employed in a coupled climate system model(the Flexible Global Ocean–Atmosphere–Land System Model,version f3-L(FGOALS-f3-L))to assimilate sea-ice concentration(SIC)and sea-ice thickness(SIT)data for melting-season ice predictions.The scheme is applied through the following steps:(1)initialization for generating initial ensembles;(2)analysis for assimilating observed data;(3)adoption for dividing ice states into five thickness categories;(4)forecast for evolving the model;(5)resampling for updating model uncertainties.Several experiments were conducted to examine its results and impacts.Compared with the control experiment,the continuous assimilation experiments(CTNs)indicate assimilations improve model SICs and SITs persistently and generate realistic initials.Assimilating SIC+SIT data better corrects overestimated model SITs spatially than when only assimilating SIC data.The continuous assimilation restart experiments indicate the initials from the CTNs correct the overestimated marginal SICs and overall SITs remarkably well,as well as the cold biases in the oceanic and atmospheric models.The initials with SIC+SIT assimilated show more reasonable spatial improvements.Nevertheless,the SICs in the central Arctic undergo abnormal summer reductions,which is probably because overestimated SITs are reduced in the initials but the strong seasonal cycle(summer melting)biases are unchanged.Therefore,since systematic biases are complicated in a coupled system,for FGOALS-f3-L to make better ice predictions,oceanic and atmospheric assimilations are expected required.展开更多
Objective Humans are exposed to complex mixtures of environmental chemicals and other factors that can affect their health.Analysis of these mixture exposures presents several key challenges for environmental epidemio...Objective Humans are exposed to complex mixtures of environmental chemicals and other factors that can affect their health.Analysis of these mixture exposures presents several key challenges for environmental epidemiology and risk assessment,including high dimensionality,correlated exposure,and subtle individual effects.Methods We proposed a novel statistical approach,the generalized functional linear model(GFLM),to analyze the health effects of exposure mixtures.GFLM treats the effect of mixture exposures as a smooth function by reordering exposures based on specific mechanisms and capturing internal correlations to provide a meaningful estimation and interpretation.The robustness and efficiency was evaluated under various scenarios through extensive simulation studies.Results We applied the GFLM to two datasets from the National Health and Nutrition Examination Survey(NHANES).In the first application,we examined the effects of 37 nutrients on BMI(2011–2016 cycles).The GFLM identified a significant mixture effect,with fiber and fat emerging as the nutrients with the greatest negative and positive effects on BMI,respectively.For the second application,we investigated the association between four pre-and perfluoroalkyl substances(PFAS)and gout risk(2007–2018 cycles).Unlike traditional methods,the GFLM indicated no significant association,demonstrating its robustness to multicollinearity.Conclusion GFLM framework is a powerful tool for mixture exposure analysis,offering improved handling of correlated exposures and interpretable results.It demonstrates robust performance across various scenarios and real-world applications,advancing our understanding of complex environmental exposures and their health impacts on environmental epidemiology and toxicology.展开更多
High-dimensional and incomplete(HDI) matrices are primarily generated in all kinds of big-data-related practical applications. A latent factor analysis(LFA) model is capable of conducting efficient representation lear...High-dimensional and incomplete(HDI) matrices are primarily generated in all kinds of big-data-related practical applications. A latent factor analysis(LFA) model is capable of conducting efficient representation learning to an HDI matrix,whose hyper-parameter adaptation can be implemented through a particle swarm optimizer(PSO) to meet scalable requirements.However, conventional PSO is limited by its premature issues,which leads to the accuracy loss of a resultant LFA model. To address this thorny issue, this study merges the information of each particle's state migration into its evolution process following the principle of a generalized momentum method for improving its search ability, thereby building a state-migration particle swarm optimizer(SPSO), whose theoretical convergence is rigorously proved in this study. It is then incorporated into an LFA model for implementing efficient hyper-parameter adaptation without accuracy loss. Experiments on six HDI matrices indicate that an SPSO-incorporated LFA model outperforms state-of-the-art LFA models in terms of prediction accuracy for missing data of an HDI matrix with competitive computational efficiency.Hence, SPSO's use ensures efficient and reliable hyper-parameter adaptation in an LFA model, thus ensuring practicality and accurate representation learning for HDI matrices.展开更多
In this paper,we introduce the censored composite conditional quantile coefficient(cC-CQC)to rank the relative importance of each predictor in high-dimensional censored regression.The cCCQC takes advantage of all usef...In this paper,we introduce the censored composite conditional quantile coefficient(cC-CQC)to rank the relative importance of each predictor in high-dimensional censored regression.The cCCQC takes advantage of all useful information across quantiles and can detect nonlinear effects including interactions and heterogeneity,effectively.Furthermore,the proposed screening method based on cCCQC is robust to the existence of outliers and enjoys the sure screening property.Simulation results demonstrate that the proposed method performs competitively on survival datasets of high-dimensional predictors,particularly when the variables are highly correlated.展开更多
The estimation of covariance matrices is very important in many fields, such as statistics. In real applications, data are frequently influenced by high dimensions and noise. However, most relevant studies are based o...The estimation of covariance matrices is very important in many fields, such as statistics. In real applications, data are frequently influenced by high dimensions and noise. However, most relevant studies are based on complete data. This paper studies the optimal estimation of high-dimensional covariance matrices based on missing and noisy sample under the norm. First, the model with sub-Gaussian additive noise is presented. The generalized sample covariance is then modified to define a hard thresholding estimator , and the minimax upper bound is derived. After that, the minimax lower bound is derived, and it is concluded that the estimator presented in this article is rate-optimal. Finally, numerical simulation analysis is performed. The result shows that for missing samples with sub-Gaussian noise, if the true covariance matrix is sparse, the hard thresholding estimator outperforms the traditional estimate method.展开更多
A four-dimensional variational (4D-Var) data assimilation method is implemented in an improved intermediate coupled model (ICM) of the tropical Pacific. A twin experiment is designed to evaluate the impact of the ...A four-dimensional variational (4D-Var) data assimilation method is implemented in an improved intermediate coupled model (ICM) of the tropical Pacific. A twin experiment is designed to evaluate the impact of the 4D-Var data assimilation algorithm on ENSO analysis and prediction based on the ICM. The model error is assumed to arise only from the parameter uncertainty. The "observation" of the SST anomaly, which is sampled from a "truth" model simulation that takes default parameter values and has Gaussian noise added, is directly assimilated into the assimilation model with its parameters set erroneously. Results show that 4D-Var effectively reduces the error of ENSO analysis and therefore improves the prediction skill of ENSO events compared with the non-assimilation case. These results provide a promising way for the ICM to achieve better real-time ENSO prediction.展开更多
Predicting tropical cyclone(TC)genesis is of great societal importance but scientifically challenging.It requires fineresolution coupled models that properly represent air−sea interactions in the atmospheric responses...Predicting tropical cyclone(TC)genesis is of great societal importance but scientifically challenging.It requires fineresolution coupled models that properly represent air−sea interactions in the atmospheric responses to local warm sea surface temperatures and feedbacks,with aid from coherent coupled initialization.This study uses three sets of highresolution regional coupled models(RCMs)covering the Asia−Pacific(AP)region initialized with local observations and dynamically downscaled coupled data assimilation to evaluate the predictability of TC genesis in the West Pacific.The APRCMs consist of three sets of high-resolution configurations of the Weather Research and Forecasting−Regional Ocean Model System(WRF-ROMS):27-km WRF with 9-km ROMS,and 9-km WRF with 3-km ROMS.In this study,a 9-km WRF with 9-km ROMS coupled model system is also used in a case test for the predictability of TC genesis.Since the local sea surface temperatures and wind shear conditions that favor TC formation are better resolved,the enhanced-resolution coupled model tends to improve the predictability of TC genesis,which could be further improved by improving planetary boundary layer physics,thus resolving better air−sea and air−land interactions.展开更多
This paper studies the target controllability of multilayer complex networked systems,in which the nodes are highdimensional linear time invariant(LTI)dynamical systems,and the network topology is directed and weighte...This paper studies the target controllability of multilayer complex networked systems,in which the nodes are highdimensional linear time invariant(LTI)dynamical systems,and the network topology is directed and weighted.The influence of inter-layer couplings on the target controllability of multi-layer networks is discussed.It is found that even if there exists a layer which is not target controllable,the entire multi-layer network can still be target controllable due to the inter-layer couplings.For the multi-layer networks with general structure,a necessary and sufficient condition for target controllability is given by establishing the relationship between uncontrollable subspace and output matrix.By the derived condition,it can be found that the system may be target controllable even if it is not state controllable.On this basis,two corollaries are derived,which clarify the relationship between target controllability,state controllability and output controllability.For the multi-layer networks where the inter-layer couplings are directed chains and directed stars,sufficient conditions for target controllability of networked systems are given,respectively.These conditions are easier to verify than the classic criterion.展开更多
The performance of conventional similarity measurement methods is affected seriously by the curse of dimensionality of high-dimensional data.The reason is that data difference between sparse and noisy dimensionalities...The performance of conventional similarity measurement methods is affected seriously by the curse of dimensionality of high-dimensional data.The reason is that data difference between sparse and noisy dimensionalities occupies a large proportion of the similarity,leading to the dissimilarities between any results.A similarity measurement method of high-dimensional data based on normalized net lattice subspace is proposed.The data range of each dimension is divided into several intervals,and the components in different dimensions are mapped onto the corresponding interval.Only the component in the same or adjacent interval is used to calculate the similarity.To validate this method,three data types are used,and seven common similarity measurement methods are compared.The experimental result indicates that the relative difference of the method is increasing with the dimensionality and is approximately two or three orders of magnitude higher than the conventional method.In addition,the similarity range of this method in different dimensions is [0,1],which is fit for similarity analysis after dimensionality reduction.展开更多
Inductively coupled mooring chain transmission technology plays a crucial role in the long-distance online monitoring of marine hydrographic information.However,the impedance characteristics of the seawater medium pos...Inductively coupled mooring chain transmission technology plays a crucial role in the long-distance online monitoring of marine hydrographic information.However,the impedance characteristics of the seawater medium pose a limitation on its capability to achieve high-speed real-time transmission of underwater images.In this paper,based on the principle of inductively coupled transmission for marine applications,the selected 1920×1080 pixel images are segmented,packaged,and then transmitted to the sending node using the user datagram protocol(UDP),and the modulation and demodulation of the transmitted images are realized using the orthogonal frequency division multiplexing(OFDM)algorithm of least squares(LS)estimation,and eight combinations with different mapping modes and coding rates are tested in the bandwidth of 200-800 k Hz and the quality of the transmitted images is evaluated using three criteria:mean squared error(MSE),peak signal to noise ratio(PSNR),and structural similarity index(SSIM).The results indicate that the image transmission quality is optimal when the mapping method is 16-quadrature amplitude modulation(16QAM),the coding rate is 1/2,and the center frequency is 800 k Hz.Under these conditions,the maximum transmission rate is 0.84 Mbit s^(-1),the SSIM criterion exceeds 0.91690 d B,and the learned perceptual image patch similarity(LPIPS)is less than 0.06000.This paper provides a solution for the underwater image transmission of inductively coupled mooring chains for marine applications.展开更多
The most promising approach for studying soil moisture is the assimilation of observation data and computational modeling. However, there is much uncertainty in the assimilation process, which affects the assimilation...The most promising approach for studying soil moisture is the assimilation of observation data and computational modeling. However, there is much uncertainty in the assimilation process, which affects the assimilation results. This research developed a one-dimensional soil moisture assimilation scheme based on the Ensemble Kalman Filter (EnKF) and Genetic Algorithm (GA). A two-dimensional hydrologic model-Distributed Hydrology-Soil-Vegetation Model (DHSVM) was coupled with a semi-empirical backscattering model (Oh). The Advanced Synthetic Aperture Radar (ASAR) data were assimilated with this coupled model and the field observation data were used to validate this scheme in the soil moisture assimilation experiment. In order to improve the assimilation results, a cost function was set up based on the distance between the simulated backscattering coefficient from the coupled model and the observed backscattering coefficient from ASAR. The EnKF and GA were used to re-initialize and re-parameterize the simulation process, respectively. The assimilation results were compared with the free-run simulations from hydrologic model and the field observation data. The results obtained indicate that this assimilation scheme is practical and it can improve the accuracy of soil moisture estimation significantly.展开更多
In the inductively coupled data transmission system of the mooring buoy, the carrier signal frequency of the transmission channel is limited due to the inherent characteristics of the system, resulting in limited chan...In the inductively coupled data transmission system of the mooring buoy, the carrier signal frequency of the transmission channel is limited due to the inherent characteristics of the system, resulting in limited channel bandwidth. The limited channel bandwidth limits the increase in inductively coupled data transmission rate.In order to improve the inductively coupled data transmission rate of mooring buoy as much as possible without damaging the data transmission performance, a new method was proposed in this paper. The method is proposed to improve the data transmission rate by selecting the appropriate carrier signal frequencies based on the principle of maximizing the amplitude value of amplitude-frequency characteristic curve of the system. Research has been done according to this method as follows. Firstly, according to the inductively coupled transmission mooring buoy structure, the inductively coupled data transmission circuit model was established. The binary frequency shift keying(2FSK) digital signal modulation mode was selected. Through theoretical analysis, the relation between the carrier signal frequency and the data transmission performance, the relation between the carrier signal frequency and the 2FSK signal bandwidth were obtained. Secondly, the performance and the bandwidth of the signal transmission were studied for the inherent characteristics of the actual inductively coupled data transmission system. The amplitude-frequency characteristic of the system was analyzed by experiments. By selecting the appropriate carrier signal frequency parameters, an excellent data transmission performance was guaranteed and a large 2FSK signal bandwidth was obtained. Finally, an inductively coupled data transmission rate optimization experiment and a bit error rate analysis experiment were designed and carried out. The results show that the high-speed and reliable data transmission of the system was realized and the rate can reach 100 kbps.展开更多
In this paper a novel coupled attribute similarity learning method is proposed with the basis on the multi-label categorical data(CASonMLCD).The CASonMLCD method not only computes the correlations between different ...In this paper a novel coupled attribute similarity learning method is proposed with the basis on the multi-label categorical data(CASonMLCD).The CASonMLCD method not only computes the correlations between different attributes and multi-label sets using information gain,which can be regarded as the important degree of each attribute in the attribute learning method,but also further analyzes the intra-coupled and inter-coupled interactions between an attribute value pair for different attributes and multiple labels.The paper compared the CASonMLCD method with the OF distance and Jaccard similarity,which is based on the MLKNN algorithm according to 5common evaluation criteria.The experiment results demonstrated that the CASonMLCD method can mine the similarity relationship more accurately and comprehensively,it can obtain better performance than compared methods.展开更多
Problems existin similarity measurement and index tree construction which affect the performance of nearest neighbor search of high-dimensional data. The equidistance problem is solved using NPsim function to calculat...Problems existin similarity measurement and index tree construction which affect the performance of nearest neighbor search of high-dimensional data. The equidistance problem is solved using NPsim function to calculate similarity. And a sequential NPsim matrix is built to improve indexing performance. To sum up the above innovations,a nearest neighbor search algorithm of high-dimensional data based on sequential NPsim matrix is proposed in comparison with the nearest neighbor search algorithms based on KD-tree or SR-tree on Munsell spectral data set. Experimental results show that the proposed algorithm similarity is better than that of other algorithms and searching speed is more than thousands times of others. In addition,the slow construction speed of sequential NPsim matrix can be increased by using parallel computing.展开更多
The structure-coupled joint inversion method of gravity and magnetic data is a powerful tool for?developing improved physical property models with high resolution and compatible features;?however, the conventional pro...The structure-coupled joint inversion method of gravity and magnetic data is a powerful tool for?developing improved physical property models with high resolution and compatible features;?however, the conventional procedure is inefficient due to the truncated singular values decomposition?(SVD) process at each iteration. To improve the algorithm, a technique using damped leastsquares?is adopted to calculate the structural term of model updates, instead of the truncated SVD. This?produces structural coupled density and magnetization images with high efficiency. A so-called?coupling factor is introduced to regulate the tuning of the desired final structural similarity level.?Synthetic examples show that the joint inversion results are internally consistent and achieve?higher?resolution than separated. The acceptable runtime performance of the damped least squares?technique used in joint inversion indicates that it is more suitable for practical use than the truncated SVD method.展开更多
The datasets for the tier-1 Scenario Model Intercomparison Project(ScenarioMIP)experiments from the Chinese Academy of Sciences(CAS)Flexible Global Ocean-Atmosphere-Land System model,finite-volume version 3(CAS FGOALS...The datasets for the tier-1 Scenario Model Intercomparison Project(ScenarioMIP)experiments from the Chinese Academy of Sciences(CAS)Flexible Global Ocean-Atmosphere-Land System model,finite-volume version 3(CAS FGOALS-f3-L)are described in this study.ScenarioMIP is one of the core MIP experiments in phase 6 of the Coupled Model Intercomparison Project(CMIP6).Considering future CO2,CH4,N2O and other gases’concentrations,as well as land use,the design of ScenarioMIP involves eight pathways,including two tiers(tier-1 and tier-2)of priority.Tier-1 includes four combined Shared Socioeconomic Pathways(SSPs)with radiative forcing,i.e.,SSP1-2.6,SSP2-4.5,SSP3-7.0 and SSP5-8.5,in which the globally averaged radiative forcing at the top of the atmosphere around the year 2100 is approximately 2.6,4.5,7.0 and 8.5 W m−2,respectively.This study provides an introduction to the ScenarioMIP datasets of this model,such as their storage location,sizes,variables,etc.Preliminary analysis indicates that surface air temperatures will increase by about 1.89℃,3.07℃,4.06℃ and 5.17℃ by around 2100 under these four scenarios,respectively.Meanwhile,some other key climate variables,such as sea-ice extension,precipitation,heat content,and sea level rise,also show significant long-term trends associated with the radiative forcing increases.These datasets will help us understand how the climate will change under different anthropogenic and radiative forcings.展开更多
With the motivation to improve the simulation of the East Asian summer monsoon(EASM) in coupled climate models, oceanic data assimilation(DA) was used in CAS-ESM-C(Chinese Academy of Sciences–Earth System Model–Clim...With the motivation to improve the simulation of the East Asian summer monsoon(EASM) in coupled climate models, oceanic data assimilation(DA) was used in CAS-ESM-C(Chinese Academy of Sciences–Earth System Model–Climate Component) in this study. Observed sea surface temperature was assimilated into CAS-ESM-C. The climatology and interannual variability of the EASM simulated in CAS-ESM-C with DA were compared with a traditional AMIP-type run.Results showed that the climatological spatial pattern and annual cycle of precipitation in the western North Paci?c, and the ENSO-related and EASM-related EASM circulation and precipitation, were largely improved. As shown in this study, air–sea coupling is important for EASM simulation. In addition, oceanic DA synchronizes the coupled model with the real world without breaking the air–sea coupling process. These two successful factors make the assimilation experiment a more reasonable experimental design than traditional AMIP-type simulations.展开更多
Viticulturists traditionally have a keen interest in studying the relationship between the biochemistry of grapevines’ leaves/petioles and their associated spectral reflectance in order to understand the fruit ripeni...Viticulturists traditionally have a keen interest in studying the relationship between the biochemistry of grapevines’ leaves/petioles and their associated spectral reflectance in order to understand the fruit ripening rate, water status, nutrient levels, and disease risk. In this paper, we implement imaging spectroscopy (hyperspectral) reflectance data, for the reflective 330 - 2510 nm wavelength region (986 total spectral bands), to assess vineyard nutrient status;this constitutes a high dimensional dataset with a covariance matrix that is ill-conditioned. The identification of the variables (wavelength bands) that contribute useful information for nutrient assessment and prediction, plays a pivotal role in multivariate statistical modeling. In recent years, researchers have successfully developed many continuous, nearly unbiased, sparse and accurate variable selection methods to overcome this problem. This paper compares four regularized and one functional regression methods: Elastic Net, Multi-Step Adaptive Elastic Net, Minimax Concave Penalty, iterative Sure Independence Screening, and Functional Data Analysis for wavelength variable selection. Thereafter, the predictive performance of these regularized sparse models is enhanced using the stepwise regression. This comparative study of regression methods using a high-dimensional and highly correlated grapevine hyperspectral dataset revealed that the performance of Elastic Net for variable selection yields the best predictive ability.展开更多
基金The National Key Research and Development Program of China under contract Nos 2024YFF0808900,2023YFF0805300,and 2020YFA0608804the Civilian Space Programme of China under contract No.D040305.
文摘The El Niño-Southern Oscillation(ENSO)is a naturally recurring interannual climate fluctuation that affects the global climate system.The advent of deep learning-based approaches has led to transformative changes in ENSO forecasts,resulting in significant progress.Most deep learning-based ENSO prediction models which primarily rely solely on reanalysis data may lead to challenges in intensity underestimation in long-term forecasts,reducing the forecasting skills.To this end,we propose a deep residual-coupled model prediction(Res-CMP)model,which integrates historical reanalysis data and coupled model forecast data for multiyear ENSO prediction.The Res-CMP model is designed as a lightweight model that leverages only short-term reanalysis data and nudging assimilation prediction results of the Community Earth System Model(CESM)for effective prediction of the Niño 3.4 index.We also developed a transfer learning strategy for this model to overcome the limitations of inadequate forecast data.After determining the optimal configuration,which included selecting a suitable transfer learning rate during training,along with input variables and CESM forecast lengths,Res-CMP demonstrated a high correlation ability for 19-month lead time predictions(correlation coefficients exceeding 0.5).The Res-CMP model also alleviated the spring predictability barrier(SPB).When validated against actual ENSO events,Res-CMP successfully captured the temporal evolution of the Niño 3.4 index during La Niña events(1998/99 and 2020/21)and El Niño events(2009/10 and 2015/16).Our proposed model has the potential to further enhance ENSO prediction performance by using coupled models to assist deep learning methods.
基金jointly funded by the National Natural Science Foundation of China(NSFC)[grant number 42130608]the China Postdoctoral Science Foundation[grant number 2024M753169]。
文摘Arctic sea ice is an important component of the global climate system and has experienced rapid changes during in the past few decades,the prediction of which is a significant application for climate models.In this study,a Localized Error Subspace Transform Kalman Filter is employed in a coupled climate system model(the Flexible Global Ocean–Atmosphere–Land System Model,version f3-L(FGOALS-f3-L))to assimilate sea-ice concentration(SIC)and sea-ice thickness(SIT)data for melting-season ice predictions.The scheme is applied through the following steps:(1)initialization for generating initial ensembles;(2)analysis for assimilating observed data;(3)adoption for dividing ice states into five thickness categories;(4)forecast for evolving the model;(5)resampling for updating model uncertainties.Several experiments were conducted to examine its results and impacts.Compared with the control experiment,the continuous assimilation experiments(CTNs)indicate assimilations improve model SICs and SITs persistently and generate realistic initials.Assimilating SIC+SIT data better corrects overestimated model SITs spatially than when only assimilating SIC data.The continuous assimilation restart experiments indicate the initials from the CTNs correct the overestimated marginal SICs and overall SITs remarkably well,as well as the cold biases in the oceanic and atmospheric models.The initials with SIC+SIT assimilated show more reasonable spatial improvements.Nevertheless,the SICs in the central Arctic undergo abnormal summer reductions,which is probably because overestimated SITs are reduced in the initials but the strong seasonal cycle(summer melting)biases are unchanged.Therefore,since systematic biases are complicated in a coupled system,for FGOALS-f3-L to make better ice predictions,oceanic and atmospheric assimilations are expected required.
基金supported in part by the Young Scientists Fund of the National Natural Science Foundation of China(Grant Nos.82304253)(and 82273709)the Foundation for Young Talents in Higher Education of Guangdong Province(Grant No.2022KQNCX021)the PhD Starting Project of Guangdong Medical University(Grant No.GDMUB2022054).
文摘Objective Humans are exposed to complex mixtures of environmental chemicals and other factors that can affect their health.Analysis of these mixture exposures presents several key challenges for environmental epidemiology and risk assessment,including high dimensionality,correlated exposure,and subtle individual effects.Methods We proposed a novel statistical approach,the generalized functional linear model(GFLM),to analyze the health effects of exposure mixtures.GFLM treats the effect of mixture exposures as a smooth function by reordering exposures based on specific mechanisms and capturing internal correlations to provide a meaningful estimation and interpretation.The robustness and efficiency was evaluated under various scenarios through extensive simulation studies.Results We applied the GFLM to two datasets from the National Health and Nutrition Examination Survey(NHANES).In the first application,we examined the effects of 37 nutrients on BMI(2011–2016 cycles).The GFLM identified a significant mixture effect,with fiber and fat emerging as the nutrients with the greatest negative and positive effects on BMI,respectively.For the second application,we investigated the association between four pre-and perfluoroalkyl substances(PFAS)and gout risk(2007–2018 cycles).Unlike traditional methods,the GFLM indicated no significant association,demonstrating its robustness to multicollinearity.Conclusion GFLM framework is a powerful tool for mixture exposure analysis,offering improved handling of correlated exposures and interpretable results.It demonstrates robust performance across various scenarios and real-world applications,advancing our understanding of complex environmental exposures and their health impacts on environmental epidemiology and toxicology.
基金supported in part by the National Natural Science Foundation of China (62372385, 62272078, 62002337)the Chongqing Natural Science Foundation (CSTB2022NSCQ-MSX1486, CSTB2023NSCQ-LZX0069)the Deanship of Scientific Research at King Abdulaziz University, Jeddah, Saudi Arabia (RG-12-135-43)。
文摘High-dimensional and incomplete(HDI) matrices are primarily generated in all kinds of big-data-related practical applications. A latent factor analysis(LFA) model is capable of conducting efficient representation learning to an HDI matrix,whose hyper-parameter adaptation can be implemented through a particle swarm optimizer(PSO) to meet scalable requirements.However, conventional PSO is limited by its premature issues,which leads to the accuracy loss of a resultant LFA model. To address this thorny issue, this study merges the information of each particle's state migration into its evolution process following the principle of a generalized momentum method for improving its search ability, thereby building a state-migration particle swarm optimizer(SPSO), whose theoretical convergence is rigorously proved in this study. It is then incorporated into an LFA model for implementing efficient hyper-parameter adaptation without accuracy loss. Experiments on six HDI matrices indicate that an SPSO-incorporated LFA model outperforms state-of-the-art LFA models in terms of prediction accuracy for missing data of an HDI matrix with competitive computational efficiency.Hence, SPSO's use ensures efficient and reliable hyper-parameter adaptation in an LFA model, thus ensuring practicality and accurate representation learning for HDI matrices.
基金Outstanding Youth Foundation of Hunan Provincial Department of Education(Grant No.22B0911)。
文摘In this paper,we introduce the censored composite conditional quantile coefficient(cC-CQC)to rank the relative importance of each predictor in high-dimensional censored regression.The cCCQC takes advantage of all useful information across quantiles and can detect nonlinear effects including interactions and heterogeneity,effectively.Furthermore,the proposed screening method based on cCCQC is robust to the existence of outliers and enjoys the sure screening property.Simulation results demonstrate that the proposed method performs competitively on survival datasets of high-dimensional predictors,particularly when the variables are highly correlated.
文摘The estimation of covariance matrices is very important in many fields, such as statistics. In real applications, data are frequently influenced by high dimensions and noise. However, most relevant studies are based on complete data. This paper studies the optimal estimation of high-dimensional covariance matrices based on missing and noisy sample under the norm. First, the model with sub-Gaussian additive noise is presented. The generalized sample covariance is then modified to define a hard thresholding estimator , and the minimax upper bound is derived. After that, the minimax lower bound is derived, and it is concluded that the estimator presented in this article is rate-optimal. Finally, numerical simulation analysis is performed. The result shows that for missing samples with sub-Gaussian noise, if the true covariance matrix is sparse, the hard thresholding estimator outperforms the traditional estimate method.
基金supported by the National Natural Science Foundation of China(Grant Nos.41490644,41475101 and 41421005)the CAS Strategic Priority Project(the Western Pacific Ocean System+2 种基金Project Nos.XDA11010105,XDA11020306 and XDA11010301)the NSFC-Shandong Joint Fund for Marine Science Research Centers(Grant No.U1406401)the NSFC Innovative Group Grant(Project No.41421005)
文摘A four-dimensional variational (4D-Var) data assimilation method is implemented in an improved intermediate coupled model (ICM) of the tropical Pacific. A twin experiment is designed to evaluate the impact of the 4D-Var data assimilation algorithm on ENSO analysis and prediction based on the ICM. The model error is assumed to arise only from the parameter uncertainty. The "observation" of the SST anomaly, which is sampled from a "truth" model simulation that takes default parameter values and has Gaussian noise added, is directly assimilated into the assimilation model with its parameters set erroneously. Results show that 4D-Var effectively reduces the error of ENSO analysis and therefore improves the prediction skill of ENSO events compared with the non-assimilation case. These results provide a promising way for the ICM to achieve better real-time ENSO prediction.
基金supported by the National Key Research&Development Program of China(Grant Nos.2017YFC1404100 and 2017YFC1404104)the National Natural Science Foundation of China(Grant Nos.41775100 and 41830964)。
文摘Predicting tropical cyclone(TC)genesis is of great societal importance but scientifically challenging.It requires fineresolution coupled models that properly represent air−sea interactions in the atmospheric responses to local warm sea surface temperatures and feedbacks,with aid from coherent coupled initialization.This study uses three sets of highresolution regional coupled models(RCMs)covering the Asia−Pacific(AP)region initialized with local observations and dynamically downscaled coupled data assimilation to evaluate the predictability of TC genesis in the West Pacific.The APRCMs consist of three sets of high-resolution configurations of the Weather Research and Forecasting−Regional Ocean Model System(WRF-ROMS):27-km WRF with 9-km ROMS,and 9-km WRF with 3-km ROMS.In this study,a 9-km WRF with 9-km ROMS coupled model system is also used in a case test for the predictability of TC genesis.Since the local sea surface temperatures and wind shear conditions that favor TC formation are better resolved,the enhanced-resolution coupled model tends to improve the predictability of TC genesis,which could be further improved by improving planetary boundary layer physics,thus resolving better air−sea and air−land interactions.
基金supported by the National Natural Science Foundation of China (U1808205)Hebei Natural Science Foundation (F2000501005)。
文摘This paper studies the target controllability of multilayer complex networked systems,in which the nodes are highdimensional linear time invariant(LTI)dynamical systems,and the network topology is directed and weighted.The influence of inter-layer couplings on the target controllability of multi-layer networks is discussed.It is found that even if there exists a layer which is not target controllable,the entire multi-layer network can still be target controllable due to the inter-layer couplings.For the multi-layer networks with general structure,a necessary and sufficient condition for target controllability is given by establishing the relationship between uncontrollable subspace and output matrix.By the derived condition,it can be found that the system may be target controllable even if it is not state controllable.On this basis,two corollaries are derived,which clarify the relationship between target controllability,state controllability and output controllability.For the multi-layer networks where the inter-layer couplings are directed chains and directed stars,sufficient conditions for target controllability of networked systems are given,respectively.These conditions are easier to verify than the classic criterion.
基金Supported by the National Natural Science Foundation of China(No.61502475)the Importation and Development of High-Caliber Talents Project of the Beijing Municipal Institutions(No.CIT&TCD201504039)
文摘The performance of conventional similarity measurement methods is affected seriously by the curse of dimensionality of high-dimensional data.The reason is that data difference between sparse and noisy dimensionalities occupies a large proportion of the similarity,leading to the dissimilarities between any results.A similarity measurement method of high-dimensional data based on normalized net lattice subspace is proposed.The data range of each dimension is divided into several intervals,and the components in different dimensions are mapped onto the corresponding interval.Only the component in the same or adjacent interval is used to calculate the similarity.To validate this method,three data types are used,and seven common similarity measurement methods are compared.The experimental result indicates that the relative difference of the method is increasing with the dimensionality and is approximately two or three orders of magnitude higher than the conventional method.In addition,the similarity range of this method in different dimensions is [0,1],which is fit for similarity analysis after dimensionality reduction.
基金supported by grants from the National Natural Science Foundation of China(No.62071329)the National Science Foundation of Tianjin(No.23JCZD JC00440)。
文摘Inductively coupled mooring chain transmission technology plays a crucial role in the long-distance online monitoring of marine hydrographic information.However,the impedance characteristics of the seawater medium pose a limitation on its capability to achieve high-speed real-time transmission of underwater images.In this paper,based on the principle of inductively coupled transmission for marine applications,the selected 1920×1080 pixel images are segmented,packaged,and then transmitted to the sending node using the user datagram protocol(UDP),and the modulation and demodulation of the transmitted images are realized using the orthogonal frequency division multiplexing(OFDM)algorithm of least squares(LS)estimation,and eight combinations with different mapping modes and coding rates are tested in the bandwidth of 200-800 k Hz and the quality of the transmitted images is evaluated using three criteria:mean squared error(MSE),peak signal to noise ratio(PSNR),and structural similarity index(SSIM).The results indicate that the image transmission quality is optimal when the mapping method is 16-quadrature amplitude modulation(16QAM),the coding rate is 1/2,and the center frequency is 800 k Hz.Under these conditions,the maximum transmission rate is 0.84 Mbit s^(-1),the SSIM criterion exceeds 0.91690 d B,and the learned perceptual image patch similarity(LPIPS)is less than 0.06000.This paper provides a solution for the underwater image transmission of inductively coupled mooring chains for marine applications.
基金Under the auspices of Major State Basic Research Development Program of China (973 Program) (No. 2007CB714400)the Program of One Hundred Talents of the Chinese Academy of Sciences (No. 99T3005WA2)
文摘The most promising approach for studying soil moisture is the assimilation of observation data and computational modeling. However, there is much uncertainty in the assimilation process, which affects the assimilation results. This research developed a one-dimensional soil moisture assimilation scheme based on the Ensemble Kalman Filter (EnKF) and Genetic Algorithm (GA). A two-dimensional hydrologic model-Distributed Hydrology-Soil-Vegetation Model (DHSVM) was coupled with a semi-empirical backscattering model (Oh). The Advanced Synthetic Aperture Radar (ASAR) data were assimilated with this coupled model and the field observation data were used to validate this scheme in the soil moisture assimilation experiment. In order to improve the assimilation results, a cost function was set up based on the distance between the simulated backscattering coefficient from the coupled model and the observed backscattering coefficient from ASAR. The EnKF and GA were used to re-initialize and re-parameterize the simulation process, respectively. The assimilation results were compared with the free-run simulations from hydrologic model and the field observation data. The results obtained indicate that this assimilation scheme is practical and it can improve the accuracy of soil moisture estimation significantly.
基金supported by the National Natural Science Foundation of China [Grant number 61733012]Qingdao Ocean Engineering and Technology Think Tank Joint Fund Project [Grant number 20190131-2]the Shandong Provincial Natural Science Fund Project [Grant number ZR2017MEE072]。
文摘In the inductively coupled data transmission system of the mooring buoy, the carrier signal frequency of the transmission channel is limited due to the inherent characteristics of the system, resulting in limited channel bandwidth. The limited channel bandwidth limits the increase in inductively coupled data transmission rate.In order to improve the inductively coupled data transmission rate of mooring buoy as much as possible without damaging the data transmission performance, a new method was proposed in this paper. The method is proposed to improve the data transmission rate by selecting the appropriate carrier signal frequencies based on the principle of maximizing the amplitude value of amplitude-frequency characteristic curve of the system. Research has been done according to this method as follows. Firstly, according to the inductively coupled transmission mooring buoy structure, the inductively coupled data transmission circuit model was established. The binary frequency shift keying(2FSK) digital signal modulation mode was selected. Through theoretical analysis, the relation between the carrier signal frequency and the data transmission performance, the relation between the carrier signal frequency and the 2FSK signal bandwidth were obtained. Secondly, the performance and the bandwidth of the signal transmission were studied for the inherent characteristics of the actual inductively coupled data transmission system. The amplitude-frequency characteristic of the system was analyzed by experiments. By selecting the appropriate carrier signal frequency parameters, an excellent data transmission performance was guaranteed and a large 2FSK signal bandwidth was obtained. Finally, an inductively coupled data transmission rate optimization experiment and a bit error rate analysis experiment were designed and carried out. The results show that the high-speed and reliable data transmission of the system was realized and the rate can reach 100 kbps.
基金Supported by Australian Research Council Discovery(DP130102691)the National Science Foundation of China(61302157)+1 种基金China National 863 Project(2012AA12A308)China Pre-research Project of Nuclear Industry(FZ1402-08)
文摘In this paper a novel coupled attribute similarity learning method is proposed with the basis on the multi-label categorical data(CASonMLCD).The CASonMLCD method not only computes the correlations between different attributes and multi-label sets using information gain,which can be regarded as the important degree of each attribute in the attribute learning method,but also further analyzes the intra-coupled and inter-coupled interactions between an attribute value pair for different attributes and multiple labels.The paper compared the CASonMLCD method with the OF distance and Jaccard similarity,which is based on the MLKNN algorithm according to 5common evaluation criteria.The experiment results demonstrated that the CASonMLCD method can mine the similarity relationship more accurately and comprehensively,it can obtain better performance than compared methods.
基金Supported by the National Natural Science Foundation of China(No.61300078)the Importation and Development of High-Caliber Talents Project of Beijing Municipal Institutions(No.CIT&TCD201504039)+1 种基金Funding Project for Academic Human Resources Development in Beijing Union University(No.BPHR2014A03,Rk100201510)"New Start"Academic Research Projects of Beijing Union University(No.Hzk10201501)
文摘Problems existin similarity measurement and index tree construction which affect the performance of nearest neighbor search of high-dimensional data. The equidistance problem is solved using NPsim function to calculate similarity. And a sequential NPsim matrix is built to improve indexing performance. To sum up the above innovations,a nearest neighbor search algorithm of high-dimensional data based on sequential NPsim matrix is proposed in comparison with the nearest neighbor search algorithms based on KD-tree or SR-tree on Munsell spectral data set. Experimental results show that the proposed algorithm similarity is better than that of other algorithms and searching speed is more than thousands times of others. In addition,the slow construction speed of sequential NPsim matrix can be increased by using parallel computing.
文摘The structure-coupled joint inversion method of gravity and magnetic data is a powerful tool for?developing improved physical property models with high resolution and compatible features;?however, the conventional procedure is inefficient due to the truncated singular values decomposition?(SVD) process at each iteration. To improve the algorithm, a technique using damped leastsquares?is adopted to calculate the structural term of model updates, instead of the truncated SVD. This?produces structural coupled density and magnetization images with high efficiency. A so-called?coupling factor is introduced to regulate the tuning of the desired final structural similarity level.?Synthetic examples show that the joint inversion results are internally consistent and achieve?higher?resolution than separated. The acceptable runtime performance of the damped least squares?technique used in joint inversion indicates that it is more suitable for practical use than the truncated SVD method.
基金supported by the Strategic Priority Research Program of the Chinese Academy of Sciences(Grant Nos.XDA19060102 and XDB42000000)the National Natural Science Foundation of China(Grants Nos.41530426 and 91958201)。
文摘The datasets for the tier-1 Scenario Model Intercomparison Project(ScenarioMIP)experiments from the Chinese Academy of Sciences(CAS)Flexible Global Ocean-Atmosphere-Land System model,finite-volume version 3(CAS FGOALS-f3-L)are described in this study.ScenarioMIP is one of the core MIP experiments in phase 6 of the Coupled Model Intercomparison Project(CMIP6).Considering future CO2,CH4,N2O and other gases’concentrations,as well as land use,the design of ScenarioMIP involves eight pathways,including two tiers(tier-1 and tier-2)of priority.Tier-1 includes four combined Shared Socioeconomic Pathways(SSPs)with radiative forcing,i.e.,SSP1-2.6,SSP2-4.5,SSP3-7.0 and SSP5-8.5,in which the globally averaged radiative forcing at the top of the atmosphere around the year 2100 is approximately 2.6,4.5,7.0 and 8.5 W m−2,respectively.This study provides an introduction to the ScenarioMIP datasets of this model,such as their storage location,sizes,variables,etc.Preliminary analysis indicates that surface air temperatures will increase by about 1.89℃,3.07℃,4.06℃ and 5.17℃ by around 2100 under these four scenarios,respectively.Meanwhile,some other key climate variables,such as sea-ice extension,precipitation,heat content,and sea level rise,also show significant long-term trends associated with the radiative forcing increases.These datasets will help us understand how the climate will change under different anthropogenic and radiative forcings.
基金supported by the Strategic Priority Research Program of the Chinese Academy of Sciences [grant number XDA19030403]the National Natural Science Foundation of China [grant numbers 41606027 and 41706028]+1 种基金the National Key R&D Program of China [grant number2017YFA0604201]the China Postdoctoral Science Foundation [grant number 2015M571095]
文摘With the motivation to improve the simulation of the East Asian summer monsoon(EASM) in coupled climate models, oceanic data assimilation(DA) was used in CAS-ESM-C(Chinese Academy of Sciences–Earth System Model–Climate Component) in this study. Observed sea surface temperature was assimilated into CAS-ESM-C. The climatology and interannual variability of the EASM simulated in CAS-ESM-C with DA were compared with a traditional AMIP-type run.Results showed that the climatological spatial pattern and annual cycle of precipitation in the western North Paci?c, and the ENSO-related and EASM-related EASM circulation and precipitation, were largely improved. As shown in this study, air–sea coupling is important for EASM simulation. In addition, oceanic DA synchronizes the coupled model with the real world without breaking the air–sea coupling process. These two successful factors make the assimilation experiment a more reasonable experimental design than traditional AMIP-type simulations.
文摘Viticulturists traditionally have a keen interest in studying the relationship between the biochemistry of grapevines’ leaves/petioles and their associated spectral reflectance in order to understand the fruit ripening rate, water status, nutrient levels, and disease risk. In this paper, we implement imaging spectroscopy (hyperspectral) reflectance data, for the reflective 330 - 2510 nm wavelength region (986 total spectral bands), to assess vineyard nutrient status;this constitutes a high dimensional dataset with a covariance matrix that is ill-conditioned. The identification of the variables (wavelength bands) that contribute useful information for nutrient assessment and prediction, plays a pivotal role in multivariate statistical modeling. In recent years, researchers have successfully developed many continuous, nearly unbiased, sparse and accurate variable selection methods to overcome this problem. This paper compares four regularized and one functional regression methods: Elastic Net, Multi-Step Adaptive Elastic Net, Minimax Concave Penalty, iterative Sure Independence Screening, and Functional Data Analysis for wavelength variable selection. Thereafter, the predictive performance of these regularized sparse models is enhanced using the stepwise regression. This comparative study of regression methods using a high-dimensional and highly correlated grapevine hyperspectral dataset revealed that the performance of Elastic Net for variable selection yields the best predictive ability.