AIM:To investigate the clinical features and prognosis of patients with orbital inflammatory myofibroblastic tumor(IMT).METHODS:This retrospective study collected clinical data from 22 patients diagnosed with orbital ...AIM:To investigate the clinical features and prognosis of patients with orbital inflammatory myofibroblastic tumor(IMT).METHODS:This retrospective study collected clinical data from 22 patients diagnosed with orbital IMT based on histopathological examination.The patients were followed up to assess their prognosis.Clinical data from patients,including age,gender,course of disease,past medical history,primary symptoms,ophthalmologic examination findings,general condition,as well as imaging,laboratory,histopathological,and immunohistochemical results from digital records were collected.Orbital magnetic resonance imaging(MRI)and(or)computed tomography(CT)scans were performed to assess bone destruction of the mass,invasion of surrounding tissues,and any inflammatory changes in periorbital areas.RESULTS:The mean age of patients with orbital IMT was 28.24±3.30y,with a male-to-female ratio of 1.2:1.Main clinical manifestations were proptosis,blurred vision,palpable mass,and pain.Bone destruction and surrounding tissue invasion occurred in 72.73%and 54.55%of cases,respectively.Inflammatory changes in the periorbital site were observed in 77.27%of the patients.Hematoxylin and eosin staining showed proliferation of fibroblasts and myofibroblasts,accompanied by infiltration of lymphocytes and plasma cells.Immunohistochemical staining revealed that smooth muscle actin(SMA)and vimentin were positive in 100%of cases,while anaplastic lymphoma kinase(ALK)showed positivity in 47.37%.The recurrence rate of orbital IMT was 27.27%,and sarcomatous degeneration could occur.There were no significant correlations between recurrence and factors such as age,gender,laterality,duration of the disease,periorbital tissue invasion,bone destruction,periorbital inflammation,tumor size,fever,leukocytosis,or treatment(P>0.05).However,lymphadenopathy and a Ki-67 index of 10%or higher may be risk factors for recurrence(P=0.046;P=0.023).CONCLUSION:Orbital IMT is a locally invasive disease that may recur or lead to sarcomatoid degeneration,primarily affecting young and middle-aged patients.The presence of lymphadenopathy and a Ki-67 index of 10%or higher may signify a poor prognosis.展开更多
BACKGROUND SMARCB1/INI1-deficient pancreatic undifferentiated rhabdoid carcinoma is a highly aggressive tumor,and spontaneous splenic rupture(SSR)as its presenting manifestation is rarely reported among pancreatic mal...BACKGROUND SMARCB1/INI1-deficient pancreatic undifferentiated rhabdoid carcinoma is a highly aggressive tumor,and spontaneous splenic rupture(SSR)as its presenting manifestation is rarely reported among pancreatic malignancies.CASE SUMMARY We herein report a rare case of a 59-year-old female who presented with acute left upper quadrant abdominal pain without any history of trauma.Abdominal imaging demonstrated a heterogeneous splenic lesion with hemoperitoneum,raising clinical suspicion of SSR.Emergency laparotomy revealed a pancreatic tumor invading the spleen and left kidney,with associated splenic rupture and dense adhesions,necessitating en bloc resection of the distal pancreas,spleen,and left kidney.Histopathology revealed a biphasic malignancy composed of moderately differentiated pancreatic ductal adenocarcinoma and an undifferentiated carcinoma with rhabdoid morphology and loss of SMARCB1 expression.Immunohistochemical analysis confirmed complete loss of SMARCB1/INI1 in the undifferentiated component,along with a high Ki-67 index(approximately 80%)and CD10 positivity.The ductal adenocarcinoma component retained SMARCB1/INI1 expression and was positive for CK7 and CK-pan.Transitional zones between the two tumor components suggested progressive dedifferentiation and underlying genomic instability.The patient received adjuvant chemotherapy with gemcitabine and nab-paclitaxel and maintained a satisfactory quality of life at the 6-month follow-up.CONCLUSION This study reports a rare case of SMARCB1/INI1-deficient undifferentiated rhabdoid carcinoma of the pancreas combined with ductal adenocarcinoma,presenting as SSR-an exceptionally uncommon initial manifestation of pancreatic malignancy.展开更多
In the article“A Lightweight Approach for Skin Lesion Detection through Optimal Features Fusion”by Khadija Manzoor,Fiaz Majeed,Ansar Siddique,Talha Meraj,Hafiz Tayyab Rauf,Mohammed A.El-Meligy,Mohamed Sharaf,Abd Ela...In the article“A Lightweight Approach for Skin Lesion Detection through Optimal Features Fusion”by Khadija Manzoor,Fiaz Majeed,Ansar Siddique,Talha Meraj,Hafiz Tayyab Rauf,Mohammed A.El-Meligy,Mohamed Sharaf,Abd Elatty E.Abd Elgawad Computers,Materials&Continua,2022,Vol.70,No.1,pp.1617–1630.DOI:10.32604/cmc.2022.018621,URL:https://www.techscience.com/cmc/v70n1/44361,there was an error regarding the affiliation for the author Hafiz Tayyab Rauf.Instead of“Centre for Smart Systems,AI and Cybersecurity,Staffordshire University,Stoke-on-Trent,UK”,the affiliation should be“Independent Researcher,Bradford,BD80HS,UK”.展开更多
BACKGROUND Pancreatic cancer remains one of the most lethal malignancies worldwide,with a poor prognosis often attributed to late diagnosis.Understanding the correlation between pathological type and imaging features ...BACKGROUND Pancreatic cancer remains one of the most lethal malignancies worldwide,with a poor prognosis often attributed to late diagnosis.Understanding the correlation between pathological type and imaging features is crucial for early detection and appropriate treatment planning.AIM To retrospectively analyze the relationship between different pathological types of pancreatic cancer and their corresponding imaging features.METHODS We retrospectively analyzed the data of 500 patients diagnosed with pancreatic cancer between January 2010 and December 2020 at our institution.Pathological types were determined by histopathological examination of the surgical spe-cimens or biopsy samples.The imaging features were assessed using computed tomography,magnetic resonance imaging,and endoscopic ultrasound.Statistical analyses were performed to identify significant associations between pathological types and specific imaging characteristics.RESULTS There were 320(64%)cases of pancreatic ductal adenocarcinoma,75(15%)of intraductal papillary mucinous neoplasms,50(10%)of neuroendocrine tumors,and 55(11%)of other rare types.Distinct imaging features were identified in each pathological type.Pancreatic ductal adenocarcinoma typically presents as a hypodense mass with poorly defined borders on computed tomography,whereas intraductal papillary mucinous neoplasms present as characteristic cystic lesions with mural nodules.Neuroendocrine tumors often appear as hypervascular lesions in contrast-enhanced imaging.Statistical analysis revealed significant correlations between specific imaging features and pathological types(P<0.001).CONCLUSION This study demonstrated a strong association between the pathological types of pancreatic cancer and imaging features.These findings can enhance the accuracy of noninvasive diagnosis and guide personalized treatment approaches.展开更多
During Donald Trump’s first term,the“Trump Shock”brought world politics into an era of uncertainties and pulled the transatlantic alliance down to its lowest point in history.The Trump 2.0 tsunami brewed by the 202...During Donald Trump’s first term,the“Trump Shock”brought world politics into an era of uncertainties and pulled the transatlantic alliance down to its lowest point in history.The Trump 2.0 tsunami brewed by the 2024 presidential election of the United States has plunged the U.S.-Europe relations into more gloomy waters,ushering in a more complex and turbulent period of adjustment.展开更多
The fusion of infrared and visible images should emphasize the salient targets in the infrared image while preserving the textural details of the visible images.To meet these requirements,an autoencoder-based method f...The fusion of infrared and visible images should emphasize the salient targets in the infrared image while preserving the textural details of the visible images.To meet these requirements,an autoencoder-based method for infrared and visible image fusion is proposed.The encoder designed according to the optimization objective consists of a base encoder and a detail encoder,which is used to extract low-frequency and high-frequency information from the image.This extraction may lead to some information not being captured,so a compensation encoder is proposed to supplement the missing information.Multi-scale decomposition is also employed to extract image features more comprehensively.The decoder combines low-frequency,high-frequency and supplementary information to obtain multi-scale features.Subsequently,the attention strategy and fusion module are introduced to perform multi-scale fusion for image reconstruction.Experimental results on three datasets show that the fused images generated by this network effectively retain salient targets while being more consistent with human visual perception.展开更多
Smart contracts are widely used on the blockchain to implement complex transactions,such as decentralized applications on Ethereum.Effective vulnerability detection of large-scale smart contracts is critical,as attack...Smart contracts are widely used on the blockchain to implement complex transactions,such as decentralized applications on Ethereum.Effective vulnerability detection of large-scale smart contracts is critical,as attacks on smart contracts often cause huge economic losses.Since it is difficult to repair and update smart contracts,it is necessary to find the vulnerabilities before they are deployed.However,code analysis,which requires traversal paths,and learning methods,which require many features to be trained,are too time-consuming to detect large-scale on-chain contracts.Learning-based methods will obtain detection models from a feature space compared to code analysis methods such as symbol execution.But the existing features lack the interpretability of the detection results and training model,even worse,the large-scale feature space also affects the efficiency of detection.This paper focuses on improving the detection efficiency by reducing the dimension of the features,combined with expert knowledge.In this paper,a feature extraction model Block-gram is proposed to form low-dimensional knowledge-based features from bytecode.First,the metadata is separated and the runtime code is converted into a sequence of opcodes,which are divided into segments based on some instructions(jumps,etc.).Then,scalable Block-gram features,including 4-dimensional block features and 8-dimensional attribute features,are mined for the learning-based model training.Finally,feature contributions are calculated from SHAP values to measure the relationship between our features and the results of the detection model.In addition,six types of vulnerability labels are made on a dataset containing 33,885 contracts,and these knowledge-based features are evaluated using seven state-of-the-art learning algorithms,which show that the average detection latency speeds up 25×to 650×,compared with the features extracted by N-gram,and also can enhance the interpretability of the detection model.展开更多
Fraudulent website is an important car-rier tool for telecom fraud.At present,criminals can use artificial intelligence generative content technol-ogy to quickly generate fraudulent website templates and build fraudul...Fraudulent website is an important car-rier tool for telecom fraud.At present,criminals can use artificial intelligence generative content technol-ogy to quickly generate fraudulent website templates and build fraudulent websites in batches.Accurate identification of fraudulent website will effectively re-duce the risk of public victimization.Therefore,this study developed a fraudulent website template iden-tification method based on DOM structure extraction of website fingerprint features,which solves the prob-lems of single-dimension identification,low accuracy,and the insufficient generalization ability of current fraudulent website templates.This method uses an im-proved SimHash algorithm to traverse the DOM tree of a webpage,extract website node features,calcu-late the weight of each node,and obtain the finger-print feature vector of the website through dimension-ality reduction.Finally,the random forest algorithm is used to optimize the training features for the best combination of parameters.This method automati-cally extracts fingerprint features from websites and identifies website template ownership based on these features.An experimental analysis showed that this method achieves a classification accuracy of 89.8%and demonstrates superior recognition.展开更多
Biometric characteristics are playing a vital role in security for the last few years.Human gait classification in video sequences is an important biometrics attribute and is used for security purposes.A new framework...Biometric characteristics are playing a vital role in security for the last few years.Human gait classification in video sequences is an important biometrics attribute and is used for security purposes.A new framework for human gait classification in video sequences using deep learning(DL)fusion assisted and posterior probability-based moth flames optimization(MFO)is proposed.In the first step,the video frames are resized and finetuned by two pre-trained lightweight DL models,EfficientNetB0 and MobileNetV2.Both models are selected based on the top-5 accuracy and less number of parameters.Later,both models are trained through deep transfer learning and extracted deep features fused using a voting scheme.In the last step,the authors develop a posterior probabilitybased MFO feature selection algorithm to select the best features.The selected features are classified using several supervised learning methods.The CASIA-B publicly available dataset has been employed for the experimental process.On this dataset,the authors selected six angles such as 0°,18°,90°,108°,162°,and 180°and obtained an average accuracy of 96.9%,95.7%,86.8%,90.0%,95.1%,and 99.7%.Results demonstrate comparable improvement in accuracy and significantly minimize the computational time with recent state-of-the-art techniques.展开更多
Due to the limitations of existing imaging hardware, obtaining high-resolution hyperspectral images is challenging. Hyperspectral image super-resolution(HSI SR) has been a very attractive research topic in computer vi...Due to the limitations of existing imaging hardware, obtaining high-resolution hyperspectral images is challenging. Hyperspectral image super-resolution(HSI SR) has been a very attractive research topic in computer vision, attracting the attention of many researchers. However, most HSI SR methods focus on the tradeoff between spatial resolution and spectral information, and cannot guarantee the efficient extraction of image information. In this paper, a multidimensional features network(MFNet) for HSI SR is proposed, which simultaneously learns and fuses the spatial,spectral, and frequency multidimensional features of HSI. Spatial features contain rich local details,spectral features contain the information and correlation between spectral bands, and frequency feature can reflect the global information of the image and can be used to obtain the global context of HSI. The fusion of the three features can better guide image super-resolution, to obtain higher-quality high-resolution hyperspectral images. In MFNet, we use the frequency feature extraction module(FFEM) to extract the frequency feature. On this basis, a multidimensional features extraction module(MFEM) is designed to learn and fuse multidimensional features. In addition, experimental results on two public datasets demonstrate that MFNet achieves state-of-the-art performance.展开更多
Recognizing road scene context from a single image remains a critical challenge for intelligent autonomous driving systems,particularly in dynamic and unstructured environments.While recent advancements in deep learni...Recognizing road scene context from a single image remains a critical challenge for intelligent autonomous driving systems,particularly in dynamic and unstructured environments.While recent advancements in deep learning have significantly enhanced road scene classification,simultaneously achieving high accuracy,computational efficiency,and adaptability across diverse conditions continues to be difficult.To address these challenges,this study proposes HybridLSTM,a novel and efficient framework that integrates deep learning-based,object-based,and handcrafted feature extraction methods within a unified architecture.HybridLSTM is designed to classify four distinct road scene categories—crosswalk(CW),highway(HW),overpass/tunnel(OP/T),and parking(P)—by leveraging multiple publicly available datasets,including Places-365,BDD100K,LabelMe,and KITTI,thereby promoting domain generalization.The framework fuses object-level features extracted using YOLOv5 and VGG19,scene-level global representations obtained from a modified VGG19,and fine-grained texture features captured through eight handcrafted descriptors.This hybrid feature fusion enables the model to capture both semantic context and low-level visual cues,which are critical for robust scene understanding.To model spatial arrangements and latent sequential dependencies present even in static imagery,the combined features are processed through a Long Short-Term Memory(LSTM)network,allowing the extraction of discriminative patterns across heterogeneous feature spaces.Extensive experiments conducted on 2725 annotated road scene images,with an 80:20 training-to-testing split,validate the effectiveness of the proposed model.HybridLSTM achieves a classification accuracy of 96.3%,a precision of 95.8%,a recall of 96.1%,and an F1-score of 96.0%,outperforming several existing state-of-the-art methods.These results demonstrate the robustness,scalability,and generalization capability of HybridLSTM across varying environments and scene complexities.Moreover,the framework is optimized to balance classification performance with computational efficiency,making it highly suitable for real-time deployment in embedded autonomous driving systems.Future work will focus on extending the model to multi-class detection within a single frame and optimizing it further for edge-device deployments to reduce computational overhead in practical applications.展开更多
Drug resistance remains a major challenge in breast cancer chemotherapy,yet the metabolic alterations underlying this phenomenon are not fully understood.There is much evidence indicating the cellular heterogeneity am...Drug resistance remains a major challenge in breast cancer chemotherapy,yet the metabolic alterations underlying this phenomenon are not fully understood.There is much evidence indicating the cellular heterogeneity among cancer cells,which exhibit varying degrees of metabolic reprogramming and thus may result in differential contributions to drug resistance.A home-built single-cell quantitative mass spectrometry(MS)platform,which integrates micromanipulation and electro-osmotic sampling,was developed to quantitatively profile the tricarboxylic acid(TCA)cycle metabolites at the single-cell level.Using this platform,the metabolic profiles of drug-sensitive MCF-7 breast cancer cells and their drug-resistant derivative MCF-7/ADR cells were compared.This results revealed a selective upregulation of downstream TCA cycle metabolites includingα-ketoglutarate,succinate,fumarate,and malate in drug-resistant cancer cells,while early TCA metabolites remained largely unchanged.Furthermore,notable variations in the abundance of the metabolites were observed in individual cells.The comparative analysis also revealed that not all MCF-7/ADR cells exhibit the same degree of metabolic deviation from the parental line in the metabolites during resistance acquisition.The observed metabolic profiles indicate enhanced glutaminolysis,altered mitochondrial electron transport chain activity,and increased metabolic flexibility in drug-resistant cancer cells that support their survival under chemotherapeutic stress.The findings further suggest the potential for incorporating cellular metabolic heterogeneity into future drug resistance studies.展开更多
Multi-Object Tracking(MOT)represents a fundamental but computationally demanding task in computer vision,with particular challenges arising in occluded and densely populated environments.While contemporary tracking sy...Multi-Object Tracking(MOT)represents a fundamental but computationally demanding task in computer vision,with particular challenges arising in occluded and densely populated environments.While contemporary tracking systems have demonstrated considerable progress,persistent limitations—notably frequent occlusion-induced identity switches and tracking inaccuracies—continue to impede reliable real-world deployment.This work introduces an advanced tracking framework that enhances association robustness through a two-stage matching paradigm combining spatial and appearance features.Proposed framework employs:(1)a Height Modulated and Scale Adaptive Spatial Intersection-over-Union(HMSIoU)metric for improved spatial correspondence estimation across variable object scales and partial occlusions;(2)a feature extraction module generating discriminative appearance descriptors for identity maintenance;and(3)a recovery association mechanism for refining matches between unassociated tracks and detections.Comprehensive evaluation on standard MOT17 and MOT20 benchmarks demonstrates significant improvements in tracking consistency,with state-of-the-art performance across key metrics including HOTA(64),MOTA(80.7),IDF1(79.8),and IDs(1379).These results substantiate the efficacy of our Cue-Tracker framework in complex real-world scenarios characterized by occlusions and crowd interactions.展开更多
Cyberbullying on social media poses significant psychological risks,yet most detection systems over-simplify the task by focusing on binary classification,ignoring nuanced categories like passive-aggressive remarks or...Cyberbullying on social media poses significant psychological risks,yet most detection systems over-simplify the task by focusing on binary classification,ignoring nuanced categories like passive-aggressive remarks or indirect slurs.To address this gap,we propose a hybrid framework combining Term Frequency-Inverse Document Frequency(TF-IDF),word-to-vector(Word2Vec),and Bidirectional Encoder Representations from Transformers(BERT)based models for multi-class cyberbullying detection.Our approach integrates TF-IDF for lexical specificity and Word2Vec for semantic relationships,fused with BERT’s contextual embeddings to capture syntactic and semantic complexities.We evaluate the framework on a publicly available dataset of 47,000 annotated social media posts across five cyberbullying categories:age,ethnicity,gender,religion,and indirect aggression.Among BERT variants tested,BERT Base Un-Cased achieved the highest performance with 93%accuracy(standard deviation across±1%5-fold cross-validation)and an average AUC of 0.96,outperforming standalone TF-IDF(78%)and Word2Vec(82%)models.Notably,it achieved near-perfect AUC scores(0.99)for age and ethnicity-based bullying.A comparative analysis with state-of-the-art benchmarks,including Generative Pre-trained Transformer 2(GPT-2)and Text-to-Text Transfer Transformer(T5)models highlights BERT’s superiority in handling ambiguous language.This work advances cyberbullying detection by demonstrating how hybrid feature extraction and transformer models improve multi-class classification,offering a scalable solution for moderating nuanced harmful content.展开更多
Remote sensing cross-modal image-text retrieval(RSCIR)can flexibly and subjectively retrieve remote sensing images utilizing query text,which has received more researchers’attention recently.However,with the increasi...Remote sensing cross-modal image-text retrieval(RSCIR)can flexibly and subjectively retrieve remote sensing images utilizing query text,which has received more researchers’attention recently.However,with the increasing volume of visual-language pre-training model parameters,direct transfer learning consumes a substantial amount of computational and storage resources.Moreover,recently proposed parameter-efficient transfer learning methods mainly focus on the reconstruction of channel features,ignoring the spatial features which are vital for modeling key entity relationships.To address these issues,we design an efficient transfer learning framework for RSCIR,which is based on spatial feature efficient reconstruction(SPER).A concise and efficient spatial adapter is introduced to enhance the extraction of spatial relationships.The spatial adapter is able to spatially reconstruct the features in the backbone with few parameters while incorporating the prior information from the channel dimension.We conduct quantitative and qualitative experiments on two different commonly used RSCIR datasets.Compared with traditional methods,our approach achieves an improvement of 3%-11% in sumR metric.Compared with methods finetuning all parameters,our proposed method only trains less than 1% of the parameters,while maintaining an overall performance of about 96%.展开更多
Malaria is considered one of the major causes of travel-related morbidity and mortality,especially among non-immune travelers from non-endemic countries to the endemic regions.According to a multicenter study from the...Malaria is considered one of the major causes of travel-related morbidity and mortality,especially among non-immune travelers from non-endemic countries to the endemic regions.According to a multicenter study from the GeoSentinel surveillance network,malaria was the most frequent cause of fever in 21%of returning travelers,followed by dengue,typhoid fever,chikungunya and rickettsiosis[1].Individuals traveling from regions without malaria transmission to areas where it is endemic face a heightened risk of contracting the disease due to their lack of immunity.Despite the official malaria-free status of the Russian Federation since 2010,annual cases of severe Plasmodium(P.)falciparum malaria continue to be reported[2].This underscores the necessity for heightened clinical vigilance and improved preventive strategies especially in non-endemic settings.展开更多
Spartina alterniflora is now listed among the world’s 100 most dangerous invasive species,severely affecting the ecological balance of coastal wetlands.Remote sensing technologies based on deep learning enable large-...Spartina alterniflora is now listed among the world’s 100 most dangerous invasive species,severely affecting the ecological balance of coastal wetlands.Remote sensing technologies based on deep learning enable large-scale monitoring of Spartina alterniflora,but they require large datasets and have poor interpretability.A new method is proposed to detect Spartina alterniflora from Sentinel-2 imagery.Firstly,to get the high canopy cover and dense community characteristics of Spartina alterniflora,multi-dimensional shallow features are extracted from the imagery.Secondly,to detect different objects from satellite imagery,index features are extracted,and the statistical features of the Gray-Level Co-occurrence Matrix(GLCM)are derived using principal component analysis.Then,ensemble learning methods,including random forest,extreme gradient boosting,and light gradient boosting machine models,are employed for image classification.Meanwhile,Recursive Feature Elimination with Cross-Validation(RFECV)is used to select the best feature subset.Finally,to enhance the interpretability of the models,the best features are utilized to classify multi-temporal images and SHapley Additive exPlanations(SHAP)is combined with these classifications to explain the model prediction process.The method is validated by using Sentinel-2 imageries and previous observations of Spartina alterniflora in Chongming Island,it is found that the model combining image texture features such as GLCM covariance can significantly improve the detection accuracy of Spartina alterniflora by about 8%compared with the model without image texture features.Through multiple model comparisons and feature selection via RFECV,the selected model and eight features demonstrated good classification accuracy when applied to data from different time periods,proving that feature reduction can effectively enhance model generalization.Additionally,visualizing model decisions using SHAP revealed that the image texture feature component_1_GLCMVariance is particularly important for identifying each land cover type.展开更多
Image captioning,the task of generating descriptive sentences for images,has advanced significantly with the integration of semantic information.However,traditional models still rely on static visual features that do ...Image captioning,the task of generating descriptive sentences for images,has advanced significantly with the integration of semantic information.However,traditional models still rely on static visual features that do not evolve with the changing linguistic context,which can hinder the ability to form meaningful connections between the image and the generated captions.This limitation often leads to captions that are less accurate or descriptive.In this paper,we propose a novel approach to enhance image captioning by introducing dynamic interactions where visual features continuously adapt to the evolving linguistic context.Our model strengthens the alignment between visual and linguistic elements,resulting in more coherent and contextually appropriate captions.Specifically,we introduce two innovative modules:the Visual Weighting Module(VWM)and the Enhanced Features Attention Module(EFAM).The VWM adjusts visual features using partial attention,enabling dynamic reweighting of the visual inputs,while the EFAM further refines these features to improve their relevance to the generated caption.By continuously adjusting visual features in response to the linguistic context,our model bridges the gap between static visual features and dynamic language generation.We demonstrate the effectiveness of our approach through experiments on the MS-COCO dataset,where our method outperforms state-of-the-art techniques in terms of caption quality and contextual relevance.Our results show that dynamic visual-linguistic alignment significantly enhances image captioning performance.展开更多
In July 2021,a catastrophic extreme precipitation(EP)event occurred in Henan Province,China,resulting in considerable human and economic losses.The synoptic pattern during this event is distinctive,characterized by th...In July 2021,a catastrophic extreme precipitation(EP)event occurred in Henan Province,China,resulting in considerable human and economic losses.The synoptic pattern during this event is distinctive,characterized by the presence of two typhoons and substantial water transport into Henan.However,a favorable synoptic pattern only does not guarantee the occurrence of heavy precipitation in Henan.This study investigates the key environmental features critical for EP under similar synoptic patterns to the 2021 Henan extreme event.It is found that cold clouds are better aggregated on EP days,accompanied by beneficial environment features like enhanced moisture conditions,stronger updrafts,and greater atmospheric instability.The temporal evolution of these environmental features shows a leading signal by one to three days.These results suggest the importance of combining the synoptic pattern and environmental features in the forecasting of heavy precipitation events.展开更多
Objective To explore the feasibility of constructing a lung cancer early-warning risk model based on facial image features,providing novel insights into the early screening of lung cancer.Methods This study included p...Objective To explore the feasibility of constructing a lung cancer early-warning risk model based on facial image features,providing novel insights into the early screening of lung cancer.Methods This study included patients with pulmonary nodules diagnosed at the Physical Examination Center of Shuguang Hospital Affiliated to Shanghai University of Traditional Chinese Medicine from November 1,2019 to December 31,2024,as well as patients with lung cancer diagnosed in the Oncology Departments of Yueyang Hospital of Integrated Traditional Chinese and Western Medicine and Longhua Hospital during the same period.The facial image information of patients with pulmonary nodules and lung cancer was collected using the TFDA-1 tongue and facial diagnosis instrument,and the facial diagnosis features were extracted from it by deep learning technology.Statistical analysis was conducted on the objective facial diagnosis characteristics of the two groups of participants to explore the differences in their facial image characteristics,and the least absolute shrinkage and selection operator(LASSO)regression was used to screen the characteristic variables.Based on the screened feature variables,four machine learning methods:random forest,logistic regression,support vector machine(SVM),and gradient boosting decision tree(GBDT)were used to establish lung cancer classification models independently.Meanwhile,the model performance was evaluated by indicators such as sensitivity,specificity,F1 score,precision,accuracy,the area under the receiver operating characteristic(ROC)curve(AUC),and the area under the precision-recall curve(AP).Results A total of 1275 patients with pulmonary nodules and 1623 patients with lung cancer were included in this study.After propensity score matching(PSM)to adjust for gender and age,535 patients were finally included in the pulmonary nodule group and the lung cancer group,respectively.There were significant differences in multiple color space metrics(such as R,G,B,V,L,a,b,Cr,H,Y,and Cb)and texture metrics[such as gray-levcl co-occurrence matrix(GLCM)-contrast(CON)and GLCM-inverse different moment(IDM)]between the two groups of individuals with pulmonary nodules and lung cancer(P<0.05).To construct a classification model,LASSO regression was used to select 63 key features from the initial 136 facial features.Based on this feature set,the SVM model demonstrated the best performance after 10-fold stratified cross-validation.The model achieved an average AUC of 0.8729 and average accuracy of 0.7990 on the internal test set.Further validation on an independent test set confirmed the model’s robust performance(AUC=0.8233,accuracy=0.7290),indicating its good generalization ability.Feature importance analysis demonstrated that color space indicators and the whole/lip Cr components(including color-B-0,wholecolor-Cr,and lipcolor-Cr)were the core factors in the model’s classification decisions,while texture indicators[GLCM-angular second moment(ASM)_2,GLCM-IDM_1,GLCM-CON_1,GLCM-entropy(ENT)_2]played an important auxiliary role.Conclusion The facial image features of patients with lung cancer and pulmonary nodules show significant differences in color and texture characteristics in multiple areas.The various models constructed based on facial image features all demonstrate good performance,indicating that facial image features can serve as potential biomarkers for lung cancer risk prediction,providing a non-invasive and feasible new approach for early lung cancer screening.展开更多
基金Supported by the National Key R&D Program of China(No.2023YFC2410203)Beijing Hospitals Authority Clinical Medicine Development of Special Funding Support(No.ZLRK202503).
文摘AIM:To investigate the clinical features and prognosis of patients with orbital inflammatory myofibroblastic tumor(IMT).METHODS:This retrospective study collected clinical data from 22 patients diagnosed with orbital IMT based on histopathological examination.The patients were followed up to assess their prognosis.Clinical data from patients,including age,gender,course of disease,past medical history,primary symptoms,ophthalmologic examination findings,general condition,as well as imaging,laboratory,histopathological,and immunohistochemical results from digital records were collected.Orbital magnetic resonance imaging(MRI)and(or)computed tomography(CT)scans were performed to assess bone destruction of the mass,invasion of surrounding tissues,and any inflammatory changes in periorbital areas.RESULTS:The mean age of patients with orbital IMT was 28.24±3.30y,with a male-to-female ratio of 1.2:1.Main clinical manifestations were proptosis,blurred vision,palpable mass,and pain.Bone destruction and surrounding tissue invasion occurred in 72.73%and 54.55%of cases,respectively.Inflammatory changes in the periorbital site were observed in 77.27%of the patients.Hematoxylin and eosin staining showed proliferation of fibroblasts and myofibroblasts,accompanied by infiltration of lymphocytes and plasma cells.Immunohistochemical staining revealed that smooth muscle actin(SMA)and vimentin were positive in 100%of cases,while anaplastic lymphoma kinase(ALK)showed positivity in 47.37%.The recurrence rate of orbital IMT was 27.27%,and sarcomatous degeneration could occur.There were no significant correlations between recurrence and factors such as age,gender,laterality,duration of the disease,periorbital tissue invasion,bone destruction,periorbital inflammation,tumor size,fever,leukocytosis,or treatment(P>0.05).However,lymphadenopathy and a Ki-67 index of 10%or higher may be risk factors for recurrence(P=0.046;P=0.023).CONCLUSION:Orbital IMT is a locally invasive disease that may recur or lead to sarcomatoid degeneration,primarily affecting young and middle-aged patients.The presence of lymphadenopathy and a Ki-67 index of 10%or higher may signify a poor prognosis.
文摘BACKGROUND SMARCB1/INI1-deficient pancreatic undifferentiated rhabdoid carcinoma is a highly aggressive tumor,and spontaneous splenic rupture(SSR)as its presenting manifestation is rarely reported among pancreatic malignancies.CASE SUMMARY We herein report a rare case of a 59-year-old female who presented with acute left upper quadrant abdominal pain without any history of trauma.Abdominal imaging demonstrated a heterogeneous splenic lesion with hemoperitoneum,raising clinical suspicion of SSR.Emergency laparotomy revealed a pancreatic tumor invading the spleen and left kidney,with associated splenic rupture and dense adhesions,necessitating en bloc resection of the distal pancreas,spleen,and left kidney.Histopathology revealed a biphasic malignancy composed of moderately differentiated pancreatic ductal adenocarcinoma and an undifferentiated carcinoma with rhabdoid morphology and loss of SMARCB1 expression.Immunohistochemical analysis confirmed complete loss of SMARCB1/INI1 in the undifferentiated component,along with a high Ki-67 index(approximately 80%)and CD10 positivity.The ductal adenocarcinoma component retained SMARCB1/INI1 expression and was positive for CK7 and CK-pan.Transitional zones between the two tumor components suggested progressive dedifferentiation and underlying genomic instability.The patient received adjuvant chemotherapy with gemcitabine and nab-paclitaxel and maintained a satisfactory quality of life at the 6-month follow-up.CONCLUSION This study reports a rare case of SMARCB1/INI1-deficient undifferentiated rhabdoid carcinoma of the pancreas combined with ductal adenocarcinoma,presenting as SSR-an exceptionally uncommon initial manifestation of pancreatic malignancy.
文摘In the article“A Lightweight Approach for Skin Lesion Detection through Optimal Features Fusion”by Khadija Manzoor,Fiaz Majeed,Ansar Siddique,Talha Meraj,Hafiz Tayyab Rauf,Mohammed A.El-Meligy,Mohamed Sharaf,Abd Elatty E.Abd Elgawad Computers,Materials&Continua,2022,Vol.70,No.1,pp.1617–1630.DOI:10.32604/cmc.2022.018621,URL:https://www.techscience.com/cmc/v70n1/44361,there was an error regarding the affiliation for the author Hafiz Tayyab Rauf.Instead of“Centre for Smart Systems,AI and Cybersecurity,Staffordshire University,Stoke-on-Trent,UK”,the affiliation should be“Independent Researcher,Bradford,BD80HS,UK”.
文摘BACKGROUND Pancreatic cancer remains one of the most lethal malignancies worldwide,with a poor prognosis often attributed to late diagnosis.Understanding the correlation between pathological type and imaging features is crucial for early detection and appropriate treatment planning.AIM To retrospectively analyze the relationship between different pathological types of pancreatic cancer and their corresponding imaging features.METHODS We retrospectively analyzed the data of 500 patients diagnosed with pancreatic cancer between January 2010 and December 2020 at our institution.Pathological types were determined by histopathological examination of the surgical spe-cimens or biopsy samples.The imaging features were assessed using computed tomography,magnetic resonance imaging,and endoscopic ultrasound.Statistical analyses were performed to identify significant associations between pathological types and specific imaging characteristics.RESULTS There were 320(64%)cases of pancreatic ductal adenocarcinoma,75(15%)of intraductal papillary mucinous neoplasms,50(10%)of neuroendocrine tumors,and 55(11%)of other rare types.Distinct imaging features were identified in each pathological type.Pancreatic ductal adenocarcinoma typically presents as a hypodense mass with poorly defined borders on computed tomography,whereas intraductal papillary mucinous neoplasms present as characteristic cystic lesions with mural nodules.Neuroendocrine tumors often appear as hypervascular lesions in contrast-enhanced imaging.Statistical analysis revealed significant correlations between specific imaging features and pathological types(P<0.001).CONCLUSION This study demonstrated a strong association between the pathological types of pancreatic cancer and imaging features.These findings can enhance the accuracy of noninvasive diagnosis and guide personalized treatment approaches.
文摘During Donald Trump’s first term,the“Trump Shock”brought world politics into an era of uncertainties and pulled the transatlantic alliance down to its lowest point in history.The Trump 2.0 tsunami brewed by the 2024 presidential election of the United States has plunged the U.S.-Europe relations into more gloomy waters,ushering in a more complex and turbulent period of adjustment.
基金Supported by the Henan Province Key Research and Development Project(231111211300)the Central Government of Henan Province Guides Local Science and Technology Development Funds(Z20231811005)+2 种基金Henan Province Key Research and Development Project(231111110100)Henan Provincial Outstanding Foreign Scientist Studio(GZS2024006)Henan Provincial Joint Fund for Scientific and Technological Research and Development Plan(Application and Overcoming Technical Barriers)(242103810028)。
文摘The fusion of infrared and visible images should emphasize the salient targets in the infrared image while preserving the textural details of the visible images.To meet these requirements,an autoencoder-based method for infrared and visible image fusion is proposed.The encoder designed according to the optimization objective consists of a base encoder and a detail encoder,which is used to extract low-frequency and high-frequency information from the image.This extraction may lead to some information not being captured,so a compensation encoder is proposed to supplement the missing information.Multi-scale decomposition is also employed to extract image features more comprehensively.The decoder combines low-frequency,high-frequency and supplementary information to obtain multi-scale features.Subsequently,the attention strategy and fusion module are introduced to perform multi-scale fusion for image reconstruction.Experimental results on three datasets show that the fused images generated by this network effectively retain salient targets while being more consistent with human visual perception.
基金partially supported by the National Natural Science Foundation (62272248)the Open Project Fund of State Key Laboratory of Computer Architecture,Institute of Computing Technology,Chinese Academy of Sciences (CARCHA202108,CARCH201905)+1 种基金the Natural Science Foundation of Tianjin (20JCZDJC00610)Sponsored by Zhejiang Lab (2021KF0AB04)。
文摘Smart contracts are widely used on the blockchain to implement complex transactions,such as decentralized applications on Ethereum.Effective vulnerability detection of large-scale smart contracts is critical,as attacks on smart contracts often cause huge economic losses.Since it is difficult to repair and update smart contracts,it is necessary to find the vulnerabilities before they are deployed.However,code analysis,which requires traversal paths,and learning methods,which require many features to be trained,are too time-consuming to detect large-scale on-chain contracts.Learning-based methods will obtain detection models from a feature space compared to code analysis methods such as symbol execution.But the existing features lack the interpretability of the detection results and training model,even worse,the large-scale feature space also affects the efficiency of detection.This paper focuses on improving the detection efficiency by reducing the dimension of the features,combined with expert knowledge.In this paper,a feature extraction model Block-gram is proposed to form low-dimensional knowledge-based features from bytecode.First,the metadata is separated and the runtime code is converted into a sequence of opcodes,which are divided into segments based on some instructions(jumps,etc.).Then,scalable Block-gram features,including 4-dimensional block features and 8-dimensional attribute features,are mined for the learning-based model training.Finally,feature contributions are calculated from SHAP values to measure the relationship between our features and the results of the detection model.In addition,six types of vulnerability labels are made on a dataset containing 33,885 contracts,and these knowledge-based features are evaluated using seven state-of-the-art learning algorithms,which show that the average detection latency speeds up 25×to 650×,compared with the features extracted by N-gram,and also can enhance the interpretability of the detection model.
基金This research is a phased achievement of The National Social Science Fund of China(23BGL272).
文摘Fraudulent website is an important car-rier tool for telecom fraud.At present,criminals can use artificial intelligence generative content technol-ogy to quickly generate fraudulent website templates and build fraudulent websites in batches.Accurate identification of fraudulent website will effectively re-duce the risk of public victimization.Therefore,this study developed a fraudulent website template iden-tification method based on DOM structure extraction of website fingerprint features,which solves the prob-lems of single-dimension identification,low accuracy,and the insufficient generalization ability of current fraudulent website templates.This method uses an im-proved SimHash algorithm to traverse the DOM tree of a webpage,extract website node features,calcu-late the weight of each node,and obtain the finger-print feature vector of the website through dimension-ality reduction.Finally,the random forest algorithm is used to optimize the training features for the best combination of parameters.This method automati-cally extracts fingerprint features from websites and identifies website template ownership based on these features.An experimental analysis showed that this method achieves a classification accuracy of 89.8%and demonstrates superior recognition.
基金King Saud University,Grant/Award Number:RSP2024R157。
文摘Biometric characteristics are playing a vital role in security for the last few years.Human gait classification in video sequences is an important biometrics attribute and is used for security purposes.A new framework for human gait classification in video sequences using deep learning(DL)fusion assisted and posterior probability-based moth flames optimization(MFO)is proposed.In the first step,the video frames are resized and finetuned by two pre-trained lightweight DL models,EfficientNetB0 and MobileNetV2.Both models are selected based on the top-5 accuracy and less number of parameters.Later,both models are trained through deep transfer learning and extracted deep features fused using a voting scheme.In the last step,the authors develop a posterior probabilitybased MFO feature selection algorithm to select the best features.The selected features are classified using several supervised learning methods.The CASIA-B publicly available dataset has been employed for the experimental process.On this dataset,the authors selected six angles such as 0°,18°,90°,108°,162°,and 180°and obtained an average accuracy of 96.9%,95.7%,86.8%,90.0%,95.1%,and 99.7%.Results demonstrate comparable improvement in accuracy and significantly minimize the computational time with recent state-of-the-art techniques.
基金supported by the Fundamental Research Funds for the Provincial Universities of Zhejiang (No.GK249909299001-036)National Key Research and Development Program of China (No. 2023YFB4502803)Zhejiang Provincial Natural Science Foundation of China (No.LDT23F01014F01)。
文摘Due to the limitations of existing imaging hardware, obtaining high-resolution hyperspectral images is challenging. Hyperspectral image super-resolution(HSI SR) has been a very attractive research topic in computer vision, attracting the attention of many researchers. However, most HSI SR methods focus on the tradeoff between spatial resolution and spectral information, and cannot guarantee the efficient extraction of image information. In this paper, a multidimensional features network(MFNet) for HSI SR is proposed, which simultaneously learns and fuses the spatial,spectral, and frequency multidimensional features of HSI. Spatial features contain rich local details,spectral features contain the information and correlation between spectral bands, and frequency feature can reflect the global information of the image and can be used to obtain the global context of HSI. The fusion of the three features can better guide image super-resolution, to obtain higher-quality high-resolution hyperspectral images. In MFNet, we use the frequency feature extraction module(FFEM) to extract the frequency feature. On this basis, a multidimensional features extraction module(MFEM) is designed to learn and fuse multidimensional features. In addition, experimental results on two public datasets demonstrate that MFNet achieves state-of-the-art performance.
文摘Recognizing road scene context from a single image remains a critical challenge for intelligent autonomous driving systems,particularly in dynamic and unstructured environments.While recent advancements in deep learning have significantly enhanced road scene classification,simultaneously achieving high accuracy,computational efficiency,and adaptability across diverse conditions continues to be difficult.To address these challenges,this study proposes HybridLSTM,a novel and efficient framework that integrates deep learning-based,object-based,and handcrafted feature extraction methods within a unified architecture.HybridLSTM is designed to classify four distinct road scene categories—crosswalk(CW),highway(HW),overpass/tunnel(OP/T),and parking(P)—by leveraging multiple publicly available datasets,including Places-365,BDD100K,LabelMe,and KITTI,thereby promoting domain generalization.The framework fuses object-level features extracted using YOLOv5 and VGG19,scene-level global representations obtained from a modified VGG19,and fine-grained texture features captured through eight handcrafted descriptors.This hybrid feature fusion enables the model to capture both semantic context and low-level visual cues,which are critical for robust scene understanding.To model spatial arrangements and latent sequential dependencies present even in static imagery,the combined features are processed through a Long Short-Term Memory(LSTM)network,allowing the extraction of discriminative patterns across heterogeneous feature spaces.Extensive experiments conducted on 2725 annotated road scene images,with an 80:20 training-to-testing split,validate the effectiveness of the proposed model.HybridLSTM achieves a classification accuracy of 96.3%,a precision of 95.8%,a recall of 96.1%,and an F1-score of 96.0%,outperforming several existing state-of-the-art methods.These results demonstrate the robustness,scalability,and generalization capability of HybridLSTM across varying environments and scene complexities.Moreover,the framework is optimized to balance classification performance with computational efficiency,making it highly suitable for real-time deployment in embedded autonomous driving systems.Future work will focus on extending the model to multi-class detection within a single frame and optimizing it further for edge-device deployments to reduce computational overhead in practical applications.
基金supported by National Natural Science Foundation of China(22374080,22174068,21722504)Primary Research&Development Plan of Jiangsu Province(BK20221303,BE2022796)+1 种基金Open Foundation of State Key Laboratory of Reproductive Medicine(SKLRM-2022BP1,JX116GSP20240507)Science and Technology Development Fund of NJMU(NJMUQY2022003)。
文摘Drug resistance remains a major challenge in breast cancer chemotherapy,yet the metabolic alterations underlying this phenomenon are not fully understood.There is much evidence indicating the cellular heterogeneity among cancer cells,which exhibit varying degrees of metabolic reprogramming and thus may result in differential contributions to drug resistance.A home-built single-cell quantitative mass spectrometry(MS)platform,which integrates micromanipulation and electro-osmotic sampling,was developed to quantitatively profile the tricarboxylic acid(TCA)cycle metabolites at the single-cell level.Using this platform,the metabolic profiles of drug-sensitive MCF-7 breast cancer cells and their drug-resistant derivative MCF-7/ADR cells were compared.This results revealed a selective upregulation of downstream TCA cycle metabolites includingα-ketoglutarate,succinate,fumarate,and malate in drug-resistant cancer cells,while early TCA metabolites remained largely unchanged.Furthermore,notable variations in the abundance of the metabolites were observed in individual cells.The comparative analysis also revealed that not all MCF-7/ADR cells exhibit the same degree of metabolic deviation from the parental line in the metabolites during resistance acquisition.The observed metabolic profiles indicate enhanced glutaminolysis,altered mitochondrial electron transport chain activity,and increased metabolic flexibility in drug-resistant cancer cells that support their survival under chemotherapeutic stress.The findings further suggest the potential for incorporating cellular metabolic heterogeneity into future drug resistance studies.
文摘Multi-Object Tracking(MOT)represents a fundamental but computationally demanding task in computer vision,with particular challenges arising in occluded and densely populated environments.While contemporary tracking systems have demonstrated considerable progress,persistent limitations—notably frequent occlusion-induced identity switches and tracking inaccuracies—continue to impede reliable real-world deployment.This work introduces an advanced tracking framework that enhances association robustness through a two-stage matching paradigm combining spatial and appearance features.Proposed framework employs:(1)a Height Modulated and Scale Adaptive Spatial Intersection-over-Union(HMSIoU)metric for improved spatial correspondence estimation across variable object scales and partial occlusions;(2)a feature extraction module generating discriminative appearance descriptors for identity maintenance;and(3)a recovery association mechanism for refining matches between unassociated tracks and detections.Comprehensive evaluation on standard MOT17 and MOT20 benchmarks demonstrates significant improvements in tracking consistency,with state-of-the-art performance across key metrics including HOTA(64),MOTA(80.7),IDF1(79.8),and IDs(1379).These results substantiate the efficacy of our Cue-Tracker framework in complex real-world scenarios characterized by occlusions and crowd interactions.
基金funded by Scientific Research Deanship at University of Hail-Saudi Arabia through Project Number RG-23092.
文摘Cyberbullying on social media poses significant psychological risks,yet most detection systems over-simplify the task by focusing on binary classification,ignoring nuanced categories like passive-aggressive remarks or indirect slurs.To address this gap,we propose a hybrid framework combining Term Frequency-Inverse Document Frequency(TF-IDF),word-to-vector(Word2Vec),and Bidirectional Encoder Representations from Transformers(BERT)based models for multi-class cyberbullying detection.Our approach integrates TF-IDF for lexical specificity and Word2Vec for semantic relationships,fused with BERT’s contextual embeddings to capture syntactic and semantic complexities.We evaluate the framework on a publicly available dataset of 47,000 annotated social media posts across five cyberbullying categories:age,ethnicity,gender,religion,and indirect aggression.Among BERT variants tested,BERT Base Un-Cased achieved the highest performance with 93%accuracy(standard deviation across±1%5-fold cross-validation)and an average AUC of 0.96,outperforming standalone TF-IDF(78%)and Word2Vec(82%)models.Notably,it achieved near-perfect AUC scores(0.99)for age and ethnicity-based bullying.A comparative analysis with state-of-the-art benchmarks,including Generative Pre-trained Transformer 2(GPT-2)and Text-to-Text Transfer Transformer(T5)models highlights BERT’s superiority in handling ambiguous language.This work advances cyberbullying detection by demonstrating how hybrid feature extraction and transformer models improve multi-class classification,offering a scalable solution for moderating nuanced harmful content.
基金supported by the National Key R&D Program of China(No.2022ZD0118402)。
文摘Remote sensing cross-modal image-text retrieval(RSCIR)can flexibly and subjectively retrieve remote sensing images utilizing query text,which has received more researchers’attention recently.However,with the increasing volume of visual-language pre-training model parameters,direct transfer learning consumes a substantial amount of computational and storage resources.Moreover,recently proposed parameter-efficient transfer learning methods mainly focus on the reconstruction of channel features,ignoring the spatial features which are vital for modeling key entity relationships.To address these issues,we design an efficient transfer learning framework for RSCIR,which is based on spatial feature efficient reconstruction(SPER).A concise and efficient spatial adapter is introduced to enhance the extraction of spatial relationships.The spatial adapter is able to spatially reconstruct the features in the backbone with few parameters while incorporating the prior information from the channel dimension.We conduct quantitative and qualitative experiments on two different commonly used RSCIR datasets.Compared with traditional methods,our approach achieves an improvement of 3%-11% in sumR metric.Compared with methods finetuning all parameters,our proposed method only trains less than 1% of the parameters,while maintaining an overall performance of about 96%.
文摘Malaria is considered one of the major causes of travel-related morbidity and mortality,especially among non-immune travelers from non-endemic countries to the endemic regions.According to a multicenter study from the GeoSentinel surveillance network,malaria was the most frequent cause of fever in 21%of returning travelers,followed by dengue,typhoid fever,chikungunya and rickettsiosis[1].Individuals traveling from regions without malaria transmission to areas where it is endemic face a heightened risk of contracting the disease due to their lack of immunity.Despite the official malaria-free status of the Russian Federation since 2010,annual cases of severe Plasmodium(P.)falciparum malaria continue to be reported[2].This underscores the necessity for heightened clinical vigilance and improved preventive strategies especially in non-endemic settings.
基金The National Key Research and Development Program of China under contract No.2023YFC3008204the National Natural Science Foundation of China under contract Nos 41977302 and 42476217.
文摘Spartina alterniflora is now listed among the world’s 100 most dangerous invasive species,severely affecting the ecological balance of coastal wetlands.Remote sensing technologies based on deep learning enable large-scale monitoring of Spartina alterniflora,but they require large datasets and have poor interpretability.A new method is proposed to detect Spartina alterniflora from Sentinel-2 imagery.Firstly,to get the high canopy cover and dense community characteristics of Spartina alterniflora,multi-dimensional shallow features are extracted from the imagery.Secondly,to detect different objects from satellite imagery,index features are extracted,and the statistical features of the Gray-Level Co-occurrence Matrix(GLCM)are derived using principal component analysis.Then,ensemble learning methods,including random forest,extreme gradient boosting,and light gradient boosting machine models,are employed for image classification.Meanwhile,Recursive Feature Elimination with Cross-Validation(RFECV)is used to select the best feature subset.Finally,to enhance the interpretability of the models,the best features are utilized to classify multi-temporal images and SHapley Additive exPlanations(SHAP)is combined with these classifications to explain the model prediction process.The method is validated by using Sentinel-2 imageries and previous observations of Spartina alterniflora in Chongming Island,it is found that the model combining image texture features such as GLCM covariance can significantly improve the detection accuracy of Spartina alterniflora by about 8%compared with the model without image texture features.Through multiple model comparisons and feature selection via RFECV,the selected model and eight features demonstrated good classification accuracy when applied to data from different time periods,proving that feature reduction can effectively enhance model generalization.Additionally,visualizing model decisions using SHAP revealed that the image texture feature component_1_GLCMVariance is particularly important for identifying each land cover type.
基金supported by the National Natural Science Foundation of China(Nos.U22A2034,62177047)High Caliber Foreign Experts Introduction Plan funded by MOST,and Central South University Research Programme of Advanced Interdisciplinary Studies(No.2023QYJC020).
文摘Image captioning,the task of generating descriptive sentences for images,has advanced significantly with the integration of semantic information.However,traditional models still rely on static visual features that do not evolve with the changing linguistic context,which can hinder the ability to form meaningful connections between the image and the generated captions.This limitation often leads to captions that are less accurate or descriptive.In this paper,we propose a novel approach to enhance image captioning by introducing dynamic interactions where visual features continuously adapt to the evolving linguistic context.Our model strengthens the alignment between visual and linguistic elements,resulting in more coherent and contextually appropriate captions.Specifically,we introduce two innovative modules:the Visual Weighting Module(VWM)and the Enhanced Features Attention Module(EFAM).The VWM adjusts visual features using partial attention,enabling dynamic reweighting of the visual inputs,while the EFAM further refines these features to improve their relevance to the generated caption.By continuously adjusting visual features in response to the linguistic context,our model bridges the gap between static visual features and dynamic language generation.We demonstrate the effectiveness of our approach through experiments on the MS-COCO dataset,where our method outperforms state-of-the-art techniques in terms of caption quality and contextual relevance.Our results show that dynamic visual-linguistic alignment significantly enhances image captioning performance.
基金supported by the National Key Research and Development Pro-gram of China(Grant No.2022YFC3003902)the National Natu-ral Science Foundation of China(Grant Nos.42075146 and 42275006).
文摘In July 2021,a catastrophic extreme precipitation(EP)event occurred in Henan Province,China,resulting in considerable human and economic losses.The synoptic pattern during this event is distinctive,characterized by the presence of two typhoons and substantial water transport into Henan.However,a favorable synoptic pattern only does not guarantee the occurrence of heavy precipitation in Henan.This study investigates the key environmental features critical for EP under similar synoptic patterns to the 2021 Henan extreme event.It is found that cold clouds are better aggregated on EP days,accompanied by beneficial environment features like enhanced moisture conditions,stronger updrafts,and greater atmospheric instability.The temporal evolution of these environmental features shows a leading signal by one to three days.These results suggest the importance of combining the synoptic pattern and environmental features in the forecasting of heavy precipitation events.
基金National Natural Science Foundation of China(82305090)Shanghai Municipal Health Commission(20234Y0168)National Key Research and Development Program of China (2017YFC1703301)。
文摘Objective To explore the feasibility of constructing a lung cancer early-warning risk model based on facial image features,providing novel insights into the early screening of lung cancer.Methods This study included patients with pulmonary nodules diagnosed at the Physical Examination Center of Shuguang Hospital Affiliated to Shanghai University of Traditional Chinese Medicine from November 1,2019 to December 31,2024,as well as patients with lung cancer diagnosed in the Oncology Departments of Yueyang Hospital of Integrated Traditional Chinese and Western Medicine and Longhua Hospital during the same period.The facial image information of patients with pulmonary nodules and lung cancer was collected using the TFDA-1 tongue and facial diagnosis instrument,and the facial diagnosis features were extracted from it by deep learning technology.Statistical analysis was conducted on the objective facial diagnosis characteristics of the two groups of participants to explore the differences in their facial image characteristics,and the least absolute shrinkage and selection operator(LASSO)regression was used to screen the characteristic variables.Based on the screened feature variables,four machine learning methods:random forest,logistic regression,support vector machine(SVM),and gradient boosting decision tree(GBDT)were used to establish lung cancer classification models independently.Meanwhile,the model performance was evaluated by indicators such as sensitivity,specificity,F1 score,precision,accuracy,the area under the receiver operating characteristic(ROC)curve(AUC),and the area under the precision-recall curve(AP).Results A total of 1275 patients with pulmonary nodules and 1623 patients with lung cancer were included in this study.After propensity score matching(PSM)to adjust for gender and age,535 patients were finally included in the pulmonary nodule group and the lung cancer group,respectively.There were significant differences in multiple color space metrics(such as R,G,B,V,L,a,b,Cr,H,Y,and Cb)and texture metrics[such as gray-levcl co-occurrence matrix(GLCM)-contrast(CON)and GLCM-inverse different moment(IDM)]between the two groups of individuals with pulmonary nodules and lung cancer(P<0.05).To construct a classification model,LASSO regression was used to select 63 key features from the initial 136 facial features.Based on this feature set,the SVM model demonstrated the best performance after 10-fold stratified cross-validation.The model achieved an average AUC of 0.8729 and average accuracy of 0.7990 on the internal test set.Further validation on an independent test set confirmed the model’s robust performance(AUC=0.8233,accuracy=0.7290),indicating its good generalization ability.Feature importance analysis demonstrated that color space indicators and the whole/lip Cr components(including color-B-0,wholecolor-Cr,and lipcolor-Cr)were the core factors in the model’s classification decisions,while texture indicators[GLCM-angular second moment(ASM)_2,GLCM-IDM_1,GLCM-CON_1,GLCM-entropy(ENT)_2]played an important auxiliary role.Conclusion The facial image features of patients with lung cancer and pulmonary nodules show significant differences in color and texture characteristics in multiple areas.The various models constructed based on facial image features all demonstrate good performance,indicating that facial image features can serve as potential biomarkers for lung cancer risk prediction,providing a non-invasive and feasible new approach for early lung cancer screening.