As a common practice,a large hydro-generator will operate in leading phase conditions to absorb the reactive power of the power grid.However,the accurate and precise prediction of the leading phase operation capacity ...As a common practice,a large hydro-generator will operate in leading phase conditions to absorb the reactive power of the power grid.However,the accurate and precise prediction of the leading phase operation capacity of a large hydro-generator has always been a formidable challenge to engineers and academicians because it is extremely hard to compute the eddy currents and losses as well as the local overheating in the pressure plate and finger.To address this problem,a full three dimensional(3D)finite element model and method of the coupled eddy current and temperature fields in the end region of a large hydro-generator are developed.The equivalent medium parameters used in the computations are comprehensively discussed.Moreover,some numerically based solution methodologies for accurate computation of the field and armature currents under different leading phase conditions are proposed.Numerical results on the coupled eddy current and temperature fields in the end regions of a 250 MW hydro-generator confirm positively the feasibility of the present work.展开更多
In this paper,we establish some strong laws of large numbers,which are for nonindependent random variables under the framework of sublinear expectations.One of our main results is for blockwise m-dependent random vari...In this paper,we establish some strong laws of large numbers,which are for nonindependent random variables under the framework of sublinear expectations.One of our main results is for blockwise m-dependent random variables,and another is for sub-orthogonal random variables.Both extend the strong law of large numbers for independent random variables under sublinear expectations to the non-independent case.展开更多
Model evaluation using benchmark datasets is an important method to measure the capability of large language models(LLMs)in specific domains,and it is mainly used to assess the knowledge and reasoning abilities of LLM...Model evaluation using benchmark datasets is an important method to measure the capability of large language models(LLMs)in specific domains,and it is mainly used to assess the knowledge and reasoning abilities of LLMs.Therefore,in order to better assess the capability of LLMs in the agricultural domain,Agri-Eval was proposed as a benchmark for assessing the knowledge and reasoning ability of LLMs in agriculture.The assessment dataset used in Agri-Eval covered seven major disciplines in the agricultural domain:crop science,horticulture,plant protection,animal husbandry,forest science,aquaculture science,and grass science,and contained a total of 2283 questions.Among domestic general-purpose LLMs,DeepSeek R1 performed best with an accuracy rate of 75.49%.In the realm of international general-purpose LLMs,Gemini 2.0 pro exp 0205 standed out as the top performer,achieving an accuracy rate of 74.28%.As an LLMs in agriculture vertical,Shennong V2.0 outperformed all the LLMs in China,and the answer accuracy rate of agricultural knowledge exceeded that of all the existing general-purpose LLMs.The launch of Agri-Eval helped the LLM developers to comprehensively evaluate the model's capability in the field of agriculture through a variety of tasks and tests to promote the development of the LLMs in the field of agriculture.展开更多
To the Editor,Artificial intelligence(AI)usage has been increasing.Many fields have implemented the use of AI and Large LanguageModels(LLMs),especially in medicine.Furthermore,manypatients have increasingly been using...To the Editor,Artificial intelligence(AI)usage has been increasing.Many fields have implemented the use of AI and Large LanguageModels(LLMs),especially in medicine.Furthermore,manypatients have increasingly been using AI;often,they will prompt AI with questions before even stepping into a physi-cian's office.The question lies in whether the information produced by AI is reliable and if this information is concise and easy to read across all patient populations.展开更多
This study evaluated the accuracy,completeness,and comprehensibility of responses from mainstream large language models(LLMs)to hepatitis C virus(HCV)-related questions,aiming to assess their performance in addressing...This study evaluated the accuracy,completeness,and comprehensibility of responses from mainstream large language models(LLMs)to hepatitis C virus(HCV)-related questions,aiming to assess their performance in addressing patient queries about disease and lifestyle behaviors.The models selected were ChatGPT-4o,Gemini 2.0 Pro,Claude 3.5 Sonnet,and DeepSeek V3,with 12 questions chosen by two HCV experts from the domains of prevention,diagnosis,and treatment.展开更多
It is known that correlation does not imply causality.Some relationships identified in the analysis of data are coincidental or unknown,and some are produced by real-world causality of the situation,which is problemat...It is known that correlation does not imply causality.Some relationships identified in the analysis of data are coincidental or unknown,and some are produced by real-world causality of the situation,which is problematic,since there is a need to differentiate between these two scenarios.Until recently,the proper−semantic−causality of the relationship could have been determined only by human experts from the area of expertise of the studied data.This has changed with the advance of large language models,which are often utilized as surrogates for such human experts,making the process automated and readily available to all data analysts.This motivates the main objective of this work,which is to introduce the design and implementation of a large language model-based semantic causality evaluator based on correlation analysis,together with its visual analysis model called Causal heatmap.After the implementation itself,the model is evaluated from the point of view of the quality of the visual model,from the point of view of the quality of causal evaluation based on large language models,and from the point of view of comparative analysis,while the results reached in the study highlight the usability of large language models in the task and the potential of the proposed approach in the analysis of unknown datasets.The results of the experimental evaluation demonstrate the usefulness of the Causal heatmap method,supported by the evident highlighting of interesting relationships,while suppressing irrelevant ones.展开更多
Background:Large language models(LLMs)have shown considerable promise in supporting clinical decision-making.However,their adoption and evaluation in dermatology remains limited.This study aimed to explore the prefere...Background:Large language models(LLMs)have shown considerable promise in supporting clinical decision-making.However,their adoption and evaluation in dermatology remains limited.This study aimed to explore the preferences of Chinese dermatologists regarding LLM-generated responses in clinical psoriasis scenarios and to assess how they prioritize key quality dimensions,including accuracy,traceability,and logicality.Methods:A cross-sectional,web-based survey was conducted between December 25,2024,and January 22,2025,following the Checklist for Reporting Results of Internet E-Surveys guidelines.A total of 1247 valid responses were collected from practicing dermatologists across 33 of China's provincial-level administrative divisions.Participants evaluated responses to five categories of clinical questions(etiology,clinical presentation,differential diagnosis,treatment,and case study)generated by five LLMs:ChatGPT-4o,Kimi.ai,Doubao,ZuoYiGPT,and Lingyi-agent.Statistical associations between participant characteristics and model preferences were examined using chi-square tests.Results:ChatGPT-4o(Model 1)emerged as the most preferred model across all clinical tasks,consistently receiving the highest number of votes in case study(n=740),clinical presentation(n=666),differential diagnosis(n=707),etiology(n=602),and treatment(n=656).Significant variation in model preference by professional title was observed only for the differential diagnosis task(χ^(2)=21.13,df=12,p=0.0485),while no significant differences were found across hospital tiers(p>0.05).In terms of evaluation dimensions,accuracy was most frequently rated as“very important”(n=635).A significant association existed between hospital tier and the most valued dimension(χ^(2)=27.667,df=9,p=0.0011),with dermatologists in primary hospitals prioritizing traceability more than their peers in higher-tier hospitals.No significant associations were found across professional titles(p=0.127).Conclusions:Chinese dermatologists suggest a strong preference for ChatGPT-4o over domestic LLMs in psoriasis-related clinical tasks.While accuracy remains the primary criterion,traceability and logicality are also critical,particularly for clinicians in lower-tier hospitals.These findings suggest that future clinical LLMs should prioritize not only content accuracy but also source transparency and structural clarity to meet the diverse needs of different clinical settings.展开更多
Large language models(LLMs)show considerable potential to revolutionize healthcare through their performance across diverse clinical applications.Given the inherent constraints of LLMs and the critical nature of medic...Large language models(LLMs)show considerable potential to revolutionize healthcare through their performance across diverse clinical applications.Given the inherent constraints of LLMs and the critical nature of medical practice,a rigorous and systematic evaluation of their medical competence is imperative.This study presents a comprehensive review of the established methodologies and benchmarks for evaluating the medical competence of LLMs,encompassing a thorough analysis of current assessment practices across medical knowledge,clinical practice competence,and ethical-safety considerations.By integrating clinician competency assessment frameworks into LLMs evaluation,we propose a structured tri-dimensional framework that systematically organizes existing evaluation approaches according to medical theoretical knowledge,clinical practice ability,and ethical-safety considerations.Furthermore,this research provides critical insights into future developmental trajectories while establishing foundational frameworks and standardization protocols for the integration of LLMs into medical practice.展开更多
Background:Despite the promise shown by large language models(LLMs)for standardized tasks,their multidimensional performance in real-world oncology decision-making remains unevaluated.This study aims to introduce a fr...Background:Despite the promise shown by large language models(LLMs)for standardized tasks,their multidimensional performance in real-world oncology decision-making remains unevaluated.This study aims to introduce a framework for evaluating LLMs and physician decisions in challenging lung cancer cases.Methods:We curated 50 challenging lung cancer cases(25 local and 25 published)classified as complex,rare,or refractory.Blinded three-dimensional,five-point Likert evaluations(1–5 for comprehensiveness,specificity,and readability)compared standalone LLMs(DeepSeek R1,Claude 3.5,Gemini 1.5,and GPT-4o),physicians by experience level(junior,intermediate,and senior),and AI-assisted juniors;intergroup differences and augmentation effects were analyzed statistically.Results:Of 50 challenging cases(18 complex,17 rare,and 15 refractory)rated by three experts,DeepSeek R1 achieved scores of 3.95±0.33,3.71±0.53,and 4.26±0.18 for comprehensiveness,specificity,and readability,respectively,positioning it between intermediate(3.68,3.68,3.75)and senior(4.50,4.64,4.53)physicians.GPT-4o and Claude 3.5 reached intermediate physician–level comprehensiveness(3.76±0.39,3.60±0.39)but junior-to-intermediate physician–level specificity(3.39±0.39,3.39±0.49).All LLMs scored higher on rare cases than intermediate physicians but fell below junior physicians in refractory-case specificity.AIassisted junior physicians showed marked gains in rare cases,with comprehensiveness rising from 2.32 to 4.29(84.8%),specificity from 2.24 to 4.26(90.8%),and readability from 2.76 to 4.59(66.0%),while specificity declined by 3.2%(3.17 to 3.07)in refractory cases.Error analysis showed complementary strengths,with physicians demonstrating reasoning stability and LLMs excelling in knowledge updating and risk management.Conclusions:LLMs performed variably in clinical decision-making tasks depending on case type,performing better in rare cases and worse in refractory cases requiring longitudinal reasoning.Complementary strengths between LLMs and physicians support case-and task-tailored human–AI collaboration.展开更多
Polyfluoroalkyl substances(PFAS)have emerged as persistent environmental contaminants because of their chemical stability,degradation-resistance and bioaccumulation potential.However,current studies mainly focus on th...Polyfluoroalkyl substances(PFAS)have emerged as persistent environmental contaminants because of their chemical stability,degradation-resistance and bioaccumulation potential.However,current studies mainly focus on the toxicity of single PFAS such as perfluorooctanoic acid(PFOA)and perfluorobutanoic acid(PFBA),the knowledge of their combined effects is relatively limited.In this study,we explored the immune response of the gut in large yellow croaker(Larimichthys crocea)under the combined stress of PFOA and PFBA.Histologicalanalyses revealed that the combined effect induced intestinal vacuolization and decreased the length of intestinal villi.And it significantly activated pro-inflammatory pathways with marked upregulation of tnfα,il1β,il6 and myd88 expressions,particularly after 14 days of exposure.Gut microbiota analysis revealed substantial dysbiosis,including 1)reduced alpha diversity,2)increased abundance of potential pathogenic taxa(Proteobacteria and Spirochaetota),and 3)depletion of beneficial Firmicutes.PICRUSt-based functional prediction indicated temporal metabolic shifts,with upregulation of DNA repair pathways at day 3 and enhanced bacterial motility protein activity at days 7 and 14 of post-exposure.The Pearson correlation analysis further indicated that these immune genes had significant positive correlations with Vibrio and Brevinema,and negative correlations with Streptococcus.Our present study will provide novel insights into the microbiome-mediated immunomodulation in the larger yellow croaker exposed to combined PFAS,which will be helpful for healthy farming of economically important marine species.展开更多
This study presents an implicit multiphysics coupling method integrating Computational Fluid Dynamics(CFD),the Multiphase Particle-in-Cell(MPPIC)model,and the Finite Element Method(FEM),implemented with OpenFOAM,Calcu...This study presents an implicit multiphysics coupling method integrating Computational Fluid Dynamics(CFD),the Multiphase Particle-in-Cell(MPPIC)model,and the Finite Element Method(FEM),implemented with OpenFOAM,CalculiX,and preCICE to simulate fluid-particle-structure interactions with large deformations.Mesh motion in the fluid field is handled using the radial basis function(RBF)method.The particle phase is modeled by MPPIC,where fluid-particle interaction is described through momentum exchange,and inter-particle collisions are characterized by collision stress.The structural field is solved by nonlinear FEM to capture large deformations induced by geometric nonlinearity.Coupling among fields is realized through a partitioned,parallel,and non-intrusive iterative strategy,ensuring stable transfer and convergence of interface forces and displacements.Notably,the influence of particles on the structure is not direct but mediated by the fluid,while structural motion directly affects particle dynamics.The results demonstrate that the proposed approach effectively captures multiphysics interaction processes and provides a valuable reference for numerical modeling of coupled fluid-particle-structure systems.展开更多
Electrochromic smart windows(ESWs)can significantly reduce building energy consumption,but the high cost hinders large-scale production.The in situ growth of tungsten oxide(WO_(3))films is only by a simple immersion p...Electrochromic smart windows(ESWs)can significantly reduce building energy consumption,but the high cost hinders large-scale production.The in situ growth of tungsten oxide(WO_(3))films is only by a simple immersion process,the silver nanowires(AgNWs)undergo oxidation to Ag^(+)ions through electron loss,and the liberated electrons provide driving force for the deposition of WO_(4)^(2-).Enabled the fabrication of large-area WO_(3)films and ESWs were fabricated under minimal laboratory conditions,demonstrating the economic feasibility,efficient and reliable nature of industrial production.Structural characterization and density functional theory calculations were combined to confirm that AgNWs effectively regulate oxygen vacancies of WO_(3)films and promote the in situ growth process.The optimized WO_(3)exhibits a maximum transmittance modulation of 90.8%and excellent cycling stability of 20,000 cycles.The largescale WO_(3)-based ESWs can save building energy up to 140.0 MJ m^(-2)compared to traditional windows in tropical regions,as verified by simulations more than40 global cities.This research provides a new approach for improving the performance and industrial production of ESW,providing the full understanding and development direction to short the distance of the ESW commercial production.展开更多
In materials science and engineering design,high-fidelity and high-efficiency numerical simulation has become a driving force for innovation and practical implementation.To address longstanding bottlenecks in the deve...In materials science and engineering design,high-fidelity and high-efficiency numerical simulation has become a driving force for innovation and practical implementation.To address longstanding bottlenecks in the development of conventional material constitutive models—such as lengthy modeling cycles and difficulties in numerical implementation—this study proposes an intelligent modeling and code generation approach powered by large languagemodels.A structured knowledge base integrating constitutive theory,numerical algorithms,and UMAT(User Material)interface specifications is constructed,and a retrieval-augmented generation strategy is employed to establish an end-to-end workflow spanning experimental data parsing,constitutive model formulation,and automatic UMAT subroutine generation.Experimental results show that the method achieves high accuracy for both a classical Johnson–Cookmodel and a physics-informed neural network(PINN)model,with key parameter identification errors below 5%.Moreover,the automatically generated UMAT subroutines yield finite element simulation results in Abaqus that are highly consistent with theoretical predictions(coefficient of determination R2>0.98)while maintaining good numerical stability.This framework is currently focused on the automatic construction of rate-dependent elastoplastic material models,and its core method also provides a clear path for extending to other constitutive categories such as hyperelasticity and viscoelasticity.This work provides an effective technical route for the rapid development and reliable numerical implementation of material constitutive models,significantly advancing the intelligence level of computational mechanics research and improving engineering application efficiency.展开更多
This study demonstrates a novel integration of large language models,machine learning,and multicriteria decision-making to investigate self-moderation in small online communities,a topic under-explored compared to use...This study demonstrates a novel integration of large language models,machine learning,and multicriteria decision-making to investigate self-moderation in small online communities,a topic under-explored compared to user behavior and platform-driven moderation on social media.The proposed methodological framework(1)utilizes large language models for social media post analysis and categorization,(2)employs k-means clustering for content characterization,and(3)incorporates the TODIM(Tomada de Decisão Interativa Multicritério)method to determine moderation strategies based on expert judgments.In general,the fully integrated framework leverages the strengths of these intelligent systems in a more systematic evaluation of large-scale decision problems.When applied in social media moderation,this approach promotes nuanced and context-sensitive self-moderation by taking into account factors such as cultural background and geographic location.The application of this framework is demonstrated within Facebook groups.Eight distinct content clusters encompassing safety,harassment,diversity,and misinformation are identified.Analysis revealed a preference for content removal across all clusters,suggesting a cautious approach towards potentially harmful content.However,the framework also highlights the use of other moderation actions,like account suspension,depending on the content category.These findings contribute to the growing body of research on self-moderation and offer valuable insights for creating safer and more inclusive online spaces within smaller communities.展开更多
Background:Assess ChatGPT and Bard's effectiveness in the initial identification of articles for Otolaryngology—Head and Neck Surgery systematic literature reviews.Methods:Three PRISMA-based systematic reviews(Ja...Background:Assess ChatGPT and Bard's effectiveness in the initial identification of articles for Otolaryngology—Head and Neck Surgery systematic literature reviews.Methods:Three PRISMA-based systematic reviews(Jabbour et al.2017,Wong et al.2018,and Wu et al.2021)were replicated using ChatGPTv3.5 and Bard.Outputs(author,title,publication year,and journal)were compared to the original references and cross-referenced with medical databases for authenticity and recall.Results:Several themes emerged when comparing Bard and ChatGPT across the three reviews.Bard generated more outputs and had greater recall in Wong et al.'s review,with a broader date range in Jabbour et al.'s review.In Wu et al.'s review,ChatGPT-2 had higher recall and identified more authentic outputs than Bard-2.Conclusion:Large language models(LLMs)failed to fully replicate peer-reviewed methodologies,producing outputs with inaccuracies but identifying relevant,especially recent,articles missed by the references.While human-led PRISMA-based reviews remain the gold standard,refining LLMs for literature reviews shows potential.展开更多
Large language models(LLMs)have revolutionized AI applications across diverse domains.However,their widespread deployment has introduced critical security vulnerabilities,particularly prompt injection attacks that man...Large language models(LLMs)have revolutionized AI applications across diverse domains.However,their widespread deployment has introduced critical security vulnerabilities,particularly prompt injection attacks that manipulate model behavior through malicious instructions.Following Kitchenham’s guidelines,this systematic review synthesizes 128 peer-reviewed studies from 2022 to 2025 to provide a unified understanding of this rapidly evolving threat landscape.Our findings reveal a swift progression from simple direct injections to sophisticated multimodal attacks,achieving over 90%success rates against unprotected systems.In response,defense mechanisms show varying effectiveness:input preprocessing achieves 60%–80%detection rates and advanced architectural defenses demonstrate up to 95%protection against known patterns,though significant gaps persist against novel attack vectors.We identified 37 distinct defense approaches across three categories,but standardized evaluation frameworks remain limited.Our analysis attributes these vulnerabilities to fundamental LLM architectural limitations,such as the inability to distinguish instructions from data and attention mechanism vulnerabilities.This highlights critical research directions such as formal verification methods,standardized evaluation protocols,and architectural innovations for inherently secure LLM designs.展开更多
War rehearsals have become increasingly important in national security due to the growing complexity of international affairs.However,traditional rehearsal methods,such as military chess simulations,are inefficient an...War rehearsals have become increasingly important in national security due to the growing complexity of international affairs.However,traditional rehearsal methods,such as military chess simulations,are inefficient and inflexible,with particularly pronounced limitations in command and decision-making.The overwhelming volume of information and high decision complexity hinder the realization of autonomous and agile command and control.To address this challenge,an intelligent warfare simulation framework named Command-Agent is proposed,which deeply integrates large language models(LLMs)with digital twin battlefields.By constructing a highly realistic battlefield environment through real-time simulation and multi-source data fusion,the natural language interaction capabilities of LLMs are leveraged to lower the command threshold and to enable autonomous command through the Observe-Orient-Decide-Act(OODA)feedback loop.Within the Command-Agent framework,a multimodel collaborative architecture is further adopted to decouple the decision-generation and command-execution functions of LLMs.By combining specialized models such as Deep Seek-R1 and MCTool,the limitations of single-model capabilities are overcome.MCTool is a lightweight execution model fine-tuned for military Function Calling tasks.The framework also introduces a Vector Knowledge Base to mitigate hallucinations commonly exhibited by LLMs.Experimental results demonstrate that Command-Agent not only enables natural language-driven simulation and control but also deeply understands commander intent.Leveraging the multi-model collaborative architecture,during red-blue UAV confrontations involving 2 to 8 UAVs,the integrated score is improved by an average of 41.8%compared to the single-agent system(MCTool),accompanied by a 161.8%optimization in the battle loss ratio.Furthermore,when compared with multi-agent systems lacking the knowledge base,the inclusion of the Vector Knowledge Base further improves overall performance by 16.8%.In comparison with the general model(Qwen2.5-7B),the fine-tuned MCTool leads by 5%in execution efficiency.Therefore,the proposed Command-Agent introduces a novel perspective to the military command system and offers a feasible solution for intelligent battlefield decision-making.展开更多
This study investigates the turbulence-induced disturbances and stall precursor triggering mechanism in NACA65-18(10)cascade based on large eddy simulations.The results indicate that the disturbances exist under vario...This study investigates the turbulence-induced disturbances and stall precursor triggering mechanism in NACA65-18(10)cascade based on large eddy simulations.The results indicate that the disturbances exist under various operating conditions along the performance curve.The shear layer is the physical structure responsible for the generation,propagation,and dissipation of disturbances.When operating near stall,the separation on the suction surface intensifies,and strong unsteady backflow occurs at the trailing edge of the passage.Under the influence of inlet disturbances,unsteady behaviors between passages form specific phase differences,leading the entire system to oscillate in a first-order mode.As the flow develops from near-stall to stall,axial momentum decreases further,reducing the main flow’s ability to drive blockages downstream through convection.Consequently,the blockage accumulates during the circumferential propagation process until the stall onset.Based on the above mechanism,this study proposes factors describing the size of the backflow zone,shedding frequency,and convection velocity to characterize blockage dynamics,identifying critical values that represent the stall onset.展开更多
The giant impact hypothesis for the Moon's origin has had difficulty explaining the nearly identical isotopic compositions of Moon rocks and rocks from Earth's silicate mantle and crust.These similarities are ...The giant impact hypothesis for the Moon's origin has had difficulty explaining the nearly identical isotopic compositions of Moon rocks and rocks from Earth's silicate mantle and crust.These similarities are instead more compatible with the Darwin-Wise hypothesis that the Moon arose by fission of a rapidly spinning Earth.To overcome problems with the fission model concerning structural stability and angular momentum conservation,some authors suggested that lunar fission was feasible on a more slowly rotating Earth if assisted by a nuclear explosion near the core-mantle boundary.In this light we consider the possible roles of the large low-velocity provinces(LLVPs).These long-lived structures have been implicated in diverse geophysical processes ranging from deep mantle plumes to continental breakup and mass extinction events.While the LLVPs have been seen as possible remnants of the giant imp actor,we propose that one of them was the site of lunar ejection.Internal heating of the liquid core is suggested to have given rise to an equatorial belt just under the core-mantle boundary analogous to the one recently detected by Ma and Tkalcic[Sci Adv 10(35):eadn5562,2024].Upwellings of heat and volatiles from this belt then generated two antipodal,equatorial bulges:the precursors of the Pacific and African LLVPs.Prior to the emergence of plate tectonics,core heat was mainly dissipated by networks of deep mantle plumes extending above the proto-LLVPs.These plume networks represent conduits of weakened mantle through which proto-lunar materials could later rise in a focused ejection.Continuing heat buildup in the core eventually triggered a cataclysmic explosion in the Pacific proto-LLVP,possibly analogous to a planetary-scale kimberlite eruption.This explosion launched LLVP and overlying mantle material into a low Earth orbit,where it coalesced to form the Moon.Some possible sources of additional energy to power the explosion are considered,including nuclear fission,bolide impacts and a hypothetical gravitational decay process culminating in a'A event'.展开更多
Recommendation systems are key to boosting user engagement,satisfaction,and retention,particularly on media platforms where personalized content is vital.Sequential recommendation systems learn from user-item interact...Recommendation systems are key to boosting user engagement,satisfaction,and retention,particularly on media platforms where personalized content is vital.Sequential recommendation systems learn from user-item interactions to predict future items of interest.However,many current methods rely on unique user and item IDs,limiting their ability to represent users and items effectively,especially in zero-shot learning scenarios where training data is scarce.With the rapid development of Large Language Models(LLMs),researchers are exploring their potential to enhance recommendation systems.However,there is a semantic gap between the linguistic semantics of LLMs and the collaborative semantics of recommendation systems,where items are typically indexed by IDs.Moreover,most research focuses on item representations,neglecting personalized user modeling.To address these issues,we propose a sequential recommendation framework using LLMs,called CIT-Rec,a model that integrates Collaborative semantics for user representation and Image and Text information for item representation to enhance Recommendations.Specifically,by aligning intuitive image information with text containing semantic features,we can more accurately represent items,improving item representation quality.We focus not only on item representations but also on user representations.To more precisely capture users’personalized preferences,we use traditional sequential recommendation models to train on users’historical interaction data,effectively capturing behavioral patterns.Finally,by combining LLMs and traditional sequential recommendation models,we allow the LLM to understand linguistic semantics while capturing collaborative semantics.Extensive evaluations on real-world datasets show that our model outperforms baseline methods,effectively combining user interaction history with item visual and textual modalities to provide personalized recommendations.展开更多
文摘As a common practice,a large hydro-generator will operate in leading phase conditions to absorb the reactive power of the power grid.However,the accurate and precise prediction of the leading phase operation capacity of a large hydro-generator has always been a formidable challenge to engineers and academicians because it is extremely hard to compute the eddy currents and losses as well as the local overheating in the pressure plate and finger.To address this problem,a full three dimensional(3D)finite element model and method of the coupled eddy current and temperature fields in the end region of a large hydro-generator are developed.The equivalent medium parameters used in the computations are comprehensively discussed.Moreover,some numerically based solution methodologies for accurate computation of the field and armature currents under different leading phase conditions are proposed.Numerical results on the coupled eddy current and temperature fields in the end regions of a 250 MW hydro-generator confirm positively the feasibility of the present work.
文摘In this paper,we establish some strong laws of large numbers,which are for nonindependent random variables under the framework of sublinear expectations.One of our main results is for blockwise m-dependent random variables,and another is for sub-orthogonal random variables.Both extend the strong law of large numbers for independent random variables under sublinear expectations to the non-independent case.
文摘Model evaluation using benchmark datasets is an important method to measure the capability of large language models(LLMs)in specific domains,and it is mainly used to assess the knowledge and reasoning abilities of LLMs.Therefore,in order to better assess the capability of LLMs in the agricultural domain,Agri-Eval was proposed as a benchmark for assessing the knowledge and reasoning ability of LLMs in agriculture.The assessment dataset used in Agri-Eval covered seven major disciplines in the agricultural domain:crop science,horticulture,plant protection,animal husbandry,forest science,aquaculture science,and grass science,and contained a total of 2283 questions.Among domestic general-purpose LLMs,DeepSeek R1 performed best with an accuracy rate of 75.49%.In the realm of international general-purpose LLMs,Gemini 2.0 pro exp 0205 standed out as the top performer,achieving an accuracy rate of 74.28%.As an LLMs in agriculture vertical,Shennong V2.0 outperformed all the LLMs in China,and the answer accuracy rate of agricultural knowledge exceeded that of all the existing general-purpose LLMs.The launch of Agri-Eval helped the LLM developers to comprehensively evaluate the model's capability in the field of agriculture through a variety of tasks and tests to promote the development of the LLMs in the field of agriculture.
文摘To the Editor,Artificial intelligence(AI)usage has been increasing.Many fields have implemented the use of AI and Large LanguageModels(LLMs),especially in medicine.Furthermore,manypatients have increasingly been using AI;often,they will prompt AI with questions before even stepping into a physi-cian's office.The question lies in whether the information produced by AI is reliable and if this information is concise and easy to read across all patient populations.
基金funded by the National Key Research and Development Program of China(No.2021YFA1100500)the National Natural Science Foundation of China(No.82370662)the Key Research&Development Plan of Zhejiang Province(No.2024C03051).
文摘This study evaluated the accuracy,completeness,and comprehensibility of responses from mainstream large language models(LLMs)to hepatitis C virus(HCV)-related questions,aiming to assess their performance in addressing patient queries about disease and lifestyle behaviors.The models selected were ChatGPT-4o,Gemini 2.0 Pro,Claude 3.5 Sonnet,and DeepSeek V3,with 12 questions chosen by two HCV experts from the domains of prevention,diagnosis,and treatment.
基金supported by University Grant Agency of Matej Bel University in Banská Bystrica project number UGA-14-PDS-2025.
文摘It is known that correlation does not imply causality.Some relationships identified in the analysis of data are coincidental or unknown,and some are produced by real-world causality of the situation,which is problematic,since there is a need to differentiate between these two scenarios.Until recently,the proper−semantic−causality of the relationship could have been determined only by human experts from the area of expertise of the studied data.This has changed with the advance of large language models,which are often utilized as surrogates for such human experts,making the process automated and readily available to all data analysts.This motivates the main objective of this work,which is to introduce the design and implementation of a large language model-based semantic causality evaluator based on correlation analysis,together with its visual analysis model called Causal heatmap.After the implementation itself,the model is evaluated from the point of view of the quality of the visual model,from the point of view of the quality of causal evaluation based on large language models,and from the point of view of comparative analysis,while the results reached in the study highlight the usability of large language models in the task and the potential of the proposed approach in the analysis of unknown datasets.The results of the experimental evaluation demonstrate the usefulness of the Causal heatmap method,supported by the evident highlighting of interesting relationships,while suppressing irrelevant ones.
基金National Key Research and Development Program of China,Grant/Award Number:2024YFF0507404Special Clinical Business Fund for High-Level Hospitals of China-Japan Friendship Hospital,Grant/Award Number:2024-NHLHCRF-TS-01。
文摘Background:Large language models(LLMs)have shown considerable promise in supporting clinical decision-making.However,their adoption and evaluation in dermatology remains limited.This study aimed to explore the preferences of Chinese dermatologists regarding LLM-generated responses in clinical psoriasis scenarios and to assess how they prioritize key quality dimensions,including accuracy,traceability,and logicality.Methods:A cross-sectional,web-based survey was conducted between December 25,2024,and January 22,2025,following the Checklist for Reporting Results of Internet E-Surveys guidelines.A total of 1247 valid responses were collected from practicing dermatologists across 33 of China's provincial-level administrative divisions.Participants evaluated responses to five categories of clinical questions(etiology,clinical presentation,differential diagnosis,treatment,and case study)generated by five LLMs:ChatGPT-4o,Kimi.ai,Doubao,ZuoYiGPT,and Lingyi-agent.Statistical associations between participant characteristics and model preferences were examined using chi-square tests.Results:ChatGPT-4o(Model 1)emerged as the most preferred model across all clinical tasks,consistently receiving the highest number of votes in case study(n=740),clinical presentation(n=666),differential diagnosis(n=707),etiology(n=602),and treatment(n=656).Significant variation in model preference by professional title was observed only for the differential diagnosis task(χ^(2)=21.13,df=12,p=0.0485),while no significant differences were found across hospital tiers(p>0.05).In terms of evaluation dimensions,accuracy was most frequently rated as“very important”(n=635).A significant association existed between hospital tier and the most valued dimension(χ^(2)=27.667,df=9,p=0.0011),with dermatologists in primary hospitals prioritizing traceability more than their peers in higher-tier hospitals.No significant associations were found across professional titles(p=0.127).Conclusions:Chinese dermatologists suggest a strong preference for ChatGPT-4o over domestic LLMs in psoriasis-related clinical tasks.While accuracy remains the primary criterion,traceability and logicality are also critical,particularly for clinicians in lower-tier hospitals.These findings suggest that future clinical LLMs should prioritize not only content accuracy but also source transparency and structural clarity to meet the diverse needs of different clinical settings.
基金Guangzhou Science and Technology Program,Grant/Award Numbers:2025B03J0110,2024A03J1074,2024A03J0927。
文摘Large language models(LLMs)show considerable potential to revolutionize healthcare through their performance across diverse clinical applications.Given the inherent constraints of LLMs and the critical nature of medical practice,a rigorous and systematic evaluation of their medical competence is imperative.This study presents a comprehensive review of the established methodologies and benchmarks for evaluating the medical competence of LLMs,encompassing a thorough analysis of current assessment practices across medical knowledge,clinical practice competence,and ethical-safety considerations.By integrating clinician competency assessment frameworks into LLMs evaluation,we propose a structured tri-dimensional framework that systematically organizes existing evaluation approaches according to medical theoretical knowledge,clinical practice ability,and ethical-safety considerations.Furthermore,this research provides critical insights into future developmental trajectories while establishing foundational frameworks and standardization protocols for the integration of LLMs into medical practice.
文摘Background:Despite the promise shown by large language models(LLMs)for standardized tasks,their multidimensional performance in real-world oncology decision-making remains unevaluated.This study aims to introduce a framework for evaluating LLMs and physician decisions in challenging lung cancer cases.Methods:We curated 50 challenging lung cancer cases(25 local and 25 published)classified as complex,rare,or refractory.Blinded three-dimensional,five-point Likert evaluations(1–5 for comprehensiveness,specificity,and readability)compared standalone LLMs(DeepSeek R1,Claude 3.5,Gemini 1.5,and GPT-4o),physicians by experience level(junior,intermediate,and senior),and AI-assisted juniors;intergroup differences and augmentation effects were analyzed statistically.Results:Of 50 challenging cases(18 complex,17 rare,and 15 refractory)rated by three experts,DeepSeek R1 achieved scores of 3.95±0.33,3.71±0.53,and 4.26±0.18 for comprehensiveness,specificity,and readability,respectively,positioning it between intermediate(3.68,3.68,3.75)and senior(4.50,4.64,4.53)physicians.GPT-4o and Claude 3.5 reached intermediate physician–level comprehensiveness(3.76±0.39,3.60±0.39)but junior-to-intermediate physician–level specificity(3.39±0.39,3.39±0.49).All LLMs scored higher on rare cases than intermediate physicians but fell below junior physicians in refractory-case specificity.AIassisted junior physicians showed marked gains in rare cases,with comprehensiveness rising from 2.32 to 4.29(84.8%),specificity from 2.24 to 4.26(90.8%),and readability from 2.76 to 4.59(66.0%),while specificity declined by 3.2%(3.17 to 3.07)in refractory cases.Error analysis showed complementary strengths,with physicians demonstrating reasoning stability and LLMs excelling in knowledge updating and risk management.Conclusions:LLMs performed variably in clinical decision-making tasks depending on case type,performing better in rare cases and worse in refractory cases requiring longitudinal reasoning.Complementary strengths between LLMs and physicians support case-and task-tailored human–AI collaboration.
基金supported by the Ningbo Natural Science Foundation(Youth Foundation,No.2024J449)the Scientific Research Foundation for Introduced Talents of Ningbo University(Nos.ZX2022000602 and ZX2024000043)。
文摘Polyfluoroalkyl substances(PFAS)have emerged as persistent environmental contaminants because of their chemical stability,degradation-resistance and bioaccumulation potential.However,current studies mainly focus on the toxicity of single PFAS such as perfluorooctanoic acid(PFOA)and perfluorobutanoic acid(PFBA),the knowledge of their combined effects is relatively limited.In this study,we explored the immune response of the gut in large yellow croaker(Larimichthys crocea)under the combined stress of PFOA and PFBA.Histologicalanalyses revealed that the combined effect induced intestinal vacuolization and decreased the length of intestinal villi.And it significantly activated pro-inflammatory pathways with marked upregulation of tnfα,il1β,il6 and myd88 expressions,particularly after 14 days of exposure.Gut microbiota analysis revealed substantial dysbiosis,including 1)reduced alpha diversity,2)increased abundance of potential pathogenic taxa(Proteobacteria and Spirochaetota),and 3)depletion of beneficial Firmicutes.PICRUSt-based functional prediction indicated temporal metabolic shifts,with upregulation of DNA repair pathways at day 3 and enhanced bacterial motility protein activity at days 7 and 14 of post-exposure.The Pearson correlation analysis further indicated that these immune genes had significant positive correlations with Vibrio and Brevinema,and negative correlations with Streptococcus.Our present study will provide novel insights into the microbiome-mediated immunomodulation in the larger yellow croaker exposed to combined PFAS,which will be helpful for healthy farming of economically important marine species.
基金supported in part by the Mining Hydraulic Technology and Equipment Engineering Research Center,Liaoning Technical University,Fuxin,China(Grant No.MHTE23-R04)the Fundamental Research Funds for the Central Universities(ID N25BSS068).
文摘This study presents an implicit multiphysics coupling method integrating Computational Fluid Dynamics(CFD),the Multiphase Particle-in-Cell(MPPIC)model,and the Finite Element Method(FEM),implemented with OpenFOAM,CalculiX,and preCICE to simulate fluid-particle-structure interactions with large deformations.Mesh motion in the fluid field is handled using the radial basis function(RBF)method.The particle phase is modeled by MPPIC,where fluid-particle interaction is described through momentum exchange,and inter-particle collisions are characterized by collision stress.The structural field is solved by nonlinear FEM to capture large deformations induced by geometric nonlinearity.Coupling among fields is realized through a partitioned,parallel,and non-intrusive iterative strategy,ensuring stable transfer and convergence of interface forces and displacements.Notably,the influence of particles on the structure is not direct but mediated by the fluid,while structural motion directly affects particle dynamics.The results demonstrate that the proposed approach effectively captures multiphysics interaction processes and provides a valuable reference for numerical modeling of coupled fluid-particle-structure systems.
基金the National Natural Science Foundation of China(grant No.52163022,62305076)Sichuan Science and Technology Program(2024ZYD0196)+1 种基金China Postdoctoral Science Foundation(2023M740505)Sichuan Postdoctoral Science Special Foundation(No.TB2023010)。
文摘Electrochromic smart windows(ESWs)can significantly reduce building energy consumption,but the high cost hinders large-scale production.The in situ growth of tungsten oxide(WO_(3))films is only by a simple immersion process,the silver nanowires(AgNWs)undergo oxidation to Ag^(+)ions through electron loss,and the liberated electrons provide driving force for the deposition of WO_(4)^(2-).Enabled the fabrication of large-area WO_(3)films and ESWs were fabricated under minimal laboratory conditions,demonstrating the economic feasibility,efficient and reliable nature of industrial production.Structural characterization and density functional theory calculations were combined to confirm that AgNWs effectively regulate oxygen vacancies of WO_(3)films and promote the in situ growth process.The optimized WO_(3)exhibits a maximum transmittance modulation of 90.8%and excellent cycling stability of 20,000 cycles.The largescale WO_(3)-based ESWs can save building energy up to 140.0 MJ m^(-2)compared to traditional windows in tropical regions,as verified by simulations more than40 global cities.This research provides a new approach for improving the performance and industrial production of ESW,providing the full understanding and development direction to short the distance of the ESW commercial production.
基金funded by the National Natural Science Foundation of China,grant number 52405341Foundation of National Key Laboratory of Computational Physics,grant number 6142A05QN24012+1 种基金Chongqing Science and Technology Committee,grant number CSTB2023NSCQ-MSX0363The Science and Technology Research Program of Chongqing Municipal Education Commission,grant number KJQN202301117.
文摘In materials science and engineering design,high-fidelity and high-efficiency numerical simulation has become a driving force for innovation and practical implementation.To address longstanding bottlenecks in the development of conventional material constitutive models—such as lengthy modeling cycles and difficulties in numerical implementation—this study proposes an intelligent modeling and code generation approach powered by large languagemodels.A structured knowledge base integrating constitutive theory,numerical algorithms,and UMAT(User Material)interface specifications is constructed,and a retrieval-augmented generation strategy is employed to establish an end-to-end workflow spanning experimental data parsing,constitutive model formulation,and automatic UMAT subroutine generation.Experimental results show that the method achieves high accuracy for both a classical Johnson–Cookmodel and a physics-informed neural network(PINN)model,with key parameter identification errors below 5%.Moreover,the automatically generated UMAT subroutines yield finite element simulation results in Abaqus that are highly consistent with theoretical predictions(coefficient of determination R2>0.98)while maintaining good numerical stability.This framework is currently focused on the automatic construction of rate-dependent elastoplastic material models,and its core method also provides a clear path for extending to other constitutive categories such as hyperelasticity and viscoelasticity.This work provides an effective technical route for the rapid development and reliable numerical implementation of material constitutive models,significantly advancing the intelligence level of computational mechanics research and improving engineering application efficiency.
基金funded by the Office of the Vice-President for Research and Development of Cebu Technological University.
文摘This study demonstrates a novel integration of large language models,machine learning,and multicriteria decision-making to investigate self-moderation in small online communities,a topic under-explored compared to user behavior and platform-driven moderation on social media.The proposed methodological framework(1)utilizes large language models for social media post analysis and categorization,(2)employs k-means clustering for content characterization,and(3)incorporates the TODIM(Tomada de Decisão Interativa Multicritério)method to determine moderation strategies based on expert judgments.In general,the fully integrated framework leverages the strengths of these intelligent systems in a more systematic evaluation of large-scale decision problems.When applied in social media moderation,this approach promotes nuanced and context-sensitive self-moderation by taking into account factors such as cultural background and geographic location.The application of this framework is demonstrated within Facebook groups.Eight distinct content clusters encompassing safety,harassment,diversity,and misinformation are identified.Analysis revealed a preference for content removal across all clusters,suggesting a cautious approach towards potentially harmful content.However,the framework also highlights the use of other moderation actions,like account suspension,depending on the content category.These findings contribute to the growing body of research on self-moderation and offer valuable insights for creating safer and more inclusive online spaces within smaller communities.
文摘Background:Assess ChatGPT and Bard's effectiveness in the initial identification of articles for Otolaryngology—Head and Neck Surgery systematic literature reviews.Methods:Three PRISMA-based systematic reviews(Jabbour et al.2017,Wong et al.2018,and Wu et al.2021)were replicated using ChatGPTv3.5 and Bard.Outputs(author,title,publication year,and journal)were compared to the original references and cross-referenced with medical databases for authenticity and recall.Results:Several themes emerged when comparing Bard and ChatGPT across the three reviews.Bard generated more outputs and had greater recall in Wong et al.'s review,with a broader date range in Jabbour et al.'s review.In Wu et al.'s review,ChatGPT-2 had higher recall and identified more authentic outputs than Bard-2.Conclusion:Large language models(LLMs)failed to fully replicate peer-reviewed methodologies,producing outputs with inaccuracies but identifying relevant,especially recent,articles missed by the references.While human-led PRISMA-based reviews remain the gold standard,refining LLMs for literature reviews shows potential.
基金supported by 2023 Higher Education Scientific Research Planning Project of China Society of Higher Education(No.23PG0408)2023 Philosophy and Social Science Research Programs in Jiangsu Province(No.2023SJSZ0993)+2 种基金Nantong Science and Technology Project(No.JC2023070)Key Project of Jiangsu Province Education Science 14th Five-Year Plan(Grant No.B-b/2024/02/41)the Open Fund of Advanced Cryptography and System Security Key Laboratory of Sichuan Province(Grant No.SKLACSS-202407).
文摘Large language models(LLMs)have revolutionized AI applications across diverse domains.However,their widespread deployment has introduced critical security vulnerabilities,particularly prompt injection attacks that manipulate model behavior through malicious instructions.Following Kitchenham’s guidelines,this systematic review synthesizes 128 peer-reviewed studies from 2022 to 2025 to provide a unified understanding of this rapidly evolving threat landscape.Our findings reveal a swift progression from simple direct injections to sophisticated multimodal attacks,achieving over 90%success rates against unprotected systems.In response,defense mechanisms show varying effectiveness:input preprocessing achieves 60%–80%detection rates and advanced architectural defenses demonstrate up to 95%protection against known patterns,though significant gaps persist against novel attack vectors.We identified 37 distinct defense approaches across three categories,but standardized evaluation frameworks remain limited.Our analysis attributes these vulnerabilities to fundamental LLM architectural limitations,such as the inability to distinguish instructions from data and attention mechanism vulnerabilities.This highlights critical research directions such as formal verification methods,standardized evaluation protocols,and architectural innovations for inherently secure LLM designs.
文摘War rehearsals have become increasingly important in national security due to the growing complexity of international affairs.However,traditional rehearsal methods,such as military chess simulations,are inefficient and inflexible,with particularly pronounced limitations in command and decision-making.The overwhelming volume of information and high decision complexity hinder the realization of autonomous and agile command and control.To address this challenge,an intelligent warfare simulation framework named Command-Agent is proposed,which deeply integrates large language models(LLMs)with digital twin battlefields.By constructing a highly realistic battlefield environment through real-time simulation and multi-source data fusion,the natural language interaction capabilities of LLMs are leveraged to lower the command threshold and to enable autonomous command through the Observe-Orient-Decide-Act(OODA)feedback loop.Within the Command-Agent framework,a multimodel collaborative architecture is further adopted to decouple the decision-generation and command-execution functions of LLMs.By combining specialized models such as Deep Seek-R1 and MCTool,the limitations of single-model capabilities are overcome.MCTool is a lightweight execution model fine-tuned for military Function Calling tasks.The framework also introduces a Vector Knowledge Base to mitigate hallucinations commonly exhibited by LLMs.Experimental results demonstrate that Command-Agent not only enables natural language-driven simulation and control but also deeply understands commander intent.Leveraging the multi-model collaborative architecture,during red-blue UAV confrontations involving 2 to 8 UAVs,the integrated score is improved by an average of 41.8%compared to the single-agent system(MCTool),accompanied by a 161.8%optimization in the battle loss ratio.Furthermore,when compared with multi-agent systems lacking the knowledge base,the inclusion of the Vector Knowledge Base further improves overall performance by 16.8%.In comparison with the general model(Qwen2.5-7B),the fine-tuned MCTool leads by 5%in execution efficiency.Therefore,the proposed Command-Agent introduces a novel perspective to the military command system and offers a feasible solution for intelligent battlefield decision-making.
基金support of the National Natural Science Foundation of China(Nos.52322603 and U24A20141)the Science Center for Gas Turbine Project of China(No.P2023-B-Ⅱ-001-001)the Fundamental Research Funds for the Central Universities of China and the Beijing Nova Program of China(Nos.20220484074 and 20230484479)。
文摘This study investigates the turbulence-induced disturbances and stall precursor triggering mechanism in NACA65-18(10)cascade based on large eddy simulations.The results indicate that the disturbances exist under various operating conditions along the performance curve.The shear layer is the physical structure responsible for the generation,propagation,and dissipation of disturbances.When operating near stall,the separation on the suction surface intensifies,and strong unsteady backflow occurs at the trailing edge of the passage.Under the influence of inlet disturbances,unsteady behaviors between passages form specific phase differences,leading the entire system to oscillate in a first-order mode.As the flow develops from near-stall to stall,axial momentum decreases further,reducing the main flow’s ability to drive blockages downstream through convection.Consequently,the blockage accumulates during the circumferential propagation process until the stall onset.Based on the above mechanism,this study proposes factors describing the size of the backflow zone,shedding frequency,and convection velocity to characterize blockage dynamics,identifying critical values that represent the stall onset.
文摘The giant impact hypothesis for the Moon's origin has had difficulty explaining the nearly identical isotopic compositions of Moon rocks and rocks from Earth's silicate mantle and crust.These similarities are instead more compatible with the Darwin-Wise hypothesis that the Moon arose by fission of a rapidly spinning Earth.To overcome problems with the fission model concerning structural stability and angular momentum conservation,some authors suggested that lunar fission was feasible on a more slowly rotating Earth if assisted by a nuclear explosion near the core-mantle boundary.In this light we consider the possible roles of the large low-velocity provinces(LLVPs).These long-lived structures have been implicated in diverse geophysical processes ranging from deep mantle plumes to continental breakup and mass extinction events.While the LLVPs have been seen as possible remnants of the giant imp actor,we propose that one of them was the site of lunar ejection.Internal heating of the liquid core is suggested to have given rise to an equatorial belt just under the core-mantle boundary analogous to the one recently detected by Ma and Tkalcic[Sci Adv 10(35):eadn5562,2024].Upwellings of heat and volatiles from this belt then generated two antipodal,equatorial bulges:the precursors of the Pacific and African LLVPs.Prior to the emergence of plate tectonics,core heat was mainly dissipated by networks of deep mantle plumes extending above the proto-LLVPs.These plume networks represent conduits of weakened mantle through which proto-lunar materials could later rise in a focused ejection.Continuing heat buildup in the core eventually triggered a cataclysmic explosion in the Pacific proto-LLVP,possibly analogous to a planetary-scale kimberlite eruption.This explosion launched LLVP and overlying mantle material into a low Earth orbit,where it coalesced to form the Moon.Some possible sources of additional energy to power the explosion are considered,including nuclear fission,bolide impacts and a hypothetical gravitational decay process culminating in a'A event'.
基金supported by the National Key R&D Program of China[2022YFF0902703]the State Administration for Market Regulation Science and Technology Plan Project(2024MK033).
文摘Recommendation systems are key to boosting user engagement,satisfaction,and retention,particularly on media platforms where personalized content is vital.Sequential recommendation systems learn from user-item interactions to predict future items of interest.However,many current methods rely on unique user and item IDs,limiting their ability to represent users and items effectively,especially in zero-shot learning scenarios where training data is scarce.With the rapid development of Large Language Models(LLMs),researchers are exploring their potential to enhance recommendation systems.However,there is a semantic gap between the linguistic semantics of LLMs and the collaborative semantics of recommendation systems,where items are typically indexed by IDs.Moreover,most research focuses on item representations,neglecting personalized user modeling.To address these issues,we propose a sequential recommendation framework using LLMs,called CIT-Rec,a model that integrates Collaborative semantics for user representation and Image and Text information for item representation to enhance Recommendations.Specifically,by aligning intuitive image information with text containing semantic features,we can more accurately represent items,improving item representation quality.We focus not only on item representations but also on user representations.To more precisely capture users’personalized preferences,we use traditional sequential recommendation models to train on users’historical interaction data,effectively capturing behavioral patterns.Finally,by combining LLMs and traditional sequential recommendation models,we allow the LLM to understand linguistic semantics while capturing collaborative semantics.Extensive evaluations on real-world datasets show that our model outperforms baseline methods,effectively combining user interaction history with item visual and textual modalities to provide personalized recommendations.