期刊文献+
共找到369,339篇文章
< 1 2 250 >
每页显示 20 50 100
Agri-Eval:Multi-level Large Language Model Valuation Benchmark for Agriculture
1
作者 WANG Yaojun GE Mingliang +2 位作者 XU Guowei ZHANG Qiyu BIE Yuhui 《农业机械学报》 北大核心 2026年第1期290-299,共10页
Model evaluation using benchmark datasets is an important method to measure the capability of large language models(LLMs)in specific domains,and it is mainly used to assess the knowledge and reasoning abilities of LLM... Model evaluation using benchmark datasets is an important method to measure the capability of large language models(LLMs)in specific domains,and it is mainly used to assess the knowledge and reasoning abilities of LLMs.Therefore,in order to better assess the capability of LLMs in the agricultural domain,Agri-Eval was proposed as a benchmark for assessing the knowledge and reasoning ability of LLMs in agriculture.The assessment dataset used in Agri-Eval covered seven major disciplines in the agricultural domain:crop science,horticulture,plant protection,animal husbandry,forest science,aquaculture science,and grass science,and contained a total of 2283 questions.Among domestic general-purpose LLMs,DeepSeek R1 performed best with an accuracy rate of 75.49%.In the realm of international general-purpose LLMs,Gemini 2.0 pro exp 0205 standed out as the top performer,achieving an accuracy rate of 74.28%.As an LLMs in agriculture vertical,Shennong V2.0 outperformed all the LLMs in China,and the answer accuracy rate of agricultural knowledge exceeded that of all the existing general-purpose LLMs.The launch of Agri-Eval helped the LLM developers to comprehensively evaluate the model's capability in the field of agriculture through a variety of tasks and tests to promote the development of the LLMs in the field of agriculture. 展开更多
关键词 large language models assessment systems agricultural knowledge agricultural datasets
在线阅读 下载PDF
LinguTimeX a Framework for Multilingual CTC Detection Using Explainable AI and Natural Language Processing
2
作者 Omar Darwish Shorouq Al-Eidi +4 位作者 Abdallah Al-Shorman Majdi Maabreh Anas Alsobeh Plamen Zahariev Yahya Tashtoush 《Computers, Materials & Continua》 2026年第1期2231-2251,共21页
Covert timing channels(CTC)exploit network resources to establish hidden communication pathways,posing signi cant risks to data security and policy compliance.erefore,detecting such hidden and dangerous threats remain... Covert timing channels(CTC)exploit network resources to establish hidden communication pathways,posing signi cant risks to data security and policy compliance.erefore,detecting such hidden and dangerous threats remains one of the security challenges. is paper proposes LinguTimeX,a new framework that combines natural language processing with arti cial intelligence,along with explainable Arti cial Intelligence(AI)not only to detect CTC but also to provide insights into the decision process.LinguTimeX performs multidimensional feature extraction by fusing linguistic attributes with temporal network patterns to identify covert channels precisely.LinguTimeX demonstrates strong e ectiveness in detecting CTC across multiple languages;namely English,Arabic,and Chinese.Speci cally,the LSTM and RNN models achieved F1 scores of 90%on the English dataset,89%on the Arabic dataset,and 88%on the Chinese dataset,showcasing their superior performance and ability to generalize across multiple languages. is highlights their robustness in detecting CTCs within security systems,regardless of the language or cultural context of the data.In contrast,the DeepForest model produced F1-scores ranging from 86%to 87%across the same datasets,further con rming its e ectiveness in CTC detection.Although other algorithms also showed reasonable accuracy,the LSTM and RNN models consistently outperformed them in multilingual settings,suggesting that deep learning models might be better suited for this particular problem. 展开更多
关键词 Arabic language chinese language covert timing channel cYBERSEcURITY deep learning English language language processing machine learning
在线阅读 下载PDF
CIT-Rec:Enhancing Sequential Recommendation System with Large Language Models
3
作者 Ziyu Li Zhen Chen +2 位作者 Xuejing Fu Tong Mo Weiping Li 《Computers, Materials & Continua》 2026年第3期2328-2343,共16页
Recommendation systems are key to boosting user engagement,satisfaction,and retention,particularly on media platforms where personalized content is vital.Sequential recommendation systems learn from user-item interact... Recommendation systems are key to boosting user engagement,satisfaction,and retention,particularly on media platforms where personalized content is vital.Sequential recommendation systems learn from user-item interactions to predict future items of interest.However,many current methods rely on unique user and item IDs,limiting their ability to represent users and items effectively,especially in zero-shot learning scenarios where training data is scarce.With the rapid development of Large Language Models(LLMs),researchers are exploring their potential to enhance recommendation systems.However,there is a semantic gap between the linguistic semantics of LLMs and the collaborative semantics of recommendation systems,where items are typically indexed by IDs.Moreover,most research focuses on item representations,neglecting personalized user modeling.To address these issues,we propose a sequential recommendation framework using LLMs,called CIT-Rec,a model that integrates Collaborative semantics for user representation and Image and Text information for item representation to enhance Recommendations.Specifically,by aligning intuitive image information with text containing semantic features,we can more accurately represent items,improving item representation quality.We focus not only on item representations but also on user representations.To more precisely capture users’personalized preferences,we use traditional sequential recommendation models to train on users’historical interaction data,effectively capturing behavioral patterns.Finally,by combining LLMs and traditional sequential recommendation models,we allow the LLM to understand linguistic semantics while capturing collaborative semantics.Extensive evaluations on real-world datasets show that our model outperforms baseline methods,effectively combining user interaction history with item visual and textual modalities to provide personalized recommendations. 展开更多
关键词 Large language models vision language models sequential recommendation instruction tuning
在线阅读 下载PDF
Detection of Maliciously Disseminated Hate Speech in Spanish Using Fine-Tuning and In-Context Learning Techniques with Large Language Models
4
作者 Tomás Bernal-Beltrán RonghaoPan +3 位作者 JoséAntonio García-Díaz María del Pilar Salas-Zárate Mario Andrés Paredes-Valverde Rafael Valencia-García 《Computers, Materials & Continua》 2026年第4期353-390,共38页
The malicious dissemination of hate speech via compromised accounts,automated bot networks and malware-driven social media campaigns has become a growing cybersecurity concern.Automatically detecting such content in S... The malicious dissemination of hate speech via compromised accounts,automated bot networks and malware-driven social media campaigns has become a growing cybersecurity concern.Automatically detecting such content in Spanish is challenging due to linguistic complexity and the scarcity of annotated resources.In this paper,we compare two predominant AI-based approaches for the forensic detection of malicious hate speech:(1)finetuning encoder-only models that have been trained in Spanish and(2)In-Context Learning techniques(Zero-and Few-Shot Learning)with large-scale language models.Our approach goes beyond binary classification,proposing a comprehensive,multidimensional evaluation that labels each text by:(1)type of speech,(2)recipient,(3)level of intensity(ordinal)and(4)targeted group(multi-label).Performance is evaluated using an annotated Spanish corpus,standard metrics such as precision,recall and F1-score and stability-oriented metrics to evaluate the stability of the transition from zero-shot to few-shot prompting(Zero-to-Few Shot Retention and Zero-to-Few Shot Gain)are applied.The results indicate that fine-tuned encoder-only models(notably MarIA and BETO variants)consistently deliver the strongest and most reliable performance:in our experiments their macro F1-scores lie roughly in the range of approximately 46%–66%depending on the task.Zero-shot approaches are much less stable and typically yield substantially lower performance(observed F1-scores range approximately 0%–39%),often producing invalid outputs in practice.Few-shot prompting(e.g.,Qwen 38B,Mistral 7B)generally improves stability and recall relative to pure zero-shot,bringing F1-scores into a moderate range of approximately 20%–51%but still falling short of fully fine-tuned models.These findings highlight the importance of supervised adaptation and discuss the potential of both paradigms as components in AI-powered cybersecurity and malware forensics systems designed to identify and mitigate coordinated online hate campaigns. 展开更多
关键词 Hate speech detection malicious communication campaigns AI-driven cybersecurity social media analytics large language models prompt-tuning fine-tuning in-context learning natural language processing
在线阅读 下载PDF
On the Evolutionary Logic of Chinese Culture’s Integration Into Foreign Language Education in China:A Bibliometric Study of CSSCI Source Journals(1980-2025)
5
作者 ZOU Yanqun 《Sino-US English Teaching》 2026年第1期1-9,共9页
This paper undertakes a systematic combing of the development of research on integrating Chinese culture into foreign language education in China from the 1980s to 2025,dividing it into three stages:cultural attachmen... This paper undertakes a systematic combing of the development of research on integrating Chinese culture into foreign language education in China from the 1980s to 2025,dividing it into three stages:cultural attachment,cultural compensation,and cultural symbiosis,and reveals the logical shift of the research from the dominance of target language culture to the construction of the subjectivity of Chinese culture.Through quantitative and qualitative analysis of 435 CSSCI papers,three core themes are extracted:what to integrate,why to integrate,and how to integrate.This paper critically analyzes three pairs of contradictions:the imbalance between instrumentality and humanism,the separation of national narrative and individual expression,and the disconnection between traditional inheritance and modern transformation.It is proposed that future research should reconstruct the educational logic based on the Chinese context,integrate the national and individual dimensions,and build a dialogue mechanism between tradition and modernity,so as to provide theoretical and practical reference for the construction of a foreign language education system with Chinese characteristics. 展开更多
关键词 chinese culture foreign language education cultural integration
在线阅读 下载PDF
When Large Language Models and Machine Learning Meet Multi-Criteria Decision Making: Fully Integrated Approach for Social Media Moderation
6
作者 Noreen Fuentes Janeth Ugang +4 位作者 Narcisan Galamiton Suzette Bacus Samantha Shane Evangelista Fatima Maturan Lanndon Ocampo 《Computers, Materials & Continua》 2026年第1期2137-2162,共26页
This study demonstrates a novel integration of large language models,machine learning,and multicriteria decision-making to investigate self-moderation in small online communities,a topic under-explored compared to use... This study demonstrates a novel integration of large language models,machine learning,and multicriteria decision-making to investigate self-moderation in small online communities,a topic under-explored compared to user behavior and platform-driven moderation on social media.The proposed methodological framework(1)utilizes large language models for social media post analysis and categorization,(2)employs k-means clustering for content characterization,and(3)incorporates the TODIM(Tomada de Decisão Interativa Multicritério)method to determine moderation strategies based on expert judgments.In general,the fully integrated framework leverages the strengths of these intelligent systems in a more systematic evaluation of large-scale decision problems.When applied in social media moderation,this approach promotes nuanced and context-sensitive self-moderation by taking into account factors such as cultural background and geographic location.The application of this framework is demonstrated within Facebook groups.Eight distinct content clusters encompassing safety,harassment,diversity,and misinformation are identified.Analysis revealed a preference for content removal across all clusters,suggesting a cautious approach towards potentially harmful content.However,the framework also highlights the use of other moderation actions,like account suspension,depending on the content category.These findings contribute to the growing body of research on self-moderation and offer valuable insights for creating safer and more inclusive online spaces within smaller communities. 展开更多
关键词 Self-moderation user-generated content k-means clustering TODIM large language models
在线阅读 下载PDF
Task-Structured Curriculum Learning for Multi-Task Distillation:Enhancing Step-by-Step Knowledge Transfer in Language Models
7
作者 Ahmet Ezgi Aytug Onan 《Computers, Materials & Continua》 2026年第3期1647-1673,共27页
Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Re... Knowledge distillation has become a standard technique for compressing large language models into efficient student models,but existing methods often struggle to balance prediction accuracy with explanation quality.Recent approaches such as Distilling Step-by-Step(DSbS)introduce explanation supervision,yet they apply it in a uniform manner that may not fully exploit the different learning dynamics of prediction and explanation.In this work,we propose a task-structured curriculum learning(TSCL)framework that structures training into three sequential phases:(i)prediction-only,to establish stable feature representations;(ii)joint prediction-explanation,to align task outputs with rationale generation;and(iii)explanation-only,to refine the quality of rationales.This design provides a simple but effective modification to DSbS,requiring no architectural changes and adding negligible training cost.We justify the phase scheduling with ablation studies and convergence analysis,showing that an initial prediction-heavy stage followed by a balanced joint phase improves both stability and explanation alignment.Extensive experiments on five datasets(e-SNLI,ANLI,CommonsenseQA,SVAMP,and MedNLI)demonstrate that TSCL consistently outperforms strong baselines,achieving gains of+1.7-2.6 points in accuracy and 0.8-1.2 in ROUGE-L,corresponding to relative error reductions of up to 21%.Beyond lexical metrics,human evaluation and ERASERstyle faithfulness diagnostics confirm that TSCL produces more faithful and informative explanations.Comparative training curves further reveal faster convergence and lower variance across seeds.Efficiency analysis shows less than 3%overhead in wall-clock training time and no additional inference cost,making the approach practical for realworld deployment.This study demonstrates that a simple task-structured curriculum can significantly improve the effectiveness of knowledge distillation.By separating and sequencing objectives,TSCL achieves a better balance between accuracy,stability,and explanation quality.The framework generalizes across domains,including medical NLI,and offers a principled recipe for future applications in multimodal reasoning and reinforcement learning. 展开更多
关键词 Knowledge distillation curriculum learning language models multi-task learning step-by-step learning
在线阅读 下载PDF
Command-agent:Reconstructing warfare simulation and command decision-making using large language models
8
作者 Mengwei Zhang Minchi Kuang +3 位作者 Heng Shi Jihong Zhu Jingyu Zhu Xiao Jiang 《Defence Technology(防务技术)》 2026年第2期294-313,共20页
War rehearsals have become increasingly important in national security due to the growing complexity of international affairs.However,traditional rehearsal methods,such as military chess simulations,are inefficient an... War rehearsals have become increasingly important in national security due to the growing complexity of international affairs.However,traditional rehearsal methods,such as military chess simulations,are inefficient and inflexible,with particularly pronounced limitations in command and decision-making.The overwhelming volume of information and high decision complexity hinder the realization of autonomous and agile command and control.To address this challenge,an intelligent warfare simulation framework named Command-Agent is proposed,which deeply integrates large language models(LLMs)with digital twin battlefields.By constructing a highly realistic battlefield environment through real-time simulation and multi-source data fusion,the natural language interaction capabilities of LLMs are leveraged to lower the command threshold and to enable autonomous command through the Observe-Orient-Decide-Act(OODA)feedback loop.Within the Command-Agent framework,a multimodel collaborative architecture is further adopted to decouple the decision-generation and command-execution functions of LLMs.By combining specialized models such as Deep Seek-R1 and MCTool,the limitations of single-model capabilities are overcome.MCTool is a lightweight execution model fine-tuned for military Function Calling tasks.The framework also introduces a Vector Knowledge Base to mitigate hallucinations commonly exhibited by LLMs.Experimental results demonstrate that Command-Agent not only enables natural language-driven simulation and control but also deeply understands commander intent.Leveraging the multi-model collaborative architecture,during red-blue UAV confrontations involving 2 to 8 UAVs,the integrated score is improved by an average of 41.8%compared to the single-agent system(MCTool),accompanied by a 161.8%optimization in the battle loss ratio.Furthermore,when compared with multi-agent systems lacking the knowledge base,the inclusion of the Vector Knowledge Base further improves overall performance by 16.8%.In comparison with the general model(Qwen2.5-7B),the fine-tuned MCTool leads by 5%in execution efficiency.Therefore,the proposed Command-Agent introduces a novel perspective to the military command system and offers a feasible solution for intelligent battlefield decision-making. 展开更多
关键词 Digital twin battlefield Large language models Multi-agent system Military command
在线阅读 下载PDF
The Xu-Argument:An Innovative Approach to Second Language Acquisition—An Interview With Prof.Wang Chuming
9
作者 Min Wang 《Chinese Journal of Applied Linguistics》 2026年第1期8-20,159,共14页
This interview examines the theoretical foundations,pedagogical applications,developmental trajectory,and future directions of the xu-argument.Professor Wang Chuming offers a comprehensive account of the xu-argument,c... This interview examines the theoretical foundations,pedagogical applications,developmental trajectory,and future directions of the xu-argument.Professor Wang Chuming offers a comprehensive account of the xu-argument,clarifying its theoretical framework,the learning mechanisms underlying xu,and its interface with international theories of second language acquisition(SLA).From the perspective of the xu-argument,he proposes novel interpretations of core issues in SLA.Drawing on the development of the xu-argument,Wang further discusses the essence,directions,and methodology of innovation in SLA theory.He emphasizes that theoretical advances must capture and illuminate underlying natural laws,arguing that innovative approaches are typically rooted in deep reflection on common sense.He also calls for theoretical innovation in SLA in the Chinese context,advocating a robust research paradigm that shifts from local observation to global theoretical generalization,thereby promoting bottom-up theoretical development.In closing,he highlights the promising prospects for SLA theory in the era of artificial intelligence. 展开更多
关键词 Wang chuming the xu-argument second language acquisition theoretical innovation
在线阅读 下载PDF
Prompt Injection Attacks on Large Language Models:A Survey of Attack Methods,Root Causes,and Defense Strategies
10
作者 Tongcheng Geng Zhiyuan Xu +1 位作者 Yubin Qu W.Eric Wong 《Computers, Materials & Continua》 2026年第4期134-185,共52页
Large language models(LLMs)have revolutionized AI applications across diverse domains.However,their widespread deployment has introduced critical security vulnerabilities,particularly prompt injection attacks that man... Large language models(LLMs)have revolutionized AI applications across diverse domains.However,their widespread deployment has introduced critical security vulnerabilities,particularly prompt injection attacks that manipulate model behavior through malicious instructions.Following Kitchenham’s guidelines,this systematic review synthesizes 128 peer-reviewed studies from 2022 to 2025 to provide a unified understanding of this rapidly evolving threat landscape.Our findings reveal a swift progression from simple direct injections to sophisticated multimodal attacks,achieving over 90%success rates against unprotected systems.In response,defense mechanisms show varying effectiveness:input preprocessing achieves 60%–80%detection rates and advanced architectural defenses demonstrate up to 95%protection against known patterns,though significant gaps persist against novel attack vectors.We identified 37 distinct defense approaches across three categories,but standardized evaluation frameworks remain limited.Our analysis attributes these vulnerabilities to fundamental LLM architectural limitations,such as the inability to distinguish instructions from data and attention mechanism vulnerabilities.This highlights critical research directions such as formal verification methods,standardized evaluation protocols,and architectural innovations for inherently secure LLM designs. 展开更多
关键词 Prompt injection attacks large language models defense mechanisms security evaluation
在线阅读 下载PDF
LLMKB:Large Language Models with Knowledge Base Augmentation for Conversational Recommendation
11
作者 FANG Xiu QIU Sijia +1 位作者 SUN Guohao LU Jinhu 《Journal of Donghua University(English Edition)》 2026年第1期91-103,共13页
Conversational recommender systems(CRSs)focus on refining preferences and providing personalized recommendations through natural language interactions and dialogue history.Large language models(LLMs)have shown outstan... Conversational recommender systems(CRSs)focus on refining preferences and providing personalized recommendations through natural language interactions and dialogue history.Large language models(LLMs)have shown outstanding performance across various domains,thereby prompting researchers to investigate their applicability in recommendation systems.However,due to the lack of task-specific knowledge and an inefficient feature extraction process,LLMs still have suboptimal performance in recommendation tasks.Therefore,external knowledge sources,such as knowledge graphs(KGs)and knowledge bases(KBs),are often introduced to address the issue of data sparsity.Compared to KGs,KBs possess higher retrieval efficiency,making them more suitable for scenarios where LLMs serve as recommenders.To this end,we introduce a novel framework integrating LLMs with KBs for enhanced retrieval generation,namely LLMKB.LLMKB initially leverages structured knowledge to create mapping dictionaries,extracting entity-relation information from heterogeneous knowledge to construct KBs.Then,LLMKB achieves the embedding calibration between user information representations and documents in KBs through retrieval model fine-tuning.Finally,LLMKB employs retrievalaugmented generation to produce recommendations based on fused text inputs,followed by post-processing.Experiment results on two public CRS datasets demonstrate the effectiveness of our framework.Our code is publicly available at the link:https://anonymous.4open.science/r/LLMKB-6FD0. 展开更多
关键词 recommender system large language model(LLM) knowledge base(KB)
在线阅读 下载PDF
OPOR-Bench:Evaluating Large Language Models on Online Public Opinion Report Generation
12
作者 Jinzheng Yu Yang Xu +4 位作者 Haozhen Li Junqi Li Ligu Zhu Hao Shen Lei Shi 《Computers, Materials & Continua》 2026年第4期1403-1427,共25页
Online Public Opinion Reports consolidate news and social media for timely crisis management by governments and enterprises.While large language models(LLMs)enable automated report generation,this specific domain lack... Online Public Opinion Reports consolidate news and social media for timely crisis management by governments and enterprises.While large language models(LLMs)enable automated report generation,this specific domain lacks formal task definitions and corresponding benchmarks.To bridge this gap,we define the Automated Online Public Opinion Report Generation(OPOR-Gen)task and construct OPOR-Bench,an event-centric dataset with 463 crisis events across 108 countries(comprising 8.8 K news articles and 185 K tweets).To evaluate report quality,we propose OPOR-Eval,a novel agent-based framework that simulates human expert evaluation.Validation experiments show OPOR-Eval achieves a high Spearman’s correlation(ρ=0.70)with human judgments,though challenges in temporal reasoning persist.This work establishes an initial foundation for advancing automated public opinion reporting research. 展开更多
关键词 Online public opinion reports crisis management large language models agent-based evaluation
在线阅读 下载PDF
Foreign Language Learning and the Cultivation of National Consciousness in the Age of Intelligence-A Case Study Through the Appreciation of The Wild Robot
13
作者 ZHANG Xiaoling WANG Yongli 《Cultural and Religious Studies》 2026年第1期22-25,共4页
This study examines how foreign language education in the artificial intelligence(AI)era could assist the cultivation of national consciousness through a technology-enhanced pedagogy of film appreciation.Using The Wil... This study examines how foreign language education in the artificial intelligence(AI)era could assist the cultivation of national consciousness through a technology-enhanced pedagogy of film appreciation.Using The Wild Robot as a case study,we argue that cinematic narratives serve as cultural mirrors,offering immersive,reflective,and affective sites for intercultural learning.We propose a three-layered pedagogical framework-progressing from semiotic decoding,through narrative and value comparison,to creative identity construction-that integrates intelligent tools to develop both communicative competence and an agentive sense of belonging.The approach exemplifies a humanistic turn in language teaching,aiming to form“rooted global communicators”who can engage in cross-civilization dialogue with cultural confidence and critical awareness. 展开更多
关键词 foreign language learning cultivation of national consciousness The Wild Robot
在线阅读 下载PDF
Unlocking Edge Fine-Tuning:A Sample-Efficient Language-Empowered Split Fine-Tuning Framework
14
作者 Zuyi Huang Yue Wang +4 位作者 Jia Liu Haodong Yi Lejun Ai Min Chen Salman A.AlQahtani 《Computers, Materials & Continua》 2026年第4期1584-1606,共23页
The personalized fine-tuning of large languagemodels(LLMs)on edge devices is severely constrained by limited computation resources.Although split federated learning alleviates on-device burdens,its effectiveness dimin... The personalized fine-tuning of large languagemodels(LLMs)on edge devices is severely constrained by limited computation resources.Although split federated learning alleviates on-device burdens,its effectiveness diminishes in few-shot reasoning scenarios due to the low data efficiency of conventional supervised fine-tuning,which leads to excessive communication overhead.To address this,we propose Language-Empowered Split Fine-Tuning(LESFT),a framework that integrates split architectures with a contrastive-inspired fine-tuning paradigm.LESFT simultaneously learns frommultiple logically equivalent but linguistically diverse reasoning chains,providing richer supervisory signals and improving data efficiency.This process-oriented training allows more effective reasoning adaptation with fewer samples.Extensive experiments demonstrate that LESFT consistently outperforms strong baselines such as SplitLoRA in task accuracy.LESFT consistently outperforms strong baselines on GSM8K,CommonsenseQA,and AQUA_RAT,with the largest gains observed on Qwen2.5-3B.These results indicate that LESFT can effectively adapt large language models for reasoning tasks under the computational and communication constraints of edge environments. 展开更多
关键词 Large language models edge computing efficient fine-tuning few-shot fine-tuning split federated learning
在线阅读 下载PDF
Exploring Recovery through Life Narratives in Psychiatric Home-Visit Nursing:A Natural Language Processing Approach Using BERTopic
15
作者 Ichiro Kutsuna Masanao Ikeya +2 位作者 Akane Fujii Aiko Hoshino Kazuya Sakai 《International Journal of Mental Health Promotion》 2026年第2期31-47,共17页
Background:In mental health,recovery is emphasized,and qualitative analyses of service users’narratives have accumulated;however,while qualitative approaches excel at capturing rich context and generating new concept... Background:In mental health,recovery is emphasized,and qualitative analyses of service users’narratives have accumulated;however,while qualitative approaches excel at capturing rich context and generating new concepts,they are limited in generalizability and feasible data volume.This study aimed to quantify the subjective life history narratives of users of psychiatric home-visit nursing using natural language processing(NLP)and to clarify the relationships between linguistic features and recovery-related indicators.Methods:We conducted audio-recorded and transcribed semi-structured interviews on daily life verbatim and collected self-report questionnaires(Recovery Assessment Scale[RAS])and clinician ratings(Global Assessment of Functioning[GAF])from Japanese users of psychiatric home-visit nursing.Using the artificial intelligence-based topic-modeling method BERTopic,we extracted topics from the interview texts and calculated each participant’s topic proportions,and then examined associations between topic proportions and recovery-related indicators using Pearson correlation analyses.Results:“School”showed a significant positive correlation with RAS(r=0.39,p=0.05),whereas“Family”showed a significant negative correlation(r=–0.46,p=0.02).GAF was positively correlated with word count(r=0.44,p=0.02)and“Hospital”(r=0.42,p=0.03),and negatively correlated with“Backchannels”(aizuchi)(r=–0.41,p=0.03).Conclusion:The present results suggest that the quantity,quality,and content of narratives can serve as useful indicators of mental health and recovery,and that objective NLP-based analysis of service users’narratives can complement traditional self-report scales and clinician ratings to inform the design of recovery-oriented care in psychiatric home-visit nursing. 展开更多
关键词 Personal recovery life history narratives natural language processing psychiatric home-visit nursing artificial intelligence
暂未订购
亚洲罕见伴VCP基因变异的肌病1例并文献复习
16
作者 朱琳 赵珺 +2 位作者 王爽 邱峰 陈娟 《解放军医学杂志》 北大核心 2026年第1期65-69,共5页
目的总结含缬酪肽蛋白(VCP)基因c.464G>A变异的临床特征并进行文献复习,以提高对该突变相关疾病的认识。方法回顾性分析1例随访7年仅表现为肌病的VCP基因c.464G>A变异病例的临床资料;检索PubMed、万方数据和中国知网数据库截至202... 目的总结含缬酪肽蛋白(VCP)基因c.464G>A变异的临床特征并进行文献复习,以提高对该突变相关疾病的认识。方法回顾性分析1例随访7年仅表现为肌病的VCP基因c.464G>A变异病例的临床资料;检索PubMed、万方数据和中国知网数据库截至2025年4月发表的病例文献,总结同类型基因突变病例的临床特征。结果患者女,43岁,病程7年,无家族史和发育畸形,隐袭缓慢进展四肢近端肌无力,由腰部至双下肢、双上肢,后期出现肉跳,肌酸肌酶正常。MRI见双侧臀部及大腿肌肉脂肪浸润,伴肌肉萎缩。组织病理学检查显示骨骼肌出现肌纤维内镶边空泡、肌纤维坏死、再生和肥大,部分肌纤维可见胞浆体,核内包涵体形成,符合空泡肌病样病理改变。头颅MRI、认知功能筛查和骨密度未见异常。依据基因检测结果确诊为VCP基因c.464G>A(p.Arg155His)突变所致包涵体肌病。通过文献检索,共收集8例携带VCP 155位点突变(R155H)的患者,早期多表现为肌肉疾病或额颞叶痴呆。结论VCP基因c.464G>A变异临床表现具有高度异质性。VCP基因表型对相关包涵体肌病诊断起关键作用;临床缺少特效治疗,需长期随访观察。 展开更多
关键词 肌肉病 R155H突变 含缬酪肽蛋白 c.464G>A 包涵体肌病
暂未订购
片上操作系统应用C语言子集及扩展设计
17
作者 曹国顺 耿力 +3 位作者 付睿 高健 何碧波 袁桃鸿 《信息技术与标准化》 2026年第1期116-119,共4页
针对片上操作系统(COS)采用标准C语言存在的异常处理能力不足、数据存储管理与生命周期表达等问题,提出了一种C语言子集及语言扩展设计。该设计对标准C语言进行裁剪以适应受限的COS环境;设计异常处理关键字,实现在运行期间对异常的结构... 针对片上操作系统(COS)采用标准C语言存在的异常处理能力不足、数据存储管理与生命周期表达等问题,提出了一种C语言子集及语言扩展设计。该设计对标准C语言进行裁剪以适应受限的COS环境;设计异常处理关键字,实现在运行期间对异常的结构化捕获与受控处理;设计应用数据类型关键字,实现对应用数据的存储管理、作用域范围及生命周期的语言级表达。工程实践表明,该语言扩展设计提高了C语言应用开发效率和应用运行安全性。 展开更多
关键词 片上操作系统 c 语言子集 语言扩展 异常处理 应用数据类型 关键词
在线阅读 下载PDF
C反应蛋白/白蛋白比值与D-二聚体联合检测对妊娠合并急性胰腺炎病情严重程度的预测价值
18
作者 刘英男 严丽梅 《中国医科大学学报》 北大核心 2026年第1期75-79,共5页
目的探讨血清C反应蛋白/白蛋白比值(CAR)、D-二聚体及二者联合检测预测妊娠合并急性胰腺炎(APIP)病情严重程度的价值。方法收集2021年1月至2023年12月间我院APIP患者及同期分娩的正常妊娠者(健康对照组)的临床资料。根据APIP病情严重程... 目的探讨血清C反应蛋白/白蛋白比值(CAR)、D-二聚体及二者联合检测预测妊娠合并急性胰腺炎(APIP)病情严重程度的价值。方法收集2021年1月至2023年12月间我院APIP患者及同期分娩的正常妊娠者(健康对照组)的临床资料。根据APIP病情严重程度将患者分为非重症组、重症组;根据病因分为胆源型组、高脂血症型组、特发型组。用logistic回归分析影响APIP病情严重程度的危险因素及其诊断效能。结果CAR、D-二聚体是影响APIP病情严重程度的独立危险因素。CAR和D-二聚体联合检测受试者操作特征曲线下面积最大,且灵敏度、特异度等诊断效能均优于单独检测。结论CAR及D-二聚体是APIP病情严重程度的独立危险因素,且二者联合检测可作为APIP病情严重程度的良好预测指标。 展开更多
关键词 妊娠合并急性胰腺炎 c反应蛋白/白蛋白比值 D-二聚体 联合预测
暂未订购
血清胱抑素C及C反应蛋白-甘油三酯-空腹血糖指数与2型糖尿病患者视网膜病变的相关性分析
19
作者 付朝泓 杨姣 +1 位作者 齐娟 刘航 《微循环学杂志》 2026年第1期66-70,共5页
目的:探讨血清胱抑素C(CysC)及C反应蛋白(CRP)-甘油三酯(TG)-空腹血糖(FPG)指数(CTI)与2型糖尿病视网膜病变(DR)的相关性。方法:纳入2024年12月至2025年9月本院确诊的2型糖尿病(T2DM)患者233例,根据有无DR分为T2DM-DR组(n=76)和T2DM-no... 目的:探讨血清胱抑素C(CysC)及C反应蛋白(CRP)-甘油三酯(TG)-空腹血糖(FPG)指数(CTI)与2型糖尿病视网膜病变(DR)的相关性。方法:纳入2024年12月至2025年9月本院确诊的2型糖尿病(T2DM)患者233例,根据有无DR分为T2DM-DR组(n=76)和T2DM-nonDR组(n=157)。检测两组患者血清CysC、CRP、TG、FPG等指标,计算甘油三酯-空腹血糖指数(TyG)及CTI。比较两组差异,并采用单因素和多因素Logistic回归分析T2DM并发DR的危险因素。结果:与T2DM-nonDR组相比,T2DM-DR组血脂代谢紊乱、血糖升高及肾功能异常更明显,血清CysC、TyG、CRP及CTI水平均显著升高(P<0.05)。多因素Logistic回归分析显示,CysC(OR=10.509,95%CI:2.034-54.288,P=0.005)和CTI(OR=3.925,95%CI:2.192-7.026,P<0.001)是T2DM患者发生DR的独立危险因素。结论:血清CysC及CTI水平是T2DM患者DR发生的独立危险因素。 展开更多
关键词 2型糖尿病 视网膜病变 胱抑素c c反应蛋白-TyG指数
暂未订购
上一页 1 2 250 下一页 到第
使用帮助 返回顶部