期刊文献+
共找到15,261篇文章
< 1 2 250 >
每页显示 20 50 100
Construction and Application Practice of the Data-driven Comprehensive Management Platform for Regional Air Quality
1
作者 Tongxing ZHANG Yun WU Yongwen LI 《Meteorological and Environmental Research》 2026年第1期21-28,共8页
To address the severe challenges of PM_(2.5) and ozone co-control during the"14^(th) Five-Year Plan"period and to enhance the precision and intelligence level of air environment governance,it is imperative t... To address the severe challenges of PM_(2.5) and ozone co-control during the"14^(th) Five-Year Plan"period and to enhance the precision and intelligence level of air environment governance,it is imperative to build an efficient comprehensive management platform for regional air quality.In this paper,the specific practice in Zibo City,Shandong Province is as an example to systematically analyze the top-level design,technical implementation,and innovative application of a comprehensive management platform for regional air quality integrating"perception monitoring,data fusion,research judgment of early warnings,analysis of sources,collaborative dispatching,and evaluation assessment".Through the construction of an"sky-air-ground"integrated three-dimensional monitoring network,the platform integrates multi-source heterogeneous environmental data,and employs big data,cloud computing,artificial intelligence,CALPUFF/CMAQ,and other numerical model technologies to achieve comprehensive perception,precise prediction,intelligent source tracing,and closed-loop management of air pollution.The platform innovatively establishes a full-process closed-loop management mechanism of"data-early warning-disposition-evaluation",and achieves a fundamental transformation from passive response to active anticipation and from experience-based judgment to data driving in environmental supervision.The application results show that this platform significantly improves the scientific decision-making ability and collaborative execution efficiency of air pollution governance in Zibo City,providing a replicable and scalable comprehensive solution for similar industrial cities to achieve the continuous improvement of air quality. 展开更多
关键词 Comprehensive management of air quality Big data Internet of Things Closed-loop management data driving Off-site supervision
在线阅读 下载PDF
Toward Secure and Auditable Data Sharing:A Cross-Chain CP-ABE Framework
2
作者 Ye Tian Zhuokun Fan Yifeng Zhang 《Computers, Materials & Continua》 2026年第4期1509-1529,共21页
Amid the increasing demand for data sharing,the need for flexible,secure,and auditable access control mechanisms has garnered significant attention in the academic community.However,blockchain-based ciphertextpolicy a... Amid the increasing demand for data sharing,the need for flexible,secure,and auditable access control mechanisms has garnered significant attention in the academic community.However,blockchain-based ciphertextpolicy attribute-based encryption(CP-ABE)schemes still face cumbersome ciphertext re-encryption and insufficient oversight when handling dynamic attribute changes and cross-chain collaboration.To address these issues,we propose a dynamic permission attribute-encryption scheme for multi-chain collaboration.This scheme incorporates a multiauthority architecture for distributed attribute management and integrates an attribute revocation and granting mechanism that eliminates the need for ciphertext re-encryption,effectively reducing both computational and communication overhead.It leverages the InterPlanetary File System(IPFS)for off-chain data storage and constructs a cross-chain regulatory framework—comprising a Hyperledger Fabric business chain and a FISCO BCOS regulatory chain—to record changes in decryption privileges and access behaviors in an auditable manner.Security analysis shows selective indistinguishability under chosen-plaintext attack(sIND-CPA)security under the decisional q-Parallel Bilinear Diffie-Hellman Exponent Assumption(q-PBDHE).In the performance and experimental evaluations,we compared the proposed scheme with several advanced schemes.The results show that,while preserving security,the proposed scheme achieves higher encryption/decryption efficiency and lower storage overhead for ciphertexts and keys. 展开更多
关键词 data sharing blockchain attribute-based encryption dynamic permissions
在线阅读 下载PDF
Design,Realization,and Evaluation of Faster End-to-End Data Transmission over Voice Channels
3
作者 Jian Huang Ming weiLi +2 位作者 Yulong Tian Yi Yao Hao Han 《Computers, Materials & Continua》 2026年第4期1650-1675,共26页
With the popularization of new technologies,telephone fraud has become the main means of stealing money and personal identity information.Taking inspiration from the website authentication mechanism,we propose an end-... With the popularization of new technologies,telephone fraud has become the main means of stealing money and personal identity information.Taking inspiration from the website authentication mechanism,we propose an end-to-end datamodem scheme that transmits the caller’s digital certificates through a voice channel for the recipient to verify the caller’s identity.Encoding useful information through voice channels is very difficult without the assistance of telecommunications providers.For example,speech activity detection may quickly classify encoded signals as nonspeech signals and reject input waveforms.To address this issue,we propose a novel modulation method based on linear frequency modulation that encodes 3 bits per symbol by varying its frequency,shape,and phase,alongside a lightweightMobileNetV3-Small-based demodulator for efficient and accurate signal decoding on resource-constrained devices.This method leverages the unique characteristics of linear frequency modulation signals,making them more easily transmitted and decoded in speech channels.To ensure reliable data delivery over unstable voice links,we further introduce a robust framing scheme with delimiter-based synchronization,a sample-level position remedying algorithm,and a feedback-driven retransmission mechanism.We have validated the feasibility and performance of our system through expanded real-world evaluations,demonstrating that it outperforms existing advanced methods in terms of robustness and data transfer rate.This technology establishes the foundational infrastructure for reliable certificate delivery over voice channels,which is crucial for achieving strong caller authentication and preventing telephone fraud at its root cause. 展开更多
关键词 Deep learning modulation CHIRP data over voice
在线阅读 下载PDF
Towards energy-efficient data centers:A comprehensive review of passive and active cooling strategies
4
作者 Senhong Cai Zhonghua Gou 《Energy and Built Environment》 2026年第1期206-226,共21页
With the rapid growth of cloud computing,the number of data centers(DCs)continuously increases,leading to a high-energy consumption dilemma.Cooling,apart from IT equipment,represents the largest energy consumption in ... With the rapid growth of cloud computing,the number of data centers(DCs)continuously increases,leading to a high-energy consumption dilemma.Cooling,apart from IT equipment,represents the largest energy consumption in DCs.Passive design(PD)and active design(AD)are two important approaches in architectural design to reduce energy consumption.However,for DC cooling,few studies have summarized AD,and there are almost no studies on PD.Based on existing international research(2005-2024),this paper summarizes the current state of cooling strategies for DCs.PD encompasses floors,ceilings,and layout and zoning of racks.Additionally,other passive strategies not yet studied in DCs are critically examined.AD includes air,liquid,free,and two-phase cooling.This paper systematically compares the performance of different AD technologies on various KPIs,including energy,economic,and environmental indicators.This paper also explores the application of different cooling design strategies through best-practice examples and presents advanced algorithms for energy management in operational DCs.This study reveals that free cooling is widely employed,with Artificial Neural Networks emerging as the most popular algorithm for managing cooling energy.Finally,this paper suggests four future directions for reducing cooling energy in DCs,with a focus on the development of passive strategies.This paper provides an overview and guide to DC energy-consumption issues,emphasizes the importance of implementing passive and active design strategies to reduce DC cooling energy consumption,and provides directions and references for future energy-efficient DC designs. 展开更多
关键词 data center Passive design Active design Cooling load Energy efficiency
在线阅读 下载PDF
A Composite Loss-Based Autoencoder for Accurate and Scalable Missing Data Imputation
5
作者 Thierry Mugenzi Cahit Perkgoz 《Computers, Materials & Continua》 2026年第1期1985-2005,共21页
Missing data presents a crucial challenge in data analysis,especially in high-dimensional datasets,where missing data often leads to biased conclusions and degraded model performance.In this study,we present a novel a... Missing data presents a crucial challenge in data analysis,especially in high-dimensional datasets,where missing data often leads to biased conclusions and degraded model performance.In this study,we present a novel autoencoder-based imputation framework that integrates a composite loss function to enhance robustness and precision.The proposed loss combines(i)a guided,masked mean squared error focusing on missing entries;(ii)a noise-aware regularization term to improve resilience against data corruption;and(iii)a variance penalty to encourage expressive yet stable reconstructions.We evaluate the proposed model across four missingness mechanisms,such as Missing Completely at Random,Missing at Random,Missing Not at Random,and Missing Not at Random with quantile censorship,under systematically varied feature counts,sample sizes,and missingness ratios ranging from 5%to 60%.Four publicly available real-world datasets(Stroke Prediction,Pima Indians Diabetes,Cardiovascular Disease,and Framingham Heart Study)were used,and the obtained results show that our proposed model consistently outperforms baseline methods,including traditional and deep learning-based techniques.An ablation study reveals the additive value of each component in the loss function.Additionally,we assessed the downstream utility of imputed data through classification tasks,where datasets imputed by the proposed method yielded the highest receiver operating characteristic area under the curve scores across all scenarios.The model demonstrates strong scalability and robustness,improving performance with larger datasets and higher feature counts.These results underscore the capacity of the proposed method to produce not only numerically accurate but also semantically useful imputations,making it a promising solution for robust data recovery in clinical applications. 展开更多
关键词 Missing data imputation autoencoder deep learning missing mechanisms
在线阅读 下载PDF
Distributed Quasi-Newton Algorithm for Non-Randomly Stored Data
6
作者 LIU Xirui WU Mixia LIU Bangshu 《Journal of Systems Science & Complexity》 2026年第1期456-480,共25页
Distributed learning is a well-established method for estimation tasks over extensively distributed datasets.However,non-randomly stored data can introduce bias into local parameter estimates,leading to significant pe... Distributed learning is a well-established method for estimation tasks over extensively distributed datasets.However,non-randomly stored data can introduce bias into local parameter estimates,leading to significant performance degradation in classical distributed algorithms.In this paper,the authors propose a novel Distributed Quasi-Newton Pilot(DQNP)method for distributed learning with non-randomly distributed data.The proposed approach accommodates both randomly and non-randomly distributed data settings and imposes no constraints on the uniformity of local sample sizes.Additionally,it avoids the need to transfer the Hessian matrix or compute its inversion,thereby greatly reducing computational and communication complexity.The authors theoretically demonstrate that the resulting estimator achieves statistical efficiency under mild conditions.Extensive numerical experiments on synthetic and real-world data validate the theoretical findings and illustrate the effectiveness of the proposed method. 展开更多
关键词 Communication-efficient computation efficiency distributed inference non-randomly distributed data quasi-Newton algorithm
原文传递
Current Situation of Application and Development Prospects of the Statistical Analysis of Big Data
7
作者 Zhuoran LI 《Meteorological and Environmental Research》 2026年第1期45-47,共3页
With the advent of the big data era,modern statistics has enjoyed unprecedented development opportunities and also faced numerous new challenges.Traditional statistical computing methods are often limited by issues su... With the advent of the big data era,modern statistics has enjoyed unprecedented development opportunities and also faced numerous new challenges.Traditional statistical computing methods are often limited by issues such as computer memory capacity and distributed storage of data across different locations,and are unable to directly apply to large-scale data sets.Therefore,in the context of big data,designing efficient and theoretically guaranteed statistical learning and inference algorithms has become a key issue that the current field of statistics urgently needs to address.In this paper,the application status of statistical analysis methods in the big data environment was systematically reviewed,and its future development directions were analyzed to provide reference and support for the further development of theory and methods of the statistical analysis of big data. 展开更多
关键词 Big data Statistical analysis Current status Development prospects
在线阅读 下载PDF
Photoacoustic-computed tomography 3D data compression method and system based on Wavelet-Transformer
8
作者 Jialin Li Tingting Li +2 位作者 Yiming Ma Yi Shen Mingjian Sun 《Journal of Innovative Optical Health Sciences》 2026年第1期110-125,共16页
Photoacoustic-computed tomography is a novel imaging technique that combines high absorption contrast and deep tissue penetration capability,enabling comprehensive three-dimensional imaging of biological targets.Howev... Photoacoustic-computed tomography is a novel imaging technique that combines high absorption contrast and deep tissue penetration capability,enabling comprehensive three-dimensional imaging of biological targets.However,the increasing demand for higher resolution and real-time imaging results in significant data volume,limiting data storage,transmission and processing efficiency of system.Therefore,there is an urgent need for an effective method to compress the raw data without compromising image quality.This paper presents a photoacoustic-computed tomography 3D data compression method and system based on Wavelet-Transformer.This method is based on the cooperative compression framework that integrates wavelet hard coding with deep learning-based soft decoding.It combines the multiscale analysis capability of wavelet transforms with the global feature modeling advantage of Transformers,achieving high-quality data compression and reconstruction.Experimental results using k-wave simulation suggest that the proposed compression system has advantages under extreme compression conditions,achieving a raw data compression ratio of up to 1:40.Furthermore,three-dimensional data compression experiment using in vivo mouse demonstrated that the maximum peak signal-to-noise ratio(PSNR)and structural similarity index(SSIM)values of reconstructed images reached 38.60 and 0.9583,effectively overcoming detail loss and artifacts introduced by raw data compression.All the results suggest that the proposed system can significantly reduce storage requirements and hardware cost,enhancing computational efficiency and image quality.These advantages support the development of photoacoustic-computed tomography toward higher efficiency,real-time performance and intelligent functionality. 展开更多
关键词 Photoacoustic-computed tomography data compression TRANSFORMER
原文传递
Model Construction for Complex and Heterogeneous Data of Urban Road Traffic Congestion
9
作者 Jianchun Wen Minghao Zhu +2 位作者 Bo Gao Zhaojian Liu Xuehan Li 《Computers, Materials & Continua》 2026年第2期1354-1370,共17页
Urban traffic generates massive and diverse data,yet most systems remain fragmented.Current approaches to congestion management suffer from weak data consistency and poor scalability.This study addresses this gap by p... Urban traffic generates massive and diverse data,yet most systems remain fragmented.Current approaches to congestion management suffer from weak data consistency and poor scalability.This study addresses this gap by proposing the Urban Traffic Congestion Unified Metadata Model(UTC-UMM).The goal is to provide a standardized and extensible framework for describing,extracting,and storing multisource traffic data in smart cities.The model defines a two-tier specification that organizes nine core traffic resource classes.It employs an eXtensible Markup Language(XML)Schema that connects general elements with resource-specific elements.This design ensures both syntactic and semantic interoperability across siloed datasets.Extension principles allow new elements or constraints to be introducedwithout breaking backward compatibility.Adistributed pipeline is implemented usingHadoop Distributed File System(HDFS)and HBase.It integrates computer vision for video and natural language processing for text to automate metadata extraction.Optimized row-key designs enable low-latency queries.Performance is tested with the Yahoo!Cloud Serving Benchmark(YCSB),which shows linear scalability and high throughput.The results demonstrate that UTC-UMM can unify heterogeneous traffic data while supporting real-time analytics.The discussion highlights its potential to improve data reuse,portability,and scalability in urban congestion studies.Future research will explore integration with association rulemining and advanced knowledge representation to capture richer spatiotemporal traffic patterns. 展开更多
关键词 METAdata urban road traffic heterogeneous data HBASE semantic description framework
在线阅读 下载PDF
Key Technologies for AI-Driven Network Traffic Classification Workflow and Data Distribution Shift
10
作者 Zhao Jianchao Geng Zhaosen +1 位作者 Li Zeyi Wang Pan 《ZTE Communications》 2026年第1期34-44,共11页
With the evolution of next-generation network technologies,the complexity of network management has significantly increased,and the means of network attacks are diversified,bringing new challenges to network traffic c... With the evolution of next-generation network technologies,the complexity of network management has significantly increased,and the means of network attacks are diversified,bringing new challenges to network traffic classification.This paper presents a general AIdriven network traffic classification workflow and elaborates on a traffic data and feature engineering framework.Most importantly,it analyzes the concept and causes of data distribution shifts in ne twork traffic,proposing detection methods and countermeasures.Experimental results on real traffic collected at different time intervals show that application evolution can induce data distribution shifts,which in turn lead to a noticeable degradation in traffic classification performance.Comparative drift detection experiments further confirm that such shifts are more evident over long-term intervals,while short-term traffic remains relatively stable.These findings demonstrate the necessity of incorporating drift-aware mechanisms into AI-driven network traffic classification systems. 展开更多
关键词 traffic classification traffic identification deep learning data distribution shift concept shifting
在线阅读 下载PDF
AI and Big Data in Oncology:A Physician-Centered Perspective on Emerging Clinical and Research Applications
11
作者 Binliang Liu Qingyao Shang +5 位作者 Jun Li Shuna Yao Meishuo Ouyang Yu Wang Sheng Luo Quchang Ouyang 《Cancer Innovation》 2026年第1期33-46,共14页
The convergence of artificial intelligence(AI)and big data is reshaping contemporary oncology by enabling the integration of multimodal information across imaging,pathology,genomics,and clinical records.From a physici... The convergence of artificial intelligence(AI)and big data is reshaping contemporary oncology by enabling the integration of multimodal information across imaging,pathology,genomics,and clinical records.From a physician-centered perspective,these technologies can potentially be used to improve diagnostic precision,support individualized treatment planning,enhance longitudinal patient management,and accelerate both clinical and translational research.In this review,we synthesize the core AI methodologies most relevant to oncology-machine learning,deep learning,and large language models-and examine how they interact with established and emerging oncology data platforms.We further highlight practical use cases in clinical workflows and research pipelines,emphasizing opportunities for advancing precision cancer care while also addressing challenges associated with data heterogeneity,model generalizability,privacy protection,and real-world implementation.By underscoring the synergistic value of AI and big data,this review aims to inform the development of clinically meaningful,context-adapted strategies that promote translational innovation in both global and locally resourced healthcare environments. 展开更多
关键词 artificial intelligence big data cancer challenges and solutions clinical applications research design
在线阅读 下载PDF
Resilient Class-Incremental Learning:On the Interplay of Drifting,Unlabeled and Imbalanced Data Streams
12
作者 Jin Li Kleanthis Malialis Marios M.Polycarpou 《Artificial Intelligence Science and Engineering》 2026年第1期49-65,共17页
In today's connected world,the generation of massive streaming data across diverse domains has become commonplace.In the presence of concept drift,class imbalance,label scarcity,and new class emergence,these chall... In today's connected world,the generation of massive streaming data across diverse domains has become commonplace.In the presence of concept drift,class imbalance,label scarcity,and new class emergence,these challenges jointly degrade representation stability,bias learning toward outdated distributions,and reduce the resilience and reliability of detection in dynamic environments.This paper proposes a streaming classincremental learning(SCIL)framework to address these issues.The SCIL framework integrates an autoencoder(AE)with a multi-layer perceptron for multi-class prediction,employs a dual-loss strategy(classification and reconstruction)for prediction and new class detection,uses corrected pseudo-labels for online training,manages classes with queues,and applies oversampling to handle imbalance.The rationale behind the method's structure is elucidated through ablation studies,and a comprehensive experimental evaluation is performed using both real-world and synthetic datasets that feature class imbalance,incremental classes,and concept drifts.Our results demonstrate that SCIL outperforms strong baselines and state-of-the-art methods.In line with our commitment to Open Science,we make our code and datasets available to the community. 展开更多
关键词 concept drift data stream mining class-incremental learning class imbalance
在线阅读 下载PDF
Graph-Based Unified Settlement Framework for Complex Electricity Markets:Data Integration and Automated Refund Clearing
13
作者 Xiaozhe Guo Suyan Long +4 位作者 Ziyu Yue Yifan Wang Guanting Yin Yuyang Wang Zhaoyuan Wu 《Energy Engineering》 2026年第1期56-90,共35页
The increasing complexity of China’s electricity market creates substantial challenges for settlement automation,data consistency,and operational scalability.Existing provincial settlement systems are fragmented,lack... The increasing complexity of China’s electricity market creates substantial challenges for settlement automation,data consistency,and operational scalability.Existing provincial settlement systems are fragmented,lack a unified data structure,and depend heavily on manual intervention to process high-frequency and retroactive transactions.To address these limitations,a graph-based unified settlement framework is proposed to enhance automation,flexibility,and adaptability in electricity market settlements.A flexible attribute-graph model is employed to represent heterogeneousmulti-market data,enabling standardized integration,rapid querying,and seamless adaptation to evolving business requirements.An extensible operator library is designed to support configurable settlement rules,and a suite of modular tools—including dataset generation,formula configuration,billing templates,and task scheduling—facilitates end-to-end automated settlement processing.A robust refund-clearing mechanism is further incorporated,utilizing sandbox execution,data-version snapshots,dynamic lineage tracing,and real-time changecapture technologies to enable rapid and accurate recalculations under dynamic policy and data revisions.Case studies based on real-world data from regional Chinese markets validate the effectiveness of the proposed approach,demonstrating marked improvements in computational efficiency,system robustness,and automation.Moreover,enhanced settlement accuracy and high temporal granularity improve price-signal fidelity,promote cost-reflective tariffs,and incentivize energy-efficient and demand-responsive behavior among market participants.The method not only supports equitable and transparent market operations but also provides a generalizable,scalable foundation for modern electricity settlement platforms in increasingly complex and dynamic market environments. 展开更多
关键词 Electricity market market settlement data model graph database market refund clearing
在线阅读 下载PDF
Multi-modal data analysis for autism spectrum disorder in children:State of the art and trends
14
作者 Lukai Pang Xiaoke Zhao +4 位作者 Lulu Zhao Jianqing Li Fengyi Kuo Hongxing Wang Chengyu Liu 《EngMedicine》 2026年第1期47-56,共10页
Autism spectrum disorder(AsD)is a highly heterogeneous neurodevelopmental disorder.Early diagnosis and intervention are crucial for improving outcomes.Traditional single-modality diagnostic methods are subjective,limi... Autism spectrum disorder(AsD)is a highly heterogeneous neurodevelopmental disorder.Early diagnosis and intervention are crucial for improving outcomes.Traditional single-modality diagnostic methods are subjective,limited,and struggle to reveal the underlying pathological mechanisms.In contrast,multimodal data analysis integrates behavioral,physiological,and neuroimaging information with advanced machine-learning and deeplearning algorithms to overcome these limitations.In this review,we surveyed the recent pediatric AsD literature,highlighting artificial intelligence-driven diagnostic techniques,multimodal data fusion strategies,and emerging trends in ASD assessment.We surveyed studies that integrated two or more modalities and summarized the fusion levels,learning paradigms,tasks,datasets,and metrics.Multimodal approaches outperform singlemodality baselines in classification,severity estimation,and subtyping by leveraging complementary information and reducing modality-specific biases.Multimodal approaches significantly enhance diagnostic accuracy and comprehensiveness,enabling early screening of AsD,symptom subtyping,severity assessment,and personalized interventions.Advances in multimodal fusion techniques have promoted progress in precision medicine for the treatment of ASD. 展开更多
关键词 Autism spectrum disorder Multi-modal data Machine learning Early screening Symptom subtyping
暂未订购
Impact of Data Processing Techniques on AI Models for Attack-Based Imbalanced and Encrypted Traffic within IoT Environments
15
作者 Yeasul Kim Chaeeun Won Hwankuk Kim 《Computers, Materials & Continua》 2026年第1期247-274,共28页
With the increasing emphasis on personal information protection,encryption through security protocols has emerged as a critical requirement in data transmission and reception processes.Nevertheless,IoT ecosystems comp... With the increasing emphasis on personal information protection,encryption through security protocols has emerged as a critical requirement in data transmission and reception processes.Nevertheless,IoT ecosystems comprise heterogeneous networks where outdated systems coexist with the latest devices,spanning a range of devices from non-encrypted ones to fully encrypted ones.Given the limited visibility into payloads in this context,this study investigates AI-based attack detection methods that leverage encrypted traffic metadata,eliminating the need for decryption and minimizing system performance degradation—especially in light of these heterogeneous devices.Using the UNSW-NB15 and CICIoT-2023 dataset,encrypted and unencrypted traffic were categorized according to security protocol,and AI-based intrusion detection experiments were conducted for each traffic type based on metadata.To mitigate the problem of class imbalance,eight different data sampling techniques were applied.The effectiveness of these sampling techniques was then comparatively analyzed using two ensemble models and three Deep Learning(DL)models from various perspectives.The experimental results confirmed that metadata-based attack detection is feasible using only encrypted traffic.In the UNSW-NB15 dataset,the f1-score of encrypted traffic was approximately 0.98,which is 4.3%higher than that of unencrypted traffic(approximately 0.94).In addition,analysis of the encrypted traffic in the CICIoT-2023 dataset using the same method showed a significantly lower f1-score of roughly 0.43,indicating that the quality of the dataset and the preprocessing approach have a substantial impact on detection performance.Furthermore,when data sampling techniques were applied to encrypted traffic,the recall in the UNSWNB15(Encrypted)dataset improved by up to 23.0%,and in the CICIoT-2023(Encrypted)dataset by 20.26%,showing a similar level of improvement.Notably,in CICIoT-2023,f1-score and Receiver Operation Characteristic-Area Under the Curve(ROC-AUC)increased by 59.0%and 55.94%,respectively.These results suggest that data sampling can have a positive effect even in encrypted environments.However,the extent of the improvement may vary depending on data quality,model architecture,and sampling strategy. 展开更多
关键词 Encrypted traffic attack detection data sampling technique AI-based detection IoT environment
在线阅读 下载PDF
Korean Sign Language Recognition and Sentence Generation through Data Augmentation
16
作者 Soo-Yeon Jeong Ho-Yeon Jeong Sun-Young Ihm 《Computers, Materials & Continua》 2026年第5期2005-2019,共15页
Sign language is a primary mode of communication for individuals with hearing impairments,conveying meaning through hand shapes and hand movements.Contrary to spoken or written languages,sign language relies on the re... Sign language is a primary mode of communication for individuals with hearing impairments,conveying meaning through hand shapes and hand movements.Contrary to spoken or written languages,sign language relies on the recognition and interpretation of hand gestures captured in video data.However,sign language datasets remain relatively limited compared to those of other languages,which hinders the training and performance of deep learning models.Additionally,the distinct word order of sign language,unlike that of spoken language,requires context-aware and natural sentence generation.To address these challenges,this study applies data augmentation techniques to build a Korean Sign Language dataset and train recognition models.Recognized words are then reconstructed into complete sentences.The sign recognition process uses OpenCV and MediaPipe to extract hand landmarks from sign language videos and analyzes hand position,orientation,and motion.The extracted features are converted into time-series data and fed into a Long Short-Term Memory(LSTM)model.The proposed recognition framework achieved an accuracy of up to 81.25%,while the sentence generation achieved an accuracy of up to 95%.The proposed approach is expected to be applicable not only to Korean Sign Language but also to other low-resource sign languages for recognition and translation tasks. 展开更多
关键词 Korean sign language recognition LSTM data augmentation sentence completion
在线阅读 下载PDF
Computing the Planet:Integrating Machine Learning,Remote Sensing,and Sensor Data Fusion for Environmental Insights
17
作者 Kai Mao 《Journal of Environmental & Earth Sciences》 2026年第1期277-297,共21页
Indeed,a range of systems in the environment requires timely,spatially explicit,and credible information to support its environmental decision-making,but no one observing system can give the complete and reliable meas... Indeed,a range of systems in the environment requires timely,spatially explicit,and credible information to support its environmental decision-making,but no one observing system can give the complete and reliable measures of the Earth system across scales.This review summarizes how the realization of the Compute the Planet is underway in the form of machine learning,remote sensing,and sensor data fusion to generate decision-ready environmental insights.We use the application-first approach,which considers remote sensing,in situ and Internet of Things(IoT)sensing,and physics-based models as complementary streams of evidence with similar strengths and failures.We look critically at how an integrated system can convert heterogeneous observations to action products across three high impact application areas:atmosphere and air quality,water–land–ecosystem dynamics,and hazards.Rapid-response situational awareness,ecosystem condition metrics,drought and flood indicators,exposure maps,and hazard/extreme indicators are key products.The integrated systems to environment interface in three high impact application areas:atmosphere and air quality,water-land-ecosystem dynamics,and hazard Examine Our operational requirements can often determine real-life value such as latency,time stability,smooth degradation in the presence of missing or degraded inputs,and calibrated uncertainty usable in thresholdbased decisions.These pitfalls are common across fields:mismatch in the scale between a point sensor and a gridded product,objectives on proxies in remotely sensed measurements,domain shift in the extremes and changing baselines,and evaluation aspects,which overestimate generalization because of spatiotemporal autocorrelation.Based on these lessons,we present cross-domain proposals for strong validation,uncertainty quantification,provenance,and versioning,as well as fair performance evaluation.We conclude that the next era of environmental intelligence will see a reduction in average accuracy improvement and an increase in terms of robustness,transparency,and operational responsibility,thus allowing the integrated environmental intelligence system to be deployed,which may be relied on to monitor human health,resource allocation,and survival in a more climate-adapted world. 展开更多
关键词 Machine Learning Remote Sensing Sensor data Fusion Environmental Monitoring Uncertainty Quantification
在线阅读 下载PDF
Automated Machine Learning for Fault Diagnosis Using Multimodal Mel-Spectrogram and Vibration Data
18
作者 Zehao Li Xuting Zhang +4 位作者 Hongqi Lin Wu Qin Junyu Qi Zhuyun Chen Qiang Liu 《Computer Modeling in Engineering & Sciences》 2026年第2期471-498,共28页
To ensure the safe and stable operation of rotating machinery,intelligent fault diagnosis methods hold significant research value.However,existing diagnostic approaches largely rely on manual feature extraction and ex... To ensure the safe and stable operation of rotating machinery,intelligent fault diagnosis methods hold significant research value.However,existing diagnostic approaches largely rely on manual feature extraction and expert experience,which limits their adaptability under variable operating conditions and strong noise environments,severely affecting the generalization capability of diagnostic models.To address this issue,this study proposes a multimodal fusion fault diagnosis framework based on Mel-spectrograms and automated machine learning(AutoML).The framework first extracts fault-sensitive Mel time–frequency features from acoustic signals and fuses them with statistical features of vibration signals to construct complementary fault representations.On this basis,automated machine learning techniques are introduced to enable end-to-end diagnostic workflow construction and optimal model configuration acquisition.Finally,diagnostic decisions are achieved by automatically integrating the predictions of multiple high-performance base models.Experimental results on a centrifugal pump vibration and acoustic dataset demonstrate that the proposed framework achieves high diagnostic accuracy under noise-free conditions and maintains strong robustness under noisy interference,validating its efficiency,scalability,and practical value for rotating machinery fault diagnosis. 展开更多
关键词 Automated machine learning mechanical fault diagnosis feature engineering multimodal data
在线阅读 下载PDF
Efficient Arabic Essay Scoring with Hybrid Models: Feature Selection, Data Optimization, and Performance Trade-Offs
19
作者 Mohamed Ezz Meshrif Alruily +4 位作者 Ayman Mohamed Mostafa Alaa SAlaerjan Bader Aldughayfiq Hisham Allahem Abdulaziz Shehab 《Computers, Materials & Continua》 2026年第1期2274-2301,共28页
Automated essay scoring(AES)systems have gained significant importance in educational settings,offering a scalable,efficient,and objective method for evaluating student essays.However,developing AES systems for Arabic... Automated essay scoring(AES)systems have gained significant importance in educational settings,offering a scalable,efficient,and objective method for evaluating student essays.However,developing AES systems for Arabic poses distinct challenges due to the language’s complex morphology,diglossia,and the scarcity of annotated datasets.This paper presents a hybrid approach to Arabic AES by combining text-based,vector-based,and embeddingbased similarity measures to improve essay scoring accuracy while minimizing the training data required.Using a large Arabic essay dataset categorized into thematic groups,the study conducted four experiments to evaluate the impact of feature selection,data size,and model performance.Experiment 1 established a baseline using a non-machine learning approach,selecting top-N correlated features to predict essay scores.The subsequent experiments employed 5-fold cross-validation.Experiment 2 showed that combining embedding-based,text-based,and vector-based features in a Random Forest(RF)model achieved an R2 of 88.92%and an accuracy of 83.3%within a 0.5-point tolerance.Experiment 3 further refined the feature selection process,demonstrating that 19 correlated features yielded optimal results,improving R2 to 88.95%.In Experiment 4,an optimal data efficiency training approach was introduced,where training data portions increased from 5%to 50%.The study found that using just 10%of the data achieved near-peak performance,with an R2 of 85.49%,emphasizing an effective trade-off between performance and computational costs.These findings highlight the potential of the hybrid approach for developing scalable Arabic AES systems,especially in low-resource environments,addressing linguistic challenges while ensuring efficient data usage. 展开更多
关键词 Automated essay scoring text-based features vector-based features embedding-based features feature selection optimal data efficiency
在线阅读 下载PDF
Individual Software Expertise Formalization and Assessment from Project Management Tool Databases
20
作者 Traian-Radu Plosca Alexandru-Mihai Pescaru +1 位作者 Bianca-Valeria Rus Daniel-Ioan Curiac 《Computers, Materials & Continua》 2026年第1期389-411,共23页
Objective expertise evaluation of individuals,as a prerequisite stage for team formation,has been a long-term desideratum in large software development companies.With the rapid advancements in machine learning methods... Objective expertise evaluation of individuals,as a prerequisite stage for team formation,has been a long-term desideratum in large software development companies.With the rapid advancements in machine learning methods,based on reliable existing data stored in project management tools’datasets,automating this evaluation process becomes a natural step forward.In this context,our approach focuses on quantifying software developer expertise by using metadata from the task-tracking systems.For this,we mathematically formalize two categories of expertise:technology-specific expertise,which denotes the skills required for a particular technology,and general expertise,which encapsulates overall knowledge in the software industry.Afterward,we automatically classify the zones of expertise associated with each task a developer has worked on using Bidirectional Encoder Representations from Transformers(BERT)-like transformers to handle the unique characteristics of project tool datasets effectively.Finally,our method evaluates the proficiency of each software specialist across already completed projects from both technology-specific and general perspectives.The method was experimentally validated,yielding promising results. 展开更多
关键词 Expertise formalization transformer-based models natural language processing augmented data project management tool skill classification
在线阅读 下载PDF
上一页 1 2 250 下一页 到第
使用帮助 返回顶部