This review provides an insightful and comprehensive exploration of the emerging 2D material borophene,both pristine and modified,emphasizing its unique attributes and potential for sustainable applications.Borophene...This review provides an insightful and comprehensive exploration of the emerging 2D material borophene,both pristine and modified,emphasizing its unique attributes and potential for sustainable applications.Borophene’s distinctive properties include its anisotropic crystal structures that contribute to its exceptional mechanical and electronic properties.The material exhibits superior electrical and thermal conductivity,surpassing many other 2D materials.Borophene’s unique atomic spin arrangements further diversify its potential application for magnetism.Surface and interface engineering,through doping,functionalization,and synthesis of hybridized and nanocomposite borophene-based systems,is crucial for tailoring borophene’s properties to specific applications.This review aims to address this knowledge gap through a comprehensive and critical analysis of different synthetic and functionalisation methods,to enhance surface reactivity by increasing active sites through doping and surface modifications.These approaches optimize diffusion pathways improving accessibility for catalytic reactions,and tailor the electronic density to tune the optical and electronic behavior.Key applications explored include energy systems(batteries,supercapacitors,and hydrogen storage),catalysis for hydrogen and oxygen evolution reactions,sensors,and optoelectronics for advanced photonic devices.The key to all these applications relies on strategies to introduce heteroatoms for tuning electronic and catalytic properties,employ chemical modifications to enhance stability and leverage borophene’s conductivity and reactivity for advanced photonics.Finally,the review addresses challenges and proposes solutions such as encapsulation,functionalization,and integration with composites to mitigate oxidation sensitivity and overcome scalability barriers,enabling sustainable,commercial-scale applications.展开更多
Over the past years,many efforts have been accomplished to achieve fast and accurate meta-heuristic algorithms to optimize a variety of real-world problems.This study presents a new optimization method based on an unu...Over the past years,many efforts have been accomplished to achieve fast and accurate meta-heuristic algorithms to optimize a variety of real-world problems.This study presents a new optimization method based on an unusual geological phenomenon in nature,named Geyser inspired Algorithm(GEA).The mathematical modeling of this geological phenomenon is carried out to have a better understanding of the optimization process.The efficiency and accuracy of GEA are verified using statistical examination and convergence rate comparison on numerous CEC 2005,CEC 2014,CEC 2017,and real-parameter benchmark functions.Moreover,GEA has been applied to several real-parameter engineering optimization problems to evaluate its effectiveness.In addition,to demonstrate the applicability and robustness of GEA,a comprehensive investigation is performed for a fair comparison with other standard optimization methods.The results demonstrate that GEA is noticeably prosperous in reaching the optimal solutions with a high convergence rate in comparison with other well-known nature-inspired algorithms,including ABC,BBO,PSO,and RCGA.Note that the source code of the GEA is publicly available at https://www.optim-app.com/projects/gea.展开更多
DURING our discussion at workshops for writing“What Does ChatGPT Say:The DAO from Algorithmic Intelligence to Linguistic Intelligence”[1],we had expected the next milestone for Artificial Intelligence(AI)would be in...DURING our discussion at workshops for writing“What Does ChatGPT Say:The DAO from Algorithmic Intelligence to Linguistic Intelligence”[1],we had expected the next milestone for Artificial Intelligence(AI)would be in the direction of Imaginative Intelligence(II),i.e.,something similar to automatic wordsto-videos generation or intelligent digital movies/theater technology that could be used for conducting new“Artificiofactual Experiments”[2]to replace conventional“Counterfactual Experiments”in scientific research and technical development for both natural and social studies[2]-[6].Now we have OpenAI’s Sora,so soon,but this is not the final,actually far away,and it is just the beginning.展开更多
With the deployment of ultra-dense low earth orbit(LEO)satellite constellations,LEO satellite access network(LEO-SAN)is envisioned to achieve global Internet coverage.Meanwhile,the civil aviation communications have i...With the deployment of ultra-dense low earth orbit(LEO)satellite constellations,LEO satellite access network(LEO-SAN)is envisioned to achieve global Internet coverage.Meanwhile,the civil aviation communications have increased dramatically,especially for providing airborne Internet services.However,due to dynamic service demands and onboard LEO resources over time and space,it poses huge challenges in satellite-aircraft access and service management in ultra-dense LEO satellite networks(UDLSN).In this paper,we propose a deep reinforcement learning-based approach for ultra-dense LEO satellite-aircraft access and service management.Firstly,we develop an airborne Internet architecture based on UDLSN and design a management mechanism including medium earth orbit satellites to guarantee lightweight management.Secondly,considering latency-sensitive and latency-tolerant services,we formulate the problem of satellite-aircraft access and service management for civil aviation to ensure service continuity.Finally,we propose a proximal policy optimization-based access and service management algorithm to solve the formulated problem.Simulation results demonstrate the convergence and effectiveness of the proposed algorithm with satisfying the service continuity when applying to the UDLSN.展开更多
Plasmonic nanoantennas provide unique opportunities for precise control of light–matter coupling in surface-enhanced infrared absorption(SEIRA)spectroscopy,but most of the resonant systems realized so far suffer from...Plasmonic nanoantennas provide unique opportunities for precise control of light–matter coupling in surface-enhanced infrared absorption(SEIRA)spectroscopy,but most of the resonant systems realized so far suffer from the obstacles of low sensitivity,narrow bandwidth,and asymmetric Fano resonance perturbations.Here,we demonstrated an overcoupled resonator with a high plasmon-molecule coupling coefficient(μ)(OC-Hμresonator)by precisely controlling the radiation loss channel,the resonator-oscillator coupling channel,and the frequency detuning channel.We observed a strong dependence of the sensing performance on the coupling state,and demonstrated that OC-Hμresonator has excellent sensing properties of ultra-sensitive(7.25%nm^(−1)),ultra-broadband(3–10μm),and immune asymmetric Fano lineshapes.These characteristics represent a breakthrough in SEIRA technology and lay the foundation for specific recognition of biomolecules,trace detection,and protein secondary structure analysis using a single array(array size is 100×100μm^(2)).In addition,with the assistance of machine learning,mixture classification,concentration prediction and spectral reconstruction were achieved with the highest accuracy of 100%.Finally,we demonstrated the potential of OC-Hμresonator for SARS-CoV-2 detection.These findings will promote the wider application of SEIRA technology,while providing new ideas for other enhanced spectroscopy technologies,quantum photonics and studying light–matter interactions.展开更多
Transformer models have emerged as pivotal tools within the realm of drug discovery,distinguished by their unique architectural features and exceptional performance in managing intricate data landscapes.Leveraging the...Transformer models have emerged as pivotal tools within the realm of drug discovery,distinguished by their unique architectural features and exceptional performance in managing intricate data landscapes.Leveraging the innate capabilities of transformer architectures to comprehend intricate hierarchical dependencies inherent in sequential data,these models showcase remarkable efficacy across various tasks,including new drug design and drug target identification.The adaptability of pre-trained trans-former-based models renders them indispensable assets for driving data-centric advancements in drug discovery,chemistry,and biology,furnishing a robust framework that expedites innovation and dis-covery within these domains.Beyond their technical prowess,the success of transformer-based models in drug discovery,chemistry,and biology extends to their interdisciplinary potential,seamlessly combining biological,physical,chemical,and pharmacological insights to bridge gaps across diverse disciplines.This integrative approach not only enhances the depth and breadth of research endeavors but also fosters synergistic collaborations and exchange of ideas among disparate fields.In our review,we elucidate the myriad applications of transformers in drug discovery,as well as chemistry and biology,spanning from protein design and protein engineering,to molecular dynamics(MD),drug target iden-tification,transformer-enabled drug virtual screening(VS),drug lead optimization,drug addiction,small data set challenges,chemical and biological image analysis,chemical language understanding,and single cell data.Finally,we conclude the survey by deliberating on promising trends in transformer models within the context of drug discovery and other sciences.展开更多
Electrical energy is essential for modern society to sustain economic growths.The soaring demand for the electrical energy,together with an awareness of the environmental impact of fossil fuels,has been driving a shif...Electrical energy is essential for modern society to sustain economic growths.The soaring demand for the electrical energy,together with an awareness of the environmental impact of fossil fuels,has been driving a shift towards the utilization of solar energy.However,traditional solar energy solutions often require extensive spaces for a panel installation,limiting their practicality in a dense urban environment.To overcome the spatial constraint,researchers have developed transparent photovoltaics(TPV),enabling windows and facades in vehicles and buildings to generate electric energy.Current TPV advancements are focused on improving both transparency and power output to rival commercially available silicon solar panels.In this review,we first briefly introduce wavelength-and non-wavelengthselective strategies to achieve transparency.Figures of merit and theoretical limits of TPVs are discussed to comprehensively understand the status of current TPV technology.Then we highlight recent progress in different types of TPVs,with a particular focus on solution-processed thin-film photovoltaics(PVs),including colloidal quantum dot PVs,metal halide perovskite PVs and organic PVs.The applications of TPVs are also reviewed,with emphasis on agrivoltaics,smart windows and facades.Finally,current challenges and future opportunities in TPV research are pointed out.展开更多
Although there are numerous optical spectroscopy techniques and methods that have been used to extract the fundamental bandgap of a semiconductor,most of them belong to one of these three approaches:(1)the excitonic a...Although there are numerous optical spectroscopy techniques and methods that have been used to extract the fundamental bandgap of a semiconductor,most of them belong to one of these three approaches:(1)the excitonic absorption,(2)modulation spectroscopy,and(3)the most widely used Tauc-plot.The excitonic absorption is based on a many-particle theory,which is physically the most correct approach,but requires more stringent crystalline quality and appropriate sample preparation and experimental implementation.The Tauc-plot is based on a single-particle theo⁃ry that neglects the many-electron effects.Modulation spectroscopy analyzes the spectroscopy features in the derivative spectrum,typically,of the reflectance and transmission under an external perturbation.Empirically,the bandgap ener⁃gy derived from the three approaches follow the order of E_(ex)>E_(MS)>E_(TP),where three transition energies are from exci⁃tonic absorption,modulation spectroscopy,and Tauc-plot,respectively.In principle,defining E_(g) as the single-elec⁃tron bandgap,we expect E_(g)>E_(ex),thus,E_(g)>E_(TP).In the literature,E_(TP) is often interpreted as E_(g),which is conceptual⁃ly problematic.However,in many cases,because the excitonic peaks are not readily identifiable,the inconsistency be⁃tween E_(g) and E_(TP) becomes invisible.In this brief review,real world examples are used(1)to illustrate how excitonic absorption features depend sensitively on the sample and measurement conditions;(2)to demonstrate the differences between E_(ex),E_(MS),and E_(TP) when they can be extracted simultaneously for one sample;and(3)to show how the popular⁃ly adopted Tauc-plot could lead to misleading results.Finally,it is pointed out that if the excitonic absorption is not ob⁃servable,the modulation spectroscopy can often yield a more useful and reasonable bandgap than Tauc-plot.展开更多
W-type barium-nickel ferrite(BaNi_(2)Fe_(16)O_(27))is a highly promising material for electromagnetic wave(EMW)absorption be-cause of its magnetic loss capability for EMW,low cost,large-scale production potential,high...W-type barium-nickel ferrite(BaNi_(2)Fe_(16)O_(27))is a highly promising material for electromagnetic wave(EMW)absorption be-cause of its magnetic loss capability for EMW,low cost,large-scale production potential,high-temperature resistance,and excellent chemical stability.However,the poor dielectric loss of magnetic ferrites hampers their utilization,hindering enhancement in their EMW-absorption performance.Developing efficient strategies that improve the EMW-absorption performance of ferrite is highly desired but re-mains challenging.Here,an efficient strategy substituting Ba^(2+)with rare earth La^(3+)in W-type ferrite was proposed for the preparation of novel La-substituted ferrites(Ba_(1-x)LaxNi_(2)Fe_(15.4)O_(27)).The influences of La^(3+)substitution on ferrites’EMW-absorption performance and the dissipative mechanism toward EMW were systematically explored and discussed.La^(3+)efficiently induced lattice defects,enhanced defect-induced polarization,and slightly reduced the ferrites’bandgap,enhancing the dielectric properties of the ferrites.La^(3+)also enhanced the ferromagnetic resonance loss and strengthened magnetic properties.These effects considerably improved the EMW-absorption perform-ance of Ba_(1-x)LaxNi_(2)Fe_(15.4)O_(27)compared with pure W-type ferrites.When x=0.2,the best EMW-absorption performance was achieved with a minimum reflection loss of-55.6 dB and effective absorption bandwidth(EAB)of 3.44 GHz.展开更多
Biometric template protection is essential for finger-based authentication systems,as template tampering and adversarial attacks threaten the security.This paper proposes a DCT-based fragile watermarking scheme incorp...Biometric template protection is essential for finger-based authentication systems,as template tampering and adversarial attacks threaten the security.This paper proposes a DCT-based fragile watermarking scheme incorporating AI-based tamper detection to improve the integrity and robustness of finger authentication.The system was tested against NIST SD4 and Anguli fingerprint datasets,wherein 10,000 watermarked fingerprints were employed for training.The designed approach recorded a tamper detection rate of 98.3%,performing 3–6%better than current DCT,SVD,and DWT-based watermarking approaches.The false positive rate(≤1.2%)and false negative rate(≤1.5%)were much lower compared to previous research,which maintained high reliability for template change detection.The system showed real-time performance,averaging 12–18 ms processing time per template,and is thus suitable for real-world biometric authentication scenarios.Quality analysis of fingerprints indicated that NFIQ scores were enhanced from 2.07 to 1.81,reflecting improved minutiae clarity and ridge structure preservation.The approach also exhibited strong resistance to compression and noise distortions,with the improvements in PSNR being 2 dB(JPEG compression Q=80)and the SSIM values rising by 3%–5%under noise attacks.Comparative assessment demonstrated that training with NIST SD4 data greatly improved the ridge continuity and quality of fingerprints,resulting in better match scores(260–295)when tested against Bozorth3.Smaller batch sizes(batch=2)also resulted in improved ridge clarity,whereas larger batch sizes(batch=8)resulted in distortions.The DCNN-based tamper detection model supported real-time classification,which greatly minimized template exposure to adversarial attacks and synthetic fingerprint forgeries.Results demonstrate that fragile watermarking with AI indeed greatly enhances fingerprint security,providing privacy-preserving biometric authentication with high robustness,accuracy,and computational efficiency.展开更多
The rapid growth of digital data necessitates advanced natural language processing(NLP)models like BERT(Bidi-rectional Encoder Representations from Transformers),known for its superior performance in text classificati...The rapid growth of digital data necessitates advanced natural language processing(NLP)models like BERT(Bidi-rectional Encoder Representations from Transformers),known for its superior performance in text classification.However,BERT’s size and computational demands limit its practicality,especially in resource-constrained settings.This research compresses the BERT base model for Bengali emotion classification through knowledge distillation(KD),pruning,and quantization techniques.Despite Bengali being the sixth most spoken language globally,NLP research in this area is limited.Our approach addresses this gap by creating an efficient BERT-based model for Bengali text.We have explored 20 combinations for KD,quantization,and pruning,resulting in improved speedup,fewer parameters,and reduced memory size.Our best results demonstrate significant improvements in both speed and efficiency.For instance,in the case of mBERT,we achieved a 3.87×speedup and 4×compression ratio with a combination of Distil+Prune+Quant that reduced parameters from 178 to 46 M,while the memory size decreased from 711 to 178 MB.These results offer scalable solutions for NLP tasks in various languages and advance the field of model compression,making these models suitable for real-world applications in resource-limited environments.展开更多
Acute lymphoblastic leukemia(ALL)is characterized by overgrowth of immature lymphoid cells in the bone marrow at the expense of normal hematopoiesis.One of the most prioritized tasks is the early and correct diagnosis...Acute lymphoblastic leukemia(ALL)is characterized by overgrowth of immature lymphoid cells in the bone marrow at the expense of normal hematopoiesis.One of the most prioritized tasks is the early and correct diagnosis of this malignancy;however,manual observation of the blood smear is very time-consuming and requires labor and expertise.Transfer learning in deep neural networks is of growing importance to intricate medical tasks such as medical imaging.Our work proposes an application of a novel ensemble architecture that puts together Vision Transformer and EfficientNetV2.This approach fuses deep and spatial features to optimize discriminative power by selecting features accurately,reducing redundancy,and promoting sparsity.Besides the architecture of the ensemble,the advanced feature selection is performed by the Frog-Snake Prey-Predation Relationship Optimization(FSRO)algorithm.FSRO prioritizes the most relevant features while dynamically reducing redundant and noisy data,hence improving the efficiency and accuracy of the classification model.We have compared our method for feature selection against state-of-the-art techniques and recorded an accuracy of 94.88%,a recall of 94.38%,a precision of 96.18%,and an F1-score of 95.63%.These figures are therefore better than the classical methods for deep learning.Though our dataset,collected from four different hospitals,is non-standard and heterogeneous,making the analysis more challenging,although computationally expensive,our approach proves diagnostically superior in cancer detection.Source codes and datasets are available on GitHub.展开更多
Air pollution,specifically fine particulate matter(PM2.5),represents a critical environmental and public health concern due to its adverse effects on respiratory and cardiovascular systems.Accurate forecasting of PM2....Air pollution,specifically fine particulate matter(PM2.5),represents a critical environmental and public health concern due to its adverse effects on respiratory and cardiovascular systems.Accurate forecasting of PM2.5 concentrations is essential for mitigating health risks;however,the inherent nonlinearity and dynamic variability of air quality data present significant challenges.This study conducts a systematic evaluation of deep learning algorithms including Convolutional Neural Network(CNN),Long Short-Term Memory(LSTM),and the hybrid CNN-LSTM as well as statistical models,AutoRegressive Integrated Moving Average(ARIMA)and Maximum Likelihood Estimation(MLE)for hourly PM2.5 forecasting.Model performance is quantified using Root Mean Squared Error(RMSE),Mean Absolute Error(MAE),Mean Absolute Percentage Error(MAPE),and the Coefficient of Determination(R^(2))metrics.The comparative analysis identifies optimal predictive approaches for air quality modeling,emphasizing computational efficiency and accuracy.Additionally,CNN classification performance is evaluated using a confusion matrix,accuracy,precision,and F1-score.The results demonstrate that the Hybrid CNN-LSTM model outperforms standalone models,exhibiting lower error rates and higher R^(2) values,thereby highlighting the efficacy of deep learning-based hybrid architectures in achieving robust and precise PM2.5 forecasting.This study underscores the potential of advanced computational techniques in enhancing air quality prediction systems for environmental and public health applications.展开更多
The application of generative artificial intelligence(AI)is bringing about notable changes in anime creation.This paper surveys recent advancements and applications of diffusion and language models in anime generation...The application of generative artificial intelligence(AI)is bringing about notable changes in anime creation.This paper surveys recent advancements and applications of diffusion and language models in anime generation,focusing on their demonstrated potential to enhance production efficiency through automation and personalization.Despite these benefits,it is crucial to acknowledge the substantial initial computational investments required for training and deploying these models.We conduct an in-depth survey of cutting-edge generative AI technologies,encompassing models such as Stable Diffusion and GPT,and appraise pivotal large-scale datasets alongside quantifiable evaluation metrics.Review of the surveyed literature indicates the achievement of considerable maturity in the capacity of AI models to synthesize high-quality,aesthetically compelling anime visual images from textual prompts,alongside discernible progress in the generation of coherent narratives.However,achieving perfect long-form consistency,mitigating artifacts like flickering in video sequences,and enabling fine-grained artistic control remain critical ongoing challenges.Building upon these advancements,research efforts have increasingly pivoted towards the synthesis of higher-dimensional content,such as video and three-dimensional assets,with recent studies demonstrating significant progress in this burgeoning field.Nevertheless,formidable challenges endure amidst these advancements.Foremost among these are the substantial computational exigencies requisite for training and deploying these sophisticated models,particularly pronounced in the realm of high-dimensional generation such as video synthesis.Additional persistent hurdles include maintaining spatial-temporal consistency across complex scenes and mitigating ethical considerations surrounding bias and the preservation of human creative autonomy.This research underscores the transformative potential and inherent complexities of AI-driven synergy within the creative industries.We posit that future research should be dedicated to the synergistic fusion of diffusion and autoregressive models,the integration of multimodal inputs,and the balanced consideration of ethical implications,particularly regarding bias and the preservation of human creative autonomy,thereby establishing a robust foundation for the advancement of anime creation and the broader landscape of AI-driven content generation.展开更多
Wireless Sensor Networks(WSNs)have emerged as crucial tools for real-time environmental monitoring through distributed sensor nodes(SNs).However,the operational lifespan of WSNs is significantly constrained by the lim...Wireless Sensor Networks(WSNs)have emerged as crucial tools for real-time environmental monitoring through distributed sensor nodes(SNs).However,the operational lifespan of WSNs is significantly constrained by the limited energy resources of SNs.Current energy efficiency strategies,such as clustering,multi-hop routing,and data aggregation,face challenges,including uneven energy depletion,high computational demands,and suboptimal cluster head(CH)selection.To address these limitations,this paper proposes a hybrid methodology that optimizes energy consumption(EC)while maintaining network performance.The proposed approach integrates the Low Energy Adaptive Clustering Hierarchy with Deterministic(LEACH-D)protocol using an Artificial Neural Network(ANN)and Bayesian Regularization Algorithm(BRA).LEACH-D improves upon conventional LEACH by ensuring more uniform energy usage across SNs,mitigating inefficiencies from random CH selection.The ANN further enhances CH selection and routing processes,effectively reducing data transmission overhead and idle listening.Simulation results reveal that the LEACH-D-ANN model significantly reduces EC and extends the network’s lifespan compared to existing protocols.This framework offers a promising solution to the energy efficiency challenges in WSNs,paving the way for more sustainable and reliable network deployments.展开更多
Integrating mobile edge computing (MEC) and digital twin (DT) technologies to enhance network performance through predictive, adaptive control for energy-efficient and low-latency communication is a significant challe...Integrating mobile edge computing (MEC) and digital twin (DT) technologies to enhance network performance through predictive, adaptive control for energy-efficient and low-latency communication is a significant challenge. This paper presents the EcoEdgeTwin model, an innovative framework that harnesses the harmony between MEC and DT technologies to ensure an efficient network operation. We optimise the utility function to balance enhancing users' quality of experience (QoE) and minimising latency and energy consumption at edge servers. This approach ensures efficient and adaptable network operations, utilising DT to synchronise and integrate real-time data seamlessly. Our framework implements robust mechanisms for task offloading, service caching and cost-effective service migration. Additionally, it manages energy consumption related to task processing, communication and the influence of DT predictions, all essential for optimising latency and minimising energy usage. Through the utility model, we also prioritise QoE, fostering a user-centric approach to network management that balances network efficiency with user satisfaction. A cornerstone of our approach is integrating the advantage actor-critic algorithm, marking a pioneering use of deep reinforcement learning for dynamic network management. This strategy addresses challenges in service mobility and network variability, ensuring optimal network performance matrices. Our extensive simulations demonstrate that compared to benchmark models, the EcoEdgeTwin framework significantly reduces energy usage and latency while enhancing QoE.展开更多
Accurate Electric Load Forecasting(ELF)is crucial for optimizing production capacity,improving operational efficiency,and managing energy resources effectively.Moreover,precise ELF contributes to a smaller environment...Accurate Electric Load Forecasting(ELF)is crucial for optimizing production capacity,improving operational efficiency,and managing energy resources effectively.Moreover,precise ELF contributes to a smaller environmental footprint by reducing the risks of disruption,downtime,and waste.However,with increasingly complex energy consumption patterns driven by renewable energy integration and changing consumer behaviors,no single approach has emerged as universally effective.In response,this research presents a hybrid modeling framework that combines the strengths of Random Forest(RF)and Autoregressive Integrated Moving Average(ARIMA)models,enhanced with advanced feature selection—Minimum Redundancy Maximum Relevancy and Maximum Synergy(MRMRMS)method—to produce a sparse model.Additionally,the residual patterns are analyzed to enhance forecast accuracy.High-resolution weather data from Weather Underground and historical energy consumption data from PJM for Duke Energy Ohio and Kentucky(DEO&K)are used in this application.This methodology,termed SP-RF-ARIMA,is evaluated against existing approaches;it demonstrates more than 40%reduction in mean absolute error and root mean square error compared to the second-best method.展开更多
In Power Line Communications(PLC),there are regulatory masks that restrict the transmit power spectral density for electromagnetic compatibility reasons,which creates coverage issues despite the not too long distances...In Power Line Communications(PLC),there are regulatory masks that restrict the transmit power spectral density for electromagnetic compatibility reasons,which creates coverage issues despite the not too long distances.Hence,PLC networks often employ repeaters/relays,especially in smart grid neighborhood area networks.Even in broadband indoor PLC systems that offer a notable data rate,relaying may pave the way to new applications like being the backbone for wireless technologies in a cost-effective manner to support the Internet-of-things paradigm.In this paper,we study Multiple-Input Multiple-Output(MIMO)PLC systems that incorporate inband full-duplex functionality in relaying networks.We present several MIMO configurations that allow end-to-end half-duplex or full-duplex operations and analyze the achievable performance with state-of-the-art PLC systems.To reach this analysis,we get channel realizations from random network layouts for indoor and outdoor scenarios.We adopt realistic MIMO channel and noise models and consider transmission techniques according to PLC standards.The concepts discussed in this work can be useful in the design of future PLC relay-aided networks for different applications that look for a coverage extension and/or throughput:smart grids with enhanced communications in outdoor scenarios,and“last meter”systems for high-speed connections everywhere in indoor ones.展开更多
Structured illumination,a wide-field imaging approach used in microscopy to enhance image resolution beyond the system's diffraction limits,is a well-studied technique that has gained significant traction over the...Structured illumination,a wide-field imaging approach used in microscopy to enhance image resolution beyond the system's diffraction limits,is a well-studied technique that has gained significant traction over the last two decades.However,when translated to endoscopic systems,severe deformations of illumination patterns occur due to the large depth of field(DOF)and the 3D nature of the targets,introducing significant implementation challenges.Hence,this study explores a speckle-based system that best suits endoscopic practices to enhance image resolution by using random illumination patterns.The study presents a prototypic model of an endoscopic add-on,its design,and fabrication facilitated by using the speckle structured illumination endoscopic(SSIE)system.The imaging results of the SSIE are explained on a colon phantom model at different imaging planes with a wide field of view(FOV)and DOF.The obtained imaging metrics are elucidated and compared with state-of-the-art(SOA)high-resolution endoscopic techniques.Moreover,the potential for a clinical translation of the prototypic SSIE model is also explored in this work.The incorporation of the add-on and its subsequent results on the colon phantom model could potentially pave the way for its successful integration and use in futuristic clinical endoscopic trials.展开更多
Cyber-Physical System (CPS) devices are increasing exponentially. Lacking confidentiality creates a vulnerable network. Thus, demanding the overall system with the latest and robust solutions for the defence mechanism...Cyber-Physical System (CPS) devices are increasing exponentially. Lacking confidentiality creates a vulnerable network. Thus, demanding the overall system with the latest and robust solutions for the defence mechanisms with low computation cost, increased integrity, and surveillance. The proposal of a mechanism that utilizes the features of authenticity measures using the Destination Sequence Distance Vector (DSDV) routing protocol which applies to the multi-WSN (Wireless Sensor Network) of IoT devices in CPS which is developed for the Device-to-Device (D2D) authentication developed from the local-chain and public chain respectively combined with the Software Defined Networking (SDN) control and monitoring system using switches and controllers that will route the packets through the network, identify any false nodes, take preventive measures against them and preventing them for any future problems. Next, the system is powered by Blockchain cryptographic features by utilizing the TrustChain features to create a private, secure, and temper-free ledger of the transactions performed inside the network. Results are achieved in the legitimate devices connecting to the network, transferring their packets to their destination under supervision, reporting whenever a false node is causing hurdles, and recording the transactions for temper-proof records. Evaluation results based on 1000+ transactions illustrate that the proposed mechanism not only outshines most aspects of Cyber-Physical systems but also consumes less computation power with a low latency of 0.1 seconds only.展开更多
基金the Engineering and Physical Sciences Research Council(EPSRC)for funding the researchUK India Education Research Initiative(UKIERI)for funding support.
文摘This review provides an insightful and comprehensive exploration of the emerging 2D material borophene,both pristine and modified,emphasizing its unique attributes and potential for sustainable applications.Borophene’s distinctive properties include its anisotropic crystal structures that contribute to its exceptional mechanical and electronic properties.The material exhibits superior electrical and thermal conductivity,surpassing many other 2D materials.Borophene’s unique atomic spin arrangements further diversify its potential application for magnetism.Surface and interface engineering,through doping,functionalization,and synthesis of hybridized and nanocomposite borophene-based systems,is crucial for tailoring borophene’s properties to specific applications.This review aims to address this knowledge gap through a comprehensive and critical analysis of different synthetic and functionalisation methods,to enhance surface reactivity by increasing active sites through doping and surface modifications.These approaches optimize diffusion pathways improving accessibility for catalytic reactions,and tailor the electronic density to tune the optical and electronic behavior.Key applications explored include energy systems(batteries,supercapacitors,and hydrogen storage),catalysis for hydrogen and oxygen evolution reactions,sensors,and optoelectronics for advanced photonic devices.The key to all these applications relies on strategies to introduce heteroatoms for tuning electronic and catalytic properties,employ chemical modifications to enhance stability and leverage borophene’s conductivity and reactivity for advanced photonics.Finally,the review addresses challenges and proposes solutions such as encapsulation,functionalization,and integration with composites to mitigate oxidation sensitivity and overcome scalability barriers,enabling sustainable,commercial-scale applications.
文摘Over the past years,many efforts have been accomplished to achieve fast and accurate meta-heuristic algorithms to optimize a variety of real-world problems.This study presents a new optimization method based on an unusual geological phenomenon in nature,named Geyser inspired Algorithm(GEA).The mathematical modeling of this geological phenomenon is carried out to have a better understanding of the optimization process.The efficiency and accuracy of GEA are verified using statistical examination and convergence rate comparison on numerous CEC 2005,CEC 2014,CEC 2017,and real-parameter benchmark functions.Moreover,GEA has been applied to several real-parameter engineering optimization problems to evaluate its effectiveness.In addition,to demonstrate the applicability and robustness of GEA,a comprehensive investigation is performed for a fair comparison with other standard optimization methods.The results demonstrate that GEA is noticeably prosperous in reaching the optimal solutions with a high convergence rate in comparison with other well-known nature-inspired algorithms,including ABC,BBO,PSO,and RCGA.Note that the source code of the GEA is publicly available at https://www.optim-app.com/projects/gea.
基金the National Natural Science Foundation of China(62271485,61903363,U1811463,62103411,62203250)the Science and Technology Development Fund of Macao SAR(0093/2023/RIA2,0050/2020/A1)。
文摘DURING our discussion at workshops for writing“What Does ChatGPT Say:The DAO from Algorithmic Intelligence to Linguistic Intelligence”[1],we had expected the next milestone for Artificial Intelligence(AI)would be in the direction of Imaginative Intelligence(II),i.e.,something similar to automatic wordsto-videos generation or intelligent digital movies/theater technology that could be used for conducting new“Artificiofactual Experiments”[2]to replace conventional“Counterfactual Experiments”in scientific research and technical development for both natural and social studies[2]-[6].Now we have OpenAI’s Sora,so soon,but this is not the final,actually far away,and it is just the beginning.
基金supported in part by the National Key R&D Program of China under Grant 2020YFB1806104in part by Innovation and Entrepreneurship of Jiangsu Province High-level Talent Program+1 种基金in part by Natural Sciences and Engineering Research Council of Canada (NSERC)the support from Huawei
文摘With the deployment of ultra-dense low earth orbit(LEO)satellite constellations,LEO satellite access network(LEO-SAN)is envisioned to achieve global Internet coverage.Meanwhile,the civil aviation communications have increased dramatically,especially for providing airborne Internet services.However,due to dynamic service demands and onboard LEO resources over time and space,it poses huge challenges in satellite-aircraft access and service management in ultra-dense LEO satellite networks(UDLSN).In this paper,we propose a deep reinforcement learning-based approach for ultra-dense LEO satellite-aircraft access and service management.Firstly,we develop an airborne Internet architecture based on UDLSN and design a management mechanism including medium earth orbit satellites to guarantee lightweight management.Secondly,considering latency-sensitive and latency-tolerant services,we formulate the problem of satellite-aircraft access and service management for civil aviation to ensure service continuity.Finally,we propose a proximal policy optimization-based access and service management algorithm to solve the formulated problem.Simulation results demonstrate the convergence and effectiveness of the proposed algorithm with satisfying the service continuity when applying to the UDLSN.
基金supported by A*STAR under the“Nanosystems at the Edge”program(Grant No.A18A4b0055)Ministry of Education(MOE)under the research grant of R-263-000-F18-112/A-0009520-01-00+1 种基金National Research Foundation Singapore grant CRP28-2022-0038the Reimagine Re-search Scheme(RRSC)Project(Grant A-0009037-02-00&A0009037-03-00)at National University of Singapore.
文摘Plasmonic nanoantennas provide unique opportunities for precise control of light–matter coupling in surface-enhanced infrared absorption(SEIRA)spectroscopy,but most of the resonant systems realized so far suffer from the obstacles of low sensitivity,narrow bandwidth,and asymmetric Fano resonance perturbations.Here,we demonstrated an overcoupled resonator with a high plasmon-molecule coupling coefficient(μ)(OC-Hμresonator)by precisely controlling the radiation loss channel,the resonator-oscillator coupling channel,and the frequency detuning channel.We observed a strong dependence of the sensing performance on the coupling state,and demonstrated that OC-Hμresonator has excellent sensing properties of ultra-sensitive(7.25%nm^(−1)),ultra-broadband(3–10μm),and immune asymmetric Fano lineshapes.These characteristics represent a breakthrough in SEIRA technology and lay the foundation for specific recognition of biomolecules,trace detection,and protein secondary structure analysis using a single array(array size is 100×100μm^(2)).In addition,with the assistance of machine learning,mixture classification,concentration prediction and spectral reconstruction were achieved with the highest accuracy of 100%.Finally,we demonstrated the potential of OC-Hμresonator for SARS-CoV-2 detection.These findings will promote the wider application of SEIRA technology,while providing new ideas for other enhanced spectroscopy technologies,quantum photonics and studying light–matter interactions.
基金supported in part by National Institute of Health(NIH),USA(Grant Nos.:R01GM126189,R01AI164266,and R35GM148196)the National Science Foundation,USA(Grant Nos.DMS2052983,DMS-1761320,and IIS-1900473)+3 种基金National Aero-nautics and Space Administration(NASA),USA(Grant No.:80NSSC21M0023)Michigan State University(MSU)Foundation,USA,Bristol-Myers Squibb(Grant No.:65109)USA,and Pfizer,USAsupported by the National Natural Science Foundation of China(Grant Nos.:11971367,12271416,and 11972266).
文摘Transformer models have emerged as pivotal tools within the realm of drug discovery,distinguished by their unique architectural features and exceptional performance in managing intricate data landscapes.Leveraging the innate capabilities of transformer architectures to comprehend intricate hierarchical dependencies inherent in sequential data,these models showcase remarkable efficacy across various tasks,including new drug design and drug target identification.The adaptability of pre-trained trans-former-based models renders them indispensable assets for driving data-centric advancements in drug discovery,chemistry,and biology,furnishing a robust framework that expedites innovation and dis-covery within these domains.Beyond their technical prowess,the success of transformer-based models in drug discovery,chemistry,and biology extends to their interdisciplinary potential,seamlessly combining biological,physical,chemical,and pharmacological insights to bridge gaps across diverse disciplines.This integrative approach not only enhances the depth and breadth of research endeavors but also fosters synergistic collaborations and exchange of ideas among disparate fields.In our review,we elucidate the myriad applications of transformers in drug discovery,as well as chemistry and biology,spanning from protein design and protein engineering,to molecular dynamics(MD),drug target iden-tification,transformer-enabled drug virtual screening(VS),drug lead optimization,drug addiction,small data set challenges,chemical and biological image analysis,chemical language understanding,and single cell data.Finally,we conclude the survey by deliberating on promising trends in transformer models within the context of drug discovery and other sciences.
基金supported by the National Natural Science Foundation of China(Grant number W2432035)financial support from the EPSRC SWIMS(EP/V039717/1)+3 种基金Royal Society(RGS\R1\221009 and IEC\NSFC\211201)Leverhulme Trust(RPG-2022-263)Ser Cymru programme–Enhancing Competitiveness Equipment Awards 2022-23(MA/VG/2715/22-PN66)the financial support from Kingdom of Saudi Arabia Ministry of Higher Education.
文摘Electrical energy is essential for modern society to sustain economic growths.The soaring demand for the electrical energy,together with an awareness of the environmental impact of fossil fuels,has been driving a shift towards the utilization of solar energy.However,traditional solar energy solutions often require extensive spaces for a panel installation,limiting their practicality in a dense urban environment.To overcome the spatial constraint,researchers have developed transparent photovoltaics(TPV),enabling windows and facades in vehicles and buildings to generate electric energy.Current TPV advancements are focused on improving both transparency and power output to rival commercially available silicon solar panels.In this review,we first briefly introduce wavelength-and non-wavelengthselective strategies to achieve transparency.Figures of merit and theoretical limits of TPVs are discussed to comprehensively understand the status of current TPV technology.Then we highlight recent progress in different types of TPVs,with a particular focus on solution-processed thin-film photovoltaics(PVs),including colloidal quantum dot PVs,metal halide perovskite PVs and organic PVs.The applications of TPVs are also reviewed,with emphasis on agrivoltaics,smart windows and facades.Finally,current challenges and future opportunities in TPV research are pointed out.
基金Supported by Bissell Distinguished Professor Endowment Fund at UNC-Charlotte。
文摘Although there are numerous optical spectroscopy techniques and methods that have been used to extract the fundamental bandgap of a semiconductor,most of them belong to one of these three approaches:(1)the excitonic absorption,(2)modulation spectroscopy,and(3)the most widely used Tauc-plot.The excitonic absorption is based on a many-particle theory,which is physically the most correct approach,but requires more stringent crystalline quality and appropriate sample preparation and experimental implementation.The Tauc-plot is based on a single-particle theo⁃ry that neglects the many-electron effects.Modulation spectroscopy analyzes the spectroscopy features in the derivative spectrum,typically,of the reflectance and transmission under an external perturbation.Empirically,the bandgap ener⁃gy derived from the three approaches follow the order of E_(ex)>E_(MS)>E_(TP),where three transition energies are from exci⁃tonic absorption,modulation spectroscopy,and Tauc-plot,respectively.In principle,defining E_(g) as the single-elec⁃tron bandgap,we expect E_(g)>E_(ex),thus,E_(g)>E_(TP).In the literature,E_(TP) is often interpreted as E_(g),which is conceptual⁃ly problematic.However,in many cases,because the excitonic peaks are not readily identifiable,the inconsistency be⁃tween E_(g) and E_(TP) becomes invisible.In this brief review,real world examples are used(1)to illustrate how excitonic absorption features depend sensitively on the sample and measurement conditions;(2)to demonstrate the differences between E_(ex),E_(MS),and E_(TP) when they can be extracted simultaneously for one sample;and(3)to show how the popular⁃ly adopted Tauc-plot could lead to misleading results.Finally,it is pointed out that if the excitonic absorption is not ob⁃servable,the modulation spectroscopy can often yield a more useful and reasonable bandgap than Tauc-plot.
基金financially supported by the National Key R&D Program of China(No.2021YFB3502500)the Natur-al Science Foundation of Shandong Province,China(No.2022HYYQ-014)+5 种基金the“20 Clauses about Colleges and Uni-versities(new)”(Independent Training of Innovation Team)Program of Jinan,China(No.2021GXRC036)the Provin-cial Key Research and Development Program of Shandong,China(No.2021ZLGX01)the National Natural Science Foundation of China(No.22375115)the Joint Laboratory project of Electromagnetic Structure Technology(No.637-2022-70-F-037)the Discipline Construction Expenditure for Distinguished Young Scholars of Shandong University,China(No.31370089963141)the Qilu Young Scholar Program of Shandong University,China(No.31370082163127).
文摘W-type barium-nickel ferrite(BaNi_(2)Fe_(16)O_(27))is a highly promising material for electromagnetic wave(EMW)absorption be-cause of its magnetic loss capability for EMW,low cost,large-scale production potential,high-temperature resistance,and excellent chemical stability.However,the poor dielectric loss of magnetic ferrites hampers their utilization,hindering enhancement in their EMW-absorption performance.Developing efficient strategies that improve the EMW-absorption performance of ferrite is highly desired but re-mains challenging.Here,an efficient strategy substituting Ba^(2+)with rare earth La^(3+)in W-type ferrite was proposed for the preparation of novel La-substituted ferrites(Ba_(1-x)LaxNi_(2)Fe_(15.4)O_(27)).The influences of La^(3+)substitution on ferrites’EMW-absorption performance and the dissipative mechanism toward EMW were systematically explored and discussed.La^(3+)efficiently induced lattice defects,enhanced defect-induced polarization,and slightly reduced the ferrites’bandgap,enhancing the dielectric properties of the ferrites.La^(3+)also enhanced the ferromagnetic resonance loss and strengthened magnetic properties.These effects considerably improved the EMW-absorption perform-ance of Ba_(1-x)LaxNi_(2)Fe_(15.4)O_(27)compared with pure W-type ferrites.When x=0.2,the best EMW-absorption performance was achieved with a minimum reflection loss of-55.6 dB and effective absorption bandwidth(EAB)of 3.44 GHz.
文摘Biometric template protection is essential for finger-based authentication systems,as template tampering and adversarial attacks threaten the security.This paper proposes a DCT-based fragile watermarking scheme incorporating AI-based tamper detection to improve the integrity and robustness of finger authentication.The system was tested against NIST SD4 and Anguli fingerprint datasets,wherein 10,000 watermarked fingerprints were employed for training.The designed approach recorded a tamper detection rate of 98.3%,performing 3–6%better than current DCT,SVD,and DWT-based watermarking approaches.The false positive rate(≤1.2%)and false negative rate(≤1.5%)were much lower compared to previous research,which maintained high reliability for template change detection.The system showed real-time performance,averaging 12–18 ms processing time per template,and is thus suitable for real-world biometric authentication scenarios.Quality analysis of fingerprints indicated that NFIQ scores were enhanced from 2.07 to 1.81,reflecting improved minutiae clarity and ridge structure preservation.The approach also exhibited strong resistance to compression and noise distortions,with the improvements in PSNR being 2 dB(JPEG compression Q=80)and the SSIM values rising by 3%–5%under noise attacks.Comparative assessment demonstrated that training with NIST SD4 data greatly improved the ridge continuity and quality of fingerprints,resulting in better match scores(260–295)when tested against Bozorth3.Smaller batch sizes(batch=2)also resulted in improved ridge clarity,whereas larger batch sizes(batch=8)resulted in distortions.The DCNN-based tamper detection model supported real-time classification,which greatly minimized template exposure to adversarial attacks and synthetic fingerprint forgeries.Results demonstrate that fragile watermarking with AI indeed greatly enhances fingerprint security,providing privacy-preserving biometric authentication with high robustness,accuracy,and computational efficiency.
文摘The rapid growth of digital data necessitates advanced natural language processing(NLP)models like BERT(Bidi-rectional Encoder Representations from Transformers),known for its superior performance in text classification.However,BERT’s size and computational demands limit its practicality,especially in resource-constrained settings.This research compresses the BERT base model for Bengali emotion classification through knowledge distillation(KD),pruning,and quantization techniques.Despite Bengali being the sixth most spoken language globally,NLP research in this area is limited.Our approach addresses this gap by creating an efficient BERT-based model for Bengali text.We have explored 20 combinations for KD,quantization,and pruning,resulting in improved speedup,fewer parameters,and reduced memory size.Our best results demonstrate significant improvements in both speed and efficiency.For instance,in the case of mBERT,we achieved a 3.87×speedup and 4×compression ratio with a combination of Distil+Prune+Quant that reduced parameters from 178 to 46 M,while the memory size decreased from 711 to 178 MB.These results offer scalable solutions for NLP tasks in various languages and advance the field of model compression,making these models suitable for real-world applications in resource-limited environments.
文摘Acute lymphoblastic leukemia(ALL)is characterized by overgrowth of immature lymphoid cells in the bone marrow at the expense of normal hematopoiesis.One of the most prioritized tasks is the early and correct diagnosis of this malignancy;however,manual observation of the blood smear is very time-consuming and requires labor and expertise.Transfer learning in deep neural networks is of growing importance to intricate medical tasks such as medical imaging.Our work proposes an application of a novel ensemble architecture that puts together Vision Transformer and EfficientNetV2.This approach fuses deep and spatial features to optimize discriminative power by selecting features accurately,reducing redundancy,and promoting sparsity.Besides the architecture of the ensemble,the advanced feature selection is performed by the Frog-Snake Prey-Predation Relationship Optimization(FSRO)algorithm.FSRO prioritizes the most relevant features while dynamically reducing redundant and noisy data,hence improving the efficiency and accuracy of the classification model.We have compared our method for feature selection against state-of-the-art techniques and recorded an accuracy of 94.88%,a recall of 94.38%,a precision of 96.18%,and an F1-score of 95.63%.These figures are therefore better than the classical methods for deep learning.Though our dataset,collected from four different hospitals,is non-standard and heterogeneous,making the analysis more challenging,although computationally expensive,our approach proves diagnostically superior in cancer detection.Source codes and datasets are available on GitHub.
文摘Air pollution,specifically fine particulate matter(PM2.5),represents a critical environmental and public health concern due to its adverse effects on respiratory and cardiovascular systems.Accurate forecasting of PM2.5 concentrations is essential for mitigating health risks;however,the inherent nonlinearity and dynamic variability of air quality data present significant challenges.This study conducts a systematic evaluation of deep learning algorithms including Convolutional Neural Network(CNN),Long Short-Term Memory(LSTM),and the hybrid CNN-LSTM as well as statistical models,AutoRegressive Integrated Moving Average(ARIMA)and Maximum Likelihood Estimation(MLE)for hourly PM2.5 forecasting.Model performance is quantified using Root Mean Squared Error(RMSE),Mean Absolute Error(MAE),Mean Absolute Percentage Error(MAPE),and the Coefficient of Determination(R^(2))metrics.The comparative analysis identifies optimal predictive approaches for air quality modeling,emphasizing computational efficiency and accuracy.Additionally,CNN classification performance is evaluated using a confusion matrix,accuracy,precision,and F1-score.The results demonstrate that the Hybrid CNN-LSTM model outperforms standalone models,exhibiting lower error rates and higher R^(2) values,thereby highlighting the efficacy of deep learning-based hybrid architectures in achieving robust and precise PM2.5 forecasting.This study underscores the potential of advanced computational techniques in enhancing air quality prediction systems for environmental and public health applications.
基金supported by the National Natural Science Foundation of China(Grant No.62202210).
文摘The application of generative artificial intelligence(AI)is bringing about notable changes in anime creation.This paper surveys recent advancements and applications of diffusion and language models in anime generation,focusing on their demonstrated potential to enhance production efficiency through automation and personalization.Despite these benefits,it is crucial to acknowledge the substantial initial computational investments required for training and deploying these models.We conduct an in-depth survey of cutting-edge generative AI technologies,encompassing models such as Stable Diffusion and GPT,and appraise pivotal large-scale datasets alongside quantifiable evaluation metrics.Review of the surveyed literature indicates the achievement of considerable maturity in the capacity of AI models to synthesize high-quality,aesthetically compelling anime visual images from textual prompts,alongside discernible progress in the generation of coherent narratives.However,achieving perfect long-form consistency,mitigating artifacts like flickering in video sequences,and enabling fine-grained artistic control remain critical ongoing challenges.Building upon these advancements,research efforts have increasingly pivoted towards the synthesis of higher-dimensional content,such as video and three-dimensional assets,with recent studies demonstrating significant progress in this burgeoning field.Nevertheless,formidable challenges endure amidst these advancements.Foremost among these are the substantial computational exigencies requisite for training and deploying these sophisticated models,particularly pronounced in the realm of high-dimensional generation such as video synthesis.Additional persistent hurdles include maintaining spatial-temporal consistency across complex scenes and mitigating ethical considerations surrounding bias and the preservation of human creative autonomy.This research underscores the transformative potential and inherent complexities of AI-driven synergy within the creative industries.We posit that future research should be dedicated to the synergistic fusion of diffusion and autoregressive models,the integration of multimodal inputs,and the balanced consideration of ethical implications,particularly regarding bias and the preservation of human creative autonomy,thereby establishing a robust foundation for the advancement of anime creation and the broader landscape of AI-driven content generation.
文摘Wireless Sensor Networks(WSNs)have emerged as crucial tools for real-time environmental monitoring through distributed sensor nodes(SNs).However,the operational lifespan of WSNs is significantly constrained by the limited energy resources of SNs.Current energy efficiency strategies,such as clustering,multi-hop routing,and data aggregation,face challenges,including uneven energy depletion,high computational demands,and suboptimal cluster head(CH)selection.To address these limitations,this paper proposes a hybrid methodology that optimizes energy consumption(EC)while maintaining network performance.The proposed approach integrates the Low Energy Adaptive Clustering Hierarchy with Deterministic(LEACH-D)protocol using an Artificial Neural Network(ANN)and Bayesian Regularization Algorithm(BRA).LEACH-D improves upon conventional LEACH by ensuring more uniform energy usage across SNs,mitigating inefficiencies from random CH selection.The ANN further enhances CH selection and routing processes,effectively reducing data transmission overhead and idle listening.Simulation results reveal that the LEACH-D-ANN model significantly reduces EC and extends the network’s lifespan compared to existing protocols.This framework offers a promising solution to the energy efficiency challenges in WSNs,paving the way for more sustainable and reliable network deployments.
文摘Integrating mobile edge computing (MEC) and digital twin (DT) technologies to enhance network performance through predictive, adaptive control for energy-efficient and low-latency communication is a significant challenge. This paper presents the EcoEdgeTwin model, an innovative framework that harnesses the harmony between MEC and DT technologies to ensure an efficient network operation. We optimise the utility function to balance enhancing users' quality of experience (QoE) and minimising latency and energy consumption at edge servers. This approach ensures efficient and adaptable network operations, utilising DT to synchronise and integrate real-time data seamlessly. Our framework implements robust mechanisms for task offloading, service caching and cost-effective service migration. Additionally, it manages energy consumption related to task processing, communication and the influence of DT predictions, all essential for optimising latency and minimising energy usage. Through the utility model, we also prioritise QoE, fostering a user-centric approach to network management that balances network efficiency with user satisfaction. A cornerstone of our approach is integrating the advantage actor-critic algorithm, marking a pioneering use of deep reinforcement learning for dynamic network management. This strategy addresses challenges in service mobility and network variability, ensuring optimal network performance matrices. Our extensive simulations demonstrate that compared to benchmark models, the EcoEdgeTwin framework significantly reduces energy usage and latency while enhancing QoE.
基金supported by the Startup Grant(PG18929)awarded to F.Shokoohi.
文摘Accurate Electric Load Forecasting(ELF)is crucial for optimizing production capacity,improving operational efficiency,and managing energy resources effectively.Moreover,precise ELF contributes to a smaller environmental footprint by reducing the risks of disruption,downtime,and waste.However,with increasingly complex energy consumption patterns driven by renewable energy integration and changing consumer behaviors,no single approach has emerged as universally effective.In response,this research presents a hybrid modeling framework that combines the strengths of Random Forest(RF)and Autoregressive Integrated Moving Average(ARIMA)models,enhanced with advanced feature selection—Minimum Redundancy Maximum Relevancy and Maximum Synergy(MRMRMS)method—to produce a sparse model.Additionally,the residual patterns are analyzed to enhance forecast accuracy.High-resolution weather data from Weather Underground and historical energy consumption data from PJM for Duke Energy Ohio and Kentucky(DEO&K)are used in this application.This methodology,termed SP-RF-ARIMA,is evaluated against existing approaches;it demonstrates more than 40%reduction in mean absolute error and root mean square error compared to the second-best method.
基金supported by the Spanish Government and EU,under project PID2019-109842RB-I00/AEI/10.13039/501100011033。
文摘In Power Line Communications(PLC),there are regulatory masks that restrict the transmit power spectral density for electromagnetic compatibility reasons,which creates coverage issues despite the not too long distances.Hence,PLC networks often employ repeaters/relays,especially in smart grid neighborhood area networks.Even in broadband indoor PLC systems that offer a notable data rate,relaying may pave the way to new applications like being the backbone for wireless technologies in a cost-effective manner to support the Internet-of-things paradigm.In this paper,we study Multiple-Input Multiple-Output(MIMO)PLC systems that incorporate inband full-duplex functionality in relaying networks.We present several MIMO configurations that allow end-to-end half-duplex or full-duplex operations and analyze the achievable performance with state-of-the-art PLC systems.To reach this analysis,we get channel realizations from random network layouts for indoor and outdoor scenarios.We adopt realistic MIMO channel and noise models and consider transmission techniques according to PLC standards.The concepts discussed in this work can be useful in the design of future PLC relay-aided networks for different applications that look for a coverage extension and/or throughput:smart grids with enhanced communications in outdoor scenarios,and“last meter”systems for high-speed connections everywhere in indoor ones.
文摘Structured illumination,a wide-field imaging approach used in microscopy to enhance image resolution beyond the system's diffraction limits,is a well-studied technique that has gained significant traction over the last two decades.However,when translated to endoscopic systems,severe deformations of illumination patterns occur due to the large depth of field(DOF)and the 3D nature of the targets,introducing significant implementation challenges.Hence,this study explores a speckle-based system that best suits endoscopic practices to enhance image resolution by using random illumination patterns.The study presents a prototypic model of an endoscopic add-on,its design,and fabrication facilitated by using the speckle structured illumination endoscopic(SSIE)system.The imaging results of the SSIE are explained on a colon phantom model at different imaging planes with a wide field of view(FOV)and DOF.The obtained imaging metrics are elucidated and compared with state-of-the-art(SOA)high-resolution endoscopic techniques.Moreover,the potential for a clinical translation of the prototypic SSIE model is also explored in this work.The incorporation of the add-on and its subsequent results on the colon phantom model could potentially pave the way for its successful integration and use in futuristic clinical endoscopic trials.
基金funded by Ajman University,AU-Funded Research Grant 2023-IRG-ENIT-22.
文摘Cyber-Physical System (CPS) devices are increasing exponentially. Lacking confidentiality creates a vulnerable network. Thus, demanding the overall system with the latest and robust solutions for the defence mechanisms with low computation cost, increased integrity, and surveillance. The proposal of a mechanism that utilizes the features of authenticity measures using the Destination Sequence Distance Vector (DSDV) routing protocol which applies to the multi-WSN (Wireless Sensor Network) of IoT devices in CPS which is developed for the Device-to-Device (D2D) authentication developed from the local-chain and public chain respectively combined with the Software Defined Networking (SDN) control and monitoring system using switches and controllers that will route the packets through the network, identify any false nodes, take preventive measures against them and preventing them for any future problems. Next, the system is powered by Blockchain cryptographic features by utilizing the TrustChain features to create a private, secure, and temper-free ledger of the transactions performed inside the network. Results are achieved in the legitimate devices connecting to the network, transferring their packets to their destination under supervision, reporting whenever a false node is causing hurdles, and recording the transactions for temper-proof records. Evaluation results based on 1000+ transactions illustrate that the proposed mechanism not only outshines most aspects of Cyber-Physical systems but also consumes less computation power with a low latency of 0.1 seconds only.