期刊文献+
共找到3篇文章
< 1 >
每页显示 20 50 100
Research on dual-command operation path optimization based on Flying-V warehouse layout
1
作者 刘建胜 YUAN Bin +1 位作者 YANG Zan ZHONG RAY Y 《High Technology Letters》 EI CAS 2023年第4期388-396,共9页
To enhance the efficiency of warehouse order management,this study investigates a dual-com-mand operation mode in the Flying-V non-traditional warehouse layout.Three dual-command opera-tion strategies are designed,and... To enhance the efficiency of warehouse order management,this study investigates a dual-com-mand operation mode in the Flying-V non-traditional warehouse layout.Three dual-command opera-tion strategies are designed,and a dual-command operation path optimization model is established with the shortest path as the optimization goal.Furthermore,a genetic algorithm based on a dynamic decoding strategy is proposed.Simulation results demonstrate that the Flying-V layout warehouse management and access cooperation operation can reduce the operation time by an average of 25%-35%compared with the single access operation path,and by an average of 13%-23%compared with the‘deposit first and then pick’operation path.These findings provide evidence for the effec-tiveness of the optimization model and algorithm. 展开更多
关键词 Flying-V access collaboration path optimization dynamic decoding genetic al-gorithm
在线阅读 下载PDF
Dynamic Batch Processing with FlexiDecode Scheduler for Efficient LLM Inference in IIoT
2
作者 Xiaocong Jia Bruce Gu +5 位作者 Jinjun Chen Longxiang Gao Weiguang Pang Guangtong Lv Youyang Qu Lei Cui 《Big Data Mining and Analytics》 2025年第6期1307-1323,共17页
Large Language Models(LLMs)are expanding their applications across various fields,including Industrial Internet of Things(IIoT),where they analyze sensor data,automate diagnostics,and enhance predictive maintenance.LL... Large Language Models(LLMs)are expanding their applications across various fields,including Industrial Internet of Things(IIoT),where they analyze sensor data,automate diagnostics,and enhance predictive maintenance.LLM inference is provided by service providers to users,with each inference request undergoing two phases:prefill and decode.Due to the autoregressive nature of generation,only one token can be produced per iteration,necessitating multiple iterations to complete a request.Typically,batch processing groups multiple requests into a single batch for inference,improving throughput and hardware utilization.However,in service systems,a fixed batch size presents challenges under fluctuating request volumes,particularly in IIoT environments,where data flow can vary significantly.Specifically,during the high-load periods,a fixed batch size may lead to underutilization of resources,while during the low-load periods,it may result in resource wastage.In this paper,we introduce FlexiDecode Scheduler(FDS)to address these challenges by dynamically adjusting the decoding batch size based on system load conditions,improving resource utilization,and reducing wait time during high-load periods.FDS prioritizes prefilling new requests to maximize decoding efficiency and employs a request output length predictor to optimize request scheduling,minimizing End-to-End(E2E)latency.Compared to virtual Large Language Model(vLLM)and Sarathi,our approach achieves a 23%and 16%reduction in E2E latency,improves actual request execution time by 34%and 15%,respectively,and increases computational utilization by 10%. 展开更多
关键词 virtual Large Language Model(vLLM)inference batch scheduling dynamic decoding batches calculating utilization
原文传递
Efficient Array Coding Scheme for Large-Scale Storage Systems
3
作者 Dan Tang 《Journal of Electronic Science and Technology》 CAS CSCD 2015年第2期102-106,共5页
A family of array codes with a maximum distance separable(MDS) property, named L codes, is proposed. The greatest strength of L codes is that the number of rows(columns) in a disk array does not be restricted by t... A family of array codes with a maximum distance separable(MDS) property, named L codes, is proposed. The greatest strength of L codes is that the number of rows(columns) in a disk array does not be restricted by the prime number, and more disks can be dynamically appended in a running storage system. L codes can tolerate at least two disk erasures and some sector loss simultaneously, and can tolerate multiple disk erasures(greater than or equal to three) under a certain condition. Because only XOR operations are needed in the process of encoding and decoding, L codes have very high computing efficiency which is roughly equivalent to X codes. Analysis shows that L codes are particularly suitable for large-scale storage systems. 展开更多
关键词 codes decoding dynamically parity restricted sector encoding operations running correcting
在线阅读 下载PDF
上一页 1 下一页 到第
使用帮助 返回顶部