In order to improve the performance of peer-to-peer files sharing system under mobile distributed en- vironments, a novel always-optimally-coordinated (AOC) criterion and corresponding candidate selection algorithm ...In order to improve the performance of peer-to-peer files sharing system under mobile distributed en- vironments, a novel always-optimally-coordinated (AOC) criterion and corresponding candidate selection algorithm are proposed in this paper. Compared with the traditional min-hops criterion, the new approach introduces a fuzzy knowledge combination theory to investigate several important factors that influence files transfer success rate and efficiency. Whereas the min-hops based protocols only ask the nearest candidate peer for desired files, the selection algorithm based on AOC comprehensively considers users' preferences and network requirements with flexible balancing rules. Furthermore, its advantage also expresses in the independence of specified resource discovering protocols, allowing for scalability. The simulation results show that when using the AOC based peer selection algorithm, system performance is much better than the rain-hops scheme, with files successful transfer rate improved more than 50% and transfer time re- duced at least 20%.展开更多
Byte-addressable non-volatile memory(NVM),as a new participant in the storage hierarchy,gives extremely high performance in storage,which forces changes to be made on current filesystem designs.Page cache,once a signi...Byte-addressable non-volatile memory(NVM),as a new participant in the storage hierarchy,gives extremely high performance in storage,which forces changes to be made on current filesystem designs.Page cache,once a significant mechanism filling the performance gap between Dynamic Random Access Memory(DRAM)and block devices,is now a liability that heavily hinders the writing performance of NVM filesystems.Therefore state-of-the-art NVM filesystems leverage the direct access(DAX)technology to bypass the page cache entirely.However,the DRAM still provides higher bandwidth than NVM,which prevents skewed read workloads from benefiting from a higher bandwidth of the DRAM and leads to sub-optimal performance for the system.In this paper,we propose RCache,a readintensive workload-aware page cache for NVM filesystems.Different from traditional caching mechanisms where all reads go through DRAM,RCache uses a tiered page cache design,including assigning DRAM and NVM to hot and cold data separately,and reading data from both sides.To avoid copying data to DRAM in a critical path,RCache migrates data from NVM to DRAM in a background thread.Additionally,RCache manages data in DRAM in a lock-free manner for better latency and scalability.Evaluations on Intel Optane Data Center(DC)Persistent Memory Modules show that,compared with NOVA,RCache achieves 3 times higher bandwidth for read-intensive workloads and introduces little performance loss for write operations.展开更多
In this paper, we present a distributed multi-level cache system based on cloud storage, which is aimed at the low access efficiency of small spatio-temporal data files in information service system of Smart City. Tak...In this paper, we present a distributed multi-level cache system based on cloud storage, which is aimed at the low access efficiency of small spatio-temporal data files in information service system of Smart City. Taking classification attribute of small spatio-temporal data files in Smart City as the basis of cache content selection, the cache system adopts different cache pool management strategies in different levels of cache. The results of experiment in prototype system indicate that multi-level cache in this paper effectively increases the access bandwidth of small spatio-temporal files in Smart City and greatly improves service quality of multiple concurrent access in system.展开更多
IIn order to improve the performance of wireless distributed peer-to-peer(P2P)files sharing systems,a general system architecture and a novel peer selecting model based on fuzzy cognitive maps(FCM)are proposed in this...IIn order to improve the performance of wireless distributed peer-to-peer(P2P)files sharing systems,a general system architecture and a novel peer selecting model based on fuzzy cognitive maps(FCM)are proposed in this paper.The new model provides an effective approach on choosing an optimal peer from several resource discovering results for the best file transfer.Compared with the traditional min-hops scheme that uses hops as the only selecting criterion,the proposed model uses FCM to investigate the complex relationships among various relative factors in wireless environments and gives an overall evaluation score on the candidate.It also has strong scalability for being independent of specified P2P resource discovering protocols.Furthermore,a complete implementation is explained in concrete modules.The simulation results show that the proposed model is effective and feasible compared with min-hops scheme,with the success transfer rate increased by at least 20% and transfer time improved as high as 34%.展开更多
A casting process CAD is put forward to design and draw casting process. The 2D casting process CAD, most of the current systems are developed based on one certain version of the AutoCAD system. However the applicatio...A casting process CAD is put forward to design and draw casting process. The 2D casting process CAD, most of the current systems are developed based on one certain version of the AutoCAD system. However the application of these 2D casting process CAD systems in foundry enterprises are restricted because they have several deficiencies, such as being overly dependent on the AutoCAD system, and some part files based on PDF format can not be opened directly. To overcome these deficiencies, for the first time an innovative 2D casting process CAD system based on PDF and image format file has been proposed, which breaks through the traditional research and application notion of the 2D casting process CAD system based on AutoCAD. Several key technologies of this system such as coordinate transformation, CAD interactive drawing, file storage, PDF and image format files display, and image recognition technologies were described in detail. A practical 2D CAD casting process system named HZCAD2D(PDF) was developed, which is capable of designing and drawing the casting process on the part drawing based on the PDF format directly, without spending time on drawing the part produced by AutoCAD system. Final y, taking two actual castings as examples, the casting processes were drawn using this system, demonstrating that this system can significantly shorten the cycle of casting process designing.展开更多
The size of the Audio and Video(AV)content of the 8K program is four times larger than that of 4K content,providing viewers with a more ideal audiovisual experience while placing higher demands on the capability and e...The size of the Audio and Video(AV)content of the 8K program is four times larger than that of 4K content,providing viewers with a more ideal audiovisual experience while placing higher demands on the capability and efficiency of document preparation and processing,signal transmission,and scheduling.However,it is difficult to meet the high robustness requirements of 8K broadcast services because the existing broadcast system architecture is limited by efficiency,cost,and other factors.In this study,an 8K Ultra-High-Definition(UHD)TV program broadcast scheme was designed.The verification results show that the scheme is high quality,highly efficient,and robust.In particular,in the research,the file format normalizing module was first placed in the broadcast area instead of the file preparation area,and the low-compression transmission scheme of the all-IP signal JPEG XS was designed in the signal transmission area for improving the efficiency of the scheme.Next,to reduce the impact on the robustness of broadcast services,the broadcast control logic of the broadcast core components is optimized.Finally,a series of 8K TV program broadcasting systems have been implemented and their performance has been verified.The results show that the system meets the efficiency and robustness requirements of a high-quality 8K AV broadcast system,and thus has a high degree of practicability.展开更多
With the increase in the quantity and scale of Static Random-Access Memory Field Programmable Gate Arrays (SRAM-based FPGAs) for aerospace application, the volume of FPGA configuration bit files that must be stored ha...With the increase in the quantity and scale of Static Random-Access Memory Field Programmable Gate Arrays (SRAM-based FPGAs) for aerospace application, the volume of FPGA configuration bit files that must be stored has increased dramatically. The use of compression techniques for these bitstream files is emerging as a key strategy to alleviate the burden on storage resources. Due to the severe resource constraints of space-based electronics and the unique application environment, the simplicity, efficiency and robustness of the decompression circuitry is also a key design consideration. Through comparative analysis current bitstream file compression technologies, this research suggests that the Lempel Ziv Oberhumer (LZO) compression algorithm is more suitable for satellite applications. This paper also delves into the compression process and format of the LZO compression algorithm, as well as the inherent characteristics of configuration bitstream files. We propose an improved algorithm based on LZO for bitstream file compression, which optimises the compression process by refining the format and reducing the offset. Furthermore, a low-cost, robust decompression hardware architecture is proposed based on this method. Experimental results show that the compression speed of the improved LZO algorithm is increased by 3%, the decompression hardware cost is reduced by approximately 60%, and the compression ratio is slightly reduced by 0.47%.展开更多
The multi-stage development strategy is often adopted in the gas field.However,when the productivity decline occurs,many large processing stations will be severely idle and underutilized,significantly reducing operati...The multi-stage development strategy is often adopted in the gas field.However,when the productivity decline occurs,many large processing stations will be severely idle and underutilized,significantly reducing operating efficiency and revenue.This study proposes a novel operation mode of multiple gathering production systems for gas field multi-stage development,integrating the decisions about processing capacity allocation and infrastructure construction to share processing stations and improve multi-system operating efficiency.A multi-period mixed integer linear programming model for multisystem operation optimization is established to optimize the net present value(NPV),considering the production of gas wells,time-varying gas prices,and the capacity of processing stations.The decision of processing capacity,location,construction timing,and capacity expansion of processing stations,as well as transmission capacity of pipelines and processing capacity allocation schemes,can be obtained to meet long-term production demand.Furthermore,a real case study indicates that the proposed processing capacity allocation approach not only has a shorter payback period and increases NPV by 4.8%,but also increases the utilization efficiency of processing stations from 27.37% to 48.94%.This work demonstrates that the synergy between the processing capacity allocation and infrastructure construction can hedge against production fluctuations and increase potential profits.展开更多
[Objective]In response to the issue of insufficient integrity in hourly routine meteorological element data files,this paper aims to improve the availability and reliability of data files,and provide high-quality data...[Objective]In response to the issue of insufficient integrity in hourly routine meteorological element data files,this paper aims to improve the availability and reliability of data files,and provide high-quality data file support for meteorological forecasting and services.[Method]In this paper,an efficient and accurate method for data file quality control and fusion processing is developed.By locating the missing measurement time,data are extracted from the"AWZ.db"database and the minute routine meteorological element data file,and merged into the hourly routine meteorological element data file.[Result]Data processing efficiency and accuracy are significantly improved,and the problem of incomplete hourly routine meteorological element data files is solved.At the same time,it emphasizes the importance of ensuring the accuracy of the files used and carefully checking and verifying the fusion results,and proposes strategies to improve data quality.[Conclusion]This method provides convenience for observation personnel and effectively improves the integrity and accuracy of data files.In the future,it is expected to provide more reliable data support for meteorological forecasting and services.展开更多
The healthcare sector involves many steps to ensure efficient care for patients,such as appointment scheduling,consultation plans,online follow-up,and more.However,existing healthcare mechanisms are unable to facilita...The healthcare sector involves many steps to ensure efficient care for patients,such as appointment scheduling,consultation plans,online follow-up,and more.However,existing healthcare mechanisms are unable to facilitate a large number of patients,as these systems are centralized and hence vulnerable to various issues,including single points of failure,performance bottlenecks,and substantial monetary costs.Furthermore,these mechanisms are unable to provide an efficient mechanism for saving data against unauthorized access.To address these issues,this study proposes a blockchain-based authentication mechanism that authenticates all healthcare stakeholders based on their credentials.Furthermore,also utilize the capabilities of the InterPlanetary File System(IPFS)to store the Electronic Health Record(EHR)in a distributed way.This IPFS platform addresses not only the issue of high data storage costs on blockchain but also the issue of a single point of failure in the traditional centralized data storage model.The simulation results demonstrate that our model outperforms the benchmark schemes and provides an efficient mechanism for managing healthcare sector operations.The results show that it takes approximately 3.5 s for the smart contract to authenticate the node and provide it with the decryption key,which is ultimately used to access the data.The simulation results show that our proposed model outperforms existing solutions in terms of execution time and scalability.The execution time of our model smart contract is around 9000 transactions in just 6.5 s,while benchmark schemes require approximately 7 s for the same number of transactions.展开更多
We now differentiate between the requirements for new and revised submissions.You may choose to submit your manuscript as a single Word or PDF file to be used in the refereeing process.Only when your paper is at the r...We now differentiate between the requirements for new and revised submissions.You may choose to submit your manuscript as a single Word or PDF file to be used in the refereeing process.Only when your paper is at the revision stage,will you be requested to put your paper into a'correct format'for acceptance and provide the items required for the publication of your article.展开更多
In the era of Big Data, typical architecture of distributed real-time stream processing systems is the combination of Flume, Kafka, and Storm. As a kind of distributed message system, Kafka has the characteristics of ...In the era of Big Data, typical architecture of distributed real-time stream processing systems is the combination of Flume, Kafka, and Storm. As a kind of distributed message system, Kafka has the characteristics of horizontal scalability and high throughput, which is manly deployed in many areas in order to address the problem of speed mismatch between message producers and consumers. When using Kafka, we need to quickly receive data sent by producers. In addition, we need to send data to consumers quickly. Therefore, the performance of Kafka is of critical importance to the performance of the whole stream processing system. In this paper, we propose the improved design of real-time stream processing systems, and focus on improving the Kafka's data loading process.We use Kafka cat to transfer data from the source to Kafka topic directly, which can reduce the network transmission. We also utilize the memory file system to accelerate the process of data loading, which can address the bottleneck and performance problems caused by disk I/O. Extensive experiments are conducted to evaluate the performance, which show the superiority of our improved design.展开更多
ABSTRACT: This paper generalizes the makeup and forming dynamic mechanism of natural disaster systems, principles and methods of comprehensive division of natural disasters, as well as structure, function and up-build...ABSTRACT: This paper generalizes the makeup and forming dynamic mechanism of natural disaster systems, principles and methods of comprehensive division of natural disasters, as well as structure, function and up-build routes of map and file information visualization system (MFIVS). Taking the Changjiang(Yangtze) Valley as an example, on the basis of revealing up the integrated mechanism on the formations of its natural disasters and its distributing law, thereafter, the paper relies on the MFIVS technique, adopts two top-down and bottom-up approaches to study a comprehensive division of natural disasters. It is relatively objective and precise that the required division results include three natural disaster sections and nine natural disaster sub-sections, which can not only provide a scientific basis for utilizing natural resources and controlling natural disaster and environmental degradation, but also be illuminated to a concise, practical and effective technique on comprehensive division.展开更多
In consultative committee for space data systems(CCSDS)file delivery protocol(CFDP)recommendation of reliable transmission,there are no detail transmission procedure and delay calculation of prompted negative acknowle...In consultative committee for space data systems(CCSDS)file delivery protocol(CFDP)recommendation of reliable transmission,there are no detail transmission procedure and delay calculation of prompted negative acknowledge and asynchronous negative acknowledge models.CFDP is designed to provide data and storage management,story and forward,custody transfer and reliable end-to-end delivery over deep space characterized by huge latency,intermittent link,asymmetric bandwidth and big bit error rate(BER).Four reliable transmission models are analyzed and an expected file-delivery time is calculated with different trans-mission rates,numbers and sizes of packet data units,BERs and frequencies of external events,etc.By comparison of four CFDP models,the requirement of BER for typical missions in deep space is obtained and rules of choosing CFDP models under different uplink state informations are given,which provides references for protocol models selection,utilization and modification.展开更多
In distributed storage systems,file access efficiency has an important impact on the real-time nature of information forensics.As a popular approach to improve file accessing efficiency,prefetching model can fetches d...In distributed storage systems,file access efficiency has an important impact on the real-time nature of information forensics.As a popular approach to improve file accessing efficiency,prefetching model can fetches data before it is needed according to the file access pattern,which can reduce the I/O waiting time and increase the system concurrency.However,prefetching model needs to mine the degree of association between files to ensure the accuracy of prefetching.In the massive small file situation,the sheer volume of files poses a challenge to the efficiency and accuracy of relevance mining.In this paper,we propose a massive files prefetching model based on LSTM neural network with cache transaction strategy to improve file access efficiency.Firstly,we propose a file clustering algorithm based on temporal locality and spatial locality to reduce the computational complexity.Secondly,we propose a definition of cache transaction according to files occurrence in cache instead of time-offset distance based methods to extract file block feature accurately.Lastly,we innovatively propose a file access prediction algorithm based on LSTM neural network which predict the file that have high possibility to be accessed.Experiments show that compared with the traditional LRU and the plain grouping methods,the proposed model notably increase the cache hit rate and effectively reduces the I/O wait time.展开更多
To overcome the traditional disadvantages, an improved FPGA-core based real-time video image acqui- sition and storage system is designed. The modular designed by Verilog programming is used for video decoding of A/D ...To overcome the traditional disadvantages, an improved FPGA-core based real-time video image acqui- sition and storage system is designed. The modular designed by Verilog programming is used for video decoding of A/D configuration, video image capturing logic control and image storage logic control modules. And IDE in- terface bard disk as storage medium and FAT32 file system as record form are used for real-time image storage. Experimental results show that the system has the advantages of strong real-time capability, high integration, powerful storage, easy expansibility and so on.展开更多
基金supported by the National Nature Science Foundation of China(No.60672124)the National High Technology Research and Development Programme the of China(No.2007AA01Z221)
文摘In order to improve the performance of peer-to-peer files sharing system under mobile distributed en- vironments, a novel always-optimally-coordinated (AOC) criterion and corresponding candidate selection algorithm are proposed in this paper. Compared with the traditional min-hops criterion, the new approach introduces a fuzzy knowledge combination theory to investigate several important factors that influence files transfer success rate and efficiency. Whereas the min-hops based protocols only ask the nearest candidate peer for desired files, the selection algorithm based on AOC comprehensively considers users' preferences and network requirements with flexible balancing rules. Furthermore, its advantage also expresses in the independence of specified resource discovering protocols, allowing for scalability. The simulation results show that when using the AOC based peer selection algorithm, system performance is much better than the rain-hops scheme, with files successful transfer rate improved more than 50% and transfer time re- duced at least 20%.
基金supported by ZTE Industry⁃University⁃Institute Coopera⁃tion Funds under Grant No.HC⁃CN⁃20181128026.
文摘Byte-addressable non-volatile memory(NVM),as a new participant in the storage hierarchy,gives extremely high performance in storage,which forces changes to be made on current filesystem designs.Page cache,once a significant mechanism filling the performance gap between Dynamic Random Access Memory(DRAM)and block devices,is now a liability that heavily hinders the writing performance of NVM filesystems.Therefore state-of-the-art NVM filesystems leverage the direct access(DAX)technology to bypass the page cache entirely.However,the DRAM still provides higher bandwidth than NVM,which prevents skewed read workloads from benefiting from a higher bandwidth of the DRAM and leads to sub-optimal performance for the system.In this paper,we propose RCache,a readintensive workload-aware page cache for NVM filesystems.Different from traditional caching mechanisms where all reads go through DRAM,RCache uses a tiered page cache design,including assigning DRAM and NVM to hot and cold data separately,and reading data from both sides.To avoid copying data to DRAM in a critical path,RCache migrates data from NVM to DRAM in a background thread.Additionally,RCache manages data in DRAM in a lock-free manner for better latency and scalability.Evaluations on Intel Optane Data Center(DC)Persistent Memory Modules show that,compared with NOVA,RCache achieves 3 times higher bandwidth for read-intensive workloads and introduces little performance loss for write operations.
基金Supported by the Natural Science Foundation of Hubei Province(2012FFC034,2014CFC1100)
文摘In this paper, we present a distributed multi-level cache system based on cloud storage, which is aimed at the low access efficiency of small spatio-temporal data files in information service system of Smart City. Taking classification attribute of small spatio-temporal data files in Smart City as the basis of cache content selection, the cache system adopts different cache pool management strategies in different levels of cache. The results of experiment in prototype system indicate that multi-level cache in this paper effectively increases the access bandwidth of small spatio-temporal files in Smart City and greatly improves service quality of multiple concurrent access in system.
基金Sponsored by the National Natural Science Foundation of China(Grant No.60672124 and 60832009)Hi-Tech Research and Development Program(National 863 Program)(Grant No.2007AA01Z221)
文摘IIn order to improve the performance of wireless distributed peer-to-peer(P2P)files sharing systems,a general system architecture and a novel peer selecting model based on fuzzy cognitive maps(FCM)are proposed in this paper.The new model provides an effective approach on choosing an optimal peer from several resource discovering results for the best file transfer.Compared with the traditional min-hops scheme that uses hops as the only selecting criterion,the proposed model uses FCM to investigate the complex relationships among various relative factors in wireless environments and gives an overall evaluation score on the candidate.It also has strong scalability for being independent of specified P2P resource discovering protocols.Furthermore,a complete implementation is explained in concrete modules.The simulation results show that the proposed model is effective and feasible compared with min-hops scheme,with the success transfer rate increased by at least 20% and transfer time improved as high as 34%.
基金financially supported by the Program for New Century Excellent Talents in University (No.NCET-09-0396)the National Science&Technology Key Projects of Numerical Control (No.2012ZX04012-011)the Fundamental Research Funds for the Central Universities (2014-IV-016)
文摘A casting process CAD is put forward to design and draw casting process. The 2D casting process CAD, most of the current systems are developed based on one certain version of the AutoCAD system. However the application of these 2D casting process CAD systems in foundry enterprises are restricted because they have several deficiencies, such as being overly dependent on the AutoCAD system, and some part files based on PDF format can not be opened directly. To overcome these deficiencies, for the first time an innovative 2D casting process CAD system based on PDF and image format file has been proposed, which breaks through the traditional research and application notion of the 2D casting process CAD system based on AutoCAD. Several key technologies of this system such as coordinate transformation, CAD interactive drawing, file storage, PDF and image format files display, and image recognition technologies were described in detail. A practical 2D CAD casting process system named HZCAD2D(PDF) was developed, which is capable of designing and drawing the casting process on the part drawing based on the PDF format directly, without spending time on drawing the part produced by AutoCAD system. Final y, taking two actual castings as examples, the casting processes were drawn using this system, demonstrating that this system can significantly shorten the cycle of casting process designing.
文摘The size of the Audio and Video(AV)content of the 8K program is four times larger than that of 4K content,providing viewers with a more ideal audiovisual experience while placing higher demands on the capability and efficiency of document preparation and processing,signal transmission,and scheduling.However,it is difficult to meet the high robustness requirements of 8K broadcast services because the existing broadcast system architecture is limited by efficiency,cost,and other factors.In this study,an 8K Ultra-High-Definition(UHD)TV program broadcast scheme was designed.The verification results show that the scheme is high quality,highly efficient,and robust.In particular,in the research,the file format normalizing module was first placed in the broadcast area instead of the file preparation area,and the low-compression transmission scheme of the all-IP signal JPEG XS was designed in the signal transmission area for improving the efficiency of the scheme.Next,to reduce the impact on the robustness of broadcast services,the broadcast control logic of the broadcast core components is optimized.Finally,a series of 8K TV program broadcasting systems have been implemented and their performance has been verified.The results show that the system meets the efficiency and robustness requirements of a high-quality 8K AV broadcast system,and thus has a high degree of practicability.
基金supported in part by the National Key Laboratory of Science and Technology on Space Microwave(Grant Nos.HTKJ2022KL504009 and HTKJ2022KL5040010).
文摘With the increase in the quantity and scale of Static Random-Access Memory Field Programmable Gate Arrays (SRAM-based FPGAs) for aerospace application, the volume of FPGA configuration bit files that must be stored has increased dramatically. The use of compression techniques for these bitstream files is emerging as a key strategy to alleviate the burden on storage resources. Due to the severe resource constraints of space-based electronics and the unique application environment, the simplicity, efficiency and robustness of the decompression circuitry is also a key design consideration. Through comparative analysis current bitstream file compression technologies, this research suggests that the Lempel Ziv Oberhumer (LZO) compression algorithm is more suitable for satellite applications. This paper also delves into the compression process and format of the LZO compression algorithm, as well as the inherent characteristics of configuration bitstream files. We propose an improved algorithm based on LZO for bitstream file compression, which optimises the compression process by refining the format and reducing the offset. Furthermore, a low-cost, robust decompression hardware architecture is proposed based on this method. Experimental results show that the compression speed of the improved LZO algorithm is increased by 3%, the decompression hardware cost is reduced by approximately 60%, and the compression ratio is slightly reduced by 0.47%.
基金supported by Zhejiang Provincial Natural Science Foundation of China under Grant No.LQ23E040004。
文摘The multi-stage development strategy is often adopted in the gas field.However,when the productivity decline occurs,many large processing stations will be severely idle and underutilized,significantly reducing operating efficiency and revenue.This study proposes a novel operation mode of multiple gathering production systems for gas field multi-stage development,integrating the decisions about processing capacity allocation and infrastructure construction to share processing stations and improve multi-system operating efficiency.A multi-period mixed integer linear programming model for multisystem operation optimization is established to optimize the net present value(NPV),considering the production of gas wells,time-varying gas prices,and the capacity of processing stations.The decision of processing capacity,location,construction timing,and capacity expansion of processing stations,as well as transmission capacity of pipelines and processing capacity allocation schemes,can be obtained to meet long-term production demand.Furthermore,a real case study indicates that the proposed processing capacity allocation approach not only has a shorter payback period and increases NPV by 4.8%,but also increases the utilization efficiency of processing stations from 27.37% to 48.94%.This work demonstrates that the synergy between the processing capacity allocation and infrastructure construction can hedge against production fluctuations and increase potential profits.
基金the Fifth Batch of Innovation Teams of Wuzhou Meteorological Bureau"Wuzhou Innovation Team for Enhancing the Comprehensive Meteorological Observation Ability through Digitization and Intelligence"Wuzhou Science and Technology Planning Project(202402122,202402119).
文摘[Objective]In response to the issue of insufficient integrity in hourly routine meteorological element data files,this paper aims to improve the availability and reliability of data files,and provide high-quality data file support for meteorological forecasting and services.[Method]In this paper,an efficient and accurate method for data file quality control and fusion processing is developed.By locating the missing measurement time,data are extracted from the"AWZ.db"database and the minute routine meteorological element data file,and merged into the hourly routine meteorological element data file.[Result]Data processing efficiency and accuracy are significantly improved,and the problem of incomplete hourly routine meteorological element data files is solved.At the same time,it emphasizes the importance of ensuring the accuracy of the files used and carefully checking and verifying the fusion results,and proposes strategies to improve data quality.[Conclusion]This method provides convenience for observation personnel and effectively improves the integrity and accuracy of data files.In the future,it is expected to provide more reliable data support for meteorological forecasting and services.
基金supported by the Ongoing Research Funding program(ORF-2025-636),King Saud University,Riyadh,Saudi Arabia.
文摘The healthcare sector involves many steps to ensure efficient care for patients,such as appointment scheduling,consultation plans,online follow-up,and more.However,existing healthcare mechanisms are unable to facilitate a large number of patients,as these systems are centralized and hence vulnerable to various issues,including single points of failure,performance bottlenecks,and substantial monetary costs.Furthermore,these mechanisms are unable to provide an efficient mechanism for saving data against unauthorized access.To address these issues,this study proposes a blockchain-based authentication mechanism that authenticates all healthcare stakeholders based on their credentials.Furthermore,also utilize the capabilities of the InterPlanetary File System(IPFS)to store the Electronic Health Record(EHR)in a distributed way.This IPFS platform addresses not only the issue of high data storage costs on blockchain but also the issue of a single point of failure in the traditional centralized data storage model.The simulation results demonstrate that our model outperforms the benchmark schemes and provides an efficient mechanism for managing healthcare sector operations.The results show that it takes approximately 3.5 s for the smart contract to authenticate the node and provide it with the decryption key,which is ultimately used to access the data.The simulation results show that our proposed model outperforms existing solutions in terms of execution time and scalability.The execution time of our model smart contract is around 9000 transactions in just 6.5 s,while benchmark schemes require approximately 7 s for the same number of transactions.
文摘We now differentiate between the requirements for new and revised submissions.You may choose to submit your manuscript as a single Word or PDF file to be used in the refereeing process.Only when your paper is at the revision stage,will you be requested to put your paper into a'correct format'for acceptance and provide the items required for the publication of your article.
基金supported by the Research Fund of National Key Laboratory of Computer Architecture under Grant No.CARCH201501the Open Project Program of the State Key Laboratory of Mathematical Engineering and Advanced Computing under Grant No.2016A09
文摘In the era of Big Data, typical architecture of distributed real-time stream processing systems is the combination of Flume, Kafka, and Storm. As a kind of distributed message system, Kafka has the characteristics of horizontal scalability and high throughput, which is manly deployed in many areas in order to address the problem of speed mismatch between message producers and consumers. When using Kafka, we need to quickly receive data sent by producers. In addition, we need to send data to consumers quickly. Therefore, the performance of Kafka is of critical importance to the performance of the whole stream processing system. In this paper, we propose the improved design of real-time stream processing systems, and focus on improving the Kafka's data loading process.We use Kafka cat to transfer data from the source to Kafka topic directly, which can reduce the network transmission. We also utilize the memory file system to accelerate the process of data loading, which can address the bottleneck and performance problems caused by disk I/O. Extensive experiments are conducted to evaluate the performance, which show the superiority of our improved design.
基金Under the auspices of President Foundation of the Chinese Academy of Sciences(1999).
文摘ABSTRACT: This paper generalizes the makeup and forming dynamic mechanism of natural disaster systems, principles and methods of comprehensive division of natural disasters, as well as structure, function and up-build routes of map and file information visualization system (MFIVS). Taking the Changjiang(Yangtze) Valley as an example, on the basis of revealing up the integrated mechanism on the formations of its natural disasters and its distributing law, thereafter, the paper relies on the MFIVS technique, adopts two top-down and bottom-up approaches to study a comprehensive division of natural disasters. It is relatively objective and precise that the required division results include three natural disaster sections and nine natural disaster sub-sections, which can not only provide a scientific basis for utilizing natural resources and controlling natural disaster and environmental degradation, but also be illuminated to a concise, practical and effective technique on comprehensive division.
基金supported by the National Natural Science Fandation of China (60672089,60772075)
文摘In consultative committee for space data systems(CCSDS)file delivery protocol(CFDP)recommendation of reliable transmission,there are no detail transmission procedure and delay calculation of prompted negative acknowledge and asynchronous negative acknowledge models.CFDP is designed to provide data and storage management,story and forward,custody transfer and reliable end-to-end delivery over deep space characterized by huge latency,intermittent link,asymmetric bandwidth and big bit error rate(BER).Four reliable transmission models are analyzed and an expected file-delivery time is calculated with different trans-mission rates,numbers and sizes of packet data units,BERs and frequencies of external events,etc.By comparison of four CFDP models,the requirement of BER for typical missions in deep space is obtained and rules of choosing CFDP models under different uplink state informations are given,which provides references for protocol models selection,utilization and modification.
基金This work is supported by‘The Fundamental Research Funds for the Central Universities(Grant No.HIT.NSRIF.201714)’‘Weihai Science and Technology Development Program(2016DXGJMS15)’‘Key Research and Development Program in Shandong Provincial(2017GGX90103)’.
文摘In distributed storage systems,file access efficiency has an important impact on the real-time nature of information forensics.As a popular approach to improve file accessing efficiency,prefetching model can fetches data before it is needed according to the file access pattern,which can reduce the I/O waiting time and increase the system concurrency.However,prefetching model needs to mine the degree of association between files to ensure the accuracy of prefetching.In the massive small file situation,the sheer volume of files poses a challenge to the efficiency and accuracy of relevance mining.In this paper,we propose a massive files prefetching model based on LSTM neural network with cache transaction strategy to improve file access efficiency.Firstly,we propose a file clustering algorithm based on temporal locality and spatial locality to reduce the computational complexity.Secondly,we propose a definition of cache transaction according to files occurrence in cache instead of time-offset distance based methods to extract file block feature accurately.Lastly,we innovatively propose a file access prediction algorithm based on LSTM neural network which predict the file that have high possibility to be accessed.Experiments show that compared with the traditional LRU and the plain grouping methods,the proposed model notably increase the cache hit rate and effectively reduces the I/O wait time.
基金Supported by the Research Fund of Shaanxi University of Technology(SLG0619)~~
文摘To overcome the traditional disadvantages, an improved FPGA-core based real-time video image acqui- sition and storage system is designed. The modular designed by Verilog programming is used for video decoding of A/D configuration, video image capturing logic control and image storage logic control modules. And IDE in- terface bard disk as storage medium and FAT32 file system as record form are used for real-time image storage. Experimental results show that the system has the advantages of strong real-time capability, high integration, powerful storage, easy expansibility and so on.