【Objective】Medical imaging data has great value,but it contains a significant amount of sensitive information about patients.At present,laws and regulations regarding to the de-identification of medical imaging data...【Objective】Medical imaging data has great value,but it contains a significant amount of sensitive information about patients.At present,laws and regulations regarding to the de-identification of medical imaging data are not clearly defined around the world.This study aims to develop a tool that meets compliance-driven desensitization requirements tailored to diverse research needs.【Methods】To enhance the security of medical image data,we designed and implemented a DICOM format medical image de-identification system on the Windows operating system.【Results】Our custom de-identification system is adaptable to the legal standards of different countries and can accommodate specific research demands.The system offers both web-based online and desktop offline de-identification capabilities,enabling customization of de-identification rules and facilitating batch processing to improve efficiency.【Conclusions】This medical image de-identification system robustly strengthens the stewardship of sensitive medical data,aligning with data security protection requirements while facilitating the sharing and utilization of medical image data.This approach unlocks the intrinsic value inherent in such datasets.展开更多
The integration of the Internet of Things(IoT)into healthcare systems improves patient care,boosts operational efficiency,and contributes to cost-effective healthcare delivery.However,overcoming several associated cha...The integration of the Internet of Things(IoT)into healthcare systems improves patient care,boosts operational efficiency,and contributes to cost-effective healthcare delivery.However,overcoming several associated challenges,such as data security,interoperability,and ethical concerns,is crucial to realizing the full potential of IoT in healthcare.Real-time anomaly detection plays a key role in protecting patient data and maintaining device integrity amidst the additional security risks posed by interconnected systems.In this context,this paper presents a novelmethod for healthcare data privacy analysis.The technique is based on the identification of anomalies in cloud-based Internet of Things(IoT)networks,and it is optimized using explainable artificial intelligence.For anomaly detection,the Radial Boltzmann Gaussian Temporal Fuzzy Network(RBGTFN)is used in the process of doing information privacy analysis for healthcare data.Remora Colony SwarmOptimization is then used to carry out the optimization of the network.The performance of the model in identifying anomalies across a variety of healthcare data is evaluated by an experimental study.This evaluation suggested that themodel measures the accuracy,precision,latency,Quality of Service(QoS),and scalability of themodel.A remarkable 95%precision,93%latency,89%quality of service,98%detection accuracy,and 96%scalability were obtained by the suggested model,as shown by the subsequent findings.展开更多
In the digital era,the free cross-border flow of data and the development of digital trade are complementary.Consequently,as an inherent demand for data privacy,trade liberalization is closely linked to the right to d...In the digital era,the free cross-border flow of data and the development of digital trade are complementary.Consequently,as an inherent demand for data privacy,trade liberalization is closely linked to the right to data privacy,and data privacy protection is increasingly becoming a trade issue.However,conflicting rule settings between the two create discrepancies and result in differing rule-making approaches.The concept of the right to data privacy provides guidance and evaluative functions for the development of trade liberalization,facilitating the healthy development of digital trade.It is appropriate to treat the interaction between trade liberalization and data privacy protection in a rational way and to place them within independent systems at this stage.Data localization measures are an effective way to balance digital trade liberalization with the right to data privacy.As a data privacy protection measure,data localization has legitimacy within the trade law framework.Looking ahead,to achieve a harmonious advancement of digital trade liberalization and protection of the right to data privacy,all parties should uphold the premise of the national regulatory autonomy,and respect the data localization measures adopted by countries based on their own national conditions and personal data protection considerations.展开更多
Multi-Source data plays an important role in the evolution of media convergence.Its fusion processing enables the further mining of data and utilization of data value and broadens the path for the sharing and dissemin...Multi-Source data plays an important role in the evolution of media convergence.Its fusion processing enables the further mining of data and utilization of data value and broadens the path for the sharing and dissemination of media data.However,it also faces serious problems in terms of protecting user and data privacy.Many privacy protectionmethods have been proposed to solve the problemof privacy leakage during the process of data sharing,but they suffer fromtwo flaws:1)the lack of algorithmic frameworks for specific scenarios such as dynamic datasets in the media domain;2)the inability to solve the problem of the high computational complexity of ciphertext in multi-source data privacy protection,resulting in long encryption and decryption times.In this paper,we propose a multi-source data privacy protection method based on homomorphic encryption and blockchain technology,which solves the privacy protection problem ofmulti-source heterogeneous data in the dissemination ofmedia and reduces ciphertext processing time.We deployed the proposedmethod on theHyperledger platformfor testing and compared it with the privacy protection schemes based on k-anonymity and differential privacy.The experimental results showthat the key generation,encryption,and decryption times of the proposedmethod are lower than those in data privacy protection methods based on k-anonymity technology and differential privacy technology.This significantly reduces the processing time ofmulti-source data,which gives it potential for use in many applications.展开更多
In order to investigate the enhancement of data privacy by distributing data packets via multiple paths, this paper fommlates a security model and analyzes the privacy problem in mul- tipath scenarios leveraging infor...In order to investigate the enhancement of data privacy by distributing data packets via multiple paths, this paper fommlates a security model and analyzes the privacy problem in mul- tipath scenarios leveraging inforrmtion theoretic concept. Based on proposed model, a privacy function related to the path number is discussed. We heuristically recommend the optin^al path num- ber and analyze the tradeoff among the perform- ance, resource consumption and privacy. For re- ducing the inforlmtion leakage, the data schedule algorithms are also proposed. The analytical model can provide guidelines for the multipath protocol design.展开更多
Nowadays,smart wearable devices are used widely in the Social Internet of Things(IoT),which record human physiological data in real time.To protect the data privacy of smart devices,researchers pay more attention to f...Nowadays,smart wearable devices are used widely in the Social Internet of Things(IoT),which record human physiological data in real time.To protect the data privacy of smart devices,researchers pay more attention to federated learning.Although the data leakage problem is somewhat solved,a new challenge has emerged.Asynchronous federated learning shortens the convergence time,while it has time delay and data heterogeneity problems.Both of the two problems harm the accuracy.To overcome these issues,we propose an asynchronous federated learning scheme based on double compensation to solve the problem of time delay and data heterogeneity problems.The scheme improves the Delay Compensated Asynchronous Stochastic Gradient Descent(DC-ASGD)algorithm based on the second-order Taylor expansion as the delay compensation.It adds the FedProx operator to the objective function as the heterogeneity compensation.Besides,the proposed scheme motivates the federated learning process by adjusting the importance of the participants and the central server.We conduct multiple sets of experiments in both conventional and heterogeneous scenarios.The experimental results show that our scheme improves the accuracy by about 5%while keeping the complexity constant.We can find that our scheme converges more smoothly during training and adapts better in heterogeneous environments through numerical experiments.The proposed double-compensation-based federated learning scheme is highly accurate,flexible in terms of participants and smooth the training process.Hence it is deemed suitable for data privacy protection of smart wearable devices.展开更多
Most studies have conducted experiments on predicting energy consumption by integrating data formodel training.However, the process of centralizing data can cause problems of data leakage.Meanwhile,many laws and regul...Most studies have conducted experiments on predicting energy consumption by integrating data formodel training.However, the process of centralizing data can cause problems of data leakage.Meanwhile,many laws and regulationson data security and privacy have been enacted, making it difficult to centralize data, which can lead to a datasilo problem. Thus, to train the model while maintaining user privacy, we adopt a federated learning framework.However, in all classical federated learning frameworks secure aggregation, the Federated Averaging (FedAvg)method is used to directly weight the model parameters on average, which may have an adverse effect on te model.Therefore, we propose the Federated Reinforcement Learning (FedRL) model, which consists of multiple userscollaboratively training the model. Each household trains a local model on local data. These local data neverleave the local area, and only the encrypted parameters are uploaded to the central server to participate in thesecure aggregation of the global model. We improve FedAvg by incorporating a Q-learning algorithm to assignweights to each locally uploaded local model. And the model has improved predictive performance. We validatethe performance of the FedRL model by testing it on a real-world dataset and compare the experimental results withother models. The performance of our proposed method in most of the evaluation metrics is improved comparedto both the centralized and distributed models.展开更多
Speech data publishing breaches users'data privacy,thereby causing more privacy disclosure.Existing work sanitizes content,voice,and voiceprint of speech data without considering the consistence among these three ...Speech data publishing breaches users'data privacy,thereby causing more privacy disclosure.Existing work sanitizes content,voice,and voiceprint of speech data without considering the consistence among these three features,and thus is susceptible to inference attacks.To address the problem,we design a privacy-preserving protocol for speech data publishing(P3S2)that takes the corrections among the three factors into consideration.To concrete,we first propose a three-dimensional sanitization that uses feature learning to capture characteristics in each dimension,and then sanitize speech data using the learned features.As a result,the correlations among the three dimensions of the sanitized speech data are guaranteed.Furthermore,the(ε,δ)-differential privacy is used to theoretically prove both the data privacy preservation and the data utility guarantee of P3S2,filling the gap of algorithm design and performance evaluation.Finally,simulations on two real world datasets have demonstrated both the data privacy preservation and the data utility guarantee.展开更多
COVID-19 has swept the whole our country and the world in the beginning of 2020.31 provinces and municipalities across the country have launched the first-level response to major public health emergencies since Januar...COVID-19 has swept the whole our country and the world in the beginning of 2020.31 provinces and municipalities across the country have launched the first-level response to major public health emergencies since January 24,and China has carried out intensive epidemic control.It is critical for effectively responding to COVID-19 to collect,collate and analyze people’s personal data.What’s more,obtaining identity information,travel records and health information of confirmed cases,suspected cases and close contacts has become a crucial step in epidemic investigation.All regions have made full use of big data to carry out personnel screening,travel records analysis and other related work in epidemic prevention and control,effectively improving the efficiency of epidemic prevention and control.However,data leakage,personnel privacy data exposure,and personal attack frequently occurred in the process of personnel travel records analysis and epidemic prevention and control.It even happened in the WeChat group to forward a person’s name,phone number,address,ID number and other sensitive information.It brought discrimination,telephone and SMS harassment to the parties,which caused great harm to individuals.Based on these,lack of information security and data security awareness and other issues were exposed.Therefore,while big data has been widely concerned and applied,attention should be paid to protecting personal privacy.It is urgent to pay more attention to data privacy and information security in order to effectively protect the legitimate rights of the people.Therefore,measures can be taken to achieve this goal,such as improving the relevant legal system,strengthening technical means to enhance the supervision and management of information security and data protection.展开更多
Large Language Models(LLMs)are complex artificial intelligence systems,which can understand,generate,and translate human languages.By analyzing large amounts of textual data,these models learn language patterns to per...Large Language Models(LLMs)are complex artificial intelligence systems,which can understand,generate,and translate human languages.By analyzing large amounts of textual data,these models learn language patterns to perform tasks such as writing,conversation,and summarization.Agents built on LLMs(LLM agents)further extend these capabilities,allowing them to process user interactions and perform complex operations in diverse task environments.However,during the processing and generation of massive data,LLMs and LLM agents pose a risk of sensitive information leakage,potentially threatening data privacy.This paper aims to demonstrate data privacy issues associated with LLMs and LLM agents to facilitate a comprehensive understanding.Specifically,we conduct an in-depth survey about privacy threats,encompassing passive privacy leakage and active privacy attacks.Subsequently,we introduce the privacy protection mechanisms employed by LLMs and LLM agents and provide a detailed analysis of their effectiveness.Finally,we explore the privacy protection challenges for LLMs and LLM agents as well as outline potential directions for future developments in this domain.展开更多
With increasing demand for data circulation,ensuring data security and privacy is paramount,specifically protecting privacy while maximizing utility.Blockchain,while decentralized and transparent,faces challenges in p...With increasing demand for data circulation,ensuring data security and privacy is paramount,specifically protecting privacy while maximizing utility.Blockchain,while decentralized and transparent,faces challenges in privacy protection and data verification,especially for sensitive data.Existing schemes often suffer from inefficiency and high overhead.We propose a privacy protection scheme using BGV homomorphic encryption and Pedersen Secret Sharing.This scheme enables secure computation on encrypted data,with Pedersen sharding and verifying the private key,ensuring data consistency and immutability.The blockchain framework manages key shards,verifies secrets,and aids security auditing.This approach allows for trusted computation without revealing the underlying data.Preliminary results demonstrate the scheme's feasibility in ensuring data privacy and security,making data available but not visible.This study provides an effective solution for data sharing and privacy protection in blockchain applications.展开更多
The overgeneralisation may happen because most studies on data publishing for multiple sensitive attributes(SAs)have not considered the personalised privacy requirement.Furthermore,sensitive information disclosure may...The overgeneralisation may happen because most studies on data publishing for multiple sensitive attributes(SAs)have not considered the personalised privacy requirement.Furthermore,sensitive information disclosure may also be caused by these personalised requirements.To address the matter,this article develops a personalised data publishing method for multiple SAs.According to the requirements of individuals,the new method partitions SAs values into two categories:private values and public values,and breaks the association between them for privacy guarantees.For the private values,this paper takes the process of anonymisation,while the public values are released without this process.An algorithm is designed to achieve the privacy mode,where the selectivity is determined by the sensitive value frequency and undesirable objects.The experimental results show that the proposed method can provide more information utility when compared with previous methods.The theoretic analyses and experiments also indicate that the privacy can be guaranteed even though the public values are known to an adversary.The overgeneralisation and privacy breach caused by the personalised requirement can be avoided by the new method.展开更多
Advances in technology require upgrades in the law. One such area involves data brokers, which have thus far gone unregulated. Data brokers use artificial intelligence to aggregate information into data profiles about...Advances in technology require upgrades in the law. One such area involves data brokers, which have thus far gone unregulated. Data brokers use artificial intelligence to aggregate information into data profiles about individual Americans derived from consumer use of the internet and connected devices. Data profiles are then sold for profit. Government investigators use a legal loophole to purchase this data instead of obtaining a search warrant, which the Fourth Amendment would otherwise require. Consumers have lacked a reasonable means to fight or correct the information data brokers collect. Americans may not even be aware of the risks of data aggregation, which upends the test of reasonable expectations used in a search warrant analysis. Data aggregation should be controlled and regulated, which is the direction some privacy laws take. Legislatures must step forward to safeguard against shadowy data-profiling practices, whether abroad or at home. In the meantime, courts can modify their search warrant analysis by including data privacy principles.展开更多
The rapid adoption of machine learning in sensitive domains,such as healthcare,finance,and government services,has heightened the need for robust,privacy-preserving techniques.Traditional machine learning approaches l...The rapid adoption of machine learning in sensitive domains,such as healthcare,finance,and government services,has heightened the need for robust,privacy-preserving techniques.Traditional machine learning approaches lack built-in privacy mechanisms,exposing sensitive data to risks,which motivates the development of Privacy-Preserving Machine Learning(PPML)methods.Despite significant advances in PPML,a comprehensive and focused exploration of Secure Multi-Party Computing(SMPC)within this context remains underdeveloped.This review aims to bridge this knowledge gap by systematically analyzing the role of SMPC in PPML,offering a structured overviewof current techniques,challenges,and future directions.Using a semi-systematicmapping studymethodology,this paper surveys recent literature spanning SMPC protocols,PPML frameworks,implementation approaches,threat models,and performance metrics.Emphasis is placed on identifying trends,technical limitations,and comparative strengths of leading SMPC-based methods.Our findings reveal thatwhile SMPCoffers strong cryptographic guarantees for privacy,challenges such as computational overhead,communication costs,and scalability persist.The paper also discusses critical vulnerabilities,practical deployment issues,and variations in protocol efficiency across use cases.展开更多
This research introduces a unique approach to segmenting breast cancer images using a U-Net-based architecture.However,the computational demand for image processing is very high.Therefore,we have conducted this resear...This research introduces a unique approach to segmenting breast cancer images using a U-Net-based architecture.However,the computational demand for image processing is very high.Therefore,we have conducted this research to build a system that enables image segmentation training with low-power machines.To accomplish this,all data are divided into several segments,each being trained separately.In the case of prediction,the initial output is predicted from each trained model for an input,where the ultimate output is selected based on the pixel-wise majority voting of the expected outputs,which also ensures data privacy.In addition,this kind of distributed training system allows different computers to be used simultaneously.That is how the training process takes comparatively less time than typical training approaches.Even after completing the training,the proposed prediction system allows a newly trained model to be included in the system.Thus,the prediction is consistently more accurate.We evaluated the effectiveness of the ultimate output based on four performance matrices:average pixel accuracy,mean absolute error,average specificity,and average balanced accuracy.The experimental results show that the scores of average pixel accuracy,mean absolute error,average specificity,and average balanced accuracy are 0.9216,0.0687,0.9477,and 0.8674,respectively.In addition,the proposed method was compared with four other state-of-the-art models in terms of total training time and usage of computational resources.And it outperformed all of them in these aspects.展开更多
In this paper,the application of agricultural big data in agricultural economic management is deeply explored,and its potential in promoting profit growth and innovation is analyzed.However,challenges persist in data ...In this paper,the application of agricultural big data in agricultural economic management is deeply explored,and its potential in promoting profit growth and innovation is analyzed.However,challenges persist in data collection and integration,limitations of analytical technologies,talent development,team building,and policy support when applying agricultural big data.Effective application strategies are proposed,including data-driven precision agriculture practices,construction of data integration and management platforms,data security and privacy protection strategies,as well as long-term planning and development strategies for agricultural big data,to maximize its impact on agricultural economic management.Future advancements require collaborative efforts in technological innovation,talent cultivation,and policy support,to realize the extensive application of agricultural big data in agricultural economic management and ensure sustainable industrial development.展开更多
The unmanned aerial vehicle(UAV)self-organizing network is composed of multiple UAVs with autonomous capabilities according to a certain structure and scale,which can quickly and accurately complete complex tasks such...The unmanned aerial vehicle(UAV)self-organizing network is composed of multiple UAVs with autonomous capabilities according to a certain structure and scale,which can quickly and accurately complete complex tasks such as path planning,situational awareness,and information transmission.Due to the openness of the network,the UAV cluster is more vulnerable to passive eavesdropping,active interference,and other attacks,which makes the system face serious security threats.This paper proposes a Blockchain-Based Data Acquisition(BDA)scheme with privacy protection to address the data privacy and identity authentication problems in the UAV-assisted data acquisition scenario.Each UAV cluster has an aggregate unmanned aerial vehicle(AGV)that can batch-verify the acquisition reports within its administrative domain.After successful verification,AGV adds its signcrypted ciphertext to the aggregation and uploads it to the blockchain for storage.There are two chains in the blockchain that store the public key information of registered entities and the aggregated reports,respectively.The security analysis shows that theBDAconstruction can protect the privacy and authenticity of acquisition data,and effectively resist a malicious key generation center and the public-key substitution attack.It also provides unforgeability to acquisition reports under the Elliptic Curve Discrete Logarithm Problem(ECDLP)assumption.The performance analysis demonstrates that compared with other schemes,the proposed BDA construction has lower computational complexity and is more suitable for the UAV cluster network with limited computing power and storage capacity.展开更多
Imagine numerous clients,each with personal data;individual inputs are severely corrupt,and a server only concerns the collective,statistically essential facets of this data.In several data mining methods,privacy has ...Imagine numerous clients,each with personal data;individual inputs are severely corrupt,and a server only concerns the collective,statistically essential facets of this data.In several data mining methods,privacy has become highly critical.As a result,various privacy-preserving data analysis technologies have emerged.Hence,we use the randomization process to reconstruct composite data attributes accurately.Also,we use privacy measures to estimate how much deception is required to guarantee privacy.There are several viable privacy protections;however,determining which one is the best is still a work in progress.This paper discusses the difficulty of measuring privacy while also offering numerous random sampling procedures and statistical and categorized data results.Further-more,this paper investigates the use of arbitrary nature with perturbations in privacy preservation.According to the research,arbitrary objects(most notably random matrices)have"predicted"frequency patterns.It shows how to recover crucial information from a sample damaged by a random number using an arbi-trary lattice spectral selection strategy.Thisfiltration system's conceptual frame-work posits,and extensive practicalfindings indicate that sparse data distortions preserve relatively modest privacy protection in various situations.As a result,the research framework is efficient and effective in maintaining data privacy and security.展开更多
Privacy protection for big data linking is discussed here in relation to the Central Statistics Office (CSO), Ireland's, big data linking project titled the 'Structure of Earnings Survey - Administrative Data Proj...Privacy protection for big data linking is discussed here in relation to the Central Statistics Office (CSO), Ireland's, big data linking project titled the 'Structure of Earnings Survey - Administrative Data Project' (SESADP). The result of the project was the creation of datasets and statistical outputs for the years 2011 to 2014 to meet Eurostat's annual earnings statistics requirements and the Structure of Earnings Survey (SES) Regulation. Record linking across the Census and various public sector datasets enabled the necessary information to be acquired to meet the Eurostat earnings requirements. However, the risk of statistical disclosure (i.e. identifying an individual on the dataset) is high unless privacy and confidentiality safe-guards are built into the data matching process. This paper looks at the three methods of linking records on big datasets employed on the SESADP, and how to anonymise the data to protect the identity of the individuals, where potentially disclosive variables exist.展开更多
Existing traffic flow prediction frameworks have already achieved enormous success due to large traffic datasets and capability of deep learning models.However,data privacy and security are always a challenge in every...Existing traffic flow prediction frameworks have already achieved enormous success due to large traffic datasets and capability of deep learning models.However,data privacy and security are always a challenge in every field where data need to be uploaded to the cloud.Federated learning(FL)is an emerging trend for distributed training of data.The primary goal of FL is to train an efficient communication model without compromising data privacy.The traffic data have a robust spatio-temporal correlation,but various approaches proposed earlier have not considered spatial correlation of the traffic data.This paper presents FL-based traffic flow prediction with spatio-temporal correlation.This work uses a differential privacy(DP)scheme for privacy preservation of participant's data.To the best of our knowledge,this is the first time that FL is used for vehicular traffic prediction while considering the spatio-temporal correlation of traffic data with DP preservation.The proposed framework trains the data locally at the client-side with DP.It then uses the model aggregation mechanism federated graph convolutional network(FedGCN)at the server-side to find the average of locally trained models.The results of the proposed work show that the FedGCN model accurately predicts the traffic.DP scheme at client-side helps clients to set a budget for privacy loss.展开更多
基金CAMS Innovation Fund for Medical Sciences(CIFMS):“Construction of an Intelligent Management and Efficient Utilization Technology System for Big Data in Population Health Science.”(2021-I2M-1-057)Key Projects of the Innovation Fund of the National Clinical Research Center for Orthopedics and Sports Rehabilitation:“National Orthopedics and Sports Rehabilitation Real-World Research Platform System Construction”(23-NCRC-CXJJ-ZD4)。
文摘【Objective】Medical imaging data has great value,but it contains a significant amount of sensitive information about patients.At present,laws and regulations regarding to the de-identification of medical imaging data are not clearly defined around the world.This study aims to develop a tool that meets compliance-driven desensitization requirements tailored to diverse research needs.【Methods】To enhance the security of medical image data,we designed and implemented a DICOM format medical image de-identification system on the Windows operating system.【Results】Our custom de-identification system is adaptable to the legal standards of different countries and can accommodate specific research demands.The system offers both web-based online and desktop offline de-identification capabilities,enabling customization of de-identification rules and facilitating batch processing to improve efficiency.【Conclusions】This medical image de-identification system robustly strengthens the stewardship of sensitive medical data,aligning with data security protection requirements while facilitating the sharing and utilization of medical image data.This approach unlocks the intrinsic value inherent in such datasets.
基金funded by Deanship of Scientific Research(DSR)at King Abdulaziz University,Jeddah under grant No.(RG-6-611-43)the authors,therefore,acknowledge with thanks DSR technical and financial support.
文摘The integration of the Internet of Things(IoT)into healthcare systems improves patient care,boosts operational efficiency,and contributes to cost-effective healthcare delivery.However,overcoming several associated challenges,such as data security,interoperability,and ethical concerns,is crucial to realizing the full potential of IoT in healthcare.Real-time anomaly detection plays a key role in protecting patient data and maintaining device integrity amidst the additional security risks posed by interconnected systems.In this context,this paper presents a novelmethod for healthcare data privacy analysis.The technique is based on the identification of anomalies in cloud-based Internet of Things(IoT)networks,and it is optimized using explainable artificial intelligence.For anomaly detection,the Radial Boltzmann Gaussian Temporal Fuzzy Network(RBGTFN)is used in the process of doing information privacy analysis for healthcare data.Remora Colony SwarmOptimization is then used to carry out the optimization of the network.The performance of the model in identifying anomalies across a variety of healthcare data is evaluated by an experimental study.This evaluation suggested that themodel measures the accuracy,precision,latency,Quality of Service(QoS),and scalability of themodel.A remarkable 95%precision,93%latency,89%quality of service,98%detection accuracy,and 96%scalability were obtained by the suggested model,as shown by the subsequent findings.
基金the phased outcome of the project“Research on China’s Rule of Law Path for Maintaining the Security and Stability of Global Supply Chain”(Approval Number 2024M751358)that received funding from the 75th general grant of the China Postdoctoral Science Foundation.
文摘In the digital era,the free cross-border flow of data and the development of digital trade are complementary.Consequently,as an inherent demand for data privacy,trade liberalization is closely linked to the right to data privacy,and data privacy protection is increasingly becoming a trade issue.However,conflicting rule settings between the two create discrepancies and result in differing rule-making approaches.The concept of the right to data privacy provides guidance and evaluative functions for the development of trade liberalization,facilitating the healthy development of digital trade.It is appropriate to treat the interaction between trade liberalization and data privacy protection in a rational way and to place them within independent systems at this stage.Data localization measures are an effective way to balance digital trade liberalization with the right to data privacy.As a data privacy protection measure,data localization has legitimacy within the trade law framework.Looking ahead,to achieve a harmonious advancement of digital trade liberalization and protection of the right to data privacy,all parties should uphold the premise of the national regulatory autonomy,and respect the data localization measures adopted by countries based on their own national conditions and personal data protection considerations.
基金funded by the High-Quality and Cutting-Edge Discipline Construction Project for Universities in Beijing (Internet Information,Communication University of China).
文摘Multi-Source data plays an important role in the evolution of media convergence.Its fusion processing enables the further mining of data and utilization of data value and broadens the path for the sharing and dissemination of media data.However,it also faces serious problems in terms of protecting user and data privacy.Many privacy protectionmethods have been proposed to solve the problemof privacy leakage during the process of data sharing,but they suffer fromtwo flaws:1)the lack of algorithmic frameworks for specific scenarios such as dynamic datasets in the media domain;2)the inability to solve the problem of the high computational complexity of ciphertext in multi-source data privacy protection,resulting in long encryption and decryption times.In this paper,we propose a multi-source data privacy protection method based on homomorphic encryption and blockchain technology,which solves the privacy protection problem ofmulti-source heterogeneous data in the dissemination ofmedia and reduces ciphertext processing time.We deployed the proposedmethod on theHyperledger platformfor testing and compared it with the privacy protection schemes based on k-anonymity and differential privacy.The experimental results showthat the key generation,encryption,and decryption times of the proposedmethod are lower than those in data privacy protection methods based on k-anonymity technology and differential privacy technology.This significantly reduces the processing time ofmulti-source data,which gives it potential for use in many applications.
基金This paper was partially supported by the National Basic Re-search Program of China under Grant No. 2007CB307100 the National Natural Science Foundation of China under Crant No 60972010 the Fundamental Research Funds for the Central U- niversities under Grant No. 2011JBM018.
文摘In order to investigate the enhancement of data privacy by distributing data packets via multiple paths, this paper fommlates a security model and analyzes the privacy problem in mul- tipath scenarios leveraging inforrmtion theoretic concept. Based on proposed model, a privacy function related to the path number is discussed. We heuristically recommend the optin^al path num- ber and analyze the tradeoff among the perform- ance, resource consumption and privacy. For re- ducing the inforlmtion leakage, the data schedule algorithms are also proposed. The analytical model can provide guidelines for the multipath protocol design.
基金supported by the National Natural Science Foundation of China,No.61977006.
文摘Nowadays,smart wearable devices are used widely in the Social Internet of Things(IoT),which record human physiological data in real time.To protect the data privacy of smart devices,researchers pay more attention to federated learning.Although the data leakage problem is somewhat solved,a new challenge has emerged.Asynchronous federated learning shortens the convergence time,while it has time delay and data heterogeneity problems.Both of the two problems harm the accuracy.To overcome these issues,we propose an asynchronous federated learning scheme based on double compensation to solve the problem of time delay and data heterogeneity problems.The scheme improves the Delay Compensated Asynchronous Stochastic Gradient Descent(DC-ASGD)algorithm based on the second-order Taylor expansion as the delay compensation.It adds the FedProx operator to the objective function as the heterogeneity compensation.Besides,the proposed scheme motivates the federated learning process by adjusting the importance of the participants and the central server.We conduct multiple sets of experiments in both conventional and heterogeneous scenarios.The experimental results show that our scheme improves the accuracy by about 5%while keeping the complexity constant.We can find that our scheme converges more smoothly during training and adapts better in heterogeneous environments through numerical experiments.The proposed double-compensation-based federated learning scheme is highly accurate,flexible in terms of participants and smooth the training process.Hence it is deemed suitable for data privacy protection of smart wearable devices.
基金supported by National Key R&D Program of China(No.2020YFC2006602)National Natural Science Foundation of China(Nos.62172324,62072324,61876217,6187612)+2 种基金University Natural Science Foundation of Jiangsu Province(No.21KJA520005)Primary Research and Development Plan of Jiangsu Province(No.BE2020026)Natural Science Foundation of Jiangsu Province(No.BK20190942).
文摘Most studies have conducted experiments on predicting energy consumption by integrating data formodel training.However, the process of centralizing data can cause problems of data leakage.Meanwhile,many laws and regulationson data security and privacy have been enacted, making it difficult to centralize data, which can lead to a datasilo problem. Thus, to train the model while maintaining user privacy, we adopt a federated learning framework.However, in all classical federated learning frameworks secure aggregation, the Federated Averaging (FedAvg)method is used to directly weight the model parameters on average, which may have an adverse effect on te model.Therefore, we propose the Federated Reinforcement Learning (FedRL) model, which consists of multiple userscollaboratively training the model. Each household trains a local model on local data. These local data neverleave the local area, and only the encrypted parameters are uploaded to the central server to participate in thesecure aggregation of the global model. We improve FedAvg by incorporating a Q-learning algorithm to assignweights to each locally uploaded local model. And the model has improved predictive performance. We validatethe performance of the FedRL model by testing it on a real-world dataset and compare the experimental results withother models. The performance of our proposed method in most of the evaluation metrics is improved comparedto both the centralized and distributed models.
基金National Natural Science Foundation of China(No.61902060)Shanghai Sailing Program,China(No.19YF1402100)Fundamental Research Funds for the Central Universities,China(No.2232019D3-51)。
文摘Speech data publishing breaches users'data privacy,thereby causing more privacy disclosure.Existing work sanitizes content,voice,and voiceprint of speech data without considering the consistence among these three features,and thus is susceptible to inference attacks.To address the problem,we design a privacy-preserving protocol for speech data publishing(P3S2)that takes the corrections among the three factors into consideration.To concrete,we first propose a three-dimensional sanitization that uses feature learning to capture characteristics in each dimension,and then sanitize speech data using the learned features.As a result,the correlations among the three dimensions of the sanitized speech data are guaranteed.Furthermore,the(ε,δ)-differential privacy is used to theoretically prove both the data privacy preservation and the data utility guarantee of P3S2,filling the gap of algorithm design and performance evaluation.Finally,simulations on two real world datasets have demonstrated both the data privacy preservation and the data utility guarantee.
基金This paper is support by:In 2019,Liaoning Provincial Department of Education Project named“Study on the Path Selection of Rural Revitalization in Ethnic Autonomous Areas of Liaoning Province”The 3rd Azure Talent Project of Dalian Ocean University in 2018+1 种基金In 2019,Liaoning Province’s overseas training project“China-Canada Cooperation Research Plan on Marine Law and Policy”(2019GJWYB019)The Ministry of Education filed the 2017 National and Regional Research Center Project“Northeast Asia Research Center for Marine Law and Policy”(GQ17091).
文摘COVID-19 has swept the whole our country and the world in the beginning of 2020.31 provinces and municipalities across the country have launched the first-level response to major public health emergencies since January 24,and China has carried out intensive epidemic control.It is critical for effectively responding to COVID-19 to collect,collate and analyze people’s personal data.What’s more,obtaining identity information,travel records and health information of confirmed cases,suspected cases and close contacts has become a crucial step in epidemic investigation.All regions have made full use of big data to carry out personnel screening,travel records analysis and other related work in epidemic prevention and control,effectively improving the efficiency of epidemic prevention and control.However,data leakage,personnel privacy data exposure,and personal attack frequently occurred in the process of personnel travel records analysis and epidemic prevention and control.It even happened in the WeChat group to forward a person’s name,phone number,address,ID number and other sensitive information.It brought discrimination,telephone and SMS harassment to the parties,which caused great harm to individuals.Based on these,lack of information security and data security awareness and other issues were exposed.Therefore,while big data has been widely concerned and applied,attention should be paid to protecting personal privacy.It is urgent to pay more attention to data privacy and information security in order to effectively protect the legitimate rights of the people.Therefore,measures can be taken to achieve this goal,such as improving the relevant legal system,strengthening technical means to enhance the supervision and management of information security and data protection.
基金supported in part by the National Natural Science Foundation of China(62402288 and 62302063)the China Postdoctoral Science Foundation,China(2024M751811).
文摘Large Language Models(LLMs)are complex artificial intelligence systems,which can understand,generate,and translate human languages.By analyzing large amounts of textual data,these models learn language patterns to perform tasks such as writing,conversation,and summarization.Agents built on LLMs(LLM agents)further extend these capabilities,allowing them to process user interactions and perform complex operations in diverse task environments.However,during the processing and generation of massive data,LLMs and LLM agents pose a risk of sensitive information leakage,potentially threatening data privacy.This paper aims to demonstrate data privacy issues associated with LLMs and LLM agents to facilitate a comprehensive understanding.Specifically,we conduct an in-depth survey about privacy threats,encompassing passive privacy leakage and active privacy attacks.Subsequently,we introduce the privacy protection mechanisms employed by LLMs and LLM agents and provide a detailed analysis of their effectiveness.Finally,we explore the privacy protection challenges for LLMs and LLM agents as well as outline potential directions for future developments in this domain.
基金supported by the National Key Research and Development Plan in China(Grant No.2020YFB1005500)。
文摘With increasing demand for data circulation,ensuring data security and privacy is paramount,specifically protecting privacy while maximizing utility.Blockchain,while decentralized and transparent,faces challenges in privacy protection and data verification,especially for sensitive data.Existing schemes often suffer from inefficiency and high overhead.We propose a privacy protection scheme using BGV homomorphic encryption and Pedersen Secret Sharing.This scheme enables secure computation on encrypted data,with Pedersen sharding and verifying the private key,ensuring data consistency and immutability.The blockchain framework manages key shards,verifies secrets,and aids security auditing.This approach allows for trusted computation without revealing the underlying data.Preliminary results demonstrate the scheme's feasibility in ensuring data privacy and security,making data available but not visible.This study provides an effective solution for data sharing and privacy protection in blockchain applications.
基金Doctoral research start-up fund of Guangxi Normal UniversityGuangzhou Research Institute of Communication University of China Common Construction Project,Sunflower-the Aging Intelligent CommunityGuangxi project of improving Middle-aged/Young teachers'ability,Grant/Award Number:2020KY020323。
文摘The overgeneralisation may happen because most studies on data publishing for multiple sensitive attributes(SAs)have not considered the personalised privacy requirement.Furthermore,sensitive information disclosure may also be caused by these personalised requirements.To address the matter,this article develops a personalised data publishing method for multiple SAs.According to the requirements of individuals,the new method partitions SAs values into two categories:private values and public values,and breaks the association between them for privacy guarantees.For the private values,this paper takes the process of anonymisation,while the public values are released without this process.An algorithm is designed to achieve the privacy mode,where the selectivity is determined by the sensitive value frequency and undesirable objects.The experimental results show that the proposed method can provide more information utility when compared with previous methods.The theoretic analyses and experiments also indicate that the privacy can be guaranteed even though the public values are known to an adversary.The overgeneralisation and privacy breach caused by the personalised requirement can be avoided by the new method.
文摘Advances in technology require upgrades in the law. One such area involves data brokers, which have thus far gone unregulated. Data brokers use artificial intelligence to aggregate information into data profiles about individual Americans derived from consumer use of the internet and connected devices. Data profiles are then sold for profit. Government investigators use a legal loophole to purchase this data instead of obtaining a search warrant, which the Fourth Amendment would otherwise require. Consumers have lacked a reasonable means to fight or correct the information data brokers collect. Americans may not even be aware of the risks of data aggregation, which upends the test of reasonable expectations used in a search warrant analysis. Data aggregation should be controlled and regulated, which is the direction some privacy laws take. Legislatures must step forward to safeguard against shadowy data-profiling practices, whether abroad or at home. In the meantime, courts can modify their search warrant analysis by including data privacy principles.
文摘The rapid adoption of machine learning in sensitive domains,such as healthcare,finance,and government services,has heightened the need for robust,privacy-preserving techniques.Traditional machine learning approaches lack built-in privacy mechanisms,exposing sensitive data to risks,which motivates the development of Privacy-Preserving Machine Learning(PPML)methods.Despite significant advances in PPML,a comprehensive and focused exploration of Secure Multi-Party Computing(SMPC)within this context remains underdeveloped.This review aims to bridge this knowledge gap by systematically analyzing the role of SMPC in PPML,offering a structured overviewof current techniques,challenges,and future directions.Using a semi-systematicmapping studymethodology,this paper surveys recent literature spanning SMPC protocols,PPML frameworks,implementation approaches,threat models,and performance metrics.Emphasis is placed on identifying trends,technical limitations,and comparative strengths of leading SMPC-based methods.Our findings reveal thatwhile SMPCoffers strong cryptographic guarantees for privacy,challenges such as computational overhead,communication costs,and scalability persist.The paper also discusses critical vulnerabilities,practical deployment issues,and variations in protocol efficiency across use cases.
基金the Researchers Supporting Project,King Saud University,Saudi Arabia,for funding this research work through Project No.RSPD2025R951.
文摘This research introduces a unique approach to segmenting breast cancer images using a U-Net-based architecture.However,the computational demand for image processing is very high.Therefore,we have conducted this research to build a system that enables image segmentation training with low-power machines.To accomplish this,all data are divided into several segments,each being trained separately.In the case of prediction,the initial output is predicted from each trained model for an input,where the ultimate output is selected based on the pixel-wise majority voting of the expected outputs,which also ensures data privacy.In addition,this kind of distributed training system allows different computers to be used simultaneously.That is how the training process takes comparatively less time than typical training approaches.Even after completing the training,the proposed prediction system allows a newly trained model to be included in the system.Thus,the prediction is consistently more accurate.We evaluated the effectiveness of the ultimate output based on four performance matrices:average pixel accuracy,mean absolute error,average specificity,and average balanced accuracy.The experimental results show that the scores of average pixel accuracy,mean absolute error,average specificity,and average balanced accuracy are 0.9216,0.0687,0.9477,and 0.8674,respectively.In addition,the proposed method was compared with four other state-of-the-art models in terms of total training time and usage of computational resources.And it outperformed all of them in these aspects.
基金Supported by Research and Application of Soil Collection Software and Soil Ecological Big Data Platform in Guangxi Woodland(GUILINKEYAN[2022ZC]44)Construction of Soil Information Database and Visualization System for Artificial Forests in Central Guangxi(2023GXZCLK62).
文摘In this paper,the application of agricultural big data in agricultural economic management is deeply explored,and its potential in promoting profit growth and innovation is analyzed.However,challenges persist in data collection and integration,limitations of analytical technologies,talent development,team building,and policy support when applying agricultural big data.Effective application strategies are proposed,including data-driven precision agriculture practices,construction of data integration and management platforms,data security and privacy protection strategies,as well as long-term planning and development strategies for agricultural big data,to maximize its impact on agricultural economic management.Future advancements require collaborative efforts in technological innovation,talent cultivation,and policy support,to realize the extensive application of agricultural big data in agricultural economic management and ensure sustainable industrial development.
基金supported in part by the National Key R&D Program of China under Project 2020YFB1006004the Guangxi Natural Science Foundation under Grants 2019GXNSFFA245015 and 2019GXNSFGA245004+2 种基金the National Natural Science Foundation of China under Projects 62162017,61862012,61962012,and 62172119the Major Key Project of PCL under Grants PCL2021A09,PCL2021A02 and PCL2022A03the Innovation Project of Guangxi Graduate Education YCSW2021175.
文摘The unmanned aerial vehicle(UAV)self-organizing network is composed of multiple UAVs with autonomous capabilities according to a certain structure and scale,which can quickly and accurately complete complex tasks such as path planning,situational awareness,and information transmission.Due to the openness of the network,the UAV cluster is more vulnerable to passive eavesdropping,active interference,and other attacks,which makes the system face serious security threats.This paper proposes a Blockchain-Based Data Acquisition(BDA)scheme with privacy protection to address the data privacy and identity authentication problems in the UAV-assisted data acquisition scenario.Each UAV cluster has an aggregate unmanned aerial vehicle(AGV)that can batch-verify the acquisition reports within its administrative domain.After successful verification,AGV adds its signcrypted ciphertext to the aggregation and uploads it to the blockchain for storage.There are two chains in the blockchain that store the public key information of registered entities and the aggregated reports,respectively.The security analysis shows that theBDAconstruction can protect the privacy and authenticity of acquisition data,and effectively resist a malicious key generation center and the public-key substitution attack.It also provides unforgeability to acquisition reports under the Elliptic Curve Discrete Logarithm Problem(ECDLP)assumption.The performance analysis demonstrates that compared with other schemes,the proposed BDA construction has lower computational complexity and is more suitable for the UAV cluster network with limited computing power and storage capacity.
文摘Imagine numerous clients,each with personal data;individual inputs are severely corrupt,and a server only concerns the collective,statistically essential facets of this data.In several data mining methods,privacy has become highly critical.As a result,various privacy-preserving data analysis technologies have emerged.Hence,we use the randomization process to reconstruct composite data attributes accurately.Also,we use privacy measures to estimate how much deception is required to guarantee privacy.There are several viable privacy protections;however,determining which one is the best is still a work in progress.This paper discusses the difficulty of measuring privacy while also offering numerous random sampling procedures and statistical and categorized data results.Further-more,this paper investigates the use of arbitrary nature with perturbations in privacy preservation.According to the research,arbitrary objects(most notably random matrices)have"predicted"frequency patterns.It shows how to recover crucial information from a sample damaged by a random number using an arbi-trary lattice spectral selection strategy.Thisfiltration system's conceptual frame-work posits,and extensive practicalfindings indicate that sparse data distortions preserve relatively modest privacy protection in various situations.As a result,the research framework is efficient and effective in maintaining data privacy and security.
文摘Privacy protection for big data linking is discussed here in relation to the Central Statistics Office (CSO), Ireland's, big data linking project titled the 'Structure of Earnings Survey - Administrative Data Project' (SESADP). The result of the project was the creation of datasets and statistical outputs for the years 2011 to 2014 to meet Eurostat's annual earnings statistics requirements and the Structure of Earnings Survey (SES) Regulation. Record linking across the Census and various public sector datasets enabled the necessary information to be acquired to meet the Eurostat earnings requirements. However, the risk of statistical disclosure (i.e. identifying an individual on the dataset) is high unless privacy and confidentiality safe-guards are built into the data matching process. This paper looks at the three methods of linking records on big datasets employed on the SESADP, and how to anonymise the data to protect the identity of the individuals, where potentially disclosive variables exist.
文摘Existing traffic flow prediction frameworks have already achieved enormous success due to large traffic datasets and capability of deep learning models.However,data privacy and security are always a challenge in every field where data need to be uploaded to the cloud.Federated learning(FL)is an emerging trend for distributed training of data.The primary goal of FL is to train an efficient communication model without compromising data privacy.The traffic data have a robust spatio-temporal correlation,but various approaches proposed earlier have not considered spatial correlation of the traffic data.This paper presents FL-based traffic flow prediction with spatio-temporal correlation.This work uses a differential privacy(DP)scheme for privacy preservation of participant's data.To the best of our knowledge,this is the first time that FL is used for vehicular traffic prediction while considering the spatio-temporal correlation of traffic data with DP preservation.The proposed framework trains the data locally at the client-side with DP.It then uses the model aggregation mechanism federated graph convolutional network(FedGCN)at the server-side to find the average of locally trained models.The results of the proposed work show that the FedGCN model accurately predicts the traffic.DP scheme at client-side helps clients to set a budget for privacy loss.