典型文献
Improving Parameter Estimation and Defensive Ability of Latent Dirichlet Allocation Model Training Under Rényi Differential Privacy
文献摘要:
Latent Dirichlet allocation(LDA)is a topic model widely used for discovering hidden semantics in massive text corpora.Collapsed Gibbs sampling(CGS),as a widely-used algorithm for learning the parameters of LDA,has the risk of privacy leakage.Specifically,word count statistics and updates of latent topics in CGS,which are essential for parameter estimation,could be employed by adversaries to conduct effective membership inference attacks(MIAs).Till now,there are two kinds of methods exploited in CGS to defend against MIAs:adding noise to word count statistics and utilizing inherent privacy.These two kinds of methods have their respective limitations.Noise sampled from the Laplacian distribution sometimes produces negative word count statistics,which render terrible parameter estimation in CGS.Utilizing inherent privacy could only provide weak guaranteed privacy when defending against MIAs.It is promising to propose an effective framework to obtain accurate parameter estimations with guaranteed differential privacy.The key issue of obtaining accurate parameter estimations when introducing differential privacy in CGS is making good use of the privacy budget such that a precise noise scale is derived.It is the first time that Rényi differential privacy(RDP)has been introduced into CGS and we propose RDP-LDA,an effective framework for analyzing the privacy loss of any differentially private CGS.RDP-LDA could be used to derive a tighter upper bound of privacy loss than the overestimated results of existing differentially private CGS obtained by ε-DP.In RDP-LDA,we propose a novel truncated-Gaussian mechanism that keeps word count statistics non-negative.And we propose distribution perturbation which could provide more rigorous guaranteed privacy than utilizing inherent privacy.Experiments validate that our proposed methods produce more accurate parameter estimation under the JS-divergence metric and obtain lower precision and recall when defending against MIAs.
文献关键词:
中图分类号:
作者姓名:
Tao Huang;Su-Yun Zhao;Hong Chen;Yi-Xuan Liu
作者机构:
Key Laboratory of Data Engineering and Knowledge Engineering(Renmin University of China),Ministry of Education Beijing 100087,China;School of Information,Renmin University of China,Beijing 100087,China;School of Information,Renmin University of China,Beijing;computer application at Renmin University of China,Beijing
文献出处:
引用格式:
[1]Tao Huang;Su-Yun Zhao;Hong Chen;Yi-Xuan Liu-.Improving Parameter Estimation and Defensive Ability of Latent Dirichlet Allocation Model Training Under Rényi Differential Privacy)[J].计算机科学技术学报(英文版),2022(06):1382-1397
A类:
Collapsed,MIAs
B类:
Improving,Parameter,Estimation,Defensive,Ability,Latent,Dirichlet,Allocation,Model,Training,Under,nyi,Differential,Privacy,allocation,LDA,model,widely,used,discovering,hidden,semantics,massive,text,corpora,Gibbs,sampling,CGS,algorithm,learning,parameters,has,risk,privacy,leakage,Specifically,word,count,statistics,updates,latent,topics,which,are,essential,could,employed,by,adversaries,conduct,effective,membership,inference,attacks,Till,now,there,two,kinds,methods,exploited,against,adding,noise,utilizing,inherent,These,have,their,respective,limitations,Noise,sampled,from,Laplacian,distribution,sometimes,produces,negative,render,terrible,Utilizing,only,provide,weak,guaranteed,when,defending,It,promising,framework,accurate,estimations,key,issue,obtaining,introducing,making,good,budget,such,that,precise,scale,derived,first,RDP,been,introduced,into,analyzing,loss,any,differentially,private,tighter,upper,bound,than,overestimated,results,existing,obtained,In,novel,truncated,Gaussian,mechanism,keeps,And,perturbation,more,rigorous,Experiments,validate,our,proposed,under,JS,divergence,metric,lower,precision,recall
AB值:
0.507218
相似文献
机标中图分类号,由域田数据科技根据网络公开资料自动分析生成,仅供学习研究参考。