Differential Privacy Preservation for Deep Auto-Encoders: an Application of Human Behavior Prediction
|
|
- Maud Francis
- 5 years ago
- Views:
Transcription
1 Differentia Privacy Preservation for Deep Auto-Encoders: an Appication of Human Behavior Prediction NhatHai Phan 1, Yue Wang 2, Xintao Wu 3, Dejing Dou 1 {haiphan,dou}@csuoregonedu, ywang91@unccedu, xintaowu@uarkedu 1 University of Oregon, USA; 2 University of North Caroina at Charotte, USA; 3 University of Arkansas, USA Abstract In recent years, deep earning has spread beyond both academia and industry with many exciting rea-word appications The deveopment of deep earning has presented obvious privacy issues However, there has been ack of scientific study about privacy preservation in deep earning In this paper, we concentrate on the autoencoder, a fundamenta component in deep earning, and propose the deep private auto-encoder (dpa Our main idea is to enforce ɛ-differentia privacy by perturbing the objective functions of the traditiona deep autoencoder, rather than its resuts We appy the dpa to human behavior prediction in a heath socia network Theoretica anaysis and thorough experimenta evauations show that the dpa is highy effective and efficient, and it significanty outperforms existing soutions Introduction Deep earning is a timey and promising area of machine earning research In genera, deep earning is about earning mutipe eves of representation and abstraction that hep to make sense of data such as images, sound, and text (Deng and Yu 2014 With significant success in recent years, the appications of deep earning are being expanded into other fieds such as socia media (Yuan et a 2013, socia network anaysis (Perozzi, A-Rfou, and Skiena 2014, bioinformatics (Chicco, Sadowski, and Badi 2014, medicine and heathcare (Song et a 2014 This presents obvious issues about protecting privacy in deep earning modes when they are buit on users persona and highy sensitive data, such as cinica records, user profies, photo, etc The present research efforts ack sufficient deep earning techniques that incorporate privacy concerns Therefore, deveoping a privacy preserving deep earning mode is an urgent demand Reeasing sensitive resuts of statistica anaysis and data mining whie protecting privacy has been studied in the past few decades One state-of-the-art approach to the probem is ɛ-differentia privacy (Dwork et a 200, which works by injecting random noise into the reeased statistica resuts computed from the underying sensitive data, such that the distribution of the noisy resuts is reativey insensitive to Copyright c 201, Association for the Advancement of Artificia Inteigence (wwwaaaiorg A rights reserved any change of a singe record in the origina dataset This ensures that the adversary cannot infer any information about any particuar record with high confidence (controed by parameter ɛ, even if the adversary possesses a the remaining tupes of the sensitive data In this paper, we propose to deveop an ɛ-differentia privacy preserving deep earning mode The structure of deep earning modes, eg, deep auto-encoders (Bengio 2009, deep beief networks (Hinton, Osindero, and Teh 200, usuay contains mutipe ayers of neurons At each ayer, they use different objective functions (eg, cross-entropy error, energy based functions and agorithms (eg, contrastive divergence (Hinton 2002, backpropagation to earn the optima parameters for output tasks such as cassification and prediction In addition, the designs of deep earning modes are varied and dependent on appication domains It is difficut, therefore, to design a unified ɛ-differentia privacy soution that covers a deep earning modes Deep auto-encoders (das (Bengio 2009 are one of the fundamenta deep earning modes which have been used in many appications such as heathcare and medicine, natura anguage processing, etc (Deng and Yu 2014 In this paper, we aim at deriving an ɛ-differentia privacy approach to deep auto-encoders (das for a binomia cassification task There are severa potentia soutions for ɛ-differentia privacy that target regression, which is simiar to the data reconstruction and cross-entropy error functions in deep autoencoders (Zhang et a 2012; Chaudhuri and Monteeoni 2008; Chaudhuri, Monteeoni, and Sarwate 2011; Lei 2011; Smith 2011 The most pertinent agorithm that we can adapt in our work is functiona mechanism (FM (Zhang et a 2012 The FM enforces ɛ-differentia privacy by perturbing the objective function of the optimization probem, eg, ogistic regression, rather than its resuts By everaging the functiona mechanism to perturb the objective functions in deep auto-encoders so that the ɛ-differentia privacy is preserved, we propose a nove ɛ-differentia Private Autoencoder (PA First, the cross-entropy error functions of the data reconstruction and softmax ayer are approximated to poynomia forms by using Tayor Expansion (Arfken 1985 Second, we inject noise into these poynomia forms so that the ɛ-differentia privacy is satisfied in the training phases Third, in the PA, we add a normaization ayer on top of the hidden ayer to protect the ɛ-differentia privacy when stacking mutipe PAs Mutipe PAs can be stacked on each other
2 to produce a deep private auto-encoder (dpa To evauate its performance, we appy the dpa mode for human behavior prediction in a rea heath socia network Conducted experimenta resuts demonstrate that the dpa mode achieves accurate resuts with prediction power comparabe to the unperturbed resuts, and it outperforms the existing soutions Preiminaries and Reated Works In this section, we briefy revisit the differentia privacy definition, Functiona Mechanism (Zhang et a 2012, and formay introduce the deep auto-encoder Let D be a database that contains n tupes t 1, t 2,, t n and d+1 attributes X 1, X 2,, X d, Y For each tupe t i = (x i1, x i2,, x id, y i, we assume without oss of generaity d x2 ij 1 where x ij 0, y i foows a binomia distribution Our objective is to construct a deep auto-encoder ρ from D that (i takes x i = (x i1, x i2,, x id as input and (ii outputs a prediction of y i that is as accurate as possibe The mode function ρ contains a mode parameter vector ω To evauate whether ω eads to an accurate mode, a cost function f D (ω is often used to measure the difference between the origina and predicted vaues of y i As the reeased mode parameter ω may discose sensitive information of D, to protect the privacy, we require that the mode training shoud be performed with an agorithm which satisfies ɛ-differentia privacy Differentia Privacy Differentia privacy research has been significanty studied from the theoretica perspective, eg, (Chaudhuri and Monteeoni 2008; Kifer and Machanavajjhaa 2011 There are aso extensive studies on the appicabiity of enforcing differentia privacy in some particuar anaysis tasks, eg, coaborative recommendation (McSherry and Mironov 2009, ogistic regression (Chaudhuri and Monteeoni 2008, pubishing contingency tabes (Xiao, Wang, and Gehrke 2010, and spectra graph anaysis (Wang, Wu, and Wu 2013 in socia network anaysis The mechanisms of achieving differentia privacy mainy incude the cassic approach of adding Lapacian noise (Dwork et a 200, the exponentia mechanism (McSherry and Tawar 2007, and the functiona perturbation approach (Chaudhuri and Monteeoni 2008 Definition 1 (ɛ-different Privacy (Dwork et a 200 A randomized agorithm A fufis ɛ-differentia privacy, iff for any two databases D and D differing at most one tupe, and for any output O of A, we have: P r(a(d = O e ɛ P r(a(d = O (1 The privacy budget ɛ contros the amount by which the distributions induced by two neighboring data sets may differ Smaer vaues of ɛ enforce a stronger privacy guarantee of A A genera method for computing an approximation to any function f (on D whie preserving ɛ-differentia privacy is the Lapace mechanism (Dwork et a 200, where the output of f is a vector of rea numbers In particuar, the mechanism expoits the goba sensitivity of f over any two neighboring data sets (differing at most one record, which is denoted as S f (D Given S f (D, the Lapace mechanism ensures ɛ-differentia privacy by injecting noise η into each vaue in the output of f(d where η is drawn iid from Lapace distribution with zero mean and scae S f (D/ɛ In (Shokri and Shmatikov 2015, Shokri et a proposed a distributed training method to preserve privacy in neura networks They extended their method to preserve differentia privacy for distributed deep neura networks Their method aims at preserving epsion-differentia privacy for each training epoch of each participant Even their method is reasonabe given the context of distributed deep neura networks; their approach may consume too much privacy budget to ensure mode accuracy when the number of training epochs, the number of participants, and the number of shared parameters are arge Functiona Mechanism Revisited Functiona mechanism (Zhang et a 2012 is an extension of the Lapace mechanism It achieves ɛ-differentia privacy by perturbing the objective function f D (ω and then reeasing the mode parameter ω that minimizes the perturbed objective function f D (ω instead of the origina one The functiona mechanism expoits the poynomia representation of f D (ω The mode parameter ω is a vector that contains d vaues ω 1,, ω d Let φ(ω denote a product of ω 1,, ω d, namey, φ(ω = ω c1 1 ωc2 2 ωc d d for some c 1,, c d N Let Φ j (j N denote the set of a products of ω 1,, ω d with degree j, ie, Φ j = { ω c1 1 ωc2 2 ωc d d d =1 c = j } By the Stone-Weierstrass Theorem, any continuous and differentiabe f(t i, ω can aways be written as a poynomia of ω 1,, ω d, for some J [0, ], ie, f(t i, ω = J j=0 φ Φ j λ φti φ(ω where λ φti R denotes the coefficient of φ(ω in the poynomia For instance, the poynomia expression of the oss function in the inear regression is as foows: f(x i, ω = (y i x T i ω2 = yi 2 d (2y ix ij ω j + 1 j, d (x ijx i ω j ω We can see that it ony invoves monomias in Φ 0 = {1}, Φ 1 = {ω 1,, ω d }, and Φ 2 = {ω i ω j i, j [1, d]} Each φ(ω has its own coefficient, eg, for ω j, its poynomia coefficient λ φti = 2y i x ij Simiary, f D (ω can aso be expressed as a poynomia of ω 1,, ω d f D (ω = J λ φti φ(ω (2 j=0 φ Φ j t i D Lemma 1 (Zhang et a 2012 Let D and D be any two neighboring datasets Let f D (ω and f D (ω be the objective functions of regression anaysis on D and D, respectivey The foowing inequaity hods = J 2 φ Φ j t i D λ φti t i D λ φt i where t i, t i or t is an arbitrary tupe 2 max 1 t J φ Φ j λ φt 1 To achieve ɛ-differentia privacy, f D (ω is perturbed by injecting Lapace noise Lap( ɛ into its poynomia coefficients λ φ, and then the mode parameter ω is derived
3 to minimize the perturbed function f D (ω, where = J 2 max t φ Φ j λ φt 1, according to the Lemma 1 When the poynomia form of an objective function (eg, ogistic regression contains terms with unbounded degrees, Zhang et a (Zhang et a 2012 deveoped an approximation poynomia form based on Tayor Expansion (Arfken 1985 For instance, given the cost function f(t i, ω of regression anaysis, assume that there exist 2m functions f 1,, f m and g 1,, g m such that f(t i, ω = =1 f (g (t i, ω, and each g (t i, ω is a poynomia function of ω, where m N is the number of functions f and g Given the above decomposition of f(t i, ω, we can appy the Tayor Expansion (Arfken 1985 on each f ( to obtain the foowing equation: f(t i, ω = m =1 R=0 (z ( R g (t i, ω z (3 f (R where each z is a rea number Thus, the objective function f(d, ω can be written as a poynomia function, ie, f D (ω = m i=1 =1 R=0 (z ( R g (t i, ω z (4 f (R Deep Auto-Encoders The deep auto-encoder (Bengio 2009 is one of the fundamenta modes in deep earning (Deng and Yu 2014 An auto-encoder has a ayer of input units, and a ayer of data reconstruction units fuy connected to a ayer of hidden units, but no connections within a ayer (Fig 1a An auto-encoder is trained to encode the input in some representation x i = {x i1,, x id } so that the input can be reconstructed from that representation The auto-encoder prefers to minimize the negative og-ikeihood of the reconstruction, given the encoding x i, RE(t i, W = og P (x i x i, W ( = xij og x ij + (1 x ij og(1 x ij (5 where W is a weight matrix (ie, from now on W wi be used to denote the parameters instead of ω, x = σ(hw, h = σ(w T x ( σ( is the sigmoid function The oss function on the whoe dataset D is the summation over a the data points x: = i=1 RE(D, W = RE(x i, W = (7 i=1 [ ] x ij og(1+e Wjhi +(1 x ij og(1+e Wjhi In the foowing, we bur the distinction between x i and t i, which means t i indicates the set of attributes in x i t i and x i wi be used interchangeaby We can stack mutipe auto-encoders to produce a deep auto-encoder On top of the deep auto-encoder, we add an output ayer which incudes a singe binomia variabe to predict Y The output variabe ŷ Agorithm 1: Pseudo Code of a dpa mode 1 Derive poynomia approximation of data reconstruction function RE(D, W (Eq 7, denoted as RE(D, W 2 The function RE(D, W is perturbed by using functiona mechanism (FM (Zhang et a 2012, the perturbed function is denoted as RE(D, W 3 Compute W = arg min W RE(D, W 4 Private Auto-encoder (PA stacking 5 Derive and perturb the poynomia approximation of cross-entropy error C(θ (Eq 8, the perturbed function is denoted as C(θ Compute θ = arg min θ C(θ; Return θ is fuy inked to the hidden variabes of the highest hidden ayer, denoted h (k, by weighted connections W (k, where k is the number of hidden ayers in the deep auto-encoder We use the ogistic function as an activation function of ŷ, ie, ŷ = σ(w (k h (k A oss function to appropriatey dea with the binomia probem is cross-entropy error Let Y T be a set of abeed data points used to train the mode, the cross-entropy error function is given by Y T ( C(Y T, θ = y i og ŷ i + (1 y i og(1 ŷ i i=1 We can use the ayer-wise unsupervised training agorithm (Bengio et a 2007 and back-propagation to train deep auto-encoders Deep Private Auto-Encoder (dpa In this section, we formay present our dpa mode, a deep auto-encoder mode under ɛ-differentia privacy Our agorithm to construct the dpa mode incudes six main steps (Agorithm 1 We everage the functiona mechanism to enforce ɛ-differentia privacy in our dpa mode Therefore, in the first step (Agorithm 1, we derive a poynomia approximation of data reconstruction function RE(D, W (Eq 7 The poynomia approximation is denoted as RE(D, W This is a non-trivia task due to the foowing key chaenges and requirements: 1 The data reconstruction function is compicated (compared with inear or ogistic regression anaysis; and 2 The error of approximation methods must be bounded and independent of the data size The second requirement guarantees the potentia to use unabeed data in a dpa mode As shown in the next sections, our approximation method theoreticay satisfies this requirement In the second step, the functiona mechanism is used to perturb the approximation function RE(D, W, the perturbed function is denoted as RE(D, W In this step, we introduce a new resut of goba sensitivity computation for auto-encoders In the third step, we train the mode to obtain the optima perturbed parameters W by using gradient descent That resuts in private auto-encoders (PAs In the fourth step, we stack private auto-encoders (PAs to construct the deep private auto-encoders (dpas Before each stacking operation, we introduce a normaization ayer, de- (8
4 noted as h (Fig 1b, which guarantees a the data assumptions and ɛ-differentia privacy wi be satisfied when training the next hidden ayer The normaization ayer is paced on top of each private auto-encoder In the fifth step, we introduce new resuts in deriving and perturbing (ie, by using functiona mechanism the poynomia approximation of the cross-entropy error C(θ (Eq 8 in the softmax ayer for prediction tasks The perturbed function is denoted as C(θ At the sixth step, back-propagation agorithm is used to finetune a the parameters in the deep private auto-encoders In our framework, ɛ-differentia privacy is preserved in our dpa mode, since it is enforced at every ayer and training step Let us first derive the poynomia approximation form of RE(D, W, denoted RE(D, W, as foows Poynomia Approximation To expain how we appy the Tayor Expansion, and how Equations 3 and 4 are reated to RE(D, W, reca the oss function shown in Eq 7 j {1,, d}, et f 1j, f 2j, g 1j, and g 2j be four functions defined as foows: g 1j (t i, W j = W j h i, g 2j (t i, W j = W j h i (9 f 1j (z j = x ij og(1 + e zj, f 2j (z j = (1 x ij og(1 + e zj Then, we have RE(t i, W = ( (f 1j g1j (t i, W j ( + f 2j g2j (t i, W j By Equations 3 and 4, we have RE(D, W = 2 i=1 =1 R=0 f (R (z ( g (t i, W j z R where each z is a rea number By setting z = 0, the above equation can be simpified as: RE(D, W = 2 i=1 =1 R=0 f (R (0 ( Wj h i R (10 There are two compications in Eq 10 that prevent us from appying it for private data reconstruction anaysis First, the equation invoves an infinite summation Second, the term f (R (0 invoved in the equation does not have a cosed form soution We address these two issues by presenting an approximation approach that reduces the degree of the summation Our approximation approach works by truncating the Tayor series in Eq 10 to remove a poynomia terms with order arger than 2 This eads to a new objective function with ow order poynomias as foows: RE(D, W = i=1 ( 2 =1 + ( 2 =1 f (0 (0+( 2 f (2 (0 2! =1 f (1 (0 W j h i (Wj h i 2 (11 (a Auto-encoder (b PA mode (c dpa mode Figure 1: Exampes of Auto-encoder, Private Auto-encoder, Deep Private Auto-encoder Perturbation of Objective Functions In this section, we empoy the functiona mechanism to perturb the objective function RE(D, W (ie, in Eq 11 by injecting Lapace noise into its poynomia coefficients The perturbed function is denoted as RE(D, W Then we derive the mode parameter W that minimizes the perturbed function RE(D, W First of a, we need to expore the sensitivity, denoted as, of RE on D In Eq 11, 2 =1 f (0 (0, 2 =1 f (1 (0, and 2 f (2 (0 =1 2! essentiay are the poynomia coefficients of the function RE given the database D To be concise, we denote {λ (0, λ (1, λ (2 } as a set of the coefficients where t i D, λ (0 = 2 =1 f (0 (0, λ(1 = 2 =1 f (1 (0, and λ(2 = 2 f (2 (0 =1 2! We are now ready to state the foowing emma Lemma 2 Let D and D be any two neighboring databases Let RE(D, W and RE(D, W be the objective functions of auto-encoder earning on D and D respectivey, then the goba sensitivity of RE over any two neighboring databases is as foows: = 2 max t 2 R=0 λ (R jt d(b b2 (12 where b is the number of hidden variabes We eave the detaied proof of Lemma 2 in the Appendix We use gradient descent to train the perturbed mode RE(D, W Stacking Private Auto-Encoders We have presented Private Auto-Encoders (PAs, the ɛ-differentia privacy preserving agorithm for auto-encoders To construct a deep Private Auto-Encoder (dpa in which ɛ-differentia privacy is satisfied, we need to stack mutipe PAs on top of each other The hidden units of the ower ayer wi be considered as the input of the next PA (Fig 1 To guarantee that this input to b the next PA satisfies our assumption (ie, h2 ij 1 may not hod, we add a normaization ayer, denoted as h, on top of the PA Each variabe h ij can be directy computed from the h ij as foows: h ij γ j h ij = (ϕ j γ j b (13
5 where γ j and ϕ j denote the minimum and maximum vaues in the domain of H j = {h 1j,, h j } Now we have b h2 ij 1 On top of the dpa mode, we add a softmax ayer for either cassification or prediction To guarantee that privacy is preserved at every ayer, we need to deveop an ɛ-differentia privacy preserving softmax ayer Perturbation Softmax Layer We focus on a binomia prediction probem for the softmax ayer This ayer contains a singe binomia variabe ŷ to predict Y (Fig 1c A oss function to appropriatey dea with this task is the cross-entropy error, which is given in Eq 8 We can preserve the ɛ-differentia privacy for the softmax ayer by deriving a poynomia approximation of the cross-entropy error function, which wi then be perturbed by using the functiona mechanism Function C(Y T, θ in Eq 8 can be rewritten as Y T ( C(Y T, θ = y i og(1 + e W (kh i(k i=1 + (1 y i og(1 + e W (kh i(k where W (k and h i(k are the states of weights W and hidden variabes h at the k-th normaization ayer, the h i(k is derived from the tupe t i by aggregating the tupe t i through the structure of the dpa (Fig 1b In addition, the four functions in Eq 9 become: g 1 (t i, W (k = W (k h i(k, g 2 (t i, W (k = W (k h i(k f 1 (z = y i og(1 + e z, f 2 (z = (1 y i og(1 + e z The poynomia approximation of C(Y T, θ is as foows: Y T 2 2 f (R (0( R Ĉ(Y T, θ = W(k h i(k (14 i=1 =1 R=0 The goba sensitivity of the function of Ĉ(Y T, θ over Y T is given in the foowing emma Lemma 3 Let Y T and Y T be any two neighboring sets of abeed data points Let Ĉ(Y T, θ and Ĉ(Y T, θ be the objective functions of auto-encoder earning on Y T and Y T respectivey, then the goba sensitivity of Ĉ over Y T is as: C = h (k h (k 2 (15 where h (k is the number of hidden variabes in h (k We eave the proof of Lemma 3 to the Appendix After computing the sensitivity C, functiona mechanism is used to achieve the differentia privacy for the objective function Ĉ(Y T, θ The perturbed function is denoted as C(Y T, θ The back-propagation agorithm is used to train a the parameters in the deep Private Auto-Encoders (dpas that aim at optimizing the function C(Y T, θ Approximation Error Bounds The foowing emma shows the resut of how much error our approximation approaches, RE(D, W (Eq 11 and Ĉ(Y T, θ (Eq 14, incur The error ony depends on the structure of the function and the number of attributes of the dataset In addition, the average error of the approximations is aways bounded Lemma 4 Given four poynomia functions RE(D, W, RE(D, W, C(YT, θ, and Ĉ(Y T, θ, the average error of the approximations is aways bounded as foows: RE(D, Ŵ RE(D, W e2 + 2e 1 e(1 + e 2 d (1 C(Y T, θ C(Y T, θ e2 + 2e 1 e(1 + e 2 (17 where W = arg minw RE(D, W, Ŵ = arg min W RE(D, W, θ = arg minθ C(YT, θ, and θ = arg min θ Ĉ(Y T, θ We eave the detaied proof of Lemma 4 to the Appendix Lemma 4 shows that the error of the approximation of the data reconstruction in an auto-encoder is a product of a sma constant and the number of attributes d (Eq 1 It is reasonabe, since the more attributes need to be reconstructed, the more errors are injected into the resuts The error is competey independent of the number of tupes or data instances This sufficienty guarantees that our approximation of the auto-encoder can be appied in arge datasets In the experiment section, we wi show that our approximation approach eads to accurate resuts In addition, the error of the approximation of cross-entropy error function is a sma constant Therefore, the error is independent of the number of abeed data points in this supervised training step Experiments To vaidate the proposed dpa mode, we have deveoped a dpa-based human behavior prediction mode (dpah on a rea heath socia network Our starting observation is that a human behavior is the outcome of behavior determinants such as sef-motivation, socia infuences, and environmenta events This observation is rooted in human agency in socia cognitive theory (Bandura 1989 In the dpah mode, these human behavior determinants are integrated together Our Heath Socia Network Data were coected from Oct 2010 to Aug 2011 as a coaboration between Peace- Heath Laboratories, SK Teecom Americas, and the University of Oregon to record daiy physica activities, socia activities (text messages, competitions, etc, biomarkers, and biometric measures (choestero, BMI, etc for a group of 254 overweight and obese individuas Physica activities, incuding information about the number of waking and running steps, were reported via a mobie device carried by each user A users enroed in an onine socia network, aowing them to friend and communicate with each other Users biomarkers and biometric measures were recorded via daiy/weeky/monthy medica tests performed at home individuay or at our aboratories In tota, we consider three groups of attributes: 1 Behaviors: #joining competitions, #exercising days, #goas set, #goas achieved, (distances, avg(speeds; 2 #Inbox Messages: Encouragement, Fitness, Foowup, Competition, Games, Persona, Study protoco, Progress report, Technique, Socia network, Meetups, Goa, Weness meter, Feedback, Hecking, Expanation, Invitation, Notice, Technica fitness, Physica; and 3 Biomarkers and Biometric Measures: Weness Score, BMI, BMI sope, Weness Score sope
6 (a Weeky Dataset (b Daiy Dataset Figure 2: Prediction accuracy vs dataset cardinaity dpa-based Human Behavior Prediction (dpah Given a tupe t i, x i1,, x id are the persona attributes and y i is a binomia parameter which indicates whether a user increases or decreases his/her exercises To describe the dpah mode, we adjust the notations x i1 and y i a itte bit to denote the tempora dimension, and our socia network information Specificay, x ut = {x 1ut,, x dut } is used to denote the d attributes of user u at time point t Meanwhie, y ut is used to denote the status of the user u at time point t y ut = 1 denotes u increases exercises at time t, otherwise y ut = 0 The human behavior prediction task is to predict the statuses of a the users in the next time point t + 1 given M past time points t M + 1,, t Given a user u, to mode sef-motivation and socia infuences, we add an aggregation of his/her past attributes and the effects from his/her friends into the activation function of the hidden units at the current time t The hidden variabes h at time t now become h t = σ(w1 T x ut + â ut (18 where N â ut = { A T e,t kx u,t k 1 e b} + η u ψ t (v, u F u v F u k=1 In the above Equation, â ut incudes a set of dynamic biases for each hidden unit in h t, ie, h t = â ut In addition, A T e,t k is a matrix of weight which connects x u,t k with the e-th hidden variabe in h t ψ t (v, u is the probabiity that v infuences u on physica activity at time t ψ t (v, u is derived from the CPP mode (Phan et a 2014 which is an efficient socia infuence mode in heath socia networks F u is a set of friends of u in the socia network η u is a parameter which presents the abiity to observe the expicit socia infuences from neighboring users of user u N is the number of time points in the past, ie, N < M In essence, the first aggregation term is the effect of user u s past attributes on the current attributes The second aggregation term captures the effect of socia infuences from neighboring users on user u The environmenta events (ie, #competitions, #meet-up events are incuded in persona attributes Thus, the effect of environmenta events is we embedded into the mode The parameters and input of our dpah mode are W = {W 1, A, η u } and x = {x ut, k x u,t k, mode incudes two PAs and a softmax ayer for the human behavior prediction task The mode has been trained on two granuar eves of our heath socia network which v Fu ψt(v,u F u } The (a Weeky Dataset (b Daiy Dataset Figure 3: Prediction accuracy vs privacy budget ɛ are daiy and weeky datasets Both datasets contain 30 attributes, 254 users, 3,7 messages, 1,883 friend connections, 11,458 competitions, etc In addition, the daiy dataset contains 300 time points, whie the weeky dataset contains 48 time points The number of hidden units and the number of previous time intervas N are set to 200 and 3, respectivey A the earning rates are set to 10 3 Competitive Modes We compare the proposed dpah mode with three types of modes 1 Deep earning modes for human behavior prediction The CRBM (Tayor, Hinton, and Roweis 200 and SctRBM (Li et a 2014; Phan et a 2015 are competitive modes None of these modes enforces ɛ-differentia privacy 2 Deep Auto-Encoder (da and Truncated Deep Auto-Encoder (TdA da and TdA are two agorithms derived from our anaysis in Section 4 They perform anaysis but do not enforce ɛ-differentia privacy da directy outputs the mode parameters that minimize the objective functions, and TdA returns the parameters obtained from approximate objective functions with truncated poynomia terms 3 Methods for regression anaysis under ɛ-differentia privacy This experimenta evauation incudes 3 regression anaysis approaches which are under ɛ- differentia privacy, namey, Functiona Mechanism (FM (Zhang et a 2012, DPME (Lei 2011, and Fiter-Priority (FP (Cormode 2011 FM and DPME are the state-of-theart methods for regression anaysis under ɛ-differentia privacy, whie FP is an ɛ-differentiay private technique for generating synthetic data that can aso be used for regression tasks For FM, DPME and FP, we use the impementations provided by their respective authors, and we set a interna parameters to their recommended vaues Accuracy vs Dataset Cardinaity Fig 2 shows the prediction accuracy of each agorithm as a function of the dataset cardinaity In this experiment, we vary the size of M, which aso can be considered as the samping rate of the dataset ɛ is 10 in this experiment In both datasets, there is a gap between the prediction accuracy of dpah and that of da and Truncated one TdA, but the gap gets smaer rapidy with the increase of dataset cardinaity In addition, the performance of FM, FP, and DPME improves with the dataset cardinaity, which is consistent with the theoretica resuts in (Zhang et a 2012; Lei 2011; Cormode 2011 Nevertheess, even when we use a tupes in the dataset, the accuracies of FM, FP, and DPME are sti ower than that of dpah, da, and TdA With very sma samping rate, the performance of the dpah mode is sighty ower than the SctRBM, which
7 is a non-privacy-enforcing deep earning mode for human behavior prediction However, the dpah significanty is better than the SctRBM when the samping rate goes just a bit higher, ie, > 03 This is a significant resut, since 03 is a sma samping rate Accuracy vs Privacy Budget Fig 3 pots the prediction accuracy of each mode as a function of the privacy budget ɛ The prediction accuracies of da, TdA, CRBM, and SctRBM remain unchanged for a ɛ This is because none of them enforces ɛ-differentia privacy Since a smaer ɛ requires a arger amount of noise to be injected, the other four modes incur higher inaccurate prediction resuts when ɛ decreases dpah outperforms FM, FP, and DPME in a cases In addition, it is reativey robust against the change of ɛ In fact, the dpah mode is competitive even with privacy non-enforcing modes such as CRBM and SctRBM Concusions and Future Works We have deveoped an approach for differentiay private deep auto-encoders Our approach conducts both sensitivity anaysis and noise insertion on the data reconstruction and the cross-entropy error objective functions The proposed dpa mode achieves more accurate prediction resuts, when the objective functions can be represented as finite poynomias Our experiments in an appication of human behavior prediction in rea heath socia networks vaidate our theoretica resuts, and demonstrate the effectiveness of our approach Our work can be extended on the foowing directions First, our current mechanism has the potentia to be extended to different deep earning modes; since it can be appied to each ayer, and stacked Second, we pan to study whether there may exist aternative anaytica toos which can be used to better approximate the objective functions Third, it is worthy to study how we might be abe to extract private information from deep neura networks Acknowedgment This work is supported by the NIH grant R01GM to the SMASH project Wu is aso supported by NSF grant and Dou is aso supported by NSF grant We thank Xiao Xiao and Rebeca Sacks for their contributions References Aposto, T 197 Cacuus John Wiey & Sons Arfken, G 1985 In Mathematica Methods for Physicists (Third Edition Academic Press Bandura, A 1989 Human agency in socia cognitive theory The American Psychoogist Bengio, Y; Lambin, P; Popovici, D; Larochee, H; Montrea, U D; and Quebec, M 2007 Greedy ayer-wise training of deep networks In NIPS Bengio, Y 2009 Learning deep architectures for ai Found Trends Mach Learn 2(1:1 127 Chaudhuri, K, and Monteeoni, C 2008 Privacy-preserving ogistic regression In NIPS 08, Chaudhuri, K; Monteeoni, C; and Sarwate, A D 2011 Differentiay private empirica risk minimization J Mach Learn Res 12 Chicco, D; Sadowski, P; and Badi, P 2014 Deep autoencoder neura networks for gene ontoogy annotation predictions In ACM BCB 14, Cormode, G 2011 Persona privacy vs popuation privacy: earning to attack anonymization In KDD 11, Deng, L, and Yu, D 2014 Deep earning: Methods and appications Technica Report MSR-TR Dwork, C; McSherry, F; Nissim, K; and Smith, A 200 Caibrating noise to sensitivity in private data anaysis Theory of Cryptography Hinton, G; Osindero, S; and Teh, Y-W 200 A fast earning agorithm for deep beief nets Neura Comput 18(7: Hinton, G E 2002 Training products of experts by minimizing contrastive divergence Neura Comput 14(8 Kifer, D, and Machanavajjhaa, A 2011 No free unch in data privacy In SIGMOD 11, Lei, J 2011 Differentiay private m-estimators In NIPS, Li, X; Du, N; Li, H; Li, K; Gao, J; and Zhang, A 2014 A deep earning approach to ink prediction in dynamic networks In SIAM 14, McSherry, F, and Mironov, I 2009 Differentiay Private Recommender Systems In KDD 09 ACM McSherry, F, and Tawar, K 2007 Mechanism design via differentia privacy In FOCS 07, Perozzi, B; A-Rfou, R; and Skiena, S 2014 Deepwak: Onine earning of socia representations In KDD 14, Phan, N; Dou, D; Xiao, X; Piniewski, B; and Ki, D 2014 Anaysis of physica activity propagation in a heath socia network In CIKM 14, Phan, N; Dou, D; Piniewski, B; and Ki, D 2015 Socia restricted botzmann machine: Human behavior prediction in heath socia networks In ASONAM 15, Shokri, R, and Shmatikov, V 2015 Privacy-preserving deep earning In CCS 15, Smith, A 2011 Privacy-preserving statistica estimation with optima convergence rates In STOC 11, Song, Y; Ni, D; Zeng, Z; He, L; et a 2014 Automatic vagina bacteria segmentation and cassification based on superpixe and deep earning Journa of Medica Imaging and Heath Informatics 4(5: Tayor, G; Hinton, G; and Roweis, S 200 Modeing human motion using binary atent variabes In NIPS 0, Wang, Y; Wu, X; and Wu, L 2013 Differentia privacy preserving spectra graph anaysis In PAKDD (2, Xiao, X; Wang, G; and Gehrke, J 2010 Differentia privacy via waveet transforms In ICDE 10, Yuan, Z; Sang, J; Liu, Y; and Xu, C 2013 Latent feature earning in socia media network In ACM MM 13,
8 Zhang, J; Zhang, Z; Xiao, X; Yang, Y; and Winsett, M 2012 Functiona mechanism: regression anaysis under differentia privacy PVLDB 5(11: Appendices Proof of Lemma 2 Proof 1 Assume that D and D differ in the ast tupe Let t n (t n be the ast tupe in D (D Then, = 2 R=0 t i D λ (R λ (R jt t i i D = 2 R=0 λ (R λ (R jt n We can show that λ (0 = 2 =1 f (0 (0 = x nj og 2 + (1 x nj og 2 = og 2 Simiary, we can show that λ (0 jt = n og 2 As a resut, λ (0 = λ (0 jt Therefore n = 2 R=0 R=1 λ (R λ (R = jt n 2 ( λ (R + λ (R jt n 2 R=1 2 max t [ 2 max ( 1 m t 2 x tj h te + e=1 2( 1 2 d b d b2 = d(b b2 λ (R λ (R 2 jt n λ (R jt R=1 (1 8 ] h te h tq where h te is the state of e-th hidden variabe derived from the tupe t, ie, h = σ(w T x Proof of Lemma 3 Proof 2 By appying Lemma 2, we can compute the sensitivity C of Ĉ(Y T, θ on the set of abeed data points Y T as foows: C 2 max t [ h (k h (k h (k 2 e,q ( 1 2 y i + 1 ] h te(k h tq(k 8 where h te(k is the state of the e-th hidden variabe at the k-th normaization ayer, and it is derived from the tupe t h (k is the number of hidden variabes in h (k Proof of Lemma 4 Proof 3 Let W = arg minw RE(D, W and Ŵ = arg min W RE(D, W, U = maxw ( RE(D, W RE(D, W and S = min W ( RE(D, W RE(D, W We have that U RE(D, Ŵ RE(D, Ŵ and W : S RE(D, W RE(D, W Therefore, we have RE(D, Ŵ RE(D, Ŵ RE(D, W + RE(D, W U S RE(D, Ŵ RE(D, W U S + ( RE(D, Ŵ RE(D, W In addition, RE(D, Ŵ RE(D, W 0, so RE(D, Ŵ RE(D, W U S If U 0 and S 0 then we have: RE(D, Ŵ RE(D, W U S (19 e,q Eq 19 hods for every W Therefore, it sti hods for W Eq 19 shows that the error incurred by truncating the Tayor series approximate function depends on the maximum and minimum vaues of RE(D, W RE(D, W To quantify the magnitude of the error, we first rewrite RE(D, W RE(D, W as RE(D, W RE(D, W = = ( 2 i=1 =1 R=3 f (R ( RE(D, Wj RE(D, W j (z ( g (t i, W j z R To derive the minimum and maximum vaues of the function above, we ook into the remainder of Tayor Expansion for each j Let z j [z 1, z + 1] According to the we-known resut (Aposto 197, 1 ( RE(D, Wj RE(D, W j must be in the interva [ min zj f (3 (zj(zj z 3, ] max zj f (3 (zj(zj z 3 max zj f (3 If (zj(zj z 3 0 and min zj f (3 (zj(zj z 3 0, then we have that ( RE(D, 1 W RE(D, W d max zj f (3 (zj(zj z 3 min zj f (3 (zj(zj z 3 This anaysis appies to the case of an auto-encoder as foows First, for the functions f 1j (z j = x ij og(1 + e zj and f 2j (z j = (1 x ij og(1 + e zj, we have f (3 1j (z j = 2xijez j (1+e z j and f (3 3 2j (z j = (1 x ij e z j (e z j 1 (1+e z j 3 It can be verified that arg min zj f (3 1j (z j = 2e (1+e < 0, 3 arg max zj f (3 1j (z j = 2e (1+e > 0, arg min 3 zj f (3 2j (z j = 1 e e(1+e < 0, and arg max 3 zj f (3 2j (z j = e(e 1 (1+e > 0 Thus, 3 the average error of the approximation is at most RE(D, Ŵ RE(D, W [ ( 2e (1 + e 3 2e (1 + e 3 + ( e(e 1 (1 + e 3 1 e e(1 + e 3 ] d = e2 + 2e 1 e(1 + e 2 d Therefore, Eq 1 hods Eq 17 can be proved in the simiar way The main difference is that there is ony one output variabe which is ŷ in the function Ĉ(Y T, θ Meanwhie, there are d output variabes (ie, x = d in the data reconstruction function RE(D, W Thus, by repacing the functions RE(D, W and RE(D, W with C(Y T, θ and Ĉ(Y T, θ in the proof of Eq 1, we obtain that C(Y T, θ C(Y T, θ [ ( 2e (1 + e 3 2e (1 + e 3 + ( e(e 1 (1 + e 3 1 e e(1 + e 3 ] = e2 + 2e 1 e(1 + e 2
arxiv: v1 [cs.db] 1 Aug 2012
Functiona Mechanism: Regression Anaysis under Differentia Privacy arxiv:208.029v [cs.db] Aug 202 Jun Zhang Zhenjie Zhang 2 Xiaokui Xiao Yin Yang 2 Marianne Winsett 2,3 ABSTRACT Schoo of Computer Engineering
More informationStochastic Variational Inference with Gradient Linearization
Stochastic Variationa Inference with Gradient Linearization Suppementa Materia Tobias Pötz * Anne S Wannenwetsch Stefan Roth Department of Computer Science, TU Darmstadt Preface In this suppementa materia,
More informationBayesian Learning. You hear a which which could equally be Thanks or Tanks, which would you go with?
Bayesian Learning A powerfu and growing approach in machine earning We use it in our own decision making a the time You hear a which which coud equay be Thanks or Tanks, which woud you go with? Combine
More informationStatistical Learning Theory: A Primer
Internationa Journa of Computer Vision 38(), 9 3, 2000 c 2000 uwer Academic Pubishers. Manufactured in The Netherands. Statistica Learning Theory: A Primer THEODOROS EVGENIOU, MASSIMILIANO PONTIL AND TOMASO
More informationBP neural network-based sports performance prediction model applied research
Avaiabe onine www.jocpr.com Journa of Chemica and Pharmaceutica Research, 204, 6(7:93-936 Research Artice ISSN : 0975-7384 CODEN(USA : JCPRC5 BP neura networ-based sports performance prediction mode appied
More informationCS229 Lecture notes. Andrew Ng
CS229 Lecture notes Andrew Ng Part IX The EM agorithm In the previous set of notes, we taked about the EM agorithm as appied to fitting a mixture of Gaussians. In this set of notes, we give a broader view
More informationDo Schools Matter for High Math Achievement? Evidence from the American Mathematics Competitions Glenn Ellison and Ashley Swanson Online Appendix
VOL. NO. DO SCHOOLS MATTER FOR HIGH MATH ACHIEVEMENT? 43 Do Schoos Matter for High Math Achievement? Evidence from the American Mathematics Competitions Genn Eison and Ashey Swanson Onine Appendix Appendix
More informationExplicit overall risk minimization transductive bound
1 Expicit overa risk minimization transductive bound Sergio Decherchi, Paoo Gastado, Sandro Ridea, Rodofo Zunino Dept. of Biophysica and Eectronic Engineering (DIBE), Genoa University Via Opera Pia 11a,
More informationA. Distribution of the test statistic
A. Distribution of the test statistic In the sequentia test, we first compute the test statistic from a mini-batch of size m. If a decision cannot be made with this statistic, we keep increasing the mini-batch
More information(This is a sample cover image for this issue. The actual cover is not yet available at this time.)
(This is a sampe cover image for this issue The actua cover is not yet avaiabe at this time) This artice appeared in a journa pubished by Esevier The attached copy is furnished to the author for interna
More informationA Brief Introduction to Markov Chains and Hidden Markov Models
A Brief Introduction to Markov Chains and Hidden Markov Modes Aen B MacKenzie Notes for December 1, 3, &8, 2015 Discrete-Time Markov Chains You may reca that when we first introduced random processes,
More informationHow the backpropagation algorithm works Srikumar Ramalingam School of Computing University of Utah
How the backpropagation agorithm works Srikumar Ramaingam Schoo of Computing University of Utah Reference Most of the sides are taken from the second chapter of the onine book by Michae Nieson: neuranetworksanddeepearning.com
More information8 APPENDIX. E[m M] = (n S )(1 exp( exp(s min + c M))) (19) E[m M] n exp(s min + c M) (20) 8.1 EMPIRICAL EVALUATION OF SAMPLING
8 APPENDIX 8.1 EMPIRICAL EVALUATION OF SAMPLING We wish to evauate the empirica accuracy of our samping technique on concrete exampes. We do this in two ways. First, we can sort the eements by probabiity
More informationAn Algorithm for Pruning Redundant Modules in Min-Max Modular Network
An Agorithm for Pruning Redundant Modues in Min-Max Moduar Network Hui-Cheng Lian and Bao-Liang Lu Department of Computer Science and Engineering, Shanghai Jiao Tong University 1954 Hua Shan Rd., Shanghai
More informationSeparation of Variables and a Spherical Shell with Surface Charge
Separation of Variabes and a Spherica She with Surface Charge In cass we worked out the eectrostatic potentia due to a spherica she of radius R with a surface charge density σθ = σ cos θ. This cacuation
More informationMoreau-Yosida Regularization for Grouped Tree Structure Learning
Moreau-Yosida Reguarization for Grouped Tree Structure Learning Jun Liu Computer Science and Engineering Arizona State University J.Liu@asu.edu Jieping Ye Computer Science and Engineering Arizona State
More informationFirst-Order Corrections to Gutzwiller s Trace Formula for Systems with Discrete Symmetries
c 26 Noninear Phenomena in Compex Systems First-Order Corrections to Gutzwier s Trace Formua for Systems with Discrete Symmetries Hoger Cartarius, Jörg Main, and Günter Wunner Institut für Theoretische
More informationMultilayer Kerceptron
Mutiayer Kerceptron Zotán Szabó, András Lőrincz Department of Information Systems, Facuty of Informatics Eötvös Loránd University Pázmány Péter sétány 1/C H-1117, Budapest, Hungary e-mai: szzoi@csetehu,
More informationThe EM Algorithm applied to determining new limit points of Mahler measures
Contro and Cybernetics vo. 39 (2010) No. 4 The EM Agorithm appied to determining new imit points of Maher measures by Souad E Otmani, Georges Rhin and Jean-Marc Sac-Épée Université Pau Veraine-Metz, LMAM,
More informationA Better Way to Pretrain Deep Boltzmann Machines
A Better Way to Pretrain Deep Botzmann Machines Rusan Saakhutdino Department of Statistics and Computer Science Uniersity of Toronto rsaakhu@cs.toronto.edu Geoffrey Hinton Department of Computer Science
More information$, (2.1) n="# #. (2.2)
Chapter. Eectrostatic II Notes: Most of the materia presented in this chapter is taken from Jackson, Chap.,, and 4, and Di Bartoo, Chap... Mathematica Considerations.. The Fourier series and the Fourier
More informationASummaryofGaussianProcesses Coryn A.L. Bailer-Jones
ASummaryofGaussianProcesses Coryn A.L. Baier-Jones Cavendish Laboratory University of Cambridge caj@mrao.cam.ac.uk Introduction A genera prediction probem can be posed as foows. We consider that the variabe
More informationMATH 172: MOTIVATION FOR FOURIER SERIES: SEPARATION OF VARIABLES
MATH 172: MOTIVATION FOR FOURIER SERIES: SEPARATION OF VARIABLES Separation of variabes is a method to sove certain PDEs which have a warped product structure. First, on R n, a inear PDE of order m is
More informationA Novel Learning Method for Elman Neural Network Using Local Search
Neura Information Processing Letters and Reviews Vo. 11, No. 8, August 2007 LETTER A Nove Learning Method for Eman Neura Networ Using Loca Search Facuty of Engineering, Toyama University, Gofuu 3190 Toyama
More informationStatistical Learning Theory: a Primer
??,??, 1 6 (??) c?? Kuwer Academic Pubishers, Boston. Manufactured in The Netherands. Statistica Learning Theory: a Primer THEODOROS EVGENIOU AND MASSIMILIANO PONTIL Center for Bioogica and Computationa
More informationExpectation-Maximization for Estimating Parameters for a Mixture of Poissons
Expectation-Maximization for Estimating Parameters for a Mixture of Poissons Brandon Maone Department of Computer Science University of Hesini February 18, 2014 Abstract This document derives, in excrutiating
More informationLecture Note 3: Stationary Iterative Methods
MATH 5330: Computationa Methods of Linear Agebra Lecture Note 3: Stationary Iterative Methods Xianyi Zeng Department of Mathematica Sciences, UTEP Stationary Iterative Methods The Gaussian eimination (or
More informationAsymptotic Properties of a Generalized Cross Entropy Optimization Algorithm
1 Asymptotic Properties of a Generaized Cross Entropy Optimization Agorithm Zijun Wu, Michae Koonko, Institute for Appied Stochastics and Operations Research, Caustha Technica University Abstract The discrete
More informationChapter 7 PRODUCTION FUNCTIONS. Copyright 2005 by South-Western, a division of Thomson Learning. All rights reserved.
Chapter 7 PRODUCTION FUNCTIONS Copyright 2005 by South-Western, a division of Thomson Learning. A rights reserved. 1 Production Function The firm s production function for a particuar good (q) shows the
More informationAsynchronous Control for Coupled Markov Decision Systems
INFORMATION THEORY WORKSHOP (ITW) 22 Asynchronous Contro for Couped Marov Decision Systems Michae J. Neey University of Southern Caifornia Abstract This paper considers optima contro for a coection of
More informationActive Learning & Experimental Design
Active Learning & Experimenta Design Danie Ting Heaviy modified, of course, by Lye Ungar Origina Sides by Barbara Engehardt and Aex Shyr Lye Ungar, University of Pennsyvania Motivation u Data coection
More informationCryptanalysis of PKP: A New Approach
Cryptanaysis of PKP: A New Approach Éiane Jaumes and Antoine Joux DCSSI 18, rue du Dr. Zamenhoff F-92131 Issy-es-Mx Cedex France eiane.jaumes@wanadoo.fr Antoine.Joux@ens.fr Abstract. Quite recenty, in
More informationMARKOV CHAINS AND MARKOV DECISION THEORY. Contents
MARKOV CHAINS AND MARKOV DECISION THEORY ARINDRIMA DATTA Abstract. In this paper, we begin with a forma introduction to probabiity and expain the concept of random variabes and stochastic processes. After
More informationOnline Appendices for The Economics of Nationalism (Xiaohuan Lan and Ben Li)
Onine Appendices for The Economics of Nationaism Xiaohuan Lan and Ben Li) A. Derivation of inequaities 9) and 10) Consider Home without oss of generaity. Denote gobaized and ungobaized by g and ng, respectivey.
More informationFitting Algorithms for MMPP ATM Traffic Models
Fitting Agorithms for PP AT Traffic odes A. Nogueira, P. Savador, R. Vaadas University of Aveiro / Institute of Teecommunications, 38-93 Aveiro, Portuga; e-mai: (nogueira, savador, rv)@av.it.pt ABSTRACT
More informationConvolutional Networks 2: Training, deep convolutional networks
Convoutiona Networks 2: Training, deep convoutiona networks Hakan Bien Machine Learning Practica MLP Lecture 8 30 October / 6 November 2018 MLP Lecture 8 / 30 October / 6 November 2018 Convoutiona Networks
More informationNOISE-INDUCED STABILIZATION OF STOCHASTIC DIFFERENTIAL EQUATIONS
NOISE-INDUCED STABILIZATION OF STOCHASTIC DIFFERENTIAL EQUATIONS TONY ALLEN, EMILY GEBHARDT, AND ADAM KLUBALL 3 ADVISOR: DR. TIFFANY KOLBA 4 Abstract. The phenomenon of noise-induced stabiization occurs
More informationMelodic contour estimation with B-spline models using a MDL criterion
Meodic contour estimation with B-spine modes using a MDL criterion Damien Loive, Ney Barbot, Oivier Boeffard IRISA / University of Rennes 1 - ENSSAT 6 rue de Kerampont, B.P. 80518, F-305 Lannion Cedex
More informationRelated Topics Maxwell s equations, electrical eddy field, magnetic field of coils, coil, magnetic flux, induced voltage
Magnetic induction TEP Reated Topics Maxwe s equations, eectrica eddy fied, magnetic fied of cois, coi, magnetic fux, induced votage Principe A magnetic fied of variabe frequency and varying strength is
More information17 Lecture 17: Recombination and Dark Matter Production
PYS 652: Astrophysics 88 17 Lecture 17: Recombination and Dark Matter Production New ideas pass through three periods: It can t be done. It probaby can be done, but it s not worth doing. I knew it was
More informationSupplement of Limited-memory Common-directions Method for Distributed Optimization and its Application on Empirical Risk Minimization
Suppement of Limited-memory Common-directions Method for Distributed Optimization and its Appication on Empirica Risk Minimization Ching-pei Lee Po-Wei Wang Weizhu Chen Chih-Jen Lin I Introduction In this
More informationarxiv: v1 [math.ca] 6 Mar 2017
Indefinite Integras of Spherica Besse Functions MIT-CTP/487 arxiv:703.0648v [math.ca] 6 Mar 07 Joyon K. Boomfied,, Stephen H. P. Face,, and Zander Moss, Center for Theoretica Physics, Laboratory for Nucear
More informationSTA 216 Project: Spline Approach to Discrete Survival Analysis
: Spine Approach to Discrete Surviva Anaysis November 4, 005 1 Introduction Athough continuous surviva anaysis differs much from the discrete surviva anaysis, there is certain ink between the two modeing
More informationNotes on Backpropagation with Cross Entropy
Notes on Backpropagation with Cross Entropy I-Ta ee, Dan Gowasser, Bruno Ribeiro Purue University October 3, 07. Overview This note introuces backpropagation for a common neura network muti-cass cassifier.
More informationFrom Margins to Probabilities in Multiclass Learning Problems
From Margins to Probabiities in Muticass Learning Probems Andrea Passerini and Massimiiano Ponti 2 and Paoo Frasconi 3 Abstract. We study the probem of muticass cassification within the framework of error
More informationHow the backpropagation algorithm works Srikumar Ramalingam School of Computing University of Utah
How the backpropagation agorithm works Srikumar Ramaingam Schoo of Computing University of Utah Reference Most of the sides are taken from the second chapter of the onine book by Michae Nieson: neuranetworksanddeepearning.com
More informationAppendix for Stochastic Gradient Monomial Gamma Sampler
3 4 5 6 7 8 9 3 4 5 6 7 8 9 3 4 5 6 7 8 9 3 3 3 33 34 35 36 37 38 39 4 4 4 43 44 45 46 47 48 49 5 5 5 53 54 Appendix for Stochastic Gradient Monomia Gamma Samper A The Main Theorem We provide the foowing
More informationTheory of Generalized k-difference Operator and Its Application in Number Theory
Internationa Journa of Mathematica Anaysis Vo. 9, 2015, no. 19, 955-964 HIKARI Ltd, www.m-hiari.com http://dx.doi.org/10.12988/ijma.2015.5389 Theory of Generaized -Difference Operator and Its Appication
More informationA SIMPLIFIED DESIGN OF MULTIDIMENSIONAL TRANSFER FUNCTION MODELS
A SIPLIFIED DESIGN OF ULTIDIENSIONAL TRANSFER FUNCTION ODELS Stefan Petrausch, Rudof Rabenstein utimedia Communications and Signa Procesg, University of Erangen-Nuremberg, Cauerstr. 7, 958 Erangen, GERANY
More informationDetermining The Degree of Generalization Using An Incremental Learning Algorithm
Determining The Degree of Generaization Using An Incrementa Learning Agorithm Pabo Zegers Facutad de Ingeniería, Universidad de os Andes San Caros de Apoquindo 22, Las Condes, Santiago, Chie pzegers@uandes.c
More information14 Separation of Variables Method
14 Separation of Variabes Method Consider, for exampe, the Dirichet probem u t = Du xx < x u(x, ) = f(x) < x < u(, t) = = u(, t) t > Let u(x, t) = T (t)φ(x); now substitute into the equation: dt
More informationA Fundamental Storage-Communication Tradeoff in Distributed Computing with Straggling Nodes
A Fundamenta Storage-Communication Tradeoff in Distributed Computing with Stragging odes ifa Yan, Michèe Wigger LTCI, Téécom ParisTech 75013 Paris, France Emai: {qifa.yan, michee.wigger} @teecom-paristech.fr
More informationAST 418/518 Instrumentation and Statistics
AST 418/518 Instrumentation and Statistics Cass Website: http://ircamera.as.arizona.edu/astr_518 Cass Texts: Practica Statistics for Astronomers, J.V. Wa, and C.R. Jenkins, Second Edition. Measuring the
More informationLearning Fully Observed Undirected Graphical Models
Learning Fuy Observed Undirected Graphica Modes Sides Credit: Matt Gormey (2016) Kayhan Batmangheich 1 Machine Learning The data inspires the structures we want to predict Inference finds {best structure,
More informationAppendix for Stochastic Gradient Monomial Gamma Sampler
Appendix for Stochastic Gradient Monomia Gamma Samper A The Main Theorem We provide the foowing theorem to characterize the stationary distribution of the stochastic process with SDEs in (3) Theorem 3
More informationFORECASTING TELECOMMUNICATIONS DATA WITH AUTOREGRESSIVE INTEGRATED MOVING AVERAGE MODELS
FORECASTING TEECOMMUNICATIONS DATA WITH AUTOREGRESSIVE INTEGRATED MOVING AVERAGE MODES Niesh Subhash naawade a, Mrs. Meenakshi Pawar b a SVERI's Coege of Engineering, Pandharpur. nieshsubhash15@gmai.com
More informationUniprocessor Feasibility of Sporadic Tasks with Constrained Deadlines is Strongly conp-complete
Uniprocessor Feasibiity of Sporadic Tasks with Constrained Deadines is Strongy conp-compete Pontus Ekberg and Wang Yi Uppsaa University, Sweden Emai: {pontus.ekberg yi}@it.uu.se Abstract Deciding the feasibiity
More informationHaar Decomposition and Reconstruction Algorithms
Jim Lambers MAT 773 Fa Semester 018-19 Lecture 15 and 16 Notes These notes correspond to Sections 4.3 and 4.4 in the text. Haar Decomposition and Reconstruction Agorithms Decomposition Suppose we approximate
More informationSmoothness equivalence properties of univariate subdivision schemes and their projection analogues
Numerische Mathematik manuscript No. (wi be inserted by the editor) Smoothness equivaence properties of univariate subdivision schemes and their projection anaogues Phiipp Grohs TU Graz Institute of Geometry
More informationA Solution to the 4-bit Parity Problem with a Single Quaternary Neuron
Neura Information Processing - Letters and Reviews Vo. 5, No. 2, November 2004 LETTER A Soution to the 4-bit Parity Probem with a Singe Quaternary Neuron Tohru Nitta Nationa Institute of Advanced Industria
More informationSVM-based Supervised and Unsupervised Classification Schemes
SVM-based Supervised and Unsupervised Cassification Schemes LUMINITA STATE University of Pitesti Facuty of Mathematics and Computer Science 1 Targu din Vae St., Pitesti 110040 ROMANIA state@cicknet.ro
More informationCentralized Coded Caching of Correlated Contents
Centraized Coded Caching of Correated Contents Qianqian Yang and Deniz Gündüz Information Processing and Communications Lab Department of Eectrica and Eectronic Engineering Imperia Coege London arxiv:1711.03798v1
More informationII. PROBLEM. A. Description. For the space of audio signals
CS229 - Fina Report Speech Recording based Language Recognition (Natura Language) Leopod Cambier - cambier; Matan Leibovich - matane; Cindy Orozco Bohorquez - orozcocc ABSTRACT We construct a rea time
More informationAn approximate method for solving the inverse scattering problem with fixed-energy data
J. Inv. I-Posed Probems, Vo. 7, No. 6, pp. 561 571 (1999) c VSP 1999 An approximate method for soving the inverse scattering probem with fixed-energy data A. G. Ramm and W. Scheid Received May 12, 1999
More information1. Measurements and error calculus
EV 1 Measurements and error cacuus 11 Introduction The goa of this aboratory course is to introduce the notions of carrying out an experiment, acquiring and writing up the data, and finay anayzing the
More informationApproximated MLC shape matrix decomposition with interleaf collision constraint
Approximated MLC shape matrix decomposition with intereaf coision constraint Thomas Kainowski Antje Kiese Abstract Shape matrix decomposition is a subprobem in radiation therapy panning. A given fuence
More informationAlgorithms to solve massively under-defined systems of multivariate quadratic equations
Agorithms to sove massivey under-defined systems of mutivariate quadratic equations Yasufumi Hashimoto Abstract It is we known that the probem to sove a set of randomy chosen mutivariate quadratic equations
More informationUnconditional security of differential phase shift quantum key distribution
Unconditiona security of differentia phase shift quantum key distribution Kai Wen, Yoshihisa Yamamoto Ginzton Lab and Dept of Eectrica Engineering Stanford University Basic idea of DPS-QKD Protoco. Aice
More informationSchedulability Analysis of Deferrable Scheduling Algorithms for Maintaining Real-Time Data Freshness
1 Scheduabiity Anaysis of Deferrabe Scheduing Agorithms for Maintaining Rea-Time Data Freshness Song Han, Deji Chen, Ming Xiong, Kam-yiu Lam, Aoysius K. Mok, Krithi Ramamritham UT Austin, Emerson Process
More informationParagraph Topic Classification
Paragraph Topic Cassification Eugene Nho Graduate Schoo of Business Stanford University Stanford, CA 94305 enho@stanford.edu Edward Ng Department of Eectrica Engineering Stanford University Stanford, CA
More informationChemical Kinetics Part 2
Integrated Rate Laws Chemica Kinetics Part 2 The rate aw we have discussed thus far is the differentia rate aw. Let us consider the very simpe reaction: a A à products The differentia rate reates the rate
More informationSupport Vector Machine and Its Application to Regression and Classification
BearWorks Institutiona Repository MSU Graduate Theses Spring 2017 Support Vector Machine and Its Appication to Regression and Cassification Xiaotong Hu As with any inteectua project, the content and views
More informationOptimality of Inference in Hierarchical Coding for Distributed Object-Based Representations
Optimaity of Inference in Hierarchica Coding for Distributed Object-Based Representations Simon Brodeur, Jean Rouat NECOTIS, Département génie éectrique et génie informatique, Université de Sherbrooke,
More informationConsistent linguistic fuzzy preference relation with multi-granular uncertain linguistic information for solving decision making problems
Consistent inguistic fuzzy preference reation with muti-granuar uncertain inguistic information for soving decision making probems Siti mnah Binti Mohd Ridzuan, and Daud Mohamad Citation: IP Conference
More informationA Comparison Study of the Test for Right Censored and Grouped Data
Communications for Statistica Appications and Methods 2015, Vo. 22, No. 4, 313 320 DOI: http://dx.doi.org/10.5351/csam.2015.22.4.313 Print ISSN 2287-7843 / Onine ISSN 2383-4757 A Comparison Study of the
More informationProceedings of Meetings on Acoustics
Proceedings of Meetings on Acoustics Voume 19, 213 http://acousticasociety.org/ ICA 213 Montrea Montrea, Canada 2-7 June 213 Underwater Acoustics Session 1pUW: Seabed Scattering: Measurements and Mechanisms
More informationMath 124B January 31, 2012
Math 124B January 31, 212 Viktor Grigoryan 7 Inhomogeneous boundary vaue probems Having studied the theory of Fourier series, with which we successfuy soved boundary vaue probems for the homogeneous heat
More informationSoft Clustering on Graphs
Soft Custering on Graphs Kai Yu 1, Shipeng Yu 2, Voker Tresp 1 1 Siemens AG, Corporate Technoogy 2 Institute for Computer Science, University of Munich kai.yu@siemens.com, voker.tresp@siemens.com spyu@dbs.informatik.uni-muenchen.de
More informationBayesian Unscented Kalman Filter for State Estimation of Nonlinear and Non-Gaussian Systems
Bayesian Unscented Kaman Fiter for State Estimation of Noninear and Non-aussian Systems Zhong Liu, Shing-Chow Chan, Ho-Chun Wu and iafei Wu Department of Eectrica and Eectronic Engineering, he University
More informationResearch of Data Fusion Method of Multi-Sensor Based on Correlation Coefficient of Confidence Distance
Send Orders for Reprints to reprints@benthamscience.ae 340 The Open Cybernetics & Systemics Journa, 015, 9, 340-344 Open Access Research of Data Fusion Method of Muti-Sensor Based on Correation Coefficient
More informationCONJUGATE GRADIENT WITH SUBSPACE OPTIMIZATION
CONJUGATE GRADIENT WITH SUBSPACE OPTIMIZATION SAHAR KARIMI AND STEPHEN VAVASIS Abstract. In this paper we present a variant of the conjugate gradient (CG) agorithm in which we invoke a subspace minimization
More informationReichenbachian Common Cause Systems
Reichenbachian Common Cause Systems G. Hofer-Szabó Department of Phiosophy Technica University of Budapest e-mai: gszabo@hps.ete.hu Mikós Rédei Department of History and Phiosophy of Science Eötvös University,
More informationSection 6: Magnetostatics
agnetic fieds in matter Section 6: agnetostatics In the previous sections we assumed that the current density J is a known function of coordinates. In the presence of matter this is not aways true. The
More informationAppendix of the Paper The Role of No-Arbitrage on Forecasting: Lessons from a Parametric Term Structure Model
Appendix of the Paper The Roe of No-Arbitrage on Forecasting: Lessons from a Parametric Term Structure Mode Caio Ameida cameida@fgv.br José Vicente jose.vaentim@bcb.gov.br June 008 1 Introduction In this
More informationChemical Kinetics Part 2. Chapter 16
Chemica Kinetics Part 2 Chapter 16 Integrated Rate Laws The rate aw we have discussed thus far is the differentia rate aw. Let us consider the very simpe reaction: a A à products The differentia rate reates
More informationSome Measures for Asymmetry of Distributions
Some Measures for Asymmetry of Distributions Georgi N. Boshnakov First version: 31 January 2006 Research Report No. 5, 2006, Probabiity and Statistics Group Schoo of Mathematics, The University of Manchester
More informationOn the Goal Value of a Boolean Function
On the Goa Vaue of a Booean Function Eric Bach Dept. of CS University of Wisconsin 1210 W. Dayton St. Madison, WI 53706 Lisa Heerstein Dept of CSE NYU Schoo of Engineering 2 Metrotech Center, 10th Foor
More informationMat 1501 lecture notes, penultimate installment
Mat 1501 ecture notes, penutimate instament 1. bounded variation: functions of a singe variabe optiona) I beieve that we wi not actuay use the materia in this section the point is mainy to motivate the
More informationVALIDATED CONTINUATION FOR EQUILIBRIA OF PDES
SIAM J. NUMER. ANAL. Vo. 0, No. 0, pp. 000 000 c 200X Society for Industria and Appied Mathematics VALIDATED CONTINUATION FOR EQUILIBRIA OF PDES SARAH DAY, JEAN-PHILIPPE LESSARD, AND KONSTANTIN MISCHAIKOW
More informationT.C. Banwell, S. Galli. {bct, Telcordia Technologies, Inc., 445 South Street, Morristown, NJ 07960, USA
ON THE SYMMETRY OF THE POWER INE CHANNE T.C. Banwe, S. Gai {bct, sgai}@research.tecordia.com Tecordia Technoogies, Inc., 445 South Street, Morristown, NJ 07960, USA Abstract The indoor power ine network
More informationPartial permutation decoding for MacDonald codes
Partia permutation decoding for MacDonad codes J.D. Key Department of Mathematics and Appied Mathematics University of the Western Cape 7535 Bevie, South Africa P. Seneviratne Department of Mathematics
More informationSUPPLEMENTARY MATERIAL TO INNOVATED SCALABLE EFFICIENT ESTIMATION IN ULTRA-LARGE GAUSSIAN GRAPHICAL MODELS
ISEE 1 SUPPLEMENTARY MATERIAL TO INNOVATED SCALABLE EFFICIENT ESTIMATION IN ULTRA-LARGE GAUSSIAN GRAPHICAL MODELS By Yingying Fan and Jinchi Lv University of Southern Caifornia This Suppementary Materia
More informationResearch Article On the Lower Bound for the Number of Real Roots of a Random Algebraic Equation
Appied Mathematics and Stochastic Anaysis Voume 007, Artice ID 74191, 8 pages doi:10.1155/007/74191 Research Artice On the Lower Bound for the Number of Rea Roots of a Random Agebraic Equation Takashi
More informationFRST Multivariate Statistics. Multivariate Discriminant Analysis (MDA)
1 FRST 531 -- Mutivariate Statistics Mutivariate Discriminant Anaysis (MDA) Purpose: 1. To predict which group (Y) an observation beongs to based on the characteristics of p predictor (X) variabes, using
More informationC. Fourier Sine Series Overview
12 PHILIP D. LOEWEN C. Fourier Sine Series Overview Let some constant > be given. The symboic form of the FSS Eigenvaue probem combines an ordinary differentia equation (ODE) on the interva (, ) with a
More informationProceedings of Meetings on Acoustics
Proceedings of Meetings on Acoustics Voume 9, 23 http://acousticasociety.org/ ICA 23 Montrea Montrea, Canada 2-7 June 23 Architectura Acoustics Session 4pAAa: Room Acoustics Computer Simuation II 4pAAa9.
More informationGeneral Decay of Solutions in a Viscoelastic Equation with Nonlinear Localized Damping
Journa of Mathematica Research with Appications Jan.,, Vo. 3, No., pp. 53 6 DOI:.377/j.issn:95-65...7 Http://jmre.dut.edu.cn Genera Decay of Soutions in a Viscoeastic Equation with Noninear Locaized Damping
More informationFormulas for Angular-Momentum Barrier Factors Version II
BNL PREPRINT BNL-QGS-06-101 brfactor1.tex Formuas for Anguar-Momentum Barrier Factors Version II S. U. Chung Physics Department, Brookhaven Nationa Laboratory, Upton, NY 11973 March 19, 2015 abstract A
More informationVALIDATED CONTINUATION FOR EQUILIBRIA OF PDES
VALIDATED CONTINUATION FOR EQUILIBRIA OF PDES SARAH DAY, JEAN-PHILIPPE LESSARD, AND KONSTANTIN MISCHAIKOW Abstract. One of the most efficient methods for determining the equiibria of a continuous parameterized
More informationTwo view learning: SVM-2K, Theory and Practice
Two view earning: SVM-2K, Theory and Practice Jason D.R. Farquhar jdrf99r@ecs.soton.ac.uk Hongying Meng hongying@cs.york.ac.uk David R. Hardoon drh@ecs.soton.ac.uk John Shawe-Tayor jst@ecs.soton.ac.uk
More information