Differential Privacy Preservation for Deep Auto-Encoders: an Application of Human Behavior Prediction

Size: px
Start display at page:

Download "Differential Privacy Preservation for Deep Auto-Encoders: an Application of Human Behavior Prediction"

Transcription

1 Differentia Privacy Preservation for Deep Auto-Encoders: an Appication of Human Behavior Prediction NhatHai Phan 1, Yue Wang 2, Xintao Wu 3, Dejing Dou 1 {haiphan,dou}@csuoregonedu, ywang91@unccedu, xintaowu@uarkedu 1 University of Oregon, USA; 2 University of North Caroina at Charotte, USA; 3 University of Arkansas, USA Abstract In recent years, deep earning has spread beyond both academia and industry with many exciting rea-word appications The deveopment of deep earning has presented obvious privacy issues However, there has been ack of scientific study about privacy preservation in deep earning In this paper, we concentrate on the autoencoder, a fundamenta component in deep earning, and propose the deep private auto-encoder (dpa Our main idea is to enforce ɛ-differentia privacy by perturbing the objective functions of the traditiona deep autoencoder, rather than its resuts We appy the dpa to human behavior prediction in a heath socia network Theoretica anaysis and thorough experimenta evauations show that the dpa is highy effective and efficient, and it significanty outperforms existing soutions Introduction Deep earning is a timey and promising area of machine earning research In genera, deep earning is about earning mutipe eves of representation and abstraction that hep to make sense of data such as images, sound, and text (Deng and Yu 2014 With significant success in recent years, the appications of deep earning are being expanded into other fieds such as socia media (Yuan et a 2013, socia network anaysis (Perozzi, A-Rfou, and Skiena 2014, bioinformatics (Chicco, Sadowski, and Badi 2014, medicine and heathcare (Song et a 2014 This presents obvious issues about protecting privacy in deep earning modes when they are buit on users persona and highy sensitive data, such as cinica records, user profies, photo, etc The present research efforts ack sufficient deep earning techniques that incorporate privacy concerns Therefore, deveoping a privacy preserving deep earning mode is an urgent demand Reeasing sensitive resuts of statistica anaysis and data mining whie protecting privacy has been studied in the past few decades One state-of-the-art approach to the probem is ɛ-differentia privacy (Dwork et a 200, which works by injecting random noise into the reeased statistica resuts computed from the underying sensitive data, such that the distribution of the noisy resuts is reativey insensitive to Copyright c 201, Association for the Advancement of Artificia Inteigence (wwwaaaiorg A rights reserved any change of a singe record in the origina dataset This ensures that the adversary cannot infer any information about any particuar record with high confidence (controed by parameter ɛ, even if the adversary possesses a the remaining tupes of the sensitive data In this paper, we propose to deveop an ɛ-differentia privacy preserving deep earning mode The structure of deep earning modes, eg, deep auto-encoders (Bengio 2009, deep beief networks (Hinton, Osindero, and Teh 200, usuay contains mutipe ayers of neurons At each ayer, they use different objective functions (eg, cross-entropy error, energy based functions and agorithms (eg, contrastive divergence (Hinton 2002, backpropagation to earn the optima parameters for output tasks such as cassification and prediction In addition, the designs of deep earning modes are varied and dependent on appication domains It is difficut, therefore, to design a unified ɛ-differentia privacy soution that covers a deep earning modes Deep auto-encoders (das (Bengio 2009 are one of the fundamenta deep earning modes which have been used in many appications such as heathcare and medicine, natura anguage processing, etc (Deng and Yu 2014 In this paper, we aim at deriving an ɛ-differentia privacy approach to deep auto-encoders (das for a binomia cassification task There are severa potentia soutions for ɛ-differentia privacy that target regression, which is simiar to the data reconstruction and cross-entropy error functions in deep autoencoders (Zhang et a 2012; Chaudhuri and Monteeoni 2008; Chaudhuri, Monteeoni, and Sarwate 2011; Lei 2011; Smith 2011 The most pertinent agorithm that we can adapt in our work is functiona mechanism (FM (Zhang et a 2012 The FM enforces ɛ-differentia privacy by perturbing the objective function of the optimization probem, eg, ogistic regression, rather than its resuts By everaging the functiona mechanism to perturb the objective functions in deep auto-encoders so that the ɛ-differentia privacy is preserved, we propose a nove ɛ-differentia Private Autoencoder (PA First, the cross-entropy error functions of the data reconstruction and softmax ayer are approximated to poynomia forms by using Tayor Expansion (Arfken 1985 Second, we inject noise into these poynomia forms so that the ɛ-differentia privacy is satisfied in the training phases Third, in the PA, we add a normaization ayer on top of the hidden ayer to protect the ɛ-differentia privacy when stacking mutipe PAs Mutipe PAs can be stacked on each other

2 to produce a deep private auto-encoder (dpa To evauate its performance, we appy the dpa mode for human behavior prediction in a rea heath socia network Conducted experimenta resuts demonstrate that the dpa mode achieves accurate resuts with prediction power comparabe to the unperturbed resuts, and it outperforms the existing soutions Preiminaries and Reated Works In this section, we briefy revisit the differentia privacy definition, Functiona Mechanism (Zhang et a 2012, and formay introduce the deep auto-encoder Let D be a database that contains n tupes t 1, t 2,, t n and d+1 attributes X 1, X 2,, X d, Y For each tupe t i = (x i1, x i2,, x id, y i, we assume without oss of generaity d x2 ij 1 where x ij 0, y i foows a binomia distribution Our objective is to construct a deep auto-encoder ρ from D that (i takes x i = (x i1, x i2,, x id as input and (ii outputs a prediction of y i that is as accurate as possibe The mode function ρ contains a mode parameter vector ω To evauate whether ω eads to an accurate mode, a cost function f D (ω is often used to measure the difference between the origina and predicted vaues of y i As the reeased mode parameter ω may discose sensitive information of D, to protect the privacy, we require that the mode training shoud be performed with an agorithm which satisfies ɛ-differentia privacy Differentia Privacy Differentia privacy research has been significanty studied from the theoretica perspective, eg, (Chaudhuri and Monteeoni 2008; Kifer and Machanavajjhaa 2011 There are aso extensive studies on the appicabiity of enforcing differentia privacy in some particuar anaysis tasks, eg, coaborative recommendation (McSherry and Mironov 2009, ogistic regression (Chaudhuri and Monteeoni 2008, pubishing contingency tabes (Xiao, Wang, and Gehrke 2010, and spectra graph anaysis (Wang, Wu, and Wu 2013 in socia network anaysis The mechanisms of achieving differentia privacy mainy incude the cassic approach of adding Lapacian noise (Dwork et a 200, the exponentia mechanism (McSherry and Tawar 2007, and the functiona perturbation approach (Chaudhuri and Monteeoni 2008 Definition 1 (ɛ-different Privacy (Dwork et a 200 A randomized agorithm A fufis ɛ-differentia privacy, iff for any two databases D and D differing at most one tupe, and for any output O of A, we have: P r(a(d = O e ɛ P r(a(d = O (1 The privacy budget ɛ contros the amount by which the distributions induced by two neighboring data sets may differ Smaer vaues of ɛ enforce a stronger privacy guarantee of A A genera method for computing an approximation to any function f (on D whie preserving ɛ-differentia privacy is the Lapace mechanism (Dwork et a 200, where the output of f is a vector of rea numbers In particuar, the mechanism expoits the goba sensitivity of f over any two neighboring data sets (differing at most one record, which is denoted as S f (D Given S f (D, the Lapace mechanism ensures ɛ-differentia privacy by injecting noise η into each vaue in the output of f(d where η is drawn iid from Lapace distribution with zero mean and scae S f (D/ɛ In (Shokri and Shmatikov 2015, Shokri et a proposed a distributed training method to preserve privacy in neura networks They extended their method to preserve differentia privacy for distributed deep neura networks Their method aims at preserving epsion-differentia privacy for each training epoch of each participant Even their method is reasonabe given the context of distributed deep neura networks; their approach may consume too much privacy budget to ensure mode accuracy when the number of training epochs, the number of participants, and the number of shared parameters are arge Functiona Mechanism Revisited Functiona mechanism (Zhang et a 2012 is an extension of the Lapace mechanism It achieves ɛ-differentia privacy by perturbing the objective function f D (ω and then reeasing the mode parameter ω that minimizes the perturbed objective function f D (ω instead of the origina one The functiona mechanism expoits the poynomia representation of f D (ω The mode parameter ω is a vector that contains d vaues ω 1,, ω d Let φ(ω denote a product of ω 1,, ω d, namey, φ(ω = ω c1 1 ωc2 2 ωc d d for some c 1,, c d N Let Φ j (j N denote the set of a products of ω 1,, ω d with degree j, ie, Φ j = { ω c1 1 ωc2 2 ωc d d d =1 c = j } By the Stone-Weierstrass Theorem, any continuous and differentiabe f(t i, ω can aways be written as a poynomia of ω 1,, ω d, for some J [0, ], ie, f(t i, ω = J j=0 φ Φ j λ φti φ(ω where λ φti R denotes the coefficient of φ(ω in the poynomia For instance, the poynomia expression of the oss function in the inear regression is as foows: f(x i, ω = (y i x T i ω2 = yi 2 d (2y ix ij ω j + 1 j, d (x ijx i ω j ω We can see that it ony invoves monomias in Φ 0 = {1}, Φ 1 = {ω 1,, ω d }, and Φ 2 = {ω i ω j i, j [1, d]} Each φ(ω has its own coefficient, eg, for ω j, its poynomia coefficient λ φti = 2y i x ij Simiary, f D (ω can aso be expressed as a poynomia of ω 1,, ω d f D (ω = J λ φti φ(ω (2 j=0 φ Φ j t i D Lemma 1 (Zhang et a 2012 Let D and D be any two neighboring datasets Let f D (ω and f D (ω be the objective functions of regression anaysis on D and D, respectivey The foowing inequaity hods = J 2 φ Φ j t i D λ φti t i D λ φt i where t i, t i or t is an arbitrary tupe 2 max 1 t J φ Φ j λ φt 1 To achieve ɛ-differentia privacy, f D (ω is perturbed by injecting Lapace noise Lap( ɛ into its poynomia coefficients λ φ, and then the mode parameter ω is derived

3 to minimize the perturbed function f D (ω, where = J 2 max t φ Φ j λ φt 1, according to the Lemma 1 When the poynomia form of an objective function (eg, ogistic regression contains terms with unbounded degrees, Zhang et a (Zhang et a 2012 deveoped an approximation poynomia form based on Tayor Expansion (Arfken 1985 For instance, given the cost function f(t i, ω of regression anaysis, assume that there exist 2m functions f 1,, f m and g 1,, g m such that f(t i, ω = =1 f (g (t i, ω, and each g (t i, ω is a poynomia function of ω, where m N is the number of functions f and g Given the above decomposition of f(t i, ω, we can appy the Tayor Expansion (Arfken 1985 on each f ( to obtain the foowing equation: f(t i, ω = m =1 R=0 (z ( R g (t i, ω z (3 f (R where each z is a rea number Thus, the objective function f(d, ω can be written as a poynomia function, ie, f D (ω = m i=1 =1 R=0 (z ( R g (t i, ω z (4 f (R Deep Auto-Encoders The deep auto-encoder (Bengio 2009 is one of the fundamenta modes in deep earning (Deng and Yu 2014 An auto-encoder has a ayer of input units, and a ayer of data reconstruction units fuy connected to a ayer of hidden units, but no connections within a ayer (Fig 1a An auto-encoder is trained to encode the input in some representation x i = {x i1,, x id } so that the input can be reconstructed from that representation The auto-encoder prefers to minimize the negative og-ikeihood of the reconstruction, given the encoding x i, RE(t i, W = og P (x i x i, W ( = xij og x ij + (1 x ij og(1 x ij (5 where W is a weight matrix (ie, from now on W wi be used to denote the parameters instead of ω, x = σ(hw, h = σ(w T x ( σ( is the sigmoid function The oss function on the whoe dataset D is the summation over a the data points x: = i=1 RE(D, W = RE(x i, W = (7 i=1 [ ] x ij og(1+e Wjhi +(1 x ij og(1+e Wjhi In the foowing, we bur the distinction between x i and t i, which means t i indicates the set of attributes in x i t i and x i wi be used interchangeaby We can stack mutipe auto-encoders to produce a deep auto-encoder On top of the deep auto-encoder, we add an output ayer which incudes a singe binomia variabe to predict Y The output variabe ŷ Agorithm 1: Pseudo Code of a dpa mode 1 Derive poynomia approximation of data reconstruction function RE(D, W (Eq 7, denoted as RE(D, W 2 The function RE(D, W is perturbed by using functiona mechanism (FM (Zhang et a 2012, the perturbed function is denoted as RE(D, W 3 Compute W = arg min W RE(D, W 4 Private Auto-encoder (PA stacking 5 Derive and perturb the poynomia approximation of cross-entropy error C(θ (Eq 8, the perturbed function is denoted as C(θ Compute θ = arg min θ C(θ; Return θ is fuy inked to the hidden variabes of the highest hidden ayer, denoted h (k, by weighted connections W (k, where k is the number of hidden ayers in the deep auto-encoder We use the ogistic function as an activation function of ŷ, ie, ŷ = σ(w (k h (k A oss function to appropriatey dea with the binomia probem is cross-entropy error Let Y T be a set of abeed data points used to train the mode, the cross-entropy error function is given by Y T ( C(Y T, θ = y i og ŷ i + (1 y i og(1 ŷ i i=1 We can use the ayer-wise unsupervised training agorithm (Bengio et a 2007 and back-propagation to train deep auto-encoders Deep Private Auto-Encoder (dpa In this section, we formay present our dpa mode, a deep auto-encoder mode under ɛ-differentia privacy Our agorithm to construct the dpa mode incudes six main steps (Agorithm 1 We everage the functiona mechanism to enforce ɛ-differentia privacy in our dpa mode Therefore, in the first step (Agorithm 1, we derive a poynomia approximation of data reconstruction function RE(D, W (Eq 7 The poynomia approximation is denoted as RE(D, W This is a non-trivia task due to the foowing key chaenges and requirements: 1 The data reconstruction function is compicated (compared with inear or ogistic regression anaysis; and 2 The error of approximation methods must be bounded and independent of the data size The second requirement guarantees the potentia to use unabeed data in a dpa mode As shown in the next sections, our approximation method theoreticay satisfies this requirement In the second step, the functiona mechanism is used to perturb the approximation function RE(D, W, the perturbed function is denoted as RE(D, W In this step, we introduce a new resut of goba sensitivity computation for auto-encoders In the third step, we train the mode to obtain the optima perturbed parameters W by using gradient descent That resuts in private auto-encoders (PAs In the fourth step, we stack private auto-encoders (PAs to construct the deep private auto-encoders (dpas Before each stacking operation, we introduce a normaization ayer, de- (8

4 noted as h (Fig 1b, which guarantees a the data assumptions and ɛ-differentia privacy wi be satisfied when training the next hidden ayer The normaization ayer is paced on top of each private auto-encoder In the fifth step, we introduce new resuts in deriving and perturbing (ie, by using functiona mechanism the poynomia approximation of the cross-entropy error C(θ (Eq 8 in the softmax ayer for prediction tasks The perturbed function is denoted as C(θ At the sixth step, back-propagation agorithm is used to finetune a the parameters in the deep private auto-encoders In our framework, ɛ-differentia privacy is preserved in our dpa mode, since it is enforced at every ayer and training step Let us first derive the poynomia approximation form of RE(D, W, denoted RE(D, W, as foows Poynomia Approximation To expain how we appy the Tayor Expansion, and how Equations 3 and 4 are reated to RE(D, W, reca the oss function shown in Eq 7 j {1,, d}, et f 1j, f 2j, g 1j, and g 2j be four functions defined as foows: g 1j (t i, W j = W j h i, g 2j (t i, W j = W j h i (9 f 1j (z j = x ij og(1 + e zj, f 2j (z j = (1 x ij og(1 + e zj Then, we have RE(t i, W = ( (f 1j g1j (t i, W j ( + f 2j g2j (t i, W j By Equations 3 and 4, we have RE(D, W = 2 i=1 =1 R=0 f (R (z ( g (t i, W j z R where each z is a rea number By setting z = 0, the above equation can be simpified as: RE(D, W = 2 i=1 =1 R=0 f (R (0 ( Wj h i R (10 There are two compications in Eq 10 that prevent us from appying it for private data reconstruction anaysis First, the equation invoves an infinite summation Second, the term f (R (0 invoved in the equation does not have a cosed form soution We address these two issues by presenting an approximation approach that reduces the degree of the summation Our approximation approach works by truncating the Tayor series in Eq 10 to remove a poynomia terms with order arger than 2 This eads to a new objective function with ow order poynomias as foows: RE(D, W = i=1 ( 2 =1 + ( 2 =1 f (0 (0+( 2 f (2 (0 2! =1 f (1 (0 W j h i (Wj h i 2 (11 (a Auto-encoder (b PA mode (c dpa mode Figure 1: Exampes of Auto-encoder, Private Auto-encoder, Deep Private Auto-encoder Perturbation of Objective Functions In this section, we empoy the functiona mechanism to perturb the objective function RE(D, W (ie, in Eq 11 by injecting Lapace noise into its poynomia coefficients The perturbed function is denoted as RE(D, W Then we derive the mode parameter W that minimizes the perturbed function RE(D, W First of a, we need to expore the sensitivity, denoted as, of RE on D In Eq 11, 2 =1 f (0 (0, 2 =1 f (1 (0, and 2 f (2 (0 =1 2! essentiay are the poynomia coefficients of the function RE given the database D To be concise, we denote {λ (0, λ (1, λ (2 } as a set of the coefficients where t i D, λ (0 = 2 =1 f (0 (0, λ(1 = 2 =1 f (1 (0, and λ(2 = 2 f (2 (0 =1 2! We are now ready to state the foowing emma Lemma 2 Let D and D be any two neighboring databases Let RE(D, W and RE(D, W be the objective functions of auto-encoder earning on D and D respectivey, then the goba sensitivity of RE over any two neighboring databases is as foows: = 2 max t 2 R=0 λ (R jt d(b b2 (12 where b is the number of hidden variabes We eave the detaied proof of Lemma 2 in the Appendix We use gradient descent to train the perturbed mode RE(D, W Stacking Private Auto-Encoders We have presented Private Auto-Encoders (PAs, the ɛ-differentia privacy preserving agorithm for auto-encoders To construct a deep Private Auto-Encoder (dpa in which ɛ-differentia privacy is satisfied, we need to stack mutipe PAs on top of each other The hidden units of the ower ayer wi be considered as the input of the next PA (Fig 1 To guarantee that this input to b the next PA satisfies our assumption (ie, h2 ij 1 may not hod, we add a normaization ayer, denoted as h, on top of the PA Each variabe h ij can be directy computed from the h ij as foows: h ij γ j h ij = (ϕ j γ j b (13

5 where γ j and ϕ j denote the minimum and maximum vaues in the domain of H j = {h 1j,, h j } Now we have b h2 ij 1 On top of the dpa mode, we add a softmax ayer for either cassification or prediction To guarantee that privacy is preserved at every ayer, we need to deveop an ɛ-differentia privacy preserving softmax ayer Perturbation Softmax Layer We focus on a binomia prediction probem for the softmax ayer This ayer contains a singe binomia variabe ŷ to predict Y (Fig 1c A oss function to appropriatey dea with this task is the cross-entropy error, which is given in Eq 8 We can preserve the ɛ-differentia privacy for the softmax ayer by deriving a poynomia approximation of the cross-entropy error function, which wi then be perturbed by using the functiona mechanism Function C(Y T, θ in Eq 8 can be rewritten as Y T ( C(Y T, θ = y i og(1 + e W (kh i(k i=1 + (1 y i og(1 + e W (kh i(k where W (k and h i(k are the states of weights W and hidden variabes h at the k-th normaization ayer, the h i(k is derived from the tupe t i by aggregating the tupe t i through the structure of the dpa (Fig 1b In addition, the four functions in Eq 9 become: g 1 (t i, W (k = W (k h i(k, g 2 (t i, W (k = W (k h i(k f 1 (z = y i og(1 + e z, f 2 (z = (1 y i og(1 + e z The poynomia approximation of C(Y T, θ is as foows: Y T 2 2 f (R (0( R Ĉ(Y T, θ = W(k h i(k (14 i=1 =1 R=0 The goba sensitivity of the function of Ĉ(Y T, θ over Y T is given in the foowing emma Lemma 3 Let Y T and Y T be any two neighboring sets of abeed data points Let Ĉ(Y T, θ and Ĉ(Y T, θ be the objective functions of auto-encoder earning on Y T and Y T respectivey, then the goba sensitivity of Ĉ over Y T is as: C = h (k h (k 2 (15 where h (k is the number of hidden variabes in h (k We eave the proof of Lemma 3 to the Appendix After computing the sensitivity C, functiona mechanism is used to achieve the differentia privacy for the objective function Ĉ(Y T, θ The perturbed function is denoted as C(Y T, θ The back-propagation agorithm is used to train a the parameters in the deep Private Auto-Encoders (dpas that aim at optimizing the function C(Y T, θ Approximation Error Bounds The foowing emma shows the resut of how much error our approximation approaches, RE(D, W (Eq 11 and Ĉ(Y T, θ (Eq 14, incur The error ony depends on the structure of the function and the number of attributes of the dataset In addition, the average error of the approximations is aways bounded Lemma 4 Given four poynomia functions RE(D, W, RE(D, W, C(YT, θ, and Ĉ(Y T, θ, the average error of the approximations is aways bounded as foows: RE(D, Ŵ RE(D, W e2 + 2e 1 e(1 + e 2 d (1 C(Y T, θ C(Y T, θ e2 + 2e 1 e(1 + e 2 (17 where W = arg minw RE(D, W, Ŵ = arg min W RE(D, W, θ = arg minθ C(YT, θ, and θ = arg min θ Ĉ(Y T, θ We eave the detaied proof of Lemma 4 to the Appendix Lemma 4 shows that the error of the approximation of the data reconstruction in an auto-encoder is a product of a sma constant and the number of attributes d (Eq 1 It is reasonabe, since the more attributes need to be reconstructed, the more errors are injected into the resuts The error is competey independent of the number of tupes or data instances This sufficienty guarantees that our approximation of the auto-encoder can be appied in arge datasets In the experiment section, we wi show that our approximation approach eads to accurate resuts In addition, the error of the approximation of cross-entropy error function is a sma constant Therefore, the error is independent of the number of abeed data points in this supervised training step Experiments To vaidate the proposed dpa mode, we have deveoped a dpa-based human behavior prediction mode (dpah on a rea heath socia network Our starting observation is that a human behavior is the outcome of behavior determinants such as sef-motivation, socia infuences, and environmenta events This observation is rooted in human agency in socia cognitive theory (Bandura 1989 In the dpah mode, these human behavior determinants are integrated together Our Heath Socia Network Data were coected from Oct 2010 to Aug 2011 as a coaboration between Peace- Heath Laboratories, SK Teecom Americas, and the University of Oregon to record daiy physica activities, socia activities (text messages, competitions, etc, biomarkers, and biometric measures (choestero, BMI, etc for a group of 254 overweight and obese individuas Physica activities, incuding information about the number of waking and running steps, were reported via a mobie device carried by each user A users enroed in an onine socia network, aowing them to friend and communicate with each other Users biomarkers and biometric measures were recorded via daiy/weeky/monthy medica tests performed at home individuay or at our aboratories In tota, we consider three groups of attributes: 1 Behaviors: #joining competitions, #exercising days, #goas set, #goas achieved, (distances, avg(speeds; 2 #Inbox Messages: Encouragement, Fitness, Foowup, Competition, Games, Persona, Study protoco, Progress report, Technique, Socia network, Meetups, Goa, Weness meter, Feedback, Hecking, Expanation, Invitation, Notice, Technica fitness, Physica; and 3 Biomarkers and Biometric Measures: Weness Score, BMI, BMI sope, Weness Score sope

6 (a Weeky Dataset (b Daiy Dataset Figure 2: Prediction accuracy vs dataset cardinaity dpa-based Human Behavior Prediction (dpah Given a tupe t i, x i1,, x id are the persona attributes and y i is a binomia parameter which indicates whether a user increases or decreases his/her exercises To describe the dpah mode, we adjust the notations x i1 and y i a itte bit to denote the tempora dimension, and our socia network information Specificay, x ut = {x 1ut,, x dut } is used to denote the d attributes of user u at time point t Meanwhie, y ut is used to denote the status of the user u at time point t y ut = 1 denotes u increases exercises at time t, otherwise y ut = 0 The human behavior prediction task is to predict the statuses of a the users in the next time point t + 1 given M past time points t M + 1,, t Given a user u, to mode sef-motivation and socia infuences, we add an aggregation of his/her past attributes and the effects from his/her friends into the activation function of the hidden units at the current time t The hidden variabes h at time t now become h t = σ(w1 T x ut + â ut (18 where N â ut = { A T e,t kx u,t k 1 e b} + η u ψ t (v, u F u v F u k=1 In the above Equation, â ut incudes a set of dynamic biases for each hidden unit in h t, ie, h t = â ut In addition, A T e,t k is a matrix of weight which connects x u,t k with the e-th hidden variabe in h t ψ t (v, u is the probabiity that v infuences u on physica activity at time t ψ t (v, u is derived from the CPP mode (Phan et a 2014 which is an efficient socia infuence mode in heath socia networks F u is a set of friends of u in the socia network η u is a parameter which presents the abiity to observe the expicit socia infuences from neighboring users of user u N is the number of time points in the past, ie, N < M In essence, the first aggregation term is the effect of user u s past attributes on the current attributes The second aggregation term captures the effect of socia infuences from neighboring users on user u The environmenta events (ie, #competitions, #meet-up events are incuded in persona attributes Thus, the effect of environmenta events is we embedded into the mode The parameters and input of our dpah mode are W = {W 1, A, η u } and x = {x ut, k x u,t k, mode incudes two PAs and a softmax ayer for the human behavior prediction task The mode has been trained on two granuar eves of our heath socia network which v Fu ψt(v,u F u } The (a Weeky Dataset (b Daiy Dataset Figure 3: Prediction accuracy vs privacy budget ɛ are daiy and weeky datasets Both datasets contain 30 attributes, 254 users, 3,7 messages, 1,883 friend connections, 11,458 competitions, etc In addition, the daiy dataset contains 300 time points, whie the weeky dataset contains 48 time points The number of hidden units and the number of previous time intervas N are set to 200 and 3, respectivey A the earning rates are set to 10 3 Competitive Modes We compare the proposed dpah mode with three types of modes 1 Deep earning modes for human behavior prediction The CRBM (Tayor, Hinton, and Roweis 200 and SctRBM (Li et a 2014; Phan et a 2015 are competitive modes None of these modes enforces ɛ-differentia privacy 2 Deep Auto-Encoder (da and Truncated Deep Auto-Encoder (TdA da and TdA are two agorithms derived from our anaysis in Section 4 They perform anaysis but do not enforce ɛ-differentia privacy da directy outputs the mode parameters that minimize the objective functions, and TdA returns the parameters obtained from approximate objective functions with truncated poynomia terms 3 Methods for regression anaysis under ɛ-differentia privacy This experimenta evauation incudes 3 regression anaysis approaches which are under ɛ- differentia privacy, namey, Functiona Mechanism (FM (Zhang et a 2012, DPME (Lei 2011, and Fiter-Priority (FP (Cormode 2011 FM and DPME are the state-of-theart methods for regression anaysis under ɛ-differentia privacy, whie FP is an ɛ-differentiay private technique for generating synthetic data that can aso be used for regression tasks For FM, DPME and FP, we use the impementations provided by their respective authors, and we set a interna parameters to their recommended vaues Accuracy vs Dataset Cardinaity Fig 2 shows the prediction accuracy of each agorithm as a function of the dataset cardinaity In this experiment, we vary the size of M, which aso can be considered as the samping rate of the dataset ɛ is 10 in this experiment In both datasets, there is a gap between the prediction accuracy of dpah and that of da and Truncated one TdA, but the gap gets smaer rapidy with the increase of dataset cardinaity In addition, the performance of FM, FP, and DPME improves with the dataset cardinaity, which is consistent with the theoretica resuts in (Zhang et a 2012; Lei 2011; Cormode 2011 Nevertheess, even when we use a tupes in the dataset, the accuracies of FM, FP, and DPME are sti ower than that of dpah, da, and TdA With very sma samping rate, the performance of the dpah mode is sighty ower than the SctRBM, which

7 is a non-privacy-enforcing deep earning mode for human behavior prediction However, the dpah significanty is better than the SctRBM when the samping rate goes just a bit higher, ie, > 03 This is a significant resut, since 03 is a sma samping rate Accuracy vs Privacy Budget Fig 3 pots the prediction accuracy of each mode as a function of the privacy budget ɛ The prediction accuracies of da, TdA, CRBM, and SctRBM remain unchanged for a ɛ This is because none of them enforces ɛ-differentia privacy Since a smaer ɛ requires a arger amount of noise to be injected, the other four modes incur higher inaccurate prediction resuts when ɛ decreases dpah outperforms FM, FP, and DPME in a cases In addition, it is reativey robust against the change of ɛ In fact, the dpah mode is competitive even with privacy non-enforcing modes such as CRBM and SctRBM Concusions and Future Works We have deveoped an approach for differentiay private deep auto-encoders Our approach conducts both sensitivity anaysis and noise insertion on the data reconstruction and the cross-entropy error objective functions The proposed dpa mode achieves more accurate prediction resuts, when the objective functions can be represented as finite poynomias Our experiments in an appication of human behavior prediction in rea heath socia networks vaidate our theoretica resuts, and demonstrate the effectiveness of our approach Our work can be extended on the foowing directions First, our current mechanism has the potentia to be extended to different deep earning modes; since it can be appied to each ayer, and stacked Second, we pan to study whether there may exist aternative anaytica toos which can be used to better approximate the objective functions Third, it is worthy to study how we might be abe to extract private information from deep neura networks Acknowedgment This work is supported by the NIH grant R01GM to the SMASH project Wu is aso supported by NSF grant and Dou is aso supported by NSF grant We thank Xiao Xiao and Rebeca Sacks for their contributions References Aposto, T 197 Cacuus John Wiey & Sons Arfken, G 1985 In Mathematica Methods for Physicists (Third Edition Academic Press Bandura, A 1989 Human agency in socia cognitive theory The American Psychoogist Bengio, Y; Lambin, P; Popovici, D; Larochee, H; Montrea, U D; and Quebec, M 2007 Greedy ayer-wise training of deep networks In NIPS Bengio, Y 2009 Learning deep architectures for ai Found Trends Mach Learn 2(1:1 127 Chaudhuri, K, and Monteeoni, C 2008 Privacy-preserving ogistic regression In NIPS 08, Chaudhuri, K; Monteeoni, C; and Sarwate, A D 2011 Differentiay private empirica risk minimization J Mach Learn Res 12 Chicco, D; Sadowski, P; and Badi, P 2014 Deep autoencoder neura networks for gene ontoogy annotation predictions In ACM BCB 14, Cormode, G 2011 Persona privacy vs popuation privacy: earning to attack anonymization In KDD 11, Deng, L, and Yu, D 2014 Deep earning: Methods and appications Technica Report MSR-TR Dwork, C; McSherry, F; Nissim, K; and Smith, A 200 Caibrating noise to sensitivity in private data anaysis Theory of Cryptography Hinton, G; Osindero, S; and Teh, Y-W 200 A fast earning agorithm for deep beief nets Neura Comput 18(7: Hinton, G E 2002 Training products of experts by minimizing contrastive divergence Neura Comput 14(8 Kifer, D, and Machanavajjhaa, A 2011 No free unch in data privacy In SIGMOD 11, Lei, J 2011 Differentiay private m-estimators In NIPS, Li, X; Du, N; Li, H; Li, K; Gao, J; and Zhang, A 2014 A deep earning approach to ink prediction in dynamic networks In SIAM 14, McSherry, F, and Mironov, I 2009 Differentiay Private Recommender Systems In KDD 09 ACM McSherry, F, and Tawar, K 2007 Mechanism design via differentia privacy In FOCS 07, Perozzi, B; A-Rfou, R; and Skiena, S 2014 Deepwak: Onine earning of socia representations In KDD 14, Phan, N; Dou, D; Xiao, X; Piniewski, B; and Ki, D 2014 Anaysis of physica activity propagation in a heath socia network In CIKM 14, Phan, N; Dou, D; Piniewski, B; and Ki, D 2015 Socia restricted botzmann machine: Human behavior prediction in heath socia networks In ASONAM 15, Shokri, R, and Shmatikov, V 2015 Privacy-preserving deep earning In CCS 15, Smith, A 2011 Privacy-preserving statistica estimation with optima convergence rates In STOC 11, Song, Y; Ni, D; Zeng, Z; He, L; et a 2014 Automatic vagina bacteria segmentation and cassification based on superpixe and deep earning Journa of Medica Imaging and Heath Informatics 4(5: Tayor, G; Hinton, G; and Roweis, S 200 Modeing human motion using binary atent variabes In NIPS 0, Wang, Y; Wu, X; and Wu, L 2013 Differentia privacy preserving spectra graph anaysis In PAKDD (2, Xiao, X; Wang, G; and Gehrke, J 2010 Differentia privacy via waveet transforms In ICDE 10, Yuan, Z; Sang, J; Liu, Y; and Xu, C 2013 Latent feature earning in socia media network In ACM MM 13,

8 Zhang, J; Zhang, Z; Xiao, X; Yang, Y; and Winsett, M 2012 Functiona mechanism: regression anaysis under differentia privacy PVLDB 5(11: Appendices Proof of Lemma 2 Proof 1 Assume that D and D differ in the ast tupe Let t n (t n be the ast tupe in D (D Then, = 2 R=0 t i D λ (R λ (R jt t i i D = 2 R=0 λ (R λ (R jt n We can show that λ (0 = 2 =1 f (0 (0 = x nj og 2 + (1 x nj og 2 = og 2 Simiary, we can show that λ (0 jt = n og 2 As a resut, λ (0 = λ (0 jt Therefore n = 2 R=0 R=1 λ (R λ (R = jt n 2 ( λ (R + λ (R jt n 2 R=1 2 max t [ 2 max ( 1 m t 2 x tj h te + e=1 2( 1 2 d b d b2 = d(b b2 λ (R λ (R 2 jt n λ (R jt R=1 (1 8 ] h te h tq where h te is the state of e-th hidden variabe derived from the tupe t, ie, h = σ(w T x Proof of Lemma 3 Proof 2 By appying Lemma 2, we can compute the sensitivity C of Ĉ(Y T, θ on the set of abeed data points Y T as foows: C 2 max t [ h (k h (k h (k 2 e,q ( 1 2 y i + 1 ] h te(k h tq(k 8 where h te(k is the state of the e-th hidden variabe at the k-th normaization ayer, and it is derived from the tupe t h (k is the number of hidden variabes in h (k Proof of Lemma 4 Proof 3 Let W = arg minw RE(D, W and Ŵ = arg min W RE(D, W, U = maxw ( RE(D, W RE(D, W and S = min W ( RE(D, W RE(D, W We have that U RE(D, Ŵ RE(D, Ŵ and W : S RE(D, W RE(D, W Therefore, we have RE(D, Ŵ RE(D, Ŵ RE(D, W + RE(D, W U S RE(D, Ŵ RE(D, W U S + ( RE(D, Ŵ RE(D, W In addition, RE(D, Ŵ RE(D, W 0, so RE(D, Ŵ RE(D, W U S If U 0 and S 0 then we have: RE(D, Ŵ RE(D, W U S (19 e,q Eq 19 hods for every W Therefore, it sti hods for W Eq 19 shows that the error incurred by truncating the Tayor series approximate function depends on the maximum and minimum vaues of RE(D, W RE(D, W To quantify the magnitude of the error, we first rewrite RE(D, W RE(D, W as RE(D, W RE(D, W = = ( 2 i=1 =1 R=3 f (R ( RE(D, Wj RE(D, W j (z ( g (t i, W j z R To derive the minimum and maximum vaues of the function above, we ook into the remainder of Tayor Expansion for each j Let z j [z 1, z + 1] According to the we-known resut (Aposto 197, 1 ( RE(D, Wj RE(D, W j must be in the interva [ min zj f (3 (zj(zj z 3, ] max zj f (3 (zj(zj z 3 max zj f (3 If (zj(zj z 3 0 and min zj f (3 (zj(zj z 3 0, then we have that ( RE(D, 1 W RE(D, W d max zj f (3 (zj(zj z 3 min zj f (3 (zj(zj z 3 This anaysis appies to the case of an auto-encoder as foows First, for the functions f 1j (z j = x ij og(1 + e zj and f 2j (z j = (1 x ij og(1 + e zj, we have f (3 1j (z j = 2xijez j (1+e z j and f (3 3 2j (z j = (1 x ij e z j (e z j 1 (1+e z j 3 It can be verified that arg min zj f (3 1j (z j = 2e (1+e < 0, 3 arg max zj f (3 1j (z j = 2e (1+e > 0, arg min 3 zj f (3 2j (z j = 1 e e(1+e < 0, and arg max 3 zj f (3 2j (z j = e(e 1 (1+e > 0 Thus, 3 the average error of the approximation is at most RE(D, Ŵ RE(D, W [ ( 2e (1 + e 3 2e (1 + e 3 + ( e(e 1 (1 + e 3 1 e e(1 + e 3 ] d = e2 + 2e 1 e(1 + e 2 d Therefore, Eq 1 hods Eq 17 can be proved in the simiar way The main difference is that there is ony one output variabe which is ŷ in the function Ĉ(Y T, θ Meanwhie, there are d output variabes (ie, x = d in the data reconstruction function RE(D, W Thus, by repacing the functions RE(D, W and RE(D, W with C(Y T, θ and Ĉ(Y T, θ in the proof of Eq 1, we obtain that C(Y T, θ C(Y T, θ [ ( 2e (1 + e 3 2e (1 + e 3 + ( e(e 1 (1 + e 3 1 e e(1 + e 3 ] = e2 + 2e 1 e(1 + e 2

arxiv: v1 [cs.db] 1 Aug 2012

arxiv: v1 [cs.db] 1 Aug 2012 Functiona Mechanism: Regression Anaysis under Differentia Privacy arxiv:208.029v [cs.db] Aug 202 Jun Zhang Zhenjie Zhang 2 Xiaokui Xiao Yin Yang 2 Marianne Winsett 2,3 ABSTRACT Schoo of Computer Engineering

More information

Stochastic Variational Inference with Gradient Linearization

Stochastic Variational Inference with Gradient Linearization Stochastic Variationa Inference with Gradient Linearization Suppementa Materia Tobias Pötz * Anne S Wannenwetsch Stefan Roth Department of Computer Science, TU Darmstadt Preface In this suppementa materia,

More information

Bayesian Learning. You hear a which which could equally be Thanks or Tanks, which would you go with?

Bayesian Learning. You hear a which which could equally be Thanks or Tanks, which would you go with? Bayesian Learning A powerfu and growing approach in machine earning We use it in our own decision making a the time You hear a which which coud equay be Thanks or Tanks, which woud you go with? Combine

More information

Statistical Learning Theory: A Primer

Statistical Learning Theory: A Primer Internationa Journa of Computer Vision 38(), 9 3, 2000 c 2000 uwer Academic Pubishers. Manufactured in The Netherands. Statistica Learning Theory: A Primer THEODOROS EVGENIOU, MASSIMILIANO PONTIL AND TOMASO

More information

BP neural network-based sports performance prediction model applied research

BP neural network-based sports performance prediction model applied research Avaiabe onine www.jocpr.com Journa of Chemica and Pharmaceutica Research, 204, 6(7:93-936 Research Artice ISSN : 0975-7384 CODEN(USA : JCPRC5 BP neura networ-based sports performance prediction mode appied

More information

CS229 Lecture notes. Andrew Ng

CS229 Lecture notes. Andrew Ng CS229 Lecture notes Andrew Ng Part IX The EM agorithm In the previous set of notes, we taked about the EM agorithm as appied to fitting a mixture of Gaussians. In this set of notes, we give a broader view

More information

Do Schools Matter for High Math Achievement? Evidence from the American Mathematics Competitions Glenn Ellison and Ashley Swanson Online Appendix

Do Schools Matter for High Math Achievement? Evidence from the American Mathematics Competitions Glenn Ellison and Ashley Swanson Online Appendix VOL. NO. DO SCHOOLS MATTER FOR HIGH MATH ACHIEVEMENT? 43 Do Schoos Matter for High Math Achievement? Evidence from the American Mathematics Competitions Genn Eison and Ashey Swanson Onine Appendix Appendix

More information

Explicit overall risk minimization transductive bound

Explicit overall risk minimization transductive bound 1 Expicit overa risk minimization transductive bound Sergio Decherchi, Paoo Gastado, Sandro Ridea, Rodofo Zunino Dept. of Biophysica and Eectronic Engineering (DIBE), Genoa University Via Opera Pia 11a,

More information

A. Distribution of the test statistic

A. Distribution of the test statistic A. Distribution of the test statistic In the sequentia test, we first compute the test statistic from a mini-batch of size m. If a decision cannot be made with this statistic, we keep increasing the mini-batch

More information

(This is a sample cover image for this issue. The actual cover is not yet available at this time.)

(This is a sample cover image for this issue. The actual cover is not yet available at this time.) (This is a sampe cover image for this issue The actua cover is not yet avaiabe at this time) This artice appeared in a journa pubished by Esevier The attached copy is furnished to the author for interna

More information

A Brief Introduction to Markov Chains and Hidden Markov Models

A Brief Introduction to Markov Chains and Hidden Markov Models A Brief Introduction to Markov Chains and Hidden Markov Modes Aen B MacKenzie Notes for December 1, 3, &8, 2015 Discrete-Time Markov Chains You may reca that when we first introduced random processes,

More information

How the backpropagation algorithm works Srikumar Ramalingam School of Computing University of Utah

How the backpropagation algorithm works Srikumar Ramalingam School of Computing University of Utah How the backpropagation agorithm works Srikumar Ramaingam Schoo of Computing University of Utah Reference Most of the sides are taken from the second chapter of the onine book by Michae Nieson: neuranetworksanddeepearning.com

More information

8 APPENDIX. E[m M] = (n S )(1 exp( exp(s min + c M))) (19) E[m M] n exp(s min + c M) (20) 8.1 EMPIRICAL EVALUATION OF SAMPLING

8 APPENDIX. E[m M] = (n S )(1 exp( exp(s min + c M))) (19) E[m M] n exp(s min + c M) (20) 8.1 EMPIRICAL EVALUATION OF SAMPLING 8 APPENDIX 8.1 EMPIRICAL EVALUATION OF SAMPLING We wish to evauate the empirica accuracy of our samping technique on concrete exampes. We do this in two ways. First, we can sort the eements by probabiity

More information

An Algorithm for Pruning Redundant Modules in Min-Max Modular Network

An Algorithm for Pruning Redundant Modules in Min-Max Modular Network An Agorithm for Pruning Redundant Modues in Min-Max Moduar Network Hui-Cheng Lian and Bao-Liang Lu Department of Computer Science and Engineering, Shanghai Jiao Tong University 1954 Hua Shan Rd., Shanghai

More information

Separation of Variables and a Spherical Shell with Surface Charge

Separation of Variables and a Spherical Shell with Surface Charge Separation of Variabes and a Spherica She with Surface Charge In cass we worked out the eectrostatic potentia due to a spherica she of radius R with a surface charge density σθ = σ cos θ. This cacuation

More information

Moreau-Yosida Regularization for Grouped Tree Structure Learning

Moreau-Yosida Regularization for Grouped Tree Structure Learning Moreau-Yosida Reguarization for Grouped Tree Structure Learning Jun Liu Computer Science and Engineering Arizona State University J.Liu@asu.edu Jieping Ye Computer Science and Engineering Arizona State

More information

First-Order Corrections to Gutzwiller s Trace Formula for Systems with Discrete Symmetries

First-Order Corrections to Gutzwiller s Trace Formula for Systems with Discrete Symmetries c 26 Noninear Phenomena in Compex Systems First-Order Corrections to Gutzwier s Trace Formua for Systems with Discrete Symmetries Hoger Cartarius, Jörg Main, and Günter Wunner Institut für Theoretische

More information

Multilayer Kerceptron

Multilayer Kerceptron Mutiayer Kerceptron Zotán Szabó, András Lőrincz Department of Information Systems, Facuty of Informatics Eötvös Loránd University Pázmány Péter sétány 1/C H-1117, Budapest, Hungary e-mai: szzoi@csetehu,

More information

The EM Algorithm applied to determining new limit points of Mahler measures

The EM Algorithm applied to determining new limit points of Mahler measures Contro and Cybernetics vo. 39 (2010) No. 4 The EM Agorithm appied to determining new imit points of Maher measures by Souad E Otmani, Georges Rhin and Jean-Marc Sac-Épée Université Pau Veraine-Metz, LMAM,

More information

A Better Way to Pretrain Deep Boltzmann Machines

A Better Way to Pretrain Deep Boltzmann Machines A Better Way to Pretrain Deep Botzmann Machines Rusan Saakhutdino Department of Statistics and Computer Science Uniersity of Toronto rsaakhu@cs.toronto.edu Geoffrey Hinton Department of Computer Science

More information

$, (2.1) n="# #. (2.2)

$, (2.1) n=# #. (2.2) Chapter. Eectrostatic II Notes: Most of the materia presented in this chapter is taken from Jackson, Chap.,, and 4, and Di Bartoo, Chap... Mathematica Considerations.. The Fourier series and the Fourier

More information

ASummaryofGaussianProcesses Coryn A.L. Bailer-Jones

ASummaryofGaussianProcesses Coryn A.L. Bailer-Jones ASummaryofGaussianProcesses Coryn A.L. Baier-Jones Cavendish Laboratory University of Cambridge caj@mrao.cam.ac.uk Introduction A genera prediction probem can be posed as foows. We consider that the variabe

More information

MATH 172: MOTIVATION FOR FOURIER SERIES: SEPARATION OF VARIABLES

MATH 172: MOTIVATION FOR FOURIER SERIES: SEPARATION OF VARIABLES MATH 172: MOTIVATION FOR FOURIER SERIES: SEPARATION OF VARIABLES Separation of variabes is a method to sove certain PDEs which have a warped product structure. First, on R n, a inear PDE of order m is

More information

A Novel Learning Method for Elman Neural Network Using Local Search

A Novel Learning Method for Elman Neural Network Using Local Search Neura Information Processing Letters and Reviews Vo. 11, No. 8, August 2007 LETTER A Nove Learning Method for Eman Neura Networ Using Loca Search Facuty of Engineering, Toyama University, Gofuu 3190 Toyama

More information

Statistical Learning Theory: a Primer

Statistical Learning Theory: a Primer ??,??, 1 6 (??) c?? Kuwer Academic Pubishers, Boston. Manufactured in The Netherands. Statistica Learning Theory: a Primer THEODOROS EVGENIOU AND MASSIMILIANO PONTIL Center for Bioogica and Computationa

More information

Expectation-Maximization for Estimating Parameters for a Mixture of Poissons

Expectation-Maximization for Estimating Parameters for a Mixture of Poissons Expectation-Maximization for Estimating Parameters for a Mixture of Poissons Brandon Maone Department of Computer Science University of Hesini February 18, 2014 Abstract This document derives, in excrutiating

More information

Lecture Note 3: Stationary Iterative Methods

Lecture Note 3: Stationary Iterative Methods MATH 5330: Computationa Methods of Linear Agebra Lecture Note 3: Stationary Iterative Methods Xianyi Zeng Department of Mathematica Sciences, UTEP Stationary Iterative Methods The Gaussian eimination (or

More information

Asymptotic Properties of a Generalized Cross Entropy Optimization Algorithm

Asymptotic Properties of a Generalized Cross Entropy Optimization Algorithm 1 Asymptotic Properties of a Generaized Cross Entropy Optimization Agorithm Zijun Wu, Michae Koonko, Institute for Appied Stochastics and Operations Research, Caustha Technica University Abstract The discrete

More information

Chapter 7 PRODUCTION FUNCTIONS. Copyright 2005 by South-Western, a division of Thomson Learning. All rights reserved.

Chapter 7 PRODUCTION FUNCTIONS. Copyright 2005 by South-Western, a division of Thomson Learning. All rights reserved. Chapter 7 PRODUCTION FUNCTIONS Copyright 2005 by South-Western, a division of Thomson Learning. A rights reserved. 1 Production Function The firm s production function for a particuar good (q) shows the

More information

Asynchronous Control for Coupled Markov Decision Systems

Asynchronous Control for Coupled Markov Decision Systems INFORMATION THEORY WORKSHOP (ITW) 22 Asynchronous Contro for Couped Marov Decision Systems Michae J. Neey University of Southern Caifornia Abstract This paper considers optima contro for a coection of

More information

Active Learning & Experimental Design

Active Learning & Experimental Design Active Learning & Experimenta Design Danie Ting Heaviy modified, of course, by Lye Ungar Origina Sides by Barbara Engehardt and Aex Shyr Lye Ungar, University of Pennsyvania Motivation u Data coection

More information

Cryptanalysis of PKP: A New Approach

Cryptanalysis of PKP: A New Approach Cryptanaysis of PKP: A New Approach Éiane Jaumes and Antoine Joux DCSSI 18, rue du Dr. Zamenhoff F-92131 Issy-es-Mx Cedex France eiane.jaumes@wanadoo.fr Antoine.Joux@ens.fr Abstract. Quite recenty, in

More information

MARKOV CHAINS AND MARKOV DECISION THEORY. Contents

MARKOV CHAINS AND MARKOV DECISION THEORY. Contents MARKOV CHAINS AND MARKOV DECISION THEORY ARINDRIMA DATTA Abstract. In this paper, we begin with a forma introduction to probabiity and expain the concept of random variabes and stochastic processes. After

More information

Online Appendices for The Economics of Nationalism (Xiaohuan Lan and Ben Li)

Online Appendices for The Economics of Nationalism (Xiaohuan Lan and Ben Li) Onine Appendices for The Economics of Nationaism Xiaohuan Lan and Ben Li) A. Derivation of inequaities 9) and 10) Consider Home without oss of generaity. Denote gobaized and ungobaized by g and ng, respectivey.

More information

Fitting Algorithms for MMPP ATM Traffic Models

Fitting Algorithms for MMPP ATM Traffic Models Fitting Agorithms for PP AT Traffic odes A. Nogueira, P. Savador, R. Vaadas University of Aveiro / Institute of Teecommunications, 38-93 Aveiro, Portuga; e-mai: (nogueira, savador, rv)@av.it.pt ABSTRACT

More information

Convolutional Networks 2: Training, deep convolutional networks

Convolutional Networks 2: Training, deep convolutional networks Convoutiona Networks 2: Training, deep convoutiona networks Hakan Bien Machine Learning Practica MLP Lecture 8 30 October / 6 November 2018 MLP Lecture 8 / 30 October / 6 November 2018 Convoutiona Networks

More information

NOISE-INDUCED STABILIZATION OF STOCHASTIC DIFFERENTIAL EQUATIONS

NOISE-INDUCED STABILIZATION OF STOCHASTIC DIFFERENTIAL EQUATIONS NOISE-INDUCED STABILIZATION OF STOCHASTIC DIFFERENTIAL EQUATIONS TONY ALLEN, EMILY GEBHARDT, AND ADAM KLUBALL 3 ADVISOR: DR. TIFFANY KOLBA 4 Abstract. The phenomenon of noise-induced stabiization occurs

More information

Melodic contour estimation with B-spline models using a MDL criterion

Melodic contour estimation with B-spline models using a MDL criterion Meodic contour estimation with B-spine modes using a MDL criterion Damien Loive, Ney Barbot, Oivier Boeffard IRISA / University of Rennes 1 - ENSSAT 6 rue de Kerampont, B.P. 80518, F-305 Lannion Cedex

More information

Related Topics Maxwell s equations, electrical eddy field, magnetic field of coils, coil, magnetic flux, induced voltage

Related Topics Maxwell s equations, electrical eddy field, magnetic field of coils, coil, magnetic flux, induced voltage Magnetic induction TEP Reated Topics Maxwe s equations, eectrica eddy fied, magnetic fied of cois, coi, magnetic fux, induced votage Principe A magnetic fied of variabe frequency and varying strength is

More information

17 Lecture 17: Recombination and Dark Matter Production

17 Lecture 17: Recombination and Dark Matter Production PYS 652: Astrophysics 88 17 Lecture 17: Recombination and Dark Matter Production New ideas pass through three periods: It can t be done. It probaby can be done, but it s not worth doing. I knew it was

More information

Supplement of Limited-memory Common-directions Method for Distributed Optimization and its Application on Empirical Risk Minimization

Supplement of Limited-memory Common-directions Method for Distributed Optimization and its Application on Empirical Risk Minimization Suppement of Limited-memory Common-directions Method for Distributed Optimization and its Appication on Empirica Risk Minimization Ching-pei Lee Po-Wei Wang Weizhu Chen Chih-Jen Lin I Introduction In this

More information

arxiv: v1 [math.ca] 6 Mar 2017

arxiv: v1 [math.ca] 6 Mar 2017 Indefinite Integras of Spherica Besse Functions MIT-CTP/487 arxiv:703.0648v [math.ca] 6 Mar 07 Joyon K. Boomfied,, Stephen H. P. Face,, and Zander Moss, Center for Theoretica Physics, Laboratory for Nucear

More information

STA 216 Project: Spline Approach to Discrete Survival Analysis

STA 216 Project: Spline Approach to Discrete Survival Analysis : Spine Approach to Discrete Surviva Anaysis November 4, 005 1 Introduction Athough continuous surviva anaysis differs much from the discrete surviva anaysis, there is certain ink between the two modeing

More information

Notes on Backpropagation with Cross Entropy

Notes on Backpropagation with Cross Entropy Notes on Backpropagation with Cross Entropy I-Ta ee, Dan Gowasser, Bruno Ribeiro Purue University October 3, 07. Overview This note introuces backpropagation for a common neura network muti-cass cassifier.

More information

From Margins to Probabilities in Multiclass Learning Problems

From Margins to Probabilities in Multiclass Learning Problems From Margins to Probabiities in Muticass Learning Probems Andrea Passerini and Massimiiano Ponti 2 and Paoo Frasconi 3 Abstract. We study the probem of muticass cassification within the framework of error

More information

How the backpropagation algorithm works Srikumar Ramalingam School of Computing University of Utah

How the backpropagation algorithm works Srikumar Ramalingam School of Computing University of Utah How the backpropagation agorithm works Srikumar Ramaingam Schoo of Computing University of Utah Reference Most of the sides are taken from the second chapter of the onine book by Michae Nieson: neuranetworksanddeepearning.com

More information

Appendix for Stochastic Gradient Monomial Gamma Sampler

Appendix for Stochastic Gradient Monomial Gamma Sampler 3 4 5 6 7 8 9 3 4 5 6 7 8 9 3 4 5 6 7 8 9 3 3 3 33 34 35 36 37 38 39 4 4 4 43 44 45 46 47 48 49 5 5 5 53 54 Appendix for Stochastic Gradient Monomia Gamma Samper A The Main Theorem We provide the foowing

More information

Theory of Generalized k-difference Operator and Its Application in Number Theory

Theory of Generalized k-difference Operator and Its Application in Number Theory Internationa Journa of Mathematica Anaysis Vo. 9, 2015, no. 19, 955-964 HIKARI Ltd, www.m-hiari.com http://dx.doi.org/10.12988/ijma.2015.5389 Theory of Generaized -Difference Operator and Its Appication

More information

A SIMPLIFIED DESIGN OF MULTIDIMENSIONAL TRANSFER FUNCTION MODELS

A SIMPLIFIED DESIGN OF MULTIDIMENSIONAL TRANSFER FUNCTION MODELS A SIPLIFIED DESIGN OF ULTIDIENSIONAL TRANSFER FUNCTION ODELS Stefan Petrausch, Rudof Rabenstein utimedia Communications and Signa Procesg, University of Erangen-Nuremberg, Cauerstr. 7, 958 Erangen, GERANY

More information

Determining The Degree of Generalization Using An Incremental Learning Algorithm

Determining The Degree of Generalization Using An Incremental Learning Algorithm Determining The Degree of Generaization Using An Incrementa Learning Agorithm Pabo Zegers Facutad de Ingeniería, Universidad de os Andes San Caros de Apoquindo 22, Las Condes, Santiago, Chie pzegers@uandes.c

More information

14 Separation of Variables Method

14 Separation of Variables Method 14 Separation of Variabes Method Consider, for exampe, the Dirichet probem u t = Du xx < x u(x, ) = f(x) < x < u(, t) = = u(, t) t > Let u(x, t) = T (t)φ(x); now substitute into the equation: dt

More information

A Fundamental Storage-Communication Tradeoff in Distributed Computing with Straggling Nodes

A Fundamental Storage-Communication Tradeoff in Distributed Computing with Straggling Nodes A Fundamenta Storage-Communication Tradeoff in Distributed Computing with Stragging odes ifa Yan, Michèe Wigger LTCI, Téécom ParisTech 75013 Paris, France Emai: {qifa.yan, michee.wigger} @teecom-paristech.fr

More information

AST 418/518 Instrumentation and Statistics

AST 418/518 Instrumentation and Statistics AST 418/518 Instrumentation and Statistics Cass Website: http://ircamera.as.arizona.edu/astr_518 Cass Texts: Practica Statistics for Astronomers, J.V. Wa, and C.R. Jenkins, Second Edition. Measuring the

More information

Learning Fully Observed Undirected Graphical Models

Learning Fully Observed Undirected Graphical Models Learning Fuy Observed Undirected Graphica Modes Sides Credit: Matt Gormey (2016) Kayhan Batmangheich 1 Machine Learning The data inspires the structures we want to predict Inference finds {best structure,

More information

Appendix for Stochastic Gradient Monomial Gamma Sampler

Appendix for Stochastic Gradient Monomial Gamma Sampler Appendix for Stochastic Gradient Monomia Gamma Samper A The Main Theorem We provide the foowing theorem to characterize the stationary distribution of the stochastic process with SDEs in (3) Theorem 3

More information

FORECASTING TELECOMMUNICATIONS DATA WITH AUTOREGRESSIVE INTEGRATED MOVING AVERAGE MODELS

FORECASTING TELECOMMUNICATIONS DATA WITH AUTOREGRESSIVE INTEGRATED MOVING AVERAGE MODELS FORECASTING TEECOMMUNICATIONS DATA WITH AUTOREGRESSIVE INTEGRATED MOVING AVERAGE MODES Niesh Subhash naawade a, Mrs. Meenakshi Pawar b a SVERI's Coege of Engineering, Pandharpur. nieshsubhash15@gmai.com

More information

Uniprocessor Feasibility of Sporadic Tasks with Constrained Deadlines is Strongly conp-complete

Uniprocessor Feasibility of Sporadic Tasks with Constrained Deadlines is Strongly conp-complete Uniprocessor Feasibiity of Sporadic Tasks with Constrained Deadines is Strongy conp-compete Pontus Ekberg and Wang Yi Uppsaa University, Sweden Emai: {pontus.ekberg yi}@it.uu.se Abstract Deciding the feasibiity

More information

Haar Decomposition and Reconstruction Algorithms

Haar Decomposition and Reconstruction Algorithms Jim Lambers MAT 773 Fa Semester 018-19 Lecture 15 and 16 Notes These notes correspond to Sections 4.3 and 4.4 in the text. Haar Decomposition and Reconstruction Agorithms Decomposition Suppose we approximate

More information

Smoothness equivalence properties of univariate subdivision schemes and their projection analogues

Smoothness equivalence properties of univariate subdivision schemes and their projection analogues Numerische Mathematik manuscript No. (wi be inserted by the editor) Smoothness equivaence properties of univariate subdivision schemes and their projection anaogues Phiipp Grohs TU Graz Institute of Geometry

More information

A Solution to the 4-bit Parity Problem with a Single Quaternary Neuron

A Solution to the 4-bit Parity Problem with a Single Quaternary Neuron Neura Information Processing - Letters and Reviews Vo. 5, No. 2, November 2004 LETTER A Soution to the 4-bit Parity Probem with a Singe Quaternary Neuron Tohru Nitta Nationa Institute of Advanced Industria

More information

SVM-based Supervised and Unsupervised Classification Schemes

SVM-based Supervised and Unsupervised Classification Schemes SVM-based Supervised and Unsupervised Cassification Schemes LUMINITA STATE University of Pitesti Facuty of Mathematics and Computer Science 1 Targu din Vae St., Pitesti 110040 ROMANIA state@cicknet.ro

More information

Centralized Coded Caching of Correlated Contents

Centralized Coded Caching of Correlated Contents Centraized Coded Caching of Correated Contents Qianqian Yang and Deniz Gündüz Information Processing and Communications Lab Department of Eectrica and Eectronic Engineering Imperia Coege London arxiv:1711.03798v1

More information

II. PROBLEM. A. Description. For the space of audio signals

II. PROBLEM. A. Description. For the space of audio signals CS229 - Fina Report Speech Recording based Language Recognition (Natura Language) Leopod Cambier - cambier; Matan Leibovich - matane; Cindy Orozco Bohorquez - orozcocc ABSTRACT We construct a rea time

More information

An approximate method for solving the inverse scattering problem with fixed-energy data

An approximate method for solving the inverse scattering problem with fixed-energy data J. Inv. I-Posed Probems, Vo. 7, No. 6, pp. 561 571 (1999) c VSP 1999 An approximate method for soving the inverse scattering probem with fixed-energy data A. G. Ramm and W. Scheid Received May 12, 1999

More information

1. Measurements and error calculus

1. Measurements and error calculus EV 1 Measurements and error cacuus 11 Introduction The goa of this aboratory course is to introduce the notions of carrying out an experiment, acquiring and writing up the data, and finay anayzing the

More information

Approximated MLC shape matrix decomposition with interleaf collision constraint

Approximated MLC shape matrix decomposition with interleaf collision constraint Approximated MLC shape matrix decomposition with intereaf coision constraint Thomas Kainowski Antje Kiese Abstract Shape matrix decomposition is a subprobem in radiation therapy panning. A given fuence

More information

Algorithms to solve massively under-defined systems of multivariate quadratic equations

Algorithms to solve massively under-defined systems of multivariate quadratic equations Agorithms to sove massivey under-defined systems of mutivariate quadratic equations Yasufumi Hashimoto Abstract It is we known that the probem to sove a set of randomy chosen mutivariate quadratic equations

More information

Unconditional security of differential phase shift quantum key distribution

Unconditional security of differential phase shift quantum key distribution Unconditiona security of differentia phase shift quantum key distribution Kai Wen, Yoshihisa Yamamoto Ginzton Lab and Dept of Eectrica Engineering Stanford University Basic idea of DPS-QKD Protoco. Aice

More information

Schedulability Analysis of Deferrable Scheduling Algorithms for Maintaining Real-Time Data Freshness

Schedulability Analysis of Deferrable Scheduling Algorithms for Maintaining Real-Time Data Freshness 1 Scheduabiity Anaysis of Deferrabe Scheduing Agorithms for Maintaining Rea-Time Data Freshness Song Han, Deji Chen, Ming Xiong, Kam-yiu Lam, Aoysius K. Mok, Krithi Ramamritham UT Austin, Emerson Process

More information

Paragraph Topic Classification

Paragraph Topic Classification Paragraph Topic Cassification Eugene Nho Graduate Schoo of Business Stanford University Stanford, CA 94305 enho@stanford.edu Edward Ng Department of Eectrica Engineering Stanford University Stanford, CA

More information

Chemical Kinetics Part 2

Chemical Kinetics Part 2 Integrated Rate Laws Chemica Kinetics Part 2 The rate aw we have discussed thus far is the differentia rate aw. Let us consider the very simpe reaction: a A à products The differentia rate reates the rate

More information

Support Vector Machine and Its Application to Regression and Classification

Support Vector Machine and Its Application to Regression and Classification BearWorks Institutiona Repository MSU Graduate Theses Spring 2017 Support Vector Machine and Its Appication to Regression and Cassification Xiaotong Hu As with any inteectua project, the content and views

More information

Optimality of Inference in Hierarchical Coding for Distributed Object-Based Representations

Optimality of Inference in Hierarchical Coding for Distributed Object-Based Representations Optimaity of Inference in Hierarchica Coding for Distributed Object-Based Representations Simon Brodeur, Jean Rouat NECOTIS, Département génie éectrique et génie informatique, Université de Sherbrooke,

More information

Consistent linguistic fuzzy preference relation with multi-granular uncertain linguistic information for solving decision making problems

Consistent linguistic fuzzy preference relation with multi-granular uncertain linguistic information for solving decision making problems Consistent inguistic fuzzy preference reation with muti-granuar uncertain inguistic information for soving decision making probems Siti mnah Binti Mohd Ridzuan, and Daud Mohamad Citation: IP Conference

More information

A Comparison Study of the Test for Right Censored and Grouped Data

A Comparison Study of the Test for Right Censored and Grouped Data Communications for Statistica Appications and Methods 2015, Vo. 22, No. 4, 313 320 DOI: http://dx.doi.org/10.5351/csam.2015.22.4.313 Print ISSN 2287-7843 / Onine ISSN 2383-4757 A Comparison Study of the

More information

Proceedings of Meetings on Acoustics

Proceedings of Meetings on Acoustics Proceedings of Meetings on Acoustics Voume 19, 213 http://acousticasociety.org/ ICA 213 Montrea Montrea, Canada 2-7 June 213 Underwater Acoustics Session 1pUW: Seabed Scattering: Measurements and Mechanisms

More information

Math 124B January 31, 2012

Math 124B January 31, 2012 Math 124B January 31, 212 Viktor Grigoryan 7 Inhomogeneous boundary vaue probems Having studied the theory of Fourier series, with which we successfuy soved boundary vaue probems for the homogeneous heat

More information

Soft Clustering on Graphs

Soft Clustering on Graphs Soft Custering on Graphs Kai Yu 1, Shipeng Yu 2, Voker Tresp 1 1 Siemens AG, Corporate Technoogy 2 Institute for Computer Science, University of Munich kai.yu@siemens.com, voker.tresp@siemens.com spyu@dbs.informatik.uni-muenchen.de

More information

Bayesian Unscented Kalman Filter for State Estimation of Nonlinear and Non-Gaussian Systems

Bayesian Unscented Kalman Filter for State Estimation of Nonlinear and Non-Gaussian Systems Bayesian Unscented Kaman Fiter for State Estimation of Noninear and Non-aussian Systems Zhong Liu, Shing-Chow Chan, Ho-Chun Wu and iafei Wu Department of Eectrica and Eectronic Engineering, he University

More information

Research of Data Fusion Method of Multi-Sensor Based on Correlation Coefficient of Confidence Distance

Research of Data Fusion Method of Multi-Sensor Based on Correlation Coefficient of Confidence Distance Send Orders for Reprints to reprints@benthamscience.ae 340 The Open Cybernetics & Systemics Journa, 015, 9, 340-344 Open Access Research of Data Fusion Method of Muti-Sensor Based on Correation Coefficient

More information

CONJUGATE GRADIENT WITH SUBSPACE OPTIMIZATION

CONJUGATE GRADIENT WITH SUBSPACE OPTIMIZATION CONJUGATE GRADIENT WITH SUBSPACE OPTIMIZATION SAHAR KARIMI AND STEPHEN VAVASIS Abstract. In this paper we present a variant of the conjugate gradient (CG) agorithm in which we invoke a subspace minimization

More information

Reichenbachian Common Cause Systems

Reichenbachian Common Cause Systems Reichenbachian Common Cause Systems G. Hofer-Szabó Department of Phiosophy Technica University of Budapest e-mai: gszabo@hps.ete.hu Mikós Rédei Department of History and Phiosophy of Science Eötvös University,

More information

Section 6: Magnetostatics

Section 6: Magnetostatics agnetic fieds in matter Section 6: agnetostatics In the previous sections we assumed that the current density J is a known function of coordinates. In the presence of matter this is not aways true. The

More information

Appendix of the Paper The Role of No-Arbitrage on Forecasting: Lessons from a Parametric Term Structure Model

Appendix of the Paper The Role of No-Arbitrage on Forecasting: Lessons from a Parametric Term Structure Model Appendix of the Paper The Roe of No-Arbitrage on Forecasting: Lessons from a Parametric Term Structure Mode Caio Ameida cameida@fgv.br José Vicente jose.vaentim@bcb.gov.br June 008 1 Introduction In this

More information

Chemical Kinetics Part 2. Chapter 16

Chemical Kinetics Part 2. Chapter 16 Chemica Kinetics Part 2 Chapter 16 Integrated Rate Laws The rate aw we have discussed thus far is the differentia rate aw. Let us consider the very simpe reaction: a A à products The differentia rate reates

More information

Some Measures for Asymmetry of Distributions

Some Measures for Asymmetry of Distributions Some Measures for Asymmetry of Distributions Georgi N. Boshnakov First version: 31 January 2006 Research Report No. 5, 2006, Probabiity and Statistics Group Schoo of Mathematics, The University of Manchester

More information

On the Goal Value of a Boolean Function

On the Goal Value of a Boolean Function On the Goa Vaue of a Booean Function Eric Bach Dept. of CS University of Wisconsin 1210 W. Dayton St. Madison, WI 53706 Lisa Heerstein Dept of CSE NYU Schoo of Engineering 2 Metrotech Center, 10th Foor

More information

Mat 1501 lecture notes, penultimate installment

Mat 1501 lecture notes, penultimate installment Mat 1501 ecture notes, penutimate instament 1. bounded variation: functions of a singe variabe optiona) I beieve that we wi not actuay use the materia in this section the point is mainy to motivate the

More information

VALIDATED CONTINUATION FOR EQUILIBRIA OF PDES

VALIDATED CONTINUATION FOR EQUILIBRIA OF PDES SIAM J. NUMER. ANAL. Vo. 0, No. 0, pp. 000 000 c 200X Society for Industria and Appied Mathematics VALIDATED CONTINUATION FOR EQUILIBRIA OF PDES SARAH DAY, JEAN-PHILIPPE LESSARD, AND KONSTANTIN MISCHAIKOW

More information

T.C. Banwell, S. Galli. {bct, Telcordia Technologies, Inc., 445 South Street, Morristown, NJ 07960, USA

T.C. Banwell, S. Galli. {bct, Telcordia Technologies, Inc., 445 South Street, Morristown, NJ 07960, USA ON THE SYMMETRY OF THE POWER INE CHANNE T.C. Banwe, S. Gai {bct, sgai}@research.tecordia.com Tecordia Technoogies, Inc., 445 South Street, Morristown, NJ 07960, USA Abstract The indoor power ine network

More information

Partial permutation decoding for MacDonald codes

Partial permutation decoding for MacDonald codes Partia permutation decoding for MacDonad codes J.D. Key Department of Mathematics and Appied Mathematics University of the Western Cape 7535 Bevie, South Africa P. Seneviratne Department of Mathematics

More information

SUPPLEMENTARY MATERIAL TO INNOVATED SCALABLE EFFICIENT ESTIMATION IN ULTRA-LARGE GAUSSIAN GRAPHICAL MODELS

SUPPLEMENTARY MATERIAL TO INNOVATED SCALABLE EFFICIENT ESTIMATION IN ULTRA-LARGE GAUSSIAN GRAPHICAL MODELS ISEE 1 SUPPLEMENTARY MATERIAL TO INNOVATED SCALABLE EFFICIENT ESTIMATION IN ULTRA-LARGE GAUSSIAN GRAPHICAL MODELS By Yingying Fan and Jinchi Lv University of Southern Caifornia This Suppementary Materia

More information

Research Article On the Lower Bound for the Number of Real Roots of a Random Algebraic Equation

Research Article On the Lower Bound for the Number of Real Roots of a Random Algebraic Equation Appied Mathematics and Stochastic Anaysis Voume 007, Artice ID 74191, 8 pages doi:10.1155/007/74191 Research Artice On the Lower Bound for the Number of Rea Roots of a Random Agebraic Equation Takashi

More information

FRST Multivariate Statistics. Multivariate Discriminant Analysis (MDA)

FRST Multivariate Statistics. Multivariate Discriminant Analysis (MDA) 1 FRST 531 -- Mutivariate Statistics Mutivariate Discriminant Anaysis (MDA) Purpose: 1. To predict which group (Y) an observation beongs to based on the characteristics of p predictor (X) variabes, using

More information

C. Fourier Sine Series Overview

C. Fourier Sine Series Overview 12 PHILIP D. LOEWEN C. Fourier Sine Series Overview Let some constant > be given. The symboic form of the FSS Eigenvaue probem combines an ordinary differentia equation (ODE) on the interva (, ) with a

More information

Proceedings of Meetings on Acoustics

Proceedings of Meetings on Acoustics Proceedings of Meetings on Acoustics Voume 9, 23 http://acousticasociety.org/ ICA 23 Montrea Montrea, Canada 2-7 June 23 Architectura Acoustics Session 4pAAa: Room Acoustics Computer Simuation II 4pAAa9.

More information

General Decay of Solutions in a Viscoelastic Equation with Nonlinear Localized Damping

General Decay of Solutions in a Viscoelastic Equation with Nonlinear Localized Damping Journa of Mathematica Research with Appications Jan.,, Vo. 3, No., pp. 53 6 DOI:.377/j.issn:95-65...7 Http://jmre.dut.edu.cn Genera Decay of Soutions in a Viscoeastic Equation with Noninear Locaized Damping

More information

Formulas for Angular-Momentum Barrier Factors Version II

Formulas for Angular-Momentum Barrier Factors Version II BNL PREPRINT BNL-QGS-06-101 brfactor1.tex Formuas for Anguar-Momentum Barrier Factors Version II S. U. Chung Physics Department, Brookhaven Nationa Laboratory, Upton, NY 11973 March 19, 2015 abstract A

More information

VALIDATED CONTINUATION FOR EQUILIBRIA OF PDES

VALIDATED CONTINUATION FOR EQUILIBRIA OF PDES VALIDATED CONTINUATION FOR EQUILIBRIA OF PDES SARAH DAY, JEAN-PHILIPPE LESSARD, AND KONSTANTIN MISCHAIKOW Abstract. One of the most efficient methods for determining the equiibria of a continuous parameterized

More information

Two view learning: SVM-2K, Theory and Practice

Two view learning: SVM-2K, Theory and Practice Two view earning: SVM-2K, Theory and Practice Jason D.R. Farquhar jdrf99r@ecs.soton.ac.uk Hongying Meng hongying@cs.york.ac.uk David R. Hardoon drh@ecs.soton.ac.uk John Shawe-Tayor jst@ecs.soton.ac.uk

More information