A Machine Learning Approach for Dead-Reckoning Navigation at Sea Using a Single Accelerometer

Size: px
Start display at page:

Download "A Machine Learning Approach for Dead-Reckoning Navigation at Sea Using a Single Accelerometer"

Transcription

1 A Machine Learning Approach for Dead-Reckoning Navigation at Sea Using a Single Accelerometer 1 Roee Diamant, Yunye Jin Electrical and Computer Engineering, The University of British Columbia Institute for Infocomm Research, Singapore roeed@ece.ubc.ca, yjin@i2r.a-star.edu.sg Abstract Dead-reckoning (DR) navigation is used when GPS reception is not available or its accuracy is not sufficient. At sea, DR requires the use of inertial sensors, usually a gyrocompass and an accelerometer, to estimate the orientation and distance traveled by the tracked object with respect to a reference coordinate system. In this paper, we consider the problem of DR navigation for vessels located close to or on the sea surface, where motion is affected by ocean waves. In such cases, the vessel pitch angle is fast timevarying and its estimation via direct measurements of orientation is prone to drifts and noises of the gyroscope in use. Regarding this problem, we suggest a method to compensate on the vessel pitch angle using a single acceleration sensor. Using a constraint expectation maximization (EM) algorithm, our method classifies acceleration measurements into states of similar pitch angles. Then, for each class, we project acceleration measurements into the reference coordinate system along the vessel heading direction, and obtain distance estimations by integrating the projected measurements. Results in both simulated and actual sea environments demonstrate that by using only acceleration measurements, our method achieves accurate results. Index Terms Naval navigation, dead-reckoning, EM classification.

2 2 I. INTRODUCTION Navigation capability is crucial for any vessel at sea. While navigation at sea often involves a GPS receiver, in many cases its accuracy may not be sufficient (for example, navigating in harbor environment or collision avoidance in near shore exploration). Moreover, for vessels like autonomous underwater vehicles (AUV), GPS reception is not available when the vessel is submerged. In such scenarios, beacon-free navigation using dead-reckoning (DR) is a viable alternative [1]. Terrestrial DR navigation usually requires the use of inertial systems including an accelerometer and a three-dimensional (3-D) digital compass or gyroscope. The former is used to estimate the displacement of the tracked object, while the latter is used to measure its 3-D orientation. At sea, due to interference from metallic and electrical instruments on board vessels, instead of magnetic compass gyrocompass is preferred [2]. In DR navigation, a key challenge is to determine the orientation of the inertial sensor with respect to a reference coordinate system (e.g. the East-North-Up world coordinate system). Only then the distance traveled by the tracked object can be estimated. When the vessel is located close to or on the sea surface such that its motion is affected by the ocean waves, DR navigation poses an additional challenge as the vessel pitch angle is fast time-varying. For pedestrian DR, where the sensor is attached to the person foot or heap, this is usually dealt with by identifying reference points for which velocity is zero (for example, when the foot is placed on ground). In this way, timely orientation measurements from 3-D compass or gyroscope are calibrated [3]. However, at sea, since vessels constantly move, it is hard to identify such reference points. Furthermore, when orientation measurements of gyrocompass are directly used to compensate on the vessel pitch angle, accuracy is affected by the drifts and noises of the gyroscope in use [1]. For this reason, traditionally, DR navigation at sea involves integrating a large number of inertial sensors and applying Bayesian filtering methods to mitigate oscillatory wave-dependent components and reduce measurement drifts [4]. However, this approach limits the resolution of DR navigation, as measurements are filtered over the period of several ocean waves. Moreover, accurate gyrocompasses are expensive and may not be available. In this paper, we take a different approach and using a single 3-D acceleration sensor we estimate the heading and distance traveled by a vessel whose motion is affected by the ocean

3 3 Estimate heading angle Align measurements with vessel motion to compansate on heading angle Clasify measurements to M pitch states Per state: estimate pitch angle Per state: compansate on pitch angle Combine per state projected measurements Integrate measurements to estimate distance Fig. 1: Flow chart for the DR-A algorithm. waves. Our method, denoted as the DR-navigation-using-a-single-accelerometer (DR-A), is based on the observation that, due to ocean waves, the vessel pitch angle is periodic in nature. The intuition behind our method is that, when the vessel pitch angle is fixed, direction can be determined without using orientation measurements. Hence, by identifying classes of acceleration measurements such that in each class the pitch angle is approximately the same, we can readily project acceleration measurements into a reference coordinate system. While our method can be used also for AUVs submerged deep at sea, it is mostly designated to compensate on wave induced motions. The main contribution of our work is therefore in employing machine learning classification techniques for DR navigation of vessels located close to the sea surface, and obviating the need to compensate on the vessel time-varying pitch angle via direct measurement of the vessel orientation using dedicated hardware, e.g., gyrocompass. We offer a sequential approach as follows. First, we estimate the vessel heading angle and project acceleration measurements such that these are aligned with the vessel motion. Next, we classify the projected measurements into pitch-states of similar pitch angles. This is performed using a constraint expectation maximization (EM) algorithm. Third, per pitch-state, we project acceleration measurements (whose pitch angles are similar) to compensate on the vessel pitch angle. After this step, the vessel s projected local coordinate system is considered aligned with the the vessel heading direction and the East-North-Up coordinates system. Finally, we integrate the projected measurements to evaluate the distance traveled by the vessel. A flow chart of the above process is given in Figure 1. The remainder of this paper is organized as follows. Related work on DR navigation is

4 4 described in Section II. Our system model and assumptions are introduced in Section III. In Section IV, we present the steps of our DR-A method. Section V includes performance evaluation of our algorithm in simulations (Section V-A) and using realistic data obtained during a designated sea trial (Section V-B). Finally, conclusions are offered in Section VI. The key notations used in this paper are summarized in Table I. TABLE I: List of key notations Notation Explanation d i,j distance traveled by the vessel at time period [t i, t j ] N T c n α n ρ n number of acceleration measurements coherence time of acceleration along the horizontal plane the vessel orientation with respect to the reference system in time t n the vessel heading angle with respect to n the vessel pitch angle in time t n [a n,x, a n,y, a n,z] the vessel 3-D acceleration at time t n in the horizontal plane coordinate system â n   h  h,p M l L δ l Θ p acceleration measurement with elements [â n,x, â n,y, â n,z ] obtain at time t n vector of acceleration measurements obtained between t i and t j Matrix  projected to match the vessel heading Matrix  projected into the horizontal plane along the vessel heading pre-defined number of pitch-states group of acceleration measurements which must be classified into the same pitch-state number of groups l classifier for group l the distribution parameters of Â, estimated at the pth EM iteration µ m,x mean of the mth class for the x axis σ m,x k m standard deviation of the mth class for the x axis prior probability of the mth class II. RELATED WORK DR navigation finds its applications both for pedestrian [2] or vessel navigation [1]. DR navigation requires inertial sensors to estimate the distance traveled by the tracked object through integration of acceleration measurements while evaluating the sensor orientation. The main challenge in DR navigation is the possible drift and measurement noise of inertial sensors, which may lead to errors in the order of 10% of the traveled distance, depending on the technology

5 5 employed [5]. In [6], it was suggested to reduce sensor drift by posing constraints on the estimated distance based on an evaluated route of the tracked object. In [7], the periodic motions when walking or running are used as reference in mitigating acceleration measurement errors. For pedestrian applications, using the fact that velocity and orientation can be set to zero when the foot is on ground, it is customary to mount inertial sensors on foot or heap and estimate distance and orientation separately for each step, e.g., [8], [3], [9]. However, while at sea we can identify time instances where the vessel pitch angle is zero (i.e., at the top or bottom of the ocean wave, see Figure 2), velocity of the vessel cannot be assumed zero at these points. Instead, in areas of shallow water, Doppler velocity loggers (DVL) are used to serve as reference for DR navigation [10], and measurement drifts are controlled through fusion with range estimations to anchor nodes [11]. At areas of deep sea, DR navigation is performed by integrating measurements from large number of inertial sensors [4]. Since motion of vessels located close to the sea surface is affected by wind, currents, and wave forces, ship s DR navigation involves dynamic positioning, heading autopilots, and thruster-assisted position mooring [1]. On the sea surface, navigation also includes signals quality checking through comparing results with, e.g., GPS receivers and surface positioning references [12]. Angular velocity measurements from gyrocompass can be used to compensate over distortions of acceleration data caused by the ocean waves [2]. To compensate on noise, measurements from inertial sensors are filtered using, e.g., extended Kalman filter or particle filters (cf. [5], [13]) over a period of several ocean waves [1]. However, often the vessel pitch angle is fast timevarying and the navigation system avoids correcting motion for every single wave [4], affecting resolution. A different approach to compensate on the vessel s time-varying pitch angle is to project acceleration measurements using the principal component analysis (PCA) method [14]. PCA gives the axis coordinate system (or the transformation matrix to that system) along whose axes the variation of change is smallest [15]. If the vessel acceleration is constant in a certain plane, PCA projection aligns acceleration measurements with that plane. However, this method relies on large variance of acceleration measurements in the y and z axes, otherwise large errors occur.

6 6 ẑ ẑ z ẑ ˆx ˆx ẑ ˆx ẑ ˆx ρ n ˆx x y α n ŷ Fig. 2: Illustration of the vessel s wave-induced motion. III. SYSTEM MODEL Our system comprises of a three-axis accelerometer device attached to a vessel, and producing N i.i.d acceleration measurements during a time period of interest, [t start, t end ]. We assume that at time t start, the vessel s initial speed and heading with respect to a reference coordinate system (e.g., UTM) is given as v i, and init, respectively. Our objective is to track the vessel heading angle between times t start and t end, and to estimate the distance traveled by the vessel during the considered time period 1, di,j. Let a n,x, n, and α n, n = 1,..., N be the vessel s acceleration along the vessel heading direction, the vessel heading direction with respect to the reference coordinate system, and the vessel orientation (i.e., the direction in which the vessel bow is pointing) with respect to its heading direction n, respectively. We assume that the change in α n is only due to timely changes in n. That is, during period [t start, t end ] the vessel orientation with respect to the reference system does not change. In addition, we assume both a n,x and n are slowly timevarying such that the considered time period can be divided into W = tend t start T c duration T c (representing the coherence time of the system), in each N slot = NTc t end t start time slots of acceleration measurements are acquired, where in each time slot w, n and a n,x are assumed fixed and equal w and a w,x, respectively. For simplicity, we assume the vessel motion is perfectly correlated with ocean waves such that the vessel roll angle can be neglected. However, extension is suggested also for the case of time-varying pitch and roll angles (see Section IV-E). 1 Throughout the paper upper case bold symbols represent matrices, bold symbols represent vectors, ˆθ represents the measurement of true parameter θ, and θ represents the estimation of θ.

7 7 In the wth time slot, on time t n, n = 1,..., N slot, t start t n t end, the accelerometer device produces three-axis acceleration measurement vector, â n = [â n,x, â n,y, â n,z ], at local coordinates, which are grouped into a matrix Â. Vector â n is a projection of a vector a n = [a n,x, a n,y, a n,z ] T representing the vessel s acceleration in a 3-D coordinate system referred to as the horizontal plane. The coordinates of the horizontal plane are defined such that the x axis is aligned with the vessel s heading direction, assumably there is no movement in the y axis (since within the time slot the heading is assumed fixed), and the z axis is as in the East-North-Up coordinates system. Thus, the horizontal plane is separately defined for each time slot of duration T c and assumed fixed heading and acceleration. Referring to the illustration on Figure 2, the projection from a n into â n is modeled by rotating vector, [a n,x, a n,y, a n,z ] T about its y-axis by the vessel time-varying pitch angle, ρ n, then its z-axis by angle α n, and adding a zero-mean Gaussian noise, e n, with a per-axis standard deviation of ς 3 m/(sec)2, such that â n,x cos ρ n 0 sin ρ n cos α n sin α n 0 â n,y = sin α n cos α n 0 â n,z sin ρ n 0 cos ρ n a n,x a n,y a n,z + e n,x e n,y e n,z. (1) We note that while a n,y is assumed zero, a n,z can be non-zero. This is due to the up and down movements of the vessel with the ocean waves. However, we expect a small value for a n,z. This is required to solve an ambiguity in estimating the pitch angle 2. In fact, we are interested in cases where a n,x >> a n,z. Otherwise, the vessel is moving in almost fixed speed, v i, and projection is not required. However, it is important to note that by (1), even when a n,z is zero compensation over both α n and ρ n is required. For each time instance t n, we aim to estimate ρ n and α n, and to project the vessel s local coordinate system about the z-axis followed by its y-axis. The first projection converts the set of measurements  into a projected set Âh with elements â h n, and the second into a projected set  h,p with elements â h,p n. Since after the second projection the local coordinate system is expected to be aligned with the horizontal plane, we can readily integrate elements â h,p n estimate d i,j. Furthermore, since the vessel orientation with respect to the reference coordinate system is assumed fixed during [t start, t end ], such projection allows us to estimate the per-time slot change in the vessel s heading direction with respect to the initial heading, init. to 2 This assumption is validated using a realistic wave model as well as results from a sea trial.

8 8 By (1), the set of measurements  are affected by the pitch angle of the vessel. Much like the ocean waves, the vessel pitch angle is periodic in nature and likely, π < ρ 4 n < π. However, being 4 induced from ocean waves, we do not assume a fixed period for the vessel pitch angle. Given this periodicity, we can identify time instances for which the vessel pitch angle is approximately the same. This observation sets the stage for classifying measurements into pitch-states of assumed fixed pitch angle, given the following model. Let Â(w) with elements â n (w) = [â n,x (w), â n,z (w)] T, n = 1,..., N slot be the set of 2-D acceleration measurements acquired during the wth time slot, w = 1,..., W, and sorted by their measurement time. We consider a Gaussian mixture with M pitch-states for the distribution of Â(w) such that for Θ(w) = {ω 1,x (w), ω 1,z (w), k 1 (w),..., ω M,x (w), ω M,z (w), k M (w)}, P (Â(w) Θ(w)) = M k m (w)p m (â n,x (w) ω m,x (w))p m (â n,z (w) ω m,z (w)), (2) â n (w) Â(w) m=1 where ω m,x (w) = [ µ m,x (w), σ 2 m,x(w) ], and ω m,z (w) = [ µ m,z (w), σ 2 m,z(w) ], are the distribution m=1 parameters of the mth pitch-state with mean µ m,x (w) and µ m,z (w), and variance σm,x(w) 2 and σm,z(w), 2 respectively, and k m (w) ( M k m (w) = 1) is the a-priori probability of the mth pitchstate. Model (2) is used to classify acceleration measurements into pitch-states using a constraint EM algorithm. For clarity, in the following we drop the time slot subindex, w, and consider projection of measurements acquired during a single time slot. IV. THE DR-A METHOD Refereing to model (1), in the DR-A method we estimate angles α n and ρ n to project measurement in  into the horizontal plane and obtain an estimation for a n,x, n = 1,..., N slot. While the orientation angle, α n, is assumed fixed during the time slot, it is not the case for the pitch angle, ρ n. For this reason, we separately compensate on the vessel heading and pitch angles. The former is done jointly for all measurements acquired during a single time slot, resulting in a projected set  h. For the latter, we use machine learning approach to classify acceleration measurements into states of similar pitch angles, and per-state compensate on the pitch angle to form a projected set  h,p, considered aligned with the horizontal plane. Finally, we combine projected measurements of all time slots, and, using the initial (given) velocity v i, integrate them over the time period [t start, t end ] to obtain d i,j. In the following sections we describe in details the steps of our method.

9 9 A. Forming Âh : Estimation of the Heading Angle Before forming classes of similar pitch angles, we project â n Â, n = 1,..., N slot to form â h n = [â h n,x, â h n,y, â h n,z], for which â h n,z = â n,z and (if the vessel s pitch angle is zero) â h n,x aligns with the heading of the vessel. This projection is performed as follows. Refereing to Figure 2, α n denotes the angle between the heading direction of the vessel, n, and the accelerometer s local x-axis. Since per time slot the vessel is assumed moving in a fixed heading, without measurement noise, there should be no acceleration, caused by the vessel motion, along the axis which is perpendicular to the heading direction. Therefore, we expect â n,x sin α n â n,y cos α n 0, (3) and estimation α n = α which minimizes N â n,x sin α â n,y cos α is obtained by n=1 tan α = N â n,y n=1. (4) N â n,x Once α is determined through (4), we project measurements in  and form matrix Âh. This is performed by multiplying â n Â, n = 1,..., N slot with the rotation matrix cos α sin α 0 sin α cos α 0. (5) In the wth time slot, the estimated angle α w from (4) is compared with α w 1 to estimate the heading direction of the vessel, w, assumed fixed in each time slot. Let w be the change in the heading direction of the vessel between the w 1th time slot and the wth one. Then, w = w 1 w. Recall we assume the direction of the vessel s bow with respect to the reference coordinate system (e.g., UTM) does not change during the considered time period, [t start, t end ]. Thus, w = α w α w 1, and w = w 1 w, (6) where 0 = init. n=1

10 10 B. Forming Âh,p : Estimating the Time-varying Pitch Angle After forming matrix Âh, we estimate the vessel pitch angle to form matrix Âh,p, which will be used to readily calculate d i,j. Since the vessel pitch angle is time-varying, different from estimation (4), elements of Âh cannot be directly projected to Âh,p. For this reason, traditionally, a direct measurement of the vessel orientation is used via, e.g., gyrocompass, which, as stated in Section I, may not be available or is too noisy. In this section we describe an alternative approach, where based on the observation that the vessel pitch angle is periodic in nature, we form matrix Âh,p by grouping acceleration measurements to classes of similar pitch angles. Based on model (2), to identify measurements of similar pitch angles we use a constraint EM algorithm and classify the elements in Âh to M pitch-states, where M is a pre-defined number of states. For an assumed coherence time, T c, classification is performed separately for each time slot. Since the effect of the pitch angle is similar in the x and z axes, classification is performed jointly for both axes. In classifying Âh, we can use side information in the form of positive constraints for measurements which must belong to the same pitch-state, and negative constraint for measurements which must be classified into different pitch-states, as described in the following. 1) Formulating Positive Constraints: Positive constraints allow us to mitigate measurement noise by relaxing the element-wise clustering problem, and instead clustering sets of measurements. Since the vessel pitch angle is continuous, we limit positive constraints to consecutive measurements, â h n, â h n+1,..., â h n+ε of similar values, where ε is limited. To formulate positive constraint, let V pos be a predefined multiplication of the sensitivity level of the used accelerometer device. Also let T p be a pitchcoherence time used as an upper bound for the time period positive constraints can be defined for two acceleration measurements. Denote relation a n a n if a n,x a n,x < V pos (7a) a n,z a n,z < V pos (7b) t n t n T p, (7c) otherwise a n a n. We classify measurements â h n and â h n into the same pitch-state if âh n â h n, and if there exist no element â h j, such that t n < t end < t n and â h n â h j or â h n âh j. Since the

11 11 vessel pitch angle is affected by ocean waves, we set T p = 1 2f wave, where f wave is the fundamental frequency of the ocean waves, which can be estimated using, e.g., cyclostationary analysis (cf. [16]). Positive constraints are used to form matrixes l, l = 1,..., L, each including a chain of consecutive measurements which pair-wise are classified into the same pitch-state. The above definition of positive constrains guarantees that l are disjoint matrixes. We obtain l by grouping consecutive acceleration measurements satisfying the above constraints in disjoint matrixes such that a new matrix is formed every time positive constraints are not met for two consecutive measurements. For each vector l we assign classifier δ l = {1,..., M} such that all elements in l are classified into the same δ l pitch-state. To this end, we reduce the problem of classifying â h n Âh into classifying l. 2) Formulating Negative Constraints: Negative constraints reflect the expected change in the vessel pitch angle. Since the last element of l and the first element of l+1 do not satisfy (7), negative constraints are formed as the complementary of positive constraints. More specifically, negative constraints ensure that two consecutive subsets l and l+1 cannot share the same pitch-state. Interestingly, since the rank of vectors l, l = 1,..., L is limited by T p, if T p < T c (recall T c is the duration of each time slot) then negative constraints renders L 2. We represent negative constraints by operator 1 δ l = δ l+1 η δl,δ l+1 =, l = 1,..., L. (8) 0 otherwise Shental [17] suggested formalizing negative constraints by a directed Markov graph, where in our case classifier δ l is a hidden (decision) node, connected to both δ l 1 and δ l+1, as well as to an observation node, â h n l. Since the pitch angle associated with l+1 depends on the pitch angle associated with l, hidden nodes form a directed chain from parent, δ l, to child, δ l+1, and from hidden nodes to observation nodes, as illustrated in Figure 3. This directed chain is a form of a Bayesian belief network [18], where for Y being the set of classifiers δ l, l = 1,..., L (with δ 0 = ), we obtain the joint distribution ( ) L ) P Y Âh, Θ = P (δ l δ l 1, Θ, Âh. (9) l=1 The relation in (9) would serve us in formalizing the likelihood function to determine δ l using a constraint EM algorithm, as described next.

12 12 δ l 1 δ l δ l+1 l 1 l l+1 Fig. 3: Graph representation for positive and negative constraints. 3) Formulating the Likelihood Function: Considering both positive and negative constraints, the data likelihood function for the event of correct classification of the elements of  h is P (Âh, Y Θ) = 1 L 1 L (1 η δl,δ c l+1 ) 1 where c 1 = δ 1 =1 L 1 δ L =1 l=1 l=1 l=1 â h n l (1 η δl,δ l+1 ) L P (δ l Θ). l=1 P (δ l Θ)P (â h n,x δ l, Θ)P (â h n,z δ l, Θ), (10) For classifying subsets l, we use the EM algorithm to iteratively maximize (10). In the pth iteration, we obtain estimation Θ p of Θ, with elements µ p m,x, µ p m,z, σ p m,x, σ p m,z, and k p m, m = 1,..., M. For the Gaussian mixture model (2), given Θ p we obtain [19] Furthermore, µ p+1 m,ω = σ p+1 m,ω = k p+1 p+1 1,..., k M s.t. m=1 L P (δ l = m Âh, Θ p ) l=1 â h n,ω l â h n,ω L P (δ l = m Âh, Θ p ) l=1 L P (δ l = m Âh, Θ p (âh ) n,ω µ p+1 m â h n,ω l l=1 k m = 1. L P (δ l = m Âh, Θ p ) l=1 ( ) 1 = argmax log + k 1,...,k M c 1 m=1 log(k m ), ω {x, z} (11a) ) 2, ω {x, z}. (11b) L P (δ l = m Âh x, Âh z, Θ p ) l=1 (12a) (12b)

13 13 Since c 1 is a function of k m (recall k m = P (δ l = m Θ p )), it is hard to maximize (12). Instead, we approximate c 1. Assuming negative constraints are mutually exclusive, c 1 (1 j=1 k 2 j ) L 1 [17]. Clearly, this approximation does not hold in our case. However, since our Markov graph is sparse, this assumption has little affect on performance (as our numerical simulations show). From (12) and using a lagrange multiplier, we get ( 2(L 1) k p+1 m + P (δ l = m Âh x, Âh z, Θ p ) k p+1 m The expression in (13) can be solved for (and k p j 1 k p+1 m ( k p+1 j=1 j ) 2 ) +(2L 1) j=1 either by approximating ( k p+1 j ) 2 1 = 0. (13) ( k p+1 j j=1 ) 2 as ( k p j )2 j=1 is known from previous iteration), or using alternating optimization technique (cf. [20]). Equations (11a), (11b), and (13), are used to determine Θ p+1, which in turn is used for the next EM iteration till convergence of (10) is reached 3. In the last EM iteration, p last, we determine classifiers δ l, l = 1,..., L of vectors l by numerically solving [ δ l = argmax P ( δ l = m Âh x, Âh z, Θ ] p last ) δ l, (14) and construct matrix Âh m, m = 1,..., M including elements â h n for which δ l = m and ρ n = ρ m. Considering model (1), to project the elements in matrix Âh m into Âh,p m, we guess a solution, ρ m, and multiply â h n Âh m, n = 1,..., N by the rotation matrix cos ρ m 0 sin ρ m 0 1 0, (15) sin ρ m 0 cos ρ m Observing model (1), we note that when the noise vector e n is zero and heading compensation is ideal, multiplying â h n Âh m by the matrix in (15) gives the vector an,x( ρ h,p m ) a n,x cos ( ρ m ρ m ) + a n,z sin (ρ m ρ m ) a h,p n,y( ρ m ) = a n,y. (16) a h,p n,z( ρ m ) a n,x sin ( ρ m ρ m ) + a n,z cos ( ρ m ρ m ) 3 Note that the EM algorithm is proven to converge to a local maxima of the log-likelihood function [19].

14 14 Since a n,z can be non-zero, ρ m cannot be estimated following the same approach as in (4). Unfortunately, (16) sets a degree of freedom in choosing ρ m. Instead, since we assumed a n,x >> a n,z (see Section III), the term a n,x (cos ( ρ m ρ m ) + sin ( ρ m ρ m )) is approximated by a h,p n,x( ρ m ) + a h,p n,z( ρ m ). Since a h,p n,x( ρ m ) + a h,p n,z( ρ m ) achieves its maximum for ρ m ρ m = π 4 when a n,x > 0, and for ρ m ρ m = π 4 + π when a n,x < 0, we set ρ m,1 = argmax ρ m ρ m,2 = argmax ρ m n: â h n Âh m n: â h n Âh m a h,p n,x( ρ m ) + a h,p n,z( ρ m ) π 4 a h,p n,x( ρ m ) + a h,p n,z( ρ m ) π 4 π. (17) From the two solutions in (17), we choose the one satisfying condition π < ρ 4 m < π, and using 4 (16) form projection a h,p n,x, considered aligned with the horizontal plane along the vessel heading direction. Finally, we combine projected measurements from all time slots and pitch-states in matrix Âh,p. Due to the dependency between hidden nodes δ l, the difficulty in (11), (13), and (14) lies in calculating P (δ l = m Âh x, Âh z, Θ p ). We next describe an efficient way to obtain the posterior probabilities. 4) Finding the Posterior Probability: Finding the posterior probability is simple when only positive constraints are imposed. Here, P (δ l = m Âh, Θ p ) = P (δ l = m l, Θ p ), and since measurements in Âh are assumed independent (see (2)), P (δ l = m l, Θ p ) = k p mp ( l ω p m) P ( l Θ p ) k p m P (â h n,x ω p m)p (â h n,z ω p m) â = h n l M k p j=1 j P (â h n,x ω p j )P (âh n,z ω p j ). (18) â h n l However, at the presence of negative constraints, δ l depends on other hidden nodes, and where P (Y Âh, Θ p ) = c 2 = δ 1 =1 (1 η δl,δ l+1 ) L L 1 l=1 L 1 δ L =1 l=1 (1 η δl,δ l+1 ) l=1 â h n l L l=1 â h n l P (δ l â h n,x, Θ p )P (δ l â h n,z, Θ p ) c 2, (19) P (δ l â h n,x, Θ p )P (δ l â h n,z, Θ p ).

15 15 Here, the posterior probability can be obtained by marginalizing the joint probability (19). However, in general, this is an NP-hard problem (with similarities to the graph coloring problem) ( ) with complexity O M L 1 [18]. Instead, since the Markov graph illustrated in Figure 3 is a directed chain, belief propagation techniques can be used to significantly reduce complexity to O (LM 2 ) [18]. The general idea in belief propagation is based on observation (9), that assuming P (δ l δ l 1 ) is known, belief BEL δl (m) = P (δ l Âh, Θ p ) can be exactly calculated by receiving belief BEL δl 1 (m) from parent δ l 1 and likelihood P (Âh δ l+1, Θ p ) from child δ l+1. Following [17], we use Pearl s belief propagation algorithm for trees [18], which adaptation to our system is as follows. In Pearl s algorithm, each hidden node l exchanges messages λ δl (m) and π δl (m) with his parents and children, respectively. We start by initializing lists π δl (m) = λ δl (m) = 1, l, m, l = 1,..., L, m = 1,..., M. Upon activation, each hidden node, δ l, receives π δl 1 (m) and λ δl+1 (m) from its parent and child, respectively. It then calculates λ δl (m) = π δl (m) = n=1 n=1 λ δl+1 (n) â h l P π δl 1 (n) â h l P where π δ0 (m) = 1 (recall δ 0 = ), and sets Next, hidden node δ l sends message (δ l = n δ l 1 = m, â hn,x, â hn,z, Θ p ) ( δ l = m δ l 1 = n, â h n,x, â h n,z, Θ ) p ) BEL δl (m) = λ δ l+1 (m)π δl (m) j=1 π δl (m) π δl (j) j=1 BEL δl (j) to its child, δ l+1, and (20a), (20b). (21) λ δ l (m) λ δl (j) j=1 to its parent, δ l 1. Since our network is a directed Markov chain, we execute (20) and (21) for the sequence δ 1 δ 2 δ L, and reach convergence after L steps. Since the above procedure is performed for each pth EM iteration and wth time slot, the complexity of the DR-A algorithm is O (LM 2 p last W ), where W = tend t start T c and L N W. 5) Initial Estimation Θ 0 : The EM algorithm in (11) and (12) requires initialization Θ 0. With no prior knowledge of the effect of ocean waves on acceleration, we use the K-means clustering algorithm (cf. [19]) to

16 16 initially classify elements in each time slot into M pitch-states and form matrixes Âh m. Similar to the EM algorithm, the K-means algorithm is executed jointly for Âh x and Âh z. After this initial classification, based on the statistical mean and standard deviation of elements â h n,x, â h n,z Âh m we calculate µ 0 m,x, µ 0 m,z, and σ m,x, 0 σ m,z, 0 m = 1,..., M, respectively, and k m 0 is estimated as the ratio between rank Âh m and the number of elements in each time slot. C. Distance Estimation After projecting vector  into Âh,p with elements â h,p n,x, n = 1,..., N, considered aligned with the vessel heading direction (see Section IV-A) and horizontal plane (see Section IV-B), we can readily estimate the distance traveled by the vessel at time period [t start, t end ], d i,j. Given the initial velocity v i at time instance t start, we obtain estimation d i,j by integrating the projected acceleration measurements over the period [t start, t end ]. First, we obtain the mean velocity by, Then, we set d i,j = ˆv i,j (t end t start ). ˆv i,j = v i N n=2 â h,p n,x(t n t n 1 ). (22) D. Summarizing the Operation of the DR-A Method We now summarize the operation of our DR-A method. Referring to the pseudo-code in Algorithm 1, we first form vectors Âw, w = 1,..., W for W = tend t start T c time slots (line 1). For each wth time slot, we estimate α n and w, and project acceleration measurements in Âw into  h,w (line 3). To assist in classification of pitch-states, we set positive and negative constraints to form subsets l and Λ l, respectively (lines 4-11). Next, we perform initial classification (line 12) to calculate Θ 0 (line 13), and perform p last EM iterations to classify l (lines 14-18). For each wth time slot and mth pitch-state, we project elements in matrix Âh,w m to compensate on the vessel pitch angle (line 19), and group the projected elements to form vector  h,p (line 21). Finally, we evaluate distance d i,j (line 22). E. Discussion In this paper, we specifically assumed the vessel roll angle is fixed. However, due to ocean waves, much like the vessel pitch angle, the former can also be time-varying. While the roll

17 17 Algorithm 1 Evaluate d i,j from vector  1: Divide  into W time slots Âw 2: for w := 1 to W do {do for each time slot} 3: Estimate α n using (4) and w using (6), and project  w into Âh,w using (5) 4: l := 1, l â h,w 1, Λ l := 5: for n := 1 to Âh,w 1 do 6: if â h,w n,x Âh,w and â h,w n+1,x Âh,w satisfy positive constraints then 7: l â h,w n+1 8: else 9: l := l + 1, l â h,w n+1, Λ l 1 (n, n + 1) 10: end if 11: end for 12: classify Âh,w using the K-means algorithm to initially form matrixes (Âh,w Âh,w m 2], 13: µ 0 m,x := E[Âh,w m,x], µ 0 m,z := E[Âh,w m,z], σ m,x 0 = E[ m,x µ m,x) 0 ) 2], σ m,z 0 = E[(Âh,w m,z µ 0 m,z k0 m = Âh,w m {initial estimation} N slot 14: for p := 0 to p last 1 do {EM iterations} 15: Iteratively calculate (21) to obtain P (δ l = m Âh,w, Θ p ) 16: Solve (11) and (13) to obtain 17: end for Θ p+1 18: Classify subsets l using (14) and form Âh,w m, m = 1,..., M 19: Project elements in matrixes Âh,w m using (17) 20: end for 21: group all projected measurements to form vector  h,p 22: Evaluate d i,j using (22) angle is expected to change much slower than the pitch angle, the former may still change within a single time slot. As a result, our assumption that (per time slot) changes in acceleration are only due to the time-varying pitch angle, does not hold and classification to pitch-states cannot be made. A possible solution to this problem would be to identify the frequency of change of the vessel roll angle, and to define shorter time slots for which both acceleration in

18 18 the horizontal plane and vessel roll angle can be considered fixed. Classification of acceleration measurements is then performed for each of these (shorter) time slots. Since changes in the vessel roll angle also affects acceleration in the projected y-axis (which otherwise would be close to zero), the rate of change in the vessel roll angle can be estimated by observing periodic changes in the y component of Â. Clearly, this approach introduces more noise in estimating d i,j since classification is based on fewer measurements. V. PERFORMANCE EVALUATION We now evaluate the performance of our DR-A algorithm. The results are presented in terms of ρ err = d i,j d i,j and ρ angle = α n α n. Since the DR-A method is based on the Gaussian mixture model (2) as well as on the assumption that per-time slot acceleration in the horizontal plane, vessel heading direction, and vessel roll angle are fixed, to validate simulation results we also present results from a designated sea trial where we tested our algorithm. A. Simulation Results Our simulation setting includes a Monte-Carlo set of channel realizations. In each simulation, a vessel is moving for 60 sec in the x y plane with initial speed uniformly distributed between [0, 5] m/sec. The vector of acceleration in the horizontal plane, a n,x, n = 1,..., N, is sampled at rate 0.1 sec (i.e., N = 600), and generated as zero-mean colored Gaussian process with standard deviation of 1 m/(sec) 2 and a cross-correlation factor between adjacent samples of e 0.1 Tc, where T c = 6 sec. Likewise, the vessel heading, n is generated uniformly between [0, 2π] with a similar cross-correlation factor between adjacent samples. Furthermore, per-simulation, the (fixed) vessel orientation angle with respect to the reference coordinate system is uniformly randomized within the interval [0, 2π]. Based on the latter and n, we form vector α n, n = 1,..., N. Note that the initial speed and heading direction are known. Let h(x, y, t) be the time domain t function for the height of the sea surface for a modeled three-dimensional ocean wave in the x y plane. To simulate wave-based acceleration in the vertical plane, for t n being exactly in the middle of the time period before a local maxima and after a local minima of h(x, y, t), we uniformly randomize sample a n,z between [0.01, 0.05] g. Similarly, a n,z is generated uniformly between [ 0.05, 0.01] g for t n exactly in between a local maxima and a local minima. Acceleration then changes linearly with time such that it

19 19 6 Wave elevation [m] y [m] x [m] Fig. 4: Example of a modeled wave in the x y plane. t = 10 sec. reaches zero at both local maxima and minima of h(x, y, t). Since the derivative of h(x, y, t) is also the slope of the wave surface, the pitch angle ρ n at time sample t n and coordinates (x n, y n ) is computed from tan ρ n = the vector of acceleration measurements, Â. t n w(x, y, t n ). Using α n and ρ n and based on model (1), we form Current literature offers multiple models for the wind-based ocean surface wave function h(x, y, t) (e.g., [21], [22], [23],). In our simulations we use the analytical wave model offered in [23]. We note that similar results were obtained also for other wave models. For the ith wave frequency and jth directional angle, let θ i, φ i, and ψ i,j be the spreading directional angle, wave frequency, and an initial phase angle, respectively. The wave height is modeled as I J h(x, y, t) = 2S(φ i ) φ i θ j cos(c i x cos θ j + c i y sin θ j φ i t ψ i,j ), (23) i=1 j=1 where c i = φ2 i is the wave number, φ and θ are the increments of angles φ and θ, respectively, g ( ) ] 4 and we use the directional spectrum function S(φ) = b 1g 2 exp [ b 2 with b 1 = , b 2 = 0.74, and a wind speed U = 5 knots. For each channel realization, parameters φ i, θ i, and ψ i,j are uniformly randomized in intervals [1, 5] rad/sec, [0, 2π] rad, and [0, 2π] rad, respectively, and we use I = 10, J = 10, φ = 0.4 rad/sec, and θ = π 5 h(x, y, t) for t = 10 sec is shown in Figure 4. φ 5 g Uφ rad. An example for Since current methods for DR navigation use both acceleration and orientation measurements, we benchmark our algorithm by showing results of ρ err when an ideal gyroscope is used, i.e.,

20 20 Empirical Pr(ρ angle x) /ς=50dB 1/ς=30dB 1/ς=20dB 1/ς=10dB x [rad] Fig. 5: C-CDF results of ρ angle. perfect compensation of the vessel pitch and heading angles (Ideal-Gyro), and when orientation measurements are noisy (Noisy-Gyro). In addition, we compare results of DR-A with a direct integration of acceleration measurements with no heading and pitch compensation (Naive), and with an alternative method in which after heading estimation, using the PCA method (see Section II), per time slot we obtain measurements aligned with the horizontal plane (PCA). We also demonstrate tradeoffs between complexity and performance by replacing the constraint EM algorithm with i) the initial K-means classifier (K-means), ii) a simple slicing classifier (Slice), and iii) non-constraint EM (NC-EM). For the Ideal-Gyro and Noisy-Gyro methods, we estimate a n,x by multiplying â n with matrices (5) and (15) for the measured heading and pitch angles, ˆα i,j and ˆρ n, respectively. For the former we use ˆρ n = ρ n and ˆα n = α n, while for the latter we set ˆρ n = ρ n + ē n and ˆα n = α n + ē n, where ē n is a zero-mean Gaussian noise with variance ς rad 2, which is the same variance considered for the acceleration measurement noise. In the following, results for the DR-A method are shown for p last = 10 EM iterations and using V pos = 0.05 m/sec 2 for the positive constraints in (7). In Figure 5, we show the complementary cumulative density function (C-CDF) of ρ angle using estimation (4) for different values of 1. We observe that reasonable performance are obtained ς for relatively low noise values (around 20 db). Next, we show tradeoffs of performance vs. the number of pitch-states, M, and the assumed coherence time, Tc. In Figure 6a we show performance of the four considered classifying method in terms of ρ err for M {2, 10, 20} and

21 DR A NC EM Kmeans Slice DR A NC EM Kmeans Slice ρ err [m] 10 ρ err [m] M=2 M=5 M=10 M=20 0 T_c=3sec T_c=6sec T_c=12sec (a) (b) Fig. 6: Average results of ρ err as a function of (a) M (for Tc = T c ), (b) T c (for M = 10). T c = 6 sec. 1 ς = 20 db. We observe that for the Slice method performance are not linear with M. This is due to the fact that in the Slice method measurements are classified to all assumed M pitchstates, and thus, for each pitch-state, fewer measurements are available affecting accuracy of estimation (17). While accuracy increases with M for the DR-A, NC-EM, and Kmeans (which allow empty pitch-states), we observe that little is gained for M > 10. Next, for M = 10, in Figure 6b we show ρ err as a function of the assumed coherence time, Tc, which is used to obtain W time slices of assumed fixed acceleration in the horizontal plane. Here we observe a slight performance degradation for mismatch coherence time (i.e., when T c T c = 6 sec). When T c < T c, this is because fewer measurements are available to estimate the pitch angle in (17), while for T c > T c, performance degrade since our assumption of fixed acceleration in the horizontal plane (required for classification) does not hold. Interestingly, we observe that performance are less affected in the latter case. That is, having enough statistics to estimate the pitch angle is more important, as was also observed for the Slice method in Figure 6a. In the following we use M = 10 and T c = T c. In Figure 7, we show average results of ρ err for the considered classification methods and the benchmark methods as a function of 1. We observe a significant performance degradation for the ς

22 22 Noisy-Gyro method compared to the Ideal-Gyro one. We consider this performance gap as the maximal gain available for using our method. We also observe that for the Noisy-Gyro method, performance is not linear with 1 ς (in the logarithmic scale). This is because of the noisy orientation measurements which add non-gaussian noise to the projected acceleration measurements. As expected, performance for the Naive approach is poor, and is in fact fixed for different noise values. The latter is due to the periodic nature of the vessel pitch angle, which averages out positive and negative acceleration measurements in the x axis. Comparing the performance of the PCA method to those of our method, we observe that PCA is better than the Slice method, mostly due to the naive classification performed in the Slice method which is largely affected by measurement noise. However, using the better classification capabilities of the EM and K- means algorithm, we observe considerable performance gain compared to the PCA method. This is due to the underline assumption in PCA of the variance of acceleration measurements (see Section II), which might not hold for all modeled ocean waves. As expected, performance of the EM algorithm, which is matched to the Gaussian mixture model in (2), outperforms that of the K-means algorithm. Moreover, significant improvement is achieved using our DR-A method compared to that of the non-constraint EM algorithm, NC-EM. From Figure 7, we observe that the performance of our DR-A method is close to that of the (unrealistic) Ideal-Gyro method. Thus, we conclude that, without using orientation measurements, the DR-A method almost entirely compensates on the vessel s heading angle and time-varying pitch angle. To comment on the distribution of the performance, in Figure 8 we show the C-CDF of ρ err for 1 ς = 20 db. Results show that the above conclusions, drawn for the average ρ err results, hold true for all simulated scenarios. In the following we present results of offline processing of 3-D acceleration measurements obtained in a sea trial. B. Sea-Trial Results In this section, we describe results of our DR-A method obtained from real sea environment. The sea trial was conducted on Nov in the Singapore straight. The experiment lasted for two hours and included two boats. During the experiment, the waves height was about 0.5 m and the boats drifted with the ocean current at about 0.7 m/sec. At each boat, we obtained around N = 33.5 k 3-D acceleration measurements using the Libelium Wasp Mote s on board

23 ρ err [m] DR A NC EM Kmeans Slice PCA Ideal Gyro Noisy Gyro Naive /ς [db] Fig. 7: Results of ρ err as a function of 1 ς. M = 10, Tc = T c. Empirical Pr(ρ err x) DR A NC EM Kmeans Slice PCA Ideal Gyro Noisy Gyro x [m] Fig. 8: C-CDF results of ρ err. 1 ς = 20 db, M = 10, Tc = T c. accelerometer at rate of f = 4.8 Hz, and used its serial port for data logging purpose and offline decoding. Throughout the experiment, the location of the boats was monitored using GPS receivers at rate of 3 sec and with expected accuracy of 5 m. The boats GPS-based location is shown in Figure 9. From the figure we observe that, as a result of the ocean current, the boats changed their heading direction. However, other than around time t = 25 min (where boat 2 had to maneuver around an obstacle), this heading change is slow and fits our underline assumption. Furthermore, from Figure 9 we observe a slow change in the boats speed (with variance of 0.1 m/sec 2 ), which allows using time slices of assumed fixed acceleration in the horizontal

24 time [min] Boat 1 Boat y [m] x [m] Fig. 9: Sea trial: nodes location in Cartesian coordinates. plane. In Figure 10, we show the measured acceleration along the x axis as a function of earth gravity, g (the results for the projected accelerations are discussed further below). We observe that acceleration measurements follow a wave pattern, and that both frequency and amplitude of these waves are different for the two boats. The periodic nature of the measurements shown in Figure 10 emphasize the need for compensating on the vessel pitch angle, as direct orientation measurements may be too noisy. Moreover, due to the observed fast time-varying measurements (caused by ripples), mitigating the effect of the vessel s pitch angle using filtering is not possible without loss of resolution. First, in Figure 11 we show C-CDF performance of the heading estimate for the two boats, where the vessel heading direction with respect to the UTM coordinate system is calculated based on the boats GPS-based location. Results are compared for consecutive time slots of duration T slot such that N estimations of ρ T slot f angle are obtained. Since accuracy of estimation (4) improves with the number of measurements but decreases if the heading angle changes within the time slot, results tradeoff for T slot. We observe that best results are obtained for T slot = 200 sec. From the figure, we observe a good fit of less than 1 degree between the GPS-based heading angle and the estimated one. However, we note that when the vessels heading angle changes within the time slot, estimation accuracy is poor. In Figure 10, for the first 50 sec of the experiment we show the measured and projected acceleration along the x axis, where the latter are obtained using M = 5 and T c = 40 sec. We

25 Acceleration x axis (g) Boat 1 (projected) Boat 1 (measured) 0.04 Boat 2 (projected) Boat 2 (measured) Time (sec) Fig. 10: Sea trial: measured and projected acceleration along the x axis relative to g for M = 5 and T c = 40 sec. Empirical Pr(ρ α x) T slot =60sec T slot =120sec T slot =200sec T slot =240sec x [rad] Fig. 11: Sea trial: C-CDF results of ρ angle for the two boats. observe that the projected measurements are almost constant, which matches the expected slow change of the boats acceleration in the horizontal plane. Finally, in Figure 12 we present C-CDF results of ρ err, comparing estimation ˆd i,j to the GPS-based distance for time slots of duration T slot = 200 sec and for different values of M and Tc. For each time slot, the initial speed v i is calculated based on the first two GPS-based locations of the boats. From the figure we observe that in some cases ρ err > 10 m (which is above the expected error due to GPS uncertainty). As in the estimation of the heading angle, these cases occur when there is a sudden shift in

26 26 Empirical Pr(ρ err x) M=2,T c =20sec M=8,T c =40sec M=5,T c =60sec M=5,T c =40sec x [m] Fig. 12: Sea trial: C-CDF results of ρ err for the two boats. T slot = 200 sec the boats heading direction. A possible way to reduce the effect of this shortcoming of our method is to avoid relying on heading and distance estimations whenever a sudden change from the previous estimation is observed. However, for M = 5 and Tc = 40 sec, we observe that in more than 96% of cases ρ err is lower than the expected error. Considering this result and the low estimation error observed for the heading angle, we conclude that when the boats heading is constant, our method fully compensates for the vessel s time-varying pitch angle using only a single accelerometer device. VI. CONCLUSION In this paper, for DR navigation of a vessel whose motion is affected by the ocean waves, we described a method to estimate the vessel heading and the distance traveled by the vessel using only a single 3-D accelerometer. This is required when measurement of the vessel orientation using, e.g., gyrocompass, is not available or is too noisy to directly compensate for the vessel pitch and heading angles. Using only 3-D acceleration measurements, the major challenge in such estimating is the waves-induced vessel s time-varying pitch angle. Considering this problem, based on the periodic nature of the vessel pitch angle we described a machine learning classification approach that forms classes of acceleration measurements for which the pitch angle is similar. Per-class, our method estimates the vessel pitch angle, and projects the available acceleration measurements into the horizontal plane. The projected measurements are

Linear Dynamical Systems

Linear Dynamical Systems Linear Dynamical Systems Sargur N. srihari@cedar.buffalo.edu Machine Learning Course: http://www.cedar.buffalo.edu/~srihari/cse574/index.html Two Models Described by Same Graph Latent variables Observations

More information

SLAM for Ship Hull Inspection using Exactly Sparse Extended Information Filters

SLAM for Ship Hull Inspection using Exactly Sparse Extended Information Filters SLAM for Ship Hull Inspection using Exactly Sparse Extended Information Filters Matthew Walter 1,2, Franz Hover 1, & John Leonard 1,2 Massachusetts Institute of Technology 1 Department of Mechanical Engineering

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2016 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2016 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

Mobile Robot Localization

Mobile Robot Localization Mobile Robot Localization 1 The Problem of Robot Localization Given a map of the environment, how can a robot determine its pose (planar coordinates + orientation)? Two sources of uncertainty: - observations

More information

1 Kalman Filter Introduction

1 Kalman Filter Introduction 1 Kalman Filter Introduction You should first read Chapter 1 of Stochastic models, estimation, and control: Volume 1 by Peter S. Maybec (available here). 1.1 Explanation of Equations (1-3) and (1-4) Equation

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project

More information

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014

Bayesian Networks: Construction, Inference, Learning and Causal Interpretation. Volker Tresp Summer 2014 Bayesian Networks: Construction, Inference, Learning and Causal Interpretation Volker Tresp Summer 2014 1 Introduction So far we were mostly concerned with supervised learning: we predicted one or several

More information

Notes on Machine Learning for and

Notes on Machine Learning for and Notes on Machine Learning for 16.410 and 16.413 (Notes adapted from Tom Mitchell and Andrew Moore.) Choosing Hypotheses Generally want the most probable hypothesis given the training data Maximum a posteriori

More information

Artificial Intelligence

Artificial Intelligence Artificial Intelligence Roman Barták Department of Theoretical Computer Science and Mathematical Logic Summary of last lecture We know how to do probabilistic reasoning over time transition model P(X t

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistical Sciences! rsalakhu@cs.toronto.edu! h0p://www.cs.utoronto.ca/~rsalakhu/ Lecture 7 Approximate

More information

Approximate Inference Part 1 of 2

Approximate Inference Part 1 of 2 Approximate Inference Part 1 of 2 Tom Minka Microsoft Research, Cambridge, UK Machine Learning Summer School 2009 http://mlg.eng.cam.ac.uk/mlss09/ Bayesian paradigm Consistent use of probability theory

More information

LEARNING DYNAMIC SYSTEMS: MARKOV MODELS

LEARNING DYNAMIC SYSTEMS: MARKOV MODELS LEARNING DYNAMIC SYSTEMS: MARKOV MODELS Markov Process and Markov Chains Hidden Markov Models Kalman Filters Types of dynamic systems Problem of future state prediction Predictability Observability Easily

More information

Probabilistic Graphical Models

Probabilistic Graphical Models 2016 Robert Nowak Probabilistic Graphical Models 1 Introduction We have focused mainly on linear models for signals, in particular the subspace model x = Uθ, where U is a n k matrix and θ R k is a vector

More information

Application of state observers in attitude estimation using low-cost sensors

Application of state observers in attitude estimation using low-cost sensors Application of state observers in attitude estimation using low-cost sensors Martin Řezáč Czech Technical University in Prague, Czech Republic March 26, 212 Introduction motivation for inertial estimation

More information

LOS and NLOS Classification for Underwater Acoustic Localization

LOS and NLOS Classification for Underwater Acoustic Localization LOS and NLOS Classification for Underwater Acoustic Localization Roee Diamant, Hwee-Pink Tan and Lutz Lampe Electrical and Computer Engineering, The University of British Columbia Institute for Infocomm

More information

Mobile Robot Localization

Mobile Robot Localization Mobile Robot Localization 1 The Problem of Robot Localization Given a map of the environment, how can a robot determine its pose (planar coordinates + orientation)? Two sources of uncertainty: - observations

More information

Approximate Inference Part 1 of 2

Approximate Inference Part 1 of 2 Approximate Inference Part 1 of 2 Tom Minka Microsoft Research, Cambridge, UK Machine Learning Summer School 2009 http://mlg.eng.cam.ac.uk/mlss09/ 1 Bayesian paradigm Consistent use of probability theory

More information

The Effect of Stale Ranging Data on Indoor 2-D Passive Localization

The Effect of Stale Ranging Data on Indoor 2-D Passive Localization The Effect of Stale Ranging Data on Indoor 2-D Passive Localization Chen Xia and Lance C. Pérez Department of Electrical Engineering University of Nebraska-Lincoln, USA chenxia@mariner.unl.edu lperez@unl.edu

More information

Recent Advances in Bayesian Inference Techniques

Recent Advances in Bayesian Inference Techniques Recent Advances in Bayesian Inference Techniques Christopher M. Bishop Microsoft Research, Cambridge, U.K. research.microsoft.com/~cmbishop SIAM Conference on Data Mining, April 2004 Abstract Bayesian

More information

Inertial Odometry on Handheld Smartphones

Inertial Odometry on Handheld Smartphones Inertial Odometry on Handheld Smartphones Arno Solin 1 Santiago Cortés 1 Esa Rahtu 2 Juho Kannala 1 1 Aalto University 2 Tampere University of Technology 21st International Conference on Information Fusion

More information

VN-100 Velocity Compensation

VN-100 Velocity Compensation VN-100 Velocity Compensation Velocity / Airspeed Aiding for AHRS Applications Application Note Abstract This application note describes how the VN-100 can be used in non-stationary applications which require

More information

MEMS Gyroscope Control Systems for Direct Angle Measurements

MEMS Gyroscope Control Systems for Direct Angle Measurements MEMS Gyroscope Control Systems for Direct Angle Measurements Chien-Yu Chi Mechanical Engineering National Chiao Tung University Hsin-Chu, Taiwan (R.O.C.) 3 Email: chienyu.me93g@nctu.edu.tw Tsung-Lin Chen

More information

Pattern Recognition and Machine Learning

Pattern Recognition and Machine Learning Christopher M. Bishop Pattern Recognition and Machine Learning ÖSpri inger Contents Preface Mathematical notation Contents vii xi xiii 1 Introduction 1 1.1 Example: Polynomial Curve Fitting 4 1.2 Probability

More information

Autonomous Helicopter Flight via Reinforcement Learning

Autonomous Helicopter Flight via Reinforcement Learning Autonomous Helicopter Flight via Reinforcement Learning Authors: Andrew Y. Ng, H. Jin Kim, Michael I. Jordan, Shankar Sastry Presenters: Shiv Ballianda, Jerrolyn Hebert, Shuiwang Ji, Kenley Malveaux, Huy

More information

Bayesian Machine Learning

Bayesian Machine Learning Bayesian Machine Learning Andrew Gordon Wilson ORIE 6741 Lecture 4 Occam s Razor, Model Construction, and Directed Graphical Models https://people.orie.cornell.edu/andrew/orie6741 Cornell University September

More information

Machine Learning for Data Science (CS4786) Lecture 24

Machine Learning for Data Science (CS4786) Lecture 24 Machine Learning for Data Science (CS4786) Lecture 24 Graphical Models: Approximate Inference Course Webpage : http://www.cs.cornell.edu/courses/cs4786/2016sp/ BELIEF PROPAGATION OR MESSAGE PASSING Each

More information

UAV Navigation: Airborne Inertial SLAM

UAV Navigation: Airborne Inertial SLAM Introduction UAV Navigation: Airborne Inertial SLAM Jonghyuk Kim Faculty of Engineering and Information Technology Australian National University, Australia Salah Sukkarieh ARC Centre of Excellence in

More information

Introduction to Machine Learning

Introduction to Machine Learning Introduction to Machine Learning Brown University CSCI 1950-F, Spring 2012 Prof. Erik Sudderth Lecture 25: Markov Chain Monte Carlo (MCMC) Course Review and Advanced Topics Many figures courtesy Kevin

More information

A Tree Search Approach to Target Tracking in Clutter

A Tree Search Approach to Target Tracking in Clutter 12th International Conference on Information Fusion Seattle, WA, USA, July 6-9, 2009 A Tree Search Approach to Target Tracking in Clutter Jill K. Nelson and Hossein Roufarshbaf Department of Electrical

More information

Clustering. Professor Ameet Talwalkar. Professor Ameet Talwalkar CS260 Machine Learning Algorithms March 8, / 26

Clustering. Professor Ameet Talwalkar. Professor Ameet Talwalkar CS260 Machine Learning Algorithms March 8, / 26 Clustering Professor Ameet Talwalkar Professor Ameet Talwalkar CS26 Machine Learning Algorithms March 8, 217 1 / 26 Outline 1 Administration 2 Review of last lecture 3 Clustering Professor Ameet Talwalkar

More information

p(d θ ) l(θ ) 1.2 x x x

p(d θ ) l(θ ) 1.2 x x x p(d θ ).2 x 0-7 0.8 x 0-7 0.4 x 0-7 l(θ ) -20-40 -60-80 -00 2 3 4 5 6 7 θ ˆ 2 3 4 5 6 7 θ ˆ 2 3 4 5 6 7 θ θ x FIGURE 3.. The top graph shows several training points in one dimension, known or assumed to

More information

Massachusetts Institute of Technology

Massachusetts Institute of Technology Massachusetts Institute of Technology 6.867 Machine Learning, Fall 2006 Problem Set 5 Due Date: Thursday, Nov 30, 12:00 noon You may submit your solutions in class or in the box. 1. Wilhelm and Klaus are

More information

Multi-Robotic Systems

Multi-Robotic Systems CHAPTER 9 Multi-Robotic Systems The topic of multi-robotic systems is quite popular now. It is believed that such systems can have the following benefits: Improved performance ( winning by numbers ) Distributed

More information

Lecture Notes 4 Vector Detection and Estimation. Vector Detection Reconstruction Problem Detection for Vector AGN Channel

Lecture Notes 4 Vector Detection and Estimation. Vector Detection Reconstruction Problem Detection for Vector AGN Channel Lecture Notes 4 Vector Detection and Estimation Vector Detection Reconstruction Problem Detection for Vector AGN Channel Vector Linear Estimation Linear Innovation Sequence Kalman Filter EE 278B: Random

More information

Fundamentals of attitude Estimation

Fundamentals of attitude Estimation Fundamentals of attitude Estimation Prepared by A.Kaviyarasu Assistant Professor Department of Aerospace Engineering Madras Institute Of Technology Chromepet, Chennai Basically an IMU can used for two

More information

EE565:Mobile Robotics Lecture 6

EE565:Mobile Robotics Lecture 6 EE565:Mobile Robotics Lecture 6 Welcome Dr. Ahmad Kamal Nasir Announcement Mid-Term Examination # 1 (25%) Understand basic wheel robot kinematics, common mobile robot sensors and actuators knowledge. Understand

More information

We provide two sections from the book (in preparation) Intelligent and Autonomous Road Vehicles, by Ozguner, Acarman and Redmill.

We provide two sections from the book (in preparation) Intelligent and Autonomous Road Vehicles, by Ozguner, Acarman and Redmill. We provide two sections from the book (in preparation) Intelligent and Autonomous Road Vehicles, by Ozguner, Acarman and Redmill. 2.3.2. Steering control using point mass model: Open loop commands We consider

More information

CS281 Section 4: Factor Analysis and PCA

CS281 Section 4: Factor Analysis and PCA CS81 Section 4: Factor Analysis and PCA Scott Linderman At this point we have seen a variety of machine learning models, with a particular emphasis on models for supervised learning. In particular, we

More information

Part 1: Expectation Propagation

Part 1: Expectation Propagation Chalmers Machine Learning Summer School Approximate message passing and biomedicine Part 1: Expectation Propagation Tom Heskes Machine Learning Group, Institute for Computing and Information Sciences Radboud

More information

Locating and supervising relief forces in buildings without the use of infrastructure

Locating and supervising relief forces in buildings without the use of infrastructure Locating and supervising relief forces in buildings without the use of infrastructure Tracking of position with low-cost inertial sensors Martin Trächtler 17.10.2014 18th Leibniz Conference of advanced

More information

with Application to Autonomous Vehicles

with Application to Autonomous Vehicles Nonlinear with Application to Autonomous Vehicles (Ph.D. Candidate) C. Silvestre (Supervisor) P. Oliveira (Co-supervisor) Institute for s and Robotics Instituto Superior Técnico Portugal January 2010 Presentation

More information

Gaussian Mixture Models

Gaussian Mixture Models Gaussian Mixture Models Pradeep Ravikumar Co-instructor: Manuela Veloso Machine Learning 10-701 Some slides courtesy of Eric Xing, Carlos Guestrin (One) bad case for K- means Clusters may overlap Some

More information

Probability Map Building of Uncertain Dynamic Environments with Indistinguishable Obstacles

Probability Map Building of Uncertain Dynamic Environments with Indistinguishable Obstacles Probability Map Building of Uncertain Dynamic Environments with Indistinguishable Obstacles Myungsoo Jun and Raffaello D Andrea Sibley School of Mechanical and Aerospace Engineering Cornell University

More information

CS491/691: Introduction to Aerial Robotics

CS491/691: Introduction to Aerial Robotics CS491/691: Introduction to Aerial Robotics Topic: Midterm Preparation Dr. Kostas Alexis (CSE) Areas of Focus Coordinate system transformations (CST) MAV Dynamics (MAVD) Navigation Sensors (NS) State Estimation

More information

Probabilistic & Unsupervised Learning

Probabilistic & Unsupervised Learning Probabilistic & Unsupervised Learning Week 2: Latent Variable Models Maneesh Sahani maneesh@gatsby.ucl.ac.uk Gatsby Computational Neuroscience Unit, and MSc ML/CSML, Dept Computer Science University College

More information

GWAS V: Gaussian processes

GWAS V: Gaussian processes GWAS V: Gaussian processes Dr. Oliver Stegle Christoh Lippert Prof. Dr. Karsten Borgwardt Max-Planck-Institutes Tübingen, Germany Tübingen Summer 2011 Oliver Stegle GWAS V: Gaussian processes Summer 2011

More information

Dynamic Multipath Estimation by Sequential Monte Carlo Methods

Dynamic Multipath Estimation by Sequential Monte Carlo Methods Dynamic Multipath Estimation by Sequential Monte Carlo Methods M. Lentmaier, B. Krach, P. Robertson, and T. Thiasiriphet German Aerospace Center (DLR) Slide 1 Outline Multipath problem and signal model

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Lecture 12 Dynamical Models CS/CNS/EE 155 Andreas Krause Homework 3 out tonight Start early!! Announcements Project milestones due today Please email to TAs 2 Parameter learning

More information

Introduction to Bayesian methods in inverse problems

Introduction to Bayesian methods in inverse problems Introduction to Bayesian methods in inverse problems Ville Kolehmainen 1 1 Department of Applied Physics, University of Eastern Finland, Kuopio, Finland March 4 2013 Manchester, UK. Contents Introduction

More information

Miscellaneous. Regarding reading materials. Again, ask questions (if you have) and ask them earlier

Miscellaneous. Regarding reading materials. Again, ask questions (if you have) and ask them earlier Miscellaneous Regarding reading materials Reading materials will be provided as needed If no assigned reading, it means I think the material from class is sufficient Should be enough for you to do your

More information

9 Forward-backward algorithm, sum-product on factor graphs

9 Forward-backward algorithm, sum-product on factor graphs Massachusetts Institute of Technology Department of Electrical Engineering and Computer Science 6.438 Algorithms For Inference Fall 2014 9 Forward-backward algorithm, sum-product on factor graphs The previous

More information

CSCI-567: Machine Learning (Spring 2019)

CSCI-567: Machine Learning (Spring 2019) CSCI-567: Machine Learning (Spring 2019) Prof. Victor Adamchik U of Southern California Mar. 19, 2019 March 19, 2019 1 / 43 Administration March 19, 2019 2 / 43 Administration TA3 is due this week March

More information

The Expectation-Maximization Algorithm

The Expectation-Maximization Algorithm 1/29 EM & Latent Variable Models Gaussian Mixture Models EM Theory The Expectation-Maximization Algorithm Mihaela van der Schaar Department of Engineering Science University of Oxford MLE for Latent Variable

More information

Overview of Statistical Tools. Statistical Inference. Bayesian Framework. Modeling. Very simple case. Things are usually more complicated

Overview of Statistical Tools. Statistical Inference. Bayesian Framework. Modeling. Very simple case. Things are usually more complicated Fall 3 Computer Vision Overview of Statistical Tools Statistical Inference Haibin Ling Observation inference Decision Prior knowledge http://www.dabi.temple.edu/~hbling/teaching/3f_5543/index.html Bayesian

More information

11. Learning graphical models

11. Learning graphical models Learning graphical models 11-1 11. Learning graphical models Maximum likelihood Parameter learning Structural learning Learning partially observed graphical models Learning graphical models 11-2 statistical

More information

Computer Vision Group Prof. Daniel Cremers. 6. Mixture Models and Expectation-Maximization

Computer Vision Group Prof. Daniel Cremers. 6. Mixture Models and Expectation-Maximization Prof. Daniel Cremers 6. Mixture Models and Expectation-Maximization Motivation Often the introduction of latent (unobserved) random variables into a model can help to express complex (marginal) distributions

More information

Naïve Bayes classification

Naïve Bayes classification Naïve Bayes classification 1 Probability theory Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. Examples: A person s height, the outcome of a coin toss

More information

Problem 1: Ship Path-Following Control System (35%)

Problem 1: Ship Path-Following Control System (35%) Problem 1: Ship Path-Following Control System (35%) Consider the kinematic equations: Figure 1: NTNU s research vessel, R/V Gunnerus, and Nomoto model: T ṙ + r = Kδ (1) with T = 22.0 s and K = 0.1 s 1.

More information

13 : Variational Inference: Loopy Belief Propagation and Mean Field

13 : Variational Inference: Loopy Belief Propagation and Mean Field 10-708: Probabilistic Graphical Models 10-708, Spring 2012 13 : Variational Inference: Loopy Belief Propagation and Mean Field Lecturer: Eric P. Xing Scribes: Peter Schulam and William Wang 1 Introduction

More information

Practical Bayesian Optimization of Machine Learning. Learning Algorithms

Practical Bayesian Optimization of Machine Learning. Learning Algorithms Practical Bayesian Optimization of Machine Learning Algorithms CS 294 University of California, Berkeley Tuesday, April 20, 2016 Motivation Machine Learning Algorithms (MLA s) have hyperparameters that

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Brown University CSCI 2950-P, Spring 2013 Prof. Erik Sudderth Lecture 13: Learning in Gaussian Graphical Models, Non-Gaussian Inference, Monte Carlo Methods Some figures

More information

Linear Regression and Its Applications

Linear Regression and Its Applications Linear Regression and Its Applications Predrag Radivojac October 13, 2014 Given a data set D = {(x i, y i )} n the objective is to learn the relationship between features and the target. We usually start

More information

p L yi z n m x N n xi

p L yi z n m x N n xi y i z n x n N x i Overview Directed and undirected graphs Conditional independence Exact inference Latent variables and EM Variational inference Books statistical perspective Graphical Models, S. Lauritzen

More information

CS 630 Basic Probability and Information Theory. Tim Campbell

CS 630 Basic Probability and Information Theory. Tim Campbell CS 630 Basic Probability and Information Theory Tim Campbell 21 January 2003 Probability Theory Probability Theory is the study of how best to predict outcomes of events. An experiment (or trial or event)

More information

STONY BROOK UNIVERSITY. CEAS Technical Report 829

STONY BROOK UNIVERSITY. CEAS Technical Report 829 1 STONY BROOK UNIVERSITY CEAS Technical Report 829 Variable and Multiple Target Tracking by Particle Filtering and Maximum Likelihood Monte Carlo Method Jaechan Lim January 4, 2006 2 Abstract In most applications

More information

Undirected Graphical Models

Undirected Graphical Models Outline Hong Chang Institute of Computing Technology, Chinese Academy of Sciences Machine Learning Methods (Fall 2012) Outline Outline I 1 Introduction 2 Properties Properties 3 Generative vs. Conditional

More information

STA 414/2104: Machine Learning

STA 414/2104: Machine Learning STA 414/2104: Machine Learning Russ Salakhutdinov Department of Computer Science! Department of Statistics! rsalakhu@cs.toronto.edu! http://www.cs.toronto.edu/~rsalakhu/ Lecture 9 Sequential Data So far

More information

Arrow Brasil. Rodrigo Rodrigues Field Application Engineer F: Date: 30/01/2014 TM 2

Arrow Brasil. Rodrigo Rodrigues Field Application Engineer F: Date: 30/01/2014 TM 2 TM Arrow Brasil Rodrigo Rodrigues Field Application Engineer Rodrigo.rodrigues@arrowbrasil.com.br F:+55 11 3613-9331 Date: 30/01/2014 TM 2 State-of-the-art review Introduction How a Gyro Works Performance

More information

ABSTRACT INTRODUCTION

ABSTRACT INTRODUCTION ABSTRACT Presented in this paper is an approach to fault diagnosis based on a unifying review of linear Gaussian models. The unifying review draws together different algorithms such as PCA, factor analysis,

More information

Probabilistic Graphical Models

Probabilistic Graphical Models Probabilistic Graphical Models Brown University CSCI 295-P, Spring 213 Prof. Erik Sudderth Lecture 11: Inference & Learning Overview, Gaussian Graphical Models Some figures courtesy Michael Jordan s draft

More information

Inertial Navigation and Various Applications of Inertial Data. Yongcai Wang. 9 November 2016

Inertial Navigation and Various Applications of Inertial Data. Yongcai Wang. 9 November 2016 Inertial Navigation and Various Applications of Inertial Data Yongcai Wang 9 November 2016 Types of Gyroscope Mechanical Gyroscope Laser Gyroscope Sagnac Effect Stable Platform IMU and Strapdown IMU In

More information

Evaluation of different wind estimation methods in flight tests with a fixed-wing UAV

Evaluation of different wind estimation methods in flight tests with a fixed-wing UAV Evaluation of different wind estimation methods in flight tests with a fixed-wing UAV Julian Sören Lorenz February 5, 2018 Contents 1 Glossary 2 2 Introduction 3 3 Tested algorithms 3 3.1 Unfiltered Method

More information

Machine Learning 4771

Machine Learning 4771 Machine Learning 4771 Instructor: ony Jebara Kalman Filtering Linear Dynamical Systems and Kalman Filtering Structure from Motion Linear Dynamical Systems Audio: x=pitch y=acoustic waveform Vision: x=object

More information

6.867 Machine learning, lecture 23 (Jaakkola)

6.867 Machine learning, lecture 23 (Jaakkola) Lecture topics: Markov Random Fields Probabilistic inference Markov Random Fields We will briefly go over undirected graphical models or Markov Random Fields (MRFs) as they will be needed in the context

More information

A Study of Covariances within Basic and Extended Kalman Filters

A Study of Covariances within Basic and Extended Kalman Filters A Study of Covariances within Basic and Extended Kalman Filters David Wheeler Kyle Ingersoll December 2, 2013 Abstract This paper explores the role of covariance in the context of Kalman filters. The underlying

More information

Mark your answers ON THE EXAM ITSELF. If you are not sure of your answer you may wish to provide a brief explanation.

Mark your answers ON THE EXAM ITSELF. If you are not sure of your answer you may wish to provide a brief explanation. CS 189 Spring 2015 Introduction to Machine Learning Midterm You have 80 minutes for the exam. The exam is closed book, closed notes except your one-page crib sheet. No calculators or electronic items.

More information

Expectation propagation for signal detection in flat-fading channels

Expectation propagation for signal detection in flat-fading channels Expectation propagation for signal detection in flat-fading channels Yuan Qi MIT Media Lab Cambridge, MA, 02139 USA yuanqi@media.mit.edu Thomas Minka CMU Statistics Department Pittsburgh, PA 15213 USA

More information

Capacity of a Two-way Function Multicast Channel

Capacity of a Two-way Function Multicast Channel Capacity of a Two-way Function Multicast Channel 1 Seiyun Shin, Student Member, IEEE and Changho Suh, Member, IEEE Abstract We explore the role of interaction for the problem of reliable computation over

More information

10708 Graphical Models: Homework 2

10708 Graphical Models: Homework 2 10708 Graphical Models: Homework 2 Due Monday, March 18, beginning of class Feburary 27, 2013 Instructions: There are five questions (one for extra credit) on this assignment. There is a problem involves

More information

Dead Reckoning navigation (DR navigation)

Dead Reckoning navigation (DR navigation) Dead Reckoning navigation (DR navigation) Prepared by A.Kaviyarasu Assistant Professor Department of Aerospace Engineering Madras Institute Of Technology Chromepet, Chennai A Navigation which uses a Inertial

More information

PROBABILISTIC REASONING OVER TIME

PROBABILISTIC REASONING OVER TIME PROBABILISTIC REASONING OVER TIME In which we try to interpret the present, understand the past, and perhaps predict the future, even when very little is crystal clear. Outline Time and uncertainty Inference:

More information

F denotes cumulative density. denotes probability density function; (.)

F denotes cumulative density. denotes probability density function; (.) BAYESIAN ANALYSIS: FOREWORDS Notation. System means the real thing and a model is an assumed mathematical form for the system.. he probability model class M contains the set of the all admissible models

More information

ECE521 Tutorial 11. Topic Review. ECE521 Winter Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides. ECE521 Tutorial 11 / 4

ECE521 Tutorial 11. Topic Review. ECE521 Winter Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides. ECE521 Tutorial 11 / 4 ECE52 Tutorial Topic Review ECE52 Winter 206 Credits to Alireza Makhzani, Alex Schwing, Rich Zemel and TAs for slides ECE52 Tutorial ECE52 Winter 206 Credits to Alireza / 4 Outline K-means, PCA 2 Bayesian

More information

Machine Learning, Midterm Exam

Machine Learning, Midterm Exam 10-601 Machine Learning, Midterm Exam Instructors: Tom Mitchell, Ziv Bar-Joseph Wednesday 12 th December, 2012 There are 9 questions, for a total of 100 points. This exam has 20 pages, make sure you have

More information

Inference and estimation in probabilistic time series models

Inference and estimation in probabilistic time series models 1 Inference and estimation in probabilistic time series models David Barber, A Taylan Cemgil and Silvia Chiappa 11 Time series The term time series refers to data that can be represented as a sequence

More information

Sensors for mobile robots

Sensors for mobile robots ROBOTICS 01PEEQW Basilio Bona DAUIN Politecnico di Torino Mobile & Service Robotics Sensors for Robotics 2 Sensors for mobile robots Sensors are used to perceive, analyze and understand the environment

More information

Hidden Markov Models. Aarti Singh Slides courtesy: Eric Xing. Machine Learning / Nov 8, 2010

Hidden Markov Models. Aarti Singh Slides courtesy: Eric Xing. Machine Learning / Nov 8, 2010 Hidden Markov Models Aarti Singh Slides courtesy: Eric Xing Machine Learning 10-701/15-781 Nov 8, 2010 i.i.d to sequential data So far we assumed independent, identically distributed data Sequential data

More information

10-701/15-781, Machine Learning: Homework 4

10-701/15-781, Machine Learning: Homework 4 10-701/15-781, Machine Learning: Homewor 4 Aarti Singh Carnegie Mellon University ˆ The assignment is due at 10:30 am beginning of class on Mon, Nov 15, 2010. ˆ Separate you answers into five parts, one

More information

Independent Component Analysis and Unsupervised Learning

Independent Component Analysis and Unsupervised Learning Independent Component Analysis and Unsupervised Learning Jen-Tzung Chien National Cheng Kung University TABLE OF CONTENTS 1. Independent Component Analysis 2. Case Study I: Speech Recognition Independent

More information

Human Mobility Pattern Prediction Algorithm using Mobile Device Location and Time Data

Human Mobility Pattern Prediction Algorithm using Mobile Device Location and Time Data Human Mobility Pattern Prediction Algorithm using Mobile Device Location and Time Data 0. Notations Myungjun Choi, Yonghyun Ro, Han Lee N = number of states in the model T = length of observation sequence

More information

Robust Monte Carlo Methods for Sequential Planning and Decision Making

Robust Monte Carlo Methods for Sequential Planning and Decision Making Robust Monte Carlo Methods for Sequential Planning and Decision Making Sue Zheng, Jason Pacheco, & John Fisher Sensing, Learning, & Inference Group Computer Science & Artificial Intelligence Laboratory

More information

Sensor Tasking and Control

Sensor Tasking and Control Sensor Tasking and Control Sensing Networking Leonidas Guibas Stanford University Computation CS428 Sensor systems are about sensing, after all... System State Continuous and Discrete Variables The quantities

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 3 Linear

More information

MMSE-Based Filtering for Linear and Nonlinear Systems in the Presence of Non-Gaussian System and Measurement Noise

MMSE-Based Filtering for Linear and Nonlinear Systems in the Presence of Non-Gaussian System and Measurement Noise MMSE-Based Filtering for Linear and Nonlinear Systems in the Presence of Non-Gaussian System and Measurement Noise I. Bilik 1 and J. Tabrikian 2 1 Dept. of Electrical and Computer Engineering, University

More information

ENGR352 Problem Set 02

ENGR352 Problem Set 02 engr352/engr352p02 September 13, 2018) ENGR352 Problem Set 02 Transfer function of an estimator 1. Using Eq. (1.1.4-27) from the text, find the correct value of r ss (the result given in the text is incorrect).

More information

CSCE 478/878 Lecture 6: Bayesian Learning

CSCE 478/878 Lecture 6: Bayesian Learning Bayesian Methods Not all hypotheses are created equal (even if they are all consistent with the training data) Outline CSCE 478/878 Lecture 6: Bayesian Learning Stephen D. Scott (Adapted from Tom Mitchell

More information

Bayesian Learning. Two Roles for Bayesian Methods. Bayes Theorem. Choosing Hypotheses

Bayesian Learning. Two Roles for Bayesian Methods. Bayes Theorem. Choosing Hypotheses Bayesian Learning Two Roles for Bayesian Methods Probabilistic approach to inference. Quantities of interest are governed by prob. dist. and optimal decisions can be made by reasoning about these prob.

More information

Naïve Bayes classification. p ij 11/15/16. Probability theory. Probability theory. Probability theory. X P (X = x i )=1 i. Marginal Probability

Naïve Bayes classification. p ij 11/15/16. Probability theory. Probability theory. Probability theory. X P (X = x i )=1 i. Marginal Probability Probability theory Naïve Bayes classification Random variable: a variable whose possible values are numerical outcomes of a random phenomenon. s: A person s height, the outcome of a coin toss Distinguish

More information

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University

Chapter 4. Data Transmission and Channel Capacity. Po-Ning Chen, Professor. Department of Communications Engineering. National Chiao Tung University Chapter 4 Data Transmission and Channel Capacity Po-Ning Chen, Professor Department of Communications Engineering National Chiao Tung University Hsin Chu, Taiwan 30050, R.O.C. Principle of Data Transmission

More information

Basics of reinforcement learning

Basics of reinforcement learning Basics of reinforcement learning Lucian Buşoniu TMLSS, 20 July 2018 Main idea of reinforcement learning (RL) Learn a sequential decision policy to optimize the cumulative performance of an unknown system

More information