Data Streaming Algorithms for Efficient and Accurate Estimation of Flow Size Distribution

Size: px
Start display at page:

Download "Data Streaming Algorithms for Efficient and Accurate Estimation of Flow Size Distribution"

Transcription

1 Data Streaing Algoriths for Efficient and Accurate Estiation of Flow Size Distribution Abhishek Kuar Minho Sung Jun (Ji) Xu College of Coputing Georgia Institute of Technology Jia Wang AT&T Labs Research ABSTRACT Knowing the distribution of the sizes of traffic flows passing through a network link helps a network operator to characterize network resource usage, infer traffic deands, detect traffic anoalies, and accoodate new traffic deands through better traffic engineering. Previous work on estiating the distribution has been focused on aking inferences fro sapled network traffic. Its accuracy is liited by the (typically) low sapling rate required to ake the sapling operation affordable. In this paper we present a novel data streaing algorith to provide uch ore accurate estiates of flow distribution, using a lossy data structure which consists of an array of counters fitted well into SRAM. For each incoing packet, our algorith only needs to increent one underlying counter, aking the algorith fast enough even for 40 Gbps (OC-768) links. The data structure is lossy in the sense that sizes of ultiple flows ay collide into the sae counter. Our algorith uses Bayesian statistical ethods such as Expectation Maxiization to infer the ost likely distribution that results in the observed counter values after collision. Evaluations of this algorith on large Internet traces obtained fro several sources (including a tier- ISP) deonstrate that it has very high easureent accuracy (within %). Our algorith not only draatically iproves the accuracy of flow distribution easureent, but also contributes to the field of data streaing by foralizing an existing ethodology and applying it to the context of estiating the flow-distribution. Categories and Subject Descriptors C..3 [COMPUTER-COMMUNICATION NETWORKS]: Network Operations - Network Monitoring E. [DATA STRUCTURES] General Ters Algoriths, Measureent, Theory Perission to ake digital or hard copies of all or part of this work for personal or classroo use is granted without fee provided that copies are not ade or distributed for profit or coercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, to republish, to post on servers or to redistribute to lists, requires prior specific perission and/or a fee. SIGMETRICS/Perforance 04, June 6, 004, New York, NY, USA. Copyright 004 ACM /04/ $5.00. Keywords Network Measureent, Traffic Analysis, Data Streaing, Statistical Inference. INTRODUCTION The proble of estiating flow distribution on a highspeed link has received considerable attention recently [,, 3, 4, 5, 6]. In this proble, given an arbitrary s, we are interested in knowing the nuber of flows that contain s packets, within a onitoring interval. In other words, we would like to know how the total traffic volue splits into flows of different sizes. An estiate of the flow distribution contains knowledge about the nuber of flows for all possible s, including elephants (large flows), kangaroos/rabbits (ediu flows), and ice (sall flows).. Motivation Flow distribution inforation can be useful in a nuber of applications in network easureent and onitoring. First, flow distribution inforation ay allow service providers to infer the usage pattern of their networks, such as the approxiate nuber of users with dial-up or broadband access. Such inforation on usage patterns can be iportant for the purpose of pricing, billing, infrastructure engineering, and resource planning. In addition, network operators ay also infer the type of applications that are running over a network link without looking into the details of traffic such as how any users are using streaed usic, streaed video, and voice over IP. In the future, we expect ore network applications to be recognizable through flow distribution inforation. Second, flow distribution inforation can help locally detect the existence of an event that causes the transition of the global network dynaics fro one ode to another. An exaple of such ode transition is a sudden increase in the nuber of large flows (i.e., elephants) in a link. Possible events that ay cause this include link failure or route flapping. Merely looking at the total load of the link ay not detect such a transition since this link could be consistently heavily used anyway. Furtherore, flow distribution inforation ay also help us detect various types of Internet security attacks such as DDoS and Internet wors. In the case of DDoS attacks, if the attackers are using spoofed IP addresses, we will observe a significant increase in flows of size. In the case of Internet Here we iniized the overlap with the otivation provided in [].

2 wors, we ay suddenly find a large nuber of flows of a particular size in Internet links around the sae tie, if the wor is a naive one that does not change in size. Also, the historical flow distribution inforation stored at various links ay help us study its evolution over tie. Finally, knowing the flow distribution of each link ay help other network easureent applications such as traffic atrix estiation [7, 8, 9, 0]. Recent work [9, 0] show that it is possible to use toography techniques to infer the traffic atrix fro link load and aggregate input/output traffic at each node. We have preliinary evidence to believe that flow distribution at each node will ake such toography uch ore accurate, since it allows the correlation of not only the total traffic volue (load), but also the correlation of its distribution into different flows.. Proble stateent The proble of coputing the distribution of the sizes of the flows can be foralized as follows. The set of possible s is the set of all positive integers between to z. Here z is the axiu that can be deterined fro the observed data. We denote the total nuber of flows as n, and the nuber of flows that have i packets as n i. We denote the fraction of flows that have i packets as φ i, i.e., φ i = n i. The data that need to be estiated are n the values of n and φ = {φ, φ,, φ z}. Our goal is to find an efficient schee to estiate this flow distribution inforation on a high-speed link (e.g., OC-9 to OC-768) with high accuracy. A naive solution to this proble is to use a hash table of per-flow counters to keep track of all active flows. These counters will later be exained to obtain the flow distribution. Although this approach is straightforward, it is not suitable for a high-speed link for the following reasons. Each flow entry in the hash table is large ( 60 bits) because it needs to store flow labels ( 00 bits), a pointer ( 3 bits) to the next entry if chaining is used to resolve hash collision, and a packet counter ( 3 bits). Since there can be a large nuber of flows (e.g., 0.5 illion) on backbone links during a typical easureent period, a hash table of this size typically can only fit into DRAM. However, DRAM speed cannot keep up with the link rate of OC-9 and higher 3. Another possible approach [] is to saple a sall percentage of packets and then infer the flow distribution fro the sapled traffic. The algorith proposed in [] ay well be the best algorith in getting as uch inforation fro the sapled data as possible. However, its accuracy is liited by the typically low sapling rate (e.g., %) required to ake the sapling operation affordable. Recent work [] has provided theoretical insights into the liitation of inferring flow distribution fro sapled traffic..3 Our approach and contributions The ain contribution of this paper is a novel data streaing algorith to provide uch ore accurate estiates of Linear probing and double hashing will not help save space since there is a tradeoff between the occupancy ratio and probe length. 3 With an average packet size of 000 bits, per-packet processing tie can be no ore than 00 ns and 5 ns, for OC-9 and OC-768, respectively. A hash table operation in DRAM will take hundreds of nanoseconds due to the need to retrieve the correct flow entry, copare the flow labels, and increent and write back the counter. flow distribution. Our algorith uses a lossy data structure that consists of an array of counters. Its total size is sall enough to fit easily in fast SRAM. For each incoing packet, our algorith only needs to increent one underlying counter (in SRAM), aking the algorith fast enough even for 40 Gbps (OC-768) links. The data structure is lossy in the sense that, due to collision in hashing, sizes of ultiple flows ay be accuulated in the sae counter. Therefore, the raw inforation obtained fro the counters can be far away fro the actual flow distribution. Our algorith then uses Bayesian statistical ethods such as Expectation Maxiization (EM) to infer the ost likely distribution that results in the observed counter values after collision. Experients of this algorith on a nuber of large traces deonstrate that it has very high easureent accuracy (within % relative error). However, to achieve this level of accuracy, our algorith needs to know the approxiate (± 50%) value of n, the total nuber of flows, in order to provision sufficient nuber of counters for streaing. Provisioning for the worst case (i.e., when the nuber of concurrent flows are the largest) leads to unnecessary waste of precious SRAM resource in the average case. To address this challenge, we propose a ulti-resolution variant of our algorith that uses a sall and fixed aount of SRAM and does not require any prior knowledge about the approxiate range of n. It guarantees high accuracy in the average case and graceful degradation in accuracy in the worst case. Our algorith not only draatically iproves the accuracy of flow distribution easureent, but also contributes to the field of data streaing by foralizing an existing yet iplicit ethodology, and exploring it in a new direction. Data streaing [] is concerned with processing a long strea of data ites in one pass using a sall working eory in order to answer a class of queries regarding the strea. The challenge is to use this sall eory to reeber as uch inforation pertinent to the queries as possible. In designing this algorith, we foralize the following ethodology. Lossy data structure + Bayesian statistics = Accurate streaing Its ain idea is to first perfor data streaing at very high speed in a sall eory to get the streaing results that are lossy. There are two causes for this loss to be inevitable. First, due to the stringent coputational coplexity requireent of the application (e.g., 5ns per packet when processing OC-768 traffic), the streaing algorith does not have enough processing tie to put the data into the exact place. Second, the streaing algorith does not have enough space to store all the relevant data. Due to the loss, the streaing result is typically far away fro the inforation we would like to estiate. Bayesian statistics is therefore used to recover inforation fro the streaing result as uch as possible. While Bayesian statistics is typically used in existing streaing algoriths to recover the second cause of loss, our algorith uses it ainly to recover the first cause of loss. Also, to the best of our knowledge, our algorith is the first to use sophisticated Bayesian tools such as EM in this recovery. The rest of this paper is organized as follows. In the next section, we provide an overview of the data collection portion of our solution and describe the design of our streaing

3 Online Streaing Module. Update. Raw streaing result Packet strea Header Header Header Offline Processing Module 3. Flow distribution Figure : Syste odel of using data-streaing to estiate flow distribution.. Initialize. A[i] := 0, i =,,..., 3. Update 4. Upon the arrival of a packet pkt 5. ind := hash(pkt.flow label); 6. A[ind] := A[ind] + ; 7. Export data when an epoch ends 8. y j := nuber of j s in A, j =,,..., z; 9. Forward the value y i s to offline analysis; Figure : Algorith for updating the online streaing odule data structure in detail. Section 3 describes our estiation echanis. We foralize the estiation echanis and analyze its correctness in Section 4. Section 5 presents a ulti-resolution version of our echanis that can operate with an array of fixed size. Section 6 evaluates the proposed schee over a nuber of large packet header traces obtained fro various places including a tier- ISP backbone network. We present a brief look at related work with a discussion about the context of our work in Section 7 before concluding in Section 8.. DATA STREAMING USING A LOSSY DATA STRUCTURE In this section, we first give an overview of the syste odel and the design philosophy of our approach. Then we describe our online update schee (i.e., the lossy data structure ) and analyze its coputational and storage coplexity. Finally, we show how our schee interfaces with the technique in [] to reduce the storage coplexity.. Syste odel The overall architecture of our solution is shown in Figure. The online streaing odule is updated upon each packet arrival (arc in Figure ). The easureent proceeds in epochs. At the end of each easureent epoch, the counter values, which we refer to as the raw data, will be paged out fro the online streaing odule, and these counters will be reset to 0 for the next easureent epoch. This raw data will be processed by an offline processing odule (arc in Figure ) that produces a final estiate (arc 3 in Figure ) of the flow distribution 4 using statistical inference techniques. This syste odel reflects our aforeentioned design philosophy of collecting as uch pertinent inforation as possible at the streaing odule, and then copensating for the inforation loss during data collection using Bayesian statistics.. Online streaing odule Our algorith for updating the data-streaing odule upon packet arrivals is shown in Figure. The streaing data structure used by our echanis is extreely siple an array of counters. Upon arrival of a packet at the router, its flow label 5 is hashed to generate an index into this array, 4 In practice, the raw data collected at the streaing odule can also be suarized and paged to persistent storage, where it can be stored till subsequent retrieval and estiation. 5 Our design does not place any constraints on the definition and the counter at this index is increented by. Collisions due to hashing ight cause two or ore flow labels to be hashed to sae indices. Counters at such an index would contain the total nuber of packets belonging to all of the flows colliding into this index. We do not have any explicit echaniss to handle collisions as any such echanis would ipose additional processing and storage overheads that are unsustainable at high speeds. This akes the encoding process very siple and fast. Efficient ipleentations of hash functions [3] allow the online streaing odule to operate at speeds as high as OC-768 without issing any packets..3 Coplexity of online streaing odule In this section, we discuss the storage and coputational coplexities of operating the data streaing odule.. Storage coplexity. This refers to both the aount of fast eory required for ipleenting the array of counters, and the aount of space (in DRAM or disk) to store the raw counter values for later retrieval and estiation by the offline estiation odule. Leveraging on the techniques for efficient ipleentation of a counter array proposed in [], we require 9 bits of SRAM per counter (to be discussed in Section.4). This allows us to ipleent about illion counters with. MB of SRAM. Interestingly, this raw data can be suarized to a very sall size when paged to DRAM or disk. The key fact here is that our estiation echanis does not need to know the apping between counter values and indices. Instead, it only needs to know, for each possible counter value, the nuber (i.e., ) of counters that have this value. Therefore, we can suarize this raw data into a list of <counter value, > tuples. It turns out that, while the nuber of flows is large, the unique s (and consequently, unique counter values) are usually quite sall. For exaple, in a trace with.6 illion packets and 9,000 flows, we observed only about 500 unique counter values. This iplies that ost counter values do not occur (i.e., occur with a of zero) in the array, resulting in a very sall list of <counter value, > tuples. For the above exaple, the suary can be stored in 8KB on persistent storage, thus requiring less than 0.05 bits per packet, or bit for 40 packets.. Coputational coplexity. For each packet, the data streaing odule needs to copute exactly one hash of flow label. It can be any cobination of fields fro the packet header.

4 function and increent exactly one counter. This is anageable even at OC768 (40 Gbps) speeds with off-the-shelf 0ns SRAM. We will show that our efficient (copact) ipleentation of counters (discussed in Section.4) causes very little overhead, allowing operation at OC-768 speed..4 Efficient ipleentation of an array of counters Internet traffic is known to have the property that a few flows can be very large, while ost other flows are sall. Thus, the counters in our array need to be large enough to accoodate the largest. On the other hand, the counter size needs to be ade as sall as possible to save precious SRAM. Recent work on efficient ipleentation of statistical counters [] provides an ideal echanis to balance these two conflicting requireents, which we will leverage on in our schee. For each counter in the array, say 3 bits wide, this echanis uses 3 bits of slow eory (DRAM) to store a large counter and aintains a saller counter, say 7 bits wide, in fast eory (SRAM). As the counters in SRAM exceed a certain threshold value (say 64) due to increents, it increents the value of the corresponding counter in DRAM by 64 and resets the counter in SRAM to 0. There is a -bit per counter overhead that covers the cost of keeping track of counters above the threshold, bringing the total nuber of bits per counter in SRAM to 9. For suitable choices of paraeters, this schee allows an efficient ipleentation of wide counters using a sall aount of SRAM. This technique can be applied sealessly to ipleenting the array of counters required in our data streaing odule. In our algorith 6, the size of each counter in SRAM is 9 bits and in DRAM is 3. Also, since the schee in [] incurs very little extra coputational and eory access overhead, our streaing algorith running on top of it can still achieve high speeds such as OC ESTIMATION MECHANISMS In this section, we describe a collection of estiation echaniss used in the offline processing odule (shown in Figure ). They help to infer the actual flow distribution fro the counter values collected by the online streaing odule. Consider the hypothetical case where there are no hash collisions. In this case the distribution of counter values is the sae as the actual flow distribution. However, collisions do occur with real-world hash functions, thus distorting the distribution of counter values away fro the true flow distribution. This effect 7 is shown in Figure 3, where we process a traffic trace (with 560K flows in it) on arrays of 04K, 5K, 56K, and 8K counters, respectively. We can see that, as the load factor (forally defined later in this section) of the array increases, the nuber of collisions increases, which further exacerbates this distortion. 3. Estiating the total nuber of flows The first quantity that we can estiate fro our counter array is the total nuber of flows during the easureent interval. The first echanis for estiating this quantity (in a different application) using a (0-) bitap is proposed 6 We have carefully checked these paraeters against the specifications in []. 7 Our experients on other traffic traces exhibit siilar distortion effects. e =04K =5K =56K =8K Figure 3: The distribution of s and raw counter values using varies nuber of counters (both x and y axes are in log-scale). = nuber of counters. in [4]. It can be used in our context with slight adaptation. The process of inserting (with collisions) flow counts into our counter array can be odeled as a coupon collector s proble, under the assuption of unifor hashing. As shown in [4], in an array of counters, if the nuber of zero entries is 0 after the insertion of n flows, the axiu likelihood estiator for n is ˆn = ln 0 () This result is also exploited in [6] to design a ore general ulti-resolution bitap schee to estiate n using uch saller eory. 3. Estiating the nuber of flows of size The total nuber of flows containing exactly one packet is arguably the ost significant single piece of inforation hidden in the distribution of s. Fro a odeling perspective, this nuber helps affir or reject statistical hypotheses such as whether the distribution is Zipfian. More iportantly, abnoral or alicious behavior in the Internet, such as port-scanning and DDoS attacks, often anifests itself as a significant increase in the nuber of flows of size. To estiate the nuber of flows of size (denoted by n ), let us look at the process of inserting flow counts into the counter array. Note that a counter of value ust contain exactly one flow of size (i.e., no collision). Based on this insight, we can derive a very accurate estiator for n. Let ˆλ = ˆn be the estiated load factor (in ters of the average nuber of flows that are apped to the sae index) on the array. Our siple estiator for n is ˆn = y eˆλ, where y is the nuber of counters with value. This surprisingly siple estiator ˆn turns out to be very accurate. In our experients shown later, we observed an accuracy of ±% using ˆn. Next, we explain the reasoning behind ˆn. Since the order of packet or flow arrivals does not affect the final values in the counter array, we consider a hypothetical situation where all flows of size and above were inserted in the counter array first. There are altogether n n of the. At this point, none of the flows of size has been inserted. The nuber of flows hashed to an index can be odeled as a binoial distribution Bino(n n, ), which in turn

5 can be approxiated by Poisson( n n ). The total nuber of indices that are not hit by any flow at this point (i.e., indices where the counter value is 0) can be estiated as 0 e n n. Now, assue all the flows of size are inserted into this array. Due to this insertion, soe of these 0 counters will becoe non-zero. The counters with value will be those out of a total of 0 that were zero before the insertion of n flows of size, and were hit by exactly one of these new insertions. By the sae arguent as above, the total nuber of such indices is 0λ e λ, where λ = n. But this nuber should be equal to y. Therefore we have y = 0λ e λ = e n n n which can be siplified as e n = ne n n = y e n () 3.3 Estiating the flow distribution One is tepted to generalize the above process to derive an estiator for the nuber of flows of size, 3, and so on (i.e., estiating n, n 3,..., n z). However, this proves to be difficult due to the following reason. While a counter of value is definitely not involved in a collision, countervalues of and above could be caused by the collision of two or ore flows. For exaple, aong the counters of value, soe correspond to a flow of size, while the others could be two flows of size hashing to the sae index. Thus, while the estiate for n (i.e., the nuber of flows of size ) depends only on our estiate of n, the estiate of n will depend on both n and n. More generally, the estiate of n i will depend on our estiates of n, n, n,, n i. Thus for a large i, the estiate is ore susceptible to errors due to this cuulative dependence effect, resulting in a sharp increase in estiation errors. Therefore, to accurately estiate flow distribution, we will take a ore holistic approach, rather than estiating each quantity step by step. This approach, based on Expectation Maxiization (EM) ethod for coputing Maxiu Likelihood Estiation (MLE), is the sole topic of the next section. 4. ESTIMATING FLOW DISTRIBUTION US- ING EXPECTATION MAXIMIZATION In this section, we describe our Maxiu Likelihood Estiation (MLE) algorith that coputes the flow distribution that is ost likely to result in the observed counter values after the hash collisions. To find this MLE directly is difficult because there is neither a closed-for forula nor a coputation procedure for p(φ y), the distribution of the flow distribution φ conditioned on the observation y. The difficulty of coputing p(φ y) can be attributed to the fact that our observed data is incoplete. To address this proble, we adopted a powerful ethod in statistics called Expectation Maxiization (EM) to iteratively copute the local 8 MLE. EM is especially effective in finding MLE when the observation can be viewed as incoplete data. In our context, the observed counter values can be viewed as incoplete data, and the issing part is 8 EM algoriths in general can only guarantee to converge to a local axiu [5], while MLE often refers to the global axiu. With this understanding, we will oit the word local fro subsequent discussions of MLE using EM. how flows collide with each other during hashing. The evaluation in Section 6 shows that our EM algorith accurately estiates the flow distribution aong all traces we have experiented with. To the best of our knowledge, this is the first work that applies EM algorith to coputing the MLE fro a lossy data structure. 4. Background on EM Let y denote our observation and φ denote the rando variable whose value we would like to estiate. In MLE, we would like to find out φ that axiizes p(φ y). However, it is usually hard to copute such a φ because the forula p(φ y) is either coplicated or does not have a closed for due to issing data. The EM algorith, which captures our intuition on handling issing data, works as follows. It starts with a guess of the paraeters, and then replaces issing values by their expectations given the guessed paraeters, and finally estiates the paraeters assuing the issing data are equal to their estiated values. This new estiate of issing values gives us a better estiate of paraeters. This process will be iterated ultiple ties until the estiated paraeters converge to a set of values (typically a local axiu as entioned above). Forally, EM begins with a guess of the paraeter φ ini, which will serve as φ old for the first iteration. Then the following two alternating steps will be executed iteratively. Expectation step. E old (log p(γ, φ y)) = (log p(γ, φ y)) p(γ φ old, y)dγ, where the expectation averages over the conditional posterior distribution of the issing data γ, given the current estiate φ old. We use the notation Q(φ, φ old ) to denote E old (log p(γ, φ y)) per the convention in statistics literature. For any applications, both p(γ φ, y) and p(φ γ, y) inside the integration forula above are straightforward to copute. Maxiization step. Let φ new be the value of φ that axiizes Q(φ, φ old ). This φ new will serve as φ old for the next iteration. These two steps will be iterated for a nuber steps until φ old and φ new are close enough to each other, a notion that will becoe rigorous later in Section Applying EM to our context Our observation y, obtained fro the output of online streaing odule, is y i (i =,,..., z), the nuber of counters that have value i. Our goal is to estiate φ i, the fraction of flows that are of size i (i =,,..., z). Here z is the axiu counter value observed fro the array. Our EM algorith for estiating φ is shown in Figure 4. We first need a guess of the flow distribution φ ini, and the total nuber of flows n ini. In our algorith, we siply use the distribution obtained fro the raw counter values as φ ini and the total nuber of non-zero counters as n ini. Based on this φ ini and n ini, we can copute, for each possible way of splitting an observed counter value, its average nuber of occurrences. Then the counts n i for flows of corresponding sizes will be credited according to this average. For exaple, when the value of a counter is 3, there are three possible events that result in this observation: (i) 3 = 3 (no hash collision); (ii) 3 = + (a flow of size colliding with a flow of size ); and (iii) 3 = + + (three flows of size hashed to the sae index). Given a guess of the flow distribution, we can estiate the posterior probabilities of these three cases. Say the respective probabilities of these three events

6 Input: y i, nuber of counters that have value i ( i z) Output: MLE for the flow distribution φ. Initialization: pick an initial flow distribution φ (ini) and estiate the total flow count n ini fro Section 3... φ new := φ ini ; n new = n ini 3. while (convergence condition is not satisfied) 4. φ old := φ new ; n old := n new 5. for i := to z 6. foreach β Ω i 7. /*Ω i is the set of all collision patterns 8. that add up to i, defined in Theore */ 9. Suppose β is that f flows of size s, f flows of 0. size s,..., and f q flows of size s q collide into. a counter of value i, then. for j := to q 3. n sj := n sj + y i f j p(β φ old, n, V = i) 4. /* Procedure for coputing p(β φ old, n, V = i) 5. is shown in Theore and Lea.*/ 6. end 7. end 8. end 9. n new z := i= n i 0. for i:= to z. φ new i := n i /n new. end 3. /* noralize the counts n is into flow distribution φ*/. 4. end Figure 4: EM algorith for coputing flow distribution are 0.5, 0.3, and 0., and there are 000 counters with value 3. Then we estiate that, on the average, 500, 300, and 00 counters split in the three above ways, respectively. So we credit 300 * + 00 * 3 = 900 to n, the count of flows of size, and credit 300 and 500 to n and n 3, respectively. Finally, after all observed counter values are split this way, we get the new counts n, n,..., n z, and obtain n new z (= i= ni). We then renoralize the into a new (and refined) flow distribution φ new. We will prove in Section 4.3 that this progra is indeed an instance of the EM algorith. Coputing the probability p(β φ, n, v). Let both n and (size of counter array) be very large so that we can approxiate binoial distribution using Poisson. This approxiation is necessary since our estiates of flow counts can be non-integers. Let λ i denote the average nuber of size i flows (before collision) that are hashed to an (arbitrary) index in the array. In other words, λ i = n i = nφ i. We z define λ = i= λi, which is the average nuber of flows (of all sizes) that is hashed to an (arbitrary) index. Let ind be an arbitrary index into the array and v be the observed value at this index. Let β be the event that f flows of size s, f flows of size s,..., f q flows of size s q collide into this slot, where s < s <... < s q z. Lea. Given φ and n, the a priori (i.e., before observing this value v) probability that event β happens is p(β φ, n) = e λ q i= λ f i s i f i!. Proof. Let B i be the event that f i flows of size s i be apped to the counter indexed by ind. Let C be the event that all other flows have zero arrivals to ind. Since the hashing is unifor, these events B, B,..., B q, and C are independent. Therefore, p(β φ, n) = p(c φ, n) q i= p(bi φ, n). by Pois- Let I = {s, s,..., s q}. Then p(b i φ, n) = e λs i λf i s i f i! son approxiation of binoial distribution. So, q i= p (Bi φ, n) = q i= e λs i λ f i s i f i =! Also, p (C φ, n) = j I e λ j. Therefore, p (β φ, n) = j I e λ j j I e λ j q i= j I e λ j λ f i s i f i! = e λ q q λ f i s i i= f i! However, the situation changes after we have already seen v, the value at the counter indexed by ind. Theore. Let Ω v be the set of all collision patterns p(β φ,n) that add up to v. Then p(β φ, n, v) =, where p(α φ,n) α Ωv p(β φ, n) and p(α φ, n) can be coputed using Lea. Proof. Let Ω be the set of all possible collision patterns as defined before. Let us choose an arbitrary index ind and let V be the counter value at this index. By Bayes rule, p(β φ, n, V = v) = p(v = v β, φ, n)p(β φ, n) p(v = v α, φ, n)p(α φ, n) α Ω However, note that p(v = v α, φ, n) = for all α Ω v (including β) and p(v = v α, φ, n) = 0 for all α Ω Ω v. Therefore, p(β φ, n, v) = p(β φ, n) α Ω v p(α φ, n) 4.3 Our algorith is an EM algorith We next prove that the algorith shown in Figure 4 is indeed an EM algorith. This proof is iportant since the fact that the algorith is an instance of EM guarantees that the outputs fro the iterations of the algorith will converge to a set of local MLEs, according to [5]. Theore. The algorith in Figure 4 is an EM algorith. Proof. Let γ ij denote the nuber of size i flows that are collided (erged) into counters of value j ( i j z). These are the issing data that the algorith in Figure 4 needs to guess in order to estiate the flow distribution φ. The coplete data likelihood function L(φ) (i.e., p(γ, φ y) defined in Section 4.), assuing γ ij is known, is z z i= j=i γij log φi. Then in the expectation step, E (φ old,n)[l(φ) y] = i= z z i= j=i E[γij φold, y, n] log φ i λ f i s i f i! This corresponds to Q(φ, φ old ) in Section 4.. Let γ i = z j=i γij. Define nij = E[γij φold, y, n] and n i = E[γ i φ old, y, n]. z By the linearity of expectation, we know that n i = j=i nij. z Therefore, E (φ old,n)[l(φ) y] = i= ni log φi. Note that the definition of n i here atches the coputation of n i in our algorith (lines 5 to 8). Finally in the axiization step, we need to axiize φi =. Here z i= ni log φi, subject to the constraint z i=

7 n i (i =,,..., z) are constants and φ is are the variables. Using the ethod of Lagrange ultiplier, we know that the n axiu value is achieved when φ i = i z j= n j. This is exactly the renoralization step in our progra (lines 9 to 3) shown in Figure 4. Therefore, our algorith is indeed an EM algorith. 4.4 Coputational coplexity of the EM algorith. It is easy to enuerate all possible events that give rise to a sall counter value. But, for large counter values, the nuber of possible events (hash collisions) that could give rise to the observed value is iense. Thus it is not possible to exhaustively copute the probabilities for all such events. The Zipfian nature of flow-size distribution coes to our rescue here. To reduce the coplexity of enuerating all events that could give rise to a large counter value (say larger than 300), we ignore the cases involving the collision of 4 or ore flows at the corresponding index. Since the nuber of counters with a value larger than 300 is quite sall, and collisions involving 4 or ore flows occur with a low probability, this assuption has very little ipact on the overall estiation echanis. With siilar justifications we ignore events involving 5 or ore collisions for counters larger than 50 but saller than 300 and those involving 7 or ore collisions for all other counters. This reduces the asyptotic coputational coplexity of splitting a counter-value j to O(j 3 ) (for j > 300). Note that we need to do this coputation only once for all counters that have a value j, and the nuber of unique counter-values is quite sall (as discussed earlier in Section.3). Finally, since the nubers of counters with very large values (say larger than 000) is extreely sall, we can ignore splitting such counter values entirely and instead report the counter value as the size of a single flow. This will clearly lead to a slight overestiation of the size of such large flows, but since the average ( 0) is two to three orders of agnitude saller than these large flows, this error is inuscule in relative ters. These optiizations bring the overall coputational coplexity well under control. On a 3. GHz Intel Pentiu 4 desktop, each iteration of the EM takes about 0 seconds. If the easureent epoch is 00 seconds long and we terinate the estiation after five iterations, then the estiation can run as fast as the data streaing odule. 5. MULTI-RESOLUTION ESTIMATION OF FLOW DISTRIBUTION As shown in Figure 3, the raw counter value distribution deviates ore and ore fro the actual flow distribution as the size of the counter array decreases. Our experients in r R r R A A R r R r Section 6 show that the accuracy of estiation falls sharply if the size of the array is less than of the total nuber of 3 flows n. Therefore, for the accurate estiation of flow distribution, we need a counter array that contains at least n 3 entries. However, in real-world Internet traffic, the nuber of flows in the worst case can be any ties ore than in the average case. Provisioning enough counters for the worst case would result in excessive waste of precious SRAM in the average case. In this section, we present a ulti-resolution version of our solution that uses a fixed-size array of counters, and allows a graceful degradation in estiation accu R r+ A r A r A r+ Figure 5: The Multi-Resolution Array of Counters.. Initialize. r = log (M/) ( 3. R i = ( i )M, ( i )M i =,, 3,..., r r )M, M i = r + 4. Arrays A, A,, A r+ are all initialized to 0 5. Update 6. Upon the arrival of a packet pkt 7. ind := hash(pkt.flow label); 8. if (ind R j ) 9. A j [ind od ]++; Figure 6: Algorith for updating MRAC. racy when the total nuber of flows increases. This akes the schee accurate and eory-efficient for the average case while its accuracy degrades only slightly for the worst case. Our design is inspired by a ulti-resolution schee used in [6]. We apply it here to a different context. Our Multi-Resolution Array of Counters (MRAC) schee works as follows. Iagine a virtual array of counters that is large enough to accurately estiate the flow distribution even in the worst case. However, the physical (actual) counter array size is uch saller. Therefore, the virtual array needs to be apped/folded to the actual physical array as shown in Figure 5. Here we describe a base- version of our apping. Its generalization to any arbitrary base b is straightforward. In the base- version, we ap a logical array of M = r counters to r + physical arrays of size each. Half of the hash space will be apped to (folded 4 of into) array, half of the reaining hash space (i.e., the total hash space) will be apped to array, and so on. Finally, we are left with two blocks of hash space of size each. They are directly apped to arrays r and (r+). The M total space taken by the arrays is (log + ). This actual apping/folding algorith is shown in Figure 6. As described above, the arrays A, A,..., A r, A r+ cover the respective hash ranges of [0, M), [ M, 3 M), [ 3 M, M),, [( )M, ( )M), [( )M, M). If a 8 r r r hash index ind is apped to a array, the counter indexed by (ind od ) in that array will be increented. Therefore, the values of r counters in the virtual array ap to (fold into) counter in array A, and the values of r virtual counters ap to counter in array A, and so on. The (r + ) arrays together cover the entire virtual hash space. The regions covered by any two arrays are disjoint. Such a apping is iplicitly a flow sapling (not packet sapling) schee. Array A processes approxiately of the flows (i.e., every packet in approxiately half of the flows), array A processes approxiately of the flows, 4 and so on. Note that the coputational coplexity of this schee is alost the sae as the baseline approach, which

8 is one hash function coputation and one eory access to SRAM. The only additional processing here is to recognize the range that a hash value falls into and to perfor a odulo operation ( ind od in line 9 of Figure 6). Since all operations involve s powers, they can be ipleented efficiently using siple binary logic. The estiation algorith works as follows. It first picks an array that will result in the best estiate of the original flow distribution. The criteria of picking such an array will be discussed next. Suppose the array we pick is i of the size of the virtual array, that is, this array saples approxiately i fraction of the flows. The algorith first estiates the flow distribution fro the array using the baseline approach described in the previous two sections, and then scales the result by i to obtain the estiate for the overall traffic. Since the nuber of very large flows (say larger than 000 packets) is quite sall, we can use the counter values larger than 000 fro all resolutions to refine our estiation for the tail of the distribution. For each of these large counter values, we subtract the average counter value in the corresponding resolution and use the result as the estiated size of the large flows hashed to this counter. In general, the arrays where the sapling rate is high (i.e., the arrays that cover large portions of the virtual hash space) tend to be over-crowded (i.e., with higher average nuber of flows apped to the sae slot). This corresponds to using a very sall array of counters, which results in inaccurate estiation. On the other hand, when the sapling rate is low (i.e., when the array covers a very sall portion of the virtual hash space), the estiation fro the corresponding array will be accurate, but the errors due to (flow) sapling becoe high. Therefore, there is a clear tradeoff between the loss of accuracy due to over-crowding on the one hand and due to sapling on the other. We find that there exists an optial array size in the iddle that iniizes the overall loss of accuracy, which can be found using the following criteria. We pick an array with as high sapling rate as possible, under the constraint that no ore than.5 flows are apped to the sae slot on the average. The reasoning behind this rule is siilar to that used in two existing ultiresolution based schees [6, 6] (for different applications). We oit the details here in the interest of space. Finally, the above design with base- can be generalized to any arbitrary base. Choosing a base that is a power of allows efficient hardware and software ipleentation. Our ipleentation evaluated in Section 6 uses base-4. The base-4 algorith needs 50% less eory than base-, with noinal loss in estiation accuracy. 6. EVALUATION In this section, we evaluate the accuracy of our estiation echanis using real-world Internet traffic traces. We also copare our results with those obtained in [] fro sapled traffic. Our experients deonstrate that our echanis achieves very high accuracy, which is typically an order of agnitude better than sapling-based approaches. 6. Traffic traces We use three sets of traces in our evaluation. The first set coprises of two packet header traces obtained fro a tier- ISP backbone, collected by a Gigascope server [7] on a high speed link leaving a data center in October, 003. Each of the packet header traces lasts a few hours, consists of 700 Source Trace # of flows # of packets ISP Weekday,34,89 68,595,755 Weekend,39,746 8,86,457 NLANR Long 563,080,769,43 Mediu 9,380,668,69 Short 55,55 58,43 [] CAMPUS 45,70 0,065,600 COS 6,038,554 37,000,000 PEERING,89,85 0,000,000 Table : Traces used in our evaluation. illion packet headers and carries 350 GB traffic. In our experients, we used segents taken fro these two traces, one for heavier traffic load on a weekday and the other for light traffic load at a weekend. Table lists the nuber of flows and packets in each trace. The second set of traces we used are publicly available traffic traces fro NLANR. We use three NLANR traces 9 naed Long, Mediu, and Short, based on the nuber of flows in each trace (Table ). Notice that the trace Long actually has fewer packets than Mediu. However, the attribute of significance in our evaluation is the nuber of flows in each trace, and the naes are intuitive in this light. Finally, we use a set of three traces fro [] to copare with previous work on estiating flow-distribution fro sapled statistics. Trace CAMPUS was collected at a LAN near the border of a capus network during a period of 300 inutes. Trace COS was collected at an OC3 link at Colorado State University during January 5 and 6, 003. This period overlaps the onset of Slaer wor [8]. Trace PEERING was collected at a peering link for a period of 37 inutes. 6. Evaluation etrics For coparing the estiated flow distribution with the actual distribution, we considered two possible etrics, Mean Relative Difference (MRD) and Weighted Mean Relative Difference (WMRD). We eventually adopt WMRD as our evaluation etric. The rationale for this choice is given below. The etric MRD is often used in easuring the distance between two probability distributions or ass functions, defined in our context as follows. Suppose the nuber of flows of size i is n i and our estiate of this nuber is ˆn i. The relative error in estiation (i.e., relative difference) is given by n i ˆn i /( n i+ ˆn i ). The ean relative difference over all s is obtained by taking the ean of relative difference over all possible s,, 3,..., z. Therefore, the MRD between the estiated and actual distribution is given by: MRD = z i n i ˆn i n i + ˆn i However, this etric is not suitable for estiating flow distribution for the following reason. The Zipfian nature of the Internet traffic iplies that there are a large nuber of sall flows and only a few large flows. In other words, when i becoes larger, n i becoes saller, and n i ˆn i /( n i+ ˆn i ) becoes larger. Therefore, the errors in estiating the tail 9 We experiented on any other NLANR traces, which yield siilar results as reported in this paper.

9 of the distribution (i.e., n i ˆn i /( n i+ ˆn i ) for large values of i) doinate the value of MRD. This akes no sense since the ain body of the distribution is the large nuber of sall flows, the estiation accuracy of which is discounted in MRD. To reflect the errors in estiating the nuber of large and sall flows in proportion to their actual population, we adopt the aforeentioned second etric called Weighted Mean Relative Difference (WMRD). It is proposed and used in [], for the sae purpose of evaluating the accuracy of estiated flow distribution. In WMRD, we assign a weight of n i+ ˆn i to the relative error in estiating the nuber of flows of size i. Thus the value of WMRD is given by: W MRD = i n i ˆn i n i + ˆn i i n i + ˆn i n i + ˆn i = i n i ˆn i n i i + ˆn i WMRD is also used in our EM algorith to deterine how close our estiate is to the convergence point. In our algorith, we choose a threshold ɛ, and terinate our iterative estiation procedure when the WMRD of the estiates produced by two consecutive estiates falls below ɛ. The intuition here is that, as estiates get closer to the convergence point, the iproveent fro one iteration to the next becoes saller, iplying a saller WMRD between the estiates produced by two consecutive estiates. 6.3 Bucketing flow distribution As can be seen fro Figure 3, the tail of the flow distribution plot is very noisy. This is due to the fact that a sall nuber of large flows are distributed in a large size range in a very sparse way. To obtain a ore intuitive visual depiction of the flow distribution, we use a bucketing schee to sooth out the noise. Buckets are sets of one or ore consecutive integers. The total nuber of flows in a bucket is the su of the nuber of flows of each unique size in the bucket. For sall s, where there are a large nuber of flows for each unique size, we use a bucket size of, iplying no soothing. As we proceed towards large s, gaps between two s that have non-zero counts start appearing (and then widening). We scale the bucket size appropriately so that ost buckets have at least one flow. In the figures depicting flow distribution, each bucket is depicted as a data point, with the id-point of the bucket as its x-coordinate and the total nuber of flows in the bucket divided by the bucket size as its y-coordinate. We ephasize that this bucketing schee is used only for better visualization of results. Our estiation echanis and nuerical results (in WMRD) reported later in this section do not use soothing of any for. 6.4 Estiation using array of counters As entioned earlier in Section 5, we show that the estiation procedure is likely to be ore accurate when the nuber of counters is close to or larger than the total nuber of flows in the easureent epoch. Table shows the effect of the choice of nuber of counters over estiation accuracy for the NLANR traces. The deviation of both the initial guess (taken fro the observed counter value distribution) and the final estiate after 0 iterations of the EM algorith becoes larger when the nuber of counters becoe saller. However, this increase in WMRD is very sall when the nuber of counters stay larger than or equal Trace # of flows Array WMRD of WMRD of in trace size raw data final estiate 04K Long 563,080 5K K K K Mediu 9,380 56K K K K Short 55,55 64K K K WMRD of initial guesses and final esti- Table : ates. WMRD =04K =5K =56K =8K Iteration Figure 7: The WMRD of the estiate vs. the nuber of iterations of the estiation algorith, for the trace Long. to the nuber of flows. The increase becoes pronounced only after the nuber of counters drops to less than of 3 the nuber of flows. Figure 7 shows how the WMRD of the estiates decreases when the nuber of EM iterations increases. The trace used for this experient is trace Long containing 563,080 flows. Each curve corresponds to a different choice of the nuber of counters, ranging fro 8K ( 7 ) to M ( 0 ). The points for iteration 0 correspond to the WMRD of the initial guess, obtained fro the distribution of the raw counter values. All the curves show a downward trend on WMRD to approach zero, indicating progress toward convergence. The curves for = 04K and = 5K begin with uch better initial guesses, thus achieving a uch saller WMRD (in absolute value) within a sall nuber of iterations. This reinforces the notion that using approxiately the sae nuber of counters as the nuber of flows provides uch better estiation accuracy than using a significantly saller nuber of counters. We observe siilar results on other traces. Figure 8 presents the results of running our estiation echanis on the trace Long (siilar results are observed on all traces in Table ). In this experient, the nuber of counters was set to a s power that is closest to the nuber of flows n. Each figure has three curves, corresponding to the actual distribution of s, the distribution of raw counter values, and the result of our estiation echanis, respectively. The near overlap of our estiate with

Birthday Paradox Calculations and Approximation

Birthday Paradox Calculations and Approximation Birthday Paradox Calculations and Approxiation Joshua E. Hill InfoGard Laboratories -March- v. Birthday Proble In the birthday proble, we have a group of n randoly selected people. If we assue that birthdays

More information

arxiv: v3 [cs.ds] 22 Mar 2016

arxiv: v3 [cs.ds] 22 Mar 2016 A Shifting Bloo Filter Fraewor for Set Queries arxiv:1510.03019v3 [cs.ds] Mar 01 ABSTRACT Tong Yang Peing University, China yangtongeail@gail.co Yuanun Zhong Nanjing University, China un@sail.nju.edu.cn

More information

Using EM To Estimate A Probablity Density With A Mixture Of Gaussians

Using EM To Estimate A Probablity Density With A Mixture Of Gaussians Using EM To Estiate A Probablity Density With A Mixture Of Gaussians Aaron A. D Souza adsouza@usc.edu Introduction The proble we are trying to address in this note is siple. Given a set of data points

More information

Lower Bounds for Quantized Matrix Completion

Lower Bounds for Quantized Matrix Completion Lower Bounds for Quantized Matrix Copletion Mary Wootters and Yaniv Plan Departent of Matheatics University of Michigan Ann Arbor, MI Eail: wootters, yplan}@uich.edu Mark A. Davenport School of Elec. &

More information

e-companion ONLY AVAILABLE IN ELECTRONIC FORM

e-companion ONLY AVAILABLE IN ELECTRONIC FORM OPERATIONS RESEARCH doi 10.1287/opre.1070.0427ec pp. ec1 ec5 e-copanion ONLY AVAILABLE IN ELECTRONIC FORM infors 07 INFORMS Electronic Copanion A Learning Approach for Interactive Marketing to a Custoer

More information

A Simple Regression Problem

A Simple Regression Problem A Siple Regression Proble R. M. Castro March 23, 2 In this brief note a siple regression proble will be introduced, illustrating clearly the bias-variance tradeoff. Let Y i f(x i ) + W i, i,..., n, where

More information

COS 424: Interacting with Data. Written Exercises

COS 424: Interacting with Data. Written Exercises COS 424: Interacting with Data Hoework #4 Spring 2007 Regression Due: Wednesday, April 18 Written Exercises See the course website for iportant inforation about collaboration and late policies, as well

More information

Polygonal Designs: Existence and Construction

Polygonal Designs: Existence and Construction Polygonal Designs: Existence and Construction John Hegean Departent of Matheatics, Stanford University, Stanford, CA 9405 Jeff Langford Departent of Matheatics, Drake University, Des Moines, IA 5011 G

More information

13.2 Fully Polynomial Randomized Approximation Scheme for Permanent of Random 0-1 Matrices

13.2 Fully Polynomial Randomized Approximation Scheme for Permanent of Random 0-1 Matrices CS71 Randoness & Coputation Spring 018 Instructor: Alistair Sinclair Lecture 13: February 7 Disclaier: These notes have not been subjected to the usual scrutiny accorded to foral publications. They ay

More information

Model Fitting. CURM Background Material, Fall 2014 Dr. Doreen De Leon

Model Fitting. CURM Background Material, Fall 2014 Dr. Doreen De Leon Model Fitting CURM Background Material, Fall 014 Dr. Doreen De Leon 1 Introduction Given a set of data points, we often want to fit a selected odel or type to the data (e.g., we suspect an exponential

More information

Block designs and statistics

Block designs and statistics Bloc designs and statistics Notes for Math 447 May 3, 2011 The ain paraeters of a bloc design are nuber of varieties v, bloc size, nuber of blocs b. A design is built on a set of v eleents. Each eleent

More information

Support Vector Machine Classification of Uncertain and Imbalanced data using Robust Optimization

Support Vector Machine Classification of Uncertain and Imbalanced data using Robust Optimization Recent Researches in Coputer Science Support Vector Machine Classification of Uncertain and Ibalanced data using Robust Optiization RAGHAV PAT, THEODORE B. TRAFALIS, KASH BARKER School of Industrial Engineering

More information

Experimental Design For Model Discrimination And Precise Parameter Estimation In WDS Analysis

Experimental Design For Model Discrimination And Precise Parameter Estimation In WDS Analysis City University of New York (CUNY) CUNY Acadeic Works International Conference on Hydroinforatics 8-1-2014 Experiental Design For Model Discriination And Precise Paraeter Estiation In WDS Analysis Giovanna

More information

Feature Extraction Techniques

Feature Extraction Techniques Feature Extraction Techniques Unsupervised Learning II Feature Extraction Unsupervised ethods can also be used to find features which can be useful for categorization. There are unsupervised ethods that

More information

Statistical Logic Cell Delay Analysis Using a Current-based Model

Statistical Logic Cell Delay Analysis Using a Current-based Model Statistical Logic Cell Delay Analysis Using a Current-based Model Hanif Fatei Shahin Nazarian Massoud Pedra Dept. of EE-Systes, University of Southern California, Los Angeles, CA 90089 {fatei, shahin,

More information

Vulnerability of MRD-Code-Based Universal Secure Error-Correcting Network Codes under Time-Varying Jamming Links

Vulnerability of MRD-Code-Based Universal Secure Error-Correcting Network Codes under Time-Varying Jamming Links Vulnerability of MRD-Code-Based Universal Secure Error-Correcting Network Codes under Tie-Varying Jaing Links Jun Kurihara KDDI R&D Laboratories, Inc 2 5 Ohara, Fujiino, Saitaa, 356 8502 Japan Eail: kurihara@kddilabsjp

More information

Analyzing Simulation Results

Analyzing Simulation Results Analyzing Siulation Results Dr. John Mellor-Cruey Departent of Coputer Science Rice University johnc@cs.rice.edu COMP 528 Lecture 20 31 March 2005 Topics for Today Model verification Model validation Transient

More information

Estimating Entropy and Entropy Norm on Data Streams

Estimating Entropy and Entropy Norm on Data Streams Estiating Entropy and Entropy Nor on Data Streas Ait Chakrabarti 1, Khanh Do Ba 1, and S. Muthukrishnan 2 1 Departent of Coputer Science, Dartouth College, Hanover, NH 03755, USA 2 Departent of Coputer

More information

This model assumes that the probability of a gap has size i is proportional to 1/i. i.e., i log m e. j=1. E[gap size] = i P r(i) = N f t.

This model assumes that the probability of a gap has size i is proportional to 1/i. i.e., i log m e. j=1. E[gap size] = i P r(i) = N f t. CS 493: Algoriths for Massive Data Sets Feb 2, 2002 Local Models, Bloo Filter Scribe: Qin Lv Local Models In global odels, every inverted file entry is copressed with the sae odel. This work wells when

More information

Upper bound on false alarm rate for landmine detection and classification using syntactic pattern recognition

Upper bound on false alarm rate for landmine detection and classification using syntactic pattern recognition Upper bound on false alar rate for landine detection and classification using syntactic pattern recognition Ahed O. Nasif, Brian L. Mark, Kenneth J. Hintz, and Nathalia Peixoto Dept. of Electrical and

More information

Sharp Time Data Tradeoffs for Linear Inverse Problems

Sharp Time Data Tradeoffs for Linear Inverse Problems Sharp Tie Data Tradeoffs for Linear Inverse Probles Saet Oyak Benjain Recht Mahdi Soltanolkotabi January 016 Abstract In this paper we characterize sharp tie-data tradeoffs for optiization probles used

More information

Probability Distributions

Probability Distributions Probability Distributions In Chapter, we ephasized the central role played by probability theory in the solution of pattern recognition probles. We turn now to an exploration of soe particular exaples

More information

A note on the multiplication of sparse matrices

A note on the multiplication of sparse matrices Cent. Eur. J. Cop. Sci. 41) 2014 1-11 DOI: 10.2478/s13537-014-0201-x Central European Journal of Coputer Science A note on the ultiplication of sparse atrices Research Article Keivan Borna 12, Sohrab Aboozarkhani

More information

Inspection; structural health monitoring; reliability; Bayesian analysis; updating; decision analysis; value of information

Inspection; structural health monitoring; reliability; Bayesian analysis; updating; decision analysis; value of information Cite as: Straub D. (2014). Value of inforation analysis with structural reliability ethods. Structural Safety, 49: 75-86. Value of Inforation Analysis with Structural Reliability Methods Daniel Straub

More information

Fast Montgomery-like Square Root Computation over GF(2 m ) for All Trinomials

Fast Montgomery-like Square Root Computation over GF(2 m ) for All Trinomials Fast Montgoery-like Square Root Coputation over GF( ) for All Trinoials Yin Li a, Yu Zhang a, a Departent of Coputer Science and Technology, Xinyang Noral University, Henan, P.R.China Abstract This letter

More information

Symbolic Analysis as Universal Tool for Deriving Properties of Non-linear Algorithms Case study of EM Algorithm

Symbolic Analysis as Universal Tool for Deriving Properties of Non-linear Algorithms Case study of EM Algorithm Acta Polytechnica Hungarica Vol., No., 04 Sybolic Analysis as Universal Tool for Deriving Properties of Non-linear Algoriths Case study of EM Algorith Vladiir Mladenović, Miroslav Lutovac, Dana Porrat

More information

Non-Parametric Non-Line-of-Sight Identification 1

Non-Parametric Non-Line-of-Sight Identification 1 Non-Paraetric Non-Line-of-Sight Identification Sinan Gezici, Hisashi Kobayashi and H. Vincent Poor Departent of Electrical Engineering School of Engineering and Applied Science Princeton University, Princeton,

More information

Pattern Recognition and Machine Learning. Artificial Neural networks

Pattern Recognition and Machine Learning. Artificial Neural networks Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2017 Lessons 7 20 Dec 2017 Outline Artificial Neural networks Notation...2 Introduction...3 Key Equations... 3 Artificial

More information

Algorithms for parallel processor scheduling with distinct due windows and unit-time jobs

Algorithms for parallel processor scheduling with distinct due windows and unit-time jobs BULLETIN OF THE POLISH ACADEMY OF SCIENCES TECHNICAL SCIENCES Vol. 57, No. 3, 2009 Algoriths for parallel processor scheduling with distinct due windows and unit-tie obs A. JANIAK 1, W.A. JANIAK 2, and

More information

On Constant Power Water-filling

On Constant Power Water-filling On Constant Power Water-filling Wei Yu and John M. Cioffi Electrical Engineering Departent Stanford University, Stanford, CA94305, U.S.A. eails: {weiyu,cioffi}@stanford.edu Abstract This paper derives

More information

A Low-Complexity Congestion Control and Scheduling Algorithm for Multihop Wireless Networks with Order-Optimal Per-Flow Delay

A Low-Complexity Congestion Control and Scheduling Algorithm for Multihop Wireless Networks with Order-Optimal Per-Flow Delay A Low-Coplexity Congestion Control and Scheduling Algorith for Multihop Wireless Networks with Order-Optial Per-Flow Delay Po-Kai Huang, Xiaojun Lin, and Chih-Chun Wang School of Electrical and Coputer

More information

In this chapter, we consider several graph-theoretic and probabilistic models

In this chapter, we consider several graph-theoretic and probabilistic models THREE ONE GRAPH-THEORETIC AND STATISTICAL MODELS 3.1 INTRODUCTION In this chapter, we consider several graph-theoretic and probabilistic odels for a social network, which we do under different assuptions

More information

The Weierstrass Approximation Theorem

The Weierstrass Approximation Theorem 36 The Weierstrass Approxiation Theore Recall that the fundaental idea underlying the construction of the real nubers is approxiation by the sipler rational nubers. Firstly, nubers are often deterined

More information

Fairness via priority scheduling

Fairness via priority scheduling Fairness via priority scheduling Veeraruna Kavitha, N Heachandra and Debayan Das IEOR, IIT Bobay, Mubai, 400076, India vavitha,nh,debayan}@iitbacin Abstract In the context of ulti-agent resource allocation

More information

A Note on Scheduling Tall/Small Multiprocessor Tasks with Unit Processing Time to Minimize Maximum Tardiness

A Note on Scheduling Tall/Small Multiprocessor Tasks with Unit Processing Time to Minimize Maximum Tardiness A Note on Scheduling Tall/Sall Multiprocessor Tasks with Unit Processing Tie to Miniize Maxiu Tardiness Philippe Baptiste and Baruch Schieber IBM T.J. Watson Research Center P.O. Box 218, Yorktown Heights,

More information

Intelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines

Intelligent Systems: Reasoning and Recognition. Perceptrons and Support Vector Machines Intelligent Systes: Reasoning and Recognition Jaes L. Crowley osig 1 Winter Seester 2018 Lesson 6 27 February 2018 Outline Perceptrons and Support Vector achines Notation...2 Linear odels...3 Lines, Planes

More information

General Properties of Radiation Detectors Supplements

General Properties of Radiation Detectors Supplements Phys. 649: Nuclear Techniques Physics Departent Yarouk University Chapter 4: General Properties of Radiation Detectors Suppleents Dr. Nidal M. Ershaidat Overview Phys. 649: Nuclear Techniques Physics Departent

More information

Interactive Markov Models of Evolutionary Algorithms

Interactive Markov Models of Evolutionary Algorithms Cleveland State University EngagedScholarship@CSU Electrical Engineering & Coputer Science Faculty Publications Electrical Engineering & Coputer Science Departent 2015 Interactive Markov Models of Evolutionary

More information

Ch 12: Variations on Backpropagation

Ch 12: Variations on Backpropagation Ch 2: Variations on Backpropagation The basic backpropagation algorith is too slow for ost practical applications. It ay take days or weeks of coputer tie. We deonstrate why the backpropagation algorith

More information

Computable Shell Decomposition Bounds

Computable Shell Decomposition Bounds Coputable Shell Decoposition Bounds John Langford TTI-Chicago jcl@cs.cu.edu David McAllester TTI-Chicago dac@autoreason.co Editor: Leslie Pack Kaelbling and David Cohn Abstract Haussler, Kearns, Seung

More information

Figure 1: Equivalent electric (RC) circuit of a neurons membrane

Figure 1: Equivalent electric (RC) circuit of a neurons membrane Exercise: Leaky integrate and fire odel of neural spike generation This exercise investigates a siplified odel of how neurons spike in response to current inputs, one of the ost fundaental properties of

More information

Sequence Analysis, WS 14/15, D. Huson & R. Neher (this part by D. Huson) February 5,

Sequence Analysis, WS 14/15, D. Huson & R. Neher (this part by D. Huson) February 5, Sequence Analysis, WS 14/15, D. Huson & R. Neher (this part by D. Huson) February 5, 2015 31 11 Motif Finding Sources for this section: Rouchka, 1997, A Brief Overview of Gibbs Sapling. J. Buhler, M. Topa:

More information

A Better Algorithm For an Ancient Scheduling Problem. David R. Karger Steven J. Phillips Eric Torng. Department of Computer Science

A Better Algorithm For an Ancient Scheduling Problem. David R. Karger Steven J. Phillips Eric Torng. Department of Computer Science A Better Algorith For an Ancient Scheduling Proble David R. Karger Steven J. Phillips Eric Torng Departent of Coputer Science Stanford University Stanford, CA 9435-4 Abstract One of the oldest and siplest

More information

On the Communication Complexity of Lipschitzian Optimization for the Coordinated Model of Computation

On the Communication Complexity of Lipschitzian Optimization for the Coordinated Model of Computation journal of coplexity 6, 459473 (2000) doi:0.006jco.2000.0544, available online at http:www.idealibrary.co on On the Counication Coplexity of Lipschitzian Optiization for the Coordinated Model of Coputation

More information

Ph 20.3 Numerical Solution of Ordinary Differential Equations

Ph 20.3 Numerical Solution of Ordinary Differential Equations Ph 20.3 Nuerical Solution of Ordinary Differential Equations Due: Week 5 -v20170314- This Assignent So far, your assignents have tried to failiarize you with the hardware and software in the Physics Coputing

More information

A method to determine relative stroke detection efficiencies from multiplicity distributions

A method to determine relative stroke detection efficiencies from multiplicity distributions A ethod to deterine relative stroke detection eiciencies ro ultiplicity distributions Schulz W. and Cuins K. 2. Austrian Lightning Detection and Inoration Syste (ALDIS), Kahlenberger Str.2A, 90 Vienna,

More information

A Smoothed Boosting Algorithm Using Probabilistic Output Codes

A Smoothed Boosting Algorithm Using Probabilistic Output Codes A Soothed Boosting Algorith Using Probabilistic Output Codes Rong Jin rongjin@cse.su.edu Dept. of Coputer Science and Engineering, Michigan State University, MI 48824, USA Jian Zhang jian.zhang@cs.cu.edu

More information

1 Proof of learning bounds

1 Proof of learning bounds COS 511: Theoretical Machine Learning Lecturer: Rob Schapire Lecture #4 Scribe: Akshay Mittal February 13, 2013 1 Proof of learning bounds For intuition of the following theore, suppose there exists a

More information

E0 370 Statistical Learning Theory Lecture 6 (Aug 30, 2011) Margin Analysis

E0 370 Statistical Learning Theory Lecture 6 (Aug 30, 2011) Margin Analysis E0 370 tatistical Learning Theory Lecture 6 (Aug 30, 20) Margin Analysis Lecturer: hivani Agarwal cribe: Narasihan R Introduction In the last few lectures we have seen how to obtain high confidence bounds

More information

Randomized Accuracy-Aware Program Transformations For Efficient Approximate Computations

Randomized Accuracy-Aware Program Transformations For Efficient Approximate Computations Randoized Accuracy-Aware Progra Transforations For Efficient Approxiate Coputations Zeyuan Allen Zhu Sasa Misailovic Jonathan A. Kelner Martin Rinard MIT CSAIL zeyuan@csail.it.edu isailo@it.edu kelner@it.edu

More information

Tight Bounds for Maximal Identifiability of Failure Nodes in Boolean Network Tomography

Tight Bounds for Maximal Identifiability of Failure Nodes in Boolean Network Tomography Tight Bounds for axial Identifiability of Failure Nodes in Boolean Network Toography Nicola Galesi Sapienza Università di Roa nicola.galesi@uniroa1.it Fariba Ranjbar Sapienza Università di Roa fariba.ranjbar@uniroa1.it

More information

Research in Area of Longevity of Sylphon Scraies

Research in Area of Longevity of Sylphon Scraies IOP Conference Series: Earth and Environental Science PAPER OPEN ACCESS Research in Area of Longevity of Sylphon Scraies To cite this article: Natalia Y Golovina and Svetlana Y Krivosheeva 2018 IOP Conf.

More information

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation

Course Notes for EE227C (Spring 2018): Convex Optimization and Approximation Course Notes for EE227C (Spring 2018): Convex Optiization and Approxiation Instructor: Moritz Hardt Eail: hardt+ee227c@berkeley.edu Graduate Instructor: Max Sichowitz Eail: sichow+ee227c@berkeley.edu October

More information

Defect-Aware SOC Test Scheduling

Defect-Aware SOC Test Scheduling Defect-Aware SOC Test Scheduling Erik Larsson +, Julien Pouget*, and Zebo Peng + Ebedded Systes Laboratory + LIRMM* Departent of Coputer Science Montpellier 2 University Linköpings universitet CNRS Sweden

More information

Inference in the Presence of Likelihood Monotonicity for Polytomous and Logistic Regression

Inference in the Presence of Likelihood Monotonicity for Polytomous and Logistic Regression Advances in Pure Matheatics, 206, 6, 33-34 Published Online April 206 in SciRes. http://www.scirp.org/journal/ap http://dx.doi.org/0.4236/ap.206.65024 Inference in the Presence of Likelihood Monotonicity

More information

LogLog-Beta and More: A New Algorithm for Cardinality Estimation Based on LogLog Counting

LogLog-Beta and More: A New Algorithm for Cardinality Estimation Based on LogLog Counting LogLog-Beta and More: A New Algorith for Cardinality Estiation Based on LogLog Counting Jason Qin, Denys Ki, Yuei Tung The AOLP Core Data Service, AOL, 22000 AOL Way Dulles, VA 20163 E-ail: jasonqin@teaaolco

More information

Biostatistics Department Technical Report

Biostatistics Department Technical Report Biostatistics Departent Technical Report BST006-00 Estiation of Prevalence by Pool Screening With Equal Sized Pools and a egative Binoial Sapling Model Charles R. Katholi, Ph.D. Eeritus Professor Departent

More information

Optimal Resource Allocation in Multicast Device-to-Device Communications Underlaying LTE Networks

Optimal Resource Allocation in Multicast Device-to-Device Communications Underlaying LTE Networks 1 Optial Resource Allocation in Multicast Device-to-Device Counications Underlaying LTE Networks Hadi Meshgi 1, Dongei Zhao 1 and Rong Zheng 2 1 Departent of Electrical and Coputer Engineering, McMaster

More information

Constant-Space String-Matching. in Sublinear Average Time. (Extended Abstract) Wojciech Rytter z. Warsaw University. and. University of Liverpool

Constant-Space String-Matching. in Sublinear Average Time. (Extended Abstract) Wojciech Rytter z. Warsaw University. and. University of Liverpool Constant-Space String-Matching in Sublinear Average Tie (Extended Abstract) Maxie Crocheore Universite de Marne-la-Vallee Leszek Gasieniec y Max-Planck Institut fur Inforatik Wojciech Rytter z Warsaw University

More information

Chapter 6 1-D Continuous Groups

Chapter 6 1-D Continuous Groups Chapter 6 1-D Continuous Groups Continuous groups consist of group eleents labelled by one or ore continuous variables, say a 1, a 2,, a r, where each variable has a well- defined range. This chapter explores:

More information

arxiv: v1 [cs.ds] 17 Mar 2016

arxiv: v1 [cs.ds] 17 Mar 2016 Tight Bounds for Single-Pass Streaing Coplexity of the Set Cover Proble Sepehr Assadi Sanjeev Khanna Yang Li Abstract arxiv:1603.05715v1 [cs.ds] 17 Mar 2016 We resolve the space coplexity of single-pass

More information

Combining Classifiers

Combining Classifiers Cobining Classifiers Generic ethods of generating and cobining ultiple classifiers Bagging Boosting References: Duda, Hart & Stork, pg 475-480. Hastie, Tibsharini, Friedan, pg 246-256 and Chapter 10. http://www.boosting.org/

More information

Proc. of the IEEE/OES Seventh Working Conference on Current Measurement Technology UNCERTAINTIES IN SEASONDE CURRENT VELOCITIES

Proc. of the IEEE/OES Seventh Working Conference on Current Measurement Technology UNCERTAINTIES IN SEASONDE CURRENT VELOCITIES Proc. of the IEEE/OES Seventh Working Conference on Current Measureent Technology UNCERTAINTIES IN SEASONDE CURRENT VELOCITIES Belinda Lipa Codar Ocean Sensors 15 La Sandra Way, Portola Valley, CA 98 blipa@pogo.co

More information

On Poset Merging. 1 Introduction. Peter Chen Guoli Ding Steve Seiden. Keywords: Merging, Partial Order, Lower Bounds. AMS Classification: 68W40

On Poset Merging. 1 Introduction. Peter Chen Guoli Ding Steve Seiden. Keywords: Merging, Partial Order, Lower Bounds. AMS Classification: 68W40 On Poset Merging Peter Chen Guoli Ding Steve Seiden Abstract We consider the follow poset erging proble: Let X and Y be two subsets of a partially ordered set S. Given coplete inforation about the ordering

More information

Curious Bounds for Floor Function Sums

Curious Bounds for Floor Function Sums 1 47 6 11 Journal of Integer Sequences, Vol. 1 (018), Article 18.1.8 Curious Bounds for Floor Function Sus Thotsaporn Thanatipanonda and Elaine Wong 1 Science Division Mahidol University International

More information

Kinetic Theory of Gases: Elementary Ideas

Kinetic Theory of Gases: Elementary Ideas Kinetic Theory of Gases: Eleentary Ideas 17th February 2010 1 Kinetic Theory: A Discussion Based on a Siplified iew of the Motion of Gases 1.1 Pressure: Consul Engel and Reid Ch. 33.1) for a discussion

More information

Bayes Decision Rule and Naïve Bayes Classifier

Bayes Decision Rule and Naïve Bayes Classifier Bayes Decision Rule and Naïve Bayes Classifier Le Song Machine Learning I CSE 6740, Fall 2013 Gaussian Mixture odel A density odel p(x) ay be ulti-odal: odel it as a ixture of uni-odal distributions (e.g.

More information

lecture 36: Linear Multistep Mehods: Zero Stability

lecture 36: Linear Multistep Mehods: Zero Stability 95 lecture 36: Linear Multistep Mehods: Zero Stability 5.6 Linear ultistep ethods: zero stability Does consistency iply convergence for linear ultistep ethods? This is always the case for one-step ethods,

More information

OBJECTIVES INTRODUCTION

OBJECTIVES INTRODUCTION M7 Chapter 3 Section 1 OBJECTIVES Suarize data using easures of central tendency, such as the ean, edian, ode, and idrange. Describe data using the easures of variation, such as the range, variance, and

More information

Compression and Predictive Distributions for Large Alphabet i.i.d and Markov models

Compression and Predictive Distributions for Large Alphabet i.i.d and Markov models 2014 IEEE International Syposiu on Inforation Theory Copression and Predictive Distributions for Large Alphabet i.i.d and Markov odels Xiao Yang Departent of Statistics Yale University New Haven, CT, 06511

More information

Intelligent Systems: Reasoning and Recognition. Artificial Neural Networks

Intelligent Systems: Reasoning and Recognition. Artificial Neural Networks Intelligent Systes: Reasoning and Recognition Jaes L. Crowley MOSIG M1 Winter Seester 2018 Lesson 7 1 March 2018 Outline Artificial Neural Networks Notation...2 Introduction...3 Key Equations... 3 Artificial

More information

SPECTRUM sensing is a core concept of cognitive radio

SPECTRUM sensing is a core concept of cognitive radio World Acadey of Science, Engineering and Technology International Journal of Electronics and Counication Engineering Vol:6, o:2, 202 Efficient Detection Using Sequential Probability Ratio Test in Mobile

More information

Pattern Recognition and Machine Learning. Artificial Neural networks

Pattern Recognition and Machine Learning. Artificial Neural networks Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2016 Lessons 7 14 Dec 2016 Outline Artificial Neural networks Notation...2 1. Introduction...3... 3 The Artificial

More information

USEFUL HINTS FOR SOLVING PHYSICS OLYMPIAD PROBLEMS. By: Ian Blokland, Augustana Campus, University of Alberta

USEFUL HINTS FOR SOLVING PHYSICS OLYMPIAD PROBLEMS. By: Ian Blokland, Augustana Campus, University of Alberta 1 USEFUL HINTS FOR SOLVING PHYSICS OLYMPIAD PROBLEMS By: Ian Bloland, Augustana Capus, University of Alberta For: Physics Olypiad Weeend, April 6, 008, UofA Introduction: Physicists often attept to solve

More information

On the Inapproximability of Vertex Cover on k-partite k-uniform Hypergraphs

On the Inapproximability of Vertex Cover on k-partite k-uniform Hypergraphs On the Inapproxiability of Vertex Cover on k-partite k-unifor Hypergraphs Venkatesan Guruswai and Rishi Saket Coputer Science Departent Carnegie Mellon University Pittsburgh, PA 1513. Abstract. Coputing

More information

A Model for the Selection of Internet Service Providers

A Model for the Selection of Internet Service Providers ISSN 0146-4116, Autoatic Control and Coputer Sciences, 2008, Vol. 42, No. 5, pp. 249 254. Allerton Press, Inc., 2008. Original Russian Text I.M. Aliev, 2008, published in Avtoatika i Vychislitel naya Tekhnika,

More information

arxiv: v1 [math.nt] 14 Sep 2014

arxiv: v1 [math.nt] 14 Sep 2014 ROTATION REMAINDERS P. JAMESON GRABER, WASHINGTON AND LEE UNIVERSITY 08 arxiv:1409.411v1 [ath.nt] 14 Sep 014 Abstract. We study properties of an array of nubers, called the triangle, in which each row

More information

Bayesian Learning. Chapter 6: Bayesian Learning. Bayes Theorem. Roles for Bayesian Methods. CS 536: Machine Learning Littman (Wu, TA)

Bayesian Learning. Chapter 6: Bayesian Learning. Bayes Theorem. Roles for Bayesian Methods. CS 536: Machine Learning Littman (Wu, TA) Bayesian Learning Chapter 6: Bayesian Learning CS 536: Machine Learning Littan (Wu, TA) [Read Ch. 6, except 6.3] [Suggested exercises: 6.1, 6.2, 6.6] Bayes Theore MAP, ML hypotheses MAP learners Miniu

More information

Correlated Bayesian Model Fusion: Efficient Performance Modeling of Large-Scale Tunable Analog/RF Integrated Circuits

Correlated Bayesian Model Fusion: Efficient Performance Modeling of Large-Scale Tunable Analog/RF Integrated Circuits Correlated Bayesian odel Fusion: Efficient Perforance odeling of Large-Scale unable Analog/RF Integrated Circuits Fa Wang and Xin Li ECE Departent, Carnegie ellon University, Pittsburgh, PA 53 {fwang,

More information

Kinematics and dynamics, a computational approach

Kinematics and dynamics, a computational approach Kineatics and dynaics, a coputational approach We begin the discussion of nuerical approaches to echanics with the definition for the velocity r r ( t t) r ( t) v( t) li li or r( t t) r( t) v( t) t for

More information

CS Lecture 13. More Maximum Likelihood

CS Lecture 13. More Maximum Likelihood CS 6347 Lecture 13 More Maxiu Likelihood Recap Last tie: Introduction to axiu likelihood estiation MLE for Bayesian networks Optial CPTs correspond to epirical counts Today: MLE for CRFs 2 Maxiu Likelihood

More information

Kinetic Theory of Gases: Elementary Ideas

Kinetic Theory of Gases: Elementary Ideas Kinetic Theory of Gases: Eleentary Ideas 9th February 011 1 Kinetic Theory: A Discussion Based on a Siplified iew of the Motion of Gases 1.1 Pressure: Consul Engel and Reid Ch. 33.1) for a discussion of

More information

Kernel Methods and Support Vector Machines

Kernel Methods and Support Vector Machines Intelligent Systes: Reasoning and Recognition Jaes L. Crowley ENSIAG 2 / osig 1 Second Seester 2012/2013 Lesson 20 2 ay 2013 Kernel ethods and Support Vector achines Contents Kernel Functions...2 Quadratic

More information

Easy Evaluation Method of Self-Compactability of Self-Compacting Concrete

Easy Evaluation Method of Self-Compactability of Self-Compacting Concrete Easy Evaluation Method of Self-Copactability of Self-Copacting Concrete Masanori Maruoka 1 Hiroi Fujiwara 2 Erika Ogura 3 Nobu Watanabe 4 T 11 ABSTRACT The use of self-copacting concrete (SCC) in construction

More information

arxiv: v3 [cs.lg] 7 Jan 2016

arxiv: v3 [cs.lg] 7 Jan 2016 Efficient and Parsionious Agnostic Active Learning Tzu-Kuo Huang Alekh Agarwal Daniel J. Hsu tkhuang@icrosoft.co alekha@icrosoft.co djhsu@cs.colubia.edu John Langford Robert E. Schapire jcl@icrosoft.co

More information

Analysis of Impulsive Natural Phenomena through Finite Difference Methods A MATLAB Computational Project-Based Learning

Analysis of Impulsive Natural Phenomena through Finite Difference Methods A MATLAB Computational Project-Based Learning Analysis of Ipulsive Natural Phenoena through Finite Difference Methods A MATLAB Coputational Project-Based Learning Nicholas Kuia, Christopher Chariah, Mechatronics Engineering, Vaughn College of Aeronautics

More information

Fixed-to-Variable Length Distribution Matching

Fixed-to-Variable Length Distribution Matching Fixed-to-Variable Length Distribution Matching Rana Ali Ajad and Georg Böcherer Institute for Counications Engineering Technische Universität München, Gerany Eail: raa2463@gail.co,georg.boecherer@tu.de

More information

A Simplified Analytical Approach for Efficiency Evaluation of the Weaving Machines with Automatic Filling Repair

A Simplified Analytical Approach for Efficiency Evaluation of the Weaving Machines with Automatic Filling Repair Proceedings of the 6th SEAS International Conference on Siulation, Modelling and Optiization, Lisbon, Portugal, Septeber -4, 006 0 A Siplified Analytical Approach for Efficiency Evaluation of the eaving

More information

On the Maximum Likelihood Estimation of Weibull Distribution with Lifetime Data of Hard Disk Drives

On the Maximum Likelihood Estimation of Weibull Distribution with Lifetime Data of Hard Disk Drives 314 Int'l Conf. Par. and Dist. Proc. Tech. and Appl. PDPTA'17 On the Maxiu Likelihood Estiation of Weibull Distribution with Lifetie Data of Hard Disk Drives Daiki Koizui Departent of Inforation and Manageent

More information

ASSUME a source over an alphabet size m, from which a sequence of n independent samples are drawn. The classical

ASSUME a source over an alphabet size m, from which a sequence of n independent samples are drawn. The classical IEEE TRANSACTIONS ON INFORMATION THEORY Large Alphabet Source Coding using Independent Coponent Analysis Aichai Painsky, Meber, IEEE, Saharon Rosset and Meir Feder, Fellow, IEEE arxiv:67.7v [cs.it] Jul

More information

The proofs of Theorem 1-3 are along the lines of Wied and Galeano (2013).

The proofs of Theorem 1-3 are along the lines of Wied and Galeano (2013). A Appendix: Proofs The proofs of Theore 1-3 are along the lines of Wied and Galeano (2013) Proof of Theore 1 Let D[d 1, d 2 ] be the space of càdlàg functions on the interval [d 1, d 2 ] equipped with

More information

Homework 3 Solutions CSE 101 Summer 2017

Homework 3 Solutions CSE 101 Summer 2017 Hoework 3 Solutions CSE 0 Suer 207. Scheduling algoriths The following n = 2 jobs with given processing ties have to be scheduled on = 3 parallel and identical processors with the objective of iniizing

More information

Distributed Subgradient Methods for Multi-agent Optimization

Distributed Subgradient Methods for Multi-agent Optimization 1 Distributed Subgradient Methods for Multi-agent Optiization Angelia Nedić and Asuan Ozdaglar October 29, 2007 Abstract We study a distributed coputation odel for optiizing a su of convex objective functions

More information

Department of Electronic and Optical Engineering, Ordnance Engineering College, Shijiazhuang, , China

Department of Electronic and Optical Engineering, Ordnance Engineering College, Shijiazhuang, , China 6th International Conference on Machinery, Materials, Environent, Biotechnology and Coputer (MMEBC 06) Solving Multi-Sensor Multi-Target Assignent Proble Based on Copositive Cobat Efficiency and QPSO Algorith

More information

Tracking using CONDENSATION: Conditional Density Propagation

Tracking using CONDENSATION: Conditional Density Propagation Tracking using CONDENSATION: Conditional Density Propagation Goal Model-based visual tracking in dense clutter at near video frae rates M. Isard and A. Blake, CONDENSATION Conditional density propagation

More information

Pattern Recognition and Machine Learning. Artificial Neural networks

Pattern Recognition and Machine Learning. Artificial Neural networks Pattern Recognition and Machine Learning Jaes L. Crowley ENSIMAG 3 - MMIS Fall Seester 2016/2017 Lessons 9 11 Jan 2017 Outline Artificial Neural networks Notation...2 Convolutional Neural Networks...3

More information

Graphical Models in Local, Asymmetric Multi-Agent Markov Decision Processes

Graphical Models in Local, Asymmetric Multi-Agent Markov Decision Processes Graphical Models in Local, Asyetric Multi-Agent Markov Decision Processes Ditri Dolgov and Edund Durfee Departent of Electrical Engineering and Coputer Science University of Michigan Ann Arbor, MI 48109

More information

Randomized Recovery for Boolean Compressed Sensing

Randomized Recovery for Boolean Compressed Sensing Randoized Recovery for Boolean Copressed Sensing Mitra Fatei and Martin Vetterli Laboratory of Audiovisual Counication École Polytechnique Fédéral de Lausanne (EPFL) Eail: {itra.fatei, artin.vetterli}@epfl.ch

More information

Support recovery in compressed sensing: An estimation theoretic approach

Support recovery in compressed sensing: An estimation theoretic approach Support recovery in copressed sensing: An estiation theoretic approach Ain Karbasi, Ali Horati, Soheil Mohajer, Martin Vetterli School of Coputer and Counication Sciences École Polytechnique Fédérale de

More information

ma x = -bv x + F rod.

ma x = -bv x + F rod. Notes on Dynaical Systes Dynaics is the study of change. The priary ingredients of a dynaical syste are its state and its rule of change (also soeties called the dynaic). Dynaical systes can be continuous

More information