Effective Bandwidth for Traffic Engineering

Similar documents
Stochastic Network Calculus

These are special traffic patterns that create more stress on a switch

Part 2: Random Routing and Load Balancing

Efficient Nonlinear Optimizations of Queuing Systems

Queueing Theory I Summary! Little s Law! Queueing System Notation! Stationary Analysis of Elementary Queueing Systems " M/M/1 " M/M/m " M/M/1/K "

A Measurement-Analytic Approach for QoS Estimation in a Network Based on the Dominant Time Scale

COMP9334: Capacity Planning of Computer Systems and Networks

Capacity management for packet-switched networks with heterogeneous sources. Linda de Jonge. Master Thesis July 29, 2009.

Author Entropy vs. File Size in the GNOME Suite of Applications

Asymptotic Delay Distribution and Burst Size Impact on a Network Node Driven by Self-similar Traffic

Resource Allocation for Video Streaming in Wireless Environment

Performance Analysis of Priority Queueing Schemes in Internet Routers

Different Scenarios of Concatenation at Aggregate Scheduling of Multiple Nodes

Singular perturbation analysis of an additive increase multiplicative decrease control algorithm under time-varying buffering delays.

SAMPLING AND INVERSION

An Admission Control Mechanism for Providing Service Differentiation in Optical Burst-Switching Networks

A Demand Response Calculus with Perfect Batteries

Input-queued switches: Scheduling algorithms for a crossbar switch. EE 384X Packet Switch Architectures 1

Burst Scheduling Based on Time-slotting and Fragmentation in WDM Optical Burst Switched Networks

Congestion In Large Balanced Fair Links

Lecture 7: Simulation of Markov Processes. Pasi Lassila Department of Communications and Networking

CHAPTER 4. Networks of queues. 1. Open networks Suppose that we have a network of queues as given in Figure 4.1. Arrivals

Embedded Systems 23 BF - ES

Effects of Gap Open and Gap Extension Penalties

On the Impact of Link Scheduling on End-to-End Delays in Large Networks

requests/sec. The total channel load is requests/sec. Using slot as the time unit, the total channel load is 50 ( ) = 1

Dynamic Power Allocation and Routing for Time Varying Wireless Networks

Approximate Queueing Model for Multi-rate Multi-user MIMO systems.

Competitive Management of Non-Preemptive Queues with Multiple Values

IEEE TRANSACTIONS ON AUTOMATIC CONTROL, VOL. 43, NO. 3, MARCH

Maximum Likelihood Estimation of the Flow Size Distribution Tail Index from Sampled Packet Data

A Retrial Queueing model with FDL at OBS core node

A discrete-time priority queue with train arrivals

A Study of Traffic Statistics of Assembled Burst Traffic in Optical Burst Switched Networks

The Burstiness Behavior of Regulated Flows in Networks

Fuzzy Optimization and Normal Simulation for Solving Fuzzy Web Queuing System Problems

Design of IP networks with Quality of Service

Dynamic resource sharing

An Overview of Traffic Matrix Estimation Methods

Bounded Delay for Weighted Round Robin with Burst Crediting

A Generalized Processor Sharing Approach to Flow Control in Integrated Services Networks: The Single Node Case. 1

Modelling the Arrival Process for Packet Audio

On queueing in coded networks queue size follows degrees of freedom

A Virtual Queue Approach to Loss Estimation

Quiz 1 EE 549 Wednesday, Feb. 27, 2008

Manufacturing System Flow Analysis

Convexity Properties of Loss and Overflow Functions

The Analysis of Microburst (Burstiness) on Virtual Switch

CDA5530: Performance Models of Computers and Networks. Chapter 4: Elementary Queuing Theory

Discrete-event simulations

Computer Systems Modelling

M/G/FQ: STOCHASTIC ANALYSIS OF FAIR QUEUEING SYSTEMS

ring structure Abstract Optical Grid networks allow many computing sites to share their resources by connecting

Wireless Internet Exercises

A Starvation-free Algorithm For Achieving 100% Throughput in an Input- Queued Switch

SPLITTING AND MERGING OF PACKET TRAFFIC: MEASUREMENT AND MODELLING

Mice and Elephants Visualization of Internet

Link Models for Circuit Switching

Network Traffic Characteristic

Motivation. Lecture 2 Topics from Optimization and Duality. network utility maximization (NUM) problem:

Laboratory Exercise X Most Dangerous Places to Live in the United States Based on Natural Disasters

Internet Traffic Modeling and Its Implications to Network Performance and Control

Signalling Analysis for Adaptive TCD Routing in ISL Networks *

Minimal Delay Traffic Grooming for WDM Optical Star Networks

F.P. KELLY Statistical Laboratory, University of Cambridge, 16 Mill Lane, Cambridge CB2 1SB, England

Chapter 2: A Fluid Model Analysis of Streaming Media in the Presence of Time- Varying Bandwidth

A Network Calculus with Effective Bandwidth

On Tandem Blocking Queues with a Common Retrial Queue

Real-Time Calculus. LS 12, TU Dortmund

Signal Structure for a Class of Nonlinear Dynamic Systems

SUM x. 2x y x. x y x/2. (i)

Modeling and Analysis of Traffic in High Speed Networks

The Multi-Path Utility Maximization Problem

Non Markovian Queues (contd.)

Robustness of stochastic discrete-time switched linear systems

Analysis of Scalable TCP in the presence of Markovian Losses

Rate Control in Communication Networks

Statistical Per-Flow Service Bounds in a Network with Aggregate Provisioning

CONTENTION RESOLUTION IN OPTICAL BURST SWITCHES USING FIBER DELAY LINE BUFFERS

An approach to service provisioning with quality of service requirements in ATM networks

Solutions to COMP9334 Week 8 Sample Problems

Capturing Network Traffic Dynamics Small Scales. Rolf Riedi

An engineering approximation for the mean waiting time in the M/H 2 b /s queue

Analysis of Software Artifacts

CONVERGENCE TO FRACTIONAL BROWNIAN MOTION AND LOSS PROBABILITY. Jin-Chun Kim and Hee-Choon Lee

Min Congestion Control for High- Speed Heterogeneous Networks. JetMax: Scalable Max-Min

Load Balancing in Distributed Service System: A Survey

Assured Horizon A new Combined Framework for Burst Assembly and Reservation in Optical Burst Switched Networks

Sub-Optimal Scheduling of a Flexible Batch Manufacturing System using an Integer Programming Solution

Queuing Networks: Burke s Theorem, Kleinrock s Approximation, and Jackson s Theorem. Wade Trappe

Information in Aloha Networks

Technion - Computer Science Department - Technical Report CS On Centralized Smooth Scheduling

End-to-end Estimation of the Available Bandwidth Variation Range

Introduction to queuing theory

NICTA Short Course. Network Analysis. Vijay Sivaraman. Day 1 Queueing Systems and Markov Chains. Network Analysis, 2008s2 1-1

Scheduling: Queues & Computation

On the Price of Anarchy in Unbounded Delay Networks

A New Technique for Link Utilization Estimation

CHAPTER 3 STOCHASTIC MODEL OF A GENERAL FEED BACK QUEUE NETWORK

CPU SCHEDULING RONG ZHENG

Transcription:

Brigham Young University BYU ScholarsArchive All Faculty Publications 2-5- Effective Bandwidth for Traffic Engineering Mark J. Clement clement@cs.byu.edu Rob Kunz See next page for additional authors Follow this and additional works at: https://scholarsarchive.byu.edu/facpub Part of the Computer Sciences Commons Original Publication Citation Effective Bandwidth for Traffic Engineering, Rob Kunz, Seth Nielson, Mark Clement, Quinn Snell, Proceedings of the IEEE Workshop on High Performance Switching and Routing (HPSR 2), Dallas, TX, May 2. BYU ScholarsArchive Citation Clement, Mark J.; Kunz, Rob; Nielson, Seth; and Snell, Quinn O., "Effective Bandwidth for Traffic Engineering" (2). All Faculty Publications. 573. https://scholarsarchive.byu.edu/facpub/573 This Peer-Reviewed Article is brought to you for free and open access by BYU ScholarsArchive. It has been accepted for inclusion in All Faculty Publications by an authorized administrator of BYU ScholarsArchive. For more information, please contact scholarsarchive@byu.edu, ellen_amatangelo@byu.edu.

Authors Mark J. Clement, Rob Kunz, Seth Nielson, and Quinn O. Snell This peer-reviewed article is available at BYU ScholarsArchive: https://scholarsarchive.byu.edu/facpub/573

Effective Bandwidth for Traffic Engineering Rob Kunz, Seth Nielson, Mark Clement, Quinn Snell Brigham Young University Department of Computer Science Brigham Young University Provo, UT 8462 {robby, darkman, clement, snell}@cs.byu.edu Abstract- In today s Internet, demand is increasing for guarantees of speed and efficiency. Current routers are very limited in the type and quantity of observed data they can provide, making it difficult for providers to maximize utilization without the risk of degraded throughput. This research uses statistical data currents provided by router vendors to estimate the impact of changes in network configuration on the probability of link overflow. This allows service providers to calculate in advance, the effect of grooming on a network, eliminating the conservative trial-and-error approach normally used. These predictions are made using Large Deviation Theory, which focuses on the tails of the distribution, giving a better estimate than average and peak values. I. INTRODUCTION In today s Internet, demand is increasing for guarantees of speed and efficiency. Service providers are motivated to find a way to multiplex as much traffic as possible without breaking any service guarantees and to also find a way to charge for the guaranteed bandwidth they are providing. The Theory of Effective Bandwidth is designed to accurately estimate the resources that a particular source requires and the degree of statistical multiplexing that can occur while maintaining a reasonable packet drop probability. This research uses traces from AT&T frame relay trunks to determine the effective bandwidth of individual connections through the network. Individual connections can be moved to new routes within the network in order to increase traffic balance. The principle metric used is the probability of overflow for each link, computed from the effective bandwidth equations. Large trunk routers are limited in the amount of trace data that they can practically obtain and maintain. Much of this research has focused on using data that routers currently provide, rather than making optimal calculations with data that is not generally available. Much of this research focuses on converting gross trace data from several thousand trunks and several hundred thousand circuits in the AT&T network into detailed traces for each circuit. II. EFFECTIVE BANDWIDTH Using Large Deviation Theory (which focuses on the impact of the tails of a distribution), Effective Bandwidth computes a bandwidth that is a better estimate than just average and peak values. The Effective Bandwidth is computed by using source characteristics along with the link through which it is passing. The effective bandwidth of a source given the operating point of a link is given by the following equation. j ( s, t) = log E e st sx j [ ] [, t ] α () In the equation for Effective Bandwidth, the two most enigmatic variables are the s and t parameters. One source states, Particular values of (s,t) can be taken to characterize the operating point of a link, []. This same source characterizes the s, or space, parameter as the degree of statistical multiplexing potential of a link. A large value of s indicates that a low degree of statistical multiplexing can be performed, whereas a small value indicates a large degree of potential statistical multiplexing. It should be noted that the s parameters is in units of kb -. The t parameter is used to characterize the most probable duration of the buffer busy period prior to overflow and also represents the minimum granularity necessary for traces to capture the statistical properties that affect buffer overflow []. It should be noted that the t parameter is in units of milliseconds. Taken together, these parameters form the st parameter which is in units ms/kb. Conceptually, this represents the amount of multiplexing that can happen in a period prior to overflow. These two parameters are dependent on all of the characteristics of the link: buffer, capacity, QoS and scheduling. It is also interesting to note that they are dependent (also) on the sources, but with a large degree of multiplexing, this definition loop is broken []. Taking s and t into account, the effective bandwidth equation begins to become clear. The Effective Bandwidth of a source is characterized by the number of cells that it produces divided by the st parameter. It should be noted that the number of cells produced by the Effective -783-67-//$. (C) 2 IEEE 9

Bandwidth is computed using a logarithmic moment generating function [2]. A moment generating function produces the central moments about a probability density function. The first moment is the mean. In this equation, the moment generating function assumes that the distribution of cells of a source follows some probability density function or some random variable X[,t] where [,t] is the time interval in which the cells are produced. A logarithmic moment generating function is used as opposed to a moment generating function because it allows for better estimates in the tails of the distribution, or in the peaks. Once this logarithmic moment value has been computed, it is divided by the st parameter. Basically, this means that effective bandwidth is equal to the most probable number of cells produced in a given [,t] time period divided by the amount of multiplexing that can be performed in the same time period. Of course, it is difficult to know the particular probability density function of your cell distribution. Fortunately, we can obtain an estimate using empirical data. The modified Effective Bandwidth equation is defined using an average of the production of cells in various time periods []. Using this equation, the logarithmic moment generating function can be computed using trace files. The empirical form of the Effective Bandwidth bears a strong resemblance to research into Self-Similarity. For a given value of t, all of the data sent during the time interval between ( i ) t and it is aggregated into a single value. A search can then be made for the interval that produces the most likely buffer utilization prior to overflow. The empirical form for the effective bandwidth equation [5] is T / t sx [( ), ] j i t it α j ( s, t) = log e. / (2) st T t i= At this point, we integrate the Effective Bandwidth of the sources into a link with a capacity C and a buffer B. The sup-inf formula (3) is used to estimate the values of s and t given empirical data. lim log Q( Nc, Nb, Nn) N N N = sup inf st n jα j ( s, t) s( ct + b) = I t s j= (3) This formula integrates the effective bandwidths of the various sources being multiplexed and evaluates them using the capacity and buffer of the link. This equation is equal to -I and the Probability of Overflow is equal to e -NI + o(n) and o(n) goes to as N is large [5]. It should be noted that N represents the number of combined sources that are entering the link. That is, if the link has three types of sources, source A, B anc C, and A comprises 2% of the flow, B comprises 4% and C comprises 4%, then n A =.2, n B =.4 and n C =.4 and N is the number of these streams entering the system. Once the s and t parameters are computed from the sup-inf formula (3), the effective bandwidth equation (2) can be completed. It has been shown through theoretical and empirical studies that changes in traffic mix have a very light impact on these parameters. Once they have been computed, as long as the traffic mix does not change drastically, new sources bandwidth could be easily computed without re-computing the entire sup-inf (which would require trace files and some computational time). As an aside, the sup-inf formula (3) requires that the formula be computed using two optimizations. First, for each t, an s is found that minimizes the inside of sup-inf. The largest value resulting from each of these t s is the result of the function. Remember that t represents the most probable time of the buffer being busy prior to overflow and s represents multiplexing. In searching for a smallest s we are actually searching for a worst case scenario: the smallest value of s that will still cause overflow for that value of t. The sup-inf formula effectively looks at the tail of the distribution rather than in the center. This allows for greater protection against over optimistic appraisals of link utilization. On the down side, these equations have been found to calculate cell loss rates and probability of overflow higher than in practice or simulation. In their defense, you probably will never loose a customer because you allow him too much bandwidth. The s parameter can be found because of the convexity of the equation. The Effective Bandwidth equation is determined by s in a logarithmic manner and has a slow increase. The s(b+ct) is determined by s but is constantly increasing. The minimal value of the difference between the effective bandwidth and the s(b+ct) terms can be computed using a Golden Section Search. Java code has been written to take a trace file and generate the effective bandwidth and the s and t parameters. These values can then be used to predict overflow probabilities and to perform other modeling tasks. We have tested the code on existing traces and have been able to calculate similar values to other researchers in order to verify our techniques. III. TRACE DATA Before the effective bandwidth of sources can be determined, trace data must be available. Chris Chase from AT&T provided us with a large quantity of trace data summarizing several thousand trunks and several hundred thousand connections. The trace data contains hourly averages and 5 minute peak values for each of the trunks. Given this data, we had to derive specific connection values from those gross trunk values. 2

Fig. 3. Network in fig. 2 represented as a matrix. IV. DERIVING CONNECTION DATA Fig.. Traffic Engineering overview. Fig. provides an overview of the trace generation process. First we extracted trunk statistics and circuit information from the statistic files. We then use Gaussian elimination and Linear Programming to calculate individual circuit values from the trunk statistics. This results in percircuit trace files. These files can be used to compute the effective bandwidth contributions of each circuit. The circuits can then be moved to a different path and the new effective bandwidth can be predicted. It is hoped that this technique can be used to come up with a more balanced network load across all of the trunks. C B A 2 25 D E C+B=2 A+B+C=25 --- A=5 D+E= D+E+B=5 --- B=5, C=5 Any values for D and E that satisfy D+E= and static constraints Fig. 2. Example network with 4 trunks with current throughput values of 2,25,,5 and 5 circuits A, B, C, D. 5 In order to derive connection data from trunk data, the contributions of each circuit must be determined from trunk statistics. Fig 2 shows a hypothetical network with trunk throughput shown in the links and five circuits (A,B,C,D,E). We must estimate the throughput contribution of each circuit given only trunk values. For example, only paths E and D cross the section of weight so we know that E+D=. Paths C and B cross the section of weight 2 so C+B=2. Paths A+B+C = 25 and paths D+B+E=5. With these 4 equations and 5 unknowns we are able to solve for all but E and D. Although this simple example can be solved by observation, a mechanized approach must be developed for systems with 5, equations and 4, unknowns The most plausible algorithm we ve found is Gaussian elimination. Fig. 2 can be represented as a matrix with trunks as rows, circuits as columns and a in the matrix if the circuit passes through the given trunk (fig. 3). Through Gaussian elimination, we can create an upper diagonal matrix where the diagonal is all ones and anything below the diagonal is zero (fig. 4). Using back substitution, we can entirely solve the system of equations, once the bottom row has been satisfied. In our small example, D and E can be set to 5 since D + E =. If there were static constraints (D is a T connection and E is a DS3) we could make better estimates for these unconstrained variables. We then proceed to previous row where C D E = 5. Since D + E =, we know that C = 5. Row 2 indicates that B + C = 2. Since C = 5, we know that B = 5. On the top row, A+B+C=25. Since B+C = 2 we know that A = 5. This is the basic algorithm. The two keys are providing good estimates for unconstrained variables and resolving inconsistencies. Estimates are made using static constraints for the connections in connection with linear programming. This gives us a proven statistical basis for our estimates. 2

T parameters T 5 4 3 2 Original Trunk T added T removed 5 5 5 Fig. 4. Network matrix after Gaussian Elimination. Buffer Size For example, an OC-3c connection will have a much higher possible value than a T circuit. An inconsistency occurs when it is impossible to solve one of the rows in the matrix while satisfying the static constraints. For example, if a row required a 56k connection to be 64k we have an inconsistency. In order to deal with this problem, we backtrack to the point where the other parameters were set and change them by raising or lowering them as well as all other nodes on the line so that each solution still works. We then work the changes into each proceeding row and repeat this process until the system of equations is solved. By setting the trunk values on the right to the contents of the traces and making each row a connection with a weight that depends on its speed, we can create trace files for each of the 4, connections. The size of the individual trace depends on the number of samples in the original trace. Since our current files contain 24 measurements * 4 days, we can generate trace files with 96 values. From these trace files we can compute the effective bandwidth and determine probabilities of overflow. We can also get an idea of the traffic flows that can be expected in the network. Our next task is to use linear programming to determine the statistically best solutions to the problem. We will use the 5-minute peak values to help determine the contribution of each circuit to trunk statistics. Fig. 5. The T parameter is the most probable period of buffer utilization prior to overflow. When a T line is removed, the T parameter is higher for a given Buffer Size showing that there is a longer buffer busy period prior to overflow. be minimized [3]. Delay and jitter are usually seen as important values to optimize in making traffic grooming choices [4]. Fig 5 shows the impact of adding and removing a T line on the T parameter. When a T line is removed, the T parameter is higher for a given Buffer Size showing that there is a longer buffer busy period prior to overflow. For a buffer size of 5 packets, the original trunk had a T value of 6. When the T was removed, the T parameter increased to 4 showing that the busy period prior to overflow was higher due to decreased utilization. When a T was added, the T parameter dropped to 7 indicating that the busy period prior to overflow was lower, roughly indicating that the probability of loss is greater with the additional T line..2 Probability of Overflow V. RESULTS We performed some experiments to determine the potential of this approach in traffic grooming. In order to do this we picked a trunk from the trace data and moved a T circuit into the trunk and out of the trunk to determine the impact as measured by effective bandwidth. These statistics can be used to determine the appropriate balance between loading on different trunks. Effective Bandwidth analysis provides information on the probability of packet loss given a specified number of buffers and line capacity. The roughly translates into queue length and delay for each circuit. If the aggregate probability of loss is minimized across the network, then delay will also Probability.8.6.4.2 5 5 5 Buffer Size Original Trunk T added T removed Fig. 6. The probability of overflow for different Buffer sizes as traffic is groomed. 22

Probability.9.8.7.6.5.4.3.2. Probability of Overflow 5 Buffer Size Original Trunk T added T removed Fig. 7. Overflow probability for larger buffer sizes. When a T line connection is added to a trunk, the probability of overflow increases. The effect is non-linear though, and is dependent on the buffer size and the burstiness of the traffic. Several examples of this effect are shown for different buffer sizes in Fig 6. (Fig 7 shows an expanded view of the overflow probability for larger buffer sizes.) VI. CONCLUSION Traffic Engineering and grooming are difficult to perform without tools to provide predicted results for potential changes. This research uses trace data that is readily available from trunk routers to determine the impact of routing changes on network balance. Large Deviation theory is used to examine the impact of changes on overflow probability, allowing a much more detailed analysis than mean values would provide. Application of this research should allow traffic engineers to increase the utilization of their trunks without degrading Quality of Service for their customers. ACKNOWLEDGMENT This research is funded by AT&T Corporation. Thanks to Chris Chase who provided trace files and helpful suggestions. REFERENCES [] V.A. Siris, Large deviation techniques for traffic engineering, http://www.ics.forth.gr/netgroup/msa, Last Access: 3/7/2 [2] C. Coutcoubetis, A. Dimakis, and G.D. Stamoulis, Application and evaluation of large deviation techniques for traffic engineering in broadband networks, In Proc. of ACM SIGMETRICS 98/PERFORMANCE 98, Madison, Wisconsin, June 998 [3] Srinivas Vutukury, J.J. Garcie-Luna-Aceves, "A traffic engineering approach based on minimum-delay routing, In Proc. of the Ninth International Conference on Computer Communications and Networks. Las Vegas, Nevada, Octover 2. [4] A. Feldmann, A. Greenberg, C. Lund, N. Reingold and J. Rexford "NetScope: Traffic engineering for IP networks," In IEEE Network, March/April 2. [5] F. P. Kelly, "Notes on effective bandwidths," Stochastic Networks: Theory and Applications Telecommunications Networks, volume 4 of Royal Statistical Society Lecture Notes Series, pp. 4-68, Oxford University Press. 23