分享
 
 
 

RFC3272 - Overview and Principles of Internet Traffic Engineering

王朝other·作者佚名  2008-05-31
窄屏简体版  字體: |||超大  

Network Working Group D. AwdUChe

Request for Comments: 3272 Movaz Networks

Category: Informational A. Chiu

Celion Networks

A. Elwalid

I. Widjaja

Lucent Technologies

X. Xiao

Redback Networks

May 2002

Overview and Principles of Internet Traffic Engineering

Status of this Memo

This memo provides information for the Internet community. It does

not specify an Internet standard of any kind. Distribution of this

memo is unlimited.

Copyright Notice

Copyright (C) The Internet Society (2002). All Rights Reserved.

Abstract

This memo describes the principles of Traffic Engineering (TE) in the

Internet. The document is intended to promote better understanding

of the issues surrounding traffic engineering in IP networks, and to

provide a common basis for the development of traffic engineering

capabilities for the Internet. The principles, architectures, and

methodologies for performance evaluation and performance optimization

of operational IP networks are discussed throughout this document.

Table of Contents

1.0 Introduction...................................................3

1.1 What is Internet Traffic Engineering?.......................4

1.2 Scope.......................................................7

1.3 Terminology.................................................8

2.0 Background....................................................11

2.1 Context of Internet Traffic Engineering....................12

2.2 Network Context............................................13

2.3 Problem Context............................................14

2.3.1 Congestion and its Ramifications......................16

2.4 Solution Context...........................................16

2.4.1 Combating the Congestion Problem......................18

2.5 Implementation and Operational Context.....................21

3.0 Traffic Engineering Process Model.............................21

3.1 Components of the Traffic Engineering Process Model........23

3.2 Measurement................................................23

3.3 Modeling, Analysis, and Simulation.........................24

3.4 Optimization...............................................25

4.0 Historical Review and Recent Developments.....................26

4.1 Traffic Engineering in Classical Telephone Networks........26

4.2 Evolution of Traffic Engineering in the Internet...........28

4.2.1 Adaptive Routing in ARPANET...........................28

4.2.2 Dynamic Routing in the Internet.......................29

4.2.3 ToS Routing...........................................30

4.2.4 Equal Cost Multi-Path.................................30

4.2.5 Nimrod................................................31

4.3 Overlay Model..............................................31

4.4 Constraint-Based Routing...................................32

4.5 Overview of Other IETF Projects Related to Traffic

Engineering................................................32

4.5.1 Integrated Services...................................32

4.5.2 RSVP..................................................33

4.5.3 Differentiated Services...............................34

4.5.4 MPLS..................................................35

4.5.5 IP Performance Metrics................................36

4.5.6 Flow Measurement......................................37

4.5.7 Endpoint Congestion Management........................37

4.6 Overview of ITU Activities Related to Traffic

Engineering................................................38

4.7 Content Distribution.......................................39

5.0 Taxonomy of Traffic Engineering Systems.......................40

5.1 Time-Dependent Versus State-Dependent......................40

5.2 Offline Versus Online......................................41

5.3 Centralized Versus Distributed.............................42

5.4 Local Versus Global........................................42

5.5 Prescriptive Versus Descriptive............................42

5.6 Open-Loop Versus Closed-Loop...............................43

5.7 Tactical vs Strategic......................................43

6.0 Recommendations for Internet Traffic Engineering..............43

6.1 Generic Non-functional Recommendations.....................44

6.2 Routing Recommendations....................................46

6.3 Traffic Mapping Recommendations............................48

6.4 Measurement Recommendations................................49

6.5 Network Survivability......................................50

6.5.1 Survivability in MPLS Based Networks..................52

6.5.2 Protection Option.....................................53

6.6 Traffic Engineering in Diffserv Environments...............54

6.7 Network Controllability....................................56

7.0 Inter-Domain Considerations...................................57

8.0 Overview of Contemporary TE Practices in Operational

IP Networks...................................................59

9.0 Conclusion....................................................63

10.0 Security Considerations......................................63

11.0 Acknowledgments..............................................63

12.0 References...................................................64

13.0 Authors' Addresses...........................................70

14.0 Full Copyright Statement.....................................71

1.0 Introduction

This memo describes the principles of Internet traffic engineering.

The objective of the document is to articulate the general issues and

principles for Internet traffic engineering; and where appropriate to

provide recommendations, guidelines, and options for the development

of online and offline Internet traffic engineering capabilities and

support systems.

This document can aid service providers in devising and implementing

traffic engineering solutions for their networks. Networking

hardware and software vendors will also find this document helpful in

the development of mechanisms and support systems for the Internet

environment that support the traffic engineering function.

This document provides a terminology for describing and understanding

common Internet traffic engineering concepts. This document also

provides a taxonomy of known traffic engineering styles. In this

context, a traffic engineering style abstracts important ASPects from

a traffic engineering methodology. Traffic engineering styles can be

viewed in different ways depending upon the specific context in which

they are used and the specific purpose which they serve. The

combination of styles and views results in a natural taxonomy of

traffic engineering systems.

Even though Internet traffic engineering is most effective when

applied end-to-end, the initial focus of this document document is

intra-domain traffic engineering (that is, traffic engineering within

a given autonomous system). However, because a preponderance of

Internet traffic tends to be inter-domain (originating in one

autonomous system and terminating in another), this document provides

an overview of aspects pertaining to inter-domain traffic

engineering.

The key Words "MUST", "MUST NOT", "REQUIRED", "SHALL", "SHALL NOT",

"SHOULD", "SHOULD NOT", "RECOMMENDED", "MAY", and "OPTIONAL" in this

document are to be interpreted as described in RFC2119.

1.1. What is Internet Traffic Engineering?

Internet traffic engineering is defined as that aspect of Internet

network engineering dealing with the issue of performance evaluation

and performance optimization of operational IP networks. Traffic

Engineering encompasses the application of technology and scientific

principles to the measurement, characterization, modeling, and

control of Internet traffic [RFC-2702, AWD2].

Enhancing the performance of an operational network, at both the

traffic and resource levels, are major objectives of Internet traffic

engineering. This is accomplished by addressing traffic oriented

performance requirements, while utilizing network resources

economically and reliably. Traffic oriented performance measures

include delay, delay variation, packet loss, and throughput.

An important objective of Internet traffic engineering is to

facilitate reliable network operations [RFC-2702]. Reliable network

operations can be facilitated by providing mechanisms that enhance

network integrity and by embracing policies emphasizing network

survivability. This results in a minimization of the vulnerability

of the network to service outages arising from errors, faults, and

failures occurring within the infrastructure.

The Internet exists in order to transfer information from source

nodes to destination nodes. Accordingly, one of the most significant

functions performed by the Internet is the routing of traffic from

ingress nodes to egress nodes. Therefore, one of the most

distinctive functions performed by Internet traffic engineering is

the control and optimization of the routing function, to steer

traffic through the network in the most effective way.

Ultimately, it is the performance of the network as seen by end users

of network services that is truly paramount. This crucial point

should be considered throughout the development of traffic

engineering mechanisms and policies. The characteristics visible to

end users are the emergent properties of the network, which are the

characteristics of the network when viewed as a whole. A central

goal of the service provider, therefore, is to enhance the emergent

properties of the network while taking economic considerations into

account.

The importance of the above observation regarding the emergent

properties of networks is that special care must be taken when

choosing network performance measures to optimize. Optimizing the

wrong measures may achieve certain local objectives, but may have

disastrous consequences on the emergent properties of the network and

thereby on the quality of service perceived by end-users of network

services.

A suBTle, but practical advantage of the systematic application of

traffic engineering concepts to operational networks is that it helps

to identify and structure goals and priorities in terms of enhancing

the quality of service delivered to end-users of network services.

The application of traffic engineering concepts also aids in the

measurement and analysis of the achievement of these goals.

The optimization aspects of traffic engineering can be achieved

through capacity management and traffic management. As used in this

document, capacity management includes capacity planning, routing

control, and resource management. Network resources of particular

interest include link bandwidth, buffer space, and computational

resources. Likewise, as used in this document, traffic management

includes (1) nodal traffic control functions such as traffic

conditioning, queue management, scheduling, and (2) other functions

that regulate traffic flow through the network or that arbitrate

Access to network resources between different packets or between

different traffic streams.

The optimization objectives of Internet traffic engineering should be

viewed as a continual and iterative process of network performance

improvement and not simply as a one time goal. Traffic engineering

also demands continual development of new technologies and new

methodologies for network performance enhancement.

The optimization objectives of Internet traffic engineering may

change over time as new requirements are imposed, as new technologies

emerge, or as new insights are brought to bear on the underlying

problems. Moreover, different networks may have different

optimization objectives, depending upon their business models,

capabilities, and operating constraints. The optimization aspects of

traffic engineering are ultimately concerned with network control

regardless of the specific optimization goals in any particular

environment.

Thus, the optimization aspects of traffic engineering can be viewed

from a control perspective. The aspect of control within the

Internet traffic engineering arena can be pro-active and/or reactive.

In the pro-active case, the traffic engineering control system takes

preventive action to obviate predicted unfavorable future network

states. It may also take perfective action to induce a more

desirable state in the future. In the reactive case, the control

system responds correctively and perhaps adaptively to events that

have already transpired in the network.

The control dimension of Internet traffic engineering responds at

multiple levels of temporal resolution to network events. Certain

aspects of capacity management, such as capacity planning, respond at

very coarse temporal levels, ranging from days to possibly years.

The introduction of automatically switched optical transport networks

(e.g., based on the Multi-protocol Lambda Switching concepts) could

significantly reduce the lifecycle for capacity planning by

eXPediting provisioning of optical bandwidth. Routing control

functions operate at intermediate levels of temporal resolution,

ranging from milliseconds to days. Finally, the packet level

processing functions (e.g., rate shaping, queue management, and

scheduling) operate at very fine levels of temporal resolution,

ranging from picoseconds to milliseconds while responding to the

real-time statistical behavior of traffic. The subsystems of

Internet traffic engineering control include: capacity augmentation,

routing control, traffic control, and resource control (including

control of service policies at network elements). When capacity is

to be augmented for tactical purposes, it may be desirable to devise

a deployment plan that expedites bandwidth provisioning while

minimizing installation costs.

Inputs into the traffic engineering control system include network

state variables, policy variables, and decision variables.

One major challenge of Internet traffic engineering is the

realization of automated control capabilities that adapt quickly and

cost effectively to significant changes in a network's state, while

still maintaining stability.

Another critical dimension of Internet traffic engineering is network

performance evaluation, which is important for assessing the

effectiveness of traffic engineering methods, and for monitoring and

verifying compliance with network performance goals. Results from

performance evaluation can be used to identify existing problems,

guide network re-optimization, and aid in the prediction of potential

future problems.

Performance evaluation can be achieved in many different ways. The

most notable techniques include analytical methods, simulation, and

empirical methods based on measurements. When analytical methods or

simulation are used, network nodes and links can be modeled to

capture relevant operational features such as topology, bandwidth,

buffer space, and nodal service policies (link scheduling, packet

prioritization, buffer management, etc.). Analytical traffic models

can be used to depict dynamic and behavioral traffic characteristics,

such as burstiness, statistical distributions, and dependence.

Performance evaluation can be quite complicated in practical network

contexts. A number of techniques can be used to simplify the

analysis, such as abstraction, decomposition, and approximation. For

example, simplifying concepts such as effective bandwidth and

effective buffer [Elwalid] may be used to approximate nodal behaviors

at the packet level and simplify the analysis at the connection

level. Network analysis techniques using, for example, queuing

models and approximation schemes based on asymptotic and

decomposition techniques can render the analysis even more tractable.

In particular, an emerging set of concepts known as network calculus

[CRUZ] based on deterministic bounds may simplify network analysis

relative to classical stochastic techniques. When using analytical

techniques, care should be taken to ensure that the models faithfully

reflect the relevant operational characteristics of the modeled

network entities.

Simulation can be used to evaluate network performance or to verify

and validate analytical approximations. Simulation can, however, be

computationally costly and may not always provide sufficient

insights. An appropriate approach to a given network performance

evaluation problem may involve a hybrid combination of analytical

techniques, simulation, and empirical methods.

As a general rule, traffic engineering concepts and mechanisms must

be sufficiently specific and well defined to address known

requirements, but simultaneously flexible and extensible to

accommodate unforeseen future demands.

1.2. Scope

The scope of this document is intra-domain traffic engineering; that

is, traffic engineering within a given autonomous system in the

Internet. This document will discuss concepts pertaining to intra-

domain traffic control, including such issues as routing control,

micro and macro resource allocation, and the control coordination

problems that arise consequently.

This document will describe and characterize techniques already in

use or in advanced development for Internet traffic engineering. The

way these techniques fit together will be discussed and scenarios in

which they are useful will be identified.

While this document considers various intra-domain traffic

engineering approaches, it focuses more on traffic engineering with

MPLS. Traffic engineering based upon manipulation of IGP metrics is

not addressed in detail. This topic may be addressed by other

working group document(s).

Although the emphasis is on intra-domain traffic engineering, in

Section 7.0, an overview of the high level considerations pertaining

to inter-domain traffic engineering will be provided. Inter-domain

Internet traffic engineering is crucial to the performance

enhancement of the global Internet infrastructure.

Whenever possible, relevant requirements from existing IETF documents

and other sources will be incorporated by reference.

1.3 Terminology

This subsection provides terminology which is useful for Internet

traffic engineering. The definitions presented apply to this

document. These terms may have other meanings elsewhere.

- Baseline analysis:

A study conducted to serve as a baseline for comparison to

the actual behavior of the network.

- Busy hour:

A one hour period within a specified interval of time

(typically 24 hours) in which the traffic load in a network

or sub-network is greatest.

- Bottleneck:

A network element whose input traffic rate tends to be

greater than its output rate.

- Congestion:

A state of a network resource in which the traffic incident

on the resource exceeds its output capacity over an interval

of time.

- Congestion avoidance:

An approach to congestion management that attempts to

obviate the occurrence of congestion.

- Congestion control:

An approach to congestion management that attempts to remedy

congestion problems that have already occurred.

- Constraint-based routing:

A class of routing protocols that take specified traffic

attributes, network constraints, and policy constraints into

account when making routing decisions. Constraint-based

routing is applicable to traffic aggregates as well as

flows. It is a generalization of QoS routing.

- Demand side congestion management:

A congestion management scheme that addresses congestion

problems by regulating or conditioning offered load.

- Effective bandwidth:

The minimum amount of bandwidth that can be assigned to a

flow or traffic aggregate in order to deliver 'acceptable

service quality' to the flow or traffic aggregate.

- Egress traffic:

Traffic exiting a network or network element.

- Hot-spot:

A network element or subsystem which is in a state of

congestion.

- Ingress traffic:

Traffic entering a network or network element.

- Inter-domain traffic:

Traffic that originates in one Autonomous system and

terminates in another.

- Loss network:

A network that does not provide adequate buffering for

traffic, so that traffic entering a busy resource within the

network will be dropped rather than queued.

- Metric:

A parameter defined in terms of standard units of

measurement.

- Measurement Methodology:

A repeatable measurement technique used to derive one or

more metrics of interest.

- Network Survivability:

The capability to provide a prescribed level of QoS for

existing services after a given number of failures occur

within the network.

- Offline traffic engineering:

A traffic engineering system that exists outside of the

network.

- Online traffic engineering:

A traffic engineering system that exists within the network,

typically implemented on or as adjuncts to operational

network elements.

- Performance measures:

Metrics that provide quantitative or qualitative measures of

the performance of systems or subsystems of interest.

- Performance management:

A systematic approach to improving effectiveness in the

accomplishment of specific networking goals related to

performance improvement.

- Performance Metric:

A performance parameter defined in terms of standard units

of measurement.

- Provisioning:

The process of assigning or configuring network resources to

meet certain requests.

- QoS routing:

Class of routing systems that selects paths to be used by a

flow based on the QoS requirements of the flow.

- Service Level Agreement:

A contract between a provider and a customer that guarantees

specific levels of performance and reliability at a certain

cost.

- Stability:

An operational state in which a network does not oscillate

in a disruptive manner from one mode to another mode.

- Supply side congestion management:

A congestion management scheme that provisions additional

network resources to address existing and/or anticipated

congestion problems.

- Transit traffic:

Traffic whose origin and destination are both outside of the

network under consideration.

- Traffic characteristic:

A description of the temporal behavior or a description of

the attributes of a given traffic flow or traffic aggregate.

- Traffic engineering system:

A collection of objects, mechanisms, and protocols that are

used conjunctively to accomplish traffic engineering

objectives.

- Traffic flow:

A stream of packets between two end-points that can be

characterized in a certain way. A micro-flow has a more

specific definition: A micro-flow is a stream of packets

with the same source and destination addresses, source and

destination ports, and protocol ID.

- Traffic intensity:

A measure of traffic loading with respect to a resource

capacity over a specified period of time. In classical

telephony systems, traffic intensity is measured in units of

Erlang.

- Traffic matrix:

A representation of the traffic demand between a set of

origin and destination abstract nodes. An abstract node can

consist of one or more network elements.

- Traffic monitoring:

The process of observing traffic characteristics at a given

point in a network and collecting the traffic information

for analysis and further action.

- Traffic trunk:

An aggregation of traffic flows belonging to the same class

which are forwarded through a common path. A traffic trunk

may be characterized by an ingress and egress node, and a

set of attributes which determine its behavioral

characteristics and requirements from the network.

2.0 Background

The Internet has quickly evolved into a very critical communications

infrastructure, supporting significant economic, educational, and

social activities. Simultaneously, the delivery of Internet

communications services has become very competitive and end-users are

demanding very high quality service from their service providers.

Consequently, performance optimization of large scale IP networks,

especially public Internet backbones, have become an important

problem. Network performance requirements are multi-dimensional,

complex, and sometimes contradictory; making the traffic engineering

problem very challenging.

The network must convey IP packets from ingress nodes to egress nodes

efficiently, expeditiously, and economically. Furthermore, in a

multiclass service environment (e.g., Diffserv capable networks), the

resource sharing parameters of the network must be appropriately

determined and configured according to prevailing policies and

service models to resolve resource contention issues arising from

mutual interference between packets traversing through the network.

Thus, consideration must be given to resolving competition for

network resources between traffic streams belonging to the same

service class (intra-class contention resolution) and traffic streams

belonging to different classes (inter-class contention resolution).

2.1 Context of Internet Traffic Engineering

The context of Internet traffic engineering pertains to the scenarios

where traffic engineering is used. A traffic engineering methodology

establishes appropriate rules to resolve traffic performance issues

occurring in a specific context. The context of Internet traffic

engineering includes:

(1) A network context defining the universe of discourse, and in

particular the situations in which the traffic engineering

problems occur. The network context includes network

structure, network policies, network characteristics,

network constraints, network quality attributes, and network

optimization criteria.

(2) A problem context defining the general and concrete issues

that traffic engineering addresses. The problem context

includes identification, abstraction of relevant features,

representation, formulation, specification of the

requirements on the solution space, and specification of the

desirable features of acceptable solutions.

(3) A solution context suggesting how to address the issues

identified by the problem context. The solution context

includes analysis, evaluation of alternatives, prescription,

and resolution.

(4) An implementation and operational context in which the

solutions are methodologically instantiated. The

implementation and operational context includes planning,

organization, and execution.

The context of Internet traffic engineering and the different problem

scenarios are discussed in the following subsections.

2.2 Network Context

IP networks range in size from small clusters of routers situated

within a given location, to thousands of interconnected routers,

switches, and other components distributed all over the world.

Conceptually, at the most basic level of abstraction, an IP network

can be represented as a distributed dynamical system consisting of:

(1) a set of interconnected resources which provide transport

services for IP traffic subject to certain constraints, (2) a demand

system representing the offered load to be transported through the

network, and (3) a response system consisting of network processes,

protocols, and related mechanisms which facilitate the movement of

traffic through the network [see also AWD2].

The network elements and resources may have specific characteristics

restricting the manner in which the demand is handled. Additionally,

network resources may be equipped with traffic control mechanisms

superintending the way in which the demand is serviced. Traffic

control mechanisms may, for example, be used to control various

packet processing activities within a given resource, arbitrate

contention for access to the resource by different packets, and

regulate traffic behavior through the resource. A configuration

management and provisioning system may allow the settings of the

traffic control mechanisms to be manipulated by external or internal

entities in order to exercise control over the way in which the

network elements respond to internal and external stimuli.

The details of how the network provides transport services for

packets are specified in the policies of the network administrators

and are installed through network configuration management and policy

based provisioning systems. Generally, the types of services

provided by the network also depends upon the technology and

characteristics of the network elements and protocols, the prevailing

service and utility models, and the ability of the network

administrators to translate policies into network configurations.

Contemporary Internet networks have three significant

characteristics: (1) they provide real-time services, (2) they have

become mission critical, and (3) their operating environments are

very dynamic. The dynamic characteristics of IP networks can be

attributed in part to fluctuations in demand, to the interaction

between various network protocols and processes, to the rapid

evolution of the infrastructure which demands the constant inclusion

of new technologies and new network elements, and to transient and

persistent impairments which occur within the system.

Packets contend for the use of network resources as they are conveyed

through the network. A network resource is considered to be

congested if the arrival rate of packets exceed the output capacity

of the resource over an interval of time. Congestion may result in

some of the arrival packets being delayed or even dropped.

Congestion increases transit delays, delay variation, packet loss,

and reduces the predictability of network services. Clearly,

congestion is a highly undesirable phenomenon.

Combating congestion at a reasonable cost is a major objective of

Internet traffic engineering.

Efficient sharing of network resources by multiple traffic streams is

a basic economic premise for packet switched networks in general and

for the Internet in particular. A fundamental challenge in network

operation, especially in a large scale public IP network, is to

increase the efficiency of resource utilization while minimizing the

possibility of congestion.

Increasingly, the Internet will have to function in the presence of

different classes of traffic with different service requirements.

The advent of Differentiated Services [RFC-2475] makes this

requirement particularly acute. Thus, packets may be grouped into

behavior aggregates such that each behavior aggregate may have a

common set of behavioral characteristics or a common set of delivery

requirements. In practice, the delivery requirements of a specific

set of packets may be specified explicitly or implicitly. Two of the

most important traffic delivery requirements are capacity constraints

and QoS constraints.

Capacity constraints can be expressed statistically as peak rates,

mean rates, burst sizes, or as some deterministic notion of effective

bandwidth. QoS requirements can be expressed in terms of (1)

integrity constraints such as packet loss and (2) in terms of

temporal constraints such as timing restrictions for the delivery of

each packet (delay) and timing restrictions for the delivery of

consecutive packets belonging to the same traffic stream (delay

variation).

2.3 Problem Context

Fundamental problems exist in association with the operation of a

network described by the simple model of the previous subsection.

This subsection reviews the problem context in relation to the

traffic engineering function.

The identification, abstraction, representation, and measurement of

network features relevant to traffic engineering is a significant

issue.

One particularly important class of problems concerns how to

explicitly formulate the problems that traffic engineering attempts

to solve, how to identify the requirements on the solution space, how

to specify the desirable features of good solutions, how to actually

solve the problems, and how to measure and characterize the

effectiveness of the solutions.

Another class of problems concerns how to measure and estimate

relevant network state parameters. Effective traffic engineering

relies on a good estimate of the offered traffic load as well as a

view of the underlying topology and associated resource constraints.

A network-wide view of the topology is also a must for offline

planning.

Still another class of problems concerns how to characterize the

state of the network and how to evaluate its performance under a

variety of scenarios. The performance evaluation problem is two-

fold. One aspect of this problem relates to the evaluation of the

system level performance of the network. The other aspect relates to

the evaluation of the resource level performance, which restricts

attention to the performance analysis of individual network

resources. In this memo, we refer to the system level

characteristics of the network as the "macro-states" and the resource

level characteristics as the "micro-states." The system level

characteristics are also known as the emergent properties of the

network as noted earlier. Correspondingly, we shall refer to the

traffic engineering schemes dealing with network performance

optimization at the systems level as "macro-TE" and the schemes that

optimize at the individual resource level as "micro-TE." Under

certain circumstances, the system level performance can be derived

from the resource level performance using appropriate rules of

composition, depending upon the particular performance measures of

interest.

Another fundamental class of problems concerns how to effectively

optimize network performance. Performance optimization may entail

translating solutions to specific traffic engineering problems into

network configurations. Optimization may also entail some degree of

resource management control, routing control, and/or capacity

augmentation.

As noted previously, congestion is an undesirable phenomena in

operational networks. Therefore, the next subsection addresses the

issue of congestion and its ramifications within the problem context

of Internet traffic engineering.

2.3.1 Congestion and its Ramifications

Congestion is one of the most significant problems in an operational

IP context. A network element is said to be congested if it

experiences sustained overload over an interval of time. Congestion

almost always results in degradation of service quality to end users.

Congestion control schemes can include demand side policies and

supply side policies. Demand side policies may restrict access to

congested resources and/or dynamically regulate the demand to

alleviate the overload situation. Supply side policies may expand or

augment network capacity to better accommodate offered traffic.

Supply side policies may also re-allocate network resources by

redistributing traffic over the infrastructure. Traffic

redistribution and resource re-allocation serve to increase the

'effective capacity' seen by the demand.

The emphasis of this memo is primarily on congestion management

schemes falling within the scope of the network, rather than on

congestion management systems dependent upon sensitivity and

adaptivity from end-systems. That is, the aspects that are

considered in this memo with respect to congestion management are

those solutions that can be provided by control entities operating on

the network and by the actions of network administrators and network

operations systems.

2.4 Solution Context

The solution context for Internet traffic engineering involves

analysis, evaluation of alternatives, and choice between alternative

courses of action. Generally the solution context is predicated on

making reasonable inferences about the current or future state of the

network, and subsequently making appropriate decisions that may

involve a preference between alternative sets of action. More

specifically, the solution context demands reasonable estimates of

traffic workload, characterization of network state, deriving

solutions to traffic engineering problems which may be implicitly or

explicitly formulated, and possibly instantiating a set of control

actions. Control actions may involve the manipulation of parameters

associated with routing, control over tactical capacity acquisition,

and control over the traffic management functions.

The following list of instruments may be applicable to the solution

context of Internet traffic engineering.

(1) A set of policies, objectives, and requirements (which may

be context dependent) for network performance evaluation and

performance optimization.

(2) A collection of online and possibly offline tools and

mechanisms for measurement, characterization, modeling, and

control of Internet traffic and control over the placement

and allocation of network resources, as well as control over

the mapping or distribution of traffic onto the

infrastructure.

(3) A set of constraints on the operating environment, the

network protocols, and the traffic engineering system

itself.

(4) A set of quantitative and qualitative techniques and

methodologies for abstracting, formulating, and solving

traffic engineering problems.

(5) A set of administrative control parameters which may be

manipulated through a Configuration Management (CM) system.

The CM system itself may include a configuration control

subsystem, a configuration repository, a configuration

accounting subsystem, and a configuration auditing

subsystem.

(6) A set of guidelines for network performance evaluation,

performance optimization, and performance improvement.

Derivation of traffic characteristics through measurement and/or

estimation is very useful within the realm of the solution space for

traffic engineering. Traffic estimates can be derived from customer

subscription information, traffic projections, traffic models, and

from actual empirical measurements. The empirical measurements may

be performed at the traffic aggregate level or at the flow level in

order to derive traffic statistics at various levels of detail.

Measurements at the flow level or on small traffic aggregates may be

performed at edge nodes, where traffic enters and leaves the network.

Measurements at large traffic aggregate levels may be performed

within the core of the network where potentially numerous traffic

flows may be in transit concurrently.

To conduct performance studies and to support planning of existing

and future networks, a routing analysis may be performed to determine

the path(s) the routing protocols will choose for various traffic

demands, and to ascertain the utilization of network resources as

traffic is routed through the network. The routing analysis should

capture the selection of paths through the network, the assignment of

traffic across multiple feasible routes, and the multiplexing of IP

traffic over traffic trunks (if such constructs exists) and over the

underlying network infrastructure. A network topology model is a

necessity for routing analysis. A network topology model may be

extracted from network architecture documents, from network designs,

from information contained in router configuration files, from

routing databases, from routing tables, or from automated tools that

discover and depict network topology information. Topology

information may also be derived from servers that monitor network

state, and from servers that perform provisioning functions.

Routing in operational IP networks can be administratively controlled

at various levels of abstraction including the manipulation of BGP

attributes and manipulation of IGP metrics. For path oriented

technologies such as MPLS, routing can be further controlled by the

manipulation of relevant traffic engineering parameters, resource

parameters, and administrative policy constraints. Within the

context of MPLS, the path of an explicit label switched path (LSP)

can be computed and established in various ways including: (1)

manually, (2) automatically online using constraint-based routing

processes implemented on label switching routers, and (3)

automatically offline using constraint-based routing entities

implemented on external traffic engineering support systems.

2.4.1 Combating the Congestion Problem

Minimizing congestion is a significant aspect of Internet traffic

engineering. This subsection gives an overview of the general

approaches that have been used or proposed to combat congestion

problems.

Congestion management policies can be categorized based upon the

following criteria (see e.g., [YARE95] for a more detailed taxonomy

of congestion control schemes): (1) Response time scale which can be

characterized as long, medium, or short; (2) reactive versus

preventive which relates to congestion control and congestion

avoidance; and (3) supply side versus demand side congestion

management schemes. These aspects are discussed in the following

paragraphs.

(1) Congestion Management based on Response Time Scales

- Long (weeks to months): Capacity planning works over a relatively

long time scale to expand network capacity based on estimates or

forecasts of future traffic demand and traffic distribution. Since

router and link provisioning take time and are generally expensive,

these upgrades are typically carried out in the weeks-to-months or

even years time scale.

- Medium (minutes to days): Several control policies fall within the

medium time scale category. Examples include: (1) Adjusting IGP

and/or BGP parameters to route traffic away or towards certain

segments of the network; (2) Setting up and/or adjusting some

explicitly routed label switched paths (ER-LSPs) in MPLS networks to

route some traffic trunks away from possibly congested resources or

towards possibly more favorable routes; (3) re-configuring the

logical topology of the network to make it correlate more closely

with the spatial traffic distribution using for example some

underlying path-oriented technology such as MPLS LSPs, ATM PVCs, or

optical channel trails. Many of these adaptive medium time scale

response schemes rely on a measurement system that monitors changes

in traffic distribution, traffic shifts, and network resource

utilization and subsequently provides feedback to the online and/or

offline traffic engineering mechanisms and tools which employ this

feedback information to trigger certain control actions to occur

within the network. The traffic engineering mechanisms and tools can

be implemented in a distributed fashion or in a centralized fashion,

and may have a hierarchical structure or a flat structure. The

comparative merits of distributed and centralized control structures

for networks are well known. A centralized scheme may have global

visibility into the network state and may produce potentially more

optimal solutions. However, centralized schemes are prone to single

points of failure and may not scale as well as distributed schemes.

Moreover, the information utilized by a centralized scheme may be

stale and may not reflect the actual state of the network. It is not

an objective of this memo to make a recommendation between

distributed and centralized schemes. This is a choice that network

administrators must make based on their specific needs.

- Short (picoseconds to minutes): This category includes packet level

processing functions and events on the order of several round trip

times. It includes router mechanisms such as passive and active

buffer management. These mechanisms are used to control congestion

and/or signal congestion to end systems so that they can adaptively

regulate the rate at which traffic is injected into the network. One

of the most popular active queue management schemes, especially for

TCP traffic, is Random Early Detection (RED) [FLJA93], which supports

congestion avoidance by controlling the average queue size. During

congestion (but before the queue is filled), the RED scheme chooses

arriving packets to "mark" according to a probabilistic algorithm

which takes into account the average queue size. For a router that

does not utilize explicit congestion notification (ECN) see e.g.,

[FLOY94], the marked packets can simply be dropped to signal the

inception of congestion to end systems. On the other hand, if the

router supports ECN, then it can set the ECN field in the packet

header. Several variations of RED have been proposed to support

different drop precedence levels in multi-class environments [RFC-

2597], e.g., RED with In and Out (RIO) and Weighted RED. There is

general consensus that RED provides congestion avoidance performance

which is not worse than traditional Tail-Drop (TD) queue management

(drop arriving packets only when the queue is full). Importantly,

however, RED reduces the possibility of global synchronization and

improves fairness among different TCP sessions. However, RED by

itself can not prevent congestion and unfairness caused by sources

unresponsive to RED, e.g., UDP traffic and some misbehaved greedy

connections. Other schemes have been proposed to improve the

performance and fairness in the presence of unresponsive traffic.

Some of these schemes were proposed as theoretical frameworks and are

typically not available in existing commercial products. Two such

schemes are Longest Queue Drop (LQD) and Dynamic Soft Partitioning

with Random Drop (RND) [SLDC98].

(2) Congestion Management: Reactive versus Preventive Schemes

- Reactive: reactive (recovery) congestion management policies react

to existing congestion problems to improve it. All the policies

described in the long and medium time scales above can be categorized

as being reactive especially if the policies are based on monitoring

and identifying existing congestion problems, and on the initiation

of relevant actions to ease a situation.

- Preventive: preventive (predictive/avoidance) policies take

proactive action to prevent congestion based on estimates and

predictions of future potential congestion problems. Some of the

policies described in the long and medium time scales fall into this

category. They do not necessarily respond immediately to existing

congestion problems. Instead forecasts of traffic demand and

workload distribution are considered and action may be taken to

prevent potential congestion problems in the future. The schemes

described in the short time scale (e.g., RED and its variations, ECN,

LQD, and RND) are also used for congestion avoidance since dropping

or marking packets before queues actually overflow would trigger

corresponding TCP sources to slow down.

(3) Congestion Management: Supply Side versus Demand Side Schemes

- Supply side: supply side congestion management policies increase

the effective capacity available to traffic in order to control or

obviate congestion. This can be accomplished by augmenting capacity.

Another way to accomplish this is to minimize congestion by having a

relatively balanced distribution of traffic over the network. For

example, capacity planning should aim to provide a physical topology

and associated link bandwidths that match estimated traffic workload

and traffic distribution based on forecasting (subject to budgetary

and other constraints). However, if actual traffic distribution does

not match the topology derived from capacity panning (due to

forecasting errors or facility constraints for example), then the

traffic can be mapped onto the existing topology using routing

control mechanisms, using path oriented technologies (e.g., MPLS LSPs

and optical channel trails) to modify the logical topology, or by

using some other load redistribution mechanisms.

- Demand side: demand side congestion management policies control or

regulate the offered traffic to alleviate congestion problems. For

example, some of the short time scale mechanisms described earlier

(such as RED and its variations, ECN, LQD, and RND) as well as

policing and rate shaping mechanisms attempt to regulate the offered

load in various ways. Tariffs may also be applied as a demand side

instrument. To date, however, tariffs have not been used as a means

of demand side congestion management within the Internet.

In summary, a variety of mechanisms can be used to address congestion

problems in IP networks. These mechanisms may operate at multiple

time-scales.

2.5 Implementation and Operational Context

The operational context of Internet traffic engineering is

characterized by constant change which occur at multiple levels of

abstraction. The implementation context demands effective planning,

organization, and execution. The planning aspects may involve

determining prior sets of actions to achieve desired objectives.

Organizing involves arranging and assigning responsibility to the

various components of the traffic engineering system and coordinating

the activities to accomplish the desired TE objectives. Execution

involves measuring and applying corrective or perfective actions to

attain and maintain desired TE goals.

3.0 Traffic Engineering Process Model(s)

This section describes a generic process model that captures the high

level practical aspects of Internet traffic engineering in an

operational context. The process model is described as a sequence of

actions that a traffic engineer, or more generally a traffic

engineering system, must perform to optimize the performance of an

operational network (see also [RFC-2702, AWD2]). The process model

described here represents the broad activities common to most traffic

engineering methodologies although the details regarding how traffic

engineering is executed may differ from network to network. This

process model may be enacted explicitly or implicitly, by an

automaton and/or by a human.

The traffic engineering process model is iterative [AWD2]. The four

phases of the process model described below are repeated continually.

The first phase of the TE process model is to define the relevant

control policies that govern the operation of the network. These

policies may depend upon many factors including the prevailing

business model, the network cost structure, the operating

constraints, the utility model, and optimization criteria.

The second phase of the process model is a feedback mechanism

involving the acquisition of measurement data from the operational

network. If empirical data is not readily available from the

network, then synthetic workloads may be used instead which reflect

either the prevailing or the expected workload of the network.

Synthetic workloads may be derived by estimation or extrapolation

using prior empirical data. Their derivation may also be obtained

using mathematical models of traffic characteristics or other means.

The third phase of the process model is to analyze the network state

and to characterize traffic workload. Performance analysis may be

proactive and/or reactive. Proactive performance analysis identifies

potential problems that do not exist, but could manifest in the

future. Reactive performance analysis identifies existing problems,

determines their cause through diagnosis, and evaluates alternative

approaches to remedy the problem, if necessary. A number of

quantitative and qualitative techniques may be used in the analysis

process, including modeling based analysis and simulation. The

analysis phase of the process model may involve investigating the

concentration and distribution of traffic across the network or

relevant subsets of the network, identifying the characteristics of

the offered traffic workload, identifying existing or potential

bottlenecks, and identifying network pathologies such as ineffective

link placement, single points of failures, etc. Network pathologies

may result from many factors including inferior network architecture,

inferior network design, and configuration problems. A traffic

matrix may be constructed as part of the analysis process. Network

analysis may also be descriptive or prescriptive.

The fourth phase of the TE process model is the performance

optimization of the network. The performance optimization phase

involves a decision process which selects and implements a set of

actions from a set of alternatives. Optimization actions may include

the use of appropriate techniques to either control the offered

traffic or to control the distribution of traffic across the network.

Optimization actions may also involve adding additional links or

increasing link capacity, deploying additional hardware such as

routers and switches, systematically adjusting parameters associated

with routing such as IGP metrics and BGP attributes, and adjusting

traffic management parameters. Network performance optimization may

also involve starting a network planning process to improve the

network architecture, network design, network capacity, network

technology, and the configuration of network elements to accommodate

current and future growth.

3.1 Components of the Traffic Engineering Process Model

The key components of the traffic engineering process model include a

measurement subsystem, a modeling and analysis subsystem, and an

optimization subsystem. The following subsections examine these

components as they apply to the traffic engineering process model.

3.2 Measurement

Measurement is crucial to the traffic engineering function. The

operational state of a network can be conclusively determined only

through measurement. Measurement is also critical to the

optimization function because it provides feedback data which is used

by traffic engineering control subsystems. This data is used to

adaptively optimize network performance in response to events and

stimuli originating within and outside the network. Measurement is

also needed to determine the quality of network services and to

evaluate the effectiveness of traffic engineering policies.

Experience suggests that measurement is most effective when acquired

and applied systematically.

When developing a measurement system to support the traffic

engineering function in IP networks, the following questions should

be carefully considered: Why is measurement needed in this particular

context? What parameters are to be measured? How should the

measurement be accomplished? Where should the measurement be

performed? When should the measurement be performed? How frequently

should the monitored variables be measured? What level of

measurement accuracy and reliability is desirable? What level of

measurement accuracy and reliability is realistically attainable? To

what extent can the measurement system permissibly interfere with the

monitored network components and variables? What is the acceptable

cost of measurement? The answers to these questions will determine

the measurement tools and methodologies appropriate in any given

traffic engineering context.

It should also be noted that there is a distinction between

measurement and evaluation. Measurement provides raw data concerning

state parameters and variables of monitored network elements.

Evaluation utilizes the raw data to make inferences regarding the

monitored system.

Measurement in support of the TE function can occur at different

levels of abstraction. For example, measurement can be used to

derive packet level characteristics, flow level characteristics, user

or customer level characteristics, traffic aggregate characteristics,

component level characteristics, and network wide characteristics.

3.3 Modeling, Analysis, and Simulation

Modeling and analysis are important aspects of Internet traffic

engineering. Modeling involves constructing an abstract or physical

representation which depicts relevant traffic characteristics and

network attributes.

A network model is an abstract representation of the network which

captures relevant network features, attributes, and characteristics,

such as link and nodal attributes and constraints. A network model

may facilitate analysis and/or simulation which can be used to

predict network performance under various conditions as well as to

guide network expansion plans.

In general, Internet traffic engineering models can be classified as

either structural or behavioral. Structural models focus on the

organization of the network and its components. Behavioral models

focus on the dynamics of the network and the traffic workload.

Modeling for Internet traffic engineering may also be formal or

informal.

Accurate behavioral models for traffic sources are particularly

useful for analysis. Development of behavioral traffic source models

that are consistent with empirical data obtained from operational

networks is a major research topic in Internet traffic engineering.

These source models should also be tractable and amenable to

analysis. The topic of source models for IP traffic is a research

topic and is therefore outside the scope of this document. Its

importance, however, must be emphasized.

Network simulation tools are extremely useful for traffic

engineering. Because of the complexity of realistic quantitative

analysis of network behavior, certain aspects of network performance

studies can only be conducted effectively using simulation. A good

network simulator can be used to mimic and visualize network

characteristics under various conditions in a safe and non-disruptive

manner. For example, a network simulator may be used to depict

congested resources and hot spots, and to provide hints regarding

possible solutions to network performance problems. A good simulator

may also be used to validate the effectiveness of planned solutions

to network issues without the need to tamper with the operational

network, or to commence an expensive network upgrade which may not

achieve the desired objectives. Furthermore, during the process of

network planning, a network simulator may reveal pathologies such as

single points of failure which may require additional redundancy, and

potential bottlenecks and hot spots which may require additional

capacity.

Routing simulators are especially useful in large networks. A

routing simulator may identify planned links which may not actually

be used to route traffic by the existing routing protocols.

Simulators can also be used to conduct scenario based and

perturbation based analysis, as well as sensitivity studies.

Simulation results can be used to initiate appropriate actions in

various ways. For example, an important application of network

simulation tools is to investigate and identify how best to make the

network evolve and grow, in order to accommodate projected future

demands.

3.4 Optimization

Network performance optimization involves resolving network issues by

transforming such issues into concepts that enable a solution,

identification of a solution, and implementation of the solution.

Network performance optimization can be corrective or perfective. In

corrective optimization, the goal is to remedy a problem that has

occurred or that is incipient. In perfective optimization, the goal

is to improve network performance even when explicit problems do not

exist and are not anticipated.

Network performance optimization is a continual process, as noted

previously. Performance optimization iterations may consist of

real-time optimization sub-processes and non-real-time network

planning sub-processes. The difference between real-time

optimization and network planning is primarily in the relative time-

scale in which they operate and in the granularity of actions. One

of the objectives of a real-time optimization sub-process is to

control the mapping and distribution of traffic over the existing

network infrastructure to avoid and/or relieve congestion, to assure

satisfactory service delivery, and to optimize resource utilization.

Real-time optimization is needed because random incidents such as

fiber cuts or shifts in traffic demand will occur irrespective of how

well a network is designed. These incidents can cause congestion and

other problems to manifest in an operational network. Real-time

optimization must solve such problems in small to medium time-scales

ranging from micro-seconds to minutes or hours. Examples of real-

time optimization include queue management, IGP/BGP metric tuning,

and using technologies such as MPLS explicit LSPs to change the paths

of some traffic trunks [XIAO].

One of the functions of the network planning sub-process is to

initiate actions to systematically evolve the architecture,

technology, topology, and capacity of a network. When a problem

exists in the network, real-time optimization should provide an

immediate remedy. Because a prompt response is necessary, the real-

time solution may not be the best possible solution. Network

planning may subsequently be needed to refine the solution and

improve the situation. Network planning is also required to expand

the network to support traffic growth and changes in traffic

distribution over time. As previously noted, a change in the

topology and/or capacity of the network may be the outcome of network

planning.

Clearly, network planning and real-time performance optimization are

mutually complementary activities. A well-planned and designed

network makes real-time optimization easier, while a systematic

approach to real-time network performance optimization allows network

planning to focus on long term issues rather than tactical

considerations. Systematic real-time network performance

optimization also provides valuable inputs and insights toward

network planning.

Stability is an important consideration in real-time network

performance optimization. This aspect will be repeatedly addressed

throughout this memo.

4.0 Historical Review and Recent Developments

This section briefly reviews different traffic engineering approaches

proposed and implemented in telecommunications and computer networks.

The discussion is not intended to be comprehensive. It is primarily

intended to illuminate pre-existing perspectives and prior art

concerning traffic engineering in the Internet and in legacy

telecommunications networks.

4.1 Traffic Engineering in Classical Telephone Networks

This subsection presents a brief overview of traffic engineering in

telephone networks which often relates to the way user traffic is

steered from an originating node to the terminating node. This

subsection presents a brief overview of this topic. A detailed

description of the various routing strategies applied in telephone

networks is included in the book by G. Ash [ASH2].

The early telephone network relied on static hierarchical routing,

whereby routing patterns remained fixed independent of the state of

the network or time of day. The hierarchy was intended to

accommodate overflow traffic, improve network reliability via

alternate routes, and prevent call looping by employing strict

hierarchical rules. The network was typically over-provisioned since

a given fixed route had to be dimensioned so that it could carry user

traffic during a busy hour of any busy day. Hierarchical routing in

the telephony network was found to be too rigid upon the advent of

digital switches and stored program control which were able to manage

more complicated traffic engineering rules.

Dynamic routing was introduced to alleviate the routing inflexibility

in the static hierarchical routing so that the network would operate

more efficiently. This resulted in significant economic gains

[HUSS87]. Dynamic routing typically reduces the overall loss

probability by 10 to 20 percent (compared to static hierarchical

routing). Dynamic routing can also improve network resilience by

recalculating routes on a per-call basis and periodically updating

routes.

There are three main types of dynamic routing in the telephone

network. They are time-dependent routing, state-dependent routing

(SDR), and event dependent routing (EDR).

In time-dependent routing, regular variations in traffic loads (such

as time of day or day of week) are exploited in pre-planned routing

tables. In state-dependent routing, routing tables are updated

online according to the current state of the network (e.g., traffic

demand, utilization, etc.). In event dependent routing, routing

changes are incepted by events (such as call setups encountering

congested or blocked links) whereupon new paths are searched out

using learning models. EDR methods are real-time adaptive, but they

do not require global state information as does SDR. Examples of EDR

schemes include the dynamic alternate routing (DAR) from BT, the

state-and-time dependent routing (STR) from NTT, and the success-to-

the-top (STT) routing from AT&T.

Dynamic non-hierarchical routing (DNHR) is an example of dynamic

routing that was introduced in the AT&T toll network in the 1980's to

respond to time-dependent information such as regular load variations

as a function of time. Time-dependent information in terms of load

may be divided into three time scales: hourly, weekly, and yearly.

Correspondingly, three algorithms are defined to pre-plan the routing

tables. The network design algorithm operates over a year-long

interval while the demand servicing algorithm operates on a weekly

basis to fine tune link sizes and routing tables to correct forecast

errors on the yearly basis. At the smallest time scale, the routing

algorithm is used to make limited adjustments based on daily traffic

variations. Network design and demand servicing are computed using

offline calculations. Typically, the calculations require extensive

searches on possible routes. On the other hand, routing may need

online calculations to handle crankback. DNHR adopts a "two-link"

approach whereby a path can consist of two links at most. The

routing algorithm presents an ordered list of route choices between

an originating switch and a terminating switch. If a call overflows,

a via switch (a tandem exchange between the originating switch and

the terminating switch) would send a crankback signal to the

originating switch. This switch would then select the next route,

and so on, until there are no alternative routes available in which

the call is blocked.

4.2 Evolution of Traffic Engineering in Packet Networks

This subsection reviews related prior work that was intended to

improve the performance of data networks. Indeed, optimization of

the performance of data networks started in the early days of the

ARPANET. Other early commercial networks such as SNA also recognized

the importance of performance optimization and service

differentiation.

In terms of traffic management, the Internet has been a best effort

service environment until recently. In particular, very limited

traffic management capabilities existed in IP networks to provide

differentiated queue management and scheduling services to packets

belonging to different classes.

In terms of routing control, the Internet has employed distributed

protocols for intra-domain routing. These protocols are highly

scalable and resilient. However, they are based on simple algorithms

for path selection which have very limited functionality to allow

flexible control of the path selection process.

In the following subsections, the evolution of practical traffic

engineering mechanisms in IP networks and its predecessors are

reviewed.

4.2.1 Adaptive Routing in the ARPANET

The early ARPANET recognized the importance of adaptive routing where

routing decisions were based on the current state of the network

[MCQ80]. Early minimum delay routing approaches forwarded each

packet to its destination along a path for which the total estimated

transit time was the smallest. Each node maintained a table of

network delays, representing the estimated delay that a packet would

experience along a given path toward its destination. The minimum

delay table was periodically transmitted by a node to its neighbors.

The shortest path, in terms of hop count, was also propagated to give

the connectivity information.

One drawback to this approach is that dynamic link metrics tend to

create "traffic magnets" causing congestion to be shifted from one

location of a network to another location, resulting in oscillation

and network instability.

4.2.2 Dynamic Routing in the Internet

The Internet evolved from the APARNET and adopted dynamic routing

algorithms with distributed control to determine the paths that

packets should take en-route to their destinations. The routing

algorithms are adaptations of shortest path algorithms where costs

are based on link metrics. The link metric can be based on static or

dynamic quantities. The link metric based on static quantities may

be assigned administratively according to local criteria. The link

metric based on dynamic quantities may be a function of a network

congestion measure such as delay or packet loss.

It was apparent early that static link metric assignment was

inadequate because it can easily lead to unfavorable scenarios in

which some links become congested while others remain lightly loaded.

One of the many reasons for the inadequacy of static link metrics is

that link metric assignment was often done without considering the

traffic matrix in the network. Also, the routing protocols did not

take traffic attributes and capacity constraints into account when

making routing decisions. This results in traffic concentration

being localized in subsets of the network infrastructure and

potentially causing congestion. Even if link metrics are assigned in

accordance with the traffic matrix, unbalanced loads in the network

can still occur due to a number factors including:

- Resources may not be deployed in the most optimal locations

from a routing perspective.

- Forecasting errors in traffic volume and/or traffic

distribution.

- Dynamics in traffic matrix due to the temporal nature of

traffic patterns, BGP policy change from peers, etc.

The inadequacy of the legacy Internet interior gateway routing system

is one of the factors motivating the interest in path oriented

technology with explicit routing and constraint-based routing

capability such as MPLS.

4.2.3 ToS Routing

Type-of-Service (ToS) routing involves different routes going to the

same destination with selection dependent upon the ToS field of an IP

packet [RFC-2474]. The ToS classes may be classified as low delay

and high throughput. Each link is associated with multiple link

costs and each link cost is used to compute routes for a particular

ToS. A separate shortest path tree is computed for each ToS. The

shortest path algorithm must be run for each ToS resulting in very

expensive computation. Classical ToS-based routing is now outdated

as the IP header field has been replaced by a Diffserv field.

Effective traffic engineering is difficult to perform in classical

ToS-based routing because each class still relies exclusively on

shortest path routing which results in localization of traffic

concentration within the network.

4.2.4 Equal Cost Multi-Path

Equal Cost Multi-Path (ECMP) is another technique that attempts to

address the deficiency in the Shortest Path First (SPF) interior

gateway routing systems [RFC-2328]. In the classical SPF algorithm,

if two or more shortest paths exist to a given destination, the

algorithm will choose one of them. The algorithm is modified

slightly in ECMP so that if two or more equal cost shortest paths

exist between two nodes, the traffic between the nodes is distributed

among the multiple equal-cost paths. Traffic distribution across the

equal-cost paths is usually performed in one of two ways: (1)

packet-based in a round-robin fashion, or (2) flow-based using

hashing on source and destination IP addresses and possibly other

fields of the IP header. The first approach can easily cause out-

of-order packets while the second approach is dependent upon the

number and distribution of flows. Flow-based load sharing may be

unpredictable in an enterprise network where the number of flows is

relatively small and less heterogeneous (for example, hashing may not

be uniform), but it is generally effective in core public networks

where the number of flows is large and heterogeneous.

In ECMP, link costs are static and bandwidth constraints are not

considered, so ECMP attempts to distribute the traffic as equally as

possible among the equal-cost paths independent of the congestion

status of each path. As a result, given two equal-cost paths, it is

possible that one of the paths will be more congested than the other.

Another drawback of ECMP is that load sharing cannot be achieved on

multiple paths which have non-identical costs.

4.2.5 Nimrod

Nimrod is a routing system developed to provide heterogeneous service

specific routing in the Internet, while taking multiple constraints

into account [RFC-1992]. Essentially, Nimrod is a link state routing

protocol which supports path oriented packet forwarding. It uses the

concept of maps to represent network connectivity and services at

multiple levels of abstraction. Mechanisms are provided to allow

restriction of the distribution of routing information.

Even though Nimrod did not enjoy deployment in the public Internet, a

number of key concepts incorporated into the Nimrod architecture,

such as explicit routing which allows selection of paths at

originating nodes, are beginning to find applications in some recent

constraint-based routing initiatives.

4.3 Overlay Model

In the overlay model, a virtual-circuit network, such as ATM, frame

relay, or WDM, provides virtual-circuit connectivity between routers

that are located at the edges of a virtual-circuit cloud. In this

mode, two routers that are connected through a virtual circuit see a

direct adjacency between themselves independent of the physical route

taken by the virtual circuit through the ATM, frame relay, or WDM

network. Thus, the overlay model essentially decouples the logical

topology that routers see from the physical topology that the ATM,

frame relay, or WDM network manages. The overlay model based on ATM

or frame relay enables a network administrator or an automaton to

employ traffic engineering concepts to perform path optimization by

re-configuring or rearranging the virtual circuits so that a virtual

circuit on a congested or sub-optimal physical link can be re-routed

to a less congested or more optimal one. In the overlay model,

traffic engineering is also employed to establish relationships

between the traffic management parameters (e.g., PCR, SCR, and MBS

for ATM) of the virtual-circuit technology and the actual traffic

that traverses each circuit. These relationships can be established

based upon known or projected traffic profiles, and some other

factors.

The overlay model using IP over ATM requires the management of two

separate networks with different technologies (IP and ATM) resulting

in increased operational complexity and cost. In the fully-meshed

overlay model, each router would peer to every other router in the

network, so that the total number of adjacencies is a quadratic

function of the number of routers. Some of the issues with the

overlay model are discussed in [AWD2].

4.4 Constrained-Based Routing

Constraint-based routing refers to a class of routing systems that

compute routes through a network subject to the satisfaction of a set

of constraints and requirements. In the most general setting,

constraint-based routing may also seek to optimize overall network

performance while minimizing costs.

The constraints and requirements may be imposed by the network itself

or by administrative policies. Constraints may include bandwidth,

hop count, delay, and policy instruments such as resource class

attributes. Constraints may also include domain specific attributes

of certain network technologies and contexts which impose

restrictions on the solution space of the routing function. Path

oriented technologies such as MPLS have made constraint-based routing

feasible and attractive in public IP networks.

The concept of constraint-based routing within the context of MPLS

traffic engineering requirements in IP networks was first defined in

[RFC-2702].

Unlike QoS routing (for example, see [RFC-2386] and [MA]) which

generally addresses the issue of routing individual traffic flows to

satisfy prescribed flow based QoS requirements subject to network

resource availability, constraint-based routing is applicable to

traffic aggregates as well as flows and may be subject to a wide

variety of constraints which may include policy restrictions.

4.5 Overview of Other IETF Projects Related to Traffic Engineering

This subsection reviews a number of IETF activities pertinent to

Internet traffic engineering. These activities are primarily

intended to evolve the IP architecture to support new service

definitions which allow preferential or differentiated treatment to

be accorded to certain types of traffic.

4.5.1 Integrated Services

The IETF Integrated Services working group developed the integrated

services (Intserv) model. This model requires resources, such as

bandwidth and buffers, to be reserved a priori for a given traffic

flow to ensure that the quality of service requested by the traffic

flow is satisfied. The integrated services model includes additional

components beyond those used in the best-effort model such as packet

classifiers, packet schedulers, and admission control. A packet

classifier is used to identify flows that are to receive a certain

level of service. A packet scheduler handles the scheduling of

service to different packet flows to ensure that QoS commitments are

met. Admission control is used to determine whether a router has the

necessary resources to accept a new flow.

Two services have been defined under the Integrated Services model:

guaranteed service [RFC-2212] and controlled-load service [RFC-2211].

The guaranteed service can be used for applications requiring bounded

packet delivery time. For this type of application, data that is

delivered to the application after a pre-defined amount of time has

elapsed is usually considered worthless. Therefore, guaranteed

service was intended to provide a firm quantitative bound on the

end-to-end packet delay for a flow. This is accomplished by

controlling the queuing delay on network elements along the data flow

path. The guaranteed service model does not, however, provide

bounds on jitter (inter-arrival times between consecutive packets).

The controlled-load service can be used for adaptive applications

that can tolerate some delay but are sensitive to traffic overload

conditions. This type of application typically functions

satisfactorily when the network is lightly loaded but its performance

degrades significantly when the network is heavily loaded.

Controlled-load service, therefore, has been designed to provide

approximately the same service as best-effort service in a lightly

loaded network regardless of actual network conditions. Controlled-

load service is described qualitatively in that no target values of

delay or loss are specified.

The main issue with the Integrated Services model has been

scalability [RFC-2998], especially in large public IP networks which

may potentially have millions of active micro-flows in transit

concurrently.

A notable feature of the Integrated Services model is that it

requires explicit signaling of QoS requirements from end systems to

routers [RFC-2753]. The Resource Reservation Protocol (RSVP)

performs this signaling function and is a critical component of the

Integrated Services model. The RSVP protocol is described next.

4.5.2 RSVP

RSVP is a soft state signaling protocol [RFC-2205]. It supports

receiver initiated establishment of resource reservations for both

multicast and unicast flows. RSVP was originally developed as a

signaling protocol within the integrated services framework for

applications to communicate QoS requirements to the network and for

the network to reserve relevant resources to satisfy the QoS

requirements [RFC-2205].

Under RSVP, the sender or source node sends a PATH message to the

receiver with the same source and destination addresses as the

traffic which the sender will generate. The PATH message contains:

(1) a sender Tspec specifying the characteristics of the traffic, (2)

a sender Template specifying the format of the traffic, and (3) an

optional Adspec which is used to support the concept of one pass with

advertising" (OPWA) [RFC-2205]. Every intermediate router along the

path forwards the PATH Message to the next hop determined by the

routing protocol. Upon receiving a PATH Message, the receiver

responds with a RESV message which includes a flow descriptor used to

request resource reservations. The RESV message travels to the

sender or source node in the opposite direction along the path that

the PATH message traversed. Every intermediate router along the path

can reject or accept the reservation request of the RESV message. If

the request is rejected, the rejecting router will send an error

message to the receiver and the signaling process will terminate. If

the request is accepted, link bandwidth and buffer space are

allocated for the flow and the related flow state information is

installed in the router.

One of the issues with the original RSVP specification was

Scalability. This is because reservations were required for micro-

flows, so that the amount of state maintained by network elements

tends to increase linearly with the number of micro-flows. These

issues are described in [RFC-2961].

Recently, RSVP has been modified and extended in several ways to

mitigate the scaling problems. As a result, it is becoming a

versatile signaling protocol for the Internet. For example, RSVP has

been extended to reserve resources for aggregation of flows, to set

up MPLS explicit label switched paths, and to perform other signaling

functions within the Internet. There are also a number of proposals

to reduce the amount of refresh messages required to maintain

established RSVP sessions [RFC-2961].

A number of IETF working groups have been engaged in activities

related to the RSVP protocol. These include the original RSVP

working group, the MPLS working group, the Resource Allocation

Protocol working group, and the Policy Framework working group.

4.5.3 Differentiated Services

The goal of the Differentiated Services (Diffserv) effort within the

IETF is to devise scalable mechanisms for categorization of traffic

into behavior aggregates, which ultimately allows each behavior

aggregate to be treated differently, especially when there is a

shortage of resources such as link bandwidth and buffer space [RFC-

2475]. One of the primary motivations for the Diffserv effort was to

devise alternative mechanisms for service differentiation in the

Internet that mitigate the scalability issues encountered with the

Intserv model.

The IETF Diffserv working group has defined a Differentiated Services

field in the IP header (DS field). The DS field consists of six bits

of the part of the IP header formerly known as TOS octet. The DS

field is used to indicate the forwarding treatment that a packet

should receive at a node [RFC-2474]. The Diffserv working group has

also standardized a number of Per-Hop Behavior (PHB) groups. Using

the PHBs, several classes of services can be defined using different

classification, policing, shaping, and scheduling rules.

For an end-user of network services to receive Differentiated

Services from its Internet Service Provider (ISP), it may be

necessary for the user to have a Service Level Agreement (SLA) with

the ISP. An SLA may explicitly or implicitly specify a Traffic

Conditioning Agreement (TCA) which defines classifier rules as well

as metering, marking, discarding, and shaping rules.

Packets are classified, and possibly policed and shaped at the

ingress to a Diffserv network. When a packet traverses the boundary

between different Diffserv domains, the DS field of the packet may be

re-marked according to existing agreements between the domains.

Differentiated Services allows only a finite number of service

classes to be indicated by the DS field. The main advantage of the

Diffserv approach relative to the Intserv model is scalability.

Resources are allocated on a per-class basis and the amount of state

information is proportional to the number of classes rather than to

the number of application flows.

It should be obvious from the previous discussion that the Diffserv

model essentially deals with traffic management issues on a per hop

basis. The Diffserv control model consists of a collection of

micro-TE control mechanisms. Other traffic engineering capabilities,

such as capacity management (including routing control), are also

required in order to deliver acceptable service quality in Diffserv

networks. The concept of Per Domain Behaviors has been introduced to

better capture the notion of differentiated services across a

complete domain [RFC-3086].

4.5.4 MPLS

MPLS is an advanced forwarding scheme which also includes extensions

to conventional IP control plane protocols. MPLS extends the

Internet routing model and enhances packet forwarding and path

control [RFC-3031].

At the ingress to an MPLS domain, label switching routers (LSRs)

classify IP packets into forwarding equivalence classes (FECs) based

on a variety of factors, including, e.g., a combination of the

information carried in the IP header of the packets and the local

routing information maintained by the LSRs. An MPLS label is then

prepended to each packet according to their forwarding equivalence

classes. In a non-ATM/FR environment, the label is 32 bits long and

contains a 20-bit label field, a 3-bit experimental field (formerly

known as Class-of-Service or CoS field), a 1-bit label stack

indicator and an 8-bit TTL field. In an ATM (FR) environment, the

label consists of information encoded in the VCI/VPI (DLCI) field.

An MPLS capable router (an LSR) examines the label and possibly the

experimental field and uses this information to make packet

forwarding decisions.

An LSR makes forwarding decisions by using the label prepended to

packets as the index into a local next hop label forwarding entry

(NHLFE). The packet is then processed as specified in the NHLFE.

The incoming label may be replaced by an outgoing label, and the

packet may be switched to the next LSR. This label-switching process

is very similar to the label (VCI/VPI) swapping process in ATM

networks. Before a packet leaves an MPLS domain, its MPLS label may

be removed. A Label Switched Path (LSP) is the path between an

ingress LSRs and an egress LSRs through which a labeled packet

traverses. The path of an explicit LSP is defined at the originating

(ingress) node of the LSP. MPLS can use a signaling protocol such as

RSVP or LDP to set up LSPs.

MPLS is a very powerful technology for Internet traffic engineering

because it supports explicit LSPs which allow constraint-based

routing to be implemented efficiently in IP networks [AWD2]. The

requirements for traffic engineering over MPLS are described in

[RFC-2702]. Extensions to RSVP to support instantiation of explicit

LSP are discussed in [RFC-3209]. Extensions to LDP, known as CR-LDP,

to support explicit LSPs are presented in [JAM].

4.5.5 IP Performance Metrics

The IETF IP Performance Metrics (IPPM) working group has been

developing a set of standard metrics that can be used to monitor the

quality, performance, and reliability of Internet services. These

metrics can be applied by network operators, end-users, and

independent testing groups to provide users and service providers

with a common understanding of the performance and reliability of the

Internet component 'clouds' they use/provide [RFC-2330]. The

criteria for performance metrics developed by the IPPM WG are

described in [RFC-2330]. Examples of performance metrics include

one-way packet

loss [RFC-2680], one-way delay [RFC-2679], and connectivity measures

between two nodes [RFC-2678]. Other metrics include second-order

measures of packet loss and delay.

Some of the performance metrics specified by the IPPM WG are useful

for specifying Service Level Agreements (SLAs). SLAs are sets of

service level objectives negotiated between users and service

providers, wherein each objective is a combination of one or more

performance metrics, possibly subject to certain constraints.

4.5.6 Flow Measurement

The IETF Real Time Flow Measurement (RTFM) working group has produced

an architecture document defining a method to specify traffic flows

as well as a number of components for flow measurement (meters, meter

readers, manager) [RFC-2722]. A flow measurement system enables

network traffic flows to be measured and analyzed at the flow level

for a variety of purposes. As noted in RFC2722, a flow measurement

system can be very useful in the following contexts: (1)

understanding the behavior of existing networks, (2) planning for

network development and expansion, (3) quantification of network

performance, (4) verifying the quality of network service, and (5)

attribution of network usage to users.

A flow measurement system consists of meters, meter readers, and

managers. A meter observes packets passing through a measurement

point, classifies them into certain groups, accumulates certain usage

data (such as the number of packets and bytes for each group), and

stores the usage data in a flow table. A group may represent a user

application, a host, a network, a group of networks, etc. A meter

reader gathers usage data from various meters so it can be made

available for analysis. A manager is responsible for configuring and

controlling meters and meter readers. The instructions received by a

meter from a manager include flow specification, meter control

parameters, and sampling techniques. The instructions received by a

meter reader from a manager include the address of the meter whose

date is to be collected, the frequency of data collection, and the

types of flows to be collected.

4.5.7 Endpoint Congestion Management

[RFC-3124] is intended to provide a set of congestion control

mechanisms that transport protocols can use. It is also intended to

develop mechanisms for unifying congestion control across a subset of

an endpoint's active unicast connections (called a congestion group).

A congestion manager continuously monitors the state of the path for

each congestion group under its control. The manager uses that

information to instruct a scheduler on how to partition bandwidth

among the connections of that congestion group.

4.6 Overview of ITU Activities Related to Traffic Engineering

This section provides an overview of prior work within the ITU-T

pertaining to traffic engineering in traditional telecommunications

networks.

ITU-T Recommendations E.600 [ITU-E600], E.701 [ITU-E701], and E.801

[ITU-E801] address traffic engineering issues in traditional

telecommunications networks. Recommendation E.600 provides a

vocabulary for describing traffic engineering concepts, while E.701

defines reference connections, Grade of Service (GOS), and traffic

parameters for ISDN. Recommendation E.701 uses the concept of a

reference connection to identify representative cases of different

types of connections without describing the specifics of their actual

realizations by different physical means. As defined in

Recommendation E.600, "a connection is an association of resources

providing means for communication between two or more devices in, or

attached to, a telecommunication network." Also, E.600 defines "a

resource as any set of physically or conceptually identifiable

entities within a telecommunication network, the use of which can be

unambiguously determined" [ITU-E600]. There can be different types

of connections as the number and types of resources in a connection

may vary.

Typically, different network segments are involved in the path of a

connection. For example, a connection may be local, national, or

international. The purposes of reference connections are to clarify

and specify traffic performance issues at various interfaces between

different network domains. Each domain may consist of one or more

service provider networks.

Reference connections provide a basis to define grade of service

(GoS) parameters related to traffic engineering within the ITU-T

framework. As defined in E.600, "GoS refers to a number of traffic

engineering variables which are used to provide a measure of the

adequacy of a group of resources under specified conditions." These

GoS variables may be probability of loss, dial tone, delay, etc.

They are essential for network internal design and operation as well

as for component performance specification.

GoS is different from quality of service (QoS) in the ITU framework.

QoS is the performance perceivable by a telecommunication service

user and expresses the user's degree of satisfaction of the service.

QoS parameters focus on performance aspects observable at the service

access points and network interfaces, rather than their causes within

the network. GoS, on the other hand, is a set of network oriented

measures which characterize the adequacy of a group of resources

under specified conditions. For a network to be effective in serving

its users, the values of both GoS and QoS parameters must be related,

with GoS parameters typically making a major contribution to the QoS.

Recommendation E.600 stipulates that a set of GoS parameters must be

selected and defined on an end-to-end basis for each major service

category provided by a network to assist the network provider with

improving efficiency and effectiveness of the network. Based on a

selected set of reference connections, suitable target values are

assigned to the selected GoS parameters under normal and high load

conditions. These end-to-end GoS target values are then apportioned

to individual resource components of the reference connections for

dimensioning purposes.

4.7 Content Distribution

The Internet is dominated by client-server interactions, especially

Web traffic (in the future, more sophisticated media servers may

become dominant). The location and performance of major information

servers has a significant impact on the traffic patterns within the

Internet as well as on the perception of service quality by end

users.

A number of dynamic load balancing techniques have been devised to

improve the performance of replicated information servers. These

techniques can cause spatial traffic characteristics to become more

dynamic in the Internet because information servers can be

dynamically picked based upon the location of the clients, the

location of the servers, the relative utilization of the servers, the

relative performance of different networks, and the relative

performance of different parts of a network. This process of

assignment of distributed servers to clients is called Traffic

Directing. It functions at the application layer.

Traffic Directing schemes that allocate servers in multiple

geographically dispersed locations to clients may require empirical

network performance statistics to make more effective decisions. In

the future, network measurement systems may need to provide this type

of information. The exact parameters needed are not yet defined.

When congestion exists in the network, Traffic Directing and Traffic

Engineering systems should act in a coordinated manner. This topic

is for further study.

The issues related to location and replication of information

servers, particularly web servers, are important for Internet traffic

engineering because these servers contribute a substantial proportion

of Internet traffic.

5.0 Taxonomy of Traffic Engineering Systems

This section presents a short taxonomy of traffic engineering

systems. A taxonomy of traffic engineering systems can be

constructed based on traffic engineering styles and views as listed

below:

- Time-dependent vs State-dependent vs Event-dependent

- Offline vs Online

- Centralized vs Distributed

- Local vs Global Information

- Prescriptive vs Descriptive

- Open Loop vs Closed Loop

- Tactical vs Strategic

These classification systems are described in greater detail in the

following subsections of this document.

5.1 Time-Dependent Versus State-Dependent Versus Event Dependent

Traffic engineering methodologies can be classified as time-

dependent, or state-dependent, or event-dependent. All TE schemes

are considered to be dynamic in this document. Static TE implies

that no traffic engineering methodology or algorithm is being

applied.

In the time-dependent TE, historical information based on periodic

variations in traffic, (such as time of day), is used to pre-program

routing plans and other TE control mechanisms. Additionally,

customer subscription or traffic projection may be used. Pre-

programmed routing plans typically change on a relatively long time

scale (e.g., diurnal). Time-dependent algorithms do not attempt to

adapt to random variations in traffic or changing network conditions.

An example of a time-dependent algorithm is a global centralized

optimizer where the input to the system is a traffic matrix and

multi-class QoS requirements as described [MR99].

State-dependent TE adapts the routing plans for packets based on the

current state of the network. The current state of the network

provides additional information on variations in actual traffic

(i.e., perturbations from regular variations) that could not be

predicted using historical information. Constraint-based routing is

an example of state-dependent TE operating in a relatively long time

scale. An example operating in a relatively short time scale is a

load-balancing algorithm described in [MATE].

The state of the network can be based on parameters such as

utilization, packet delay, packet loss, etc. These parameters can be

obtained in several ways. For example, each router may flood these

parameters periodically or by means of some kind of trigger to other

routers. Another approach is for a particular router performing

adaptive TE to send probe packets along a path to gather the state of

that path. Still another approach is for a management system to

gather relevant information from network elements.

Expeditious and accurate gathering and distribution of state

information is critical for adaptive TE due to the dynamic nature of

network conditions. State-dependent algorithms may be applied to

increase network efficiency and resilience. Time-dependent

algorithms are more suitable for predictable traffic variations. On

the other hand, state-dependent algorithms are more suitable for

adapting to the prevailing network state.

Event-dependent TE methods can also be used for TE path selection.

Event-dependent TE methods are distinct from time-dependent and

state-dependent TE methods in the manner in which paths are selected.

These algorithms are adaptive and distributed in nature and typically

use learning models to find good paths for TE in a network. While

state-dependent TE models typically use available-link-bandwidth

(ALB) flooding for TE path selection, event-dependent TE methods do

not require ALB flooding. Rather, event-dependent TE methods

typically search out capacity by learning models, as in the success-

to-the-top (STT) method. ALB flooding can be resource intensive,

since it requires link bandwidth to carry LSAs, processor capacity to

process LSAs, and the overhead can limit area/autonomous system (AS)

size. Modeling results suggest that event-dependent TE methods could

lead to a reduction in ALB flooding overhead without loss of network

throughput performance [ASH3].

5.2 Offline Versus Online

Traffic engineering requires the computation of routing plans. The

computation may be performed offline or online. The computation can

be done offline for scenarios where routing plans need not be

executed in real-time. For example, routing plans computed from

forecast information may be computed offline. Typically, offline

computation is also used to perform extensive searches on multi-

dimensional solution spaces.

Online computation is required when the routing plans must adapt to

changing network conditions as in state-dependent algorithms. Unlike

offline computation (which can be computationally demanding), online

computation is geared toward relative simple and fast calculations to

select routes, fine-tune the allocations of resources, and perform

load balancing.

5.3 Centralized Versus Distributed

Centralized control has a central authority which determines routing

plans and perhaps other TE control parameters on behalf of each

router. The central authority collects the network-state information

from all routers periodically and returns the routing information to

the routers. The routing update cycle is a critical parameter

directly impacting the performance of the network being controlled.

Centralized control may need high processing power and high bandwidth

control channels.

Distributed control determines route selection by each router

autonomously based on the routers view of the state of the network.

The network state information may be obtained by the router using a

probing method or distributed by other routers on a periodic basis

using link state advertisements. Network state information may also

be disseminated under exceptional conditions.

5.4 Local Versus Global

Traffic engineering algorithms may require local or global network-

state information.

Local information pertains to the state of a portion of the domain.

Examples include the bandwidth and packet loss rate of a particular

path. Local state information may be sufficient for certain

instances of distributed-controlled TEs.

Global information pertains to the state of the entire domain

undergoing traffic engineering. Examples include a global traffic

matrix and loading information on each link throughout the domain of

interest. Global state information is typically required with

centralized control. Distributed TE systems may also need global

information in some cases.

5.5 Prescriptive Versus Descriptive

TE systems may also be classified as prescriptive or descriptive.

Prescriptive traffic engineering evaluates alternatives and

recommends a course of action. Prescriptive traffic engineering can

be further categorized as either corrective or perfective.

Corrective TE prescribes a course of action to address an existing or

predicted anomaly. Perfective TE prescribes a course of action to

evolve and improve network performance even when no anomalies are

evident.

Descriptive traffic engineering, on the other hand, characterizes the

state of the network and assesses the impact of various policies

without recommending any particular course of action.

5.6 Open-Loop Versus Closed-Loop

Open-loop traffic engineering control is where control action does

not use feedback information from the current network state. The

control action may use its own local information for accounting

purposes, however.

Closed-loop traffic engineering control is where control action

utilizes feedback information from the network state. The feedback

information may be in the form of historical information or current

measurement.

5.7 Tactical vs Strategic

Tactical traffic engineering aims to address specific performance

problems (such as hot-spots) that occur in the network from a

tactical perspective, without consideration of overall strategic

imperatives. Without proper planning and insights, tactical TE tends

to be ad hoc in nature.

Strategic traffic engineering approaches the TE problem from a more

organized and systematic perspective, taking into consideration the

immediate and longer term consequences of specific policies and

actions.

6.0 Recommendations for Internet Traffic Engineering

This section describes high level recommendations for traffic

engineering in the Internet. These recommendations are presented in

general terms.

The recommendations describe the capabilities needed to solve a

traffic engineering problem or to achieve a traffic engineering

objective. Broadly speaking, these recommendations can be

categorized as either functional and non-functional recommendations.

Functional recommendations for Internet traffic engineering describe

the functions that a traffic engineering system should perform.

These functions are needed to realize traffic engineering objectives

by addressing traffic engineering problems.

Non-functional recommendations for Internet traffic engineering

relate to the quality attributes or state characteristics of a

traffic engineering system. These recommendations may contain

conflicting assertions and may sometimes be difficult to quantify

precisely.

6.1 Generic Non-functional Recommendations

The generic non-functional recommendations for Internet traffic

engineering include: usability, automation, scalability, stability,

visibility, simplicity, efficiency, reliability, correctness,

maintainability, extensibility, interoperability, and security. In a

given context, some of these recommendations may be critical while

others may be optional. Therefore, prioritization may be required

during the development phase of a traffic engineering system (or

components thereof) to tailor it to a specific operational context.

In the following paragraphs, some of the aspects of the non-

functional recommendations for Internet traffic engineering are

summarized.

Usability: Usability is a human factor aspect of traffic engineering

systems. Usability refers to the ease with which a traffic

engineering system can be deployed and operated. In general, it is

desirable to have a TE system that can be readily deployed in an

existing network. It is also desirable to have a TE system that is

easy to operate and maintain.

Automation: Whenever feasible, a traffic engineering system should

automate as many traffic engineering functions as possible to

minimize the amount of human effort needed to control and analyze

operational networks. Automation is particularly imperative in large

scale public networks because of the high cost of the human aspects

of network operations and the high risk of network problems caused by

human errors. Automation may entail the incorporation of automatic

feedback and intelligence into some components of the traffic

engineering system.

Scalability: Contemporary public networks are growing very fast with

respect to network size and traffic volume. Therefore, a TE system

should be scalable to remain applicable as the network evolves. In

particular, a TE system should remain functional as the network

expands with regard to the number of routers and links, and with

respect to the traffic volume. A TE system should have a scalable

architecture, should not adversely impair other functions and

processes in a network element, and should not consume too much

network resources when collecting and distributing state information

or when exerting control.

Stability: Stability is a very important consideration in traffic

engineering systems that respond to changes in the state of the

network. State-dependent traffic engineering methodologies typically

mandate a tradeoff between responsiveness and stability. It is

strongly recommended that when tradeoffs are warranted between

responsiveness and stability, that the tradeoff should be made in

favor of stability (especially in public IP backbone networks).

Flexibility: A TE system should be flexible to allow for changes in

optimization policy. In particular, a TE system should provide

sufficient configuration options so that a network administrator can

tailor the TE system to a particular environment. It may also be

desirable to have both online and offline TE subsystems which can be

independently enabled and disabled. TE systems that are used in

multi-class networks should also have options to support class based

performance evaluation and optimization.

Visibility: As part of the TE system, mechanisms should exist to

collect statistics from the network and to analyze these statistics

to determine how well the network is functioning. Derived statistics

such as traffic matrices, link utilization, latency, packet loss, and

other performance measures of interest which are determined from

network measurements can be used as indicators of prevailing network

conditions. Other examples of status information which should be

observed include existing functional routing information

(additionally, in the context of MPLS existing LSP routes), etc.

Simplicity: Generally, a TE system should be as simple as possible.

More importantly, the TE system should be relatively easy to use

(i.e., clean, convenient, and intuitive user interfaces). Simplicity

in user interface does not necessarily imply that the TE system will

use naive algorithms. When complex algorithms and internal

structures are used, such complexities should be hidden as much as

possible from the network administrator through the user interface.

Interoperability: Whenever feasible, traffic engineering systems and

their components should be developed with open standards based

interfaces to allow interoperation with other systems and components.

Security: Security is a critical consideration in traffic engineering

systems. Such traffic engineering systems typically exert control

over certain functional aspects of the network to achieve the desired

performance objectives. Therefore, adequate measures must be taken

to safeguard the integrity of the traffic engineering system.

Adequate measures must also be taken to protect the network from

vulnerabilities that originate from security breaches and other

impairments within the traffic engineering system.

The remainder of this section will focus on some of the high level

functional recommendations for traffic engineering.

6.2 Routing Recommendations

Routing control is a significant aspect of Internet traffic

engineering. Routing impacts many of the key performance measures

associated with networks, such as throughput, delay, and utilization.

Generally, it is very difficult to provide good service quality in a

wide area network without effective routing control. A desirable

routing system is one that takes traffic characteristics and network

constraints into account during route selection while maintaining

stability.

Traditional shortest path first (SPF) interior gateway protocols are

based on shortest path algorithms and have limited control

capabilities for traffic engineering [RFC-2702, AWD2]. These

limitations include :

1. The well known issues with pure SPF protocols, which do not take

network constraints and traffic characteristics into account

during route selection. For example, since IGPs always use the

shortest paths (based on administratively assigned link metrics)

to forward traffic, load sharing cannot be accomplished among

paths of different costs. Using shortest paths to forward traffic

conserves network resources, but may cause the following problems:

1) If traffic from a source to a destination exceeds the capacity

of a link along the shortest path, the link (hence the shortest

path) becomes congested while a longer path between these two

nodes may be under-utilized; 2) the shortest paths from different

sources can overlap at some links. If the total traffic from the

sources exceeds the capacity of any of these links, congestion

will occur. Problems can also occur because traffic demand

changes over time but network topology and routing configuration

cannot be changed as rapidly. This causes the network topology

and routing configuration to become sub-optimal over time, which

may result in persistent congestion problems.

2. The Equal-Cost Multi-Path (ECMP) capability of SPF IGPs supports

sharing of traffic among equal cost paths between two nodes.

However, ECMP attempts to divide the traffic as equally as

possible among the equal cost shortest paths. Generally, ECMP

does not support configurable load sharing ratios among equal cost

paths. The result is that one of the paths may carry

significantly more traffic than other paths because it may also

carry traffic from other sources. This situation can result in

congestion along the path that carries more traffic.

3. Modifying IGP metrics to control traffic routing tends to have

network-wide effect. Consequently, undesirable and unanticipated

traffic shifts can be triggered as a result. Recent work

described in Section 8.0 may be capable of better control [FT00,

FT01].

Because of these limitations, new capabilities are needed to enhance

the routing function in IP networks. Some of these capabilities have

been described elsewhere and are summarized below.

Constraint-based routing is desirable to evolve the routing

architecture of IP networks, especially public IP backbones with

complex topologies [RFC-2702]. Constraint-based routing computes

routes to fulfill requirements subject to constraints. Constraints

may include bandwidth, hop count, delay, and administrative policy

instruments such as resource class attributes [RFC-2702, RFC-2386].

This makes it possible to select routes that satisfy a given set of

requirements subject to network and administrative policy

constraints. Routes computed through constraint-based routing are

not necessarily the shortest paths. Constraint-based routing works

best with path oriented technologies that support explicit routing,

such as MPLS.

Constraint-based routing can also be used as a way to redistribute

traffic onto the infrastructure (even for best effort traffic). For

example, if the bandwidth requirements for path selection and

reservable bandwidth attributes of network links are appropriately

defined and configured, then congestion problems caused by uneven

traffic distribution may be avoided or reduced. In this way, the

performance and efficiency of the network can be improved.

A number of enhancements are needed to conventional link state IGPs,

such as OSPF and IS-IS, to allow them to distribute additional state

information required for constraint-based routing. These extensions

to OSPF were described in [KATZ] and to IS-IS in [SMIT].

Essentially, these enhancements require the propagation of additional

information in link state advertisements. Specifically, in addition

to normal link-state information, an enhanced IGP is required to

propagate topology state information needed for constraint-based

routing. Some of the additional topology state information include

link attributes such as reservable bandwidth and link resource class

attribute (an administratively specified property of the link). The

resource class attribute concept was defined in [RFC-2702]. The

additional topology state information is carried in new TLVs and

sub-TLVs in IS-IS, or in the Opaque LSA in OSPF [SMIT, KATZ].

An enhanced link-state IGP may flood information more frequently than

a normal IGP. This is because even without changes in topology,

changes in reservable bandwidth or link affinity can trigger the

enhanced IGP to initiate flooding. A tradeoff is typically required

between the timeliness of the information flooded and the flooding

frequency to avoid excessive consumption of link bandwidth and

computational resources, and more importantly, to avoid instability.

In a TE system, it is also desirable for the routing subsystem to

make the load splitting ratio among multiple paths (with equal cost

or different cost) configurable. This capability gives network

administrators more flexibility in the control of traffic

distribution across the network. It can be very useful for

avoiding/relieving congestion in certain situations. Examples can be

found in [XIAO].

The routing system should also have the capability to control the

routes of subsets of traffic without affecting the routes of other

traffic if sufficient resources exist for this purpose. This

capability allows a more refined control over the distribution of

traffic across the network. For example, the ability to move traffic

from a source to a destination away from its original path to another

path (without affecting other traffic paths) allows traffic to be

moved from resource-poor network segments to resource-rich segments.

Path oriented technologies such as MPLS inherently support this

capability as discussed in [AWD2].

Additionally, the routing subsystem should be able to select

different paths for different classes of traffic (or for different

traffic behavior aggregates) if the network supports multiple classes

of service (different behavior aggregates).

6.3 Traffic Mapping Recommendations

Traffic mapping pertains to the assignment of traffic workload onto

pre-established paths to meet certain requirements. Thus, while

constraint-based routing deals with path selection, traffic mapping

deals with the assignment of traffic to established paths which may

have been selected by constraint-based routing or by some other

means. Traffic mapping can be performed by time-dependent or state-

dependent mechanisms, as described in Section 5.1.

An important aspect of the traffic mapping function is the ability to

establish multiple paths between an originating node and a

destination node, and the capability to distribute the traffic

between the two nodes across the paths according to some policies. A

pre-condition for this scheme is the existence of flexible mechanisms

to partition traffic and then assign the traffic partitions onto the

parallel paths. This requirement was noted in [RFC-2702]. When

traffic is assigned to multiple parallel paths, it is recommended

that special care should be taken to ensure proper ordering of

packets belonging to the same application (or micro-flow) at the

destination node of the parallel paths.

As a general rule, mechanisms that perform the traffic mapping

functions should aim to map the traffic onto the network

infrastructure to minimize congestion. If the total traffic load

cannot be accommodated, or if the routing and mapping functions

cannot react fast enough to changing traffic conditions, then a

traffic mapping system may rely on short time scale congestion

control mechanisms (such as queue management, scheduling, etc.) to

mitigate congestion. Thus, mechanisms that perform the traffic

mapping functions should complement existing congestion control

mechanisms. In an operational network, it is generally desirable to

map the traffic onto the infrastructure such that intra-class and

inter-class resource contention are minimized.

When traffic mapping techniques that depend on dynamic state feedback

(e.g., MATE and such like) are used, special care must be taken to

guarantee network stability.

6.4 Measurement Recommendations

The importance of measurement in traffic engineering has been

discussed throughout this document. Mechanisms should be provided to

measure and collect statistics from the network to support the

traffic engineering function. Additional capabilities may be needed

to help in the analysis of the statistics. The actions of these

mechanisms should not adversely affect the accuracy and integrity of

the statistics collected. The mechanisms for statistical data

acquisition should also be able to scale as the network evolves.

Traffic statistics may be classified according to long-term or

short-term time scales. Long-term time scale traffic statistics are

very useful for traffic engineering. Long-term time scale traffic

statistics may capture or reflect periodicity in network workload

(such as hourly, daily, and weekly variations in traffic profiles) as

well as traffic trends. Aspects of the monitored traffic statistics

may also depict class of service characteristics for a network

supporting multiple classes of service. Analysis of the long-term

traffic statistics MAY yield secondary statistics such as busy hour

characteristics, traffic growth patterns, persistent congestion

problems, hot-spot, and imbalances in link utilization caused by

routing anomalies.

A mechanism for constructing traffic matrices for both long-term and

short-term traffic statistics should be in place. In multi-service

IP networks, the traffic matrices may be constructed for different

service classes. Each element of a traffic matrix represents a

statistic of traffic flow between a pair of abstract nodes. An

abstract node may represent a router, a collection of routers, or a

site in a VPN.

Measured traffic statistics should provide reasonable and reliable

indicators of the current state of the network on the short-term

scale. Some short term traffic statistics may reflect link

utilization and link congestion status. Examples of congestion

indicators include excessive packet delay, packet loss, and high

resource utilization. Examples of mechanisms for distributing this

kind of information include SNMP, probing techniques, FTP, IGP link

state advertisements, etc.

6.5 Network Survivability

Network survivability refers to the capability of a network to

maintain service continuity in the presence of faults. This can be

accomplished by promptly recovering from network impairments and

maintaining the required QoS for existing services after recovery.

Survivability has become an issue of great concern within the

Internet community due to the increasing demands to carry mission

critical traffic, real-time traffic, and other high priority traffic

over the Internet. Survivability can be addressed at the device

level by developing network elements that are more reliable; and at

the network level by incorporating redundancy into the architecture,

design, and operation of networks. It is recommended that a

philosophy of robustness and survivability should be adopted in the

architecture, design, and operation of traffic engineering that

control IP networks (especially public IP networks). Because

different contexts may demand different levels of survivability, the

mechanisms developed to support network survivability should be

flexible so that they can be tailored to different needs.

Failure protection and restoration capabilities have become available

from multiple layers as network technologies have continued to

improve. At the bottom of the layered stack, optical networks are

now capable of providing dynamic ring and mesh restoration

functionality at the wavelength level as well as traditional

protection functionality. At the SONET/SDH layer survivability

capability is provided with Automatic Protection Switching (APS) as

well as self-healing ring and mesh architectures. Similar

functionality is provided by layer 2 technologies such as ATM

(generally with slower mean restoration times). Rerouting is

traditionally used at the IP layer to restore service following link

and node outages. Rerouting at the IP layer occurs after a period of

routing convergence which may require seconds to minutes to complete.

Some new developments in the MPLS context make it possible to achieve

recovery at the IP layer prior to convergence [SHAR].

To support advanced survivability requirements, path-oriented

technologies such a MPLS can be used to enhance the survivability of

IP networks in a potentially cost effective manner. The advantages

of path oriented technologies such as MPLS for IP restoration becomes

even more evident when class based protection and restoration

capabilities are required.

Recently, a common suite of control plane protocols has been proposed

for both MPLS and optical transport networks under the acronym

Multi-protocol Lambda Switching [AWD1]. This new paradigm of Multi-

protocol Lambda Switching will support even more sophisticated mesh

restoration capabilities at the optical layer for the emerging IP

over WDM network architectures.

Another important aspect regarding multi-layer survivability is that

technologies at different layers provide protection and restoration

capabilities at different temporal granularities (in terms of time

scales) and at different bandwidth granularity (from packet-level to

wavelength level). Protection and restoration capabilities can also

be sensitive to different service classes and different network

utility models.

The impact of service outages varies significantly for different

service classes depending upon the effective duration of the outage.

The duration of an outage can vary from milliseconds (with minor

service impact) to seconds (with possible call drops for IP telephony

and session time-outs for connection oriented transactions) to

minutes and hours (with potentially considerable social and business

impact).

Coordinating different protection and restoration capabilities across

multiple layers in a cohesive manner to ensure network survivability

is maintained at reasonable cost is a challenging task. Protection

and restoration coordination across layers may not always be

feasible, because networks at different layers may belong to

different administrative domains.

The following paragraphs present some of the general recommendations

for protection and restoration coordination.

- Protection and restoration capabilities from different layers

should be coordinated whenever feasible and appropriate to provide

network survivability in a flexible and cost effective manner.

Minimization of function duplication across layers is one way to

achieve the coordination. Escalation of alarms and other fault

indicators from lower to higher layers may also be performed in a

coordinated manner. A temporal order of restoration trigger timing

at different layers is another way to coordinate multi-layer

protection/restoration.

- Spare capacity at higher layers is often regarded as working

traffic at lower layers. Placing protection/restoration functions in

many layers may increase redundancy and robustness, but it should not

result in significant and avoidable inefficiencies in network

resource utilization.

- It is generally desirable to have protection and restoration

schemes that are bandwidth efficient.

- Failure notification throughout the network should be timely and

reliable.

- Alarms and other fault monitoring and reporting capabilities

should be provided at appropriate layers.

6.5.1 Survivability in MPLS Based Networks

MPLS is an important emerging technology that enhances IP networks in

terms of features, capabilities, and services. Because MPLS is

path-oriented, it can potentially provide faster and more predictable

protection and restoration capabilities than conventional hop by hop

routed IP systems. This subsection describes some of the basic

aspects and recommendations for MPLS networks regarding protection

and restoration. See [SHAR] for a more comprehensive discussion on

MPLS based recovery.

Protection types for MPLS networks can be categorized as link

protection, node protection, path protection, and segment protection.

- Link Protection: The objective for link protection is to protect

an LSP from a given link failure. Under link protection, the path

of the protection or backup LSP (the secondary LSP) is disjoint

from the path of the working or operational LSP at the particular

link over which protection is required. When the protected link

fails, traffic on the working LSP is switched over to the

protection LSP at the head-end of the failed link. This is a

local repair method which can be fast. It might be more

appropriate in situations where some network elements along a

given path are less reliable than others.

- Node Protection: The objective of LSP node protection is to

protect an LSP from a given node failure. Under node protection,

the path of the protection LSP is disjoint from the path of the

working LSP at the particular node to be protected. The secondary

path is also disjoint from the primary path at all links

associated with the node to be protected. When the node fails,

traffic on the working LSP is switched over to the protection LSP

at the upstream LSR directly connected to the failed node.

- Path Protection: The goal of LSP path protection is to protect an

LSP from failure at any point along its routed path. Under path

protection, the path of the protection LSP is completely disjoint

from the path of the working LSP. The advantage of path

protection is that the backup LSP protects the working LSP from

all possible link and node failures along the path, except for

failures that might occur at the ingress and egress LSRs, or for

correlated failures that might impact both working and backup

paths simultaneously. Additionally, since the path selection is

end-to-end, path protection might be more efficient in terms of

resource usage than link or node protection. However, path

protection may be slower than link and node protection in general.

- Segment Protection: An MPLS domain may be partitioned into

multiple protection domains whereby a failure in a protection

domain is rectified within that domain. In cases where an LSP

traverses multiple protection domains, a protection mechanism

within a domain only needs to protect the segment of the LSP that

lies within the domain. Segment protection will generally be

faster than path protection because recovery generally occurs

closer to the fault.

6.5.2 Protection Option

Another issue to consider is the concept of protection options. The

protection option uses the notation m:n protection, where m is the

number of protection LSPs used to protect n working LSPs. Feasible

protection options follow.

- 1:1: one working LSP is protected/restored by one protection LSP.

- 1:n: one protection LSP is used to protect/restore n working LSPs.

- n:1: one working LSP is protected/restored by n protection LSPs,

possibly with configurable load splitting ratio. When more than

one protection LSP is used, it may be desirable to share the

traffic across the protection LSPs when the working LSP fails to

satisfy the bandwidth requirement of the traffic trunk associated

with the working LSP. This may be especially useful when it is

not feasible to find one path that can satisfy the bandwidth

requirement of the primary LSP.

- 1+1: traffic is sent concurrently on both the working LSP and the

protection LSP. In this case, the egress LSR selects one of the

two LSPs based on a local traffic integrity decision process,

which compares the traffic received from both the working and the

protection LSP and identifies discrepancies. It is unlikely that

this option would be used extensively in IP networks due to its

resource utilization inefficiency. However, if bandwidth becomes

plentiful and cheap, then this option might become quite viable

and attractive in IP networks.

6.6 Traffic Engineering in Diffserv Environments

This section provides an overview of the traffic engineering features

and recommendations that are specifically pertinent to Differentiated

Services (Diffserv) [RFC-2475] capable IP networks.

Increasing requirements to support multiple classes of traffic, such

as best effort and mission critical data, in the Internet calls for

IP networks to differentiate traffic according to some criteria, and

to accord preferential treatment to certain types of traffic. Large

numbers of flows can be aggregated into a few behavior aggregates

based on some criteria in terms of common performance requirements in

terms of packet loss ratio, delay, and jitter; or in terms of common

fields within the IP packet headers.

As Diffserv evolves and becomes deployed in operational networks,

traffic engineering will be critical to ensuring that SLAs defined

within a given Diffserv service model are met. Classes of service

(CoS) can be supported in a Diffserv environment by concatenating

per-hop behaviors (PHBs) along the routing path, using service

provisioning mechanisms, and by appropriately configuring edge

functionality such as traffic classification, marking, policing, and

shaping. PHB is the forwarding behavior that a packet receives at a

DS node (a Diffserv-compliant node). This is accomplished by means

of buffer management and packet scheduling mechanisms. In this

context, packets belonging to a class are those that are members of a

corresponding ordering aggregate.

Traffic engineering can be used as a compliment to Diffserv

mechanisms to improve utilization of network resources, but not as a

necessary element in general. When traffic engineering is used, it

can be operated on an aggregated basis across all service classes

[RFC-3270] or on a per service class basis. The former is used to

provide better distribution of the aggregate traffic load over the

network resources. (See [RFC-3270] for detailed mechanisms to

support aggregate traffic engineering.) The latter case is discussed

below since it is specific to the Diffserv environment, with so

called Diffserv-aware traffic engineering [DIFF_TE].

For some Diffserv networks, it may be desirable to control the

performance of some service classes by enforcing certain

relationships between the traffic workload contributed by each

service class and the amount of network resources allocated or

provisioned for that service class. Such relationships between

demand and resource allocation can be enforced using a combination

of, for example: (1) traffic engineering mechanisms on a per service

class basis that enforce the desired relationship between the amount

of traffic contributed by a given service class and the resources

allocated to that class, and (2) mechanisms that dynamically adjust

the resources allocated to a given service class to relate to the

amount of traffic contributed by that service class.

It may also be desirable to limit the performance impact of high

priority traffic on relatively low priority traffic. This can be

achieved by, for example, controlling the percentage of high priority

traffic that is routed through a given link. Another way to

accomplish this is to increase link capacities appropriately so that

lower priority traffic can still enjoy adequate service quality.

When the ratio of traffic workload contributed by different service

classes vary significantly from router to router, it may not suffice

to rely exclusively on conventional IGP routing protocols or on

traffic engineering mechanisms that are insensitive to different

service classes. Instead, it may be desirable to perform traffic

engineering, especially routing control and mapping functions, on a

per service class basis. One way to accomplish this in a domain that

supports both MPLS and Diffserv is to define class specific LSPs and

to map traffic from each class onto one or more LSPs that correspond

to that service class. An LSP corresponding to a given service class

can then be routed and protected/restored in a class dependent

manner, according to specific policies.

Performing traffic engineering on a per class basis may require

certain per-class parameters to be distributed. Note that it is

common to have some classes share some aggregate constraint (e.g.,

maximum bandwidth requirement) without enforcing the constraint on

each individual class. These classes then can be grouped into a

class-type and per-class-type parameters can be distributed instead

to improve scalability. It also allows better bandwidth sharing

between classes in the same class-type. A class-type is a set of

classes that satisfy the following two conditions:

1) Classes in the same class-type have common aggregate requirements

to satisfy required performance levels.

2) There is no requirement to be enforced at the level of individual

class in the class-type. Note that it is still possible,

nevertheless, to implement some priority policies for classes in the

same class-type to permit preferential access to the class-type

bandwidth through the use of preemption priorities.

An example of the class-type can be a low-loss class-type that

includes both AF1-based and AF2-based Ordering Aggregates. With such

a class-type, one may implement some priority policy which assigns

higher preemption priority to AF1-based traffic trunks over AF2-based

ones, vice versa, or the same priority.

See [DIFF-TE] for detailed requirements on Diffserv-aware traffic

engineering.

6.7 Network Controllability

Off-line (and on-line) traffic engineering considerations would be of

limited utility if the network could not be controlled effectively to

implement the results of TE decisions and to achieve desired network

performance objectives. Capacity augmentation is a coarse grained

solution to traffic engineering issues. However, it is simple and

may be advantageous if bandwidth is abundant and cheap or if the

current or expected network workload demands it. However, bandwidth

is not always abundant and cheap, and the workload may not always

demand additional capacity. Adjustments of administrative weights

and other parameters associated with routing protocols provide finer

grained control, but is difficult to use and imprecise because of the

routing interactions that occur across the network. In certain

network contexts, more flexible, finer grained approaches which

provide more precise control over the mapping of traffic to routes

and over the selection and placement of routes may be appropriate and

useful.

Control mechanisms can be manual (e.g., administrative

configuration), partially-automated (e.g., scripts) or fully-

automated (e.g., policy based management systems). Automated

mechanisms are particularly required in large scale networks.

Multi-vendor interoperability can be facilitated by developing and

deploying standardized management

systems (e.g., standard MIBs) and policies (PIBs) to support the

control functions required to address traffic engineering objectives

such as load distribution and protection/restoration.

Network control functions should be secure, reliable, and stable as

these are often needed to operate correctly in times of network

impairments (e.g., during network congestion or security attacks).

7.0 Inter-Domain Considerations

Inter-domain traffic engineering is concerned with the performance

optimization for traffic that originates in one administrative domain

and terminates in a different one.

Traffic exchange between autonomous systems in the Internet occurs

through exterior gateway protocols. Currently, BGP [BGP4] is the

standard exterior gateway protocol for the Internet. BGP provides a

number of attributes and capabilities (e.g., route filtering) that

can be used for inter-domain traffic engineering. More specifically,

BGP permits the control of routing information and traffic exchange

between Autonomous Systems (AS's) in the Internet. BGP incorporates

a sequential decision process which calculates the degree of

preference for various routes to a given destination network. There

are two fundamental aspects to inter-domain traffic engineering using

BGP:

- Route Redistribution: controlling the import and export of routes

between AS's, and controlling the redistribution of routes between

BGP and other protocols within an AS.

- Best path selection: selecting the best path when there are

multiple candidate paths to a given destination network. Best

path selection is performed by the BGP decision process based on a

sequential procedure, taking a number of different considerations

into account. Ultimately, best path selection under BGP boils

down to selecting preferred exit points out of an AS towards

specific destination networks. The BGP path selection process can

be influenced by manipulating the attributes associated with the

BGP decision process. These attributes include: NEXT-HOP, WEIGHT

(Cisco proprietary which is also implemented by some other

vendors), LOCAL-PREFERENCE, AS-PATH, ROUTE-ORIGIN, MULTI-EXIT-

DESCRIMINATOR (MED), IGP METRIC, etc.

Route-maps provide the flexibility to implement complex BGP policies

based on pre-configured logical conditions. In particular, Route-

maps can be used to control import and export policies for incoming

and outgoing routes, control the redistribution of routes between BGP

and other protocols, and influence the selection of best paths by

manipulating the attributes associated with the BGP decision process.

Very complex logical expressions that implement various types of

policies can be implemented using a combination of Route-maps, BGP-

attributes, Access-lists, and Community attributes.

When looking at possible strategies for inter-domain TE with BGP, it

must be noted that the outbound traffic exit point is controllable,

whereas the interconnection point where inbound traffic is received

from an EBGP peer typically is not, unless a special arrangement is

made with the peer sending the traffic. Therefore, it is up to each

individual network to implement sound TE strategies that deal with

the efficient delivery of outbound traffic from one's customers to

one's peering points. The vast majority of TE policy is based upon a

"closest exit" strategy, which offloads interdomain traffic at the

nearest outbound peer point towards the destination autonomous

system. Most methods of manipulating the point at which inbound

traffic enters a network from an EBGP peer (inconsistent route

announcements between peering points, AS pre-pending, and sending

MEDs) are either ineffective, or not accepted in the peering

community.

Inter-domain TE with BGP is generally effective, but it is usually

applied in a trial-and-error fashion. A systematic approach for

inter-domain traffic engineering is yet to be devised.

Inter-domain TE is inherently more difficult than intra-domain TE

under the current Internet architecture. The reasons for this are

both technical and administrative. Technically, while topology and

link state information are helpful for mapping traffic more

effectively, BGP does not propagate such information across domain

boundaries for stability and scalability reasons. Administratively,

there are differences in operating costs and network capacities

between domains. Generally, what may be considered a good solution

in one domain may not necessarily be a good solution in another

domain. Moreover, it would generally be considered inadvisable for

one domain to permit another domain to influence the routing and

management of traffic in its network.

MPLS TE-tunnels (explicit LSPs) can potentially add a degree of

flexibility in the selection of exit points for inter-domain routing.

The concept of relative and absolute metrics can be applied to this

purpose. The idea is that if BGP attributes are defined such that

the BGP decision process depends on IGP metrics to select exit points

for inter-domain traffic, then some inter-domain traffic destined to

a given peer network can be made to prefer a specific exit point by

establishing a TE-tunnel between the router making the selection to

the peering point via a TE-tunnel and assigning the TE-tunnel a

metric which is smaller than the IGP cost to all other peering

points. If a peer accepts and processes MEDs, then a similar MPLS

TE-tunnel based scheme can be applied to cause certain entrance

points to be preferred by setting MED to be an IGP cost, which has

been modified by the tunnel metric.

Similar to intra-domain TE, inter-domain TE is best accomplished when

a traffic matrix can be derived to depict the volume of traffic from

one autonomous system to another.

Generally, redistribution of inter-domain traffic requires

coordination between peering partners. An export policy in one

domain that results in load redistribution across peer points with

another domain can significantly affect the local traffic matrix

inside the domain of the peering partner. This, in turn, will affect

the intra-domain TE due to changes in the spatial distribution of

traffic. Therefore, it is mutually beneficial for peering partners

to coordinate with each other before attempting any policy changes

that may result in significant shifts in inter-domain traffic. In

certain contexts, this coordination can be quite challenging due to

technical and non- technical reasons.

It is a matter of speculation as to whether MPLS, or similar

technologies, can be extended to allow selection of constrained paths

across domain boundaries.

8.0 Overview of Contemporary TE Practices in Operational IP Networks

This section provides an overview of some contemporary traffic

engineering practices in IP networks. The focus is primarily on the

aspects that pertain to the control of the routing function in

operational contexts. The intent here is to provide an overview of

the commonly used practices. The discussion is not intended to be

exhaustive.

Currently, service providers apply many of the traffic engineering

mechanisms discussed in this document to optimize the performance of

their IP networks. These techniques include capacity planning for

long time scales, routing control using IGP metrics and MPLS for

medium time scales, the overlay model also for medium time scales,

and traffic management mechanisms for short time scale.

When a service provider plans to build an IP network, or expand the

capacity of an existing network, effective capacity planning should

be an important component of the process. Such plans may take the

following aspects into account: location of new nodes if any,

existing and predicted traffic patterns, costs, link capacity,

topology, routing design, and survivability.

Performance optimization of operational networks is usually an

ongoing process in which traffic statistics, performance parameters,

and fault indicators are continually collected from the network.

This empirical data is then analyzed and used to trigger various

traffic engineering mechanisms. Tools that perform what-if analysis

can also be used to assist the TE process by allowing various

scenarios to be reviewed before a new set of configurations are

implemented in the operational network.

Traditionally, intra-domain real-time TE with IGP is done by

increasing the OSPF or IS-IS metric of a congested link until enough

traffic has been diverted from that link. This approach has some

limitations as discussed in Section 6.2. Recently, some new intra-

domain TE approaches/tools have been proposed

[RR94][FT00][FT01][WANG]. Such approaches/tools take traffic matrix,

network topology, and network performance objective(s) as input, and

produce some link metrics and possibly some unequal load-sharing

ratios to be set at the head-end routers of some ECMPs as output.

These new progresses open new possibility for intra-domain TE with

IGP to be done in a more systematic way.

The overlay model (IP over ATM or IP over Frame relay) is another

approach which is commonly used in practice [AWD2]. The IP over ATM

technique is no longer viewed favorably due to recent advances in

MPLS and router hardware technology.

Deployment of MPLS for traffic engineering applications has commenced

in some service provider networks. One operational scenario is to

deploy MPLS in conjunction with an IGP (IS-IS-TE or OSPF-TE) that

supports the traffic engineering extensions, in conjunction with

constraint-based routing for explicit route computations, and a

signaling protocol (e.g., RSVP-TE or CRLDP) for LSP instantiation.

In contemporary MPLS traffic engineering contexts, network

administrators specify and configure link attributes and resource

constraints such as maximum reservable bandwidth and resource class

attributes for links (interfaces) within the MPLS domain. A link

state protocol that supports TE extensions (IS-IS-TE or OSPF-TE) is

used to propagate information about network topology and link

attribute to all routers in the routing area. Network administrators

also specify all the LSPs that are to originate each router. For

each LSP, the network administrator specifies the destination node

and the attributes of the LSP which indicate the requirements that to

be satisfied during the path selection process. Each router then

uses a local constraint-based routing process to compute explicit

paths for all LSPs originating from it. Subsequently, a signaling

protocol is used to instantiate the LSPs. By assigning proper

bandwidth values to links and LSPs, congestion caused by uneven

traffic distribution can generally be avoided or mitigated.

The bandwidth attributes of LSPs used for traffic engineering can be

updated periodically. The basic concept is that the bandwidth

assigned to an LSP should relate in some manner to the bandwidth

requirements of traffic that actually flows through the LSP. The

traffic attribute of an LSP can be modified to accommodate traffic

growth and persistent traffic shifts. If network congestion occurs

due to some unexpected events, existing LSPs can be rerouted to

alleviate the situation or network administrator can configure new

LSPs to divert some traffic to alternative paths. The reservable

bandwidth of the congested links can also be reduced to force some

LSPs to be rerouted to other paths.

In an MPLS domain, a traffic matrix can also be estimated by

monitoring the traffic on LSPs. Such traffic statistics can be used

for a variety of purposes including network planning and network

optimization. Current practice suggests that deploying an MPLS

network consisting of hundreds of routers and thousands of LSPs is

feasible. In summary, recent deployment experience suggests that

MPLS approach is very effective for traffic engineering in IP

networks [XIAO].

As mentioned previously in Section 7.0, one usually has no direct

control over the distribution of inbound traffic. Therefore, the

main goal of contemporary inter-domain TE is to optimize the

distribution of outbound traffic between multiple inter-domain links.

When operating a global network, maintaining the ability to operate

the network in a regional fashion where desired, while continuing to

take advantage of the benefits of a global network, also becomes an

important objective.

Inter-domain TE with BGP usually begins with the placement of

multiple peering interconnection points in locations that have high

peer density, are in close proximity to originating/terminating

traffic locations on one's own network, and are lowest in cost.

There are generally several locations in each region of the world

where the vast majority of major networks congregate and

interconnect. Some location-decision problems that arise in

association with inter-domain routing are discussed in [AWD5].

Once the locations of the interconnects are determined, and circuits

are implemented, one decides how best to handle the routes heard from

the peer, as well as how to propagate the peers' routes within one's

own network. One way to engineer outbound traffic flows on a network

with many EBGP peers is to create a hierarchy of peers. Generally,

the Local Preferences of all peers are set to the same value so that

the shortest AS paths will be chosen to forward traffic. Then, by

over-writing the inbound MED metric (Multi-exit-discriminator metric,

also referred to as "BGP metric". Both terms are used

interchangeably in this document) with BGP metrics to routes received

at different peers, the hierarchy can be formed. For example, all

Local Preferences can be set to 200, preferred private peers can be

assigned a BGP metric of 50, the rest of the private peers can be

assigned a BGP metric of 100, and public peers can be assigned a BGP

metric of 600. "Preferred" peers might be defined as those peers

with whom the most available capacity exists, whose customer base is

larger in comparison to other peers, whose interconnection costs are

the lowest, and with whom upgrading existing capacity is the easiest.

In a network with low utilization at the edge, this works well. The

same concept could be applied to a network with higher edge

utilization by creating more levels of BGP metrics between peers,

allowing for more granularity in selecting the exit points for

traffic bound for a dual homed customer on a peer's network.

By only replacing inbound MED metrics with BGP metrics, only equal

AS-Path length routes' exit points are being changed. (The BGP

decision considers Local Preference first, then AS-Path length, and

then BGP metric). For example, assume a network has two possible

egress points, peer A and peer B. Each peer has 40% of the

Internet's routes exclusively on its network, while the remaining 20%

of the Internet's routes are from customers who dual home between A

and B. Assume that both peers have a Local Preference of 200 and a

BGP metric of 100. If the link to peer A is congested, increasing

its BGP metric while leaving the Local Preference at 200 will ensure

that the 20% of total routes belonging to dual homed customers will

prefer peer B as the exit point. The previous example would be used

in a situation where all exit points to a given peer were close to

congestion levels, and traffic needed to be shifted away from that

peer entirely.

When there are multiple exit points to a given peer, and only one of

them is congested, it is not necessary to shift traffic away from the

peer entirely, but only from the one congested circuit. This can be

achieved by using passive IGP-metrics, AS-path filtering, or prefix

filtering.

Occasionally, more drastic changes are needed, for example, in

dealing with a "problem peer" who is difficult to work with on

upgrades or is charging high prices for connectivity to their

network. In that case, the Local Preference to that peer can be

reduced below the level of other peers. This effectively reduces the

amount of traffic sent to that peer to only originating traffic

(assuming no transit providers are involved). This type of change

can affect a large amount of traffic, and is only used after other

methods have failed to provide the desired results.

Although it is not much of an issue in regional networks, the

propagation of a peer's routes back through the network must be

considered when a network is peering on a global scale. Sometimes,

business considerations can influence the choice of BGP policies in a

given context. For example, it may be imprudent, from a business

perspective, to operate a global network and provide full access to

the global customer base to a small network in a particular country.

However, for the purpose of providing one's own customers with

quality service in a particular region, good connectivity to that

in-country network may still be necessary. This can be achieved by

assigning a set of communities at the edge of the network, which have

a known behavior when routes tagged with those communities are

propagating back through the core. Routes heard from local peers

will be prevented from propagating back to the global network,

whereas routes learned from larger peers may be allowed to propagate

freely throughout the entire global network. By implementing a

flexible community strategy, the benefits of using a single global AS

Number (ASN) can be realized, while the benefits of operating

regional networks can also be taken advantage of. An alternative to

doing this is to use different ASNs in different regions, with the

consequence that the AS path length for routes announced by that

service provider will increase.

9.0 Conclusion

This document described principles for traffic engineering in the

Internet. It presented an overview of some of the basic issues

surrounding traffic engineering in IP networks. The context of TE

was described, a TE process models and a taxonomy of TE styles were

presented. A brief historical review of pertinent developments

related to traffic engineering was provided. A survey of

contemporary TE techniques in operational networks was presented.

Additionally, the document specified a set of generic requirements,

recommendations, and options for Internet traffic engineering.

10.0 Security Considerations

This document does not introduce new security issues.

11.0 Acknowledgments

The authors would like to thank Jim Boyle for inputs on the

recommendations section, Francois Le Faucheur for inputs on Diffserv

aspects, Blaine Christian for inputs on measurement, Gerald Ash for

inputs on routing in telephone networks and for text on event-

dependent TE methods, Steven Wright for inputs on network

controllability, and Jonathan Aufderheide for inputs on inter-domain

TE with BGP. Special thanks to Randy Bush for proposing the TE

taxonomy based on "tactical vs strategic" methods. The subsection

describing an "Overview of ITU Activities Related to Traffic

Engineering" was adapted from a contribution by Waisum Lai. Useful

feedback and pointers to relevant materials were provided by J. Noel

Chiappa. Additional comments were provided by Glenn Grotefeld during

the working last call process. Finally, the authors would like to

thank Ed Kern, the TEWG co-chair, for his comments and support.

12.0 References

[ASH2] J. Ash, Dynamic Routing in Telecommunications Networks,

McGraw Hill, 1998.

[ASH3] Ash, J., "TE & QoS Methods for IP-, ATM-, & TDM-Based

Networks", Work in Progress, March 2001.

[AWD1] D. Awduche and Y. Rekhter, "Multiprocotol Lambda

Switching: Combining MPLS Traffic Engineering Control

with Optical Crossconnects", IEEE Communications

Magazine, March 2001.

[AWD2] D. Awduche, "MPLS and Traffic Engineering in IP

Networks", IEEE Communications Magazine, Dec. 1999.

[AWD5] D. Awduche et al, "An Approach to Optimal Peering Between

Autonomous Systems in the Internet", International

Conference on Computer Communications and Networks

(ICCCN'98), Oct. 1998.

[CRUZ] R. L. Cruz, "A Calculus for Network Delay, Part II:

Network Analysis", IEEE Transactions on Information

Theory, vol. 37, pp. 132-141, 1991.

[DIFF-TE] Le Faucheur, F., Nadeau, T., Tatham, M., Telkamp, T.,

Cooper, D., Boyle, J., Lai, W., Fang, L., Ash, J., Hicks,

P., Chui, A., Townsend, W. and D. Skalecki, "Requirements

for support of Diff-Serv-aware MPLS Traffic Engineering",

Work in Progress, May 2001.

[ELW95] A. Elwalid, D. Mitra and R.H. Wentworth, "A New Approach

for Allocating Buffers and Bandwidth to Heterogeneous,

Regulated Traffic in an ATM Node", IEEE IEEE Journal on

Selected Areas in Communications, 13:6, pp. 1115-1127,

Aug. 1995.

[FGLR] A. Feldmann, A. Greenberg, C. Lund, N. Reingold, and J.

Rexford, "NetScope: Traffic Engineering for IP Networks",

IEEE Network Magazine, 2000.

[FLJA93] S. Floyd and V. Jacobson, "Random Early Detection

Gateways for Congestion Avoidance", IEEE/ACM Transactions

on Networking, Vol. 1 Nov. 4., p. 387-413, Aug. 1993.

[FLOY94] S. Floyd, "TCP and Explicit Congestion Notification", ACM

Computer Communication Review, V. 24, No. 5, p. 10-23,

Oct. 1994.

[FT00] B. Fortz and M. Thorup, "Internet Traffic Engineering by

Optimizing OSPF Weights", IEEE INFOCOM 2000, Mar. 2000.

[FT01] B. Fortz and M. Thorup, "Optimizing OSPF/IS-IS Weights in

a Changing World",

www.research.att.com/~mthorup/PAPERS/papers.Html.

[HUSS87] B.R. Hurley, C.J.R. Seidl and W.F. Sewel, "A Survey of

Dynamic Routing Methods for Circuit-Switched Traffic",

IEEE Communication Magazine, Sep. 1987.

[ITU-E600] ITU-T Recommendation E.600, "Terms and Definitions of

Traffic Engineering", Mar. 1993.

[ITU-E701] ITU-T Recommendation E.701, "Reference Connections for

Traffic Engineering", Oct. 1993.

[ITU-E801] ITU-T Recommendation E.801, "Framework for Service

Quality Agreement", Oct. 1996.

[JAM] Jamoussi, B., Editior, Andersson, L., Collon, R. and R.

Dantu, "Constraint-Based LSP Setup using LDP", RFC3212,

January 2002.

[KATZ] Katz, D., Yeung, D. and K. Kompella, "Traffic Engineering

Extensions to OSPF", Work in Progress, February 2001.

[LNO96] T. Lakshman, A. Neidhardt, and T. Ott, "The Drop from

Front Strategy in TCP over ATM and its Interworking with

other Control Features", Proc. INFOCOM'96, p. 1242-1250,

1996.

[MA] Q. Ma, "Quality of Service Routing in Integrated Services

Networks", PhD Dissertation, CMU-CS-98-138, CMU, 1998.

[MATE] A. Elwalid, C. Jin, S. Low, and I. Widjaja, "MATE: MPLS

Adaptive Traffic Engineering", Proc. INFOCOM'01, Apr.

2001.

[MCQ80] J.M. McQuillan, I. Richer, and E.C. Rosen, "The New

Routing Algorithm for the ARPANET", IEEE. Trans. on

Communications, vol. 28, no. 5, pp. 711-719, May 1980.

[MR99] D. Mitra and K.G. Ramakrishnan, "A Case Study of

Multiservice, Multipriority Traffic Engineering Design

for Data Networks", Proc. Globecom'99, Dec 1999.

[RFC-1458] Braudes, R. and S. Zabele, "Requirements for Multicast

Protocols", RFC1458, May 1993.

[RFC-1771] Rekhter, Y. and T. Li, "A Border Gateway Protocol 4

(BGP-4)", RFC1771, March 1995.

[RFC-1812] Baker, F., "Requirements for IP Version 4 Routers", STD

4, RFC1812, June 1995.

[RFC-1992] Castineyra, I., Chiappa, N. and M. Steenstrup, "The

Nimrod Routing Architecture", RFC1992, August 1996.

[RFC-1997] Chandra, R., Traina, P. and T. Li, "BGP Community

Attributes", RFC1997, August 1996.

[RFC-1998] Chen, E. and T. Bates, "An Application of the BGP

Community Attribute in Multi-home Routing", RFC1998,

August 1996.

[RFC-2205] Braden, R., Zhang, L., Berson, S., Herzog, S. and S.

Jamin, "Resource Reservation Protocol (RSVP) - Version 1

Functional Specification", RFC2205, September 1997.

[RFC-2211] Wroclawski, J., "Specification of the Controlled-Load

Network Element Service", RFC2211, September 1997.

[RFC-2212] Shenker, S., Partridge, C. and R. Guerin, "Specification

of Guaranteed Quality of Service", RFC2212, September

1997.

[RFC-2215] Shenker, S. and J. Wroclawski, "General Characterization

Parameters for Integrated Service Network Elements", RFC

2215, September 1997.

[RFC-2216] Shenker, S. and J. Wroclawski, "Network Element Service

Specification Template", RFC2216, September 1997.

[RFC-2328] Moy, J., "OSPF Version 2", STD 54, RFC2328, July 1997.

[RFC-2330] Paxson, V., Almes, G., Mahdavi, J. and M. Mathis,

"Framework for IP Performance Metrics", RFC2330, May

1998.

[RFC-2386] Crawley, E., Nair, R., Rajagopalan, B. and H. Sandick, "A

Framework for QoS-based Routing in the Internet", RFC

2386, August 1998.

[RFC-2474] Nichols, K., Blake, S., Baker, F. and D. Black,

"Definition of the Differentiated Services Field (DS

Field) in the IPv4 and IPv6 Headers", RFC2474, December

1998.

[RFC-2475] Blake, S., Black, D., Carlson, M., Davies, E., Wang, Z.

and W. Weiss, "An Architecture for Differentiated

Services", RFC2475, December 1998.

[RFC-2597] Heinanen, J., Baker, F., Weiss, W. and J. Wroclawski,

"Assured Forwarding PHB Group", RFC2597, June 1999.

[RFC-2678] Mahdavi, J. and V. Paxson, "IPPM Metrics for Measuring

Connectivity", RFC2678, September 1999.

[RFC-2679] Almes, G., Kalidindi, S. and M. Zekauskas, "A One-way

Delay Metric for IPPM", RFC2679, September 1999.

[RFC-2680] Almes, G., Kalidindi, S. and M. Zekauskas, "A One-way

Packet Loss Metric for IPPM", RFC2680, September 1999.

[RFC-2702] Awduche, D., Malcolm, J., Agogbua, J., O'Dell, M. and J.

McManus, "Requirements for Traffic Engineering over

MPLS", RFC2702, September 1999.

[RFC-2722] Brownlee, N., Mills, C. and G. Ruth, "Traffic Flow

Measurement: Architecture", RFC2722, October 1999.

[RFC-2753] Yavatkar, R., Pendarakis, D. and R. Guerin, "A Framework

for Policy-based Admission Control", RFC2753, January

2000.

[RFC-2961] Berger, L., Gan, D., Swallow, G., Pan, P., Tommasi, F.

and S. Molendini, "RSVP Refresh Overhead Reduction

Extensions", RFC2961, April 2000.

[RFC-2998] Bernet, Y., Ford, P., Yavatkar, R., Baker, F., Zhang, L.,

Speer, M., Braden, R., Davie, B., Wroclawski, J. and E.

Felstaine, "A Framework for Integrated Services Operation

over Diffserv Networks", RFC2998, November 2000.

[RFC-3031] Rosen, E., Viswanathan, A. and R. Callon, "Multiprotocol

Label Switching Architecture", RFC3031, January 2001.

[RFC-3086] Nichols, K. and B. Carpenter, "Definition of

Differentiated Services Per Domain Behaviors and Rules

for their Specification", RFC3086, April 2001.

[RFC-3124] Balakrishnan, H. and S. Seshan, "The Congestion Manager",

RFC3124, June 2001.

[RFC-3209] Awduche, D., Berger, L., Gan, D., Li, T., Srinivasan, V.

and G. Swallow, "RSVP-TE: Extensions to RSVP for LSP

Tunnels", RFC3209, December 2001.

[RFC-3210] Awduche, D., Hannan, A. and X. Xiao, "Applicability

Statement for Extensions to RSVP for LSP-Tunnels", RFC

3210, December 2001.

[RFC-3213] Ash, J., Girish, M., Gray, E., Jamoussi, B. and G.

Wright, "Applicability Statement for CR-LDP", RFC3213,

January 2002.

[RFC-3270] Le Faucheur, F., Wu, L., Davie, B., Davari, S., Vaahanen,

P., Krishnan, R., Cheval, P. and J. Heinanen, "Multi-

Protocol Label Switching (MPLS) Support of Differentiated

Services", RFC3270, April 2002.

[RR94] M.A. Rodrigues and K.G. Ramakrishnan, "Optimal Routing in

Shortest Path Networks", ITS'94, Rio de Janeiro, Brazil.

[SHAR] Sharma, V., Crane, B., Owens, K., Huang, C., Hellstrand,

F., Weil, J., Anderson, L., Jamoussi, B., Cain, B.,

Civanlar, S. and A. Chui, "Framework for MPLS Based

Recovery", Work in Progress.

[SLDC98] B. Suter, T. Lakshman, D. Stiliadis, and A. Choudhury,

"Design Considerations for Supporting TCP with Per-flow

Queueing", Proc. INFOCOM'98, p. 299-306, 1998.

[SMIT] Smit, H. and T. Li, "IS-IS extensions for Traffic

Engineering", Work in Progress.

[WANG] Y. Wang, Z. Wang, L. Zhang, "Internet traffic engineering

without full mesh overlaying", Proceedings of

INFOCOM'2001, April 2001.

[XIAO] X. Xiao, A. Hannan, B. Bailey, L. Ni, "Traffic

Engineering with MPLS in the Internet", IEEE Network

magazine, Mar. 2000.

[YARE95] C. Yang and A. Reddy, "A Taxonomy for Congestion Control

Algorithms in Packet Switching Networks", IEEE Network

Magazine, p. 34-45, 1995.

13.0 Authors' Addresses

Daniel O. Awduche

Movaz Networks

7926 Jones Branch Drive, Suite 615

McLean, VA 22102

Phone: 703-298-5291

EMail: awduche@movaz.com

Angela Chiu

Celion Networks

1 Sheila Dr., Suite 2

Tinton Falls, NJ 07724

Phone: 732-747-9987

EMail: angela.chiu@celion.com

Anwar Elwalid

Lucent Technologies

Murray Hill, NJ 07974

Phone: 908 582-7589

EMail: anwar@lucent.com

Indra Widjaja

Bell Labs, Lucent Technologies

600 Mountain Avenue

Murray Hill, NJ 07974

Phone: 908 582-0435

EMail: iwidjaja@research.bell-labs.com

XiPeng Xiao

Redback Networks

300 Holger Way

San Jose, CA 95134

Phone: 408-750-5217

EMail: xipeng@redback.com

14.0 Full Copyright Statement

Copyright (C) The Internet Society (2002). All Rights Reserved.

This document and translations of it may be copied and furnished to

others, and derivative works that comment on or otherwise explain it

or assist in its implementation may be prepared, copied, published

and distributed, in whole or in part, without restriction of any

kind, provided that the above copyright notice and this paragraph are

included on all such copies and derivative works. However, this

document itself may not be modified in any way, such as by removing

the copyright notice or references to the Internet Society or other

Internet organizations, except as needed for the purpose of

developing Internet standards in which case the procedures for

copyrights defined in the Internet Standards process must be

followed, or as required to translate it into languages other than

English.

The limited permissions granted above are perpetual and will not be

revoked by the Internet Society or its successors or assigns.

This document and the information contained herein is provided on an

"AS IS" basis and THE INTERNET SOCIETY AND THE INTERNET ENGINEERING

TASK FORCE DISCLAIMS ALL WARRANTIES, EXPRESS OR IMPLIED, INCLUDING

BUT NOT LIMITED TO ANY WARRANTY THAT THE USE OF THE INFORMATION

HEREIN WILL NOT INFRINGE ANY RIGHTS OR ANY IMPLIED WARRANTIES OF

MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE.

Acknowledgement

Funding for the RFCEditor function is currently provided by the

Internet Society.

 
 
 
免责声明:本文为网络用户发布,其观点仅代表作者个人观点,与本站无关,本站仅提供信息存储服务。文中陈述内容未经本站证实,其真实性、完整性、及时性本站不作任何保证或承诺,请读者仅作参考,并请自行核实相关内容。
2023年上半年GDP全球前十五强
 百态   2023-10-24
美众议院议长启动对拜登的弹劾调查
 百态   2023-09-13
上海、济南、武汉等多地出现不明坠落物
 探索   2023-09-06
印度或要将国名改为“巴拉特”
 百态   2023-09-06
男子为女友送行,买票不登机被捕
 百态   2023-08-20
手机地震预警功能怎么开?
 干货   2023-08-06
女子4年卖2套房花700多万做美容:不但没变美脸,面部还出现变形
 百态   2023-08-04
住户一楼被水淹 还冲来8头猪
 百态   2023-07-31
女子体内爬出大量瓜子状活虫
 百态   2023-07-25
地球连续35年收到神秘规律性信号,网友:不要回答!
 探索   2023-07-21
全球镓价格本周大涨27%
 探索   2023-07-09
钱都流向了那些不缺钱的人,苦都留给了能吃苦的人
 探索   2023-07-02
倩女手游刀客魅者强控制(强混乱强眩晕强睡眠)和对应控制抗性的关系
 百态   2020-08-20
美国5月9日最新疫情:美国确诊人数突破131万
 百态   2020-05-09
荷兰政府宣布将集体辞职
 干货   2020-04-30
倩女幽魂手游师徒任务情义春秋猜成语答案逍遥观:鹏程万里
 干货   2019-11-12
倩女幽魂手游师徒任务情义春秋猜成语答案神机营:射石饮羽
 干货   2019-11-12
倩女幽魂手游师徒任务情义春秋猜成语答案昆仑山:拔刀相助
 干货   2019-11-12
倩女幽魂手游师徒任务情义春秋猜成语答案天工阁:鬼斧神工
 干货   2019-11-12
倩女幽魂手游师徒任务情义春秋猜成语答案丝路古道:单枪匹马
 干货   2019-11-12
倩女幽魂手游师徒任务情义春秋猜成语答案镇郊荒野:与虎谋皮
 干货   2019-11-12
倩女幽魂手游师徒任务情义春秋猜成语答案镇郊荒野:李代桃僵
 干货   2019-11-12
倩女幽魂手游师徒任务情义春秋猜成语答案镇郊荒野:指鹿为马
 干货   2019-11-12
倩女幽魂手游师徒任务情义春秋猜成语答案金陵:小鸟依人
 干货   2019-11-12
倩女幽魂手游师徒任务情义春秋猜成语答案金陵:千金买邻
 干货   2019-11-12
 
推荐阅读
 
 
 
>>返回首頁<<
 
靜靜地坐在廢墟上,四周的荒凉一望無際,忽然覺得,淒涼也很美
© 2005- 王朝網路 版權所有