Mpls Gmpls

  • November 2019
  • PDF

This document was uploaded by user and they confirmed that they have the permission to share it. If you are author or own the copyright of this book, please report to us by using this DMCA report form. Report DMCA


Overview

Download & View Mpls Gmpls as PDF for free.

More details

  • Words: 6,184
  • Pages: 8
DE OLIVEIRA LAYOUT

7/3/07

11:46 AM

Page 38

Path-Computation-Element-Based Architecture for Interdomain MPLS/GMPLS Traffic Engineering: Overview and Performance Sukrit Dasgupta and Jaudelice C. de Oliveira, Drexel University Jean-Philippe Vasseur, Cisco Systems Abstract The Path Computation Element Working Group at the Internet Engineering Task Force is chartered to specify a PCE-based architecture for the path computation of interdomain MPLS- and GMPLS-based traffic engineered label switched paths. In this architecture, path computation does not occur at the head-end LSR, but on another path computation entity that may not be physically located on the headend LSR. This method is greatly different from the traditional “per-domain” approach to path computation. This article presents analysis and results that compare performance of the PCE architecture with the current state-of-the-art approach. Detailed simulations are undertaken on varied and realistic scenarios where preliminary results show several performance benefits from the deployment of PCE. To provide a complete overview of significant development taking place in this area, milestones and progress at the IETF PCE WG are also discussed.

W

ith ever increasing requirements posed by significant advances in networking, service providers may use traffic engineering (TE) techniques to efficiently manage resources and provide consistent quality of service (QoS). Traffic engineering in multiprotocol label switching (MPLS) and generalized MPLS (GMPLS) networks are fundamentally based on constraint-based path computation. Briefly, constraint-based path computation involves the pruning of links that do not satisfy constraints and subsequently using the shortest path algorithm on the resulting subgraph [1]. This process is simple and efficient when the path involves only one domain, but can potentially become severely resource heavy, complex, and inefficient when multiple domains are involved. To address this problem, the Path Computation Element (PCE) Working Group (WG) of the Internet Engineering Task Force (IETF) has been developing an architecture that will allow multidomain path computation to be simple and efficient. The PCE architecture introduces a special computational entity that will cooperate with similar entities to compute the best possible path through multiple domains. A PCE is a node that has special path computation ability and receives path computation requests from entities known as path computation clients (PCCs). The PCE holds limited routing information from other domains, allowing it to possibly compute better and shorter interdomain paths than those obtained using the traditional per-domain approach. Among other purposes, PCEs are also being advocated for CPU-intensive computations, minimal-cost- based TE-LSP placement, backup path computations, and bandwidth protection. Along with the process

38

0890-8044/07/$20.00 © 2007 IEEE

of identifying the requirements and development of the architecture accordingly, a plethora of work is underway at the PCE WG on the new communication protocols that will make this architecture work. This includes the development of new interPCE communication protocols and introducing extensions to existing underlying routing protocols. Request for Comments (RFC) 4655 [2] specifies a PCE-based architecture. RFC 4657 [3] covers PCE communication protocol generic requirements, and RFC 4674 [4] discusses the requirements for PCE discovery. Several WG IETF drafts are currently underway to define PCE communication protocol in different scenarios, PCE-based interlayer TE, protocol extensions for Open Shortest Path First (OSPF) and Intermediate System to Intermediate System (ISIS) for PCE discovery, PCC-PCE communication, and policyenabled path computation. As is evident, a thorough performance study to justify and quantify the motivation for this new architecture is required. As the architecture is in its infancy, a detailed analysis comparing existing approaches to a PCE-based approach has never been undertaken. This article has been written with two goals in mind. First, it presents a brief review of the significant developments that have taken place in the PCE WG since it was conceived in the IETF. Second, it identifies the most significant performance metrics, and then presents detailed simulation results and accompanying analysis to contrast the performance of the existing approach with that of the PCE. The rest of the article is organized as follows. Several scenarios that motivate the use of a PCE-based architecture are highlighted. The PCE architecture is described. The two path computation approaches to be compared, PCE-based and per

IEEE Network • July/August 2007

DE OLIVEIRA LAYOUT

7/3/07

11:46 AM

Page 39

domain, are discussed. Performance metrics for comparison are identified, while the complete simulation scenario is described and the results are analyzed. Finally, we conclude the article.

Motivation for PCE The development of the PCE architecture has been motivated by numerous path computation requirements and scenarios arising in present-day provider networks. This section presents some scenarios that are aided by the presence of PCE. Several other scenarios are discussed in [2], as well as these: • Limited visibility: There are several situations where the node responsible for path computation has limited visibility of the network topology to the destination. In such cases, it is not possible to guarantee that the optimal (shortest) path will be computed, or even that a viable path will be discovered except, possibly, through repeated trial and error using crankback [5] or other signaling extensions. • Processing overhead: There are many situations where the computation of a path may be highly CPU-intensive (computation of Steiner trees, multicriteria path computation, etc.). In these situations, it may not be possible or desirable for some routers to perform path computation because of the constraints on their CPUs. • Limited routing capability: It is common in legacy optical networks for the network elements not to have a control plane or routing capability. Such network elements only have a data plane and a management plane, and all crossconnections are made from the management plane. It is desirable in this case to run the path computation on the PCE, and to send the cross-connection commands to each node on the computed path. • Multilayer networks: A server-layer network of one switching capability may support multiple networks of another (more granular) switching capability. The different clientand server-layer networks may be considered distinct path computation regions within a PCE domain, so the PCE architecture is useful to allow path computation from one client-layer network region, across the server-layer network, to another client-layer network region.

PCE Architecture Depending on the requirements of a provider, the PCE architecture can be deployed based on a single composite-node external and dedicated entity, or a cooperative set of multiple entities. On a PCE, a traffic engineering database (TED) is maintained and updated by the underlying routing protocol. The TED is then used to compute paths based on requests received by a PCC. When the PCE exists as a composite PCE node, the path computation and path setup request take place on the same physical network element. In contrast, a PCE that is external to the entity requesting a path setup uses the TED present on it for path computation, eventually returning the computed path as a response. Multiple PCEs can be involved in path computation in different scenarios. Depending on how PCEs across different domains are configured, several PCEs may be involved in path computation. A partial path returned by one PCE may require another partial path computation by another PCE to complete the path. Inter-PCE communication can also result in the computation of a single path in a cooperative manner. Based on the architecture, a provider can deploy a centralized or distributed computation model. The centralized model involves a single PCE for all path computation requirements in its domain. The distributed model may involve several PCEs that cooperate and share the path computation requests.

IEEE Network • July/August 2007

Path Computation and Setup This section gives a brief overview of the two path computation methodologies, the per-domain and the PCE-based approach respectively. Related issues that arise during path setup are also highlighted. Both approaches use the Constrained Shortest Path First (CSPF) algorithm [1]. CSPF first prunes all the links in the topology that do not satisfy the constraint (available bandwidth, delay, etc.) and then runs Dijkstra’s Shortest Path algorithm on the resulting subgraph. The metric used for this computation is the weight assigned to the links. In the analysis undertaken for this article, size of the TE-label switched path (LSP) (bandwidth to be reserved on every link of the path) is the only constraint considered. It should be noted that the path computed by CSPF always depends on the current load conditions of the network. Rerouting of TE-LSPs due to link failures and setup of other TE-LSPs prior to the current setup request are some factors that can result in the computation of different paths for the same request. Under any circumstance, though, CSPF will always return the shortest path satisfying the constraint corresponding to the current loading in the network. Working of the two path computation techniques is explained in Figs. 1b and 1c, respectively, and using the simple network in Fig. 1a.

Backward Recursive PCE-Based Computation Path computation across domains is particularly difficult as the visibility of a head-end router is restricted only to the domain it lies in. Backward Recursive PCE-Based Computation (BRPC) utilizes multiple PCEs to compute the shortest interdomain constrained path along a determined sequence of domains. Inherent in the PCE architecture, this technique also preserves confidentiality across domains, an important requirement when domains are managed by different service providers [6]. The BRPC procedure is backward, as the path is computed in a reverse fashion from the destination area to the source area. It is recursive since the same basic sequence of steps is repeated for every intermediate domain lying between the source and destination. Every domain has certain “entry” boundary nodes (BNs) into the domain and “exit” BNs out of the domain (into other domains). The following notation is defined for a clear explanation of the BRPC procedure. The entry BN of domain i joins domains i – 1 and i. Let the source and destination domains be denoted 0 and n, respectively, and n – 1 intermediate domains 1, ⋅ ⋅ ⋅, n – 1. For k (i), and the kth exit domain i, the kth entry BN is denoted BNen k (i). BRPC uses the concept of a virtual BN is denoted BN ex shortest path tree (VSPT). VSPT(i) denotes the multipoint-topoint (MP2P) tree formed by the set of constrained shortest paths from the entry BNs of domain i to the destination router. Each link of tree VSPT(i) represents a shortest path. The BRPC procedure is as follows. For the destination domain n, VSPT(n) is computed from the list of shortest paths to all the entry BNs from the destination node. This is illustrated in Fig. 1b. VSPT(n) is then sent to the PCEs in the previous domain (domain n – 1 in this case) where it is concatenated to their TEDs (as links). This TED is then used by the PCEs in domain n – 1 to compute VSPT(n – 1). This sequence is repeated until the source domain and subsequently the headend router of the TE-LSP areiou reached, as shown in Fig. 1b.

Per-Domain Path Computation In contrast to the BRPC procedure, the per-domain path computation technique involves the computation of individual path segments in every intermediate domain without the sharing of any path information from other domains. The complete path for a TE-LSP is obtained by concatenating the path

39

DE OLIVEIRA LAYOUT

7/3/07

11:46 AM

Page 40

segments that are computed for every domain i. Figure 1c illustrates a simple scenario of per-domain path computation. The head-end router computes the first path segment by finding the shortest path to the nearest exit BN. The second path segment is then computed for the second domain by this BN to the next nearest exit BN. This method does not necessarily always yield the shortest path across domains. Several setup issues arise after a path has been computed and the headend router of the tunnel tries to setup reservation along the route using the Resource Reservation Protocol with Traffic Engineering extensions (RSVP-TE) [7]. They result from outdated TEDs on any or all of the routers involved in the path computation process. The TED is maintained on every TE enabled router and holds information about several attributes of the links in its domain of existence. Path computation based on CSPF uses the TED and results in a path that satisfies the constraints put on one or a combination of these attributes. If the TED on the router computing the path is outdated, a call admission control (CAC) failure occurs at the entry point router of the link corresponding to the outdated information. In such a situation, if the underlying routing protocol is OSPF with TE extensions (OSPF-TE), a link state advertisement (LSA) is flooded, or if it is IS-IS, a link state packet (LSP) is flooded. This flooding updates the TEDs of all the routers with the most updated information with respect to the concerned link and is restricted only to the domain in which the CAC failure occurred. The per-domain and PCE approaches proceed differently in this situation. Details of LSA flooding and their mechanisms are explained in great detail in [1].

capacity or highly utilized links. The maximum and average path costs are observed for each TE-LSP. The distributions for the maximum and average path costs are analyzed.

PCE-Based Approach — The BRPC method [6] is used to k (i) and the destination. compute the VSPT between BN en When a CAC failure occurs, a new VSPT for that domain is computed, and path setup is retried in the domain. The number of CAC failures that take place with the PCE approach is at most equal to that with the per-domain approach.

TE-LSPs/Bandwidth Setup Capacity — Due to the different path computation techniques, there is a significant difference in the amount of TE-LSPs/bandwidth that can be set up. This metric captures the difference in the number of TE-LSPs and corresponding bandwidth that can be set up using the two techniques. The traffic matrix is continuously scaled and stopped when the first TE-LSP cannot be set up for both methods. The difference in the scaling factor gives the extra bandwidth that can be set up using the corresponding path computation technique.

Per-Domain Approach — Here, a crankback [5] is used to report a setup failure to the router that is trying to set up the path. For example, without any loss of generality, let the CAC k (i) was setting up a failure take place in domain i when BNen j path to BN ex(i). Upon receiving the failure information, k (i) tries to find a new shortest path to the next closest BN en (j+1) BNex (i). This process continues until no path can be found k (i) to any BN (i). In this situation, the setup failure from BNen ex k (i) is sent upstream to BN k (i – 1) in information from BN en en k (i – 1) then selects the domain i – 1 using a crankback. BNen (j+1) next BNex (i – 1) to enter domain i. Information about setup failures can propagate using the crankback to the source domain containing the headend router. This cycle repeats until a path can be set up or no path is found after exhausting all possibilities. It should be noted that crankback signaling is associated only with the per-domain approach to path computation. Several issues related to CAC and crankback signaling may arise, and can be found in [8, references therein].

Performance Metrics This section discusses the metrics used to capture and compare the performance of the two approaches. Path Cost — As the two path computation techniques differ significantly, paths and their corresponding costs can be very different. In our analysis, path cost is the sum of link metric/weight assigned to the links that constitute the path. A higher path cost is undesired since it could be representative of a total higher delay or may signify the traversal of low-

40

CAC Failures — As discussed in the previous section, CAC failures occur due to two reasons: a midpoint router having an outdated TED, or when an entry BN to a domain does not have a path in the domain. Although the former situation is common to both the per-domain and PCE-based approaches, the latter occurs only in the per-domain approach. This metric captures the total number of CAC failures that occur during initial setup and reroute (on link failure) of a TE-LSP. The distribution across all TE-LSPs is analyzed. Crankback Signaling — When an entry BN fails to find a route in the corresponding domain, crankback signaling takes place in the case of per-domain path computation. As discussed in the previous section, a crankback signaling message propagates to the entry BN of the previous domain, and a new entry BN to the next domain is chosen, after which path computation takes place to find a path segment to the new entry BN of the next domain. This causes a significant delay in setup time. This metric captures the distribution of the number of crankbacks and the corresponding delay in setup time for a TE-LSP when using the per-domain path computation technique. The total delay arising from the crankback signaling is proportional to the costs of the links over which the signal travels (i.e., the path which is setup from the entry BN of a domain to its exit BN). Similar to above, the distribution across TE-LSPs is analyzed.

Failed TE-LSPs/Bandwidth on Link Failure — Link failures are induced in the network during the course of the simulations conducted. This metric captures the number of TE-LSPs and the corresponding bandwidth that failed to find a route when one or more links lying its path failed.

Results and Analysis Simulation Setup A very detailed simulator has been developed to replicate a real-life network scenario accurately. Following is the set of entities used in the simulation with a brief description of their behavior. Topology Description — To obtain meaningful results applicable to present-day provider topologies, simulations have been run on two realistic topologies representative of current service provider deployments. They consists of a large backbone area to which four smaller areas are connected. For the first topology, named MESH-CORE, a highly connected backbone was obtained from RocketFuel [9]. The second topology has a symmetrical backbone and is called SYM-CORE. The four connected smaller areas are obtained from [10]. Details of the topologies are shown in Table 1 along with their layout in Fig. 2. All TE-LSPs set up on this network have their sources and destinations in different areas, and all of them need to traverse the backbone network. Table 1 also

IEEE Network • July/August 2007

DE OLIVEIRA LAYOUT

7/3/07

11:46 AM

Page 41

Domain (0) 1

Domain (1) 1

1

Domain (2)

1

1

1

1 5

1

1

S

D

3 BN0ex(i)

1

1

1

1

BN1en(i)

1

1

(a)

9 S

6

6 D

D

D

1

4 Domain (0)

1

Domain (1)

Domain (2)

VSPT(1)

VSPT(2)

Direction of path computation c

Represents a cost of c units (b)

4

3

2

S

D

Domain (0)

Domain (1)

Domain (2)

Direction of path computation c

Represents a cost of c units (c)

■ Figure 1. Interdomain path computation: a) example network with associated link costs; b) VSPTs formed by the BRPC procedure and their costs; c) per-domain path computation and costs of the path segments. shows the number of TE-LSPs that have their sources in the corresponding areas along with their size distribution. Node Behavior — Every node in the topology represents a router that maintains state for all the TE-LSPs passing through it. Each node is a source for TE-LSPs to all the other nodes in the other areas. As in a real-life scenario, where routers boot up at random points in time, the nodes in the topologies also start sending traffic on the TE-LSPs originating from them at a random start time (to take into account the different boot-up times). All nodes are up within an hour of the start of simulation. All nodes maintain a TED, which is updated using LSA updates as outlined in [1]. The LSA updates are restricted only to the domain in which they originate. The nodes have a path computation engine that operates on the routing information in the TED. In the per-domain approach, routing information is

IEEE Network • July/August 2007

limited only to the domain of existence. In the PCE approach, routing information is passed on to the domains according to the methodology outlined in the BRPC procedure [6]. TE-LSP Setup —When a node boots up, it tries to set up all the TE-LSPs that originate from it in descending order of size. The network is dimensioned such that all TE-LSPs can find a path. Once set up, all TE-LSPs stay in the network for the complete duration of the simulation. The traffic matrix represents a full mesh where there is a TE-LSP between every pair of nodes that lie in different areas.

Inducing Failures For thorough performance analysis and comparison, link failures are induced in all the areas. Each link in a domain can fail independently with a mean failure time of 24 h and be

41

DE OLIVEIRA LAYOUT

7/3/07

11:46 AM

Page 42

1d

2d

6b 1b4b 12b 67 26 9b 7b 11b 38 5b 3b 31 10b 14 29 22 2b 8b 16 1512 53 76 ABR-80-BB 32 ABR-81-BB 40 5234 68 19 37 55 45 60 54 39 28 30 47 27 20 ABR-A1-BB 3626 49 24 51 43 41 27 13 6a ABR-A0-BB 75 9a 57 33 58 83 ABR-D1-BB 81 3a 48 66 7d 6d 80 60 1a ABR-A2-BB 46 8a ABR-D0-BB 6 70 2a 7a 9 2 54 1d 2d ABR-C0-BB 14a 10a 73 4d 6c 3d 74 ABR-C2-BB 13a 10c 71 61 11a 12a 18 17 72 ABR-C1-BB 3c 1c 9c 78 11 14c 11c 13c 2c 17c 23 8c 4c 16c 7c 12c 15c

9a

5a

2a

6d 7d

4a

6a

3d 4d

ABR-D0-BB 9

11

12

12c 8c 16c 13c 14c 17c

10a

11a

1

16 17 6

2a

ABR-A0-BB 13

5

4

2

ABR-B0-BB

3

ABR-B1BB 7b 5b 3b 1b

ABR-C2BB 4c ABR-C1BB ABR-C0BB 5c 1c 7c 11c 6c 15c 3c 10c 9c 2c

(a)

1a 14a

14 15

18 8

12a

3a

ABR-A1-BB 10

7

8a 13a

ABR-A2-BB

5d ABR-D1-BB

2b

8b 10b

6b

11b

9b 4b 12b

(b)

■ Figure 2. Topologies used for the simulations: a) MESH-CORE; b) SYM-CORE. restored with a mean restore time of 15 min. Both interfailure and interrestore times are uniformly distributed. When a link fails, an LSA is flooded by the adjacent routers of the link, and all the TEDs of the routers lying in the same domain of existence are updated. When a link on the path of a TE-LSP fails, reservation is removed along the path, and the head-end router tries to set up the TE-LSP around the failure. No attempt to re-optimize the path of a TE-LSP is made when a link is restored. The links that join two domains never fail. This step has been taken to concentrate only on how link failures within domains affect performance.

and results were collected after the expiry of a transient period. Figures 3 and 4 illustrate the behavior of the metrics for topologies MESH-CORE and SYM-CORE, respectively.

Path Cost — Figures 3a and 4a show the distribution of the average path cost of the TE-LSPs for MESH-CORE and SYM-CORE, respectively. During initial setup, roughly 40 percent of TE-LSPs for MESH-CORE and 70 percent of TELSPs for SYM-CORE have path costs greater with the perdomain approach (PD-Setup) than with the PCE approach (PCE-Setup). This is due to the ability of the BRPC procedure to select the shortest available paths that satisfy the conResults and Analysis straints. Since the per-domain approach to path computation is undertaken in stages where every entry BN to a domain Simulations were carried out on the two topologies previously computes the path in the corresponding domain, the most described. The results are presented and discussed in this secoptimal route is not always found. When failures start to take tion. In the figures, PD-Setup and PCE-Setup represent results place in the network, TE-LSPs are rerouted over different corresponding to the initial setting up of TE-LSPs on an empty paths resulting in path costs that are different from the initial network using the per-domain and PCE approach, respectively. costs. PD-Failure and PCE-Failure in Figs. 3a and 4a show Similarly, PD-Failure and PCE-Failure denote the results under the distribution of the average path costs that the TE-LSPs the link failure scenario. A period of one week was simulated have over the duration of the simulation with link failures occurring. Similarly, the average path Domain description TE-LSP size costs with the per-domain approach are much higher than with the PCE approach when link # of # of OC-48 OC-192 [0,20) [20,100] failures occur. Figures 3b and 4b show similar Domain name nodes links links links Mb/s Mb/s trends and present the maximum path costs for a TE-LSP for the two topologies. It can be seen D1 17 24 18 6 125 368 that with per-domain path computation, the maximum path costs are larger for 30 percent and 100 percent of the TE-LSPs for MESH-CORE D2 14 17 12 5 76 186 and SYM-CORE, respectively. D3

19

26

20

6

14

20

D4

9

12

9

3

7

18

MESH backbone

83

167

132

35

0

0

SYM backbone

29

37

26

11

0

0

■ Table 1. Details of all the areas used to create the two topologies.

42

Crankbacks/Setup Delay — Due to crankbacks that take place in the per-domain approach to path computation, TE-LSP setup time is significantly increased. This could lead to QoS requirements not being met, especially during failures when rerouting needs to be quick in order to keep traffic disruption to a minimum. Since crankbacks do not take place during path computation with a PCE, setup delays are significantly reduced. Fig-

IEEE Network • July/August 2007

DE OLIVEIRA LAYOUT

7/3/07

11:46 AM

Page 43

45

110 PD setup PCE setup PD failure PCE failure

PD setup PCE setup PD failure PCE failure

100 90 Maximum path cost

Average path cost

40 35 30 25

80 70 60 50 40

20 15

30

0

40 60 Percentage of TE-LSPs

20

20 0

100

80

0

20

40 60 Percentage of TE-LSPs

(a)

(b)

16

Crankbacks

12

300 PD setup PCE setup PD failure PCE failure

250 Crankback delay

14

100

80

10 8 6

PD setup PCE setup PD failure PCE failure

200 150 100

4 50

2 0 60

65

70

75 80 85 Percentage of TE-LSPs

90

95

0 60

100

65

70

75 80 85 Percentage of TE-LSPs

(c)

16

PD failure PCE failure

400

1600

PD setup PCE setup PD failure PCE failure

8 6

1200 1000 800 600

4

250 200 150 100

400 2

50

200 65

70

75 80 85 Percentage of TE-LSPs (e)

90

95

100

PD failure PCE failure

300 Number of TE-LSPs

10

0 60

100

350

1400 Bandwidth (Mb/s)

CAC failures

12

95

(d)

1800

14

90

0

0 Simulation runs

Simulation runs (f)

■ Figure 3. Results for MESH-CORE path cost: a) distribution of average path costs across TELSPs; b) distribution of maximum path costs across TE-LSPs; c) distribution of number of crankbacks across TE-LSPs; d) distribution of proportional setup delay due to crankback across TE-LSPs; e) distribution of CAC failures across TE-LSPs; f) TE-LSPs and corresponding bandwidth that failed to find a route.

IEEE Network • July/August 2007

43

DE OLIVEIRA LAYOUT

7/3/07

11:46 AM

Page 44

ures 3c and 4c show the distributions of the number of crankbacks that took place during the setup of the corresponding TE-LSPs for MESH-CORE and SYM-CORE, respectively. It can be seen that all crankbacks occurred when failures were taking place in the networks. Information regarding failures never propagate across domains since the corresponding LSAs are restricted only to the domain in which they originate. As a result, BNs of one domain do not have information on a failure in the next domain, leading to a crankback when a route cannot be found in the next domain. Figures 3d and 4d illustrate the proportional setup delays experienced by the TE-LSPs due to crankbacks for the two topologies. It can be observed that for a large proportion of the TE-LSPs, the setup delays arising out of crankbacks are very large, possibly proving to be very detrimental to QoS requirements. The large delays arise out of the crankback signaling that needs to propagate back and forth from the exit BN of a domain to its entry BN. More crankbacks occur for SYM-CORE than for MESH-CORE as it is a very restricted and constrained network in terms of connectivity. This causes a lack of routes, and often several cycles of crankback signaling are required to find a route. CAC Failures — As discussed in the previous sections, CAC failures occur either due to an outdated TED or when a route cannot be found from the selected entry BN. Figures 3e and 4e show the distribution of the total number of CAC failures experienced by the TE-LSPs during setup. About 38 percent and 55 percent of TE-LSPs for MESH-CORE and SYMCORE, respectively, experience CAC failures with perdomain path computation when link failures take place in the network. In contrast, only about 3 percent of the TE-LSPs experience CAC failures with the PCE method. It should be noted that the CAC failures experienced with the PCE correspond only to the TEDs being out of date. This is because with a PCE deployment, a BN that does not have a route to the destination is never selected by the BRPC procedure. Failed TE-LSPs/Bandwidth on Link Failures — Figures 4f and 3f show the number of TE-LSPs and the associated required bandwidth that fail to find a route when link failures are taking place in the topologies. For MESH-CORE, with the per-domain approach, 395 TE-LSPs failed to find a path corresponding to 1612 Mb/s of bandwidth. For PCE, this number is lower at 374 corresponding to 1546 Mb/s of bandwidth. For SYM-CORE, with the per-domain approach, 434 TE-LSPs fail to find a route corresponding to 1893 Mb/s of bandwidth. With the PCE approach, only 192 TE-LSPs fail to find a route, corresponding to 895 Mb/s of bandwidth. It is clearly visible that the PCE allows more TE-LSPs to find a route, thus leading to better performance during link failures. This improvement in performance arises due to the difference in path computation procedures too. BRPC always includes all links that satisfy the constraint in the VSPT and builds the shortest path from the destination. Since the per-domain approach does not have all information about links in other domains, after several TE-LSPs are set up, available bandwidth is left over in fragments insufficient to route all requests, thereby resulting in several setup failures. TE-LSP/Bandwidth Setup Capacity — Since PCE and perdomain path computation differ in how path computation takes place, more bandwidth can be set up with PCE. This is primarily due to the way in which BRPC functions. To observe the extra bandwidth that can fit into the network, the traffic matrix was scaled. Scaling was stopped when the first TE-LSP failed to set up with PCE. This metric, like all the others discussed above, is topology-dependent (hence the choice of two topologies for this study). This metric highlights the ability of

44

PCE to fit more bandwidth in the network. For MESHCORE, on scaling, 1556 Mb/s more could be set up with PCE. In comparison, for SYM-CORE this value is 986 Mb/s. The amount of extra bandwidth that can be set up on SYM-CORE is less due to its restricted nature and limited capacity.

Conclusion and Future Directions In this article a performance comparison with respect to several critical metrics between a PCE-based and a per-domain path computation deployment was presented. The PCE-based approach showed better results than the per-domain path computation approach with the selected performance metrics. Keeping in mind that all evaluation metrics are topology- and traffic-dependent, this study used two significantly different but representative topologies where the difference in performance improvement was also illustrated. Although many more nuances are left to be uncovered and analyzed in future work, novel preliminary analysis and insight into factors governing interdomain path computation are brought to light. An overview of the current status of the IETF PCE Working Group was also given. The WG is currently refining drafts, as well as awaiting deployment experiences. This should bring even further insights on the benefits of the PCE architecture for current network systems.

References [1] E. Osborne and A. Simha, Traffic Engineering with MPLS, Cisco Press, 2002. [2] A. Farrel, J.-P. Vasseur, and J. Ash, “A Path Computation Element (PCE) Based Architecture,” IETF RFC 4655, Aug. 2006. [3] J. Ash and J. L. Le Roux, “A Path Computation Element (PCE) Communication Protocol Generic Requirements,” IETF RFC 4657, Sept. 2006. [4] J. L. Le Roux, “Requirements for Path Computation Element (PCE) Discovery,” IETF RFC 4674, Oct. 2006. [5] A. Farrel et al., “Crankback Signaling Extensions for MPLS and GMPLS RSVP-TE,” Internet draft, v. 05, Nov. 2005, work in progress. [6] J.-P. Vasseur et al., “A Backward Recursive PCE-Based Computation (BRPC) Procedure to Compute Shortest Interdomain Traffic Engineering Label Switched Paths,” Internet draft, v. 01, Dec. 2006, work in progress. [7] D. Awduche et al., “RSVP-TE: Extensions to RSVP for LSP Tunnels,” IETF RFC 3209, Dec. 2001 [8] D. Medhi, “Quality of Service Routing Computation with Path Caching: A Framework and Network Performance,” IEEE Commun. Mag., vol. 40, no. 12, June 2002. [9] N. Spring, R. Mahajan, and D. Wetherall, “Measuring ISP Topologies with Rocketfuel,” Proc. ACM SIGCOMM, 2002. [10] J. Guichard, F. Le Faucheur, and J.-P. Vasseur, Definitive MPLS Network Designs, Cisco Press, 2005.

Biographies SUKRIT DASGUPTA ([email protected]) is currently a Ph.D. candidate at Drexel University. He received his B.Tech. in computer engineering from Sikkim Manipal Institute of Technology, India, in 2003 and M.S. in computer engineering from Drexel University in 2006. His interests lie in several areas of traffic engineering including resource allocation, path computation, signaling, and protocol analysis. JAUDELICE CAVALCANTE DE OLIVEIRA ([email protected]) received her B.S.E.E. degree from Universidade Federal do Ceara, Ceara, Brazil, in December 1995. She received her M.S.E.E. degree from Universidade Estadual de Campinas, Sao Paulo, Brazil, in February 1998, and her Ph.D. degree in electrical and computer engineering from the Georgia Institute of Technology in May 2003. She joined Drexel University in July of 2003 as an assistant professor. Her research interests include the development of new protocols and policies to support finegrained quality of service provisioning in the future Internet, researching and developing traffic engineering strategies, and the design of solutions for efficient routing in ad hoc and sensor networks. JEAN-PHILIPPE VASSEUR ([email protected]) is a system architect at Cisco Systems where he works on IP/MPLS architecture sepcifications, focusing on IP, TE, and network recovery. He holds an engineering degree from France and an M.S. from Stevens Institute of Technology, New Jersey. Before joining Cisco, he worked for several service providers in large multiprotocol environments. He is an active member of the IETF, co-chair of the IETF PCE Working Group, and has co-authored several IETF specifications. He is a regular speaker at various international conference and is involved in various research projects in the area of IP and MPLS. He has also filed several patents in the areas of IP and MPLS, and is the co-author of Network Recovery.

IEEE Network • July/August 2007

DE OLIVEIRA LAYOUT

7/3/07

11:46 AM

Page 45

50

60 PD setup PCE setup PD failure PCE failure

PD setup PCE setup PD failure PCE failure

55 50

40 Max path cost

Average path cost

45

35 30 25

45 40 35 30 25

20 15

20 0

20

40 60 Percentage of TE-LSPs

80

15

100

0

20

40 60 Percentage of TE-LSPs

(a)

(b)

140

8

Crankbacks

6

PD setup PCE setup PD failure PCE failure

120

Crankback delay

7

5 4 3

PD setup PCE setup PD failure PCE failure

100 80 60 40

2

20

1 0 40

50

60 70 80 Percentage of TE-LSPs

90

0 40

100

50

60 70 80 Percentage of TE-LSPs

(c)

PD setup PCE setup PD failure PCE failure

100

4 3

350

1400

Number of TE-LSPs

5

PD failure PCE failure

400

1600 Bandwidth (Mb/s)

CAC failures

450

PD failure PCE failure

1800

8

6

90

(d)

2000

7

100

80

1200 1000 800

300 250 200

600

150

400

100

200

50

2 1 0 40

50

60 70 80 Percentage of TE-LSPs (e)

90

100

0

0 Simulation runs

Simulation runs (f)

■ Figure 4. Results for SYM-CORE: a) distribution of average path costs across TELSPs; b) distribution of maximum path costs across TE-LSPs; c) distribution of number of crankbacks across TE-LSPs; d) distribution of proportional setup delay due to crankback across TE-LSPs; e) distribution of CAC failures across TE-LSPs; f) TE-LSPs and corresponding bandwidth that failed to find a route.

IEEE Network • July/August 2007

45

Related Documents

Mpls Gmpls
November 2019 12
Mpls
November 2019 15
Mpls Overview
May 2020 5
Mpls Presentation
May 2020 5
Mpls Overview
November 2019 17
Introduction To Gmpls
July 2020 3