Sie sind auf Seite 1von 35

IMPLEMENTATION GUIDE

INter Data CeNter WorkloaD MobIlIty WIth VMWare

although Juniper Networks has attempted to provide accurate information in this guide, Juniper Networks does not warrant or guarantee the accuracy of the information provided herein. third party product descriptions and related technical details provided in this document are for information purposes only and such products are not supported by Juniper Networks. all information provided in this guide is provided as is, with all faults, and without warranty of any kind, either expressed or implied or statutory. Juniper Networks and its suppliers hereby disclaim all warranties related to this guide and the information contained herein, whether expressed or implied of statutory including, without limitation, those of merchantability, fitness for a particular purpose and noninfringement, or arising from a course of dealing, usage, or trade practice.

Copyright 2013, Juniper Networks, Inc.

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

Table of Contents
Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .3 Scope . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .3 Design Considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 Distance and latency . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 l2 Domain Considerations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 Congestion Point . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 alternative approaches . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4 VMware vMotion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 Network Implications . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 Maintaining State for traffic Flows . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .7 Implementation overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8 Metro vMotion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8 optimizing traffic Flows after Migration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11 Implementation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12 QFX3600-QFX003 in DC1 Configuration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13 eX4200-VC in DC2 Configuration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .14 MX960-MX13 in DC1 Configuration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 15 MX960-MX14 in DC1 Configuration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .19 Validation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30 Validation Using the vCenter Client application . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34 Summary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35 about Juniper Networks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35

List of Figures
Figure 1: Minimum required networks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6 Figure 2: redundant virtual and physical network connectivity design . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .7 Figure 3: three-tier traffic flow . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8 Figure 4: three-tier traffic flow within the same physical host . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8 Figure 5: traffic flow before metro vMotion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9 Figure 6: traffic flow during metro vMotion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9 Figure 7: traffic flow after metro vMotion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .10 Figure 8: Data center workload mobility, network configuration . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12 Figure 9: Ping traffic from client to server . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34 Figure 10: time and success of the vMotion event . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35

Copyright 2013, Juniper Networks, Inc.

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

Introduction
Implementing server virtualization results in higher utilization and reduced capital expenditure on physical servers. Server virtualization abstracts the server operating system from the server hardware, allowing physical resources to be shared among many virtual servers or virtual machines (VMs). the basic features of cloning, suspending, and migrating live VMs among a cluster of physical hosts enable new applications to be built on top of these features including applications that migrate VMs based on resource availability. VM migration technology was originally designed to accommodate a cluster of servers located near each other. however, advancements in networking technology now allow servers within a single cluster to be located across the entire data center, in an adjacent building or across the city.

Scope
this document presents design considerations and configurations for server and virtualization administrators, security and network administrators, and designers and architects who want to extend their VM resource clusters beyond a set of racks in the data center (the original design purpose) and migrate them to additional data centers. VMware vSphere, including its vMotion function, is a dominant player in the enterprise x86 server virtualization market. Juniper has used this product to verify the network design. the concepts discussed here can also be applied to other server virtualization products from IbM, Microsoft, Citrix, and others. Use Cases enterprises are leveraging live workload migration in several ways. Some use cases include: Optimizing resource utilization. the most common use case is optimizing resource utilization within a single data center. there are several reasons for this: - If a VM requires more compute resources than provided by the server, it can be migrated to another server in the cluster with available resources or migrate nonprioritized VMs to make room for the resources needed. - If fewer servers can serve the total resources required by applications in the cluster, VMs can be migrated to a smaller group of servers. Shutting down excess servers saves power and cooling costs. - If technical server problems occur or if the server is to be brought offline for maintenance purposes, the VMs can be migrated to another server. Campus migration. VMs are migrated from one data center to another for resource consumption purposes. this scenario is similar to the above use case but instead of the cluster residing in one data center, it is distributed over a short distance (multiple kilometers or miles) to different buildings within the organization. typically, the campus owner owns the backbone network and therefore has control over bandwidth and latency resources for the backbone. because only one ingress and egress point exists in the network for external traffic, the problem of route optimization is not a limiting factor for VM migration compared to data centers that are separated by a larger geographical distance and have their own connections to multiple service providers and the Internet. Hybrid cloud. the migration of VMs between private data centers and a private or public cloud service is called the hybrid cloud approach (or cloud bursting). Workloads in VMs are burst from the enterprise data center to the service providers data center when they exceed available resources in their original data center. this alternative works well for special projects with large, unpredictable workloads. typically, the enterprise forms a contract with the service provider to handle this type of traffic and possibly a VPN if traffic must be isolated from the Internet. Disaster avoidance. administrators can nondisruptively migrate machines to another data center before a natural disaster strikes, preventing loss of data and operation time. If a disaster is expected, such as a severe storm, an administrator can migrate VMs to the backup data center while still operational. to do so, administrators must reserve bandwidth on a high-bandwidth, low-latency link to handle traffic when live migration starts. It is difficult to estimate how much bandwidth is required and how long the migration can take because users are updating VMs constantly, and live migration must continue sending the VM state updates. It is expensive to have the bandwidth available just in case it is needed, so an easier and less expensive alternative is to implement a backup plan of shutting down VMs, copying, and restarting. this approach reduces the bandwidth for VMs that are not operating and the time required to transfer them. Data from VMs can be backed up as part of a scheduled plan. Follow the sun. Moving users resources (VMs) to an advantageous location as the workday progresses can lower resource costs. Workloads are transferred from data center to data center based on the time of day.

Copyright 2013, Juniper Networks, Inc.

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

Design Considerations
Distance and Latency
low latency is a requirement for clustering VMs because of the sensitivity of the heartbeat that detects a failure in the system. In a vMotion event, latency is also important because the vMotion migration process is a chatty communications stream that requires synchronization of memory pages between two VMs. the synchronization ensures that the moved VM starts in exactly the same state and time as when the initial VM stopped. as distance increases, so does latency between these two VMs, causing reduced effective bandwidth. If the VM cannot receive replies or acknowledgements, it cannot transmit its state, and memory synchronization cannot occur. VMware imposes a minimum of 250 Mbps on throughput for vMotion and a maximum of 5-ms latency in the standard version, with 10 ms of latency for versions using the Metro vMotion feature. VMware recommends a maximum of 5 ms of latency and a 200 km distance between users and the virtualized applications. the two main factors that affect latency in data center networks and interconnections are propagation delay (caused by distance) and processing delay. Serialization delay is less important because of the size and amount of bandwidth required in these use cases. electronic manipulation of the packet causes processing delay, so any switching, routing, firewalling, or load balancing causes delay. Most types of delays are measured in microseconds (s, which is 1/1000th of a millisecond). the propagation delay is a function of the speed of light, and a conservative estimate of delay over fiber optics is approximately 5 s per km. one way to decipher the delay budget of 5 ms is to look at the distance latency measured purely as propagation delay. the propagation delay over fiber with a 5-ms budget is 1,000 km. For round-trip packets, divide by two to get the 5-ms round-trip time (rtt) for 500 km. Processing delay is unique for each network and varies by device based on software, hardware, functions, features, and the age of the technology. you can assume the following requirements: 5 to 20 s per switch; 20 to 100 microseconds per router; and 20 to 300 s per firewall. a rule of thumb states a minimum of two switches, a single router and one firewall per data center are required. In a worst case scenario and only allowing a unidirectional path, processing delay totals 80 s in switch delay, 200 s in router delay, and 600 s in firewall delay. bidirectionally, this totals a 1760-s delay of the 5,000-s delay budget, leaving approximately 3,000 s or about 600-km worth of propagation delay, 300 km bidirectionally. these figures do not include packet size, serialization delay, and queuing delay. It is safe to assume that a distance below 300 km provides a safer latency target, because it allows more devices to be in the path which cause further processing delays. In addition, several devices might be implemented in the software, which traditionally incurs higher processing delays. another point to note is that the distance the fiber route travels is often not a direct line between two points. obtaining the real distance that the circuit travels is important. Different storage vendors synchronized storage solutions often recommend distances below 100 km.

L2 Domain Considerations
Stretching an l2 domainextending the geographical distancetypically by interconnecting data centers, increases the number of devices within the domain. the domain is still susceptible to misconfigured, maliciously configured, or faulty devices. the technology commonly used to provide large-scale l2 laN extension (VPlS) includes mechanisms to rate-limit this type of traffic that floods across a MaN or WaN. Maliciously configured devices can be detected and halted by intrusion prevention systems (IPS) and firewall services.

Congestion Point
When referring to data center architecture, the bandwidth congestion point is at the MaN or WaN edge for the DCI (Data Center Interconnect), because this is the most expensive link in the network. It is important to have class of service (CoS) enabled here to ensure that important latency- and bandwidth-sensitive traffic is serviced in a predictable and reliable manner. you need to ensure that vMotion traffic has a higher priority than generic traffic. If administrators do not prioritize the traffic, additional delay might affect vMotion. this type of delay is called queuing delay and occurs at each point in the network where congestion exists.

Alternative Approaches
the following alternative methods might meet the needs without requiring live workload migration.

Routed Live Migration


as an alternative to l2 stretch, you can use an l3 live migration, which many hypervisor vendors support, such as Microsoft, kVM, and Citrix. this method allows administrators to change IP and MaC addresses and use Dynamic host Configuration Protocol/Domain Name System (DhCP/DNS) updates to force quicker changes. Implementations can also use session tunneling to keep existing sessions alive, if required.

Copyright 2013, Juniper Networks, Inc.

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

Cold Migration
Migrating VMs live and maintaining sessions creates complexity in the network, as previously discussed. If live migration is not a requirement, cold migration is a good option. With a cold migration, VMs are shut down, moved, and restarted. Many organizations find that this is an excellent way to share work with global teams.

VM Migration Using Load Balancing


administrators can create or clone a new VM in the second data center, and new traffic is then forwarded to the new VM through load balancing. the original VM stops operating when all traffic has been offloaded. this is a common and well-established method of scaling traffic that does not have location or IP restrictions. however, it does not specifically move a live server with live sessions from one location to another.

VMware vMotion
the implementation described in this document uses VMware vMotion to provide live workload migration. live migration is one of the key capabilities that caused users to switch to virtualization. live migration enables many useful applications, such as Distributed resource Scheduler (DrS), Distributed Power Management (DPM), and Fault tolerance. vMotion transfers the running contents of a VMs memory to another host. the vMotion component of vSphere version 5 does the following: 1. 2. Creates a shadow VM on the destination host. Places traces on all the VMs memory.

3. Copies memory pages from the source VM host to the destination VM host. 4. Copies memory pages that have changed during the first copy. 5. Continues this process iteratively until a threshold is met where nearly all memory pages have been transferred. 6. Stops (quiesces) the source VM. 7. Copies the state of the VMs CPU, video, network, disk, and storage.

8. Copies final memory page changes, and resumes operation on the destination hosts VM. 9. Sends reverse address resolution Protocol (arP) to the network to ensure that traffic is forwarded to its location via the destination host VM. two additional points worth noting are: vMotion is not by itself a high-availability mechanism. vSphere has a feature called high availability that can restart VMs based on VM or host failures or through an additional product called Site recovery Manager (SrM). vMotion is designed to migrate a VM and all of its current execution state from one host to another. From a network perspective, you want to maintain all active sessions between the VM and other hosts on the network, as well as inherit the source VMs IP and media access control (MaC) addresses on the destination hosts. What else can be derived from the network and storage requirements from the vMotion process? as the VM execution state is being copied, the physical hosts must both be connected to the same network and that network must have sufficient bandwidth to transfer this memory state. the latency must be minimal so that state synchronization can occur. If there is insufficient bandwidth or too much latency, memory state synchronization does not occur, and the vMotion fails. In vSphere version 5, the minimum support bandwidth required is 250 Mbps, and the latency is 5 ms. In addition, vSphere version 5 has a higher licensing feature called Metro vMotion. this feature allows the latency requirement to be extended to 10 ms. VMware creates a minimum of four concurrent vMotions per 1G vMotion network and eight per 10G vMotion network. based on 250 Mbps of available bandwidth, which is the minimum requirement per vMotion, four concurrent live vMotions in a 1G network totally consume this 1G of bandwidth. another key requirement is that the VM and both physical hosts must have access to the same storage location. Performance requirements can be derived from these connectivity requirements. Performance requirements depend on the type of workload that the VM is running. a web server has different requirements compared with an e-mail server or a database server. therefore, from a memory synchronization perspective and a storage access point perspective, each server type forces different demands on the network.

Copyright 2013, Juniper Networks, Inc.

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

Network Implications
From a network perspective, a key requirement concerning vMotion, cloning, and suspending VMs is that the source and destination VM are the same. the VMs take the same virtualized network characteristics, unless manually or scripted to change on the cloned VMs. For migrating VMs, the destination VM takes the same virtual MaC and IP addresses as the source VM when migrated. therefore, when migrating machines, administrators must have that the same or similar virtual networking (vSwitch) and physical networking (switching, IP addressing, and routing) components in place to ensure connectivity to the other devices that communicate with the migrated VM. Figure 1 shows the minimum required networks when using VMware vMotion.

Public Network

VM VM VM

VM VM VM

VM VM VM

VM VM VM

Management Network

Live Migration Network


Figure 1: Minimum required networks Figure 1 shows that each VM typically connects to the public network. Using load balancing, each host or vSwitch can connect to multiple physical links, either for redundancy or extra bandwidth. VMware recommends a separate, physical network for VMkernel management (connection to vCenter) per eSXi host, and a separate network for vMotion per eSXi host. It is common practice to have redundant or additional connections for public or production traffic. It is likely that six physical connections from a physical host with VMs may be used, if not more. VMware is capable of segmenting all these networks on less links by using port groups or VlaNs. VMware and other hypervisor platforms use a construct called a vSwitch, which enables communication between VMs on a host without using the physical network. a vSwitch shares the physical ports between the guest VMs, management ports, shared storage networks, and vMotion or Fault tolerance ports. VMs connect to the switches using vNICs, and all traffic outside the host traverses the vSwitch. Multiple vSwitches can be used within a physical host to ensure isolation. the vSwitches in VMwares basic licensing form are fairly simple and do not run many protocols, such as Spanning tree Protocol (StP), but do implement forwarding rules to ensure that traffic is not looped or flooded unnecessarily on vSwitches using port groups and VlaNs. Figure 2 shows a redundant virtual and physical network connectivity design in which management and storage services connect to the eSXi host and VMs connect to the network.

Copyright 2013, Juniper Networks, Inc.

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

QFX Series or EX Series

QFX Series or EX Series

QFX Series or EX Series

QFX Series or EX Series

Physical Ports vSwitches or Port Groups

ESXi Host
Virtual NICs Server VM Serviced VM Service VM (Network Virtualization, Load Balancing, NAT, Security) IP Storage VMK NIC Interface

vMotion/FT VMK NIC Interface

ESXi Host Management VMK NIC Interface

Non-Virtualized

Virtualized

Figure 2: Redundant virtual and physical network connectivity design the vMotion process and vSwitch have the following network requirements: l2 connections between the hosts on the vMotion network Same l2 public or data network present in both data centers to ensure that network sessions are not interrupted Shared storage (replicated in the MaN to avoid DCI usage) VlaN trunking to the server if multiple VlaNs per port exist Same port groups and VlaNs for all vMotion-capable servers in an eSXi cluster When network changes are made, configuration of the server, vSwitch, switches, and routers less than 510 ms latency (less latency is better; deterministic latency is best) Minimum of 250 Mbps bandwidth link per vMotion (more bandwidth is better) Network reliability and redundancy

Maintaining State for Traffic Flows


When migrating VMs with vMotion, administrators must consider that the destination host VM assumes all the characteristics of the source VM, and the IP and MaC addresses move to a new location in the network. If relocation causes the VM to shift from one stateful service device (load balancer or firewall) to another, it is likely that the new device will not recognize the network session state of applications running on the VM and block these network communications, as they are designed to do. this causes the application to reset the session and, as a result, shows as a non-seamless VM move. to avoid this, we implement a design that forces all existing sessions through the original stateful service device, resulting in non-optimal traffic routing and bandwidth usage for the DCI. administrators can use alternate designs that maintain session state between the stateful service devices of the source VMs host and destination VMs host. typically, clustering resolves this alternate option. Clustering is another technology that typically works in laN environments or with devices located next to each other. Using clustered stateful devices in a WaN environment might not be optimal or might create similar problems compared to vMotion. In addition, clustered stateful devices have bandwidth and latency requirements, which might not be compatible with vMotion requirements.

Copyright 2013, Juniper Networks, Inc.

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

Implementation Overview
When migrating VMs between hosts, it is helpful to map the communication flows between applications that traverse the network. this mapping aids in understanding potential communication problems and how specific configurations can ensure that the vMotion event occurs successfully and all communication flows are maintained. Figure 3 shows external traffic arriving from outside the data center to a traditional three-tier application and processed traffic exiting. each VM resides on a separate physical eSXi host. this traffic flow does not include load balancers or the potential for multiple applications and database servers that might be required to provide all the data processing for a client entering the web tier.

Tra c Flow
FW LAN Web VM LAN App VM LAN DB VM LAN App VM LAN Web VM LAN FW

Figure 3: Three-tier traffic flow When migrating individual VMs, it is important to plan the outcome. Is the same traffic flow achievable? Will some VMs reside in one data center and others in another, thereby splitting or sharing the IP address range? are there additional stateful devices, such as load balancers and firewalls that can disrupt the traffic flow? Is there sufficient bandwidth to service traffic between the VMs if they reside in different data centers? Do the application communications between the VMs have strict latency requirements? the traffic flow depicted in Figure 4 does not include stateful hosts that might be located in the path.

Tra c Flow
FW LAN Virtual Switch Web VM Virtual Switch App VM Virtual Switch DB VM Virtual Switch App VM Virtual Switch Web VM Virtual Switch LAN FW

Workload-to-workload tra c ows over virtual connections within the same server

Figure 4: Three-tier traffic flow within the same physical host the traffic flow in Figures 3 and 4 are similar, except Figure 4 shows all the application VMs that reside on the same server. how will the traffic flow change when migrating one of the VMs out of the host to another data center? applications have different structures compared with those shown in Figures 3 and 4. applications do not need to leave the subnet, VlaN, or data center to an external client. administrators should also consider how to accommodate these applications and their traffic flows when planning to migrate VMs between data centers.

Metro vMotion
Figures 5 through 7 show traffic flows of a monolithic applicationall application functions reside in one VM. If the application flows are plotted out in the same way for a three-tier application, many more flows would crisscross the data centers. load balancers and other functions are not depicted, but they might be required in a typical application network call flow. Many detailed traffic flows involve more components than just a simple three-tier application. Many different application and database servers might need to communicate with each other before delivering an answer to a client. In addition, load balancers and firewalls might exist between each application tier. these flows also require consideration when analyzing a Metro vMotion and the dependencies that these types of traffic flows might warrant.

Copyright 2013, Juniper Networks, Inc.

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

Outside Client for Remote Connection Client Router Testing MX Series Single VPLS all VLANs or Multiple VPLS for load-balancing and tra c separation

DC2 PE MX Series

DC1 PE MX Series MC LAG, LAG and Virtual Chassis technologies used to improve performance and reliability

VirtualChassis

Aggregation Switch EX Series

SRX Series Cluster QFX Series

SRX Series Cluster


VirtualChassis

EX Series

vCenter ESXi Host with vGW

NIC teaming to alternate switches for production and vMotion VLANs. Multiple links each for production, storage and vMotion/management

ESXi Host with vGW

DATA CENTER 1

DATA CENTER 2

Client tra c to VM vCenter tra c to ESXi host Storage tra c to VM/ESXi

Figure 5: Traffic flow before metro vMotion


Outside Client for Remote Connection Client Router Testing MX Series Single VPLS all VLANs or Multiple VPLS for load-balancing and tra c separation

DC2 PE MX Series

DC1 PE MX Series MC LAG, LAG and Virtual Chassis technologies used to improve performance and reliability

VirtualChassis

Aggregation Switch EX Series

SRX Series Cluster QFX Series

SRX Series Cluster


VirtualChassis

EX Series

vCenter ESXi Host with vGW

NIC teaming to alternate switches for production and vMotion VLANs. Multiple links each for production, storage and vMotion/management

ESXi Host with vGW

DATA CENTER 1

DATA CENTER 2
Client tra c to VM vCenter tra c to ESXi host Storage tra c to VM/ESXi vMotion tra c ESXi host to host

Figure 6: Traffic flow during metro vMotion

Copyright 2013, Juniper Networks, Inc.

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

Figure 7 shows that after a vMotion migration, traffic continues to flow across the data center interconnect. In this simple monolithic application case, traffic patterns include: 1. 2. red-dashed flowexternal client to the server now located in DC2 Green-dotted flowvCenter management server in DC2 to vSphere eSXi host in DC1

3. orange-solid flowStorage in DC2 to eSXi host in DC1 4. black (solid) flowvCenter management server in DC1 to vSphere eSki host in DC2

these flows are problematic because the load increases through the DCI, which is the most expensive and congested point on the network. these flows allow less bandwidth to be available for existing applications and other migration traffic. Due to the latency incurred, the application might perform poorly because the storage and client are located further away from the VM. Moreover, this issue is magnified because more components reside in a three-tier or larger application. Migrating live servers between data centers -- while maintaining existing application flows is sometimes known as the hairpinning or trombone effect.
Outside Client for Remote Connection Client Router Testing MX Series Single VPLS all VLANs or Multiple VPLS for load-balancing and tra c separation

DC2 PE MX Series

DC1 PE MX Series MC LAG, LAG and Virtual Chassis technologies used to improve performance and reliability

VirtualChassis

Aggregation Switch EX Series

SRX Series Cluster QFX Series

SRX Series Cluster


VirtualChassis

EX Series

vCenter ESXi Host with vGW

NIC teaming to alternate switches for production and vMotion VLANs. Multiple links each for production, storage and vMotion/management

ESXi Host with vGW

DATA CENTER 1

DATA CENTER 2

Client tra c to VM vCenter tra c to ESXi host Storage tra c to VM/ESXi vMotion tra c ESXi host to host

Figure 7: Traffic flow after metro vMotion

10

Copyright 2013, Juniper Networks, Inc.

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

Optimizing Traffic Flows After Migration


this section reviews the methods for optimizing traffic flows after VM migration occurs between data centers.

Red Traffic Flow Solution


as shown in Figure 5, the red traffic flows from the external client to the server in DC1. this traffic flow is one of the most difficult flows to maintain and optimize because it typically traverses through stateful devices, such as firewalls and load balancers. It is also difficult to provide a more direct path to the second data center because of IP addressing. IP addressing is not designed for portability or mobility between different locations. to the outside world, there is one way to reach a serverthrough an IP address. typically, an IP address is discovered using DNS. although routing updates can be made quickly, they do not necessarily converge as rapidly as a VM can move from one location to another. the changeover time from the source VM to the destination VM can be in subseconds. Inter-data-center traffic can still be observed until the routing change occurs in the IP network. When migrating VMs on the same VlaN or subnet, network administrators must consider whether all VMs on that subnet migrate or just a few of them. If VMs are not migrated as a contiguous group or block of IP addresses, administrators can fragment the subnet. Fragmented subnets cause routing updates that are less likely to be acceptable by external autonomous systems, unless they are re-aggregated at the border routers. Sending route updates with small prefix sizes (fragmented) is acceptable in a closed network environment because administrators can control the subnet prefix size in a routing update. In the Internet environment, service providers incur limitations on the subnet size for an acceptable routing update to maintain the stability of the route tables. additional subnet route prefixes mean additional route changes, which consume more memory and CPU and lead to large resource requirements and potential instability. typically, /24 is the smallest prefix size accepted by todays Internet service providers. a single IP address prefix used by a typical VM is /32. Internet routing tables do not accept small address ranges that move around, such as those used by single VMs or small groups of VMs (under 255 IP addresses). Prefix size issues can be mitigated by providing routes that can be aggregated up to a /24 prefix at the peering provider. Changing the routing tables provides a method for optimizing ingress access. the common way of changing egress access is with a first-hop routing protocol such as Virtual router redundancy Protocol (VrrP), which allows an optimized choice of network exit in each data center as opposed to traffic crossing the DCI to exit the network.

Black Traffic Flow


the green traffic flow can be improved either by either moving the vCenter management server or deploying a clustered management server in the second data center. this solution is also applicable for other servers that interface with the VM. based on the type of traffic and communication requirements, administrators can arrange for a similar server in the second data center or move that server as part of a group to the second data center, thereby ensuring communication during the moves.

Orange Traffic Flow


the solution for the orange storage traffic flow has two possible methods. one option is to implement storage replication between the data centers. Many storage vendors already have integrated their solutions specifically with VMware to enable smooth transitions. another option is to use VMware svMotion (Storage vMotion), which allows administrators to migrate a VMs storage location from one machine to another by moving the entire data store. the difference is in replication. all data is transferred, and then the system simply makes incremental updates or delta changes to the remote site storage. this method reduces overall DCI link usage. all data is transferred at once either before or after the VMs vMotion event. therefore, heavy DCI usage must occur during the svMotion process to move all data at once. In addition, poor performance occurs due to latency between the VM and storage while they are located in different data centers. this technique avoids significant investment in a high-end, synchronous storage replication system but might not be suitable for VMs with large data stores.

Copyright 2013, Juniper Networks, Inc.

11

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

Implementation
Figure 8 illustrates the network configuration used to test inter-data-center VMware vMotion. two servers are located in Data Center 1 (DC1), and one server is located in Data Center 2 (DC2). the vCenter server also resides in DC1 and manages all three servers in the eSXi DrS cluster. all servers use the equallogic iSCSI storage array in DC1. all servers and storage use the same VlaN enabling l2 communication. In DC1, an access tier consisting of a QFX3600 is for switching. a combined core-aggregation tier consists of two eX82s in a VC configuration for redundancy and loop avoidance. the data center edge has two MX960s using VPlS multihoming and MC-laG for redundancy. In DC2, an access tier of three eX42s in a VC configuration for redundancy and loop avoidance is for switching. a combined core-aggregation tier consists of two MX480s in a VC configuration for redundancy and loop avoidance. the data center edge has two MX80s using VPlS multihoming and MC-laG for redundancy. the server and storage VlaN is trunked between the access and core-aggregation layers using 802.1Q in each data center and trunked across data centers using VPlS at the data center WaN edge.

MX960 WAN Edge VPLS PE

MC-LAG

vCenter
172.28.113.27

MC-LAG MX80
MX 24 172.28.114.131

MX80 WAN Edge VPLS PE

MX960
MX 13 172.28.114.50

MPLS Core P-Routers MX960


MX 14 172.28.114.52

MX480 Virtual Chassis Core


172.28.113.133

MX80
MX 23 172.28.114.130 xe-3/1/0

MX480
MX 07 172.28.114.132

MX480
MX 020 172.28.114.134

EX8208
EX 037 172.28.114.128

EX8216
EX 038 172.28.114.128

xe-2/1/1 Virtual Chassis 10.13.97.161

EX4200

EX8200 Core
xe-50/0/18 xe-0/0/14 xe-0/0/10 xe-0/0/15 xe-0/0/12 Eth1.0 Eth0.1 10.1.11.191 Eth1.1 10.1.11.190 WM-DC1-Vidclone-172.28.113.201, 10.1.11.194 DSL-DC1-172.28.113.201, 10.1.11.217 DSL-DC2-172.28.113.202, 10.1.11.218 xe-50/0/19 xe-0/0/21 ge-0/0/22 ge-0/0/17 10.1.11.150

xe-1/1/1 xe-1/1/0 xe-2/1/0

ge-0/0/12

QFX3500
QFX 003 172.28.114.167 ge=0/0/18 xe-0/0/13 ge-3

10.1.11.181

ge 4 10.1.11.186

ge-2 10.1.11.195

ge-2 10.1.11.180

Eth0.0

Storage Array (iSCSI) Srv121 10.13.99.103 10.1.11.179 DC2-GRP-2

Srv123 ESXi1-R710 10.13.99.102

Storage Array (iSCSI) 172.28.113.90 10.1.11.192 DC1-GRP-1

Win28K-Srv 1-10.1.11.211, 10.13.99.109 WM-VM2-710-10.1.11.212, 10.13.99.81

DATA CENTER 1

DATA CENTER 2

Figure 8: Data center workload mobility, network configuration to meet the requirements presented in the design section, administrators should create a reliable l2 connection from DC1 to DC2 by trunking the VlaNs from the vSwitches on the eSXi hosts in DC1 to the eSXi host in DC2. the VlaNs must be enabled and trunked on the top of rack (tor) switches in both locations and trunked across the VlPS connection at the data center edge. We use redundant links, link aggregation group (laG), Multichassis laG (MC-laG), MPlS, and VPlS multihoming to ensure reliability of the l2-trunked VlaN. the VlaN that we chose to trunk (used by the VMware eSXi cluster) is VLAN1101. NOTE: If you require additional VlaNs to separate traffic for vMotion traffic, management traffic, or other traffic, these specific VlaNs must be trunked in the same manner. Using Figure 8 as a reference, all eSXi host interfaces are inserted into VLAN 1101 and are trunked to the data center edge in both data centers.

12

Copyright 2013, Juniper Networks, Inc.

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

QFX3600-QFX003 in DC1 Configuration


this example configures VlaN access and trunk ports on the QFX access switch. all interfaces that require access are added to the VlaN. If you use multiple VlaNs (one for data, one for vMotion, one for management, etc,), they must also be added and trunked. this example contains a single VlaN.

vlans { VR1-vlan1101 { vlan-id 1101; interface { xe-0/0/21.0; xe-0/0/14.0; xe-0/0/15.0; xe-0/0/10.0; xe-0/0/12.0; xe-0/0/13.0; ge-0/0/17.0; ge-0/0/18.0; ge-0/0/22.0; } l3-interface vlan.1101; } }
trunks are configured on uplinks in the following code snippet. If multiple VlaNs are required from each access link to the eSXi host, port mode or tagged access should also be added to those interfaces.

interfaces { xe-0/0/14 { unit 0 { family ethernet-switching { port-mode trunk; } } } xe-0/0/21 { unit 0 { family ethernet-switching { port-mode trunk; } } }

Copyright 2013, Juniper Networks, Inc.

13

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

EX4200-VC in DC2 Configuration


this example configures VlaN access and trunk ports on the eX VC access switch. all interfaces that require access are added to the VlaN. If you use multiple VlaNs (one for data, one for vMotion, one for management, and so on), they must also to added and trunked. this example contains a single VlaN.

vlans { VR1-vlan1101 { vlan-id 1101; interface { ge-0/0/9.0; ge-0/0/12.0; xe-2/1/1.0; xe-2/1/0.0; xe-1/1/0.0; xe-1/1/1.0; } l3-interface vlan.1101; } }
trunks are configured on uplinks in the following code snippet. If multiple VlaNs are required from each access link to the eSXi host, port mode or tagged access should also be added to those interfaces.

interfaces { xe-2/1/1 { unit 0 { family ethernet-switching { port-mode trunk; } } }


If aggregated ethernet (laG) interfaces are used for more bandwidth or redundancy, they must be added to the VlaN and have port-mode trunk enabled on them.

14

Copyright 2013, Juniper Networks, Inc.

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

MX960-MX13 in DC1 Configuration


the following configuration shows the data center provider edge routers that handle the l2 stretch between the data centers. this configuration requires several steps: Configure the laN-side interfaces to trunk the VlaNs; configure the backbone interfaces for MPlS, rSVP for signaling the label-switched paths (lSPs), and lSPs to allow traffic to be carried between the two provider edge routers; and configure VPlS using separate routing instances, Multiprotocol bGP (MP-bGP), and policies.

Interfaces Configuration

xe-1/0/0 { description CONNECTED-TO-EX8208; gigether-options { 802.3ad { ae3; backup; } } } xe-1/1/0 { description CONNECTED-TO-EX8216; gigether-options { 802.3ad { ae3; primary; } } } xe-10/0/0 { description Connected to EX8200-VC via MC-LAG-1; gigether-options { 802.3ad ae1; } } xe-10/1/0 { description Connected to EX8200-VC via MC-LAG-2; gigether-options { 802.3ad ae2; } } xe-10/2/0 { description CONNECTED-TO-MX960-52-ICL-TRAFFIC; gigether-options { 802.3ad ae5; } } xe-10/3/0 { description CONNECTED-TO-MX960-52-ICL-TRAFFIC; gigether-options { 802.3ad ae5; } } interfaces { ge-11/2/0 { description CONNECTED-TO-PE2-52-ICCP-TRAFFIC; unit 0 { family inet { address 10.11.12.2/30; } } }

Copyright 2013, Juniper Networks, Inc.

15

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

MC-LAG Configuration for LAN

protocols { iccp { local-ip-addr 10.11.12.2; peer 10.11.12.1 { redundancy-group-id-list 1; liveness-detection { minimum-interval 100; multiplier 3; detection-time { threshold 500; } } } } lacp { ppm centralized; } } interfaces { ae1 { description Connecting to EX8200 MC-LAG-1; flexible-vlan-tagging; encapsulation vlan-vpls; aggregated-ether-options { lacp { active; periodic fast; system-id 00:00:00:00:00:50; admin-key 123; } mc-ae { mc-ae-id 101; redundancy-group 1; chassis-id 0; mode active-standby; status-control active; } } unit 1101 { encapsulation vlan-vpls; vlan-id 1101; family vpls; } unit 1102 { encapsulation vlan-vpls; vlan-id 1102; family vpls; } unit 1103 { encapsulation vlan-vpls; vlan-id 1103; family vpls; } unit 1201 {

16

Copyright 2013, Juniper Networks, Inc.

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

} unit 1202 { encapsulation vlan-vpls; vlan-id 1202; family vpls; } unit 1203 { encapsulation vlan-vpls; vlan-id 1203; family vpls; }

encapsulation vlan-vpls; vlan-id 1201; family vpls;

Backbone Routing and MPLS Configuration

protocols { rsvp { interface xe-1/0/0.0; interface ge-11/0/1.0; ospf { traffic-engineering; area 0.0.0.0 { interface all { } } } mpls { label-switched-path DC1.PE1-DC2.PE1 { to 11.1.1.31; fast-reroute; } label-switched-path DC1.PE1-DC2.PE2 { to 11.1.1.130; fast-reroute; } label-switched-path Campus-PE1 { to 10.10.10.65; fast-reroute; } label-switched-path Campus-PE2 { to 10.10.10.69; fast-reroute; } interface xe-1/0/0.0; interface ge-11/0/1.0;

Copyright 2013, Juniper Networks, Inc.

17

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

VPLS Configuration

routing-options { nonstop-routing; } router-id 10.10.10.50; autonomous-system 500; protocols { bgp { group DC-1 { type internal; local-address 10.10.10.50; } family l2vpn { signaling; } peer-as 500; local-as 500; neighbor 11.1.1.31; neighbor 10.10.10.52; neighbor 11.1.1.130; } } DC-VR-1 { instance-type vpls; vlan-id all; interface ae1.1101; interface ae1.1102; interface ae1.1103; interface ae1.1201; interface ae1.1202; interface ae1.1203; route-distinguisher 1:50; vrf-target target:900:1000; protocols { vpls { site-range 8; no-tunnel-services; site DC1-1 { site-identifier 1; multi-homing; site-preference primary; interface ae1.1101; interface ae1.1102; interface ae1.1103; interface ae1.1203; interface ae1.1202; interface ae1.1201; } } } }

18

Copyright 2013, Juniper Networks, Inc.

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

MX960-MX14 in DC1 Configuration


the following configuration shows the data center provider edge routers that handle the l2 stretch between the data centers. this configuration requires several steps: Configure the laN-side interfaces to trunk the VlaNs; configure the backbone interfaces for MPlS, rSVP for signaling the lSPs, and lSPs to allow traffic to be carried between the two provider edge routers; and configure the VPlS using separate routing instances, MP-bGP, and policies.

Interfaces Configuration

interfaces { xe-1/0/0 { description CONNECTED-TO-EX8216-AE5; gigether-options { 802.3ad { ae3; primary; } } } xe-1/1/0 { description CONNECTED-TO-EX8208-AE5; gigether-options { 802.3ad { ae3; backup; } } } xe-9/0/0 { description CONNECTED-TO-P4-122; mtu 2000; unit 0 { family inet { address 10.3.10.41/30; } family iso; family mpls; } } xe-9/1/0 { description CONNECTED-TO-P2-72; mtu 2000; unit 0 { family inet { address 10.3.10.45/30; } family iso; family mpls; } } xe-10/0/0 { description Connected to EX8200-VC via MC-LAG-1; gigether-options { 802.3ad ae1; } } xe-10/1/0 { description Connected to EX8200-VC via MC-LAG-2;

Copyright 2013, Juniper Networks, Inc.

19

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

} xe-10/2/0 { description CONNECTED-TO-MX960-50-ICL-TRAFFIC; gigether-options { 802.3ad ae5; } } xe-10/3/0 { description CONNECTED-TO-MX960-50-ICL-TRAFFIC; gigether-options { 802.3ad ae5; } } ge-11/2/0 { description CONNECTED-TO-PE1-50-ICCP-TRAFFIC; unit 0 { family inet { address 10.11.12.1/30; } } }
MC-LAG LAN Configuration

gigether-options { 802.3ad ae2; }

protocols { iccp { local-ip-addr 10.11.12.1; peer 10.11.12.2 { redundancy-group-id-list 1; liveness-detection { minimum-interval 100; multiplier 3; detection-time { threshold 500; } } } } lacp { ppm centralized; } } interfaces { ae1 { description Connecting to EX8200 MC-LAG-1; flexible-vlan-tagging; encapsulation vlan-vpls; aggregated-ether-options { lacp { active; periodic fast; system-id 00:00:00:00:00:50; admin-key 123; }

20

Copyright 2013, Juniper Networks, Inc.

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

} unit 1101 { encapsulation vlan-id 1101; family vpls; } unit 1102 { encapsulation vlan-id 1102; family vpls; } unit 1103 { encapsulation vlan-id 1103; family vpls; } unit 1201 { encapsulation vlan-id 1201; family vpls; } unit 1202 { encapsulation vlan-id 1202; family vpls; } unit 1203 { encapsulation vlan-id 1203; family vpls; }

mc-ae { mc-ae-id 101; redundancy-group 1; chassis-id 1; mode active-standby; status-control standby; } vlan-vpls;

vlan-vpls;

vlan-vpls;

vlan-vpls;

vlan-vpls;

vlan-vpls;

Backbone Routing and MPLS Configuration

protocols { rsvp { interface xe-1/0/0.0; interface all; } ospf { traffic-engineering; area 0.0.0.0 { interface all; } } mpls { path-mtu; label-switched-path DC1.PE2-DC2.PE1 { to 11.1.1.31;

Copyright 2013, Juniper Networks, Inc.

21

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

} label-switched-path DC1.PE2-DC2.PE2 { to 11.1.1.130; } interface xe-1/0/0.0; interface ge-11/0/0.0; interface all;


VPLS Configuration

routing-options { nonstop-routing; } router-id 10.10.10.52; autonomous-system 500; } protocols { bgp { group DC-1 { type internal; local-address 10.10.10.52; family l2vpn { signaling; } peer-as 500; local-as 500; neighbor 11.1.1.31; neighbor 11.1.1.130; neighbor 10.10.10.50; } } routing-instances { DC-VR-1 { instance-type vpls; vlan-id all; interface ae1.1101; interface ae1.1102; interface ae1.1103; route-distinguisher 1:52; vrf-target target:900:1000; protocols { vpls { site-range 8; no-tunnel-services; site DC1-1 { site-identifier 1; multi-homing; site-preference backup; interface ae1.1101; interface ae1.1102; interface ae1.1103; } } } }

22

Copyright 2013, Juniper Networks, Inc.

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

MX80-MX23 in DC2 Interface Configuration

interfaces { xe-0/0/0 { description CONNECTED-TO-MX480-134; gigether-options { 802.3ad ae1; } } xe-0/0/1 { description CONNECTED-TO-P1-71; mtu 2000; unit 0 { family inet { address 10.3.10.49/30; } family iso; family mpls; } } xe-0/0/2 { description CONNECTED-TO-MX480-132; gigether-options { 802.3ad ae2; } } ge-1/0/0 { description CONNECTED-TO-PE1-131; gigether-options { 802.3ad ae0; } } ge-1/0/1 { description CONNECTED-TO-PE1-131; disable; gigether-options { 802.3ad ae0; } } ae0 { description Connecting MX80 Multichassis LAG; aggregated-ether-options { lacp { active; periodic fast; } } unit 0 { family inet { address 10.4.10.14/30; } } }

Copyright 2013, Juniper Networks, Inc.

23

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

MC-LAG LAN Configuration

protocols { iccp { local-ip-addr 10.4.10.14; peer 10.4.10.13 { redundancy-group-id-list 1; liveness-detection { minimum-interval 100; multiplier 3; } } } } interfaces { ae2 { description Connecting to MX480 MC-LAG-2-132; flexible-vlan-tagging; encapsulation vlan-vpls; aggregated-ether-options { lacp { active; periodic fast; system-id 00:00:00:00:00:31; admin-key 123; } mc-ae { mc-ae-id 132; redundancy-group 1; chassis-id 1; mode active-standby; status-control standby; } } unit 1101 { encapsulation vlan-vpls; vlan-id 1101; family vpls; } unit 1102 { encapsulation vlan-vpls; vlan-id 1102; family vpls; } unit 1103 { encapsulation vlan-vpls; vlan-id 1103; family vpls; } unit 1201 { encapsulation vlan-vpls; vlan-id 1201; family vpls; } unit 1202 {

24

Copyright 2013, Juniper Networks, Inc.

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

} unit 1203 { encapsulation vlan-vpls; vlan-id 1203; family vpls; }

encapsulation vlan-vpls; vlan-id 1202; family vpls;

Backbone Routing and MPLS Configuration

protocols { rsvp { interface all; } mpls { label-switched-path DC2.PE2-DC1.PE1 { to 10.10.10.50; } label-switched-path DC2.PE2-DC1.PE2 { to 10.10.10.52; } interface all; } ospf { traffic-engineering; area 0.0.0.0 { interface lo0.0 { passive; } interface all; interface fxp0.0 { disable; } } }
VPLS Configuration

routing-options { router-id 11.1.1.130; autonomous-system 500; } protocols { bgp { group DC-2 { type internal; local-address 11.1.1.130; family inet { unicast; any; } family inet-vpn { unicast;

Copyright 2013, Juniper Networks, Inc.

25

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

any; } family l2vpn { signaling; } peer-as 500; local-as 500; neighbor 10.10.10.50; neighbor 10.10.10.52; neighbor 11.1.1.31;

routing-instances { DC-VR-1 { instance-type vpls; vlan-id all; interface ae2.1101; interface ae2.1102; interface ae2.1103; interface ae2.1201; interface ae2.1202; interface ae2.1203; route-distinguisher 1:130; vrf-target target:900:1000; protocols { vpls { site-range 8; no-tunnel-services; site DC2-1 { site-identifier 3; multi-homing; site-preference backup; interface ae2.1101; interface ae2.1102; interface ae2.1103; interface ae2.1203; interface ae2.1202; interface ae2.1201; } } } }
MX80-MX24 in DC2 Interface Configuration

interfaces { xe-0/0/0 { description CONNECTED-TO-MX480-132; gigether-options { 802.3ad ae1; } } xe-0/0/1 { description CONNECTED-TO-P3-92; mtu 2000; unit 0 {

26

Copyright 2013, Juniper Networks, Inc.

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

} } xe-0/0/2 { description CONNECTED-TO-MX480-134; gigether-options { 802.3ad ae2; } } xe-0/0/3 { description CONNECTED-TO-MX80-130-ICC-TRAFFIC; unit 0 { family inet { address 10.3.10.57/30; } } } ge-1/0/0 { description CONNECTED-TO-PE2-130; gigether-options { 802.3ad ae0; } } ge-1/0/2 { description CONNECTED-TO-PE2-130; gigether-options { 802.3ad ae0; } } ae0 { description Connecting MX80 Multichassis LAG; aggregated-ether-options { lacp { active; periodic fast; } } unit 0 { family inet { address 10.4.10.13/30; } } }
MC-LAG LAN Configuration

family inet { address 10.3.10.53/30; } family iso; family mpls;

protocols { iccp { local-ip-addr 10.4.10.13; peer 10.4.10.14 { redundancy-group-id-list 1; liveness-detection { minimum-interval 50;

Copyright 2013, Juniper Networks, Inc.

27

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

} interfaces { ae2 { description Connecting to MX480 MC-LAG-2; flexible-vlan-tagging; encapsulation vlan-vpls; aggregated-ether-options { lacp { active; periodic fast; system-id 00:00:00:00:00:31; admin-key 123; } mc-ae { mc-ae-id 132; redundancy-group 1; chassis-id 0; mode active-standby; status-control active; } } unit 1101 { encapsulation vlan-vpls; vlan-id 1101; family vpls; } unit 1102 { encapsulation vlan-vpls; vlan-id 1102; family vpls; } unit 1103 { encapsulation vlan-vpls; vlan-id 1103; family vpls; } unit 1201 { encapsulation vlan-vpls; vlan-id 1201; family vpls; } unit 1202 { encapsulation vlan-vpls; vlan-id 1202; family vpls; } unit 1203 { encapsulation vlan-vpls; vlan-id 1203; family vpls; } }

multiplier 3;

28

Copyright 2013, Juniper Networks, Inc.

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

Backbone Routing and MPLS Configuration

rsvp { interface all; } mpls { label-switched-path DC2.PE1-DC1.PE1 { to 10.10.10.50; } label-switched-path DC2.PE1-DC1.PE2 { to 10.10.10.52; } interface all; } ospf { traffic-engineering; area 0.0.0.0 { interface lo0.0 { passive; } interface fxp0.0 { disable; } interface all; } } protocols {
VPLS Configuration

router-id 11.1.1.31; autonomous-system 500;

protocols { bgp { group DC-2 { type internal; local-address 11.1.1.31; family inet { any; } family inet-vpn { unicast; any; } family l2vpn { signaling; } peer-as 500; local-as 500; routing-options { neighbor 10.10.10.50; neighbor 10.10.10.52; neighbor 11.1.1.130; } }

Copyright 2013, Juniper Networks, Inc.

29

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

routing-instances { DC-VR-1 { instance-type vpls; vlan-id all; interface ae2.1101; interface ae2.1102; interface ae2.1103; interface ae2.1201; interface ae2.1202; interface ae2.1203; route-distinguisher 1:131; vrf-target target:900:1000; protocols { vpls { site-range 8; no-tunnel-services; site DC2-1 { site-identifier 3; multi-homing; site-preference primary; interface ae2.1101; interface ae2.1102; interface ae2.1103; interface ae2.1201; interface ae2.1202; interface ae2.1203; } } } }

Validation
the following commands validate the MPlS configuration, mainly from the devices in DC1. For brevity, DC2 is not shown. Validation of the successful move from the vCenter VMware management console is also shown. the following validation shows MPlS and rSVP in DC1.

MX960-MX14 in DC1

root@MX960-DC-1-52> show mpls lsp Ingress LSP: 2 sessions To From State Rt P 11.1.1.31 10.10.10.52 Up 0 * 11.1.1.130 10.10.10.52 Up 0 * Total 2 displayed, Up 2, Down 0 Egress LSP: 2 sessions To From State 10.10.10.52 11.1.1.130 Up 10.10.10.52 11.1.1.31 Up Total 2 displayed, Up 2, Down 0 Transit LSP: 0 sessions Total 0 displayed, Up 0, Down 0

ActivePath

LSPname DC1.PE2-DC2.PE1 DC1.PE2-DC2.PE2

Rt Style Labelin Labelout LSPname 0 1 FF 3 - DC2.PE2-DC1.PE2 0 1 FF 3 - DC2.PE1-DC1.PE2

30

Copyright 2013, Juniper Networks, Inc.

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

root@MX960-DC-1-52> show rsvp session Ingress RSVP: 2 sessions To From State 11.1.1.31 10.10.10.52 Up 11.1.1.130 10.10.10.52 Up Total 2 displayed, Up 2, Down 0 Egress RSVP: 2 sessions To From State 10.10.10.52 11.1.1.130 Up 10.10.10.52 11.1.1.31 Up Total 2 displayed, Up 2, Down 0 Transit RSVP: 0 sessions Total 0 displayed, Up 0, Down 0
MX960-MX13 in DC1

Rt Style Labelin Labelout LSPname 0 1 FF 299776 DC1.PE2-DC2.PE1 0 1 FF 299840 DC1.PE2-DC2.PE2

Rt Style Labelin Labelout LSPname 0 1 FF 3 - DC2.PE2-DC1.PE2 0 1 FF 3 - DC2.PE1-DC1.PE2

root@MX960-DC-1-50> show mpls lsp Ingress LSP: 2 sessions To From State Rt P 11.1.1.31 10.10.10.50 Up 0 * 11.1.1.130 10.10.10.50 Up 0 * Total 2 displayed, Up 2, Down 0 Egress LSP: 2 sessions To From State 10.10.10.50 11.1.1.130 Up 10.10.10.50 11.1.1.31 Up Total 2 displayed, Up 2, Down 0 Transit LSP: 0 sessions Total 0 displayed, Up 0, Down 0 root@MX960-DC-1-50> show rsvp session Ingress RSVP: 2 sessions To From State 11.1.1.31 10.10.10.50 Up 11.1.1.130 10.10.10.50 Up Total 2 displayed, Up 2, Down 0 Egress RSVP: 2 sessions To From State 10.10.10.50 11.1.1.130 Up 10.10.10.50 11.1.1.31 Up Total 2 displayed, Up 2, Down 0 Transit RSVP: 0 sessions Total 0 displayed, Up 0, Down

ActivePath

LSPname DC1.PE1-DC2.PE1 DC1.PE1-DC2.PE2

Rt Style Labelin Labelout LSPname 0 1 FF 3 - DC2.PE2-DC1.PE1 0 1 FF 3 - DC2.PE1-DC1.PE1

Rt Style Labelin Labelout LSPname 0 1 FF 299792 DC1.PE1-DC2.PE1 0 1 FF 299808 DC1.PE1-DC2.PE2

Rt Style Labelin Labelout LSPname 0 1 FF 3 - DC2.PE2-DC1.PE1 0 1 FF 3 - DC2.PE1-DC1.PE1

Copyright 2013, Juniper Networks, Inc.

31

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

MX960-MX13 in DC1
this command shows ICCP status in DC1.

root@MX960-DC-1-50> show iccp brief Redundancy Group Information for peer 10.11.12.1 TCP Connection : Established Liveliness Detection : Up Redundancy Group ID Status 1 Up Client Application: lacpd Redundancy Group IDs Joined: 1 Client Application: l2ald_iccpd_client Redundancy Group IDs Joined: None Client Application: MCSNOOPD Redundancy Group IDs Joined: None
MX960-MX14 in DC1

root@MX960-DC-1-52> show iccp brief Redundancy Group Information for peer 10.11.12.2 TCP Connection : Established Liveliness Detection : Up Redundancy Group ID Status 1 Up Client Application: lacpd Redundancy Group IDs Joined: 1 Client Application: l2ald_iccpd_client Redundancy Group IDs Joined: None Client Application: MCSNOOPD Redundancy Group IDs Joined: None
the following validation shows that the MC-laG interfaces are in the up state and running DC1.

root@MX960-DC-1-50> show interfaces mc-ae Member Link : ae1 Current State Machines State: mcae active state Local Status : active Local State : up Peer Status : standby Peer State : up Logical Interface : ae1.1101 root@MX960-DC-1-52> show interfaces mc-ae Member Link : ae1 Current State Machines State: mcae standby state Local Status : standby Local State : up Peer Status : active Peer State : up Logical Interface : ae1.1101

32

Copyright 2013, Juniper Networks, Inc.

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

MX960-MX13 in DC1
the following validation shows that the VPlS connection is up and running in DC1 and DC2.

root@MX960-DC-1-50> show vpls connections Instance: DC-VR-1 Local site: DC1-1 (1) connection-site Type St Time last up # Up trans 1 rmt RN 3 rmt Up Aug 15 15:09:08 2011 1 Remote PE: 11.1.1.31, Negotiated control-word: No Incoming label: 262147, Outgoing label: 262145 Local interface: lsi.1049088, Status: Up, Encapsulation: VPLS Description: Intf - vpls DC-VR-1 local site 1 remote site 3
MX960-MX14 in DC1

root@MX960-DC-1-52> show vpls connections Layer-2 VPN connections: Instance: DC-VR-1 Local site: DC1-1 (1) connection-site Type St 1 rmt LN 3 rmt LN root@MX80-DC-2-130> show vpls connections Layer-2 VPN connections: Instance: DC-VR-1 Local site: DC2-1 (3) connection-site Type St 1 rmt LN 3 rmt LN
MX80-MX24 in DC2

Time last up

# Up trans

Time last up

# Up trans

root@MX80-DC-2-131> show vpls connections Layer-2 VPN connections: Instance: DC-VR-1 Local site: DC2-1 (3) connection-site Type St Time last up # Up trans 1 rmt Up Aug 15 20:41:06 2011 1 Remote PE: 10.10.10.50, Negotiated control-word: No Incoming label: 262145, Outgoing label: 262147 Local interface: lsi.1049088, Status: Up, Encapsulation: VPLS Description: Intf - vpls DC-VR-1 local site 3 remote site 1 3 rmt RN

Copyright 2013, Juniper Networks, Inc.

33

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

MX80-MX23 in DC2

root@MX80-DC-2-131> show vpls connections Layer-2 VPN connections: Instance: DC-VR-1 Local site: DC2-1 (3) connection-site Type St Time last up # Up trans 1 rmt Up Aug 15 20:41:06 2011 1 Remote PE: 10.10.10.50, Negotiated control-word: No Incoming label: 262145, Outgoing label: 262147 Local interface: lsi.1049088, Status: Up, Encapsulation: VPLS Description: Intf - vpls DC-VR-1 local site 3 remote site 1 3 rmt RN

Validation Using the vCenter Client Application


Using the vCenter Client application, the following screen captures show a VM migration. Figure 9 shows the ping traffic from the client to the server reaching a 512 ms rtt during live migration of the server. the large rtt is when the VM is stopped on one server and started on the second.

Figure 9: Ping traffic from client to server

34

Copyright 2013, Juniper Networks, Inc.

IMPLEMENTATION GUIDE - Inter Data Center Workload Mobility with VMware

Figure 10 shows the time and success of the vMotion event in the recent task pane.

Figure 10: Time and success of the vMotion event

Summary
Network administrators can achieve inter-data-center workload mobility by deploying a combination of VMware and Juniper Networks devices. VMware provides the mobility solution, and Juniper provides a high-performance, highly reliable data center laN and data center WaN experience. When planning to migrate workloads, it is critical to examine the business case, because costs and planning requirements can be expensive and complicated. as presented in the Campus and Cold Migration scenarios, data center migration for live workloads can be accomplished without implementing complicated first-hop-responder and route-optimization techniques. other scenarios are possible; however, they require implementing protocols that share state and fate, thereby increasing the complexity and reducing the reliability of the connected data centers. When designing your workload mobility solution, it is important to examine costs, distance, and latency in your environment and to evaluate the dependencies that your server application flows have on each other and the network.

About Juniper Networks


Juniper Networks is in the business of network innovation. From devices to data centers, from consumers to cloud providers, Juniper Networks delivers the software, silicon and systems that transform the experience and economics of networking. the company serves customers and partners worldwide. additional information can be found at www.juniper.net.
Corporate and Sales Headquarters Juniper Networks, Inc. 1194 North Mathilda avenue Sunnyvale, Ca 94089 USa Phone: 888.JUNIPer (888.586.4737) or 408.745.2000 Fax: 408.745.2100 www.juniper.net
Copyright 2013 Juniper Networks, Inc. all rights reserved. Juniper Networks, the Juniper Networks logo, Junos, NetScreen, and ScreenoS are registered trademarks of Juniper Networks, Inc. in the United States and other countries. all other trademarks, service marks, registered marks, or registered service marks are the property of their respective owners. Juniper Networks assumes no responsibility for any inaccuracies in this document. Juniper Networks reserves the right to change, modify, transfer, or otherwise revise this publication without notice.

APAC and EMEA Headquarters Juniper Networks International b.V. boeing avenue 240 1119 PZ Schiphol-rijk amsterdam, the Netherlands Phone: 31.0.207.125.700 Fax: 31.0.207.125.701

to purchase Juniper Networks solutions, please contact your Juniper Networks representative at 1-866-298-6428 or authorized reseller.

8010089-001-eN

Jan 2013

Printed on recycled paper

Copyright 2013, Juniper Networks, Inc.

35

Das könnte Ihnen auch gefallen