Beruflich Dokumente
Kultur Dokumente
Concepts Guide
version 11.0
MAN-0346-00
Product Version
This guide applies to product version 11.0 of the BIG-IP® Global Traffic Manager™.
Publication Date
This guide was published on August 11, 2011.
Legal Notices
Copyright
Copyright 2011, F5 Networks, Inc. All rights reserved.
F5 Networks, Inc. (F5) believes the information it furnishes to be accurate and reliable. However, F5
assumes no responsibility for the use of this information, nor any infringement of patents or other rights of
third parties which may result from its use. No license is granted by implication or otherwise under any
patent, copyright, or other intellectual property right of F5 except as specifically described by applicable
user licenses. F5 reserves the right to change specifications at any time without notice.
Trademarks
3DNS, Access Policy Manager, Acopia, Acopia Networks, Advanced Client Authentication, Advanced
Routing, APM, Application Security Manager, ARX, AskF5, ASM, BIG-IP, Cloud Extender,
CloudFucious, CMP, Data Manager, DevCentral, DevCentral [DESIGN], DNS Express, DSC, DSI, Edge
Client, Edge Gateway, Edge Portal, EM, Enterprise Manager, F5, F5 [DESIGN], F5 Management Pack, F5
Networks, F5 World, Fast Application Proxy, Fast Cache, FirePass, Global Traffic Manager, GTM, IBR,
Intelligent Browser Referencing, Intelligent Compression, IPv6 Gateway, iApps, iControl, iHealth,
iQuery, iRules, iRules OnDemand, iSession, IT agility. Your way., L7 Rate Shaping, LC, Link Controller,
Local Traffic Manager, LTM, Message Security Module, MSM, Netcelera, OneConnect, Packet Velocity,
Protocol Security Module, PSM, Real Traffic Policy Builder, ScaleN, SSL Acceleration, StrongBox,
SuperVIP, SYN Check, TCP Express, TDR, TMOS, Traffic Management Operating System,
TrafficShield, Transparent Data Reduction, VIPRION, vCMP, WA, WAN Optimization Manager,
WANJet, WebAccelerator, WOM, and ZoneRunner, are trademarks or service marks of F5 Networks, Inc.,
in the U.S. and other countries, and may not be used without F5's express written consent.
All other product and company names herein may be trademarks of their respective owners.
Patents
This product may be protected by U.S. Patents 6,374,300; 6,473,802; 6,970,733; 7,047,301; 7,707,289.
This list is believed to be current as of August 11, 2011.
RF Interference Warning
This is a Class A product. In a domestic environment this product may cause radio interference, in which
case the user may be required to take adequate measures.
FCC Compliance
This equipment has been tested and found to comply with the limits for a Class A digital device pursuant
to Part 15 of FCC rules. These limits are designed to provide reasonable protection against harmful
interference when the equipment is operated in a commercial environment. This unit generates, uses, and
can radiate radio frequency energy and, if not installed and used in accordance with the instruction manual,
may cause harmful interference to radio communications. Operation of this equipment in a residential area
is likely to cause harmful interference, in which case the user, at his own expense, will be required to take
whatever measures may be required to correct the interference.
Standards Compliance
This product conforms to the IEC, European Union, ANSI/UL and Canadian CSA standards applicable to
Information Technology products at the time of manufacture.
Acknowledgments
This product includes software developed by Gabriel Forté.
This product includes software developed by Bill Paul.
This product includes software developed by Jonathan Stone.
This product includes software developed by Manuel Bouyer.
This product includes software developed by Paul Richards.
This product includes software developed by the NetBSD Foundation, Inc. and its contributors.
This product includes software developed by the Politecnico di Torino, and its contributors.
This product includes software developed by the Swedish Institute of Computer Science and its
contributors.
This product includes software developed by the University of California, Berkeley and its contributors.
This product includes software developed by the Computer Systems Engineering Group at the Lawrence
Berkeley Laboratory.
This product includes software developed by Christopher G. Demetriou for the NetBSD Project.
This product includes software developed by Adam Glass.
This product includes software developed by Christian E. Hopps.
This product includes software developed by Dean Huxley.
This product includes software developed by John Kohl.
This product includes software developed by Paul Kranenburg.
This product includes software developed by Terrence R. Lambert.
This product includes software developed by Philip A. Nelson.
This product includes software developed by Herb Peyerl.
This product includes software developed by Jochen Pohl for the NetBSD Project.
This product includes software developed by Chris Provenzano.
This product includes software developed by Theo de Raadt.
This product includes software developed by David Muir Sharnoff.
This product includes software developed by SigmaSoft, Th. Lockert.
This product includes software developed for the NetBSD Project by Jason R. Thorpe.
This product includes software developed by Jason R. Thorpe for And Communications,
http://www.and.com.
This product includes software developed for the NetBSD Project by Frank Van der Linden.
This product includes software developed for the NetBSD Project by John M. Vinopal.
This product includes software developed by Christos Zoulas.
This product includes software developed by the University of Vermont and State Agricultural College and
Garrett A. Wollman.
In the following statement, "This software" refers to the Mitsumi CD-ROM driver: This software was
developed by Holger Veit and Brian Moore for use with "386BSD" and similar operating systems.
"Similar operating systems" includes mainly non-profit oriented systems for research and education,
including but not restricted to "NetBSD," "FreeBSD," "Mach" (by CMU).
This product includes software developed by the Apache Group for use in the Apache HTTP server project
(http://www.apache.org/).
This product includes software licensed from Richard H. Porter under the GNU Library General Public
License (© 1998, Red Hat Software), www.gnu.org/copyleft/lgpl.html.
ii
This product includes the standard version of Perl software licensed under the Perl Artistic License (©
1997, 1998 Tom Christiansen and Nathan Torkington). All rights reserved. You may find the most current
standard version of Perl at http://www.perl.com.
This product includes software developed by Jared Minch.
This product includes software developed by the OpenSSL Project for use in the OpenSSL Toolkit
(http://www.openssl.org/).
This product includes cryptographic software written by Eric Young (eay@cryptsoft.com).
This product contains software based on oprofile, which is protected under the GNU Public License.
This product includes RRDtool software developed by Tobi Oetiker (http://www.rrdtool.com/index.html)
and licensed under the GNU General Public License.
This product contains software licensed from Dr. Brian Gladman under the GNU General Public License
(GPL).
This product includes software developed by the Apache Software Foundation <http://www.apache.org/>.
This product includes Hypersonic SQL.
This product contains software developed by the Regents of the University of California, Sun
Microsystems, Inc., Scriptics Corporation, and others.
This product includes software developed by the Internet Software Consortium.
This product includes software developed by Nominum, Inc. (http://www.nominum.com).
This product contains software developed by Broadcom Corporation, which is protected under the GNU
Public License.
This product contains software developed by MaxMind LLC, and is protected under the GNU Lesser
General Public License, as published by the Free Software Foundation.
This product includes the GeoPoint Database developed by Quova, Inc. and its contributors.
This product includes software developed by Balazs Scheidler <bazsi@balabit.hu>, which is protected
under the GNU Public License.
This product includes software developed by NLnet Labs and its contributors.
This product includes software written by Steffen Beyer and licensed under the Perl Artistic License and
the GPL.
This product includes software written by Makamaka Hannyaharamitu © 2007-2008.
1
Overview
Global Traffic Manager .................................................................................................................. 1-1
Security features .................................................................................................................... 1-1
Local Traffic Manager resources ....................................................................................... 1-2
Internet protocol and network management support ................................................. 1-2
The Configuration utility .............................................................................................................. 1-3
The Traffic Management Shell (tmsh) ........................................................................................ 1-4
2
Components
Introduction ..................................................................................................................................... 2-1
Physical network components .................................................................................................... 2-2
Data centers ........................................................................................................................... 2-2
Servers ..................................................................................................................................... 2-2
Links ......................................................................................................................................... 2-2
Virtual servers ........................................................................................................................ 2-3
Logical network components ...................................................................................................... 2-4
Listeners .................................................................................................................................. 2-4
Pools ......................................................................................................................................... 2-4
Wide IPs .................................................................................................................................. 2-4
Distributed applications ....................................................................................................... 2-5
3
Setup and Configuration
Introduction ..................................................................................................................................... 3-1
Network Topology ........................................................................................................................ 3-2
Redundant system configuration ................................................................................................ 3-3
System communications ............................................................................................................... 3-4
Synchronization .............................................................................................................................. 3-6
Synchronization groups ....................................................................................................... 3-6
DNS zone file synchronization .......................................................................................... 3-7
Global monitor settings ................................................................................................................ 3-8
Heartbeat interval ................................................................................................................. 3-8
Synchronous monitor queries ............................................................................................ 3-9
Disabled resources ............................................................................................................... 3-9
Domain validation ........................................................................................................................ 3-10
4
Listeners
Introduction ..................................................................................................................................... 4-1
Node mode ............................................................................................................................ 4-1
Bridge or Router mode ....................................................................................................... 4-1
Wildcard listener .................................................................................................................. 4-1
Listeners and Network Traffic .................................................................................................... 4-3
Listeners and VLANs ..................................................................................................................... 4-4
5
The Physical Network
Introduction ..................................................................................................................................... 5-1
Data centers .................................................................................................................................... 5-2
Servers .............................................................................................................................................. 5-3
6
The Logical Network
Introduction ..................................................................................................................................... 6-1
Pools .................................................................................................................................................. 6-2
Virtual servers and Ratio mode load balancing .............................................................. 6-3
Canonical pool names .......................................................................................................... 6-3
Wide IPs ........................................................................................................................................... 6-5
Wildcard characters in wide IP names ............................................................................. 6-5
Wide IPs and pools ............................................................................................................... 6-6
Incorporating iRules ............................................................................................................. 6-7
NoError response for IPv6 resolution ............................................................................ 6-7
Distributed applications ................................................................................................................ 6-8
Dependencies for distributed applications ...................................................................... 6-8
Distributed application traffic ............................................................................................. 6-9
Persistent connections ....................................................................................................... 6-10
7
Load Balancing
About load balancing and Global Traffic Manager .................................................................. 7-1
Static load balancing modes ......................................................................................................... 7-3
Drop Packet mode ............................................................................................................... 7-3
Fallback IP mode .................................................................................................................... 7-4
Global Availability mode ...................................................................................................... 7-4
None mode ............................................................................................................................ 7-4
Ratio mode ............................................................................................................................. 7-5
Return to DNS mode .......................................................................................................... 7-5
Round Robin mode ............................................................................................................... 7-5
Static Persist mode ............................................................................................................... 7-5
Topology mode ..................................................................................................................... 7-6
Dynamic load balancing modes ................................................................................................... 7-6
Completion Rate mode ....................................................................................................... 7-6
CPU mode .............................................................................................................................. 7-6
Hops mode ............................................................................................................................. 7-7
Kilobyte/Second mode ......................................................................................................... 7-7
Least Connections mode .................................................................................................... 7-7
Packet Rate mode ................................................................................................................. 7-7
Quality of Service mode ...................................................................................................... 7-7
Round Trip Times mode ..................................................................................................... 7-8
Virtual Server Score mode ................................................................................................. 7-8
VS Capacity mode ................................................................................................................. 7-8
Dynamic Ratio option .......................................................................................................... 7-9
viii
Table of Contents
8
Connections
Connection management ............................................................................................................. 8-1
Resource health .............................................................................................................................. 8-2
Resource availability ...................................................................................................................... 8-3
Limit settings .......................................................................................................................... 8-3
Monitor availability requirements ..................................................................................... 8-3
Virtual server dependency .................................................................................................. 8-4
Restoration of availability ............................................................................................................. 8-5
Persistent connections .................................................................................................................. 8-6
Drain persistent requests option ...................................................................................... 8-6
Last resort pool .............................................................................................................................. 8-7
9
Topologies
Introduction ..................................................................................................................................... 9-1
IP geolocation data updates ......................................................................................................... 9-2
Topology records ........................................................................................................................... 9-3
Topology load balancing ............................................................................................................... 9-4
Longest Match load balancing option ............................................................................... 9-4
10
DNSSEC Keys and Zones
About DNSSEC ............................................................................................................................ 10-1
DNSSEC keys and zones ............................................................................................................ 10-1
Automatic key rollover ...................................................................................................... 10-1
DNSSEC resource records .............................................................................................. 10-3
11
Health and Performance Monitors
Introduction ................................................................................................................................... 11-1
Monitor types ...................................................................................................................... 11-2
Pre-configured and custom monitors ............................................................................ 11-2
Special configuration considerations ........................................................................................ 11-5
Monitor destinations .......................................................................................................... 11-5
Transparent and reverse modes ..................................................................................... 11-5
Virtual server status ........................................................................................................... 11-7
Monitors and resources ............................................................................................................. 11-7
Monitor associations .......................................................................................................... 11-8
12
Statistics
Introduction ................................................................................................................................... 12-1
Statistics access ............................................................................................................................. 12-2
Status Summary screen ............................................................................................................... 12-2
Types of statistics ......................................................................................................................... 12-3
Distributed application statistics ..................................................................................... 12-3
Wide IP statistics ................................................................................................................. 12-5
Pool statistics ....................................................................................................................... 12-6
13
Metric Collection
Introduction ................................................................................................................................... 13-1
About metrics ............................................................................................................................... 13-2
Probes and local DNS servers .................................................................................................. 13-3
TTL and timer values .................................................................................................................. 13-5
14
Performance Data
Introduction ................................................................................................................................... 14-1
Performance data graphs ............................................................................................................ 14-1
Performance graph ............................................................................................................. 14-1
Request Breakdown graph ................................................................................................ 14-1
15
iRules
Introduction ................................................................................................................................... 15-1
What is an iRule? .......................................................................................................................... 15-2
Event-based traffic management ............................................................................................... 15-3
Event declarations ............................................................................................................... 15-3
16
ZoneRunner
ZoneRunner utility ....................................................................................................................... 16-1
ZoneRunner tasks ............................................................................................................... 16-1
Zone files ....................................................................................................................................... 16-2
Types of zone files .............................................................................................................. 16-2
Zone file import .................................................................................................................. 16-2
Resource records ......................................................................................................................... 16-4
Types of resource records ............................................................................................... 16-4
Views ............................................................................................................................................... 16-6
Named.conf ................................................................................................................................... 16-7
A
big3d Agent
Introduction .....................................................................................................................................A-1
Metrics ..............................................................................................................................................A-2
Data collection with the big3d agent ................................................................................A-3
Data collection and broadcast sequence .........................................................................A-3
Communications ............................................................................................................................A-5
iQuery and the big3d agent ................................................................................................A-5
iQuery and firewalls .............................................................................................................A-6
Communications between Global Traffic Managers, big3d agents, and
local DNS servers .................................................................................................................A-7
x
Table of Contents
B
Probes
Introduction ..................................................................................................................................... B-1
About iQuery .................................................................................................................................. B-2
Probe responsibility ....................................................................................................................... B-3
Probes and the big3d agent .......................................................................................................... B-5
LDNS probes .................................................................................................................................. B-7
Probes and log entries .................................................................................................................. B-9
Probe information in the log file ........................................................................................ B-9
Glossary
Index
xii
1
Overview
Security features
Global Traffic Manager offers a variety of security features that can help
prevent hostile attacks on your site or equipment.
◆ Secure administrative connections
Global Traffic Manager supports Secure Shell (SSH) administrative
connections for remote administration from the command line. The web
server, which hosts the web-based Configuration utility, supports SSL
connections as well as user authentication.
◆ Secure iQuery communications
Global Traffic Manager supports web certificate authentication for
BIG-IP iQuery® protocol communications between itself and other
systems running the big3d agent.
◆ TCP wrappers
Global Traffic Manager supports the use of TCP wrappers to provide an
extra layer of security for network connections.
1-2
Overview
1-4
2
Components
• Introduction
Introduction
For the BIG-IP® Global Traffic Manager™ system to operate effectively,
you need to define the components that make up the segments of your
network. These components include physical components, such as data
centers and servers, as well as logical components, such as wide IPs,
addresses, and pools. By defining these components, you essentially build a
network map that Global Traffic Manager can use to direct Domain Name
System (DNS) traffic to the best available resource.
The most basic configuration of Global Traffic Manager includes:
• A listener that is a specific virtual server that identifies network traffic
for global traffic management
• A data center that contains at least one server
• A server that contains at least one resource or virtual server
Data centers
Data centers are the top level of your physical network setup. You must
configure one data center for each physical location in your global network.
When you create a data center in Global Traffic Manager, you define the
servers (Global Traffic Manager systems, Local Traffic Manager™ systems,
Link Controller™ systems, hosts, and routers) that reside at that location.
A data center can contain any type of server. For example, one data center
can contain a Global Traffic Manager system and a host, while another
might contain two Global Traffic Manager systems and eight Local Traffic
Manager systems.
Servers
A server is a physical device on which you can configure one or more
virtual servers. The servers that you define can include both BIG-IP systems
and third-party servers, such as Local Traffic Manager systems and systems
running Microsoft® Windows® 2000 Server.
One server that you must define is Global Traffic Manager. This places the
system on the network map.
Links
A link is a logical representation of a physical device (router) that connects
your network to the Internet. You can assign multiple links to each data
center by logically attaching links to a collection of servers in order to
manage access to your data sources. Configuring links is optional, although
they are very useful when determining resource availability.
2-2
Components
Virtual servers
Servers, excluding Global Traffic Manager systems and Link Controller
systems, contain at least one virtual server. A virtual server, in the context
of Global Traffic Manager, is a combination of an IP address and a port
number that points to a resource that provides access to an application or
data source on your network. In the case of host servers, this IP address and
port number likely point to the resource itself. With load balancing systems,
such as Local Traffic Manager, these virtual servers are often proxies that
allow the load balancing server to manage the resource request across a
multitude of resources. Virtual servers are the ultimate destination for
connection requests.
Listeners
To communicate with the rest of your network, you must configure Global
Traffic Manager so that it can correctly identify the resolution requests for
which it is responsible. A listener is an object that monitors the network for
DNS queries, and thus is critical for global traffic management. The listener
instructs the system to monitor the network traffic destined for a specific IP
address.
In most installations, when you define a listener for Global Traffic Manager,
you use the IP address of Global Traffic Manager; however, there are many
different ways you can configure listeners so that the system handles DNS
traffic correctly.
Pools
A pool is a collection of virtual servers that can reside on multiple network
servers. When you define the virtual servers to which Global Traffic
Manager directs DNS traffic, you combine those virtual servers into pools.
You can then configure Global Traffic Manager to direct traffic to a specific
virtual server within a pool, using a specific load balancing method.
You can apply a different set of options to the same resources as a virtual
server. When you add a virtual server to a pool, it becomes a pool member
to which you can apply monitors, iRules®, and other configuration options.
Wide IPs
One of the most common logical components you create in Global Traffic
Manager is a wide IP. A wide IP maps a fully-qualified domain name to one
or more pools of virtual servers that host the domain’s content.
2-4
Components
Distributed applications
A distributed application is a collection of one or more wide IPs, data
centers, and links that serve as a single application to a web site visitor. A
distributed application is the highest-level component that Global Traffic
Manager supports. You can configure Global Traffic Manager so that the
availability of distributed applications is dependent on a specific data center,
link, or server. For example, if the New York data center goes offline, this
information causes the wide IP and its corresponding distributed application
to become unavailable. Consequently, the system does not send resolution
requests to any of the distributed application resources, until the entire
application becomes available again.
2-6
3
Setup and Configuration
• Introduction
• Network Topology
• System communications
• Synchronization
• Domain validation
Setup and Configuration
Introduction
When you install a BIG-IP® Global Traffic Manager™ system on the
network, the actions you take to integrate it into the network fall into two
categories: setup tasks and configuration tasks.
Setup tasks are tasks that apply either to Global Traffic Manager itself, or
universally to all other components that you configure later, such as servers,
data centers, and wide IPs. Examples of setup tasks include running the
Setup utility. This utility guides you through licensing the product, assigning
an IP address to the management port of the system, assigning self IP
addresses, enabling high-availability, and configuring the passwords for the
root and administrator accounts.
Configuration tasks are tasks in which you define how you want Global
Traffic Manager to manage global traffic, such as load balancing methods,
pools and pool members, and iRules®. These tasks affect specific aspects of
how you want the system to manage Domain Name System (DNS) traffic.
Network Topology
Global Traffic Manager is designed to manage DNS traffic as it moves from
outside the network, to the appropriate resource, and back again. The
management capabilities of the system require that it has an accurate
definition of the sections of the network over which it has jurisdiction. You
must define network elements such as data centers, servers (including
BIG-IP systems), and virtual servers in Global Traffic Manager. Defining
these elements is similar to drawing a network diagram; you include all of
the relevant components in such a diagram in order to have an accurate
depiction of how the system works as a whole.
Note
3-2
Setup and Configuration
System communications
Before Global Traffic Manager can operate as an integrated component
within your network, you must first establish how it can communicate with
other external systems. An external system is any server with which Global
Traffic Manager must exchange information to perform its functions. In
general, system communications are established for the purpose of:
• Communicating with other BIG-IP systems
• Communicating with third-party systems
Note
Protocols iQuery
3-4
Setup and Configuration
Protocols SNMP
Synchronization
The primary goal of Global Traffic Manager is to ensure that name
resolution requests are sent to the best available resource on the network.
Consequently, it is typical for multiple Global Traffic Manager systems to
reside in several locations within a network. For example, a standard
installation might include a Global Traffic Manager system at each data
center within an organization.
When an LDNS submits a name resolution request, you cannot control to
which Global Traffic Manager the request is sent. As a result, you often
want multiple Global Traffic Manager systems to share the same
configuration values, and maintain those configurations over time.
In network configurations that contain more than one Global Traffic
Manager, synchronization means that each Global Traffic Manager
regularly compares the timestamps of its configuration files with the
timestamps of configuration files on other Global Traffic Manager systems.
If Global Traffic Manager determines that its configuration files are older
than those on another system, it acquires the newer files and begins using
them to load balance name resolution requests. With synchronization, you
can change settings on one system and have that change distributed to all
other systems.
Synchronization groups
You can separate the Global Traffic Manager systems on your network into
separate groups, called synchronization groups. A synchronization group is
a collection of multiple Global Traffic Manager systems that share and
synchronize configuration settings. These groups are identified by a
synchronization group name, and only systems that share this name also
shares configuration settings. These synchronization groups allow you to
customize the synchronization behavior. For example, Global Traffic
Manager systems residing in data centers in Europe might belong to one
synchronization group, while the systems in North America belong to
another group.
Initially, when you enable synchronization for Global Traffic Manager, the
system belongs to a synchronization group called default. However, you
can create new groups at any time to customize the synchronization process,
ensuring that only certain sets of Global Traffic Manager systems share
configuration values.
To illustrate how synchronization groups work, consider the fictional
company, SiteRequest. SiteRequest has decided to add a new data center in
Los Angeles. As part of bringing this data center online, SiteRequest has
decided that it wants the Global Traffic Manager systems installed in New
York and in Los Angeles to share configurations, and the Paris and Tokyo
data centers to share configurations. This setup exists because SiteRequest’s
network optimization processes require slightly different settings within the
United States than the rest of the world. To accommodate this new network
3-6
Setup and Configuration
While monitors supply information you need to ensure that network traffic
moves efficiently across the network, they do so at the cost of increasing
that network traffic. These settings allow you to control this increase.
Heartbeat interval
In daily operations, Global Traffic Manager frequently acquires much of its
network data from other BIG-IP systems that you employ, such as Local
Traffic Manager systems. For example, the Local Traffic Manager system
monitors the resources it manages. When Global Traffic Manager requires
this same information for load balancing DNS requests, it can query Local
Traffic Manager, instead of each resource itself. This process ensures that
the system efficiently acquires the information it needs.
Because Global Traffic Manager queries other BIG-IP systems to gather
information, you can configure the frequency at which these queries occur,
by configuring the Heartbeat Interval setting. Based on the value you
specify for this setting, Global Traffic Manager queries other BIG-IP
3-8
Setup and Configuration
Tip
F5 Networks recommends that, when configuring resource monitors, you
ensure that the frequency at which the monitor attempts to query a resource
is greater than the value of the Heartbeat Interval setting. Otherwise, the
monitor might acquire out-of-date data during a query.
Disabled resources
One of the ways a given network resource can become unavailable during
the load balancing of DNS traffic is when you manually disable the
resource. You might disable a resource because you are upgrading the server
on which it resides, or because you are modifying the resource itself and
need to remove it temporarily from service.
You can control whether Global Traffic Manager monitors these disabled
resources. In some network configurations, for example, you might want to
continue monitoring these resources when you put them offline.
Note
Domain validation
Global Traffic Manager handles traffic using DNS and BIND to translate
domain names into IP addresses. By configuring the Domain Validation
setting, you can specify which domain names Global Traffic Manager
recognizes. You can configure the system so that it accepts all domain
names, or you can restrict the use of certain characters in domain names.
3 - 10
4
Listeners
• Introduction
Introduction
Before you can fully configure Global Traffic Manager™ to handle name
resolution requests, you must determine how you want the system to
integrate with the existing network. Specifically, you must identify what
network traffic you want Global Traffic Manager to handle and how. In
general, the system performs global traffic management in two ways: Node
mode and Bridge or Router mode.
Node mode
Typically, when you add a Global Traffic Manager system to your network,
you want the system to respond to at least a subset of your incoming DNS
requests. You can configure the system to direct the requests to the wide IPs
that are configured on Global Traffic Manager; however, you can also
configure the system to respond to DNS requests for other network
resources that are not associated with a wide IP, such as other DNS servers.
When Global Traffic Manager receives traffic, processes it locally, and
sends the appropriate Domain Name System (DNS) response back to the
querying server, it is operating in Node mode. In this situation, you create a
listener that corresponds to an IP address on the system. If Global Traffic
Manager operates as a standalone unit, this IP address is the self IP address
of the system. If Global Traffic Manager is part of a redundant system
configuration for high availability purposes, this IP address is the floating IP
address that belongs to both systems.
Wildcard listener
In some cases, you might want Global Traffic Manager to handle the traffic
coming into your network, regardless of the destination IP address of the
given DNS request. In this configuration, Global Traffic Manager continues
to process and respond to requests for the wide IPs that you configure, but is
also responsible for forwarding additional DNS requests to other network
resources, such as DNS servers. To accomplish this type of configuration,
you create a wildcard listener.
4-2
Listeners
Tip
If you are familiar with Local Traffic Manager™, it might be helpful to
consider a listener as a specialized type of virtual server that is responsible
for handling traffic for Global Traffic Manager.
Note
If you configure user accounts on Local Traffic Manager, you can assign
listeners, like other virtual servers, to specific partitions. However, because
listeners play an important role in global traffic management, F5 Networks
recommends that you assign all listeners to partition Common.
As you can see from this example, the role that Global Traffic Manager
plays in managing DNS traffic varies depending on the listener through
which the traffic arrives. As a result, Global Traffic Manager becomes a
flexible system for managing DNS traffic in a variety of ways.
4-4
5
The Physical Network
• Introduction
• Data centers
• Servers
• Virtual servers
• Links
The Physical Network
Introduction
The components that make up Global Traffic Manager™ can be divided into
two categories: logical network components and physical networks
components. Logical network components are abstractions of network
resources, such as virtual servers. Physical network components have a
direct correlation with one or more physical entities on the network. This
chapter deals with the physical components of Global Traffic Manager, and
describes how to use Global Traffic Manager to define the following
physical network components that make up your network:
• Data centers
• Servers
• Virtual servers
• Links
Data centers
A data center defines the servers and links that share the same subnet on the
network. All resources on your network, whether physical or logical, are
associated in some way with a data center. Global Traffic Manager
consolidates the paths and metrics data collected from servers, virtual
servers, and links into the data center, and uses that data to conduct load
balancing operations.
Depending on your router configuration, the following data center
configurations are available:
• One data center in one physical location
• One data center that includes servers in multiple physical locations
• Multiple data centers in one physical location
5-2
The Physical Network
Servers
A server defines a specific physical system on the network. Within Global
Traffic Manager, servers are not only physical entities that you can
configure and modify as needed; servers also contain the virtual servers that
are the ultimate destinations of name resolution requests. When you
configure a server on Global Traffic Manager, unless the server is either a
Global Traffic Manager system or a Link Controller™ system, the server
must contain at least one virtual server.
Global Traffic Manager supports three types of servers:
◆ BIG-IP systems
A BIG-IP® system can be a Global Traffic Manager system, a Local
Traffic Manager™ system, a Link Controller system, or a VIPRION®
system.
◆ Third-party load balancing systems
A third-party load balancing system is any system, other than a BIG-IP
system, that supports and manages virtual servers on the network.
◆ Third-party host servers
A third-party host system is any server on the network that does not
support virtual servers.
Important
You must use a self IP address when you define Global Traffic Manager.
You cannot use the management IP address.
Note
If your network uses a load balancing server that is not found on this list,
you can use the Generic Load Balancer option.
5-4
The Physical Network
Note
If your network uses a host server that is not on this list, you can use the
Generic Host option.
Availability thresholds
When you set thresholds for availability, Global Traffic Manager can detect
when a managed server is low on resources, and redirect the traffic to
another server. Setting limits can help eliminate any negative impact on a
server's performance of tasks that may be time critical, require high
bandwidth, or put high demand on system resources. The system resources
vary depending on the monitors you assign to the server.
You can specify thresholds for the following components:
• Servers
• Virtual servers
• Pools
• Pool members
Server thresholds
When you configure a server, you can set limits for specific elements
depending upon whether the server is part of the BIG-IP product family,
such as Local Traffic Manager, or another server type. If the server is part of
the BIG-IP product family, you can base thresholds on:
• Bits per second
• Packets per second
• Current connections
If the server is not part of the BIG-IP product family, such as a generic host
server, you can base thresholds on:
• CPU
• Memory
• Bits
• Packets
• Current connections
If a server meets or exceeds its limits, both the server and the virtual servers
it manages are marked as unavailable for load balancing. You can quickly
review the availability of any of your servers or virtual servers on the
Statistics screens.
Pool thresholds
When you configure a pool, you can set thresholds for:
• Bits per second
• Packets per second
• Current connections
If a pool meets or exceeds its limits, both the pool and the pool members it
manages are marked as unavailable for load balancing. You can quickly
review the availability of any of your pools or pool members on the
Statistics screens.
5-6
The Physical Network
Virtual servers
Servers, excluding Global Traffic Manager systems and Link Controller
systems, contain at least one virtual server. A virtual server, in the context
of Global Traffic Manager, is a specific IP address and port number that
points to a resource on the network. In the case of host servers, this IP
address and port number likely point to the resource itself. With load
balancing systems, such as Local Traffic Manager, these virtual servers are
often proxies that allow the load balancing server to manage the resource
request across a multitude of resources.
You can add virtual servers in two ways:
• Automatically, through the use of the discovery feature.
• Manually, through the properties screens of the given server.
5-8
The Physical Network
Links
A link defines a physical connection to the Internet that is associated with
one or more routers on the network. Global Traffic Manager tracks the
performance of links, which in turn can dictate the overall availability of a
given pool, data center, wide IP, or distributed application.
To configure the links that you want Global Traffic Manager to load
balance, you add a link entry, and then associate one or more routers with
that entry. You can also configure monitors to check certain metrics
associated with a link, and modify how the system load balances network
traffic across links.
Important
You can use either the Ratio or Price (Dynamic Ratio) weighting option to
load balance the traffic through all of your links. You must use the same
weighting option for all links.
5 - 10
6
The Logical Network
• Introduction
• Pools
• Wide IPs
• Distributed applications
The Logical Network
Introduction
After you define the physical components of your network, such as data
centers, servers, and links, you can configure Global Traffic Manager™
with the logical components of your network. Logical components are
abstractions of network resources, such as a virtual servers. Unlike physical
components, the logical network can often span multiple physical devices,
or encompass a subsection of a single device.
Through Global Traffic Manager, you define three primary types of logical
network components:
• Pools
• Wide IPs
• Distributed applications
To better understand the interactions between pools, wide IPs, and data
centers, consider the fictional company of SiteRequest. SiteRequest is an
online application repository. Currently, its presence on the World Wide
Web consists of a main site, www.siterequest.com; a download area,
downloads.siterequest.com; and a search area, search.siterequest.com.
These three fully-qualified domain names (FQDNs), www.siterequest.com,
downloads.siterequest.com, and search.siterequest.com, are wide IPs.
Each of these wide IPs contain several pools of virtual servers. For example,
www.siterequest.com contains two pools of virtual servers: poolMain, and
poolBackup. When Global Traffic Manager receives a connection request
for www.siterequest.com, it applies its load balancing logic to select the
appropriate pool to handle the request.
After Global Traffic Manager selects a pool, it then load balances the
request to the appropriate virtual server. For example, mainPool contains
three virtual servers: 192.168.3.10:80, 192.168.4.20:80, and
192.168.5.30:80. Global Traffic Manager responds to the system that made
the connection request with the selected virtual server. At this point, Global
Traffic Manager steps out of the communication, and the system requesting
the resource communicates directly with the virtual server.
Note
Pools
A pool represents one or more virtual servers that share a common role on
the network. A virtual server, in the context of Global Traffic Manager, is a
combination of IP address and port number that points to a specific resource
on the network.
Global Traffic Manager considers any virtual servers that you add to a pool
to be pool members. A pool member is a virtual server that has specific
attributes that pertain to the virtual server only in the context of that pool.
Through this differentiation, you can customize settings, such as thresholds,
dependencies, and health monitors, for a given virtual server on a per-pool
basis.
As an example of the difference between pool members and virtual servers,
consider the fictional company SiteRequest. In the London data center, the
IT team has a virtual server that acts as a proxy for a Local Traffic
Manager™ system. This virtual server is the main resource for name
resolution requests for the company’s main web page that originate from
Europe. This same virtual server is the backup resource for name resolution
requests that originate from the United States. Because these are two
distinctly different roles, the virtual server is a pool member in two different
pools. This configuration allows the IT team to customize the virtual server
for each pool to which it belongs, without modifying the actual virtual
server itself.
Before you can add virtual servers to Global Traffic Manager, you must
define a server that represents a physical component of your network. Then
you can add virtual servers to the server, and group the virtual servers in
pools.
When you create a pool, you name it and add at least one virtual server as a
member of the pool. You can also assign specific load balancing methods, a
fallback IP address, and one or more health monitors to the pool. You assign
a fallback IP address in the event that the load balancing methods you assign
to the pool fail to return a valid virtual server. The health monitors that you
assign to the pool use various methods to determine if the virtual servers
within the pool are available.
Certain load balancing methods within Global Traffic Manager select virtual
servers based on the order in which they are listed in the pool. For example,
the load balancing method, Global Availability, instructs Global Traffic
Manager to select the first virtual server in the pool until it reaches capacity
or goes offline, at which point it selects the next virtual server until the first
pool becomes available again.
If you use a load balancing method that selects virtual servers based on the
order in which they are listed in the pool, you may want to change the order
in which the virtual servers are listed in the Member List. When you
organize your virtual servers in conjunction with these load balancing
methods, you can ensure that your most robust virtual server always
receives resolution requests, while the other virtual servers act as backups in
case the primary virtual server becomes unavailable.
6-2
The Logical Network
This pattern exists because the weight value, 50, is 50 percent of the total
weight for all virtual servers (100), while the weight value, 25, is 25 percent.
6-4
The Logical Network
Wide IPs
A wide IP maps a fully-qualified domain name (FQDN) to a set of virtual
servers that host the domain’s content, such as a web site, an e-commerce
site, or a CDN. Wide IPs use pools to organize virtual servers, which creates
a tiered load balancing effect: Global Traffic Manager first load balances
requests to the appropriate pool of a wide IP, and then load balances within
the pool to the appropriate virtual server.
◆ The asterisk ( * )
• Use the asterisk to replace multiple consecutive characters, with the
exception of dots ( . ).
• Use more than one asterisk in a wide IP name or alias.
• Use both the question mark and the asterisk in the same wide IP name
or alias.
The following examples are all valid uses of the wildcard characters for the
wide IP name, www.mydomain.net.
• ???.mydomain.net
• www.??domain.net
• www.my*.net
• www.??*.net
• www.my*.*
• ???.my*.*
• *.*.net
• www.*.???
Tip
You can assign the same pool to multiple wide IPs.
6-6
The Logical Network
Notice that the total of all the weights in this wide IP is 100. Each time
Global Traffic Manager selects this wide IP, it load balances across all three
pools. Over time, the load balancing statistics for this wide IP appear as
follows:
• Pool 1: selected 50 percent of the time
• Pool 2: selected 25 percent of the time
• Pool 3: selected 25 percent of the time
This pattern exists because the weight value, 50, is 50 percent of the total
weight for all pools, while the weight value, 25, is 25 percent of the total.
Incorporating iRules
An iRule is a set of one or more Tcl-based expressions that you can use with
wide IPs to customize how Global Traffic Manager handles network
connection requests.
You can add or remove an iRule to a wide IP at any time. When you add an
iRule to a wide IP, Global Traffic Manager uses the iRule to determine how
to load balance name resolution requests. Removing an iRule does not
delete it from Global Traffic Manager; you can still access the iRule by
clicking iRules under Global Traffic on the Main tab.
You can also customize a wide IP using more than one iRule. For example,
a wide IP might have an iRule that focuses on the geographical source of the
name resolution request, and another that focuses on redirecting specific
requests to a different wide IP. If you assign more than one iRule to a wide
IP, Global Traffic Manager applies iRules® in the order in which they are
listed in the iRules List for the wide IP.
You can change the order in which Global Traffic Manager applies iRules to
network connection requests at any time.
Distributed applications
A distributed application is a collection of wide IPs that serves as a single
application to a site visitor. Within Global Traffic Manager, distributed
applications provide you with several advantages:
◆ You can organize logical network components into groups that represent
the business environment for which these components were designed.
◆ You can configure a distributed application so that it is dependent on a
physical component of your network, such as a data center, server, or
link. If this physical component becomes unavailable, Global Traffic
Manager flags the distributed application as unavailable as well. These
dependencies ensure that a user cannot access a distributed application
that does not have all of its resources available.
◆ You can define persistence for a distributed application, ensuring that a
user accessing the distributed application uses the same network
resources until they end their session.
When you create a distributed application, you name it and add at least one
wide IP. You can also configure the distributed application so that its
availability depends on the availability of specific servers, virtual servers, or
data centers. Additionally, you can configure whether the system routes
requests coming from the same source during a specific time period to the
same pool, or to a different pool. This is known as persistence.
6-8
The Logical Network
Note
You do not have to set a dependency for a distributed application. You can
accept the default value of None. If you do not set a dependency, then
Global Traffic Manager considers the application available as long as there
is at least one wide IP to which it can load balance a name resolution
request.
Note
Persistent connections
Many distributed applications require that users access a single set of
resources until they complete their transaction. For example, customers
purchasing a product online might need to remain with the same data center
until they finish their order. In the context of Global Traffic Manager, this
requirement is called persistence. Persistence is the state in which a user of
the system remains with the same set of resources until the user closes the
connection.
When you enable persistence for a distributed application, and an LDNS
makes repetitive requests on behalf of a client, the system reconnects the
client to the same resource to which it was connected for previous requests.
For persistence to work correctly for a distributed application, you must also
specify a dependency level. This is because a connection to the distributed
application persists to the dependency object you specify (that is, the
specified wide IP, server, data center, or link), and not to a specific pool
member.
6 - 10
7
Load Balancing
Load balancing mode Use for wide IP Use for preferred Use for alternate Use for fallback
(static) load balancing method method method
Drop Packet X X X
Fallback IP X X X
Global Availability X X X X
None X X
Ratio X X X X
Return to DNS X X X
Round Robin X X X X
Static Persist X X X
Topology X X X X
Load balancing mode Use for wide IP Use for preferred Use for alternate Use for fallback
(dynamic) load balancing method method method
Completion Rate X X
CPU X X
Hops X X
Kilobytes/Second X X
7-2
Load Balancing
Load balancing mode Use for wide IP Use for preferred Use for alternate Use for fallback
(dynamic) load balancing method method method
Least Connections X X
Packet Rate X X X
Quality of Service X X
VS Capacity X X X
The None and Return to DNS modes are special modes that you can use to
skip load balancing under certain conditions. The other static load balancing
modes perform true load balancing.
Fallback IP mode
When you choose the Fallback IP mode, Global Traffic Manager answers a
query by returning the IP address that you specify as the fallback IP. The IP
address that you specify is not monitored for availability before being
returned as an answer. When you use the Fallback IP mode, you can specify
that Global Traffic Manager return a disaster recovery site when no load
balancing mode returns an available virtual server. F5 Networks
recommends that you use the Fallback IP mode only for the fallback load
balancing method. Global Traffic Manager uses the fallback method only
when the preferred and alternate methods do not provide at least one virtual
server to return as an answer to a query.
None mode
The None mode is a special mode you can use if you want to skip the
current load balancing method, or skip to the next pool in a multiple pool
configuration. For example, if you set an alternate method to None in a pool,
Global Traffic Manager skips the alternate method and immediately tries the
mode specified as the fallback method. If the fallback method is set to None,
and you have multiple pools configured, Global Traffic Manager uses the
next available pool. If all pools become unavailable, Global Traffic Manager
returns an aggregate of the IP addresses of all pool members using BIND.
Tip
If you do not want Global Traffic Manager to return multiple addresses that
are potentially unavailable, F5 Networks recommends that you set the
alternate method to Drop Packet.
You can also use this mode to limit each pool to a single load balancing
mode. For example, you can set the preferred method in each pool to the
desired mode, and then you can set both the alternate and fallback methods
to None in each pool. If the preferred method fails, the None value for both
the alternate and fallback methods forces Global Traffic Manager to go to
the next pool for a load balancing answer.
7-4
Load Balancing
Ratio mode
The Ratio mode distributes connections among a pool of virtual servers as a
weighted round robin. Weighted round robin refers to a load balancing
pattern in which Global Traffic Manager rotates connection requests among
several resources based on a priority level, or weight, assigned to each
resource. For example, you can configure the Ratio mode to send twice as
many connections to a fast, new server, and only half as many connections
to an older, slower server.
The Ratio mode requires that you define a ratio weight for each virtual
server in a pool, or for each pool if you are load balancing requests among
multiple pools. The default ratio weight for a server or a pool is set to 1.
Topology mode
The Topology mode allows you to direct or restrict traffic flow by adding
topology records to a topology statement in the configuration file. When you
use the Topology mode, you can develop proximity-based load balancing.
For example, a client request in a particular geographic region can be
directed to a data center or server within that same region. Global Traffic
Manager determines the proximity of servers by comparing location
information derived from the DNS message to the topology records.
This load balancing mode requires you to do some advanced configuration
planning, such as gathering the information you need to define the topology
records. Global Traffic Manager contains an IP classifier that accurately
maps the LDNS, so when you create topology records, you can refer to
continents and countries, instead of IP subnets.
CPU mode
The CPU load mode selects the virtual server that currently has the most
CPU processing time available to handle name resolution requests.
7-6
Load Balancing
Hops mode
The Hops mode is based on the traceroute utility, and tracks the number of
intermediate system transitions (router hops) between a client’s LDNS and
each data center. Hops mode selects a virtual server in the data center that
has the fewest router hops from the LDNS.
Kilobyte/Second mode
The Kilobytes/Second mode selects the virtual server that is currently
processing the fewest number of kilobytes per second. You can use this load
balancing mode only with servers for which Global Traffic Manager can
collect the kilobytes per second metric.
VS Capacity mode
The VS Capacity mode creates a list of the virtual servers, weighted by
capacity, then picks one of the virtual servers from the list. The virtual
servers with the greatest capacity are picked most often, but over time all
virtual servers are returned. If more than one virtual server has the same
capacity, then Global Traffic Manager load balances using the Round Robin
mode among those virtual servers.
7-8
Load Balancing
7 - 10
Load Balancing
Enabling the Ignore Path TTL option instructs Global Traffic Manager to
use path information gathered during metrics collection even if the
time-to-live value for that information has expired. This option is often used
when you want the system to continue using a dynamic load balancing mode
even if some metrics data is temporarily unavailable, and you want Global
Traffic Manager to use old metric data rather than employ an alternate load
balancing method. This option is disabled by default.
The Verify Virtual Server Availability option instructs Global Traffic
Manager to verify that a virtual server is available before returning it as a
response to a name solution request. If this option is disabled, the system
responds to a name resolution request with the virtual server’s IP address
regardless of whether the server is up or down. This option is enabled by
default,and is rarely disabled outside of a test or staging environment.
7 - 12
8
Connections
• Connection management
• Resource health
• Resource availability
• Restoration of availability
• Persistent connections
Connection management
When you integrate a Global Traffic Manager™ system into your network,
one of its primary responsibilities is to load balance incoming connection
requests to the virtual server resource that best fits the configuration
parameters you defined. However, load balancing is only one part of
managing connections to your network resources. Additional issues that you
must consider include:
◆ Resource health
Resource health refers to the ability of a given resource to handle
incoming connection requests. For example, the Configuration utility
uses a green circle to identify a resource, such as a wide IP, that has
available pools and virtual servers, while a pool that is down appears as a
red diamond. These visual clues can help you identify connection issues
quickly and efficiently.
◆ Resource availability
Resource availability refers to the settings within the Configuration
utility that you use to control when a resource is available for connection
request. For example, you can establish limit settings, which instruct
Global Traffic Manager to consider a resource as unavailable when a
statistical threshold (such as CPU usage) is reached.
◆ Restoring availability
When a resource goes offline, Global Traffic Manager immediately
sends incoming connection requests to the next applicable resource.
When you bring that resource online again, you can control how to
restore its availability to Global Traffic Manager, ensuring that
connections are sent to the resource only when it is fully ready to receive
them.
◆ Persisting connections
Certain interactions with your network require that a given user access
the same virtual server resource until their connection is completed. An
example of this situation is an online store, in which you want the user to
access the same virtual server for their shopping cart until they place
their order. With Global Traffic Manager, you can configure your load
balancing operations to take persistent connections into account.
◆ Selecting a last resort pool
Global Traffic Manager includes the ability to create a last resort pool. A
last resort pool is a collection of virtual servers that are not used during
normal load balancing operations. Instead, these virtual servers are held
in reserve unless all other pools for a given wide IP become unavailable.
Resource health
In Global Traffic Manager, resource health refers to the ability of a given
resource to handle incoming connection requests. Global Traffic Manager
determines this health through the use of limit settings, monitors, and
dependencies on other network resources.
The health of a resource is indicated by a status code in the Configuration
utility. A status code is a visual representation of the availability of a given
resource. Global Traffic Manager displays these status codes in the main
screens for a given resource. The types of status codes available for a
resource are:
◆ Blue
A blue status code indicates that the resource has not been checked. This
status often appears when you first add a resource into the Configuration
utility.
◆ Green
A green status code indicates that the resource is available and
operational. Global Traffic Manager uses this resource to manage traffic
as appropriate.
◆ Red
A red status code indicates that the resource did not respond as expected
to a monitor. Global Traffic Manager uses this resource only when two
conditions are met:
• Global Traffic Manager is using the load balancing mode specified in
the Fallback load balancing setting.
• The Fallback load balancing setting for the pool is not None.
◆ Yellow
A yellow status code indicates that the resource is operational, but has
exceeded one of its established bandwidth thresholds. Global Traffic
Manager uses a resource that has a yellow status code only if no other
resource is available.
◆ Black
A black status code indicates that the resource has been manually
disabled and is no longer available for load balancing operations.
As the preceding list illustrates, the health of a resource does not necessarily
impact the availability of that resource. For example, Global Traffic
Manager can select a virtual server that has a red status code.
8-2
Connections
Resource availability
To load balance effectively, Global Traffic Manager must determine
whether the appropriate resources are available. In the context of the
Global Traffic Manager, availability means that the resource meets one or
more sets of pre-defined requirements. These requirements can be a set of
statistical thresholds, a dependency on another resource, or set of values
returned by a monitoring agent. If a resource fails to meet one or more of
these requirements, Global Traffic Manager considers it unavailable
and attempts to select the next resource based on the load balancing
methodology you defined.
Global Traffic Manager includes three methods of determining resource
availability:
• Limit settings
• Monitor availability requirements
• Virtual server dependencies
Limit settings
One of the methods for determining the availability of a resource is to
establish limit settings. A limit setting is a threshold for a particular statistic
associated with a system.
Global Traffic Manager supports the following limit settings:
• Kilobytes
• Packets
• Total Connections
the server to verify that the main index.html page is available. If the
monitor cannot access the page, it notifies Global Traffic Manager, which
then considers the server unavailable until the monitor is successful.
Monitors provide a robust, customizable means of determining the
availability of a given resource with Global Traffic Manager. The following
procedure describes how to control the impact that a set of monitors has on
the availability of a resource.
You can also assign monitors to a specific server. In most cases, when you
assign a monitor to a server, that monitor checks all virtual servers
associated with that server.
An exception to this guideline is the SNMP monitor. If you assign an SNMP
monitor to a Cisco®, Alteon®, Extreme Networks®, Foundry®, or Radware
server, that monitor obtains information on the virtual servers associated
with that server. If you assign the SNMP monitor to any other server type,
that monitor obtains data on the server itself.
In cases where you assign a monitor to a virtual server both directly and to
its parent server, the availability information acquired from the monitor
directly assigned to the virtual server takes precedence over any other data.
8-4
Connections
Restoration of availability
When a network resource, such as a virtual server, goes offline, Global
Traffic Manager considers that resource to be unavailable and proceeds to
send name resolution requests to other resources based on the configured
load balancing mode. By default, Global Traffic Manager resumes sending
requests to an offline resource as soon as that the resource becomes
available again, provided that the resource meets the appropriate load
balancing requirements.
Under certain circumstances, you might not want Global Traffic Manager to
resume connections to a resource immediately. For example, a server for the
fictional company, SiteRequest, goes offline. Global Traffic Manager
detects that the virtual servers associated with this server are unavailable,
and proceeds to send name resolution requests to other virtual servers as
appropriate. When the server is online again, it must still run several
synchronization processes before it is fully ready to handle name resolution
requests. However, Global Traffic Manager might detect that the server is
available before these processes are complete, and send requests to the
server before that server can handle them.
To avoid this possibility, you can configure pools to use the manual resume
feature. The manual resume feature ensures that Global Traffic Manager
does not load balance requests to a virtual server within a pool until you
manually re-enable it.
Persistent connections
Most load balancing modes divide name resolution requests among
available pools or virtual servers. Each time Global Traffic Manager
receives a request, it sends that request to the most appropriate resource
based on the configuration of your network. For example, when a user visits
a web site, it results in multiple name resolution requests as that user moves
from page to page. Depending on the load balancing mode selected, the
system sends each request to a completely different server, virtual server, or
data center.
In certain circumstances, you might want to ensure that a user remains with
a given set of resources throughout the session. For example, a user
attempting to conduct a transaction through an online bank needs to remain
with the same set of resources to ensure the transaction is completed
successfully.
To ensure that users stay with a specific set of resources, Global Traffic
Manager includes a persistence option. The persistence option instructs the
system to send a user to the same set of resources until a specified period of
time has elapsed.
8-6
Connections
8-8
9
Topologies
• Introduction
• Topology records
Introduction
As the name implies, Global Traffic Manager™ handles name resolution
requests at an international level. You can use topologies to load balance
these requests. A topology is a set of characteristics that identifies the origin
of a given name resolution request. In Global Traffic Manager, topologies
belong to one of several categories, including:
• Continent
• Country
• IP Subnet
• ISP
• Region
• State
A region is a customized collection of topologies. For example, you can
create a topology for Denmark, Iceland, Finland, Norway, and Sweden.
These topologies can compose a custom region called Scandinavia.
Through topologies, you can instruct Global Traffic Manager to select a data
center or resource based on its physical proximity to the client making the
name resolution request. This process helps ensure that name resolution
requests are answered and managed in the fastest possible time.
You can also instruct Global Traffic Manager to use topologies to load
balance name resolution requests across pools at the wide IP level, and
across virtual servers at the pool level.
To better understand topologies, consider the fictional company,
SiteRequest, which allows its customers to download applications from its
web site. SiteRequest has three data centers: New York, Paris, and Tokyo.
To ensure that customers can download their purchased application as
quickly as possible, the IT department has decided to create topologies with
which to load balance name resolution requests.
The New York data center is chosen as the designated data center for any
name resolution requests originating in the western hemisphere. To ensure
that these requests go only to the New York data center, the IT department
first creates a custom region, called Western Hemisphere, that contains the
continents North America and South America. With this custom region
created, the next step is to create a topology record for Global Traffic
Manager. A topology record is a statement that tells Global Traffic Manager
how to handle name resolution requests based on topologies. In this case,
the IT department creates the record as follows:
• Request Source: Region is Western Hemisphere
• Destination Source: Data Center is New York
• Weight: 10
The final step to implement this topology is to configure the pools in the
corresponding wide IP, www.siterequest.com, to use topology load
balancing.
Tip
If you require geolocation data at the city-level, contact your F5 Networks
sales representative to purchase additional database files.
9-2
Topologies
Topology records
A topology record has several elements: a request source statement, a
destination statement, an operator, and a weight.
A request source statement defines the origin of a name resolution request.
You can define the origin of a request as a:
• Continent
• Country (based on the ISO 3166 top-level domain codes)
• Internet Service Provider (ISP)
• IP subnet (Classless Inter-Domain Routing [CIDR] format)
• Region (custom)
• State
A destination statement defines the resource to which Global Traffic
Manager directs the name resolution request. The types of resources
available for a destination statement are as follows:
• Continent
• Country (based on the ISO 3166 top-level domain codes)
• Data center
• Internet Service Provider (ISP)
• IP subnet (CDIR definition)
• Pool of virtual servers
• Region (custom)
• State
You can select one of two operators for both a request source and a
destination statement. The is operator indicates that the name resolution
request matches the statement. The is not operator indicates that the name
resolution request does not match the statement.
The last element of a topology record, called the topology score or weight,
specifies the weight of the topology record. The system finds the weight of
the first topology record that matches the server object (pool or pool
member) and the LDNS. The system then assigns that weight as the
topology score for that server object. The system load balances to the server
object with the highest topology score. If the system finds no topology
record that matches both the server object and the LDNS, then the system
assigns that server object a zero score.
Note
9-4
Topologies
Note
When you enable the Longest Match option the system gives priority to the
topology records that contain IP subnet blocks that you defined using the
CIDR format. You can create a region and define an IP subnet using the
CIDR format as a member of that region; however, the system gives a
higher priority to the IP subnet defined in the topology record.
9-6
10
DNSSEC Keys and Zones
• About DNSSEC
About DNSSEC
The Domain Name System Security Extensions (DNSSEC) is an
industry-standard protocol that functions as an extension to the Domain
Name System (DNS) protocol. The BIG-IP® Global Traffic Manager™ uses
DNSSEC to guarantee the authenticity of DNS responses to queries and to
return Denial of Existence responses.
You can use the DNSSEC feature of Global Traffic Manager to protect your
network infrastructure from DNS protocol and DNS server attacks such as
spoofing, ID hacking, cache poisoning, and denial of service.
key has an ID of 0 (zero). Each time the system dynamically creates a new
generation of the key, the ID increments by 1. When a generation of a key
expires, the system automatically removes that generation of the key from
the configuration.
Figure 10.1 illustrates this, and shows how over time each generation of a
key overlaps the previous generation of the key.
The value that you assign to the TTL (time-to-live) setting for a key
specifies how long a client resolver can cache the key. As shown in Figure
10.1, the value you assign to the TTL setting of the key must be less than
the difference between the values of the Rollover Period and Expiration
Period settings of the key; otherwise, a client can make a query and the
system can send a valid key that the client cannot recognize.
Important
To ensure that each Global Traffic Manager system is referencing the same
time when generating keys, you must synchronize the time setting on each
system with the Network Time Protocol (NTP) servers that Global Traffic
Manager references.
10 - 2
DNSSEC Keys and Zones
10 - 4
11
Health and Performance Monitors
• Introduction
Introduction
An important feature of Global Traffic Manager™ is a set of load balancing
tools called monitors. Monitors verify connections on pools and virtual
servers. A monitor can be either a health monitor or a performance monitor.
Monitors are designed to check the status of a pool or virtual server on an
ongoing basis, at a set interval. If a pool or virtual server being checked does
not respond within a specified timeout period, or the status of a pool or
virtual server indicates that performance is degraded, then Global Traffic
Manager can redirect the traffic to another resource.
Some monitors are included as part of Global Traffic Manager, while other
monitors are user-created. Monitors that Global Traffic Manager provides
are called pre-configured monitors. User-created monitors are called
custom monitors.
Before configuring and using monitors, it is helpful to understand some
basic concepts regarding monitor types, monitor settings, and monitor
implementation.
◆ Monitor types
Every monitor, whether pre-configured or custom, belongs to a certain
category, or monitor type. Each monitor type checks the status of a
particular protocol, service, or application. For example, an HTTP
monitor allows you to monitor the availability of the HTTP service on a
pool member (that is a virtual server).
◆ Monitor settings
Every monitor consists of settings with values. The settings and their
values differ depending on the type of monitor. In some cases, Global
Traffic Manager assigns default values. For example, the following are
the default values for the HTTP monitor:
• Interval: 30 seconds
• Timeout: 120 seconds
• Probe Timeout: 5 seconds
• Reverse: No
• Transparent: No
These settings specify that an HTTP monitor is configured to check the
status of an IP address every 30 seconds, to time out after 120 seconds, to
timeout the probe request every 5 seconds, and specifies that the monitor
does not operate in either Reverse or Transparent mode.
◆ Monitor implementation
The task of implementing a monitor varies depending on whether you are
using a pre-configured monitor or creating a custom monitor. If you want
to implement a pre-configured monitor, you need only associate the
monitor with a pool or virtual server. If you want to implement a custom
monitor, you must first create the custom monitor, and then associate it
with a pool or virtual server.
Monitor types
Global Traffic Manager includes many different types of monitors, each
designed to perform a specific type of monitoring. The monitors belong to
one of three categories: simple, extended content verification (ECV), and
extended application verification (EAV).
◆ Simple monitors check the health of a resource by sending a packet using
the specified protocol, and waiting for a response from the resource. If
the monitor receives a response, then the health check is successful and
the resource is considered up.
◆ ECV monitors check the health of a resource by sending a query for
content using the specified protocol, and waiting to receive the content
from the resource. If the monitor receives the correct content, then the
health check is successful and the resource is considered up.
◆ EAV monitors check the health of a resource by accessing the specified
application. If the monitor receives the correct response, then the health
check is successful and the resource is considered up.
Pre-configured monitors
For a subset of monitor types, Global Traffic Manager includes a set of
pre-configured monitors. A pre-configured monitor is an existing monitor
with default settings already configured. You use a pre-configured monitor
when the default values of the settings meet your needs.
Global Traffic Manager includes these pre-configured monitors:
• big ip
• big ip link
• gateway_icmp
• http
• https
• real_server
• snmp
• tcp
• tcp_half_open
• udp
11 - 2
Health and Performance Monitors
Custom monitors
A custom monitor is a monitor that you create based on one of the allowed
monitor types.
Like http, each of the custom monitors has a Type setting based on the type
of service it checks (for example, https, ftp, pop3), and takes that type as its
name. (Exceptions are port-specific monitors, like the external monitor,
which calls a user-supplied program.)
If a pre-configured monitor exists that corresponds to the type of custom
monitor you are creating, you can import the settings and values of that
pre-configured monitor into the custom monitor. For example, if you create
a custom monitor called my_http, the monitor can inherit the settings and
values of the pre-configured monitor http. This ability to import existing
setting values is useful when you want to retain some setting values for your
new monitor, but modify others.
The following list shows an example of a custom HTTP monitor called
my_http, which is based on the pre-configured monitor http. Note that the
value of the Interval setting has been changed from the default value of 30
to a new value of 60. The other settings retain the values defined in the
pre-configured monitor.
• Name: my_http
• Type: HTTP
• Interval: 60
• Timeout: 120
• Reverse: No
• Transparent: No
You can import settings from another custom monitor instead of from a
pre-configured monitor. This is useful when you want to use the setting
values defined in another custom monitor, or when no pre-configured
monitor exists for the type of monitor you are creating. For example, if you
create a custom monitor called my_oracle_server2, you can import settings
from an existing Oracle® monitor such as my_oracle_server1. In this case,
because Global Traffic Manager does not provide a pre-configured Oracle®
monitor, a custom monitor is the only kind of monitor from which you can
import setting values.
11 - 4
Health and Performance Monitors
Monitor destinations
By default, the value for the Alias Address setting for most monitors is set
to the wildcard * Addresses, and the Alias Service Port setting is set to the
wildcard * Ports (exceptions to this rule are the WMI and Real Server
monitors). This value causes the monitor instance created for a pool or
virtual server to take that resource’s address or address and port as its
destination. You can, however, replace either or both wildcard symbols with
an explicit destination value, by creating a custom monitor. An explicit
value for the Alias Address and/or Alias Service Port setting is used to
force the instance destination to a specific address and/or port which may
not be that of the pool or virtual server.
The ECV monitors http, https, and tcp have the settings Send String and
Receive String for the send string and receive expression, respectively.
The most common Send String value is GET /, which retrieves a default
HTML page for a web site. To retrieve a specific page from a web site, you
can enter a Send String value that is a fully qualified path name:
"GET /www/support/customer_info_form.html"
The Receive String expression is the text string the monitor looks for in the
returned resource. The most common Receive String expressions contain a
text string that is included in a particular HTML page on your site. The text
string can be regular text, HTML tags, or image names.
The sample Receive expression below searches for a standard HTML tag:
"<HEAD>"
You can also use the default null Receive String value [""]. In this case,
any content retrieved is considered a match. If both the Send String and
Receive String are left empty, only a simple connection check is performed.
For HTTP monitors, you can use the special settings get or hurl in place of
Send String and Receive String statements, respectively.
◆ Transparent setting
Sometimes it is necessary to ping the aliased destination through a
transparent pool or virtual server. When you create a custom monitor and
set the Transparent setting to Yes, Global Traffic Manager forces the
monitor to ping through the pool or virtual server with which it is
associated (usually a firewall) to the pool or virtual server. (In other
words, if there are two firewalls in a load balancing pool, the destination
pool or virtual server is always pinged through the pool or virtual server
specified and not through the pool or virtual server selected by the load
balancing method.) In this way, the transparent pool or virtual server is
tested: if there is no response, the transparent pool or virtual server is
marked as down.
Common examples are checking a router, or checking a mail or FTP
server through a firewall. For example, you might want to check the
router address 10.10.10.53:80 through a transparent firewall
10.10.10.101:80. To do this, you create a monitor called http_trans in
which you specify 10.10.10.53:80 as the monitor destination address,
and set the Transparent setting to Yes. Then you associate the monitor
http_trans with the transparent firewall (10.10.10.101:80).
This causes the monitor to check the address 10.10.10 53:80 through
10.10.10.101:80. (In other words, Global Traffic Manager routes the
check of 10.10.10.53:80 through 10.10.10.101:80.) If the correct
response is not received from 10.10.10.53:80, then 10.10.10.101:80 is
marked down.
◆ Reverse setting
In most monitor settings, Global Traffic Manager considers the resource
available when the monitor successfully probes it. However, in some
cases you may want the resource to be considered unavailable after a
successful monitor test. You accomplish this configuration with the
Reverse setting. With the Reverse setting set to Yes, the monitor marks
the pool or virtual server down when the test is successful. For example,
if the content on your web site home page is dynamic and changes
frequently, you may want to set up a reverse ECV service check that
looks for the string: Error. A match for this string means that the web
server was down.
Table 11.1 shows the monitors that contain the Transparent setting, the
Reverse setting, or both.
11 - 6
Health and Performance Monitors
For information about the command syntax you use to change this variable,
see the gtm settings component in the Traffic Management Shell (tmsh)
Reference Guide.
Monitor associations
Some monitor types are designed for association only with nodes (IP
address), while other monitor types are intended for association only with
pools and virtual servers (IP address and service port). Therefore, when you
use the Configuration utility to associate a monitor with a pool or virtual
server, the utility displays only those pre-configured monitors that are
designed for association with that object type.
The types of monitor associations are:
◆ Monitor-to-pool association
Links a monitor with an entire load balancing pool. In this case, the
monitor checks all members of the pool. For example, you can create an
instance of the monitor http for the pool my_pool, thus ensuring that all
members of that pool are checked.
◆ Monitor-to-pool member association
Links a monitor with a pool member within a given pool. For example,
you can create an instance of the monitor FTP for specific pools within
the pool my_pool, ensuring that only specific pool members are verified
as available through the FTP monitor.
◆ Monitor-to-virtual server association
Links a monitor with a specific virtual server. In this case, the monitor
checks only the virtual server itself, and not any services running on that
virtual server. For example, you can create an instance of the monitor
http for virtual server 10.10.10.10.
11 - 8
12
Statistics
• Introduction
• Statistics access
• Types of statistics
• Persistence records
Statistics
Introduction
An important part of successfully managing a network is having access to
up-to-date information about network performance. This information can
verify that Global Traffic Manager™ is handling your name resolution
requests as efficiently as possible, as well as provide data about the overall
performance of a specific resource, such as a data center or distributed
application.
Global Traffic Manager gathers and displays statistical data about multiple
aspects of your network. The types of statistics you can view include:
• Status Summary (a summary of network components, as defined in
Global Traffic Manager)
• Distributed applications
• Wide IPs
• Pools
• Pool Members
• Data centers
• Links
• Servers
• Virtual servers
• iRules
• Paths
• Local DNS
• Persistence Records
Statistics access
You can access Global Traffic Manager statistics in two ways:
• Through the Statistics option on the Main tab of the navigation pane
• Through the Statistics menu from various main screens for different
components
Both methods take you to the same screen within Global Traffic Manager.
When you access statistics through a menu on the main screen for a given
network component, the Statistics screen is pre-configured for the given
network element, although you can switch to a different set of statistics at
any time.
Additionally, you can use the search feature to locate a specific component
or group of components. The default search value is an asterisk (*), which
instructs the system to display all relevant components in a list. You can
type a string in the box, and when you click the Search button, the system
modifies the list to show only those components that match the string.
Tip
You can also access statistics from the command line using the tmsh
command show. For more information about viewing statistics using tmsh,
see the Traffic Management Shell (tmsh) Reference Guide.
12 - 2
Statistics
◆ Unavailable
The Unavailable column describes the total number of resources of the
type corresponding to the Object Type column that Global Traffic
Manager can verify as unavailable.
◆ Offline
The Offline column describes the total number of resources of the type
corresponding to the Object Type column that Global Traffic Manager
can verify as offline.
◆ Unknown
The Available column describes the total number of resources of the type
corresponding to the Object Type column that Global Traffic Manager
can verify as available.
Each value within the Total, Available, Unavailable, Offline, and Unknown
columns is a link. When you click the link, you access the main screen for
that resource, with the list of resources filtered to show only those resources
with the corresponding status. For example, if the Available column for data
centers has a value of 5, clicking the 5 brings up a filtered main screen for
data centers that shows only the five data centers that are available.
Types of statistics
You can view a variety of statistics through Global Traffic Manager as
described in the following sections.
the information they need to see how many requests are being sent for the
application, allowing them to plan additional resource allocations more
effectively.
The distributed application statistics screen consists of a Distributed
Application Statistics table. This table contains the following information:
◆ Status
The Status column indicates the current status of the wide IP. The
available status types are: Available, Unavailable, Offline, and
Unknown. Each status type is represented by a symbol; for example, the
available status type is represented by a green circle.
◆ Distributed Application
The Distributed Application column displays the name of an application
for which Global Traffic Manager is responsible. Each name appears as a
link. When you click the link, the properties screen for the distributed
application opens.
◆ Members
The Members column provides a link that opens a wide IP details screen
for the distributed application. This screen displays load balancing
statistics for each pool within the distributed application. You can return
to the main distributed application statistics screen by clicking the Back
button in the Display Options area of the screen.
◆ Requests
The Requests column displays the cumulative number of Domain Name
System (DNS) requests sent to the distributed application.
◆ Load Balancing
The Load Balancing column provides information about how Global
Traffic Manager load balanced connection requests to this resource. This
column consists of four subcolumns:
• The Preferred subcolumn displays the cumulative number of requests
that Global Traffic Manager load balanced with the preferred load
balancing method.
• The Alternate subcolumn displays the cumulative number of requests
that Global Traffic Manager load balanced with the alternate load
balancing method.
• The Fallback subcolumn displays the cumulative number of requests
that Global Traffic Manager load balanced with the Fallback load
balancing method.
• The Returned to DNS subcolumn displays the cumulative number of
requests that Global Traffic Manager did not resolve and returned to
the DNS.
12 - 4
Statistics
Wide IP statistics
Global Traffic Manager captures several statistics related to the performance
of a wide IP. These statistics primarily focus on how many resolution
requests have been sent for the wide IP, and how Global Traffic Manager
has load balanced these requests. You can access the wide IP statistics by
selecting Wide IPs from the Statistics Type list in the Statistics screen.
As an example of wide IP statistics, consider the fictional company
SiteRequest. The IT department at SiteRequest has a wide IP,
www.siterequest.com, which uses the Global Availability load balancing
mode. This mode sends all name resolution requests for this wide IP to a
specific pool until that pool is unavailable. Because the wide IP,
www.siterequest.com, is critical to SiteRequest’s operations, the IT
department wants to track traffic to the wide IP and ensure that the primary
pool is not at risk of getting overloaded. The wide IP statistics provide the
IT department the information they need to see how many requests are being
sent for the wide IP, allowing them to plan additional resource allocations
more effectively.
The wide IP statistics screen consists of a Wide IP Statistics table. This table
contains the following information:
◆ Status
The Status column indicates the current status of the wide IP. The
available status types are: Available, Unavailable, Offline, and
Unknown. Each status type is represented by a symbol; for example, the
available status type is represented by a green circle.
◆ Wide IP
The Wide IP column displays the name of a wide IP for which Global
Traffic Manager is responsible. Each name appears as a link. When you
click the link, the properties screen for the wide IP opens.
◆ Pools
The Pools column provides a link that opens a pool details screen for the
wide IP. This screen displays load balancing statistics for each pool
within the wide IP. You can return to the main wide IP statistics screen
by clicking the Back button in the Display Options area of the screen.
◆ Requests
The Requests column displays the cumulative number of DNS requests
sent to the wide IP.
◆ Requests Persisted
The Requests Persisted column displays the cumulative number of
requests that persisted. Persisted requests use the same pool during a
connection session.
◆ Load Balancing
The Load Balancing column provides information about how Global
Traffic Manager load balanced connection requests to this resource. This
column consists of four subcolumns:
Pool statistics
The pool statistics available through Global Traffic Manager focus on how
Global Traffic Manager has load balanced name resolution requests. You
can access the pool statistics by selecting Pools from the Statistics Type list
in the Statistics screen.
As an example of pool statistics, consider the fictional company
SiteRequest. The IT department at SiteRequest has a wide IP,
www.siterequest.com, which contains pools that use the dynamic load
balancing mode, Quality of Service. This mode acquires statistical data
about response times between Global Traffic Manager and an LDNS
sending a name resolution request. There has been some concern of late as
to how well this new load balancing mode is working and if Global Traffic
Manager is able to gather the statistical information it needs to load balance
with this mode, or if it has to resort to an alternate or fallback method. By
using the pool statistics screen, the IT department can track how many name
resolution requests are load balanced using the preferred Quality of Service
method, and how many are load balanced using another method.
The pool statistics screen consists of a Pool Statistics table. This table
contains the following information:
◆ Status
The Status column indicates the current status of the pool. The available
status types are: Available, Unavailable, Offline, and Unknown. Each
status type is represented by a symbol; for example, the available status
type is represented by a green circle.
◆ Pool
The Pool column displays the name of a wide IP for which Global
Traffic Manager is responsible. Each name appears as a link. When you
click the link, the properties screen for the pool opens.
◆ Members
The Members column provides a link that opens a virtual server details
screen for the pool. This screen displays connection statistics for each
virtual server within the pool, including the number of times the virtual
12 - 6
Statistics
server was selected for a name resolution request and the amount of
traffic flowing from and to the virtual server. You can return to the main
wide IP statistics screen by clicking the Back button in the Display
Options area of the screen.
◆ Load Balancing
The Load Balancing column provides information about how Global
Traffic Manager load balanced connection requests to this resource. This
column consists of four subcolumns:
• The Preferred subcolumn displays the cumulative number of requests
that Global Traffic Manager load balanced with the preferred load
balancing method.
• The Alternate subcolumn displays the cumulative number of requests
that Global Traffic Manager load balanced with the alternate load
balancing method.
• The Fallback subcolumn displays the cumulative number of requests
that Global Traffic Manager load balanced with the Fallback load
balancing method.
• The Returned to DNS subcolumn displays the cumulative number of
requests that Global Traffic Manager did not resolve and returned to
the DNS.
◆ Data Center
The Data Center column displays the name of a data center. Each name
appears as a link. When you click the link, the properties screen for the
data center opens.
◆ Servers
The Servers column provides a link that opens a server details screen for
the data center. This screen displays connection statistics for each server
at a data center, including the number of times the server was selected for
a name resolution request and the amount of traffic flowing from and to
the server. You can return to the main data center statistics screen by
clicking the Back button in the Display Options area of the screen.
◆ Connections
The Connections column displays the cumulative number of requests that
Global Traffic Manager resolved using a resource from the
corresponding data center.
◆ Throughput (bits/sec)
The Throughput (bits/sec) column contains two subcolumns:
• The In column displays the cumulative number of bits per second sent
to the data center.
• The Out column displays the cumulative number of bits per second
sent from the data center.
◆ Throughput (packets/sec)
The Throughput (packets/sec) column contains two subcolumns:
• The In column displays the cumulative number of packets per second
sent to the data center.
• The Out column displays the cumulative number of packets per
second sent from the data center.
Link statistics
Link statistics focus on how much traffic is flowing in and out through a
specific link to the Internet. This information can help you prevent a link
from getting over-used, saving your organization from higher bandwidth
costs. You can access the link statistics by selecting Links from the
Statistics Type list in the Statistics screen.
As an example of how the statistics for data centers can help you manage
your network resources, consider the fictional company SiteRequest.
SiteRequest has two links with two different Internet Service Providers
(ISPs). The primary ISP is paid in advance for a specific amount of
bandwidth usage. This allows SiteRequest to save money, but if the
bandwidth exceeds the prepaid amount, the costs increase considerably. As
a result, the IT department uses a second ISP, which has a slower connection
but considerably lower costs. By using the links statistics, the IT department
can ensure that links to the Internet are used as efficiently as possible.
12 - 8
Statistics
The link statistics screen consists of a Link Statistics table. This table
contains the following information:
◆ Status
The Status column indicates the current status of the link. The available
status types are: Available, Unavailable, Offline, and Unknown. Each
status type is represented by a symbol; for example, the available status
type is represented by a green circle.
◆ Link
The Link column displays the name of a link for which Global Traffic
Manager is responsible. Each name appears as a link. When you click the
link, the properties screen for the link opens.
◆ Throughput (bits/sec)
The Throughput (bits/sec) column contains four subcolumns:
• The In column displays the cumulative number of bits per second sent
to the data center.
• The Out column displays the cumulative number of bits per second
sent from the data center.
• The Total column displays the cumulative number of both incoming
and outgoing bits per second for the link.
• The Over Prepaid column displays the amount of traffic, in bits per
second, that has exceeded the prepaid traffic allotment for the link.
In addition to viewing the link data as a table, you can also view it in a graph
format. To use this format, click the Graph button. A graph screen opens,
which shows the amount of traffic used over time. You can change the
amount of time shown in the graph by selecting a value from the Graph
Interval list, located in the Display Options area of the screen.
Server statistics
With server statistics, you can analyze the amount of traffic flowing to and
from each server. This information can tell you if your resources are
distributed appropriately for your network. You can access the server
statistics by selecting Servers from the Statistics Type list in the Statistics
screen.
As an example of how the statistics for servers can help you manage your
network resources, consider the fictional company SiteRequest. The IT
department at SiteRequest is considering whether it needs a few more
servers to better manage name resolution requests; however, there is some
debate as to whether the servers should be consolidated at the New York
data center (which the New York team prefers) or spread out over all of the
data centers. It is also possible that an under-utilized server at one data
center might be moved to another data center. By using the server statistics,
the IT department can look at how much traffic is handled by each server,
giving them the information they need to decide where these new servers, if
any, should go.
The server statistics screen consists of a Server Statistics table. This table
contains the following information:
◆ Status
The Status column indicates the current status of the server. The
available status types are: Available, Unavailable, Offline, and
Unknown. Each status type is represented by a symbol; for example, the
available status type is represented by a green circle.
◆ Server
The Server column displays the name of a server for which Global
Traffic Manager is responsible. Each name appears as a link. When you
click the link, the properties screen for the server opens.
◆ Virtual Servers
The Virtual Servers column provides a link that opens a virtual server
details screen for the server. This screen displays connection statistics for
each virtual server at a data center, including the number of times the
virtual server was selected for a name resolution request and the amount
of traffic flowing from and to the server. You can return to the main data
center statistics screen by clicking the Back button in the Display
Options area of the screen.
◆ Picks
The Picks column displays the cumulative number of times Global
Traffic Manager picked a server to handle a name resolution request.
◆ Connections
The Connections column displays the cumulative number of requests that
Global Traffic Manager resolved using a resource from the
corresponding data center.
◆ Throughput (bits/sec)
The Throughput (bits/sec) column contains two subcolumns:
• The In column displays the cumulative number of bits per second sent
to the server.
• The Out column displays the cumulative number of bits per second
sent from the server.
◆ Throughput (packets/sec)
The Throughput (packets/sec) column contains two subcolumns:
• The In column displays the cumulative number of packets per second
sent to the server.
• The Out column displays the cumulative number of packets per
second sent from the server.
12 - 10
Statistics
◆ Throughput (bits/sec)
The Throughput (bits/sec) column contains two subcolumns:
• The In column displays the cumulative number of bits per second sent
to the server.
• The Out column displays the cumulative number of bits per second
sent from the server.
◆ Throughput (packets/sec)
The Throughput (packets/sec) column contains two subcolumns:
• The In column displays the cumulative number of packets per second
sent to the server.
• The Out column displays the cumulative number of packets per
second sent from the server.
Paths statistics
The paths statistics captured by Global Traffic Manager provide information
about how quickly traffic moves between an LDNS and a resource for
which Global Traffic Manager is responsible. Information presented in the
paths statistics screen includes details about round trip times (RTT), hops,
and completion rates. You can access the paths statistics by selecting Paths
from the Statistics Type list in the Statistics screen.
Paths statistics are primarily used when you employ a dynamic load
balancing mode for a given wide IP or pool. You can use the information in
the Paths statistics to get an overall sense of how responsive your wide IPs
are in relation to the local DNS servers that have been sending name
resolution requests to a wide IP.
The paths statistics screen consists of a paths statistics table. This table
contains the following information:
◆ Local DNS Address
The Local DNS Address column displays the IP address of each LDNS
that has sent a name resolution request for a wide IP for which Global
Traffic Manager is responsible.
◆ Link
The Link column displays the ISP link that Global Traffic Manager used
to send and receive data from the LDNS.
◆ Round Trip Time (RTT)
The Round Trip Time (RTT) column contains two subcolumns:
• The Current subcolumn displays the current round trip time between
the LDNS and Global Traffic Manager.
• The Average subcolumn displays the average round trip time between
the LDNS and Global Traffic Manager.
12 - 12
Statistics
◆ Hops
The Hops column contains two subcolumns:
• The Current subcolumn displays the current number of hops between
the LDNS and Global Traffic Manager.
• The Average subcolumn displays the average number of hops
between the LDNS and Global Traffic Manager.
◆ Completion Rate
The Completion Rate column contains two subcolumns:
• The Current subcolumn displays the current completion rate of
transactions between the LDNS and Global Traffic Manager.
• The Average subcolumn displays the average completion rate of
transactions between the LDNS and Global Traffic Manager.
◆ Last Accessed
The Last Accessed column displays the last time the LDNS attempted a
connection to Global Traffic Manager.
◆ Location
The Location column contains four subcolumns:
• The Continent subcolumn displays the continent on which the LDNS
resides.
• The Country subcolumn displays the country in which the LDNS is
located.
• The State subcolumn displays the state in which the LDNS is located.
• The City subcolumn displays the city in which the LDNS is located.
12 - 14
Statistics
Persistence records
One of the common methods of modifying name resolution requests with
Global Traffic Manager is to activate persistent connections. A persistent
connection is a connection in which Global Traffic Manager sends name
resolution requests from a specific LDNS to the same set of resources until a
time-to-live value has been reached. If you use persistent connections in
your configuration of Global Traffic Manager, you may want to see what
persistent connections are currently active on your network. You can access
the persistence records by selecting Persistence Records from the Statistics
Type list in the Statistics screen.
The persistence records screen consists of a persistence records table. This
table contains the following information:
◆ Local DNS Address
The LDNS Address column displays the IP address of each LDNS that
has sent a name resolution request for a wide IP for which Global Traffic
Manager is responsible.
◆ Level
The Level column displays the level at which the persistent connection is
based. Available types are wide IPs and distributed applications.
◆ Destination
The Destination column displays the wide IP or distributed application to
which the name resolution request was directed.
◆ Target Type
The Target Type column displays the type of resource on which
persistence is based. Examples of target types include data centers,
servers, pools, and virtual servers.
◆ Target Name
The Target Name column displays the name of the resource on which
persistence is based.
◆ Expires
The Expires column displays the time at which the persistence for the
given LDNS request expires.
12 - 16
13
Metric Collection
• Introduction
• About metrics
Introduction
Global Traffic Manager™ system uses specialized software components,
called monitors, to capture data regarding the availability of a resource, such
as a virtual server. Monitors represent one half of the statistical gathering
capabilities of Global Traffic Manager. The second half, metrics collection,
captures data about how well network traffic flows between Global Traffic
Manager and the external local DNS servers and internal resources with
which it communicates.
The resources you make available to your users over the Internet are often
critical to your organization; consequently, it is vital that these resources are
not only available, but highly responsive to your users. Typically, two main
criteria determine the responsiveness of a resource: hops and paths. A hop is
one point-to-point transmission between a host and a client server in a
network. A network path that includes a stop at a network router has two
hops: the first from the client to the router, and the second from the router to
the host server. A path is a logical network route between a data center
server and an LDNS.
It is important to remember that hops and paths can differ from each other
widely on a per-connection basis. For example, an LDNS might take a long
path to reach a specific resource, but require only a few hops to get there. On
the other hand, that same LDNS might select a short path, yet have to move
between a larger number of routers, increasing the number of hops it takes to
reach the resource. It is up to you to determine what thresholds for hops and
paths are acceptable for your network, as the needs of each network, and
even each application within the same network, can vary widely.
Through the metrics collection capabilities of Global Traffic Manager, you
can accomplish several tasks related to improving the availability and
responsiveness of your network applications and resources. You can:
• Define the types of metrics that Global Traffic Manager collects, and
how long the system keeps those metrics before acquiring fresh data.
• Assign probes to local DNS servers that attempt to acquire the metrics
information.
• Configure Time-to-Live (TTL) values for your metrics data.
• Exclude specific local DNS servers from Global Traffic Manager probes.
• Implement the Quality of Service load balancing mode, which uses
metrics to determine the best resource for a particular name resolution
request.
About metrics
When you decide to use Global Traffic Manager to collect metrics on the
local DNS servers that attempt to access your network resources, you can
define the following characteristics:
• Types of metrics collected (either hops, paths, both, or disabled)
• Time-to-live (TTL) values for each metric
• Frequency at which the system updates the data
• Size of a packet sent (relevant for hop metrics only)
• Length of time that can pass before the system times out the collection
attempt
• Number of packets sent for each collection attempt
While each of these settings is important, the ones that perhaps require the
most planning beforehand are the TTL values. In general, the lower the TTL
value, the more often Global Traffic Manager probes an LDNS. This
improves the accuracy of the data, but increases bandwidth usage.
Conversely, increasing the TTL value for a metric lowers the bandwidth
your network uses, but increases the chance that Global Traffic Manager is
basing its load balancing operations off of stale data
An additional consideration is the number of local DNS servers that Global
Traffic Manager queries. The more local DNS servers that the system
queries, the more bandwidth is required to ensure those queries are
successful. Therefore, setting the TTL values for metrics collection can
require incremental fine-tuning. F5 Networks recommends that you
periodically check the TTL values, and verify that they are appropriate for
your network.
13 - 2
Metric Collection
With these probes, it does not matter whether Global Traffic Manager
receives a valid response, such as the name of the LDNS as queried by the
DNS_REV probe, or a request refused statement. The relevant information
is the metrics generated between the probe request and the response. For
example, Global Traffic Manager uses the DNS_REV probe to query two
local DNS servers. The first LDNS responds to the probe with its name, as
per the request. The second LDNS, however, responds with a request
refused statement, because it is configured to not allow such requests. In
both cases, the probe was successful, because Global Traffic Manager was
able to acquire data about how long it took for both local DNS servers to
respond to the probe.
You can configure Global Traffic Manager to use a select number of probes,
or you can assign all five. The more probes that Global Traffic Manager
uses, the more bandwidth is required.
When Global Traffic Manager attempts to probe an LDNS, it is actively
attempting to acquire data from that LDNS. Certain Internet Service
Providers and other organizations might request that you do not probe their
local DNS servers, while other local DNS servers might be known to act as
proxies, which do not provide accurate metrics data. In these situations, you
can configure Global Traffic Manager to exclude local DNS servers from
probes. When you exclude an LDNS, Global Traffic Manager does not
probe that server; however, Global Traffic Manager is also unable to use the
Quality of Service load balancing mode to load balance name resolution
requests from that LDNS.
You can remove an LDNS from the address exclusion list at any time.
Situations in which you want to remove an LDNS include the LDNS
becoming inactive, or the IP address of the LDNS changing to a different
network subnet.
13 - 4
Metric Collection
13 - 6
14
Performance Data
• Introduction
Introduction
Global Traffic Manager™ captures data about how network traffic flows
between Global Traffic Manager and the external local DNS servers and
internal resources with which it communicates.
You can view graphs that display information about how Global Traffic
Manager is performing. You can use this information to help you determine
how to modify the configuration to obtain the best possible performance
from the system.
Performance graph
The GTM Performance graph shows the throughput of Global Traffic
Manager. The graph includes the following data:
• GTM Requests
Represents the number of incoming DNS requests.
• GTM Resolutions
Represents the number of incoming DNS requests that were resolved by
any method.
• GTM Resolutions Persisted
Represents the number of incoming DNS requests that were resolved by
a persistence record.
• GTM Resolutions Returned to DNS
Represents the number of incoming DNS requests that were not resolved
by Global Traffic Manager, but were instead passed on to the DNS server
for resolution.
14 - 2
15
iRules
• Introduction
• What is an iRule?
Introduction
As you work with Global Traffic Manager™, you might find that you want
to incorporate additional customizations beyond the available features
associated with load balancing, monitors, or other aspects of your traffic
management. For example, you might want to have the system respond to a
name resolution request with a specific CNAME record, but only when the
request is for a particular wide IP and originates from Europe. In Global
Traffic Manager, these customizations are defined through iRules®. iRules
are code snippets that are based on TCL 8.4. These snippets allow you a
great deal of flexibility in managing your global network traffic.
If you are familiar with Local Traffic Manager™, you might already be
aware of and use iRules to manage your network traffic on a local level. The
iRules in Global Traffic Manager share a similar syntax with their Local
Traffic Manager counterparts, but support a different set of events and
objects.
Due to the dynamic nature of iRules development, the following sections
focus on providing an overview of iRule operations and describe the events
and command specific to Global Traffic Manager. For additional
information about how to write iRules, visit the F5 DevCentral web site:
http://devcentral.f5.com. At this site, you can learn more about iRules
development, as well as discuss iRules functionality with others.
What is an iRule?
An iRule is a script that you write if you want individual connections to
target a pool other than the default pool defined for a virtual server. iRules
allow you to more directly specify the pools to which you want traffic to be
directed. Using iRules, you can send traffic not only to pools, but also to
individual pool members or hosts.
The iRules you create can be simple or sophisticated, depending on your
content-switching needs. Figure 15.1 shows an example of a simple iRule.
when DNS_REQUEST {
if { [IP::addr [IP::client_addr] equals 10.10.10.10] } {
pool my_pool
}
}
This iRule is triggered when a DNS request has been detected, causing
Global Traffic Manager to send the packet to the pool my_pool, if the IP
address of the local DNS making the request matches 10.10.10.10.
iRules can direct traffic not only to specific pools, but also to individual pool
members, including port numbers and URI paths, either to implement
persistence or to meet specific load balancing requirements.
The syntax that you use to write iRules is based on the Tool Command
Language (Tcl) programming standard. Thus, you can use many of the
standard Tcl commands, plus a set of extensions that Global Traffic
Manager provides to help you further increase load balancing efficiency.
For information about standard Tcl syntax, see the Tcl Reference Manual at
http://tmml.sourceforge.net/doc/tcl/index.html.
Within Global Traffic Manager, you assign iRules to the wide IPs in your
network configuration.
15 - 2
iRules
Event declarations
The iRules feature includes several types of event declarations that you can
make in an iRule. Specifying an event declaration determines when Global
Traffic Manager evaluates the iRule. The following sections list and
describe these event types. Also described is the concept of iRule context
and the use of the when keyword.
You make an event declaration in an iRule by using the when keyword,
followed by the event name. For example:
when DNS_REQUEST {
iRule details...
15 - 4
16
ZoneRunner
• ZoneRunner utility
• Zone files
• Resource records
• Views
• Named.conf
ZoneRunner
ZoneRunner utility
One of the modes in which you can operate Global Traffic Manager™
system is the node mode. In node mode, Global Traffic Manager is
responsible not only for load balancing name resolution requests and
monitoring the health of your physical and logical network; it is also
responsible for maintaining the DNS zone files that map name resolution
requests to the appropriate network resource.
In Global Traffic Manager, you create, manage, and maintain DNS files
using the ZoneRunner™ utility. The ZoneRunner utility is a zone file
management utility that can manage both DNS zone files and your BIND
configuration. With the ZoneRunner utility, you can:
• Manage the DNS zones and zone files for your network, including
importing and transferring zone files
• Manage the resource records for those zones
• Manage views
• Manage a local nameserver and its configuration file, named.conf
ZoneRunner tasks
When you use the ZoneRunner utility to manage your DNS zones and
resource records, you can accomplish several tasks, including:
• Configure a zone
• Configure the resource records that make up the zone
• Configure a view, for access control
• Configure options in the named.conf file
Note
In the Configuration utility, you must configure a zone before you configure
any other objects in the ZoneRunner utility.
Zone files
With the ZoneRunner utility, you can create, modify, and delete zone files.
Additionally, you can transfer zone files to another nameserver, or import
zone files from another nameserver. A zone file contains resource records
and directives that describe the characteristics and hosts of a zone, otherwise
known as a domain or sub-domain.
16 - 2
ZoneRunner
DNS information in case your primary DNS server goes offline, or the
primary DNS server. In either situation, you can use the ZoneRunner utility
to import existing zone files into Global Traffic Manager instead of
re-creating them manually. It is important to note that you can import only
primary zones files.
Through the ZoneRunner utility, you can import zone files using one of two
methods:
• Loading zones from a file
If you know where the zone files you want to import reside on your
server, you can load these files directly into Global Traffic Manager
through the ZoneRunner utility. After you load a zone file into Global
Traffic Manager, the ZoneRunner utility displays information about the
zone and any of its resource records within the Configuration utility.
Important
You can load only primary zones files.
Before you can transfer zone files from another server, you must ensure
that the you have configured the source server to allow transfers to the
destination server. You typically accomplish this task using the
allow-transfer option. See your DNS and BIND documentation for more
information.
Important
You can transfer only primary zones files.
Resource records
Resource records are the files that contain details about a zone. These
resource records, in a hierarchical structure, make up the domain name
system (DNS). After you have created a zone, you can use the ZoneRunner
utility to view, create, modify, and delete the resource records for that zone.
Note
Although case is preserved in names and data fields when loaded into the
nameserver, comparisons and lookups in the nameserver database are not
case-sensitive.
16 - 4
ZoneRunner
Views
In BIND, a view allows you to modify the nameserver configuration based
on the community attempting to access it. For example, if your DNS handles
requests from both inside and outside your company, you can create two
views: internal and external. Through views, you can build nameserver
configurations on the same server, and have those configurations apply
dynamically when the request originates from a specified source.
In Global Traffic Manager, a single view is created automatically within the
ZoneRunner utility: external. If you do not want to create views, all zones
that Global Traffic Manager maintains are associated with this default view.
16 - 6
ZoneRunner
Named.conf
You define the primary operational characteristics of BIND using a single
file, named.conf. The functions defined in this file include views, access
control list definitions, and zones.
You can control most of the contents of the named.conf file through the
ZoneRunner utility, as this utility updates the named.conf file to implement
any modifications that you make. However, you can also use the
ZoneRunner utility to edit the named.conf file directly.
Important
Modifying the named.conf file carries a high level of risk, as a syntax error
can prevent the entire BIND system from performing as expected. For this
reason, F5 Networks recommends that you use the user interface of the
ZoneRunner utility whenever possible, and that you exercise caution when
editing the named.conf file.
16 - 8
A
big3d Agent
• Introduction
• Metrics
• Communications
big3d Agent
Introduction
The big3d agent runs on all BIG-IP® systems, collects performance
information on behalf of the Global Traffic Manager™ system, and
continually monitors the availability of the servers that Global Traffic
Manager load balances. The utility also monitors the integrity of the
network paths between the servers that host the domain, and the various
local DNS servers that attempt to connect to the domain. Each big3d agent
broadcasts its collected data to all of the Global Traffic Manager systems
and Link Controller™ systems in your network, ensuring that these systems
work with the latest information.
You can turn off the big3d agent on any BIG-IP system at any time;
however, if you turn off the big3d agent on a server, Global Traffic
Manager can no longer check the availability of the server or its virtual
servers, and the statistics screens display the status of these servers as
unknown (blue ball).
Tip
F5 Networks recommends that you have at least one BIG-IP system running
the big3d agent in each data center in your network. This ensures that
Global Traffic Manager has timely access to the metrics associated with
network traffic.
Metrics
A big3d agent collects the following types of performance information that
the system uses for load balancing. The big3d agent broadcasts this
information to all Global Traffic Manager systems in your network.
◆ Network path round trip time
The big3d agent calculates the round trip time for the network path
between the utility’s data center and the client’s LDNS that is making the
resolution request. Global Traffic Manager uses round trip time to
determine the best virtual server to answer the request when a pool uses a
dynamic load balancing mode, such as Round Trip Time, or Quality of
Service.
◆ Network path packet loss
The big3d agent calculates the packet completion percentage for the
network path between the utility’s data center and the client’s LDNS that
is making the resolution request. Global Traffic Manager uses the packet
completion rate to determine the best virtual server to answer the request
when a wide IP or pool uses either the Completion Rate or the Quality of
Service load balancing modes.
◆ Router hops along the network path
The big3d agent calculates the number of intermediate system transitions
(router hops) between the utility’s data center and the client’s LDNS.
Global Traffic Manager uses hops to determine the best virtual server to
answer the request when a pool uses the Hops or the Quality of Service
load balancing modes.
◆ Server performance
The big3d agent returns server metrics, such as the packet rate, for
BIG-IP systems or SNMP-enabled hosts. Global Traffic Manager uses
packet rate to determine the best virtual server to answer the request
when a pool uses the Packet Rate, KBPS, Least Connections, or Quality
of Service load balancing modes.
◆ Virtual server availability and performance
The big3d agent queries virtual servers to verify whether they are up and
available to receive connections, and uses only those virtual servers that
are up for load balancing. The big3d agent also determines the number
of current connections to virtual servers that are defined on BIG-IP
systems or SNMP-enabled hosts. Global Traffic Manager uses the
number of current connections to determine the best virtual server when
a pool uses the Least Connections or VS Capacity load balancing mode.
A-2
big3d Agent
The load on the big3d agents depends on the timer settings that you assign
to the different types of data the big3d agents collect. The shorter the timers,
the more frequently the big3d agent needs to refresh the data. While short
timers guarantee that you always have valid data readily available for load
balancing, they also increase the frequency of data collection.
Another factor that can affect data collection is the number of client local
DNS servers that make name resolution requests. The more local DNS
servers that make resolution requests, the more path data that the big3d
agents have to collect. While round trip time for a given path may vary
constantly due to current network load, the number of hops along a network
path between a data center and a specific LDNS does not often change.
Consequently, you may want to set short timer settings for round trip time
data so that it refreshes more often, but set high timer settings for hops data
because it does not need to be refreshed often.
A-4
big3d Agent
Communications
In order to copy big3d agents from a Global Traffic Manager system to
BIG-IP systems, the Global Traffic Manager system must be able to
communicate with these other systems. Specifically, every BIG-IP system,
which you define as a server on the Global Traffic Manager system, must
have sufficient network privileges and configured routes to be able to probe
the virtual servers that it hosts, as well as the virtual servers hosted by other
servers defined on the Global Traffic Manager systems in a synchronization
group.
In the following configuration, every big3d agent that the Global Traffic
Manager synchronization group recognizes must be able to probe the virtual
server 10.1.0.1:80 via TCP.
server { // datacenter=DC1, #VS=1
name "Generic Host Server 1"
type generic
box {
address 10.1.0.1
unit_id 1
}
monitor "http"
vs {
name "Generic_VS1"
address 10.1.0.1:80 // http
}
}
Table A.2 shows the protocols and corresponding ports used for iQuery
communications between big3d agents and SNMP agents that run on host
servers.
big3d agent host SNMP agent UDP >1023 161 Ephemeral ports used to make
SNMP queries for host statistics
host SNMP agent big3d agent UDP 161 >1023 Ephemeral ports used to receive
host statistics using SNMP
Table A.2 Communication between big3d agents and SNMP agents on hosts
Table A.3 shows the ports used for communications between big3d agents
and virtual servers that are not hosted by a BIG-IP system.
big3d agent virtual server UDP >1024 Service Ephemeral ports used to monitor
Port host virtual server
big3d agent virtual server TCP >1024 Service Ephemeral ports used to monitor
Port host virtual servers
Table A.3 Communication between big3d agents and virtual servers not hosted by BIG-IP systems
A-6
big3d Agent
big3d LDNS TCP >1023 53 Probe using TCP (Cisco® routers: allow
establish)
LDNS big3d TCP 53 >1023 Replies using TCP (Cisco® routers: allow
establish)
LDNS big3d ICMP N/A N/A Replies to ICMP, UDP pings, or traceroute
probes
big3d LDNS dns_rev >1023 53 Probe using DNS rev or DNS dot
dns_dot
LDNS big3d dns_rev 53 >1023 Replies to DNS rev or DNS dot probes
dns_dot
Table A.4 Communications between big3d agents and local DNS servers
A-8
B
Probes
• Introduction
• About iQuery
• Probe responsibility
• LDNS probes
Introduction
When you install a Global Traffic Manager™ system in a network, that
system typically works within a larger group of BIG-IP® products. These
products include other Global Traffic Manager systems, Link Controller™
systems, and Local Traffic Manager™ systems. Global Traffic Manager
must be able to communicate with these other systems to maintain an
accurate assessment of the health and availability of different network
components. For example, Global Traffic Manager must be able to acquire
statistical data from resources that are managed by Local Traffic Manager in
a different data center. BIG-IP systems acquire this information through the
use of probes. A probe is an action a BIG-IP system takes to acquire data
from other network resources.
Probes are an essential means by which Global Traffic Manager tracks the
health and availability of network resources; however, it is equally
important that the responsibility for conducting probes be distributed across
as many BIG-IP products as possible. This distribution ensures that no one
system becomes overloaded with conducting probes, which can cause a
decrease in performance in the other tasks for which a BIG-IP system is
responsible.
To distribute probe requests effectively across multiple BIG-IP systems,
Global Traffic Manager systems employ several different technologies and
methodologies, including:
• iQuery®, which is the communication protocol used between Global
Traffic Manager systems and the big3d agents that reside on other
BIG-IP systems
• A selection methodology that determines which Global Traffic Manager
is responsible for managing the probe request
• A selection methodology that determines which big3d agent actually
conducts the probe
About iQuery
At the heart of probe management with Global Traffic Manager systems is
iQuery, the communications protocol that these systems use to send
information from one system to another. With iQuery, Global Traffic
Manager systems in the same synchronization group can share configuration
settings, assign probe requests to big3d agents, and receive data on the
status of network resources.
The iQuery protocol is an XML protocol that is sent between each system
using gzip compression and SSL. These communications can only be
allowed between systems that have a trusted relationship established, which
is why configuration tools such as big3d_install, bigip_add, and gtm_add
are critical when installing or updating Global Traffic Manager systems. If
two systems have not exchanged their SSL certificates, they cannot share
information with each other using iQuery.
In addition to requiring trusted relationships, systems send iQuery
communications only on the VLAN on which the system received the
incoming message. Also, iQuery communications occur only within the
same synchronization group. If your network consists of two
synchronization groups, with each group sharing a subset of network
resources, these groups probe the network resources and communicate with
iQuery separately.
Generally, iQuery communications require no user intervention; however,
on occasion it can be necessary to view the data transmitted between each
system. For example, you might be troubleshooting the reason that a Global
Traffic Manager system is exhibiting a particular behavior. In such a
situation, you can use the command, iqdump.
B-2
Probes
Probe responsibility
When you assign a monitor to a network resource, Global Traffic Manager
is responsible for ensuring that a big3d agent probes the selected resource. It
is important to remember that this does not necessarily mean the selected
Global Traffic Manager actually conducts the probe; it means only that a
specific Global Traffic Manager is in charge of assigning a big3d agent to
probe the resource. The big3d agent can be installed on the same system as
Global Traffic Manager, a different Global Traffic Manager, or another
BIG-IP system.
A crucial component to determining which system manages a probe request
is the data centers that you define in the Global Traffic Manager
configuration. For each probe, the Global Traffic Manager systems
determine the following:
• Is there a Global Traffic Manager system in the same data center as the
resource?
• Is there more than one Global Traffic Manager at that data center?
Now, consider that you want to acquire statistical data from a resource in the
New York data center. First, the Global Traffic Manager systems, based on
their iQuery communications with each other, identify whether there is a
Global Traffic Manager system that belongs to the New York data center. In
this case, the answer is yes; the New York data center contains a Global
Traffic Manager system. Next, the systems determine if more than one
Global Traffic Manager belongs to the New York data center. In this case,
the answer is no; the New York data center has only a stand-alone system.
Consequently, the Global Traffic Manager system in the New York data
center assumes responsibility for conducting the probe on this particular
resource.
In situations where more than one Global Traffic Manager belongs to a data
center, the systems use an algorithm to distribute the responsibility for
probes equally among Global Traffic Manager systems. This distribution
ensures that each Global Traffic Manager has an equal chance of being
responsible for managing a probe request.
To demonstrate how probe requests are delegated between two Global
Traffic Manager systems at the same data center, consider again the network
configuration at SiteRequest. This time, the company needs to acquire data
from a resource that resides at the Los Angeles data center. As with the
previous example, the first step identifies whether the Los Angeles data
center has any Global Traffic Manager systems; in this case, the answer is
yes. The next criteria is whether there is more than one Global Traffic
Manager at that data center; in this case, the answer is also yes: the Los
Angeles data center has a redundant system configuration that consists of
two Global Traffic Manager systems. Because there are two Global Traffic
Manager systems at this data center, each system compares the hash value of
the resource with its own information; whichever Global Traffic Manager
has the closest value to the resource becomes responsible for managing the
probe request.
A final consideration is if a data center does not have any Global Traffic
Manager systems at all, such as the London data center in the configuration
for SiteRequest. In this situation, the responsibility for probing a resource at
that data center is divided among the other Global Traffic Manager systems;
much in the same way as the responsibility is divided among Global Traffic
Manager systems within the same data center.
When Global Traffic Manager becomes responsible for managing a probe, it
remains responsible for that probe until the network configuration changes
in one of the following ways:
• Global Traffic Manager goes offline.
• A new Global Traffic Manager system is added to the data center.
• The network configuration of the resource (such as its IP address)
changes.
B-4
Probes
Consider that a Global Traffic Manager system in the Los Angeles data
center has assumed responsibility for managing a probe for a network
resource. At this data center, the system can assign the probe to one of four
big3d agents: one for each BIG-IP system at the data center. To select a
big3d, Global Traffic Manager looks to see which big3d agent has the
fewest number of probes for which it is responsible. The big3d agent with
the lowest number of probes is tasked with conducting the probe. Global
Traffic Manager checks this statistic each time it needs to delegate the
probe; as a result, the selected big3d can change from probe instance to
probe instance.
In situations where a big3d agent does not reside in the same data center as
the resource, the designated Global Traffic Manager selects a big3d from all
available big3d agents on the network. Again, the agent selected is the agent
with the fewest number of probe requests, and this check occurs each time
the probe is conducted.
For example, SiteRequest adds a new set of web servers in Tokyo. At this
location, the company has yet to install its BIG-IP systems; however, the
current set of Global Traffic Manager systems in Los Angeles and New
York are managing traffic to these web servers. When initiating a probe
request to determine the availability of one of these servers, a Global Traffic
Manager system is selected to manage the probe request. Then, that system
chooses a big3d agent to probe the web server, selecting any big3d agent
located in Los Angeles, New York, or London.
B-6
Probes
LDNS probes
Global Traffic Manager systems are responsible for probes of local DNS
servers (LDNS). Unlike probes conducted on internal systems, such as web
servers, probes of local DNS servers require that the Global Traffic Manager
system verifies data from a resource that exists outside the network.
Typically, this data is the path information Global Traffic Manager requires
when conducting Quality of Service, Round Trip Time, Completion Rate,
and Hops load balancing methods.
Note
If you do not use Quality of Service load balancing, Global Traffic Manager
does not conduct probes of local DNS servers.
When a given LDNS makes a DNS request for a wide IP, that request is sent
to a single Global Traffic Manager. Global Traffic Manager then creates an
LDNS entry, and assigns that entry one of the following states:
• New: Global Traffic Manager has not come across this particular LDNS
before
• Active: Global Traffic Manager already has an existing entry for this
LDNS
• Pending: Global Traffic Manager has been contacted by this LDNS
before, however, this server has yet to respond to a probe from a Global
Traffic Manager system on this network
In general, the New and Pending states are temporary states; an LDNS
remains in one of these states only until it responds to the first probe request
from Global Traffic Manager. After Global Traffic Manager receives a
response, the LDNS entry is moved to the Active state. Each Global Traffic
Manager within a given synchronization group shares the LDNS entries that
are assigned this state, resulting in the synchronization group having a
common list of known local DNS servers.
Unlike internal probes, LDNS probes are not load balanced across Global
Traffic Manager systems. Instead, the Global Traffic Manager system that
the LDNS first queries becomes responsible for the initial probe to that
LDNS. These probes are load balanced, however, across the multiple big3d
agents, with preference given to big3d agents that either belong to the same
data center as the responding Global Traffic Manager, or belong to the same
link through which Global Traffic Manager received the LDNS query. After
the initial probe, an algorithm is used to load balance subsequent probes
across the available Global Traffic Manager systems.
B-8
Probes
B - 10
Glossary
Glossary
A record
The A record is the resource record that Global Traffic Manager™ returns to
a local DNS server in response to a name resolution request. The A record
contains a variety of information, including one or more IP addresses that
resolve to the requested domain name.
active unit
In a redundant system configuration, an active unit is a system that currently
load balances name resolution requests. If the active unit in the redundant
system fails, the standby unit assumes control and begins to load balance
requests.
alternate method
The alternate method specifies the load balancing mode that Global Traffic
Manager uses to pick a virtual server if the preferred method fails. See also
fallback method, preferred method.
big3d agent
The big3d agent is a monitoring agent that collects metrics information
about server performance and network paths between a data center and a
specific local DNS server. Global Traffic Manager uses the information
collected by the big3d agent for dynamic load balancing.
BIG-IP system
A BIG-IP system can be a Global Traffic Manager system (including the
current Global Traffic Manager system), a Local Traffic Manager™ system,
or a Link Controller™ system.
bridge mode
Bridge mode instructs Global Traffic Manager to forward the traffic it
receives to another part of the network.
CNAME record
A canonical name (CNAME) record acts as an alias to another domain
name. A canonical name and its alias can belong to different zones, so the
CNAME record must always be entered as a fully qualified domain name.
CNAME records are useful for setting up logical names for network
services so that they can be easily relocated to different physical hosts.
completion rate
The completion rate is the percentage of packets that a server successfully
returns during a given session.
Configuration utility
The Configuration utility is the browser-based application that you use to
configure the BIG-IP system.
custom monitor
A custom monitor is a user-created monitor. See also monitor, health
monitor, performance monitor, pre-configured monitor.
data center
A data center is a physical location that houses one or more Global Traffic
Manager systems, BIG-IP systems, or host machines.
Glossary - 2
Glossary
destination statement
A destination statement defines the resource to which Global Traffic
Manager directs the name resolution request.
distributed application
A distributed application is a collection of wide IPs, data center, and links. It
is the highest level component that Global Traffic Manager supports.
DNSSEC zones
DNSSEC zones are containers that map a domain name to a set of DNSSEC
keys.
domain name
A domain name is a unique name that is associated with one or more IP
addresses. Domain names are used in URLs to identify particular web pages.
For example, in the URL http://www.f5.com/index.html, the domain name
is f5.com.
draining requests
Draining requests refers to allowing existing sessions to continue accessing
a specific set of resources while disallowing new connections.
EAV monitor
An EAV monitor checks the health of a resource by accessing the specified
application.
ECV monitor
An ECV monitor checks the health of a resource by sending a query for
content using the specified protocol, and waiting to receive the content from
the resource. See also monitor, health monitor, external monitor.
external monitor
An external monitor is a user-supplied health monitor. See also health
monitor.
external system
An external system is any server with which Global Traffic Manager must
exchange information to perform its functions.
failover
Failover is the process whereby a standby unit in a redundant system
configuration takes over when a software failure or hardware failure is
detected on the active unit.
failover cable
The failover cable is the cable that directly connects the two system units in
a hardware-based redundant system configuration.
fallback method
The fallback method is the third method in a load balancing hierarchy that
Global Traffic Manager uses to load balance a resolution request. Global
Traffic Manager uses the fallback method only when the load balancing
modes specified for the preferred and alternate methods fail. Unlike the
preferred method and the alternate method, the fallback method uses neither
server nor virtual server availability for load balancing calculations. See also
preferred method, alternate method.
Glossary - 4
Glossary
in the list, while Round Robin mode searches for an available server starting
with the next server in the list (with respect to the server selected for the
previous connection request).
gtmd
The gtmd utility processes communications between two Global Traffic
Manager systems.
health monitor
A health monitor checks a node to see if it is up and functioning for a given
service. If the node fails the check, it is marked down. Different monitors
exist for checking different services. See also monitor, custom monitor,
pre-configured monitor, performance monitor.
host
A host is a network server that manages one or more virtual servers that
Global Traffic Manager uses for load balancing.
iQuery
The iQuery® protocol is used to exchange information between Global
Traffic Manager systems and BIG-IP systems. The iQuery protocol is
officially registered with IANA for port 4353, and works on UDP and TCP
connections.
iRule
An iRule is a user-written script that controls the behavior of a connection
passing through the Global Traffic Manager™ system. iRules® are an F5
Networks feature and are frequently used to direct certain connections to a
non-default load balancing pool. However, iRules can perform other tasks,
such as implementing secure network address translation and enabling
session persistence.
key-signing key
Global Traffic Manager uses key signing keys to sign only the DNSKEY
record of a DNSSEC record set. See also DNSSEC (DNS Security
Extensions), DNSSEC zones, and zone-signing key.
Kilobytes/Second mode
The Kilobytes/Second mode is a dynamic load balancing mode that
distributes connections based on which available server currently processes
the fewest kilobytes per second.
LDNS
An LDNS is a server that makes name resolution requests on behalf of a
client. With respect to Global Traffic Manager, local DNS servers are the
source of name resolution requests.
link
A link is a logical representation of a physical device (router), which
connects your network to the rest of the Internet.
Link Controller
Link Controller™ is an IP application switch that manages traffic to and
from a site across multiple links, regardless of connection type or provider.
listener
A listener is an object that listens for DNS queries. A listener instructs
Global Traffic Manager to listen for network traffic destined for a specific
IP address.
Glossary - 6
Glossary
metrics information
Metrics information is the data that is typically collected about the paths
between BIG-IP systems and local DNS servers. Metrics information is also
collected about the performance and availability of virtual servers. Metrics
information is used for load balancing, and it can include statistics such as
round trip time, packet rate, and packet loss.
monitor
A monitor is a software utility that specializes in a specific metric of a
Global Traffic Manager resource. A monitor tests to see if a given resource
responds as expected. See also custom monitor, pre-configured monitor,
health monitor, performance monitor.
monitor template
A monitor template is an abstraction that exists within the Global Traffic
Manager system for each monitor type, and contains a group of settings and
default values.
named
The named daemon manages domain nameserver software.
nameserver
A nameserver is a server that maintains a DNS database, and resolves
domain name requests to IP addresses using that database.
name resolution
Name resolution is the process by which a nameserver matches a domain
name request to an IP address, and sends the information to the client
requesting the resolution.
node
A node is a logical object on the BIG-IP system that identifies the IP address
of a physical resource on the network, such as a web server.
Node mode
The Node mode instructs Global Traffic Manager to process traffic locally,
and send the appropriate DNS response back to the querying server.
NS record
A nameserver (NS) record is used to define a set of authoritative
nameservers for a DNS zone. A nameserver is considered authoritative for
some given zone when it has a complete set of data for the zone, allowing it
to answer queries about the zone on its own, without needing to consult
another nameserver.
packet rate
The packet rate is the number of data packets per second processed by a
server.
path
A path is a logical network route between a data center server and a local
DNS server.
path probing
Path probing is the process of collecting metrics data, such as round trip
time and packet rate, for a given path between a requesting LDNS and a data
center server.
performance monitor
Performance monitors check the performance of a pool or virtual server, and
dynamically load balance traffic accordingly. See also monitor,
pre-configured monitor, custom monitor, health monitor.
persistence
On Global Traffic Manager, persistence is a series of related requests
received from the same local DNS server for the same wide IP name. When
persistence is activated, Global Traffic Manager sends all requests from a
particular local DNS server for a specific wide IP to the same virtual server,
instead of load balancing the requests.
picks
Picks represent the number of times a particular virtual server is selected to
receive a load balanced connection.
Glossary - 8
Glossary
pool
A pool is a group of virtual servers managed by a BIG-IP system, or a host.
Global Traffic Manager load balances among pools (using the Pool LB
Mode), as well as among individual virtual servers.
pool ratio
A pool ratio is a ratio weight applied to pools in a wide IP. If the Pool LB
mode is set to Ratio, Global Traffic Manager uses each pool for load
balancing in proportion to the weight defined for the pool.
preferred method
The preferred method specifies the first load balancing mode that Global
Traffic Manager uses to load balance a resolution request. See also alternate
method, fallback method, and load balancing methods.
pre-configured monitor
Pre-configured monitors are monitors that Global Traffic Manager provides.
See also monitor, custom monitor, and health monitor.
probe
A probe is a specific query, initiated by a big3d agent, that attempts to
gather specific data from a given network resource. Probes are most often
employed when a health monitor attempts to verify the availability of a
resource.
ratio
A ratio is the parameter in a virtual server statement that assigns a weight to
the virtual server for load balancing purposes.
Ratio mode
The Ratio load balancing mode is a static load balancing mode that
distributes connections across an pool of virtual servers in proportion to the
ratio weight assigned to each individual virtual server.
Ratio weighting
Ratio weighting is a methodology in which the system uses a frequency that
you set to determine to which link to send traffic.
region
A region is a customized collection of topologies. See topology.
resource record
A resource record is a record in a DNS database that stores data associated
with domain names. A resource record typically includes a domain name, a
TTL, a record type, and data specific to that record type. See also A record,
CNAME record, NS record.
root nameserver
A root nameserver is a master DNS server that maintains a complete DNS
database. There are approximately 13 root nameservers in the world that
manage the DNS database for the World Wide Web.
Glossary - 10
Glossary
router hops
Router hops are intermediate system transitions along a given network path.
Router mode
Router mode instructs Global Traffic Manager to forward the traffic it
receives to another DNS server.
self IP address
A self IP address is an IP address that you define on a VLAN of a BIG-IP
system. This term does not apply to the management IP address of a BIG-IP
system, or to IP addresses on other devices.
server
A server is a physical device on which you can configure one or more
virtual servers.
Setup utility
The Setup utility is a utility that takes you through the initial system
configuration process. The Setup utility runs automatically when you turn
on a system for the first time.
Simple monitor
A Simple monitor checks the health of a resource by sending a packet using
the specified protocol, and waiting for a response from the resource. See
also health monitor.
standby unit
A standby unit is the system in a redundant system configuration that is
always prepared to become the active unit if the active unit fails.
synchronization
Synchronization means that each Global Traffic Manager regularly
compares the timestamps of its configuration files with the timestamps of
the configuration files on the other Global Traffic Manager systems on the
network.
synchronization group
A synchronization group is a group of Global Traffic Manager systems that
synchronize system configurations and zone files (if applicable). All
synchronization group members receive broadcasts of metrics data from the
big3d agents throughout the network. All synchronization group members
also receive broadcasts of updated configuration settings from Global
Traffic Manager that has the latest configuration changes.
tmsh
The Traffic Management Shell (tmsh) is a command-line utility that you
can use to configure Global Traffic Manager.
topology
A topology is a set of characteristics that identify the origin of a given name
resolution request.
Topology mode
The Topology mode is a static load balancing mode that bases the
distribution of name resolution requests on the weighted scores for topology
records. Topology records are used by the Topology load balancing mode to
redirect DNS queries to the closest virtual server, geographically, based on
location information derived from the DNS query message.
topology record
A topology record specifies a score for a local DNS server location endpoint
and a virtual server location endpoint.
Glossary - 12
Glossary
topology score
The topology score is the weight assigned to a topology record when Global
Traffic Manager is filtering the topology records to find the best virtual
server match for a DNS query.
topology statement
A topology statement is a collection of topology records.
unavailable
The unavailable status is used for data center servers and virtual servers.
When a data center server or virtual server is unavailable, Global Traffic
Manager does not use it for load balancing.
unknown
The unknown status is used for data center servers and virtual servers.
When a data center server or virtual server is new to Global Traffic Manager
and does not yet have metrics information, Global Traffic Manager marks
its status as unknown. Global Traffic Manager can use unknown servers for
load balancing, but if the load balancing mode is dynamic, Global Traffic
Manager uses default metrics information for the unknown server until it
receives live metrics data.
up
The up status is used for data center servers and virtual servers. When a data
center server or virtual server is up, the data center server or virtual server is
available to respond to name resolution requests.
virtual server
A virtual server, in the context of Global Traffic Manager, is a combination
of an IP address and a port number that, together, provide access to an
application or data source on your network.
wide IP
A wide IP is a collection of one or more domain names that maps to one or
more groups of virtual servers managed either by BIG-IP systems, or by
host servers. Global Traffic Manager load balances name resolution requests
across the virtual servers that are defined in the wide IP that is associated
with the requested domain name.
wildcard listener
A wildcard listener monitors all traffic coming into your network, regardless
of the destination IP address of the given DNS request.
zone
In DNS terms, a zone is a subset of DNS records for one or more domains.
zone file
In DNS terms, a zone file is a database set of domains with one or many
domain names, designated mail servers, a list of other nameservers that can
answer resolution requests, and a set of zone attributes, which are contained
in an SOA record.
zone-signing key
Global Traffic Manager uses a zone-signing key to sign all of the record sets
in a DNSSEC zone. See also DNSSEC (DNS Security Extensions), DNSSEC
zones, and key-signing key.
ZoneRunner
ZoneRunner™ is the utility that allows you manage your resource records,
zone files, and named configuration associated with your implementation of
DNS and BIND.
Glossary - 14
Index
Index
Index - 2
Index
Index - 4
Index