cc/td/doc/product/cable/svc_ctrl/scmgtsu
hometocprevnextglossaryfeedbacksearchhelp
PDF

Table Of Contents

Subscriber Manager Fail-Over

Information About Subscriber Manager Fail-Over

Overview

Normal Operation

Fail-Over Topology

Fail-Over Operation

How to Recover from Fail-Over

Machine Reboot

Replacing the Server

Database Duplication Recovery

Database Duplication Recovery Management


Subscriber Manager Fail-Over


This module describes topics that are related to using the Subscriber Manager together with clusters and redundancy.

As the Subscriber Manager plays a critical role in the Cisco SCA BB solution that is deployed in tier-one service provider environments, it also supports, starting with SM version 2.2, a fail-over operational mode. This feature minimizes system downtime that is caused by SM failure (as further discussed in Information About Subscriber Manager Fail-Over ).

This section introduces various concepts that are related to using a cluster of two SM nodes in a fail-over operational mode.


Note For the purposes of this section, it is assumed that the reader is familiar with the Veritas Cluster technology.


Information About Subscriber Manager Fail-Over

Overview 

Normal Operation 

Fail-Over Topology 

Fail-Over Operation 

Overview

The fail-over scheme that is implemented in the SM is based on the Veritas cluster technology. The cluster includes two machines, each of them running SM TimesTen and Veritas software. The Veritas Cluster Server (VCS) software consolidates the SMs and exposes a single entity by providing a single virtual IP address for the entire cluster.

The cluster software distinguishes an active and a standby machine: the active machine "owns" the virtual IP address and all network connections, while the standby machine is passive until a fail-over occurs. At fail-over, the IP address is passed from the failing server to the backup server, which becomes activated and re-establishes all network connections.

When a fail-over occurs, the LEGs lose their connection with the failed SM, and reconnect to the activated (backup) SM and retransmit their uncommitted messages. The activated SM connects to the SCE platforms and performs an SCE resynchronization.

The TimesTen database replication agent constantly replicates the SM database from the active node to the standby node. This enables a fast fail-over from one SM to another, since the subscriber data in the activated machine is always valid. The two SM nodes do not communicate except for passing the subscriber data.

The VCS uses software components called "cluster agents" to monitor and control the state of resources such as Network Interface Cards (NICs), disks, IP addresses, and processes. Cisco supplies cluster agents to monitor the SM and the TimesTen database daemon and replication agent.

As part of the cluster operation, the TimesTen database daemon and replication agents are up and running regardless of the fail-over state. The SM Veritas agent monitors the daemon and the replication agent process. In case one of them fails, a fail-over takes place.


Note The SM software configuration on both the active and the standby machines must be identical . Apply the same configuration file and the same application PQI module to both machines.


The following sections describe these concepts in further detail.

Normal Operation

The two SM nodes operate in hot-standby mode, where at any given time one node (the active node) receives and processes all the SM events, while the other node (the standby node) waits and is ready to go into operation on fail-over. For enabling seamless fail-over and for minimizing the fail-over time, the two SM nodes operate without an external storage device.

During the normal operation of the cluster, the active node (selected by the cluster):

Performs all SM functionality of a non-cluster environment

Provides "health" information for the cluster agent

Periodically replicates its subscriber database to the standby node

On the standby node, both the SM and the TimesTen software are running:

The SM is fully configured. (It is applied with the same configuration file and PQI application module as the active node, but does not interfere with the active node's work.)

The SM connects to the TimesTen database, but does not connect to the LEG and the SCE devices.

The TimesTen software is operating as a replication client for the subscriber database, receiving and applying updates from the active node's TimesTen software.

Fail-Over Topology

The following figure depicts an SM cluster configuration in a topology with a redundant AAA server and two SCE 2000 platforms that are cascaded for redundancy.

Figure 3-1 SM Cluster Configuration for Fail-Over Topology

As already mentioned, an SM fail-over topology includes two SM nodes connected in a cluster scheme.

Two dedicated (private) redundant networks interconnect the two nodes:

Heartbeat network—Used by the Veritas Cluster Server to perform cluster monitoring and control.

Replication network—Used by the replication process to pass the subscriber records.

The two nodes should be located in the same site, where the heartbeat network is implemented using back-to-back connectivity between the two nodes or via redundant switches. Each node in the cluster has redundant network paths (NICs) connecting it to all of the external entities with which the SM communicates (AAA, LEG, SCE).

Each node in the cluster has a minimum of six Ethernet NICs, where:

Two NICs are used for the (private) heartbeat network

Two NICs are used for the (private) replication network

Two NICs are used for the public network (connectivity to SCEs and LEGs, and management of the SM)

The cluster has a virtual IP (VIP) address used for communication with the external entities. Each node in the cluster has also an IP address for administration of the node/cluster, as well as an IP address for replication use.

Upon failure of the primary NIC of the public network, there is a fail-over to the secondary NIC on the same node, keeping the same IP addresses (VIP1), with no fail-over of the cluster. Upon failure of the primary NIC of the replication or heartbeat networks, there is fail-over to the secondary NIC on the same node, keeping the same IP addresses (VIP2 and VIP3), with no fail-over of the cluster.

The following diagram illustrates the usage of the regular and virtual IP addresses used in cluster configuration:

Administration of the nodes uses IP1/IP2 and IP3/IP4 respectively.

The cluster IP address for external clients over the public network uses VIP1.

Figure 3-2 Regular and Virtual IPs in Cluster Configuration

For further information about replication IP configuration, see Veritas Cluster Server.

Fail-Over Operation

During normal operation, the Veritas Cluster Server mechanism automatically selects one of the SM servers to be active and the other to be standby.

The active SM server performs all the normal SM functionality. The two servers maintain the heartbeat mechanism between them, and the active server continuously replicates the subscriber database to the standby server's database.

The standby SM server acts as a hot-standby machine, so it is completely ready for taking over (becoming activated) in a minimal fail-over time.

The following types of failures trigger the fail-over mechanism:

SM application failure, including failure of the TimesTen database.

Failure of the TimesTen daemon of the TimesTen replication process.

SUN server failure, due to failure of one of the resources of the server; for example, failure of both of the public network NICs.

Manual activation of fail-over.


Note Communication failure does not cause a fail-over if there is a redundant NIC. Therefore, because each SUN machine has two NICs for connecting to external devices, a failure of one of the NICs merely causes switching to the redundant NIC, without activating the fail-over mechanism.


After detecting a failure, the standby SM becomes activated, and the following occurs:

The activated SM takes over the IP resources of the virtual IP mechanism.

The LEGs reconnect to the activated SM.

The activated SM creates IP connections with the SCEs and resynchronizes with them.

The activated SM starts processing information that is sent from the different LEGs and forwards it to the SCEs.

How to Recover from Fail-Over

Different types of failures require different triggering for the recovery procedure. Some failures may recover automatically such as intra-node ports link-failure, which recovers automatically when the link revives, while others may need manual intervention.

Recovery may take place when an SM that experienced a failure is self-recovered or after it was replaced (if needed). The purpose of the recovery procedure is to take the cluster back to a fully functional mode. When the recovery procedure ends, the behavior is the same as it was after installation.

The failed SM server is recovered manually or automatically, according to the type of failure that occurred. The recovery procedures, and when they are used, are described in the following sections.

Machine Reboot 

Replacing the Server 

Database Duplication Recovery 

Database Duplication Recovery Management 

Machine Reboot

Recovering from a machine reboot is a fully automatic recovery process, where the failed SM server reboots, and after establishing a connection with the other server and synchronizing the databases, the cluster of the two SM servers is ready again for fail-over operation.


Note This steps in this procedure are automatic.



Step 1 The reboot process is run on the node.

Step 2 VCS makes the node standby.

Step 3 The node boots.

Step 4 VCS establishes intra-node communication and the new node joins the cluster.

Step 5 The TimesTen database replication process is started from the point before the reboot.

The SM in the recovered server is ready after the database recovery process is running and the SM moves from Init state to Standby state.


Replacing the Server

Replacing the server is necessary when the machine has an unrecoverable physical failure. A new machine that is installed with fresh SM, TimesTen, and VCS installations replaces the server.

Replacing the server is a manual recovery, where the failed SM server is physically replaced. After connecting the new SM server to the network, configuring it and synchronizing the two databases, the cluster of the two SM servers is ready again for fail-over operation.


Step 1 Connect a new server to the inter-node ports and intra-node ports (but leave the network ports disconnected).

Step 2 Perform the basic network and cluster configurations manually (the first time).

Step 3 Copy the configuration files from the active node.

Use the following CLU command if you need to copy only the p3sm.cfg file:

p3sm --load-config --remote= NEW-SM_IP

Step 4 Perform the TimesTen database duplication operation.

See Database Duplication Recovery.

Step 5 Start the VCS operation on the recovered node.

Step 6 Connect the network ports.

The SM in the recovered server is ready after the database recovery process is completed and the SM moves from Init state to Standby state.


Database Duplication Recovery

Database duplication recovery is a manual recovery, which is needed when the standby node database loses synchronization with the active node database. Loss of synchronization can occur when one of the SM machines is replaced or when the replication process on the active node fails to replicate all of the data inserted to its database (replication NICs were disconnected).


Step 1 Stop the cluster server (VCS) monitoring of the resources.

Use the VCS CLU hastop -localcommand to stop the VCS.

Step 2 Stop the SM, so it will not be affected by clearing the database.

Use the CLU command p3sm --stop.

Step 3 Stop the replication agent.

Use the CLU command p3db --rep-stop.

Step 4 Destroy the database.

Use the CLU command p3db --destroy-rep-db.

Step 5 Duplicate the remote database to the local machine.

Use the CLU command p3db --duplicate.

Step 6 Start the cluster server monitoring of the resources.

Use the VCS CLU hastartcommand), which will automatically start the replication process and the SM.


Database Duplication Recovery Management

The two SM servers are configured using Command-Line Utilities and a configuration file (see Configuration and Management and How to Configure a Subscriber Management Solution ). The actual configuration is performed for the active SM and then manually replicated for the standby SM.

SUMMARY STEPS

1. Establish an FTP connection between the active and standby machines.

2. Copy the configuration file.

3. Copy the application PQI file you installed on the active node to the standby node.

4. Install the PQI file.

5. Manually copy the database-related configuration files to the required location.

6. Configure and administer the Veritas Cluster Server using Veritas tools.

DETAILED STEPS


Step 1 Establish an FTP connection between the active and standby machines.

Step 2 Copy the configuration file.

Copy the configuration file from ~pcube/sm/server/root/config/p3sm.cfg on the active node to the standby node, and apply the SM configuration file by using the CLU command p3sm --load-config.

Alternatively, you can replicate the SM configuration file to the standby machine by running the CLU command p3sm --load-config --remote=NEW-SM_IP

Step 3 Copy the application PQI file you installed on the active node to the standby node.

Step 4 Install the PQI file.

Use the CLU command p3inst --install -f PQI file path.

Step 5 Manually copy the database-related configuration files to the required location.

If you have made changes in the database-related configuration files, copy the files to /etc/system (for Solaris) or to /etc/sysctl.conf (for Linux), and /var/TimesTen/sys.odbc.ini from the active node to the standby node.


Note If you this step, a reboot of the standby node is required.



Note If the database is located in different directories in the two nodes, then the files sys.odbc.ini in both nodes are not identical and the actual parameter changed in the file must be copied.


Step 6 Configure and administer the Veritas Cluster Server using Veritas tools.

Notifications are enabled through SNMP traps that the Veritas Cluster Server provides. The Veritas Cluster Server supports SNMP traps such as:

Fatal failure detected (local or remote)

Secondary node starts fail-over procedure

Secondary node is operational (end of fail-over)



hometocprevnextglossaryfeedbacksearchhelp

Posted: Wed Aug 15 17:43:37 PDT 2007
All contents are Copyright © 1992--2007 Cisco Systems, Inc. All rights reserved.
Important Notices and Privacy Statement.