RDF System Management Manual
Table Of Contents
- RDF System Management Manual
- What’s New in This Manual
- About This Manual
- 1 Introducing RDF
- RDF Subsystem Overview
- RDF Processes
- RDF Operations
- Reciprocal and Chain Replication
- Available Types of Replication to Multiple Backup Systems
- Triple Contingency
- Loopback Configuration (Single System)
- Online Product Initialization
- Online Database Synchronization
- Online Dumps
- Subvolume- and File-Level Replication
- Shared Access DDL Operations
- EMS Support
- SMF Support
- RTD Warning Thresholds
- Process-Lockstep Operation
- Support for Network Transactions
- RDF and NonStop SQL/MX
- Zero Lost Transactions (ZLT)
- Monitoring RDF Entities With ASAP
- 2 Preparing the RDF Environment
- 3 Installing and Configuring RDF
- 4 Operating and Monitoring RDF
- 5 Managing RDF
- Recovering From File System Errors
- Handling Disk Space Problems
- Responding to Operational Failures
- Stopping RDF
- Restarting RDF
- Carrying Out a Planned Switchover
- Takeover Operations
- Reading the Backup Database
- Access to Backup Databases in a Consistent State
- RDF and NonStop SQL/MP DDL Operations
- RDF and NonStop SQL/MX Operations
- Backing Up Image Trail Files
- Making Online Dumps With Updaters Running
- Doing FUP RELOAD Operations With Updaters Running
- Exception File Optimization
- Switching Disks on Updater UPDATEVOLUMES
- 6 Maintaining the Databases
- 7 Online Database Synchronization
- 8 Entering RDFCOM Commands
- 9 Entering RDFSCAN Commands
- 10 Triple Contingency
- 11 Subvolume- and File-Level Replication
- 12 Auxiliary Audit Trails
- 13 Network Transactions
- Configuration Changes
- RDF Network Control Files
- Normal RDF Processing Within a Network Environment
- RDF Takeovers Within a Network Environment
- Takeover Phase 1 – Local Undo
- Takeover Phase 2 – File Undo
- Takeover Phase 3 – Network Undo
- Takeover Phase 3 Performance
- Communication Failures During Phase 3 Takeover Processing
- Takeover Delays and Purger Restarts
- Takeover Restartability
- Takeover and File Recovery
- The Effects of Undoing Network Transactions
- Takeover and the RETAINCOUNT Value
- Network Configurations and Shared Access NonStop SQL/MP DDL Operations
- Network Validation and Considerations
- RDF Re-Initialization in a Network Environment
- RDF Networks and ABORT or STOP RDF Operations
- RDF Networks and Stop-Update-to-Time Operations
- Sample Configurations
- RDFCOM STATUS Display
- 14 Process-Lockstep Operation
- Starting a Lockstep Operation
- The DoLockstep Procedure
- The Lockstep Transaction
- RDF Lockstep File
- Multiple Concurrent Lockstep Operations
- The Lockstep Gateway Process
- Disabling Lockstep
- Reenabling Lockstep
- Lockstep Performance Ramifications
- Lockstep and Auxiliary Audit Trails
- Lockstep and Network Transactions
- Lockstep Operation Event Messages
- 15 NonStop SQL/MX and RDF
- Including and Excluding SQL/MX Objects
- Obtaining ANSI Object Names From Updater Event Messages
- Creating NonStop SQL/MX Primary and Backup Databases from Scratch
- Creating a NonStop SQL/MX Backup Database From an Existing Primary Database
- Online Database Synchronization With NonStop SQL/MX Objects
- Offline Synchronization for a Single Partition
- Online Synchronization for a Single Partition
- Correcting Incorrect NonStop SQL/MX Name Mapping
- Consideration for Creating Backup Tables
- Restoring to a Specific Location
- Comparing NonStop SQL/MX Tables
- 16 Zero Lost Transactions (ZLT)
- A RDF Command Summary
- B Additional Reference Information
- C Messages
- D Operational Limits
- E Using ASAP
- Index
Installing and Configuring RDF
HP NonStop RDF System Management Manual—524388-003
3-6
Synchronizing the Primary and Backup Databases
then duplicate this table on the backup system by using the SQLCI DUP command. In
this example, \PRIM is the primary system and \BACK is the backup system.
Notice that the catalog for this NonStop SQL/MP table is created on the backup system
before starting RDF on the primary system so that RDF will recognize the backup
catalog and not report errors when attempting to process audit data for this catalog.
1. Using SQLCI, enter a CREATE command to create the catalog on the backup
system. TMF must be up for NonStop SQL/MP catalog updating:
CREATE CATALOG \BACK.$DATA1.DBCAT;
2. Set up DEFINEs on the primary system to simplify referring to NonStop SQL/MP
tables in subsequent SQLCI commands for the primary system:
SET DEFMODE ON;
ADD DEFINE =EMPLOYEE, CLASS MAP,
FILE \PRIM.$DATA1.DB.EMPLOYEE;
ADD DEFINE =EMPLPAR2, CLASS MAP,
FILE \PRIM.$DATA2.DB.EMPLOYEE;
ADD DEFINE =EMPLNAME, CLASS MAP,
FILE \PRIM.$DATA2.DB.EMPLNAME;
3. Create the catalog on your primary system and make this the default catalog for all
partitions:
CREATE CATALOG \PRIM.$TEST.DBCAT;
CATALOG \PRIM.$TEST.DBCAT;
4. Enter a CREATE TABLE command to create the partitioned table:
CREATE TABLE =EMPLOYEE (
EMPNUM DECIMAL (5) UNSIGNED NO DEFAULT,
FIRST_NAME CHARACTER(15) NO DEFAULT,
LAST_NAME CHARACTER(20) NO DEFAULT,
PRIMARY KEY EMPNUM )
ORGANIZATION KEY SEQUENCED
PARTITION ( =EMPLPAR2 FIRST KEY 3000 );
This command creates an audited table with AUDITCOMPRESS on.
5. Enter CREATE CONSTRAINT commands for any constraints that values in
particular columns of the table must satisfy:
CREATE CONSTRAINT EMPNUM_CONSTRNT
ON =EMPLOYEE
CHECK EMPNUM BETWEEN 1 AND 99999;
6. Create the index for the NonStop SQL/MP table on the primary system:
CREATE INDEX =EMPLNAME
ON =EMPLOYEE( LAST_NAME, FIRST_NAME );