RDF System Management Manual
Table Of Contents
- RDF System Management Manual
- What’s New in This Manual
- About This Manual
- 1 Introducing RDF
- RDF Subsystem Overview
- RDF Processes
- RDF Operations
- Reciprocal and Chain Replication
- Available Types of Replication to Multiple Backup Systems
- Triple Contingency
- Loopback Configuration (Single System)
- Online Product Initialization
- Online Database Synchronization
- Online Dumps
- Subvolume- and File-Level Replication
- Shared Access DDL Operations
- EMS Support
- SMF Support
- RTD Warning Thresholds
- Process-Lockstep Operation
- Support for Network Transactions
- RDF and NonStop SQL/MX
- Zero Lost Transactions (ZLT)
- Monitoring RDF Entities With ASAP
- 2 Preparing the RDF Environment
- 3 Installing and Configuring RDF
- 4 Operating and Monitoring RDF
- 5 Managing RDF
- Recovering From File System Errors
- Handling Disk Space Problems
- Responding to Operational Failures
- Stopping RDF
- Restarting RDF
- Carrying Out a Planned Switchover
- Takeover Operations
- Reading the Backup Database
- Access to Backup Databases in a Consistent State
- RDF and NonStop SQL/MP DDL Operations
- RDF and NonStop SQL/MX Operations
- Backing Up Image Trail Files
- Making Online Dumps With Updaters Running
- Doing FUP RELOAD Operations With Updaters Running
- Exception File Optimization
- Switching Disks on Updater UPDATEVOLUMES
- 6 Maintaining the Databases
- 7 Online Database Synchronization
- 8 Entering RDFCOM Commands
- 9 Entering RDFSCAN Commands
- 10 Triple Contingency
- 11 Subvolume- and File-Level Replication
- 12 Auxiliary Audit Trails
- 13 Network Transactions
- Configuration Changes
- RDF Network Control Files
- Normal RDF Processing Within a Network Environment
- RDF Takeovers Within a Network Environment
- Takeover Phase 1 – Local Undo
- Takeover Phase 2 – File Undo
- Takeover Phase 3 – Network Undo
- Takeover Phase 3 Performance
- Communication Failures During Phase 3 Takeover Processing
- Takeover Delays and Purger Restarts
- Takeover Restartability
- Takeover and File Recovery
- The Effects of Undoing Network Transactions
- Takeover and the RETAINCOUNT Value
- Network Configurations and Shared Access NonStop SQL/MP DDL Operations
- Network Validation and Considerations
- RDF Re-Initialization in a Network Environment
- RDF Networks and ABORT or STOP RDF Operations
- RDF Networks and Stop-Update-to-Time Operations
- Sample Configurations
- RDFCOM STATUS Display
- 14 Process-Lockstep Operation
- Starting a Lockstep Operation
- The DoLockstep Procedure
- The Lockstep Transaction
- RDF Lockstep File
- Multiple Concurrent Lockstep Operations
- The Lockstep Gateway Process
- Disabling Lockstep
- Reenabling Lockstep
- Lockstep Performance Ramifications
- Lockstep and Auxiliary Audit Trails
- Lockstep and Network Transactions
- Lockstep Operation Event Messages
- 15 NonStop SQL/MX and RDF
- Including and Excluding SQL/MX Objects
- Obtaining ANSI Object Names From Updater Event Messages
- Creating NonStop SQL/MX Primary and Backup Databases from Scratch
- Creating a NonStop SQL/MX Backup Database From an Existing Primary Database
- Online Database Synchronization With NonStop SQL/MX Objects
- Offline Synchronization for a Single Partition
- Online Synchronization for a Single Partition
- Correcting Incorrect NonStop SQL/MX Name Mapping
- Consideration for Creating Backup Tables
- Restoring to a Specific Location
- Comparing NonStop SQL/MX Tables
- 16 Zero Lost Transactions (ZLT)
- A RDF Command Summary
- B Additional Reference Information
- C Messages
- D Operational Limits
- E Using ASAP
- Index
Maintaining the Databases
HP NonStop RDF System Management Manual—524388-003
6-6
NonStop SQL/MP or NonStop SQL/MX Databases
It is safe to perform these operations just like other DDL operations on the primary
system.
DDL Operation Example
This example shows the SQLCI/MXCI commands for adding an index to a table and
the order of the operations:
1. Specify the default catalog for the backup system.
CATALOG \BACK.$TEST.DBCAT;
2. Create an index on the backup system.
CREATE INDEX \BACK.$DATA1.DB.FIRST
ON \BACK.$DATA1.DB.EMPLOYEE( FIRST_NAME, LAST_NAME );
3. Specify the default catalog for the primary system.
CATALOG \PRIM.$TEST.DBCAT;
4. Create an index on the primary system that corresponds to the index created on
the backup system.
CREATE INDEX \PRIM.$DATA1.DB.FIRST
ON \PRIM.$DATA1.DB.EMPLOYEE( FIRST_NAME, LAST_NAME );
Note that you should use WITH SHARED ACCESS for the CREATE INDEX operations
in the above example if both RDF and the application are running.
Multiple Indexes On a Single Base Table
Note that these same issues also apply to NonStop SQL/MX.
If there are multiple indexes on a single base table, special considerations apply when
you use SQLCI CREATE INDEX commands on the backup system to coordinate
NonStop SQL/MP DDL operations between the primary and backup databases.
Each NonStop SQL/MP index is assigned a unique key specifier that is stored as part
of the key for that index. You can explicitly define the key specifier by including the
KEYTAG clause in the CREATE INDEX command. If you do not do so, then the
CREATE INDEX operation assigns a numeric value based on the order of index
creation (1, 2, 3, and so forth).
Because the key specifier is part of the key of every index row created on an RDF
primary system, it also becomes part of the associated TMF audit record. RDF
transmits the audit record to the backup system where it is then applied to the backup
copy of the index.
If a CREATE INDEX command on the backup system does not include the KEYTAG
clause (and if you are not extremely careful to create the indexes in the order shown by
a SQLCI FILEINFO <base table>, DETAIL command on the primary system), it is
possible for the key specifier of a backup index to be different than that of the primary
index. In such a case, the index rows transmitted from the primary system to the