0% found this document useful (0 votes)
118 views403 pages

MIMIX Operations 5250

Uploaded by

ayishamma29
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
118 views403 pages

MIMIX Operations 5250

Uploaded by

ayishamma29
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd

Version 8.

MIMIX Operations—5250
Notices

MIMIX Operations - 5250 User Guide


February 2018
Version: 8.1.23.00
© Copyright 1999, 2018 Vision Solutions®, Inc. All rights reserved.
The information in this document is subject to change without notice and is furnished under a license
agreement. This document is proprietary to Vision Solutions, Inc., and may be used only as authorized in our
license agreement. No portion of this manual may be copied or otherwise reproduced without the express
written consent of Vision Solutions, Inc.
Vision Solutions provides no expressed or implied warranty with this manual.
The following are trademarks or registered trademarks of their respective organizations or companies:
• MIMIX and Vision Solutions are registered trademarks and Data Manager, Director, Dynamic Apply,
ECS/400, IntelliStart, Integrator, iOptimize, iTERA, iTERA Availability, MIMIX AutoNotify, MIMIX Availability,
MIMIX Availability Manager, MIMIX Availability for AIX, MIMIX DB2 Replicator, MIMIX Director, MIMIX dr1,
MIMIX Enterprise, MIMIX Global, MIMIX Monitor, MIMIX Object Replicator, MIMIX for PowerHA, MIMIX
Professional, MIMIX Promoter, OMS/ODS, MIMIX DR for AIX, MIMIX Share, RJ Link, SAM/400, Switch
Assistant, Vision AutoValidate, and Vision Suite are trademarks of Vision Solutions, Inc.
• AIX, AIX 5L, AS/400, DB2, eServer, IBM, Informix, i5/OS, iSeries, OS/400, Power, PowerHA, System i,
System i5, System p, System x, System z, and WebSphere—International Business Machines Corporation.
• Adobe and Acrobat Reader—Adobe Systems, Inc.
• HP-UX—Hewlett-Packard Company.
• Teradata—Teradata Corporation.
• Intel—Intel Corporation.
• Linux—Linus Torvalds.
• Excel, Internet Explorer, Microsoft, Windows, and Windows Server—Microsoft Corporation.
• Mozilla and Firefox—Mozilla Foundation.
• Java, Solaris, Oracle—Oracle Corporation.
• Red Hat—Red Hat, Inc.
• Sybase—Sybase, Inc.
• UNIX and UNIXWare—the Open Group.
All other brands and product names are trademarks or registered trademarks of their respective owners.
If you need assistance, contact Vision Solutions’ CustomerCare team at:
CustomerCare
Vision Solutions, Inc.
Telephone: 1.800.337.8214 or 1.949.724.5465
Email: [email protected]
Web Site: http://www.visionsolutions.com/services-support/support/contact-customercare
Contents
Who this book is for................................................................................................... 11
What is in this book ............................................................................................. 11
The MIMIX documentation set .................................................................................. 11
MIMIX DR limitations................................................................................................. 12
Sources for additional information............................................................................. 14
How to contact us...................................................................................................... 15
Chapter 1 MIMIX overview 16
MIMIX concepts......................................................................................................... 18
Product concepts................................................................................................. 18
System role concepts .......................................................................................... 19
Journaling concepts ............................................................................................ 20
Configuration concepts........................................................................................ 21
Process concepts ................................................................................................ 22
Additional switching concepts ............................................................................. 23
Best practices for maintaining your MIMIX environment ........................................... 24
Authority to products and commands........................................................................ 24
Accessing the MIMIX Main Menu.............................................................................. 26
Chapter 2 MIMIX policies 28
Environment considerations for policies.................................................................... 29
Policies in environments with more than two nodes or bi-directional replication. 29
Policies for data protection reports ...................................................................... 30
When to disable automatic recovery for replication and auditing ........................ 30
Disabling audits and recovery when using the MIMIX CDP feature .............. 31
Setting policies - general ........................................................................................... 31
Changing policies for an installation .................................................................... 31
Changing policies for a data group...................................................................... 32
Resetting a data group-level policy to use the installation level value ................ 32
Policies which affect an installation ........................................................................... 33
Changing directory depth for data protection reports .......................................... 33
Changing retention criteria for procedure history ................................................ 33
Changing the system ASP percent used threshold for recoveries ...................... 34
Policies which affect replication................................................................................. 35
Changing the source capture delay for recoveries.................................................... 36
Changing the source capture delay for recoveries during replication ................. 37
Changing the source capture delay used during virtual switch testing................ 37
Policies which affect auditing behavior...................................................................... 40
Changing auditing policies ........................................................................................ 42
Changing the audit level policy when switching .................................................. 42
Changing the system where audits are performed.............................................. 42
Changing retention criteria for audit history......................................................... 43
Restricting auditing based on the state of the data group ................................... 43
Changing the severity reported for audit statuses ............................................... 44
Changing what audits are checked for compliance ............................................. 45
Preventing audits from running ........................................................................... 46
Disabling all auditing for an installation ......................................................... 46
Disabling all auditing for a data group ........................................................... 47
Policies for switching with model switch framework .................................................. 48
Specifying a default switch framework in policies ............................................... 48

3
Setting polices for MIMIX Switch Assistant ......................................................... 49
Setting policies when MIMIX Model Switch Framework is not used.................... 49
Policy descriptions..................................................................................................... 50
Chapter 3 Scheduling audits and reports 60
How audits and reports are submitted automatically................................................. 60
Scheduling criteria you can change for audits and reports ................................. 61
Default schedules for automatically submitted audits ............................................... 64
Default schedules for automatically submitted reports.............................................. 67
Changing when automatic audits are allowed to run................................................. 68
Changing scheduling criteria for automatic audits............................................... 68
Changing the selection frequency of priority auditing categories ........................ 69
Disabling or enabling automatically submitted audits.......................................... 70
Changing when automatic reports are allowed to run ............................................... 72
Changing scheduling criteria for automatic reports ............................................. 72
Disabling or enabling automatically submitted reports ........................................ 73
Chapter 4 Where to check status 75
Checking status in application group environments .................................................. 75
Checking status in data group-only environments .................................................... 78
Chapter 5 Checking application group status 80
Displaying application group status........................................................................... 80
Resolving problem statuses for an application group................................................ 82
Resolving a procedure status problem reported for an application group ........... 82
Resolving an *ATTN status for an application group........................................... 83
Resolving other common status values for an application group ........................ 84
Resolving problems reported in the Monitors field .............................................. 85
Resolving status for node entries .............................................................................. 87
Resolving status for data resource group entries...................................................... 88
Verifying the sequence of the recovery domain ........................................................ 90
Changing the sequence of backup nodes ................................................................. 91
Examples of changing the backup sequence ...................................................... 93
Chapter 6 Working with procedures and steps 97
Displaying status of procedures ................................................................................ 98
Responding to a procedure in *MSGW status......................................................... 101
Resolving a *FAILED or *CANCELED procedure status......................................... 102
Displaying status of steps within a procedure run ................................................... 103
Resolving problems with step status ....................................................................... 106
Responding to a step with a *MSGW status ..................................................... 108
Resolving *CANCEL or *FAILED step statuses ................................................ 109
Acknowledging a procedure .................................................................................... 110
Running a procedure............................................................................................... 111
Resuming a procedure ...................................................................................... 112
Overriding the attributes of a step ............................................................... 112
Canceling a procedure ............................................................................................ 114
Displaying scheduling information for automatic reports......................................... 114
Displaying available status history of procedure runs ............................................. 115

4
Chapter 7 Working with data group status 117
Displaying data group status ................................................................................... 118
Identifying and resolving problems reflected in the Data Group column ........... 120
Resolving problems highlighted in the Data Group column......................... 121
Identifying and resolving replication problems reflected in the Source and Target
columns ................................................................................................................... 123
Setting the automatic refresh interval ................................................................ 125
Displaying detailed status for a data group ............................................................. 126
Merged view ...................................................................................................... 126
Object detailed status views .............................................................................. 131
Database detailed status views ......................................................................... 133
Identifying replication processes with backlogs....................................................... 137
Data group status in environments with journal cache or journal state ................... 139
Resolving a problem with journal cache or journal state ................................... 141
Resolving a problem for a journal status of *DIFFJRN...................................... 143
Resolving a deadlock condition for apply processes............................................... 144
Chapter 8 Working with audits 145
Auditing overview .................................................................................................... 146
Components of an audit .................................................................................... 146
Phases of audit processing ............................................................................... 147
Object selection methods for automatic audits.................................................. 147
How priority auditing determines what objects to select.............................. 148
Audit status and results ..................................................................................... 149
Audit compliance ............................................................................................... 150
Guidelines and considerations for auditing ............................................................. 151
Auditing best practices ...................................................................................... 151
Considerations for specific audits...................................................................... 152
Recommendations when checking audit results ............................................... 153
Displaying scheduling information for automatic audits .......................................... 155
Displaying audit runtime status and compliance status........................................... 157
Audits - Summary view and alternate columns ................................................. 157
Audits - Compliance summary view and alternate columns .............................. 160
Resolving auditing problems ................................................................................... 162
Resolving audit runtime status problems .......................................................... 162
Checking the job log of an audit ........................................................................ 166
Resolving audit compliance status problems .................................................... 167
Running an audit immediately ................................................................................. 169
Ending an audit ....................................................................................................... 170
Displaying audit history ........................................................................................... 171
Prioritized audit runs with no selected objects .................................................. 173
Working with audited objects................................................................................... 173
Displaying audited objects from a specific audit run ......................................... 175
Displaying a customized list of audited objects ................................................. 175
Working with audited object history......................................................................... 176
Displaying the audit history for a specific object................................................ 177
Chapter 9 Working with system-level processes 178
Displaying status of system-level processes........................................................... 180
Resolving system manager problems and backlogs ............................................... 181

5
Resolving ASP 1 storage exceeded threshold problems ........................................ 183
Starting system managers or a journal manager .................................................... 184
Ending system managers........................................................................................ 184
Ending a journal manager ....................................................................................... 185
Starting collector services ....................................................................................... 185
Ending collector services......................................................................................... 186
Starting target journal inspection processes ........................................................... 186
Ending target journal inspection processes............................................................. 187
Displaying status of target journal inspection .......................................................... 188
Displaying results of target journal inspection ......................................................... 189
Displaying details associated with target journal inspection notifications.......... 190
Displaying messages for TGTJRNINSP notifications.................................. 190
Identifying the last entry inspected on the target system ........................................ 191
Displaying status of monitors on the local system................................................... 192
Starting a monitor .................................................................................................... 194
Ending a monitor ..................................................................................................... 195
Starting and ending the master monitors ................................................................ 195
Starting the master monitor on the local system ............................................... 196
Ending the master monitor on the local system ................................................ 196
Enabling or disabling a monitor ............................................................................... 197
Enabling a monitor ............................................................................................ 197
Disabling a monitor............................................................................................ 197
Displaying status of a remote journal link................................................................ 199
Starting a remote journal link independently ........................................................... 202
Ending a remote journal link independently ............................................................ 202
Chapter 10 Working with notifications and recoveries 204
What are notifications and recoveries ..................................................................... 204
Displaying notifications............................................................................................ 205
What information is available for notifications ................................................... 206
Detailed information..................................................................................... 207
Options for working with notifications ................................................................ 208
Displaying IFS entries associated with notifications from INSTALL .................. 208
Notifications for newly created objects .................................................................... 210
Displaying recoveries .............................................................................................. 211
What information is available for recoveries in the native user interface........... 212
Detailed information..................................................................................... 212
Options for working with recoveries in the native user interface ....................... 213
Orphaned recoveries in the native user interface.............................................. 213
Determining whether a recovery is orphaned.............................................. 214
Removing an orphaned recovery ................................................................ 214
Chapter 11 Starting and ending replication 215
Before starting replication........................................................................................ 217
Commands for starting replication........................................................................... 217
What is started with the STRMMX command.................................................... 217
STRMMX and ENDMMX messages............................................................ 218
What is started by the default START procedure for an application group ....... 218
Choices when starting or ending an application group...................................... 218
What occurs when a data group is started .............................................................. 221

6
Journal starting point identified on the STRDG request .................................... 222
Journal starting point when the object send process is shared ................... 222
Clear pending and clear error processing ......................................................... 223
Starting MIMIX......................................................................................................... 226
Starting an application group................................................................................... 227
Starting selected data group processes .................................................................. 228
Starting replication when open commit cycles exist ................................................ 230
Checking for open commit cycles...................................................................... 230
Resolving open commit cycles .......................................................................... 230
Before ending replication......................................................................................... 232
Commands for ending replication............................................................................ 232
Command choice by reason for ending replication ........................................... 232
Additional considerations when ending replication............................................ 234
Ending immediately or controlled ...................................................................... 234
Controlling how long to wait for a controlled end to complete ..................... 235
Ending all or selected processes....................................................................... 235
When to end RJ links ........................................................................................ 236
What is ended by the ENDMMX command ....................................................... 236
What is ended by the default END procedure for an application group ............ 237
What occurs when a data group is ended ............................................................... 238
Ending MIMIX.......................................................................................................... 240
Ending with default values................................................................................. 240
Ending by prompting the ENDMMX command.................................................. 240
After you end MIMIX products ........................................................................... 241
Ending an application group.................................................................................... 243
Ending a data group in a controlled manner ........................................................... 244
Preparing for a controlled end of a data group .................................................. 244
Performing the controlled end ........................................................................... 244
Confirming the end request completed without problems ................................. 245
Ending selected data group processes ................................................................... 247
Enabling MIMIX-disabled target constraints............................................................ 248
Enabling constraints for an *INACTIVE data group........................................... 249
Enabling constraints when ending a data group ............................................... 249
What replication processes are started by the STRDG command.......................... 251
What replication processes are ended by the ENDDG command .......................... 255
Chapter 12 Resolving common replication problems 259
Replication problems that MIMIX attempts to recover............................................. 260
Replication errors handled by automatic database recovery ............................ 260
Replication errors handled by automatic object recovery.................................. 261
Working with user journal replication errors ............................................................ 263
Working with files needing attention (replication and access path errors)......... 263
Working with journal transactions for files in error....................................... 266
Placing a file on hold ......................................................................................... 267
Ignoring a held file ............................................................................................. 267
Releasing a held file at a synchronization point ................................................ 268
Releasing a held file .......................................................................................... 268
Releasing a held file and clearing entries.......................................................... 269
Correcting file-level errors ................................................................................. 269
Correcting record-level errors............................................................................ 270

7
Record written in error ................................................................................. 270
Working with tracking entries .................................................................................. 272
Accessing the appropriate tracking entry display .............................................. 272
Holding journal entries associated with a tracking entry ................................... 274
Ignoring journal entries associated with a tracking entry................................... 275
Waiting to synchronize and release held journal entries for a tracking entry .... 275
Releasing held journal entries for a tracking entry ............................................ 276
Releasing and clearing held journal entries for a tracking entry........................ 276
Removing a tracking entry................................................................................. 276
Related tracking entries..................................................................................... 277
Working with objects in error ................................................................................... 278
Using the Work with DG Activity Entries display ............................................... 279
Retrying data group activity entries ................................................................... 281
Retrying a failed data group activity entry ................................................... 281
Determining whether an activity entry is in a delay/retry cycle .......................... 282
Working with message queues ............................................................................... 283
Working with the message log ................................................................................ 284
Removing data group activity history entries........................................................... 285
Chapter 13 Starting, ending, and verifying journaling 286
What objects need to be journaled.......................................................................... 287
Authority requirements for starting journaling.................................................... 288
MIMIX commands for starting journaling................................................................. 289
Forcing objects to use the configured journal.................................................... 290
Journaling for physical files ..................................................................................... 291
Displaying journaling status for physical files .................................................... 291
Starting journaling for physical files ................................................................... 291
Ending journaling for physical files .................................................................... 292
Verifying journaling for physical files ................................................................. 293
Journaling for IFS objects........................................................................................ 294
Displaying journaling status for IFS objects ...................................................... 294
Starting journaling for IFS objects ..................................................................... 294
Ending journaling for IFS objects ...................................................................... 295
Verifying journaling for IFS objects.................................................................... 296
Journaling for data areas and data queues............................................................. 298
Displaying journaling status for data areas and data queues............................ 298
Starting journaling for data areas and data queues .......................................... 298
Ending journaling for data areas and data queues............................................ 299
Verifying journaling for data areas and data queues ......................................... 300
Chapter 14 Switching 302
About switching ....................................................................................................... 303
Planned switch - overview ................................................................................. 303
Unplanned switch - overview............................................................................. 304
Switching application group environments with procedures.............................. 305
Switching data group environments with MIMIX Model Switch Framework ...... 307
Switching an application group................................................................................ 308
Switching a data group-only environment ............................................................... 309
Switching to the backup system ........................................................................ 309
Synchronizing data and starting MIMIX on the original production system ....... 310

8
Switching to the production system ................................................................... 310
Determining when the last switch was performed ................................................... 311
Performing a data group switch............................................................................... 312
Switch Data Group (SWTDG) command................................................................. 314
Chapter 15 Support for virtual switching 316
What is virtual switching? ........................................................................................ 316
MIMIX virtual switch overview ................................................................................. 317
MIMIX environment considerations for virtual switching ......................................... 319
Testing considerations for virtual switching............................................................. 321
Performing a precheck for virtual switching............................................................. 323
Performing a virtual switch ...................................................................................... 323
Chapter 16 Less common operations 327
Starting the TCP/IP server ...................................................................................... 328
Ending the TCP/IP server........................................................................................ 329
Working with objects ............................................................................................... 330
Displaying long object names............................................................................ 330
Considerations for working with long IFS path names ................................ 330
Displaying data group spooled file information.................................................. 330
Viewing status for active file operations .................................................................. 331
Identifying data groups that use an RJ link ............................................................. 332
Identifying journal definitions used with RJ ............................................................. 333
Disabling and enabling data groups ........................................................................ 334
Disabling a data group ...................................................................................... 335
Enabling a data group ....................................................................................... 335
Determining if non-file objects are configured for user journal replication............... 336
Determining how IFS objects are configured .................................................... 336
Determining how data areas or data queues are configured ............................ 337
Using file identifiers (FIDs) for IFS objects .............................................................. 338
Chapter 17 Troubleshooting - where to start 339
Reducing contention between MIMIX and user applications................................... 341
Data groups cannot be ended ................................................................................. 342
Verifying a communications link for system definitions ........................................... 343
Verifying the communications link for a data group................................................. 344
Verifying all communications links..................................................................... 344
Checking file entry configuration manually.............................................................. 345
Data groups cannot be started ................................................................................ 347
Cannot start or end an RJ link................................................................................. 348
Removing unconfirmed entries to free an RJ link.............................................. 348
RJ link active but data not transferring .................................................................... 349
Errors using target journal defined by RJ link.......................................................... 350
Verifying data group file entries............................................................................... 351
Verifying key attributes ............................................................................................ 351
Working with data group timestamps ...................................................................... 352
Automatically creating timestamps .................................................................... 352
Creating additional timestamps ......................................................................... 352
Creating timestamps for remote journaling processing ..................................... 353
Deleting timestamps .......................................................................................... 354
Displaying or printing timestamps ..................................................................... 354

9
Removing journaled changes.................................................................................. 355
Performing journal analysis ..................................................................................... 356
Removing journal analysis entries for a selected file ........................................ 358
Appendix A Interpreting audit results - supporting information 360
Interpreting results for configuration data - #DGFE audit........................................ 361
When the difference is “not found” .......................................................................... 363
Interpreting results of audits for record counts and file data ................................... 364
What differences were detected by #FILDTA.................................................... 364
What differences were detected by #MBRRCDCNT ......................................... 365
Interpreting results of audits that compare attributes .............................................. 367
What attribute differences were detected .......................................................... 367
Where was the difference detected................................................................... 369
What attributes were compared ........................................................................ 370
Appendix B IBM Power™ Systems operations that affect MIMIX 371
MIMIX procedures when performing an initial program load (IPL) .......................... 372
MIMIX procedures when performing an operating system upgrade........................ 374
Prerequisites for performing an OS upgrade on either system ......................... 375
MIMIX-specific steps for an OS upgrade on a backup system.......................... 375
MIMIX-specific steps for an OS upgrade on a production system with switching ...
377
MIMIX-specific steps for an OS upgrade on the production system without switch-
ing............................................................................................................................ 379
MIMIX procedures when upgrading hardware without a disk image change .......... 382
Considerations for performing a hardware system upgrade without a disk image
change..................................................................................................................... 382
MIMIX-specific steps for a hardware upgrade without a disk image change..... 382
Hardware upgrade without a disk image change - preliminary steps .......... 382
Hardware upgrade without a disk image change - subsequent steps ......... 383
MIMIX procedures when performing a hardware upgrade with a disk image change...
385
Considerations for performing a hardware system upgrade with a disk image
change..................................................................................................................... 385
MIMIX-specific steps for a hardware upgrade with a disk image change.......... 386
Hardware upgrade with a disk image change - preliminary steps ............... 386
Hardware upgrade with a disk image change - subsequent steps .............. 387
Handling MIMIX during a system restore ................................................................ 391
Prerequisites for performing a restore of MIMIX ............................................... 391
Index 392

10
Who this book is for

Who this book is for


The MIMIX Operations - 5250 book describes how to perform routine operational
tasks and basic troubleshooting for MIMIX® Enterprise™, MIMIX® Professional™, and
MIMIX® DR, and from the native user interface on systems running IBM i. Some
topics may not apply to all products.

What is in this book


The MIMIX Operations - 5250 book provides these distinct types of information:
• A summary of concepts within MIMIX
• Application group and data group status and troubleshooting
• Audit status, troubleshooting, scheduling, and history
• Procedures for starting, ending, and switching replication
• Procedures for starting, ending, and verifying journaling
• Procedures for handling MIMIX when performing operations such as IPLs or
hardware and operating system upgrades.

The MIMIX documentation set


The following documents about MIMIX® products are available:
Using License Manager
License Manager currently supports MIMIX®, iOptimize™, and iTERA®
Availability™. This book describes software requirements, system security, and
other planning considerations for installing software and software fixes for Vision
Solutions products that are supported through License Manager. The preferred
way to obtain license keys and install software is by using Vision AutoValidate™
and the product’s Installation Wizard. However, if you cannot use the wizard or
AutoValidate, this book provides instructions for obtaining licenses and installing
software from a native user interface. This book also describes how to use the
additional security functions from Vision Solutions which are available for License
Manager and MIMIX and implemented through License Manager.
MIMIX Administrator Reference
This book provides detailed conceptual, configuration, and programming
information for MIMIX products. It includes checklists for setting up several
common configurations, information for planning what to replicate, and detailed
advanced configuration topics for custom needs. It also identifies what information
can be returned in outfiles if used in automation. Some topics may not be
applicable to all MIMIX products.
Using MIMIX from Vision Solutions Portal
This book is for administrators and operators who want a better user experience
when creating or modifying configuration, monitoring status, and resolving

11
problems in replication environments for one or more instances of MIMIX® DR,
MIMIX® Enterprise™ or MIMIX® Professional™. The modern web browser
interfaces available through Vision Solutions Portal (VSP) and the MIMIX portal
application provide a better user experience for MIMIX and are required to use
this book. Some topics may not be applicable to all MIMIX products.
MIMIX Operations with IBM PowerHA for i
This book is for administrators and operators in an IBM i clustering environment
who use MIMIX® for PowerHA® to integrate cluster management with MIMIX
logical replication or supported hardware-based replication techniques. This book
focuses on addressing problems reported in MIMIX status and basic operational
procedures such as starting, ending, and switching.
MIMIX Operations - 5250
This book provides high level concepts and operational procedures for managing
your high availability environment using MIMIX products from a native user
interface. This book focuses on tasks typically performed by an operator, such as
checking status, starting or stopping replication, performing audits, and basic
problem resolution.
Using MIMIX Monitor
This book describes how to use the MIMIX Monitor user and programming
interfaces available with MIMIX products. This book also includes programming
information about MIMIX Model Switch Framework.
Using MIMIX Promoter
This book describes how to use MIMIX commands for copying and reorganizing
active files. MIMIX Promoter functionality is included with MIMIX® Enterprise™.
MIMIX for IBM WebSphere MQ
This book identifies requirements for the MIMIX for MQ feature which supports
replication in IBM WebSphere MQ environments. This book describes how to
configure MIMIX for this environment and how to perform the initial
synchronization and initial startup. Once configured and started, all other
operations are performed as described in the MIMIX Operations - 5250 book.

MIMIX DR limitations
Environments that replicate in one direction between only two systems may not
require the features offered by a high availability product. Instead, a disaster recovery
solution that enables data to be readily available is all that is needed. MIMIX® DR
features many of the capabilities found in Vision Solutions high availability products,
with the following specifications that make it strictly a disaster recovery solution:
• Processor group - MIMIX DR is limited to P05 and P10 systems with 1 to 4 CPUs
(cores). The backup system can be any P-Group model.
• MIMIX Instance - Only one MIMIX Instance is allowed on each participating
system. A MIMIX DR installation comprises two systems that transfer data and
objects between them. Both systems within the instance use the same name for

12
MIMIX DR limitations

the library in which the product is installed.


• Management system - One system in every MIMIX installation is designated as a
management system where operations such as configuration are performed.
Because MIMIX DR is a simple, two system environment, it does not support
designating multiple management systems. Multi-management allows certain
MIMIX products to be managed from multiple systems in the MIMIX instance.
• Virtual failover (virtual switching) - MIMIX DR supports virtual failover as a means
of testing that production applications can function properly on the backup system
within the context of your MIMIX DR replication environment without actually
having to end your production environment or replication. Virtual failover is
available to MIMIX DR installations running version 8.1.02.00 or higher.
• Switching - MIMIX DR does not support planned or unplanned switching.
Switching is the process by which a production environment is automatically
moved from one system to another system and replication of the production
environment data can be performed from that system.
• Keyed replication - Keyed replication is not supported for MIMIX DR. MIMIX DR
uses positional file replication to identify the target system location to be updated
with replicated data. Positional replication identifies target system data by
position, or relative record number (RRN), in the file member. Keyed replication
updates files based on key values within the data and may be appropriate in
complex environments.
• Cluster Resource Services - MIMIX DR does not support Cluster Resource
Services, which is part of the base IBM i operating system. Cluster resource
services provides the integrated services and application programming interfaces
(APIs) necessary to create and manage an IBM cluster.
• MIMIX for WebSphere MQ - MIMIX DR does not support the separately licensed
feature, MIMIX for WebSphere MQ. This feature enables an IBM WebSphere MQ
environment to be included as part of a managed availability solution. IBM
WebSphere MQ is an application that transmits messages between systems
running on a variety of platforms.
• Copy While Active/Reorganize While Active - Copy/Reorganize while active
capabilities are not supported by MIMIX DR. These capabilities copy or
reorganize database files while the files are in active production using the Copy
Active File (CPYACTF) and Reorganize Active File (RGZACTF) commands.
• Name mapping - Name mapping is not supported by MIMIX DR. Name mapping is
an advanced configuration technique that allows files to be renamed as they are
replicated. Typically this capability is used by more complex environments where
objects exist in different libraries or paths.
• INTRA - INTRA configurations are not supported in MIMIX DR because they do
not provide disaster recovery. In an INTRA configuration, replication occurs within
a single system. Intra is a special configuration that allows certain MIMIX products
to function fully within a single system environment.

13
Sources for additional information
This book refers to other published information. The following information, plus
additional technical information, can be located in the IBM Knowledge Center.

From the Information center you can access these IBM Power™ Systems topics,
books, and redbooks:
• Backup and Recovery
• Journal management
• DB2 Universal Database for IBM Power™ Systems Database Programming
• Integrated File System Introduction
• Independent disk pools
• TCP/IP Setup
• IBM redbook Striving for Optimal Journal Performance on DB2 Universal
Database for iSeries, SG24-6286
• IBM redbook AS/400 Remote Journal Function for High Availability and Data
Replication, SG24-5189
• IBM redbook Power™ Systems iASPs: A Guide to Moving Applications to
Independent ASPs, SG24-6802
The following information may also be helpful if you replicate journaled data areas,
data queues, or IFS objects:
• DB2 UDB for iSeries SQL Programming Concepts
• DB2 Universal Database for iSeries SQL Reference
• IBM redbook AS/400 Remote Journal Function for High Availability and Data
Replication, SG24-5189

14
How to contact us

How to contact us
For contact information, visit our Contact CustomerCare web page.
If you are current on maintenance, support for MIMIX products is also available when
you log in to Support Central.
It is important to include product and version information whenever you report
problems.

15
MIMIX overview

CHAPTER 1 MIMIX overview

This book provides operational information and procedures for using MIMIX®
Enterprise™, MIMIX® Professional™, or MIMIX® DR through the native user interface
user interface. For simplicity, this book uses the term MIMIX to refer to the
functionality provided unless a more specific name is necessary. Some topics may not
apply to all products.
This MIMIX version provides high availability for your critical data in a production
environment on IBM Power™ Systems through real-time replication of changes and
the ability to quickly switch your production environment to a ready backup system.
These capabilities allow your business operations to continue when you have planned
or unplanned outages in your System i environment. MIMIX also provides advanced
capabilities that can help ensure the integrity of your MIMIX environment.
Replication: MIMIX continuously captures changes to critical database files and
objects on a production system, sends the changes to a backup system, and applies
the changes to the appropriate database file or object on the backup system. The
backup system stores exact duplicates of the critical database files and objects from
the production system.
MIMIX uses two replication paths to address different pieces of your replication
needs. These paths operate with configurable levels of cooperation or can operate
independently.
• The user journal replication path captures changes to critical files and objects
configured for replication through a user journal. When configuring this path,
shipped defaults use the remote journaling function of the operating system to
simplify sending data to the remote system. In previous versions, MIMIX DB2
Replicator provided this function.
• The system journal replication path handles replication of critical system objects
(such as user profiles, program objects, or spooled files), integrated file system
(IFS) objects, and document library object (DLOs) using the system journal. In
previous versions MIMIX Object Replicator provided this function.
Configuration choices determine the degree of cooperative processing used between
the system journal and user journal replication paths when replicating database files,
IFS objects, data areas, and data queues.
Switching: One common use of MIMIX is to support a hot backup system to which
operations can be switched in the event of a planned or unplanned outage. If a
production system becomes unavailable, its backup is already prepared for users. In
the event of an outage, you can quickly switch users to the backup system where they
can continue using their applications. MIMIX captures changes on the backup system
for later synchronization with the original production system. When the original
production system is brought back online, MIMIX assists you with analysis and
synchronization of the database files and other objects.
Automatic verification and correction: MIMIX assists in maintaining availability and
switch-readiness in your replication environment by automatically detecting problems

16
during replication, verifying replicated objects as well as configuration integrity with
audits, and automatically attempting to correct any problems detected.
MIMIX is shipped with these capabilities enabled. Incorporated best practices for
maintaining availability and switch-readiness are key to ensuring that your MIMIX
environment is in tip-top shape for protecting your data. User interfaces allow you to
fine-tune to the needs of your environment.
Analysis: MIMIX also provides advanced analysis capabilities through the MIMIX
portal application for Vision Solutions Portal (VSP). When using the VSP user
interface from a web browser, you can view a list of the objects replicated by MIMIX
and a list of objects with detected problems or changes for which automatic recovery
actions have been requested or are in progress. You can use reports to view which
objects on a system are and are not identified to MIMIX configuration. You can also
check historical arrival and backlog rates for replication to help you identify trends in
your operations that may affect MIMIX performance.
Uses: MIMIX is typically used among systems in a network to support a hot backup
system. Simple environments have one production system and one backup system.
More complex environments have multiple production systems or backup systems.
MIMIX can also be used on a single system.
You can view the replicated data on the backup system at any time without affecting
productivity. This allows you to generate reports, submit (read-only) batch jobs, or
perform backups to tape from the backup system. In addition to real-time backup
capability, replicated databases and objects can be used for distributed processing,
allowing you to off-load applications to a backup system.
The topics in this chapter include:
• “MIMIX concepts” on page 18 summarizes key concepts that you need to know
about MIMIX.
• “Best practices for maintaining your MIMIX environment” on page 24 summarizes
recommendations from Vision Solutions.
• “Authority to products and commands” on page 24 identifies authority levels to
MIMIX functions when additional security features provided by Vision Solutions
are used.
• “Accessing the MIMIX Main Menu” on page 26 describes the MIMIX Basic Main
menu and the MIMIX Intermediate Main Menu.

17
MIMIX concepts
The following subtopics organize the basic concepts associated with MIMIX® into
related groups. More detailed information is available in the MIMIX Administrator
Reference book.

Product concepts
MIMIX installation - The network of IBM Power™ Systems systems that transfer data
and objects among each other using functions of a common MIMIX product. A MIMIX
installation is defined by the way in which you configure the MIMIX product for each of
the participating systems. A system can participate in multiple independent MIMIX
installations.
Replication - The activity that MIMIX performs to continuously capture changes to
critical database files and objects on a production system as they occur, send the
changes to a backup system, and apply the changes to the appropriate database file
or object on the backup system.
Switch - The process by which a production environment is moved from one system
to another system and the production environment is made available there. A switch
may be performed as part of a planned event such as for system maintenance, or an
unplanned event such as a power or equipment failure. MIMIX provides customizable
functions for switching.
Audits - Audits are predetermined programs that are used to check for differences in
replicated objects and other conditions between systems. Audits run and can correct
detected problems automatically. Policies control when audits run and many other
aspects of how audits are performed. Additional auditing concepts and
recommendations are described in the auditing chapter of this book.
Automatic recovery - MIMIX provides a set of functions that are able to detect and
automatically correct problems in a MIMIX installation. Processes that can detect
problems and initiate automatic recovery actions include database replication, object
replication, target journal inspection, audits, and virtual switch procedures. Through
policies, you have the ability to disable automatic recovery at the installation or data
group level.
Application group - A MIMIX construct used to group and control resources from a
single point in a way that maintains relationships between them. The use of
application groups is best practice for MIMIX Professional and MIMIX Enterprise and
required for MIMIX for PowerHA.
Data group - A MIMIX construct that is used to control replication activities. A data
group is a logical grouping of database files, data areas, objects, IFS objects, DLOs,
or a combination thereof that defines a unit of work by which MIMIX replication activity
is controlled. A data group may represent an application, a set of one or more
libraries, or all of the critical data on a given system. Application environments may
define a data group as a specific set of files and objects.
Prioritized status - MIMIX assigns a priority to status values to ensure that problems
with the highest priorities, those for detected problems or situations that require

18
MIMIX concepts

immediate attention or intervention, are reflected on the highest level of the user
interface. Additional detail and lower priority items can be viewed by drilling down to
the next level within the interfaces. Those interfaces are the Work with Systems
display and depending on your configuration, either the Work with Application Groups
display or the Work with Data Groups display.
Policies - A policy is a mechanism used to enable, disable, or provide input to a
function such as replication, auditing, or MIMIX Model Switch Framework. For most
policies, the initially shipped values apply to an installation. However, policies can be
changed and most can also be overridden for individual data groups. Policies that
control when audits are automatically performed can be set only for each specific
combination of audit rule and data group.
Notifications - A notification is the resulting automatic report associated with an
event that has already occurred. The severity of a notification is reflected in the
overall status of the installation. Notifications can be generated by a process,
program, command, or monitor. Because the originator of notifications varies, it is
important to note that notifications can represent both real-time events as well as
events that occurred in the past but, due to scheduling, are being reported in the
present.
Recoveries - The term recovery is used in two ways. The most common use refers to
the recovery action taken by a MIMIX process to correct a detected difference when
automatic recovery polices are enabled. The second use refers to a temporary report
that provides details about a recovery action in progress that is created when the
recovery action starts and is removed when it completes.

System role concepts


MIMIX uses several pairs of terms to refer to the role of a system within a particular
context. These terms are not interchangeable.
Production system and backup system - These terms describe the role of a system
relative to the way applications are used on that system.
A production system is the system currently running the production workload for the
applications. In normal operations, the production system is the system on which the
principal copy of the data and objects associated with the application exist.
A backup system is the system that is not currently running the production workload
for the applications. In normal operations, the backup system is the system on which
you maintain a copy of the data and objects associated with the application. These
roles are not always associated with a specific system. For example, if you switch
application processing to the backup system, the backup system temporarily
becomes the production system.
Typically, for normal operations in basic two-system environment, replicated data
flows from the system running the production workload to the backup system.
Source system and target system - These terms identify the direction in which an
activity occurs between two participating systems.
A source system is the system from which MIMIX replication activity between two
systems originates. In replication, the source system contains the journal entries.

19
Information from the journal entries is either replicated to the target system or used to
identify objects to be replicated to the target system.
A target system is the system on which MIMIX replication activity between two
systems completes.
Management system and network system - These terms define the role of a
system relative to how the products interact within a MIMIX installation. These roles
remain associated with the system within the MIMIX installation to which they are
defined. One system in the MIMIX installation is designated as the management
system and the remaining one or more systems are designated as network systems.
A management system is the system in a MIMIX installation that is designated as the
control point for all installations of the product within the MIMIX installation. The
management system is the location from which work to be performed by the product
is defined and maintained. Often the system defined as the management system also
serves as the backup system during normal operations.
A network system is any system in a MIMIX installation that is not designated as the
management system (control point) of that MIMIX installation. Work definitions are
automatically distributed from the management system to a network system. Often a
system defined as a network system also serves as the production system during
normal operations.

Journaling concepts
MIMIX uses journaling to perform replication and to support newer analysis
functionality.
Journaling and object auditing - Journaling and object auditing are techniques that
allow object activity to be logged to a journal. Journaling logs activity for selected
objects of specific object types to a user journal. Object auditing logs activity for all
objects to the security audit journal (QAUDJRN, the system journal), including those
defined to a user journal. MIMIX relies on these techniques and the entries placed in
the journal receivers for replicating logged activity.
Journal - An IBM i system object that identifies the objects being journaled and the
journal receivers associated with the journal. The system journal is a specialized
journal on the system which MIMIX uses.
Journal receiver - An IBM i system object that is associated with a journal and
contains the log of all activity for objects defined to the journal.
Journal entry - A record added to a journal receiver that identifies an event that
occurred on a journaled object. MIMIX uses file and record level journal entries to
recreate the object on a designated system.
Remote journaling - A function of IBM i that allows you to establish journals and
journal receivers on one system and associate them with specific journals and journal
receivers on another system. Once the association is established, the operating
system can use the pair of journals to replicate journal entries in one direction, from
the local journal to the remote journal on the other system. MIMIX uses remote
journaling to transfer replicated data from the source system to the target system and
to distributed information among systems.

20
MIMIX concepts

Configuration concepts
MIMIX configuration provides considerable flexibility to enable supporting a wide
variety of customer environments. Configuration is implemented through sets of
related commands. The following terms describe configuration concepts.
Definitions - MIMIX uses several types of named definitions to identify related
configuration choices.
• System definitions identify systems that participate in a MIMIX installation. Each
system definition identifies one system.
• Transfer definitions identify the communications path and protocol to be used
between systems.
• Journal definitions identify journaling environments that MIMIX uses for replication
Each journal definition identifies a system and characteristics of the journaling
environment on that system.
• Data group definitions identify the characteristics of how replication occurs
between two systems. Each data group definition determines the direction in
which replication occurs between the systems, whether that direction can be
switched, and the default processing characteristics for replication processes.
• Application group definitions identify whether the replication environment does or
does not use IBM i clustering. When clustering is used, the application group also
defines information about an application or proprietary programs necessary for
controlling operations in the clustering environment.
Data group entries - A data group entry is a configuration construct that identifies a
source of information to be replicated by or excluded from replication by a data group.
Each entry identifies at least one object and its location on the source system.
Classes of data group entries are based on object type. MIMIX uses data group
entries to determine whether a journal entry should be replicated. Data groups that
replicate from both the system journal and a user journal can have any combination of
data group entries.
Remote journal link (RJ link) - An RJ link is a MIMIX configuration element that
identifies an IBM i remote journaling environment used by user journal replication
processes. An RJ link identifies the journal definitions that define the source and
target journals, primary and secondary transfer definitions for the communications
path used by MIMIX, and whether the IBM i remote journal function sends journal
entries asynchronously or synchronously.
Cooperative processing - Cooperative processing refers to MIMIX techniques that
efficiently replicate certain object types by using a coordinated effort between the
system journal and user journal replication paths. Configuration choices in data group
definitions and data group entries determine the degree of cooperative processing
used between the system journal and user journal replication paths when replicating
database files, IFS objects, data areas, and data queues.
Tracking entries - Tracking entries identify objects that can be replicated using
advanced journaling techniques and assist with tracking the status of their replication.
A unique tracking entry is associated with each IFS object, data area, and data queue
that is eligible for replication using advanced journaling. IFS tracking entries identify

21
eligible, existing IFS objects while object tracking entries identify eligible, existing data
areas and data queues.

Process concepts
The following terms identify MIMIX processes. Some, like the system manager, are
required to allow MIMIX to function. Others, like procedures, are used only when
invoked by users.
Replication path - A replication path is a series of processes used for replication that
represent the critical path on which data to be replicated moves from its origin to its
destination. MIMIX uses two replication paths to accommodate differences in how
replication occurs for user journal and system journal entries. These paths operate
with configurable levels of cooperation or can operate independently.
• The user journal replication path captures changes to critical files and objects
configured for replication through a user journal. When configuring this path,
shipped defaults use the remote journaling function of the operating system to
simplify sending data to the remote system. The changes are applied to the target
system.
• The system journal replication path handles replication of critical system objects
(such as user profiles, program objects, or spooled files), integrated file system
(IFS) objects, and document library object (DLOs) using the system journal.
Information about the changes are sent to the target system where it is applied.
System manager - System manager processes automatically distribute configuration
and status changes among systems in a MIMIX installation. Although often thought of
as one entity per system, there are actually multiple system manager processes
associated with each system. Each system manager process consists of a remote
journal link (RJ Link) that transmits entries in one direction between a pair of systems
and a system manager processing job that applies the entries to MIMIX files on the
remote system (or target side) of the RJ link. Between each pair of communicating
systems there are always two system manager processes that transmit in opposite in
directions. In environments with more than two systems, MIMIX restricts which
systems can communicate based on their designation as a management (*MGT) or
network (*NET) system.
Journal manager - The journal manager is a job on each system that MIMIX uses to
maintain the journaling environment on that system. By default, MIMIX performs both
change management and delete management for journal receivers used by the
replication process.
Collector services - A group of jobs that are necessary for MIMIX to track historical
data and to support using the MIMIX portal application within the Vision Solutions
Portal. One or more collector service jobs collect and combine MIMIX status from all
systems.
Cluster services - When configured for use in an IBM i clustering environment,
MIMIX uses the cluster services function provided by IBM i to integrate the system
management functions needed for clustering. Cluster services must be active in order
for a cluster node to be recognized by the other nodes in the cluster. MIMIX integrates

22
MIMIX concepts

starting and stopping cluster services into status and commands for controlling
processes that run at the system level.
Target journal inspection - A set of MIMIX processes that read journals on a system
being used as the target system for replication. Each target journal inspection process
identifies people or processes other than MIMIX that created, changed, or deleted
replicated objects on the target system. The list of affected objects and any automatic
recovery actions initiated by MIMIX can be displayed in the Objects Changed on
Target window within the MIMIX portal application in Vision Solutions Portal. The
window can be accessed from multiple portlets and is filtered accordingly.
Procedures and steps - Procedures and steps are a highly customizable means of
performing operations. A set of default procedures is shipped with MIMIX for
frequently used operations. These default procedures include the ability to start or
end replication for an application group, perform pre-check activity for switching,
perform a planned or unplanned switch of an application group, perform virtual switch
testing, prepare MIMIX for running a backup, and run data protection reports for
nodes. Each operation is performed by a procedure that consists of a sequence of
steps and multiple jobs. Each step calls a predetermined step program to perform a
specific sub-task of the larger operation. Steps also identify runtime attributes that
determine how the procedure will start processing the step and how it will respond if
the step ends in error.
Log space - A MIMIX object that provides an efficient storage and manipulation
mechanism for replicated data that is temporarily stored on the target system during
the receive and apply processes.

Additional switching concepts


The following concepts are specific to switching.
In environments licensed for MIMIX DR, only virtual switching1 is supported.
Environments configured with application groups perform switching through
procedures.
Planned switch - An intentional change to the direction of replication for any of a
variety of reasons. You may need to take the system offline to perform maintenance
on its hardware or software, or you may be testing your disaster recovery plan. In a
planned switch, the production system (the source of replication) is available. When
you perform a planned switch, replication is ended on both the source and target
systems. The next time you start replication, it will be set to replicate in the opposite
direction.
Unplanned switch - A change the direction of replication as a response to a problem.
Most likely the production system is no longer available. When you perform an
unplanned switch, you must initiate the switch from the target system. Replication is
ended on the target system. The next time you start replication, it will be set to
replicate in the opposite direction.

1. MIMIX DR installations running version 8.1.02.00 or higher support virtual failover (virtual
switching)

23
Virtual switch - A way to test production applications on a node that could assume
the role of the production system in the event of a switch without actually performing a
planned switch. A virtual switch allows you to designate a backup node in an
application group for testing, sets up the MIMIX environment to allow user-performed
testing while the production environment and MIMIX remain active, and recovers the
MIMIX environment when testing is complete.
MIMIX Model Switch Framework - A set of programs and commands that provide a
consistent framework to be used when performing planned or unplanned switches in
environments that do not use application groups. Typically, a model switch framework
is customized to your environment through its exit programs.
MIMIX Switch Assistant - A guided user interface that guides you through switching
using your default MIMIX Model Switch Framework. MIMIX Switch Assistant is
accessed from the MIMIX Basic Main Menu and does not support application groups.

Best practices for maintaining your MIMIX environment


MIMIX is shipped with default settings that incorporate many best practices for
maintaining your environment. Others may require changing policies and adopting
best practices within your organization. Best practices include:
• Allow MIMIX to automatically correct differences detected during database and
object replication processes that would otherwise result in errors. If MIMIX is
unable to perform the recovery, the problem is reported as a replication error (a
file is placed in held error or an object is in error).
• Allow MIMIX to automatically perform audits and to automatically recover any
differences detected by audits. Best practice is to allow regularly scheduled audits
of all objects configured for replication and daily audits of prioritized categories of
replicated objects. User interfaces summarize audit results and indicate whether
MIMIX is unable to recover an object.
• Perform all audits with the audit level set at level 30 immediately prior to a planned
switch to the backup system and before switching back to the production system.
• Perform switches on a regular basis. Best practice is to switch every three to six
months. You need to set aside time for performing planned switches.
Environments that continue to use MIMIX Switch Assistant can use policies so
that compliance with regular switching is automatically reported in the user
interface.

Authority to products and commands


If your MIMIX environment takes advantage of the additional security available in the
product and command authority functions which Vision Solutions provides through
License Manager, you may need a higher authority level in order to perform MIMIX
daily operations.
A MIMIX administrator can change your authorization level to commands and
displays. Authorization levels typically fall into these categories:

24
Authority to products and commands

• Viewing information requires display (*DSP) authority.


• Controlling operations requires operator (*OPR) authority.
• Creating or changing configuration requires management (*MGT) authority.
For example, consider audits. You can view an audit if you have display authority,
perform audits if you have operator authority, and change policies that affect how
auditing is performed if you have management authority.
For more information about these provided security functions, see the Using License
Manager book.

25
Accessing the MIMIX Main Menu
The MIMIX command accesses the main menu for a MIMIX installation. The MIMIX
Main Menu has two assistance levels, basic and intermediate. The command defaults
to the basic assistance level, shown in Figure 1, with its options designed to simplify
day-to-day interaction with MIMIX. Figure 2 shows the intermediate assistance level.
The options on the menu vary with the assistance level. In either assistance level, the
available options also depend on the MIMIX products installed in the installation
library and their licensing. The products installed and the licensing also affect
subsequent menus and displays.
Accessing the menu - If you know the name of the MIMIX installation you want, you
can use the name to library-qualify the command, as follows:
Type the command library-name/MIMIX and press Enter. The default name of
the installation library is MIMIX.
If you do not know the name of the library, do the following:
1. Type the command LAKEVIEW/WRKPRD and press Enter.
2. Type a 9 (Display product menu) next to the product in the library you want on the
Vision Solutions Installed Products display and press Enter.
Changing the assistance level - The F21 key (Assistance level) on the main menu
toggles between basic and intermediate levels of the menu. You can also specify the
Assistance Level (ASTLVL) parameter on the MIMIX command.

Figure 1. MIMIX Basic Main Menu

MIMIX Basic Main Menu


System: SYSTEM1
MIMIX

Select one of the following:

1. Work with application groups WRKAG


2. Start MIMIX
3. End MIMIX
4. Switch all application groups
5. Start or complete switch using Switch Asst.
6. Work with data groups WRKDG

10. Availability status WRKMMXSTS


11. Configuration menu
12. Work with monitors WRKMON
13. Work with messages WRKMSGLOG
14. Cluster menu
More...
Selection or command
===>__________________________________________________________________________
______________________________________________________________________________
F3=Exit F4=Prompt F9=Retrieve F21=Assistance level F12=Cancel
(C) Copyright Vision Solutions, Inc., 1990, 2014.

26
Accessing the MIMIX Main Menu

Note: On the MIMIX Basic Main Menu, options 5 (Start or complete switch using
Switch Asst.) and 10 (Availability Status) are not recommended for
installations that use application groups.

Figure 2. MIMIX Intermediate Main Menu

MIMIX Intermediate Main Menu


System: SYSTEM1
MIMIX
Select one of the following:

1. Work with data groups WRKDG


2. Work with systems WRKSYS
3. Work with messages WRKMSGLOG
4. Work with monitors WRKMON
5. Work with application groups WRKAG
6. Work with audits WRKAUD
7. Work with procedures WRKPROC

11. Configuration menu


12. Compare, verify, and synchronize menu
13. Utilities menu
14. Cluster menu
More...
Selection or command
===>__________________________________________________________________________
______________________________________________________________________________
F3=Exit F4=Prompt F9=Retrieve F21=Assistance level F12=Cancel
(C) Copyright Vision Solutions, Inc., 1990, 2014.

27
MIMIX policies

CHAPTER 2 MIMIX policies

Each MIMIX policy is a mechanism used to enable, disable, or provide input to a


function such as replication, auditing, or MIMIX Model Switch Framework. A policy
may also determine how you are notified about certain problems that may occur.
For most policies, the initially shipped values apply to an installation. However,
policies can be changed and most can also be overridden for individual data groups.
When a policy is set for a data group, it takes precedence over the installation policy.
Policies must be changed from the management system. Changing policies requires
that you have management-level authority to the Set MIMIX Policy (SETMMXPCY)
command.
You can set policies from a command line or from the Work with Audits and the Work
with DG Definitions displays.
The topics in this chapter include:
• “Environment considerations for policies” on page 29 describes additional
considerations for setting policies for environments with more than two nodes or
bi-directional replication. Also, applications and features can conflict with policy-
controlled automatic recovery functions.
• “Setting policies - general” on page 31 provides basic procedures for changing
policies. Other topics in this chapter include more in-depth procedures for specific
policy-controlled functionality.
• “Policies which affect an installation” on page 33 identifies the policies that are set
for an installation and which cannot be overridden by a data group-level setting.
Also, this includes procedures for changing retention criteria for procedure history.
• “Policies which affect replication” on page 35 identifies the policies associated
with automatic error detection and correction during replication and identifies the
common object and file error situations that can be automatically recovered.
• “Policies which affect auditing behavior” on page 40 identifies policies that
influence the behavior of audits when they run, regardless of whether audits are
submitted manually or automatically by MIMIX.
• “Changing auditing policies” on page 42 provides additional information and
procedures for changing policies associated with auditing. This includes changing
the auditing level before switching, restricting auditing based on the state of data
groups, changing audit history retention, and disabling all auditing.
• “Policies for switching with model switch framework” on page 48 identify the
policies associated with model switch framework and includes instructions for
changing these policies.
• “Policy descriptions” on page 50 describes polices used by MIMIX.

28
Environment considerations for policies

Environment considerations for policies


Default settings for policies are chosen to address the needs of a broad set of
customer environments. However, in more complex environments, you need to
consider the effect of policies. Also, applications and other MIMIX features in some
environments can conflict with automatic recovery actions during replication and with
auditing.

Policies in environments with more than two nodes or bi-directional repli-


cation
Policy values may affect data throughout your entire environment, not just a single
installation or data group. This is of particular concern in environments that have more
than two systems (nodes) or which have replication occurring simultaneously in more
than one direction (bi-directional). Specifically, be aware of the following:
• In these environments, the value *DISABLED for the Objects only on target policy
is recommended. When the policy is disabled, audits will detect that objects exist
only on the target system but will not attempt to correct them. The commands
used by an audit are aware of all objects on the target system, not just those
which originate from the source system of the data group associated with the
audit. In these environments, the value *DELETE must be used with care. When
the policy value is Delete, audits will delete objects which may have originated
from systems not associated with the data group being audited.
• Synchronization of user profiles and authorization lists associated with an object
will occur unless the user profiles and authorization lists are explicitly excluded
from the data group configuration. In the environments mentioned, this may result
in user profiles and authorization lists being synchronized to other systems in your
configuration. This behavior occurs whenever any of the automatic recovery
policies are enabled (database, object, audit). To prevent this from occurring, you
must explicitly exclude the user profiles and authorization lists from replication for
any data group for which you do not want them synchronized.
• In a simultaneously bi-directional environment, determine which system ‘wins’ in
the event of a data conflict, that is, which system will be considered as having the
correct data. Choose one direction of replication that will be audited and allow
auditing for those data groups. Disable audits for data groups that replicate in the
opposite direction. For example, data groups AB and BA are configured for bi-
directional replication between system A and system B. Data group AB replicates
from system A to system B and data group BA replicates the opposite direction.
System B is also the management system for this installation. You chose system
A as the winning system and want to permit auditing in the direction from A to B.
The Audit level policy for data group AB must be set to a level that permits audits
to run (level 10 or higher). The Audit level policy for data group BA must be set to
disable audits. The results of audits of data group AB will be available on system
B, because system B is the management system and default policy values cause
rules to be run from the management system.
• In environments with three or more systems in the same installation, you need to
evaluate each pair of systems. For each pair of systems, evaluate the directions in

29
which replication is permitted. If any pair of systems supports simultaneous bi-
directional replication, determine the winning system in each pair and determine
the direction to be audited. Set the audit level policy to permit auditing for the data
group that replicates in the chosen direction. Disable auditing for the data group
which replicates in the other direction. You may also want to consider changing
the values of the Run rule on system policy for the installation or the audited data
groups to balance processing loads associated with auditing.
• In environments that permit multiple management systems in the same
installation, in addition to evaluating the direction of replication permitted within
each pair of systems, you must also consider whether the systems defined by
each data group are both management systems. If any pair of systems supports
simultaneous bi-directional replication, choose the winning system and change
the Audit level policies for each data group so that only one direction is audited.
You may need to change the Run rule on system policy to prevent certain data
groups from being audited from specific management systems.

Policies for data protection reports


By default, data protection reports automatically run on all systems in an installation.
The data protection report for directories uses the directory depth policy to determine
how many levels of sub-directories to check on a system when it runs. The value of
this policy may influence how often or when you want to run the directory report.
Changing the directory depth policy to include more directories for a report may cause
the report to run longer. Consider the needs of your environment and system
resources available when the report runs. For more information, see “Changing when
automatic reports are allowed to run” on page 72.

When to disable automatic recovery for replication and auditing


At times, you may need to disable automatic recoveries during replication and
auditing for certain data groups because a feature in use or an application being
replicated may interact with auditing in an undesirable way.
Features - Do not use automatic recoveries during auditing and replication in any
data group that is using functions provided by the MIMIX CDP™ feature. This feature,
which requires an additional license key, permits you to perform operations
associated with maintaining continuous data protection. By configuring a recovery
window for a data group, you introduce an automatic delay into when the apply
processes complete replication. By setting a recovery point for a data group, you
identify a point that, when reached, will cause the apply processes to be suspended.
In both cases, source system changes have been transferred to the target system but
have not been applied. In such an environment, comparisons will report differences
and automatic recoveries will attempt recovery for items that have not completed
replication. To prevent this from occurring, disable comparisons and automatic
recoveries for any data group which uses the MIMIX CDP feature. For details, see
“Disabling audits and recovery when using the MIMIX CDP feature” on page 31.
Applications - At times, data groups for some applications will encounter problems if
the application cannot acquire locks on objects that are defined to MIMIX. These data
groups may need to be excluded from auditing. MIMIX acquires locks occasionally to

30
Setting policies - general

save and restore objects within the replication environment. Some applications may
fail when they cannot acquire a lock on an object. Refer to our Support Central for
FAQs that list specific applications whose data groups should be excluded from
auditing. For those excluded data groups, you can still run compares to determine if
objects are not synchronized between source and target systems. Care must be
taken to recover from these unsynchronized conditions. The applications may need to
be ended prior to manually synchronizing the objects.
To exclude a data group from audits, use the instructions in “Preventing audits from
running” on page 46.

Disabling audits and recovery when using the MIMIX CDP feature
The functions provided by the MIMIX CDP™ feature1 create an environment in which
source system changes have been transferred to the target system but have not been
applied. Any data group which uses this feature must disable automatic comparisons
and automatic recovery actions for the data group.
Do the following from the management system:
1. From the command line type SETMMXPCY and press F4 (Prompt).
2. For the Data group definition, specify the full three-part name of the data group
that uses the MIMIX CDP feature.
3. Press Enter to see all the policies and their current values.
4. For Automatic object recovery, specify *DISABLED.
5. For Automatic database recovery, specify *DISABLED.
6. For Automatic audit recovery, specify *DISABLED.
7. For Audit level, select *DISABLED.
8. To accept the changes, press Enter.

Setting policies - general


Policies must be changed from the management system. Changing policies requires
that you have management-level authority to the Set MIMIX Policy (SETMMXPCY)
command.
The following procedures describe the basic procedures for setting policies.

Changing policies for an installation


This procedure changes a policy value at the installation level. The installation level
value will overridden if a data group level policy has been specified with a value other
than *INST.

1. The MIMIX CDP™ feature requires an additional license key.

31
Do the following from the management system:
1. From the command line type SETMMXPCY and press F4 (Prompt).
2. Verify that the value specified for Data group definition is *INST.
3. Press Enter to see all the policies and their current values.
4. Specify a value for the policy you want. Use F1 (Help) to view descriptions of
possible values.
5. To accept the changes, press Enter.

Changing policies for a data group


Do the following from the management system:
1. From the command line type SETMMXPCY and press F4 (Prompt).
2. For the Data group definition, specify the full three-part name.
3. Press Enter to see all the policies and their current values.
4. Specify a value for the policy you want defined for the data group. Use F1 (Help)
to view descriptions of possible values.
5. To accept the changes, press Enter.

Resetting a data group-level policy to use the installation level value


Do the following from the management system:
1. From the command line type SETMMXPCY and press F4 (Prompt).
2. For the Data group definition, specify the full three-part name.
3. Press Enter to see all the policies and their current values.
4. For the policy you want to reset, specify *INST.
5. To accept the changes, press Enter.

32
Policies which affect an installation

Policies which affect an installation


While many policies can be set for an installation, the policies in Table 1 cannot be
overridden for an individual data group. At the data group level, these policies always
have a value of *INST.

Table 1. Policies that can be set only at the installation level and shipped default values.

Policy Shipped Values – Installation1

Directory depth2 3

Procedure history retention


• Minimum days 7
• Minimum runs per procedure 1
• Min. runs per switch procedure 1

Independent ASP library ratio 5

ASP threshold for recoveries3 95


1. Default values for the installation level are assigned to polices at the time the installation is cre-
ated.
2. This policy only applies to data protection reports for IFS directories.
3. This policy is available on instances running MIMIX 8.1.23.00 or higher.

Changing directory depth for data protection reports


The directory depth policy determines how many levels of sub-directories to include
when data is gathered for a node by a data protection report. Data protection reports
include information about which objects and their contents are configured for
replication by MIMIX.
The directory depth policy is used by data protection report procedure for IFS
directory objects (CRTDPRDIR). By default, this policy is set to a value of 3.
To change the directory depth policy for the installation, do the following:
1. From the command line type SETMMXPCY and press F4 (Prompt).
2. Verify that the value *INST is specified for the Data group definition prompt:
3. Press Enter to see all the policies and their current values.
4. Locate the Directory depth for report policy. Specify the value you want.
5. To accept the changes, press Enter.

Changing retention criteria for procedure history


The procedure history retention policy determines how long to retain historical
information about procedure runs that completed, completed with errors, or that failed
or were canceled and have been acknowledged.
Environments configured with application groups use procedures to control
operations such as starting, ending, or switching. History information for a procedure
includes timestamps indicating when the procedure was run and detailed information

33
about each step within the procedure. The policy specifies how many days to keep
history information and the minimum number of runs to keep. You can specify a
different number of runs to keep for switch procedure runs than what is kept for other
types of procedures.
Each procedure run is evaluated individually against the policy and its history
information is retained until the specified minimum days and minimum runs are both
met. When a procedure run exceeds these criteria, system cleanup jobs will remove
the historical information for that procedure run from all systems. The values specified
at the time the cleanup jobs run are used for evaluation.
To change the procedure history retention policy for the installation, do the following:
1. From the command line type SETMMXPCY and press F4 (Prompt).
2. Verify that the value *INST is specified for the Data group definition prompt:
3. Press Enter to see all the policies and their current values.
4. Locate the Procedure history retention policy. The current values are displayed.
Specify values for the elements you want to change.
5. To accept the changes, press Enter.

Changing the system ASP percent used threshold for recoveries


To limit MIMIX from contributing to the percentage of storage used on the source
system when storage in the system auxiliary storage pool (ASP 1) is already
constrained, MIMIX automatically delays and resumes processing recoveries for
member data of *FILE objects based on the value of the ASP threshold for recoveries
(RCYASPTHLD) 1. MIMIX delays processing for this type of recovery when the
percent of storage used by ASP 1 on the source system exceeds the policy value and
automatically resumes processing these recoveries after the storage used is less than
the policy value.
You can change the value of the threshold policy for the installation, do the following.
1. From the command line type SETMMXPCY and press F4 (Prompt).
2. Verify that the value *INST is specified for the Data group definition prompt:
3. Press Enter to see all the policies and their current values.
4. Locate the ASP threshold for recoveries policy. The current values are displayed.
Specify the percentage of storage used by ASP1 at which MIMIX will delay
member data recoveries.
5. To accept the changes, press Enter.

1. The ASP threshold for recoveries (RCYASPTHLD) policy is available on instances running
MIMIX service pack 8.1.23.00 or higher.

34
Policies which affect replication

Policies which affect replication


Table 2 identifies the policies which can affect replication and their shipped default
values Click on a policy name to access its description.

Table 2. Policies associated with replication and shipped default values.

Policy Shipped Values at Creation1 Replication Processes

Installation Data Groups System User


Journal Journal

Data group definition2 *INST Name3 Yes Yes

Automatic object (system *ENABLED *INST3 Yes4 –


journal) recovery

Automatic database (user *ENABLED *INST – Yes4


journal) recovery

Object recovery notify on *YES *INST Yes –


success

DB apply cache *ENABLED *INST – Yes

Access path maintenance – Yes


• Optimize for DB apply *DISABLED *INST
• Maximum number of jobs 99 *INST

Synchronize threshold size 9,999,999 *INST Yes Yes

Number of third delay retry 100 *INST Yes –


attempts

Third delay retry interval 15 *INST Yes –

Source capture delay Yes Yes


• Replication recovery
processing *NONE *INST
• Virtual switch testing 5 *INST
1. Default values for the installation level are assigned to polices at the time the installation is created.
Default values for data group level are assigned when the data group is created.
2. A value must be specified for the data group definition (DGDFN) parameter on the SETMMXPCY
command in order to specify any policy value.
3. A data group definition value of *INST indicates the policy is installation-wide. A name indicates the
policies are in effect only for the specified data group.
4. When this policy is enabled, the other policies in the same column are in effect unless otherwise
noted.

For more information about the Automatic object recovery and Automatic database
recovery polices, see “Replication problems that MIMIX attempts to recover” on
page 260 and “When to disable automatic recovery for replication and auditing” on
page 30.

35
For information about how the Source capture delay policy is used by recoveries and
considerations for changing the policy, see “Changing the source capture delay for
recoveries” on page 36.
These topics in the MIMIX Administrator Reference book provide information about
the following policies:
• For information about the DB apply cache policy, see topic “Configuring database
apply caching”.
• For information about the Access path maintenance policy, see topic “Optimizing
access path maintenance”.
• For information about delay retry processing in object replication and the Number
of third delay retry attempts and Third delay retry interval policies, see Parameters
for automatic retry processing in topic “Tips for data group parameters”.

Changing the source capture delay for recoveries


A replication recovery is additional processing, initiated as needed by database or
object replication processes, that automatically attempts to correct a detected
replication error without placing replication activity for the affected object or file on
hold. A target journal inspection recovery attempts to correct a problem with a
replicated object on the target node that was modified by users or programs other
than MIMIX.
Some recovery actions initiated by replication processes and target journal inspection
may require synchronizing objects or data between the source and target nodes. The
recovery action gathers the needed information from the source node and submits a
journal entry to correct the detected error or changed object. If synchronization is
necessary, replication manager processes use the Source capture delay
(SRCCAPDLY) policy to determine how long to wait between detecting the need for a
replication recovery and capturing the affected file records, attributes, authorities, or
the entire object from the source node.
The policy has two delay settings, one for recoveries during normal replication, and
one for recoveries during the application testing phase of a virtual switch. (Both
settings are ignored during the recovery phase of a virtual switch). You can change
either setting if needed.
When a delay is configured for either setting, the delay starts with the first error
detected for the object or record being replicated. During the delay time, the recovery
action is not processed and subsequent recovery actions for the same type of
problem for the same object or file record range are not tracked for the duration of the
delay. When the delay time is reached, a journal entry requesting the recovery
synchronization is placed into the journal stream and enough information for the
tracked object or record range is captured from the source node and sent to the target
so that the tracked recovery action and any additional recovery actions requested
during the delay can be processed.

36
Changing the source capture delay for recoveries

Changing the source capture delay for recoveries during replication


Recovery actions initiated by replication processes or target journal inspection default
to having no delay before capturing information needed from the source node. This is
safest, but can use more communications and CPU resources. You can optionally
specify the amount of time to delay, in minutes.
A delay may reduce the number of captures and resulting communications usage.
Recoveries have a status of New for the duration of the delay. However, a delay also
creates risk because your environment may not be switch-ready. Objects and records
associated with new, in-progress, or failed recovery actions are not synchronized
between the source and target nodes, thereby affecting switch readiness.

To change the source capture delay used during replication, do the following:
Note: This procedure changes a policy value at the installation level. The installation
level value can be overridden by a data group level policy value. Therefore, if
a data group has a value other than *INST for this policy, that value remains in
effect.

1. From the command line type SETMMXPCY and press F4 (Prompt).


2. Verify that the value specified for Data group definition is *INST.
3. Press Enter to see all the policies and their current values.
4. For the Repl. recovery processing prompt under Source capture delay (min.),
specify one of the following:
• Specify the number of minutes to delay source captures.
• Specify *NONE to have recovery captures performed immediately.
5. To accept the changes, press Enter.
Changes become effective automatically as the replication manager checks for
recoveries to process.

Changing the source capture delay used during virtual switch testing
During the application testing phase of a virtual switch, apply processes are ended
and replicated transactions wait to be applied. Recovery actions initiated by
replication processes and target journal inspection during virtual switch testing use a
configured delay time with a default value of 5 minutes. Optionally, you can change
the delay time.
The source capture delay value needs to be an acceptable balance between
increased risk and resource usage for the amount of time that you expect to remain in
the application testing phase of a virtual switch. During virtual switch testing, your
environment is already exposed because transactions accumulate on the target node
and are not applied until after the testing phase of the switch completes. The volume
of transactions, recovery actions, and how frequently recovery actions capture source
information can affect apply process backlogs and resource usage. All transactions

37
that have accumulated for apply processes during testing, including those for new
recoveries, are applied during the recovery phase of a virtual switch. The virtual
switch recovery phase must complete before a planned or unplanned switch of the
application group can be performed.
During virtual switch testing:
• No delay allows the information needed for recoveries to be immediately captured
and sent to the target. This represents the least risk in the event of the loss of the
source node. However, frequent access of source objects may interfere with
applications and increase communication usage, and the increased transactions
may quickly increase apply process backlogs.
• A delay time may reduce the number of captures needed for an object or file,
reduce communications usage, and place fewer recovery-related transactions into
the queues for apply processing. However, the objects and records identified by
new recoveries created during virtual switch testing are not synchronized, and
their source information will not be sent to the target node until the delay time is
reached.
• The value Delay Until Recovery (*VRTSWTRCY) delays all source captures for
target journal inspection recovery actions created during the application testing
phase of a virtual switch until the virtual switch recovery phase starts. If your
environment has workload or storage constraints, this may be more efficient. The
source information needed to synchronize affected target node files and objects is
not captured and sent to the target node at or near the time the errors occurred,
which is an exposure in the event of the loss of the source node. Also, the virtual
switch recovery phase may be slower because it includes all delayed capture
operations as well as applying transactions that accumulated while testing.

To change the source capture delay used during the application testing phase of a
virtual switch, do the following:
Note: This procedure changes a policy value at the installation level. The installation
level value can be overridden by a data group level policy value. Therefore, if
a data group has a value other than *INST for this policy, that value remains in
effect.

1. From the command line type SETMMXPCY and press F4 (Prompt).


2. Verify that the value specified for Data group definition is *INST.
3. Press Enter to see all the policies and their current values.
4. For the Virtual switch testing prompt under Source capture delay (min.), specify
one of the following:
• Specify the number of minutes to delay source captures.
• Specify *NONE to have recovery captures performed immediately.
• Specify *VRTSWTRCY to delay the source capture until the application testing
phase of the virtual switch is complete and virtual switch recovery starts.
5. To accept the changes, press Enter.

38
Changing the source capture delay for recoveries

Changes become effective automatically as the replication manager checks for


recoveries to process.

39
Policies which affect auditing behavior
The policies identified in Table 3 affect the behavior of all audits in an installation
regardless of whether the audit are automatically submitted or manually invoked.
These policies can be set for the installation as well as overridden for a data group.
When set for a specific data group, these policies affect all audits for the data group.
The shipped initial values for both levels are indicated.
When the Set MIMIX Policies (SETMMXPCY) command specifies a data group
definition value of *INST, the policies being changed are effective for all data groups in
the installation, unless a data group-level override exists. When the data group
definition specifies a name, policies which specify the value *INST inherit their value
from the installation-level policy value and polices which specify other values are in
effect for only the specified data group.

Table 3. Policies that affect behavior of audits and their shipped default values

Policy Shipped Values at Creation1

Installation Data Groups

Data group definition2 *INST Name

Automatic audit recovery *ENABLED *INST

Audit notify on success *RULE *INST

Notification severity *RULE *INST

Object only on target *DISABLED *INST

Journal attribute differences action


• MIMIX configured higher *CHGOBJ *INST
• MIMIX configured lower *NOCHG *INST

DB apply threshold action *END *INST

Maximum rule runtime 1440 *INST

Audit warning threshold3 7 *INST

Audit action threshold3 14 *INST

Audits included for compliance *BOTH *INST

Audit level *LEVEL30 *INST

Run rule on system *MGT *INST

Action for running audits


• Inactive data group *NOTRUN4 *INST
• Repl. process in threshold *NOTRUN *INST

Audit Severity
• Not run audit status *WARNING *INST
• Ignored objects audit status *WARNING *INST

40
Policies which affect auditing behavior

Table 3. Policies that affect behavior of audits and their shipped default values

Policy Shipped Values at Creation1

Installation Data Groups

Audit history retention


• Minimum days 7 *INST
• Minimum runs per audit 1 *INST
• Object details *YES *INST
• DLO and IFS details *YES *INST

Synchronize threshold size 9,999,999 *INST

CMPRCDCNT commit threshold *NOMAX *INST


1. Default values for the installation level are assigned to polices at the time the installation is created.
Default values for data group level are assigned when the data group is created.
2. A value must be specified for the data group definition (DGDFN) parameter on the SETMMXPCY
command in order to specify any policy value.
3. These policies are not limited to recovery actions.
4. Prior to version 7.1.12.00, the shipped default value was *NONE.

Several policies affect when audits automatically run and how those audits select
objects. However, those policies are no longer preferred for setting audit scheduling
and object selection criteria. Instead, use the Set MIMIX Schedules (SETMMXSCD)
command as described in “Scheduling audits and reports” on page 60.

41
Changing auditing policies
This topic describes how to change specific policies that affect auditing behavior and
when automatic audits will run. MIMIX service providers are specifically trained to
provide a robust audit solution that meets your needs.

Changing the audit level policy when switching


Regardless of the level you use for daily operations, Vision Solutions strongly
recommends that you perform audits at audit level 30 before the following events to
ensure that 100 percent of the data is valid on the target system:
• Before performing a planned switch to the backup system.
• Before switching back to the production system.
For more information about the risks associated with lower audit levels, see
“Considerations for rules” on page 699.
From the native user interface, do the following from the management system:
1. From the command line type SETMMXPCY and press F4 (Prompt).
2. Verify that the value specified for Data group definition is *INST.
3. Press Enter to see all the policies and their current values.
4. For Audit level, specify *LEVEL30. Then press Enter.

Changing the system where audits are performed


The Run rule on system policy determines the system on which audits run. The
shipped default is to run all audits for the installation from the management system.
When changing the value of this policy, also consider your switching needs. Click this
link to see additional information about the Run rule on system policy.
Note: This procedure changes a policy value at the installation level. The installation
level value can be overridden by a data group level policy value. Therefore, if
a data group has a value other than *INST for this policy, that value remains in
effect.

To change the policy for the installation, do the following


1. On the management system type the following command and press F4 (Prompt)
installation-library/SETMMXPCY
2. Verify that the value *INST appears for the Data group definition.
3. Locate the Run rule on system policy. Specify the value you want.
4. Press Enter.

42
Changing auditing policies

Changing retention criteria for audit history


The Audit history retention policy determines whether to retain information about the
results of completed audits and the objects that were audited. The policy specifies
how many days to keep history information and how many audit runs to keep, as well
as whether details about audited library-based objects and audited DLO and IFS
objects are to be kept with the history information. Each audit is evaluated individually
against the policy values.
The policy is checked when an audit runs to determine whether to keep details about
the objects audited by that run. The policy is also checked when system manager
cleanup jobs run to determine if any audit has history information which exceeds both
specified retention criteria. The policy value in effect at the time each check occurs
determines the result.
To change the audit history retention policy, do the following:
1. From the MIMIX Intermediate Main Menu, select option 6 (Work with Audits) and
press Enter.
2. Determine whether to change the policy for the installation or at the data group
level. From the Work with Audits display, do one of the following:
• To change the policy for all audits in the installation, press F16 (Inst. policies).
Then, press Enter when the Set MIMIX Policies (SETMMXPCY) command
appears.
• To change the policy for all audits for a specific data group, type 36 (Change
DG policies) next to any audit for the data group you want and press Enter.
3. Locate the Audit history retention policy. The current values for the level you
chose in Step 2 are displayed. Specify values for the elements you want to
change.
Note: When large quantities of objects are eligible for replication, specifying
*YES to retain either Object details or DLO and IFS details may use a
significant amount of disk storage. Consider the combined effect of the
quantity of replicated objects for each data group, the number of days to
retain history, the number of audits to retain, and the frequency in which
audits are performed.
4. To accept the changes, press Enter.

Restricting auditing based on the state of the data group


You may want to control when audits are allowed to run based on the state of the data
group at the time of the audit request. For example, if you end MIMIX so that a batch
process can run, you may want to prevent audits from running while data groups are
inactive. If a data group process has a backlog during peak activity, you may want to
prevent audits from running while the backlog exists. Or, you may want to prevent
only automatic recovery from occurring during a backlog or when the data group is
inactive. The Action for running audits policy provides the ability to define what audit
activity will be permitted based on the state of the data group at the time of audit
request. This policy can be set for an installation or for a specific data group.

43
Note: Most audits check for threshold conditions in all database and object
replication processes, including the RJ link. #FILDTA audits only check for
threshold warning conditions in the RJ link and database replication
processes. #DLOATR audits only check for threshold warning conditions in
object replication processes.
Restricting audit activity in an installation based on data group state: Do the
following from the management system:
Note: This procedure changes a policy value at the installation level. The installation
level value can be overridden by a data group level policy value. Therefore, if
a data group has a value other than *INST for this policy, that value remains in
effect.

1. From the command line type SETMMXPCY and press F4 (Prompt).


2. Verify that the value specified for Data group definition is *INST.
3. Press Enter to see all the policies and their current values.
4. For Action for running audits, do the following:
a. Specify the value you want for Inactive data group that indicates the audit
actions to permit when the data group is inactive
b. Specify the value you want for Repl. process in threshold that indicates the
audit actions to permit when any replication process checked by an audit has
reached its configured threshold.
5. To accept the changes, press Enter.
Restricting audit activity for a specific data group based on its state: Do the
following from the management system:
1. From the command line type SETMMXPCY and press F4 (Prompt).
2. For the Data group definition, specify the full three-part name.
3. Press Enter to see all the policies and their current values.
4. For Action for running audits, do the following:
a. Specify the value you want for Inactive data group that indicates the audit
actions to permit when the data group is inactive
b. Specify the value you want for Repl. process in threshold that indicates the
audit actions to permit when any replication process checked by an audit has
reached its configured threshold.
5. To accept the changes, press Enter.

Changing the severity reported for audit statuses


You can control the severity associated with the audit status values *NOTRUN and
*IGNOBJ. The severity determines whether these status values are considered
errors, warnings, or informational when auditing status is rolled up to other status
values within a MIMIX instance. The severity associated with *NOTRUN and

44
Changing auditing policies

*IGNOBJ statuses affects the order in which audits are listed on the Work with Audits
display (WRKAUD command) and in the Audits portlet in Vision Solutions Portal
(VSP), and the icon associated with these statuses in the Audits portlet.
All audit statuses have an associated severity. Only the severity associated with these
statuses can be changed with the Audit severity (AUDSEV) policy:
• *NOTRUN - indicates that an audit was submitted but did not run due to the status
of replication processes or policy settings.
• *IGNOBJ - indicates than an audit could not check some objects, typically due to
replication activity, locks, or authorizations. All other objects checked were
considered equal or had detected differences that were automatically recovered.
The policy setting for the severity of the *IGNOBJ status also affects whether any
ignored objects are counted as differences. The policy’s effect on how ignored objects
are counted occurs regardless of whether the audit status is *IGNOBJ or a more
severe audit status.
Note: This procedure changes a policy value at the installation level. The installation
level value can be overridden by a data group level policy value. Therefore, if
a data group has a value other than *INST for this policy, that value remains in
effect.

1. From the command line type SETMMXPCY and press F4 (Prompt).


2. Verify that the value specified for Data group definition is *INST.
3. Press Enter to see all the policies and their current values.
4. For Audit severity, do the following, as needed:
• At the Not run audit status prompt, specify the severity value (*ERROR,
*WARNING, or *INFO) to associate with the audit status *NOTRUN.
• At the Ignored objects audit status prompt, specify the severity value
(*WARNING or *INFO) to associate with the audit status *IGNOBJ.
5. To accept the changes, press Enter.

Changing what audits are checked for compliance


You can determine what audits will be included in checks for compliance with the
Audits included for compliance (AUDINCMPL) policy1. For each audit that you have
checked for compliance, MIMIX evaluates whether the last run of the audit exceeds
the days specified for the compliance warning threshold and audit action threshold
policies. When an audit is approaching or is out-of-compliance, a compliance status
value is reflected in the overall audit status and is included when auditing status is
rolled up to other areas of a MIMIX instance.
The default behavior uses the value *BOTH, which means that both the Audit level
policy and existing configuration entries are considered when determining what audits

1. This policy is available for instances running MIMIX service pack 8.0.07.00 or higher. For
instances running earlier software levels, all audits are checked for compliance.

45
are evaluated for compliance. This prevents the potential for compliance errors for
audits that have no configured objects to audit and for the File Member Record Count
(#MBRRCDCNT) audit when the audit level policy value is *LEVEL30. For a detailed
description of this value and other possible values, press F1 (Help) for the Audits
included for compliance (AUDINCMPL) policy in the Set MIMIX Policies
(SETMMXPCY) command.
Note: This procedure changes a policy value at the installation level. The installation
level value can be overridden by a data group level policy value. Therefore, if
a data group has a value other than *INST for this policy, that value remains in
effect.

1. From the command line type SETMMXPCY and press F4 (Prompt).


2. Verify that the value specified for Data group definition is *INST.
3. Press Enter to see all the policies and their current values.
4. For Audits included for compliance, either specify a value (*BOTH, *ALL,
*AUDLVL, or *CONFIG) or use the “+ for more” field as needed to specify one or
more audit names.
5. To accept the changes, press Enter.

Preventing audits from running


There may be scenarios when you need to disable auditing completely for either an
installation or a specific data group. Auditing may not be desirable on a test data
group or during system or network maintenance.
The Audit level policy can be used to disable all auditing, including manually invoked
audits.The Audit level can be set for an installation or for specific data groups. Note
that an explicitly set value for a data group will override the installation value and may
still allow an audit to run.
You can also prevent audits from being submitted automatically for a data group while
still allowing them to be invoked manually. Automatic submission of scheduled object
audits, prioritized audits, or both, can be prevented by changing the state specified in
the audit’s schedule. For more information, see “Disabling or enabling automatically
submitted audits” on page 70.
In addition to auditing, automatic recovery during replication may need to be
prevented from running due to issues with applications or MIMIX features, For more
information, see “When to disable automatic recovery for replication and auditing” on
page 30.

Disabling all auditing for an installation


Note: This procedure changes a policy value at the installation level. The installation
level value can be overridden by a data group level policy value. Therefore, if
a data group has a value other than *INST for this policy, that value remains in
effect.

46
Changing auditing policies

Do the following from the management system:


1. From the command line type SETMMXPCY and press F4 (Prompt).
2. Verify that the value specified for Data group definition is *INST.
3. Press Enter to see all the policies and their current values.
4. Specify *DISABLED for the Audit level policy.
5. To accept the changes, press Enter.

Disabling all auditing for a data group


Do the following from the management system:
1. From the command line type SETMMXPCY and press F4 (Prompt).
2. For the Data group definition, specify the full three-part name.
3. Press Enter to see all the policies and their current values.
4. Specify *DISABLED for the Audit level policy.
5. To accept the changes, press Enter.

47
Policies for switching with model switch framework
In environments that do not use application groups, MIMIX Switch Assistant (which
implements MIMIX Model Switch Framework) is usually used for switching1. MIMIX
Model Switch Framework cannot be used to switch application groups.
Table 4 identifies the policies associated with switching using MIMIX Model Switch
Framework and the shipped default values of those policies.
For these policies, MIMIX Switch Assistant uses only the policy values specified for
the installation. If MIMIX cannot determine whether a MIMIX Model Switch
Framework is defined, the switch framework policy is *DISABLED.
If the SETMMXPCY command specifies a data group name, the switch framework is
required to be *INST. The switch thresholds are *DISABLED by default but can be
changed.
The policies in Table 4 have no effect on application group switching.

Table 4. Shipped values of policies used by MIMIX Switch Assistant.

Policy Shipped Values at Creation1

Installation Data Groups

Data group definition2 *INST Name3

Switch warning threshold 90 *DISABLED

Switch action threshold 180 *DISABLED

Default model switch framework MXMSFDFT *INST


1. Default values for the installation level are assigned to polices at the time the installation is created.
Default values for data group level are assigned when the data group is created.
2. A value must be specified for the data group definition (DGDFN) parameter on the SETMMXPCY
command in order to specify any policy value.
3. A data group definition value of *INST indicates the policy is installation-wide. A name indicates the
policies are in effect only for the specified data group.

Specifying a default switch framework in policies


MIMIX Switch Assistant requires that you have a configured MIMIX Model Switch
Framework and that you specify it in the default model switch framework policy for the
installation. You may also want to adjust policies for thresholds associated with MIMIX
Switch Assistant.
If you do not have a configured MIMIX Model Switch Framework, contact your
Certified MIMIX Consultant.
From the native user interface, do the following from the management system:
1. From the command line type SETMMXPCY and press F4 (Prompt).
2. Verify that the value specified for Data group definition is *INST.

1. Switching is not supported in environments licensed for MIMIX DR.

48
Policies for switching with model switch framework

3. Press Enter to see all the policies and their current values.
4. At the Default model switch framework prompt, specify the name of the switch
framework to use for switching this installation.
5. To accept the changes, press Enter.

Setting polices for MIMIX Switch Assistant


If the value of the installation-level policy is disabled, you must change the policy in
order to use MIMIX Switch Assistant.
Do the following from the management system:
1. From the command line type SETMMXPCY and press F4 (Prompt).
2. Verify that the value specified for Data group definition is *INST.
3. Press Enter to see all the policies and their current values.
4. Specify values for the following fields:
a. For Switch warning threshold, the value 90 is recommended.
b. For Switch action threshold, the value 180 is recommended.
c. For Default model switch framework, specify the name of your MIMIX Model
Switch Framework.
5. To accept the changes, press Enter.

Setting policies when MIMIX Model Switch Framework is not used


If you do not use MIMIX Model Switch Framework for switching, you disable the
default model switch framework policy at the installation level.
Do the following from the management system:
1. From the command line type SETMMXPCY and press F4 (Prompt).
2. Verify that the value specified for Data group definition is *INST.
3. Press Enter to see all the policies and their current values.
4. At the Default model switch framework prompt, specify *DISABLED.
5. To accept the change, press Enter.

49
Policy descriptions
For a complete description of all policy values, see online help for the Set MIMIX
Policies (SETMMXPCY) command. The Data group definition and Audit rule
parameters on the command are not technically policies, but they must have values
specified in order to set a value for any policy.
Data group definition - Select the scope of the policies to be set. When the value
*INST is specified, the policies being set by the command apply to all systems and
data groups in the installation, with the exception of any policy for which a data group-
level override exists. When a three-part qualified name of a data group is specified,
the policies being set by the command apply to only that data group and override the
installation-level policy values.
Audit rule - The audit rule must specify a value of *NONE when setting a value for a
policy.
Note: The Audit rule, Audit schedule, and Priority audit policies are no longer
preferred for changing scheduling criteria for automatically submitted audits.
For more information about the recommended way to change audit
scheduling, see “Scheduling audits and reports” on page 60.
Automatic object recovery — Determines whether to enable functions that
automatically attempt to correct common object replication errors that may occur
during replication from the system journal and changes reported by target journal
inspection. When this policy is enabled, a recovery action is automatically started
when an error is detected. Recovery actions that require synchronizing information
from the source node to the target node use the time specified in Source capture
delay (SRCCAPDLY) policy.
Automatic database recovery — Determines whether to enable functions that
automatically attempt to correct common file errors that may occur during replication
from the user journal and changes reported by target journal inspection. When this
policy is enabled and an error is detected, MIMIX will generate a recovery and attempt
to replicate the transaction again. Recovery actions that require synchronizing
information from the source node to the target node use the time specified in Source
capture delay (SRCCAPDLY) policy
Automatic audit recovery — Determines whether to enable automatic recovery
actions for audits. When enabled, audits that support automatic recovery will check
their comparison results for known problems and will attempt to correct any detected
errors by automatically starting recovery actions.
Note: An audit that is currently running is not affected by a change to this policy until
the next run of the audit.
Object recovery notify on success — Determines whether object replication
processes that successfully process an object while using the configured third
delay/retry cycle will send an informational (*INFO) notification. This policy is only
valid when the Automatic object recovery policy is enabled.
Audit notify on success — Determines whether activity initiated by audits,
including recovery actions, will automatically send an informational (*INFO)
notification upon successful completion. If an audit is run when the Automatic audit

50
Policy descriptions

recovery policy is disabled, successful notifications are sent only for the compare
phase of the audit.
Notification severity — Determines the severity level of the notifications sent when
a rule ends in error. This policy determines the severity of the notification that is sent,
not the severity of the error itself. The policy is in effect whether the rule is invoked
manually or automatically.
This policy is useful for setting up an order precedence for notifications at the data
group level. For example, if you set this policy for data group CRITICAL to be
*ERROR when the value for the installation-level policy is *WARNING, any error
notifications sent from data group CRITICAL will have a higher severity than those
from other data groups.
Object only on target — Determines how recovery actions should handle correcting
objects configured for replication that exist only on the target node. The only-on-target
error may have been detected by replication processes, target journal inspection, or
audits (#OBJATR, #IFSATR, #DLOATR, #FILATR, and #FILATRMBR). This policy is
used by the resulting recovery actions when the Automatic recovery policies
(Database, Object, and Audit) are enabled and a virtual switch is not in progress.
See “Policies in environments with more than two nodes or bi-directional replication”
on page 29 for additional information.
Journaling attribute difference action — Determines the recovery action to take for
scenarios in which audits have detected differences between the actual and
configured values of journaling attributes for objects journaled to a user journal. This
type of difference can occur for the Journal Images attribute and the Journal Omit
Open/Close attribute. Differences found on either the source or target object are
affected by this policy.
MIMIX configured higher
Determines the recovery for correcting a difference in which the MIMIX
configuration specifies an attribute value that results in a higher number of journal
transactions than the object's journaling attribute.
MIMIX configured lower
Determines the recovery action for correcting a difference in which the MIMIX
configuration specifies an attribute value that results in a lower number of journal
transactions than the object's journaling attribute.
DB apply threshold action — Determines what action to pass to the Compare File
Data (CMPFILDTA) command or the Compare Record Count (CMPRCDCNT)
command when it is invoked with *DFT specified for its DB apply threshold
(DBAPYTHLD) parameter. The command’s parameter determines what to do if the
database apply session backlog exceeds the threshold warning value configured for
the database apply process. This policy applies whenever these commands are used
and the backlog exceeds the threshold.
The shipped default for this policy causes the requested command to end and may
cause the loss of repairs in progress or inaccurate counts for members. You can also
set this policy to allow the request to continue despite the exceeded threshold.
DB apply cache — Determines whether to use database (DB) apply cache to
improve performance for database apply processes. When this policy is enabled,

51
MIMIX uses buffering technology within database apply processes in data groups that
specify *YES for journal on target (JRNTGT). This policy is not used by data groups
which specify JRNTGT(*NO) or by data groups whose target journals use journal
caching or journal standby functionality provided by the IBM feature for High
Availability Journal Performance (IBM i option 42). As of MIMIX 8.0, the
DBAPYCACHE policy is shipped so that it is enabled at the installation level.
Upgrading an existing MIMIX 7.1 or older installation to MIMIX 8.0 or later does not
change DBAPYCACHE. If the installation you are upgrading did not use the
DBAPYCACHE policy, follow the instructions in the MIMIX Administrator Reference
book to enable it.
Note: When DB apply cache is used, before and after journal images are sent to the
local journal on the target system.This will increase the amount of storage
needed for journal receivers on the target system if before images were not
previously being sent to the journal.
Access path maintenance — Determines whether MIMIX can optimize access path
maintenance during database apply processing as well as the maximum number of
jobs allowed per data group when performing delayed maintenance. Enabling
optimized access path maintenance improves performance for the database apply
process. To make any change to this policy effective, end and restart the database
apply processes for the affected data groups. For more information about optimizing
access path maintenance, see the MIMIX Administrator Reference book.
Optimize for DB apply
Specify whether to enable optimized access path maintenance. When enabled,
the database apply processes are allowed to temporarily change the value of the
access path maintenance attribute for eligible replicated files on the target system.
Eligible files include physical files, logical files, and join logical files with keyed
access paths that are not unique and that specify *IMMED for their access path
maintenance.
Maximum number of jobs
Specify the maximum number of access path maintenance jobs allowed for a data
group when optimized access path maintenance is enabled. The actual number of
jobs varies as needed between a minimum of one job and the specified value. The
default value is 99.
Maximum rule runtime — Determines the maximum number of minutes an audit can
run when the Automatic audit recovery policy is enabled. The compare phase of the
audit is always allowed to complete regardless of this policy’s value. The elapsed time
of the audit is checked when the recovery phase starts and periodically during the
recovery phase. When the time elapsed since the rule started exceeds the value
specified, any recovery actions in progress will end. This policy has no effect on the
#MBRRCDCNT audit because it has no recovery phase. The shipped default for this
policy of 1440 minutes (24 hours) prevents running multiple instances of the same
audit within the same day. Valid values are 60 minutes through 10080 minutes (1
week).
Audit warning threshold — Determines how many days can elapse after an audit
was last performed before an indicator is set. When the number of days that have
elapsed exceeds the threshold, the indicator is set to inform you that auditing needs

52
Policy descriptions

your attention. The shipped default value of 7 days is at the limit of best practices for
auditing.
Note: It is recommended that you set this value to match the frequency with which
you perform audits. It is possible for an audit to be prevented from running for
several days due to environmental conditions or the Action for running audit
policy. You may not notice that the audit did not run when expected until the
Audit warning threshold is exceeded, potentially several days later. If you run
all audits daily, specify 1 for the Audit warning threshold policy. If you do not
run audits daily, set the value to what makes sense in your MIMIX
environment. For example, if you run the #FILDTA audit once a week and run
all other audits daily, the default value of 7 would cause all audits except
#FILDTA to have exposure indicated. The value 1 would be appropriate for the
daily audits but the #FILDTA audit would be identified as approaching out of
compliance much of the time.

Audit action threshold — Determines how many days can elapse after an audit was
last performed before an indicator is set. When the number of days that have elapsed
exceeds the threshold, the indicator is set to inform you that action is required
because the audit is out of compliance. The shipped default of 14 days is the
suggested value for this threshold, which is 7 days beyond the limit of best practices
for auditing.
Note: It is recommended that you set this value to match the frequency with which
you perform audits. It is possible for an audit to be prevented from running for
several days due to environmental conditions or the Action for running audit
policy. You may not notice that the audit did not run when expected until the
Audit action threshold is exceeded, potentially several days later. If you run all
audits daily, specify 1 for the Audit action threshold policy. If you do not run
audits daily, set the value to what makes sense in your MIMIX environment.
For example, if you run the #FILDTA audit once a week and run all other
audits daily, the default value of 14 would cause all audits except #FILDTA to
have exposure indicated. The value 2 would be appropriate for the daily audits
but the #FILDTA audit would be identified as approaching out of compliance
much of the time.
Audits included for compliance — Specifies which audits will be evaluated for
compliance. For each indicated audit, MIMIX evaluates whether the last run of the
audit exceeds the days specified for the Audit warning and Audit action thresholds
(AUDWARN and AUDACT).
To specify multiple audits by name, a plus sign (+) in the + for more prompt, and press
Enter.
Audit level — Determines the level of comparison that an audit will perform when a
MIMIX rule which supports multiple levels is invoked against a data group. The policy
is in effect regardless of how the rule is invoked. The amount of checking performed
increases with the level number. This policy makes it easy to change the level of audit
performed without changing the audit scheduling or rules. No auditing is performed if
this policy is set to *DISABLED.
The audit level you choose for audits depends on your environment, and especially
on the data compared by the #FILDTA, #DLOATR, and #IFSATR audits. When

53
choosing a value, consider how much data there is to compare, how frequently it
changes, how long the audit runs, how often you run the audit, and how often you
need to be certain that data is synchronized between source and target systems.
Note: Best practice is to use level 30 to perform the most extensive audit. If you use
a lower level, consider its effect on how often you need to guarantee data
integrity between source and target systems.
Regardless of the level you use for daily operations, Vision Solutions strongly
recommends that you perform audits at audit level 30 before the following events to
ensure that 100 percent of the data is valid on the target system:
• Before performing a planned switch to the backup system.
• Before switching back to the production system.
For additional information, see “Guidelines and considerations for auditing” on
page 151 and “Changing auditing policies” on page 42.
Run rule on system — Determines the system on which to run audits. This policy is
used when audits are invoked with *YES specified for the value of the Use run rule on
system policy (USERULESYS) parameter on the Run Rule (RUNRULE) or Run Rule
Group (RUNRULEGRP) command. When *YES is specified in these commands, this
policy determines the system on which to run audits. While this policy is intended for
audits, any rule that meets the same criteria will use this policy.
The policy’s shipped default value, *MGT, runs audits from the management system.
In multi-management environments where both systems defined to a data group are
management systems, the value *MGT will run audits only on the target system.
You can also set the policy to run audits from the network system, the source or target
system, or from a list of system definitions. When both systems of a data group are in
the specified list, the target system is used.
When choosing the value for the Run rule on system policy, also consider your
switching needs.
Action for running audits — Determines the type of audit actions permitted when
certain conditions exist in the data group. If a condition exists at the time of an audit
request, audit activity is restricted to the specified action. If multiple conditions exist
and the values specified are different, only the most restrictive of the specified actions
is allowed. If none of the conditions are present, the audit requests are performed
according to other policy values in effect.
Inactive data group
Specify the type of auditing actions allowed when any replication process required
by the data group is inactive. For example, a data group of TYPE(*ALL) is
considered inactive if any of its database or object replication processes is in a
state other than active. This element has no effect on the #FILDTA and
#MBRRCDCNT audits because these audits can run only when the data group is
active.
Replication process in threshold
Specify the type of auditing actions allowed when a threshold warning condition
exists for any process used in replicating the class of objects checked by an audit.
If a checked process has reached its configured warning value, auditing is

54
Policy descriptions

restricted to the specified actions. Most audits check for threshold conditions in all
database and object replication processes, including the RJ link. #FILDTA audits
only check for threshold warning conditions in the RJ link and database replication
processes. #DLOATR audits only check for threshold warning conditions in object
replication processes.
Audit severity — Specifies the severity to associate with audits that have a status of
*NOTRUN or *IGNOBJ. The selected severity determines whether audits with these
statuses roll up into overall status of MIMIX. The severity also affects the order in
which audits are displayed in interfaces, the color displayed in interfaces that use
color, and the icon displayed with the status in Vision Solutions Portal interfaces.
Not run audit status
Specifies the severity to associate with audits that have a status of *NOTRUN.
Audits with this status are the result of the Action for running audits (RUNAUDIT)
policy, or the data group was inactive when a #FILDTA or #MBRRCDCNT audit
was requested.
Ignored objects audit status
Specifies the severity to associate with audits that have a status of *IGNOBJ.
Audits with this status indicate that some objects were ignored because they were
considered active or could not be compared, typically due to locks or
authorizations. The value for this policy also determines whether any ignored
objects in an audit are counted as differences regardless of the audit status.
Audit history retention — Determines criteria for retaining historical information
about audit results and the objects that were audited. History information for an audit
includes timestamps indicating when the audit was performed, the list of objects that
were audited, and result statistics. Each audit, a unique combination of audit rule and
data group, is evaluated separately and its history information is retained until the
specified minimum days and minimum runs are both met. When an audit exceeds
these criteria, system manager cleanup jobs will remove the historical information for
that audit from all systems and will remove the audited object details from the system
on which the audit request originated. The values specified at the time the cleanup
jobs run are used for evaluation.
Minimum days
Specify the minimum number of days to retain audit history for each completed
audit. Valid values range from 0 through 365 days.The shipped default is 7 days.
Minimum runs per audit
Specify the minimum number of completed audits for which history is to retained.
Valid values range from 1 through 365 runs. The shipped default is 1 completed
audit.
Object details
Specify whether to retain the list of audited objects and their audit status for each
completed audit of library-based objects. The specified value in effect at the time
an audit runs determines whether object details for that run are retained. The
specified value has no effect on cleanup of details for previously completed audit
runs. Cleanup of retained details occurs at the time of audit history cleanup. The
shipped default is *YES.

55
DLO and IFS details
Specify whether to retain the list of audited objects and their audit status for each
completed audit of DLO and IFS objects. The specified value in effect at the time
an audit runs determines whether object details for that run are retained. The
specified value has no effect on cleanup of details for previously completed audit
runs. Cleanup of retained details occurs at the time of audit history cleanup. The
shipped default is *YES.
Note: When large quantities of objects are eligible for replication, specifying *YES to
retain either Object details or DLO and IFS details may use a significant
amount of disk storage. Consider the combined effect of the quantity of
replicated objects for all data groups, the number of days to retain history, the
number of audits to retain, and the frequency in which audits are performed.
Synchronize threshold size — Determines the threshold, in megabytes (MB), to
use for preventing the synchronization of large objects during recovery actions
initiated by database replication, object replication, or audits. When any of the
Automatic system journal recovery, Automatic user journal recovery, or Automatic
audit recovery policies are enabled, all initiated recovery actions use this policy value
for the corresponding synchronize command's Maximum sending size (MB)
parameter. This policy is useful for preventing performance issues when
synchronizing large objects.
Number of third delay retry attempts — Determines the number of times to retry a
process during the third delay/retry interval. This policy is used when the Automatic
system journal recovery policy is enabled. Object replication processes use this policy
value when attempting recovery of an in-use condition that persists after the data
group’s configured values for the first and second delay/retry intervals are exhausted.
The shipped default is 100 attempts.
This policy and its related policy, Third delay retry interval, can be disabled so that
object replication does not attempt the third delay/retry interval but still allow
recoveries for other errors.
Third delay retry interval — Determines the delay time (in minutes) before retrying a
process in the third delay/retry interval. This policy is used when the Automatic
system journal recovery policy is enabled. Object replication processes use this policy
value when attempting recovery of an in-use condition that persists after the data
group’s configured values for the first and second delay/retry intervals are exhausted.
The shipped default is 15 minutes.
Source capture delay — Determines the delay (in minutes) between detecting the
need for a recovery action that requires a synchronization request and capturing the
affected file records, attributes, authorities, or the entire object from the source
system. Recovery actions initiated by replication processes and target journal
inspection can delay these capture operations. There are two delay settings, one for
recoveries during normal replication, and one for recoveries during the testing phase
of a virtual switch. During the recovery phase of a virtual switch, MIMIX ignores both
delay settings in this policy and immediately captures any needed source information
for recoveries that have been identified. Recovery actions are only performed when
the Automatic database recovery and Automatic object recovery policies are enabled.
Changes to either value become effective automatically as the replication manager
checks for recoveries to process.

56
Policy descriptions

Replication recovery processing


Specifies the time, in minutes, to delay before capturing information needed for a
recovery action that requires a synchronization request. During the delay, the
recovery action is not processed and any additional replication recoveries for
subsequent errors of the same type on the same object or file record range will not
initiate additional capture requests, thereby reducing the number of captures. The
shipped default is *NONE.
Virtual switch testing
Specifies the time, in minutes, to delay before capturing information needed for a
recovery action that requires a synchronization request during the application
testing phase of a virtual switch. During the delay, any additional replication
recoveries needed for subsequent errors of the same type on the same object or
file record range will not initiate additional capture requests, thereby reducing the
number of captures. The shipped default is 5 minutes.
Switch warning threshold — Determines how many days can elapse after the last
switch was performed before an indicator is set for the installation. When the number
of days that have elapsed exceeds this threshold, the indicator is set to inform you
that switching may need your attention. The shipped default is 90 days, which is
considered at the limit of best practices for switching.
The indicator is associated with the Last switch field. The Last switch field identifies
when the last completed switch was performed using the default model switch
framework (DFTMSF) policy.
Switch action threshold — Determines how many days can elapse after the last
switch was performed before an indicator is set for the installation. When the number
of days that have elapsed exceeds this threshold, the indicator is set to inform you
that action is required. The shipped default of 180 days is the suggested value for this
threshold, which beyond the limit of best practices for switching.
The indicator is associated with the Last switch field. The Last switch field identifies
when the last completed switch was performed using the default model switch
framework (DFTMSF) policy.
Default model switch framework — Determines the default MIMIX Model Switch
Framework to use for switching. This value is used by configurations which switch via
model switch framework. The shipped default value is MXMSFDFT, which is the
default model switch framework name for the installation. If the default name is not
being used, this value should be changed to the name of the MIMIX Model Switch
Framework used to switch the installation.
Independent ASP library ratio — Determines the number for n in a ratio (n:1) of
independent ASP libraries (n) on the production system to SYSBAS libraries on the
backup system1. For each switchable independent ASP defined to MIMIX by a device
resource group, a monitor with the same name as the resource group checks this
ratio. When the number of independent ASP libraries falls to a level that is below the
specified ratio, the monitor sends a notification to inform you that action may be
required. This signals that your recovery time objective could be in jeopardy because
of a prolonged independent ASP switch time.

1. The library ratio monitor and the policy it uses require a license key for MIMIX for PowerHA.

57
CMPRCDCNT commit threshold — Determines the threshold at which a request to
compare record counts (CMPRCDCNT command or #MBRRCDCNT audit) will not
perform the comparison due to commit cycle activity on the source system. The value
specified is the maximum number of uncommitted record operations that can exist for
files waiting to be applied at the time the compare request is invoked. Each database
apply session is evaluated against the threshold independently. As a result, it is
possible that record counts will be compared for files in one apply session but will not
be compared for files in another apply session. For additional information see the
MIMIX Administrator Reference book.
Directory depth — Determines the number of directory levels to be checked by the
directory data protection report. The policy value at the time the report runs
determines the number of sub-directories included in the resulting report. The default
is 3.
ASP threshold for recoveries — Specifies the threshold as a percent of storage
used by the system auxiliary storage pool (ASP 1) at which certain types of recoveries
are delayed. When the percentage of storage used by the system auxiliary storage
pool (ASP1) on the source system exceeds the specified value, MIMIX delays
processing of new recoveries for member data of *FILE objects. Other recoveries are
not affected and continue to be processed. The delayed recoveries are automatically
submitted after the storage used by ASP1 on the source system is less than the
threshold value. The default is 95.
Procedure history retention — Specifies criteria for retaining historical information
about procedure runs that completed, completed with errors, or that have an
acknowledged status. History information for a procedure includes timestamps
indicating when the procedure was run and detailed information about each step
within the procedure. System cleanup jobs use the policy values at the time they run
to evaluate whether procedure history information can be removed. Each procedure
run is evaluated separately. If the history information includes more runs than the
minimum runs setting, then any runs older than the minimum days are removed by
system cleanup jobs.
Minimum days
Specifies the minimum number of days to retain procedure run history. The default
value is 7.
Minimum runs per procedure
Specifies the minimum number of completed procedure runs for which history is to
retained. This value applies to procedures of all other types except *SWTPLAN
and *SWTUNPLAN. The default value is 1.
Min. runs per switch procedure
Specifies the minimum number of completed switch procedure runs for which
history is to retained. This value applies to procedures of type *SWTPLAN and
*SWTUNPLAN that are used to switch an application group. The default value is
12.

The Audit schedule and Priority audit policies are no longer preferred for changing
audit scheduling criteria. For more information about using the preferred method, see
“Scheduling audits and reports” on page 60.

58
Policy descriptions

59
Scheduling audits and reports

CHAPTER 3 Scheduling audits and reports

This chapter describes how MIMIX automatically submits jobs to run audits and data
protection reports and how you can change scheduling criteria. The shipped default
schedules are also identified.
• “How audits and reports are submitted automatically” on page 60 describes how
the MIMIX Scheduler job is started and how it submits automatic requests, and
identifies what scheduling criteria you can change for audits and reports.
• “Default schedules for automatically submitted audits” on page 64 identifies the
shipped default schedules for both prioritized object auditing and scheduled object
auditing.
• “Default schedules for automatically submitted reports” on page 67 identifies the
shipped default schedule for data protection reports.
• “Changing when automatic audits are allowed to run” on page 68 describes how
to change scheduling criteria for both types of automatic auditing, how to change
the selection frequency of priority auditing categories, and how to disable or
enable automatic auditing.
• “Changing when automatic reports are allowed to run” on page 72 describes how
to change scheduling criteria for data protection reports and how to disable or
enable automatic reports.

How audits and reports are submitted automatically


MIMIX uses a MIMIX Scheduler job (MXSCHED) on each system to automatically
submit requests to run audits and data protection reports at predetermined times.
MIMIX provides default schedules for all audits for every data group and all data
protection reports for every system. You can change the scheduling criteria for any
automatically scheduled audit or report.
The scheduler job is started when the master monitor on a system is started. At the
appropriate time on the local system, the scheduler job on a system initiates a job to
run the scheduled activity.
• For a data protection report, the scheduler initiates a job to run the associated
procedure on the local system.
• For an audit, the scheduler job initiates a job on each system in the data group.
Because audits run between systems, MIMIX evaluates the Run rule on system
policy to determine where the audit should run, and immediately ends the audit
job if it is not on the appropriate system.

Each audit rule1 for a data group has a separate schedule for each type of auditing,
scheduled object auditing or prioritized object auditing. The job scheduler submits
requests for each auditing type.

60
How audits and reports are submitted automatically

The scheduler job on a system remains active whenever the master monitor is active.
The master monitor is started by an autostart entry in the MIMIXSBS subsystem as
well as by the Start MIMIX (STRMMX) command. Starting other processes may not
automatically start the master monitor. You may need to manually start it on each
system with the STRMSTMON command to ensure that automatic scheduling can
occur. For example, starting data groups (STRDG command or processes which
invoke it) does not ensure that the master monitors are active.

Scheduling criteria you can change for audits and reports


The Set MIMIX Schedules (SETMMXSCD) command1 can be used to set scheduling
criteria for automatically submitting audits and reports. A change to a schedule is
effective immediately. If a change occurs while an automatically submitted run is in
progress, the change will be reflected in the next scheduled run of the report or audit.
Schedules for data protection reports and scheduled object auditing requests are
typically set to run by frequency and time of day. They can also be set to run on
specific dates or days of the week, or on relative days of the month.
Schedules for prioritized object auditing requests identify the range of time during
which requests can be submitted every day and identify other criteria for determining
how MIMIX prioritizes which objects are to be audited.
Note: If the Audit level policy in effect when an audit is submitted has a value of
*DISABLED, the audit will not run. This is true regardless of how the audit is
invoked or what is specified for its automatic scheduling criteria.
The following parameters are available on the SETMMXSCD command.
Schedule type — Specifies the type of event that the schedule is for. You must
specify a value.
• Use *AUDIT for either type of automatically submitted audits. When specified, you
must also specify values for Audit rule and Data group definition.
• Use *PROC for data protection reports. When specified you must also specify
values for Procedure and Node (system definition).
Audit rule — Specifies the audit rule for which scheduling and object selection
criteria will be set for the specified data group.
Data group definition — Specifies the three-part qualified name of the data group
definitions to which the specified scheduling and priority auditing criteria applies for
the specified audit rule.
Procedure — Specifies the name of the data protection report procedure to which the
specified scheduling criteria applies.
Node (system definition) — Specifies the node (system definition) to which the
specified procedure and scheduling criteria applies.

1. The #DGFE audit is an exception. Because it compares MIMIX configuration instead of


objects, it does not support prioritized object auditing.
1. Report schedules can only be changed by using the SETMMXSCD command or interfaces
which invoke it. For audit schedules, the SETMMXPCY command is no longer recom-
mended.

61
Scheduling audits and reports

Schedule — Specifies the scheduling information for the specified data protection
report procedure or audit. MIMIX uses this information to automatically submit
requests to run the report or audit.
• When the Schedule type is *AUDIT, the values you specify for Schedule elements
determine when audits that compare all objects selected by data group
configuration entries for the specified rule are performed.
• When the Schedule type is *PROC, the values you specify for Schedule elements
determine when the procedure will run on the node to check whether existing
libraries, directories, or folders on that node are protected by MIMIX replication.
Scheduled dates are entered and displayed in job date format. When the job date
format is Julian, the equivalent month and day are used to determine when to
schedule requests.
State
Specify whether the schedule is enabled or disabled for the specified procedure or
audit. The value *ENABLED is required to allow automatic submission.
Frequency
Specifies how often MIMIX submits a request to run the scheduled activity. The
values specified for other elements further qualify the specified frequency.
Scheduled date
Specifies a date, in job date format, on which MIMIX submits the request. If the
job is to be submitted more than once, this date is used to calculate subsequent
submission dates.
Note: Scheduled date and Scheduled day are mutually exclusive
Scheduled day
Specifies the days of the week on which MIMIX submits the request. If today is the
day of the week that is specified and the scheduled time has already passed, the
request is submitted on the next occurrence of a specified day.
For example, if *FRI is specified for Scheduled day and 12:00:00 is specified for
Scheduled time and you are setting this policy at 11:00 a.m. on a Friday, the
request is submitted the same day. If you are setting the policy at 4:00 p.m. on a
Friday or at 11:00 a.m. on a Monday, the request is submitted the following Friday.
Scheduled time
Specifies the time on the local system at which MIMIX submits the request on the
scheduled date or day. Although the time can be specified to the second, the
activity involved in submitting a job and the load on the system may affect the
exact time at which the job is submitted.
Time can be specified with or without a time separator.
Without a time separator, specify a string of 4 or 6 digits (hhmm or hhmmss)
where hh = hours, mm = minutes, and ss = seconds. Valid values for hh range
from 00 to 23. Valid values for mm and ss range from 00 to 59.
With a time separator, specify a string of 5 or 8 digits where the time separator
specified for your job is used to separate the hours, minutes, and seconds. If this
command is entered from the command line, the string must be enclosed in
apostrophes. If a time separator other than the separator specified for your job is

62
How audits and reports are submitted automatically

used, this command will fail.


Relative day of month
Specifies the relative day of the month on which MIMIX submits the request. A
relative day is only valid when the schedule Frequency is *MONTHLY and
Scheduled day is a value other than *NONE.
For example, if you specify *MONTHLY for Frequency, *TUE for Scheduled day,
and 1 for Relative day of the month, the job is submitted on the first Tuesday of
every month.
Priority audit — Specifies when priority-based audit requests for the specified data
group and audit rule are allowed to automatically start and how often replicated
objects are eligible for auditing based on their priority classification. A prioritized audit
can run multiple times during the specified range, approximately once per hour or
more often. This parameter is ignored for schedule types of *PROC.
The #DGFE rule does not support priority auditing. The priority audit value for #DGFE
audit is always set to *DISABLED and cannot be changed.
State
Specifies whether priority auditing is enabled or disabled for this data group and
audit rule. The value *ENABLED is required to allow automatic submission.
Start after
Specifies the time after which priority-based audits are allowed to start. This is the
beginning of a range of time during which priority-based audits can start each day.
The value *ANY allows priority-based audits to run repeatedly throughout the day.
Note: Times specified for Start after and Start until elements is in 24-hour format
and can be specified with or without a time separator. Without a time
separator, specify a string of 4 or 6 digits (hhmm or hhmmss) where hh =
hours, mm = minutes, and ss = seconds. Valid values for hh range from 00
to 23. Valid values for mm and ss range from 00 to 59. With a time
separator, specify a string of 5 or 8 digits where the time separator
specified for your job is used to separate the hours, minutes, and seconds.
If this command is entered from the command line, the string must be
enclosed in apostrophes. If a time separator other than the separator
specified for your job is used, this command will fail.
Start until
Specifies the end of the time range during which priority-based audits are allowed
to start. Priority-based audits can start until this time. This value is ignored when
Start after is *ANY.
New objects selected
Specifies the frequency at which new objects are considered for auditing. A new
object is one that has not been audited since it was created.
Changed objects selected
Specifies the frequency at which changed objects are considered for auditing. A
changed object is one that has been modified since the last time it was audited.
Unchanged objects selected
Specifies the frequency at which unchanged objects are considered for auditing.

63
Scheduling audits and reports

An unchanged object is one that has not been modified since the last time it was
audited.
Audited with no diff.
Specifies the frequency at which objects with no differences are considered for
auditing. An object with no differences is one that has not been modified since the
last time it was audited and has been successfully audited on at least three
consecutive audit runs.

Default schedules for automatically submitted audits


Each shipped audit rule has default values for submitting prioritized object audits as
well as for submitting scheduled object audits. These values are assigned to the
audits for a data group when the data group is created. The shipped values listed in
Table 5 are used for auditing in all new data groups.

Table 5. Initial values for automatically submitting audits shown in context of the parame-
ters on the Set MIMIX Schedules (SETMMXSCD) command.

SETMMXSCD Parameters Initial Values for Audits at Creation

Schedule (SCHEDULE)
State *ENABLED1
Frequency *WEEKLY1
Scheduled date
Scheduled day *SUN
Scheduled time Exact time varies by rule, see Table 6.
Relative day of month

Priority audit (PTYAUD)


State *ENABLED2
Start after 0300002
Start until 080000
New objects selected *DAILY
Changed objects selected *DAILY
Unchanged objects selected *WEEKLY
Audited with no differences *MONTHLY
1. In installations that upgraded from version 7.1.01.00 through 7.1.11.00 to version 8, any data groups
that existed prior to upgrading have their audit schedule values mapped as follows. If the Frequency
was a value other than *NONE, that value is preserved and the State is set to *ENABLED. If the Fre-
quency value was *NONE, it is changed to *WEEKLY and the State is set to *DISABLED.
2. In installations that upgraded from version 7.1.01.00 through 7.1.11.00 to version 8, any data groups
that existed prior to upgrading have their priority audit values mapped as follows. If the Start after
value was a value other than *NONE, that value is preserved and the State is set to *ENABLED.
However if the Start after value was *NONE, it is changed to 030000 and State is set to *DISABLED.

Every day, a prioritized audit starts one or more times each hour during the time range
specified in its scheduling criteria. Other criteria specified with an audit’s priority time
range determine the auditing frequency assigned to the auditing priority categories.

64
Default schedules for automatically submitted audits

A scheduled audit runs once at its specified time on the days or dates for its
frequency as specified in its schedule. The shipped value for start time of each audit
rule is staggered, beginning at 2 a.m.
Table 6 shows the initially set times for priority audits versus scheduled audits.

Table 6. Initial times for automatically submitted audits for a new data group.

Shipped Shipped Rule Name Description Job Name


Priority Scheduled
Start Range Time

n/a1 2:00 a.m. #DGFE Checks configuration for files using sdn_DGFE
cooperative processing.
Uses the Check Data Group File Entries
(CHKDGFE) command.

65
Scheduling audits and reports

Table 6. Initial times for automatically submitted audits for a new data group.

Shipped Shipped Rule Name Description Job Name


Priority Scheduled
Start Range Time

All other 2:05 a.m. #OBJATR Compares all attributes for all object sdn_OBJATR
audits: types supported for replication.
3 a.m. Uses the Compare Object Attributes
to (CMPOBJA) command
8 a.m. 2:10 a.m. #FILATR Compares all file attributes. sdn_FILATR
Uses the Compare File Attributes
(CMPFILA) command.

2:15 a.m. #IFSATR Compares IFS attributes. sdn_IFSATR


Uses the Compare IFS Attributes
(CMPIFSA) command.

2:20 a.m. #FILATRMBR Compares basic file attributes at the sdn_MBRATR


member level.
Uses the Compare File Attributes
(CMPFILA) command.

2:25 a.m. #DLOATR Compares all DLO attributes. sdn_DLOATR


Uses the Compare DLO Attributes
(CMPDLOA) command.

Disabled2 #MBRRCDCNT Compares the number of current sdn_RCDCNT


records (*CURRDS) and the number of
deleted records (*NBRDLTRCDS) for
physical files that are defined to an
active data group.
Uses the Compare Record Counts
(CMPRCDCNT) command.
Note: Equal record counts suggest but do
not guarantee that files are
synchronized. This audit does not
have a recovery phase. Differences
detected by this audit appear as not
recovered in the Audit Summary.

2:35 a.m. #FILDTA3 Compares file contents. sdn_FILDTA


Uses the Compare File Data
(CMPFILDTA) command.
1. The #DGFE audit is not eligible for prioritized auditing because it checks configuration data, not objects.
2. New installations running service pack 8.0.07.00 or higher will get this default value. Instances that upgraded to ser-
vice pack 8.0.07.00 or higher will retain their existing value. For service packs 8.0.06.00 and earlier, the default
shipped schedule for #MBRRCDCNT audit is 2:30 a.m.
3. The #FILDTA audit and the Compare File Data (CMPFILDTA) command require TCP/IP communications as their
communications protocol.

66
Default schedules for automatically submitted reports

Default schedules for automatically submitted reports


When a system definition is created, default schedules are also created to allow the
following data protection reports to automatically run on that system:
• CRTDPRDIR - Creates a data protection report of the directories on a system,
with some exceptions, that identifies whether the directories and their contents are
configured for MIMIX replication.
• CRTDPRFLR - Creates a data protection report of the folders on a system, with
some exceptions, that identifies whether the folders and their contents are
configured for MIMIX replication.
• CRTDPRLIB - Creates a data protection report of the libraries on a system, with
some exceptions, that identifies whether the libraries and their contents are
configured for MIMIX replication.
Each shipped data protection report has a default schedule used to submit its
associated procedure. The default scheduling values, shown in Table 7, are used for
all new system definitions. The shipped value for each report is set to run every
Sunday at 3 a.m. on the local system.

Table 7. Initial values for automatically submitting reports, shown in context of the parame-
ters on the Set MIMIX Schedules (SETMMXSCD) command.

SETMMXSCD Parameters Initial Values for Data Protection Reports at Creation

Schedule (SCHEDULE)
State *ENABLED
Frequency *WEEKLY
Scheduled date
Scheduled day *SUN
Scheduled time 3:00 a.m.
Relative day of month

67
Changing when automatic audits are allowed to run
This topic describes how to change scheduling criteria for when automatic audits will
run. To effectively audit your replication environment you may need to fine-tune when
MIMIX is allowed to automatically submit prioritized object audits and scheduled
object audits. MIMIX service providers are trained to provide a robust audit solution
that meets your needs.
For both types of auditing, consider:
• How much time or system resource can you dedicate to audit processing each
day, week, or month?
• How often should all data within the database be audited? Business requirements
as well as time and system resources need to be considered.
• Does automatic scheduling conflict with regularly scheduled backups?
• Are there jobs running at the same time as audits that could lock files needing to
be accessed during recovery?
For scheduled object auditing (which select all objects), also consider:
• Are there are a large number of objects to be compared?
• Are there a large number of objects for which a rule is expected to attempt
recovery?
• Specific audits may have additional needs. See “Considerations for specific
audits” on page 152.
• While you may decide to vary the scheduled times, it is recommended that you
maintain the same relative order indicated in “Default schedules for automatically
submitted audits” on page 64.

Changing scheduling criteria for automatic audits


Both scheduled audits and priority audits have scheduling information. A change to
an audit’s scheduling information is effective immediately. If an audit is in progress at
the time its scheduling information is changed, the change is effective on the next
automatic run of the audit.
Do the following from the management system:
1. From the MIMIX Intermediate Main Menu, select option 6 (Work with audits) and
press Enter.
2. The Work with Audits display appears. Type 37 (Change audit schedule) next to
the audit you want to change and press Enter.
3. The Set MIMIX Schedules (SETMMXSCD) command appears, showing the
schedule type as *AUDIT and the selected audit rule and data group. Press Enter.
Then complete Step 4 and Step 5 as needed.
4. To change when MIMIX is allowed to submit runs of the audit to check all
configured objects, do the following:
a. Press Page Down.

68
Changing when automatic audits are allowed to run

b. The current values for the Schedule prompts are displayed. Specify the values
you want for these prompts.
• Frequency
• Scheduled date
• Scheduled day
• Scheduled time
• Relative day of month
5. To change when MIMIX is allowed to submit priority-based runs of the audit every
day, do the following:
a. Press Page Down. If you skipped the previous step you may need to do this
twice.
b. The current values of Priority audit prompts are displayed. Specify the values
for these prompts:
• Start after
• Start until
6. To make the changes effective, press Enter.

Changing the selection frequency of priority auditing categories


When priority auditing is used, you can control how often objects within priorities are
eligible for selection. Objects which had differences in their previous audit are always
selected. For other priority classes, you can change how often objects within the class
are eligible for selection by a prioritized audit.
If an audit is in progress at the time its category frequency information is changed, the
change is effective on the next automatic run of the audit.
Note: The #DGFE audit does not support priority auditing.
Do the following from the management system:
1. From the MIMIX Intermediate Main Menu, select option 6 (Work with audits) and
press Enter.
2. The Work with Audits display appears. Type 37 (Change audit schedule) next to
the audit you want to change and press Enter.
3. The Set MIMIX Schedules (SETMMXSCD) command appears, showing the
schedule type as *AUDIT and the selected audit rule and data group. Do the
following.
a. Press Enter, then press Page Down two (2) times.
b. The current values for the selected audit and data group are displayed in the
Priority audit prompts. To change selection criteria, specify values in the
following prompts to indicate how often objects in each category are eligible for
selection by a priority audit.
• New objects selected

69
• Changed objects selected
• Unchanged objects selected
• Audited with no diff.
For descriptions of the priority classes with changeable frequencies, see “How
priority auditing determines what objects to select” on page 148.
4. To make the changes effective, press Enter.

Disabling or enabling automatically submitted audits


You can control whether each audit for a data group can be submitted automatically
by priority or by schedule. You can have both, one, or neither type of automatically
submitted auditing.
Use these instructions to disable or enable only automatically submitted audits from
running. Manually invoked audits can still be performed. If you need to temporarily
disable both manual and automatic audit submissions for an installation during
system or network maintenance, temporarily changing audit level policy is a better
choice.
The change is effective immediately. Any audit in progress at the time of the change is
not affected.
Do the following from the management system:
1. From the MIMIX Intermediate Main Menu, select option 6 (Work with audits) and
press Enter.
2. The Work with Audits display appears. Type 37 (Change audit schedule) next to
the audit you want to change and press Enter.
3. The Set MIMIX Schedules (SETMMXSCD) command appears, showing the
schedule type as *AUDIT and the selected audit rule and data group. Press Enter.
Then complete Step 4 and Step 5 as needed.
4. To disable or enable scheduled object auditing, do the following:
a. Press Page Down.
b. The current values for the Schedule prompts are displayed. At the State
prompt specify the appropriate value:
• Specify *DISABLED to disable scheduled object audit runs of the audit for
the specified data group.
• Specify *ENABLED to enable scheduled object audit runs of the audit for the
specified data group.
5. To disable or enable priority-based audits, do the following:
a. Press Page Down. If you skipped the previous step you may need to do this
twice.
b. The current values of Priority audit prompts are displayed. At the State prompt
specify the appropriate value:

70
Changing when automatic audits are allowed to run

• Specify *DISABLED to disable priority-based runs of the audit for the


specified data group.
• Specify *ENABLED to enable priority-based runs of the audit for the specified
data group.
6. To make the changes effective, press Enter.

71
Changing when automatic reports are allowed to run
This topic describes how to change scheduling criteria for when MIMIX is allowed to
submit requests to run data protection reports. Each data protection report is a
procedure that runs on a specific system.
When changing the schedule for a report, consider the following:
• Which categories of objects do you need to replicate? If you do not need to
replicate library-based objects, IFS objects in directories, or DLO objects in
folders, you can disable schedules for the respective category of data protection
report.
• On which systems do you need to check whether new libraries, directories, or
folders are protected by MIMIX replication? Data protection reports should be run
on systems in your production environment that are source systems for
replication. You might want to disable schedules for data protection reports on
systems that are not source systems or that can never become source systems
for any application group or data group.
• How frequently are new libraries, directories, or folders created on a system? This
affects how often the reports should be scheduled to run on that system to
evaluate whether data is protected by MIMIX. Business requirements as well as
time and system resources also need to be considered when determining
scheduling frequency.
• What is the setting for the Directory depth policy? This policy determines the
number of directory levels that are checked by the data protection report for
directories. The value of this policy may influence how often or when you want to
run the directory report. Changing the value to include more directories for a
report may cause the report to run longer. Consider the needs of your
environment and system resources available when the report runs.
• Does automatic scheduling conflict with regularly scheduled backups or times
when a system has high processing loads?
• Does heavy creation, deletion, renaming, or locking activity associated with
libraries, directories, or folders occur at the same time that data protection reports
run which might affect the report results?

Changing scheduling criteria for automatic reports


A report schedule is specific to a data protection report and the system on which it
runs. A change to a report’s scheduling information is effective immediately. If a report
is in progress at the time its scheduling information is changed, the change is effective
on the next automatic run of the report’s procedure.
Do the following from the management system:
1. To access the data protection reports for a specific system, do the following:
a. From the MIMIX Intermediate Main Menu, select option 2 (Work with systems)
and press Enter.

72
Changing when automatic reports are allowed to run

b. From the Work with Systems display, select option 21 (Procedure status) next
to the system you want and press Enter.
2. The Work with Procedure Status display appears showing all procedures of type
*NODE that can run on the selected system. Type 37 (Change schedule) next to
the procedure you want to change and press Enter.
Note: Only procedures named CRTDPRDIR, CRTDPRFLR, and CRTDPRLIB
are data protection reports that be scheduled for automatic submission by
MIMIX.
3. The Set MIMIX Schedules (SETMMXSCD) command appears, showing the
schedule type as *PROC and the selected procedure and node. Press Enter, then
press Page Down.
4. The current values for the Schedule prompts are displayed. Specify the values
you want for these prompts:
• Frequency
• Scheduled date
• Scheduled day
• Scheduled time
• Relative day of month
5. To make the changes effective, press Enter.

Disabling or enabling automatically submitted reports


You can control whether each data protection report on a system will be submitted
automatically. Use these instructions to disable or enable only the automatic
submission of a report. You can always manually submit a request to run a report by
running the appropriate procedure.
The change is effective immediately. Any report in progress at the time of the change
is not affected.
Do the following from the management system:
1. To access the data protection reports for a specific system, do the following:
a. From the MIMIX Intermediate Main Menu, select option 2 (Work with systems)
and press Enter.
b. From the Work with Systems display, select option 21 (Procedure status) next
to the system you want and press Enter.
2. The Work with Procedure Status display appears showing all procedures that can
run on the selected system. Type 37 (Change schedule) next to the procedure
you want to change and press Enter.
Note: Only procedures named CRTDPRDIR, CRTDPRFLR, and CRTDPRLIB
are data protection reports that be scheduled for automatic submission by
MIMIX.
3. The Set MIMIX Schedules (SETMMXSCD) command appears, showing the

73
schedule type as *PROC and the selected procedure and node. Press Enter, then
press Page Down.
4. The current values for the Schedule prompts are displayed. At the State prompt
specify the appropriate value:
• Specify *DISABLED to disable the report from being submitted automatically
on the specified system.
• Specify *ENABLED to enable the report for automatic submission on the
specified system.
5. To accept the change, press Enter.

74
Checking status in application group environments

CHAPTER 4 Where to check status

The types of statuses to check and where you check them depends on how your
MIMIX environment is configured.
• For configurations with application groups of type *NONCLU (not configured for
an IBM i clustering environment), use “Checking status in application group
environments” on page 75.
Note: This document does not address status of application groups of type *CLU
that are configured for an IBM i clustering environment. If you are using
clustering or have MIMIX for PowerHA configured, use the MIMIX
Operations with IBM PowerHA for i book to check status within a clustering
environment.
• For configurations with only data groups, use “Checking status in data group-only
environments” on page 78.
The MIMIX Availability Status display (WRKMMXSTS command) should not be used
in environments that use application groups and is no longer documented in this
book.

Checking status in application group environments


MIMIX summarizes the status of all activity associated with an application group into
broad categories for procedures, nodes (systems), and replication-related activity and
reports the summarized status on the Work with Application Groups (WRKAG)
display.
When necessary, you can drill into each category for details about a subset for a
specific application group.
Important!
Always start at the Work with Application Groups (WRKAG) display. There is a
hierarchy for resolving reported problems on this display. To assist you with
following this hierarchy, the resolution for all types of statuses shown in Table 8
intentionally begins with topic “Resolving problem statuses for an application
group” on page 82.
Also, this document does not address status of application groups of type *CLU
that are configured for an IBM i clustering environment. If you are using clustering
or have MIMIX for PowerHA configured, use the MIMIX Operations with IBM

75
Where to check status

PowerHA for i book to check status within a clustering environment.

Table 8. Checking status in environments with application groups of type *NONCLU

What to Check Where to Check and Detailed Instructions

Procedures Summary: Work with Application Groups display, option 21 (Procedure status).
Details: Work with Procedures Status display.
See: “Displaying status of procedures” on page 98.
See also: “Resolving problem statuses for an application group” on page 82.

System-level processes Summary: Work with Application Groups display, App Node Status column.
Also, Work with Node Entries display subsets to a summary status by node.
Details: Work with Systems display.
See: “Resolving problem statuses for an application group” on page 82.
See also: For additional problem determination, you may also need “Displaying
status of system-level processes” on page 180.

Replication direction Summary: Work with Application Groups display, Repl.Status column. Also,
(node roles) Work with Data Rsc. Grp. Ent. display, Replication Status column, subsets to a
summary by resource group.
Replication processes
Details: Detail locations vary depending on type of problem. Follow the
Replication errors hierarchy for resolving application group problems.
See: “Resolving problem statuses for an application group” on page 82.
Auditing errors
See also: For additional problem determination, you may also need:
Recoveries in progress • Replication direction: “Resolving status for data resource group entries” on
page 88.
• Replication processes (WRKDG): “Identifying and resolving replication
problems reflected in the Source and Target columns” on page 123.
• Replication errors (WRKDG, highlighted name in Data Group column):
“Identifying and resolving problems reflected in the Data Group column” on
page 120.
• Replication errors (WRKDG, Errors DB and Errors Obj columns): “Working
with files needing attention (replication and access path errors)” on page 263,
“Working with tracking entries” on page 272, and “Working with objects in
error” on page 278.
• Auditing errors: “Resolving auditing problems” on page 162.
• Recoveries in progress: “Displaying recoveries” on page 211.

Notifications Summary: Work with Application Groups display, Notifications field identifies the
highest severity level of any new notifications.
Details: Work with Notifications display.
See: “Displaying notifications” on page 205.

Monitors - status and Summary: Work with Application Groups display, Monitors field identifies the
errors highest severity monitor problem on the local system.
Details: Work with Monitors display.
See: “Resolving problems reported in the Monitors field” on page 85.
See also: “Displaying status of monitors on the local system” on page 192.

76
Checking status in application group environments

Table 8. Checking status in environments with application groups of type *NONCLU

What to Check Where to Check and Detailed Instructions

Switching status Summary: Work with Application Groups display, AG Proc. Status column
includes switching procedures in the summary status.
Details: Work with Procedures Status display.
See: “Resolving problem statuses for an application group” on page 82.
See also: “Displaying status of procedures” on page 98.

77
Where to check status

Checking status in data group-only environments


In environments with only data groups, the Work with Data Groups (WRKDG) display
is the primary location from which to start checking status of replication activity and
processes.

Table 9. Checking status in environments with only data groups.

What to Check Where to Check and Detailed Instructions

System-level processes Summary: Work with Data Groups display, Source Mgr and Target Mgr
columns.
Details: Work with Systems display.
See: “Displaying status of system-level processes” on page 180.

Node procedures Summary: Work with Systems display, option 21 (Proc Sts).
Details: Work with Procedures Status display.
See: “Displaying status of procedures” on page 98.

Replication processes Summary: by system where processes run, Work with Data Groups display,
Source DB, Source OBJ, Target DB, and Target OBJ columns.
See: “Identifying and resolving replication problems reflected in the Source and
Target columns” on page 123.

Replication errors Summary: Work with Data Groups display, two locations:
• Data Group column, when name is highlighted
See: “Identifying and resolving problems reflected in the Data Group column”
on page 120
• Errors DB and Errors Obj columns
See: “Working with files needing attention (replication and access path
errors)” on page 263, “Working with tracking entries” on page 272, and
“Working with objects in error” on page 278.

Auditing errors Summary: installation count on Work with Data Groups display, Audits field.
Details: Work with Audits display.
See: “Resolving auditing problems” on page 162.

Recoveries in progress Summary: installation count on Work with Data Groups display, Recov. field.
Details: Work with Recoveries display, and the Object Correction Activity
window in Vision Solutions Portal.
Note: A recovery in progress indicates that MIMIX detected a problem and is currently
attempting to correct it. It is important that no recoveries are in progress when
certain activity, such as ending MIMIX, is performed. The count on the Work with
Data Groups display includes recoveries from all possible sources. However, the
Work with Recoveries display only lists new or in-progress recoveries from audits
or that are the result of object replication third delay retry processing. Recoveries
that originated from replication processes, target journal inspection, and virtual
switch procedures can only be viewed from Vision Solutions Portal.
See: “Displaying recoveries” on page 211.

Notifications - new Summary: installation count on Work with Data Groups display, Notif. field.
errors or warnings Details: Work with Notifications display.
See: “Displaying notifications” on page 205.

78
Checking status in data group-only environments

Table 9. Checking status in environments with only data groups.

What to Check Where to Check and Detailed Instructions

Monitors (processes Details: Work with Monitors display on each system.


and errors) See: “Displaying status of monitors on the local system” on page 192.

Switching status Summary: Switch in progress indicated on Work with Data Groups display,
Source DB, Source OBJ, Target DB, and Target OBJ columns.
See: “Identifying and resolving replication problems reflected in the Source and
Target columns” on page 123.

79
Checking application group status

CHAPTER 5 Checking application group status

Monitoring status in environments that use application groups begins at the level of
the application group. Resolving reported problems will require investigation into
additional displays for more details.
Note: This chapter does not include status for application groups that are configured
for an IBM i clustering environment. If you are using clustering or have
MIMIX® for PowerHA® configured, see the MIMIX Operations with IBM
PowerHA for i book for status information within a clustering environment.

Displaying application group status


The status view of the Work with Application Groups display provides a summary of
all status associated with an environment configured with application groups.
1. Do one of the following to access the Work with Application Groups display:
• Select option 1 (Work with application groups) from the MIMIX Basic Main
Menu.
• Select option 5 (Work with application groups) from the MIMIX Intermediate
Main Menu.
• Enter the command: WRKAG
2. If necessary, use F10 to access the status view.
3. Check the values displayed. Figure 3 shows ideal status conditions. Any other
status values require investigation to determine the cause of a problem. To
ensure that you address problems in the appropriate order, use “Resolving
problem statuses for an application group” on page 82.

80
Displaying application group status

Figure 3. Status view of Work with Application Groups display

Work with Application Groups


System: SYSA
Node procedures. . : *COMP Monitors . . . . . : *ACTIVE
Notifications . . : *NONE
Type options, press Enter.
1=Create 2=Change 4=Delete 5=Display 6=Print 9=Start
10=End 12=Node entries 13=Data resource groups 15=Switch

App App App Node Data Rsc Data Node Repl. AG Proc.
Opt Group Status Status Grp Status Status Status Status
__ __________
__ SAMPLEAG *ACTIVE *ACTIVE *COMP

Bottom
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F6=Create F9=Retrieve F10=View config
F12=Cancel F13=Repeat F18=Subset F23=More options F24=More keys

Note: The App Status, Data Rsc Grp Status, and Data Node Status columns are
always blank for application groups that are not configured for an IBM i
clustering environment. If your environment is configured to support
clustering, use the MIMIX Operations with IBM PowerHA for i book.

Table 10. Work with Application Groups display -ideal status values for replication

Column or Field What is Summarized Ideal Status

App Status Always blank for non-cluster application groups.

App Node Status Status of system-level processes that run the nodes *ACTIVE
within the application group

Data Rsc Grp Status Always blank for non-cluster application groups.

Data Node Status Always blank for non-cluster application groups.

Repl. Status Status of all replication processes, replication *ACTIVE


errors, replication direction, audits, and recoveries
associated with the application group.

AG Proc Status Status of the most recent run of each procedure *COMP
associated with an application group.

Monitors Status of monitors on the local system. *ACTIVE

Notifications Status of new notifications for the installation. *NONE

81
Checking application group status

Resolving problem statuses for an application group


There is an order in which to resolve problems reported on the Work with Application
Groups display. This topic describes the order of precedence in which to resolve
problems.
Each column on the display is a summation of multiple processes (Table 10). The
fields in the upper right of the display indicate other processes with status that also
need to be checked.
Troubleshooting Tip: When investigating problems, begin with the Proc. Status
column. A problem with procedure status can affect values in other columns.
When any procedure status problems are resolved, refresh the display. Then
check the other columns beginning the left-most column that is reporting a
problem. Resolve the most severe problem in that column first, then refresh the
display. Investigate problems in the remaining columns from left to right.
To address the most common problems with status for application groups, do the
following:
1. Check the AG Proc Status column first. Resolve any problems using Table 11 in
“Resolving a procedure status problem reported for an application group” on
page 82.
2. Refresh the Work with Application Groups display.
3. Starting from the left of the display, check the other columns one at a time as
follows:
a. Resolve any *ATTN status problems in a column first, using Table 12 in
“Resolving an *ATTN status for an application group” on page 83.
b. Refresh the Work with Application Groups display.
c. Address less severe problems in the same column using Table 13 in
“Resolving other common status values for an application group” on page 84.
d. Refresh the Work with Application Groups display.
e. Repeat this step for the next column. Do this until you have resolved all
problems in all columns.
4. Check the Node procedure field. See “Working with procedures and steps” on
page 97.
5. Check the Monitor field using “Displaying status of monitors on the local system”
on page 192.
6. Check the Notifications field using “Displaying notifications” on page 205.

Resolving a procedure status problem reported for an application group


The AG Proc Status column represents a summary of the most recent run of all
procedures defined for the application group.
Important! Check the status of the AG Proc Status column and address any
problem indicated by *ATTN or *ACTIVE status before attempting to resolve any

82
Resolving problem statuses for an application group

problem reported in other columns on the Work with Application Groups display.
Status shown in other columns will likely change after active procedures complete
or procedures which require attention have been addressed.

Table 11. Summary of Procedure Status values that require attention

Column Value Description and Action

*ACTIVE One or more of the last started runs of the procedures to run are still
active or queued. Wait for the procedure to complete. Do not attempt to
correct other status problems reported on the display until the procedure
completes. Use option 21 (Procedure status) to view the status of the
last started runs of procedures.

*ATTN One or more of the last started runs of the procedures have a status that
requires attention. Resolve the procedure problems before attempting to
correct other status problems reported on the Work with Application
Groups display.
1. Use option 21 (Procedure status) to view the status of the last started
runs of the procedures.
2. The resulting procedures shown on the Work with Procedure Status
display which have status values of *ATTN, *CANCELED, *FAILED,
*MSGW, or *PENDCNL require user action.
3. It may be necessary to check status of the steps within the procedure
to resolve a step problem before the procedure can continue. Use
option 8 (Step status) to display step status for the procedure.
Note: For a virtual switch procedure (type *SWTVRT), a *MSGW status for step
MXVRTRCY is expected and should remain unanswered for the duration
of the testing phase of the virtual switch. Do not respond to the message
for step MXVRTRCY until testing is complete and you are ready for the
recovery phase of the virtual switch procedure. Return to the Work with
Application Groups display to check for other reported problems with the
application group that need to be resolved.
For detailed information, see “Working with procedures and steps” on
page 97.

Note: The status *COMP is a summary status that indicates the most recently
started run of each procedure for the application group has completed,
completed with errors, or has an acknowledged status. For any individual
procedure that completed with errors, user action is recommended to
investigate the cause of the error and assess its implications.

Resolving an *ATTN status for an application group


The value *ATTN can appear in each column of the Work with Application Groups
display to indicate that user action is required to correct a problem.
Important! Check the status of the AG Proc Status column and address any
problem indicated by *ATTN or *ACTIVE status before attempting to resolve any
problem reported in other columns. Use “Resolving a procedure status problem
reported for an application group” on page 82.

83
Checking application group status

If there are no procedure problems, each of the other columns with an *ATTN status
must be addressed individually, starting from the left-most column.

Table 12. Resolving *ATTN status for columns (except AG Proc Status) on the Work with
Application Groups display

*ATTN Status Description and Actions for *ATTN Status


in Column

App Node The App Node Status column is a summary of the status of the nodes
Status associated with the application group. The status includes the MIMIX
system manager, journal manager, target journal inspection, and collector
services jobs for the nodes in the application group.
*ATTN indicates that the node status and the MIMIX manager status
values do not match. Investigate the status of the associated nodes and
MIMIX managers using option 12 (Node entries). Then use “Resolving
status for node entries” on page 87.

Replication The Replication Status column is a summary status of data replication


Status activity for the data resource groups associated with an application group.
*ATTN indicates that data replication for at least one data group in the
data resource groups has one or more of the following conditions:
• Is active with an incorrect source system
• Has a status that does not match the status of the appropriate data
resource group
• Has a failed state
• Has an error condition
• Has a process other than apply that is inactive during a virtual switch
(possible only in *NONCLU application groups)
• Has audit errors
• Has pending recoveries
To determine the cause, use option 13 (Data resource groups) to identify
the data resource group where the problem exists. Then use “Resolving
status for data resource group entries” on page 88.

Resolving other common status values for an application group


Table 13 lists other common problems with application group status and identifies
how to begin to their resolution.

Table 13. Other problem statuses which may appear in multiple columns on the Work with Application
Groups display

Column Value Description and Action

*ATTN Each column has a unique recovery. See “Resolving an *ATTN status for an application
group” on page 83.

84
Resolving problem statuses for an application group

Table 13. Other problem statuses which may appear in multiple columns on the Work with Application
Groups display

Column Value Description and Action

*INACTIVE The current status of the resource group or node is inactive. This status is possible in the
Repl. Status column.
• If all columns with a status value are *INACTIVE, the application group may have been
ended intentionally. Use option 9 (Start) to start the application group.
• If this value appears only in the App Node Status column, the application resource
group nodes are all inactive and all MIMIX manager jobs are also inactive. Use option
12 (Node entries) to investigate further. For more information see “Resolving status for
node entries” on page 87.
• If this value appears only in the Repl. Status column, logical replication is not active.
Use option 13 (Data resource groups) to investigate. For more information see
“Resolving status for data resource group entries” on page 88.

*UNKNOWN The current status is unknown. The local node is a network node in a non-cluster
application group and does not participate in the recovery domain. Its status cannot be
determined.
When this status appears in all columns, do one of the following:
• Enter the command WRKSYS. On the Work with Systems display, check the status of
Cluster Services for the local system definition. If necessary, use option 9 (Start) to
start cluster services.
• Sign on to a node that is active and use the WRKAG command to check the
application group status. If the status is still *UNKNOWN, use option 12 (Node entries)
to check the status of Cluster Services on the node.

*VRTSWTSTR These values can appear in the Repl. Status column when a virtual switch procedure is in
*VRTSWTTST progress for an application group of type *NONCLU. These statuses are not necessarily a
*VRTSWTRCY problem. However, complete your virtual switch testing as soon as possible to limit
exposure to your environment. While virtual switch testing is in progress, replicated
transactions accumulate because apply processes and access path maintenance for
affected data groups are ended. These transactions must be processed during virtual
switch recovery, and the virtual switch procedure must complete before any other
operation such as starting, stopping, or switching can be performed for the application
group.

Resolving problems reported in the Monitors field


The Monitors field located in the upper right corner of the Work with Application
Groups display summarizes the status of the MIMIX monitors on the local system.
Each node or system in the product configuration has MIMIX monitors which run on
that system to check for specific potential problems. A status of *ACTIVE indicates
that all enabled monitors on the local system are active.

85
Checking application group status

Table 14 shows possible status values for the Monitors field that require user action.
For a complete list of possible values, press F1 (Help).

Table 14. Monitor field status values that may require user action

Monitor Status Description

*ATTN Either one or more monitors on the local system failed or there are both
active and inactive monitors on the local system.

*INACTIVE All enabled monitors on the local system are inactive.

Do the following:
1. Press F14 (Monitors) to display the list of monitors on the local system on the
Work with Monitors display.
2. Check the Status column for status values of FAILED, FAILED/ACT, and
INACTIVE.
3. If the monitor is needed on the local system as indicated in Table 15, use option 9
(Start) to start the monitor.

Table 15. Possible monitors and the nodes on which they should be active

Monitor When and Where Needed

journal-name - remote journal link monitor Primary node and the


Checks the journal message queue for indications of current Backup node of
problems with the remote journal link. A monitor exists for application groups which
both the local and remote system of the RJ link. perform logical replication.

MMIASPMON - independent ASP threshold monitor On all nodes which control


Checks the QSYSOPR message queue for indications that an independent ASP.
the independent ASP threshold has been exceeded. This
monitor improves the ability to detect overflow conditions
that put your high availability solution at risk due to
insufficient storage.

MMNFYNEWE - monitor for new object notification entries Primary node when the
Monitors the source system for the newly created libraries, application group is
folders, or directories that are not already included or configured for logical
excluded for replication by a data group configuration. replication.

MQRCDIMG - MQ record image monitor Primary node,


This monitor is optional for environments that use MIMIX for
MQ. The monitor run a time-based event which runs the
IBM WebSphere MQ command RCDMQMIMG for each
active queue manager on the system.This monitor is
available in MIMIX 8.1.17.00 and higher.

86
Resolving status for node entries

Resolving status for node entries


The Work with Node Entries display lists the nodes associated with an application
group.

Figure 4. Status view of Work with Node Entries display for an application group which does
not participate in a cluster

Work with Node Entries


System: SYSA
Application group . . . . . : SAMPLEAG

Type options, press Enter.


1=Add 2=Change 4=Remove 5=Display 6=Print 9=Start 10=End

-------------Current------------- Manager
Opt Node Role Sequence Data Provider Status
__ ________
__ SYSB *PRIMARY *PRIMARY *ACTIVE
__ SYSA *BACKUP 1 *PRIMARY *ACTIVE

Bottom
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F6=Add F7=Systems F9=Retrieve
F10=View config F11=Sort by node F12=Cancel F18=Subset F24=More keys

For each node listed, check the Manager Status column for status values that require
attention. For a complete list of status values for each field and column, press F1
(Help).
Manager Status - This column indicates the status of all of the MIMIX system
manager, journal manager, target journal inspection, and collector services jobs for
the specified node.

Table 16. Manager Status values that require user action.

Status Description and Action


Value

*ATTN At least one of the system manager, journal manager, target journal
inspection, or collector services jobs for the node has failed.
When all the nodes listed do not have the same value, use F7 (Systems) to
access the Work with Systems display.
• Check the status of the system and journal managers, target journal
inspection, and collector services.
• Use option 9 (Start) to start the managers and services that are not active
on the node.

87
Checking application group status

Table 16. Manager Status values that require user action.

Status Description and Action


Value

*INACTIVE All system manager, journal manager, target journal inspection, and
collector services jobs for the specified node are inactive. This may be
intentional when MIMIX is ended to perform certain activities.
Use F7 (Systems) to access the Work with Systems display.

For additional information, see “Displaying status of system-level processes” on


page 180.

Resolving status for data resource group entries


The Work with Data Resource Group Entries display lists the data resource groups
associated with an application group. Each entry identifies a data resource group and
the summary of the replication status from its associated data groups.

Figure 5. The Work with Data Resource Group Entries display for an application group that
does not participate in a cluster.

Work with Data Rsc. Grp. Ent.


System: SYSA
Application group . . . . . : SAMPLEAG
Virtual switch status . . . : *TESTING

Type options, press Enter.


1=Add 2=Change 4=Remove 5=Display 6=Print 8=Data groups 9=Start
10=End 12=Node entries 14=Build environment 15=Switch

Resource
Resource Group Node Replication
Opt Group Type Status Status Status
__ __________
__ AGRSGRP *DTA *ACTIVE

Bottom
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F6=Add F9=Retrieve F12=Cancel
F13=Repeat F18=Subset F19=Load F21=Print list

A Virtual switch status field can appear below the Application group field when a
virtual switch is in progress for the application group (*NONCLU only). The field
identifies the phase of virtual switch activity.
Resource Group Status - This column identifies the status of the data resource
group. In environments that do not include IBM i clustering, this column is always
blank.

88
Resolving status for data resource group entries

Node Status - This column identifies the status of the nodes for the data resource
group. In environments that do not include IBM i clustering, this column is always
blank
Replication Status - The value in this column is a summary status of data replication
activity for the data resource group. The status includes the status of all data group
processes, replication direction, replicated object and file entries, audits, and
recoveries.
Figure 17 identifies status values for replication that require user action.

Table 17. Replication Status values that require user action.

Status Value Description and Action

*ATTN One or more of the following problems exist for data groups within the
data resource group.
• The source system of an active data group is not the primary node of
its application group.
• A data group has a failed state, a process other than apply is inactive
during a virtual switch, an error condition, audit errors, or pending
recoveries.
To prevent damage to data in your environment, it is important that
you begin by determining which system should be the source for
the data groups. Do the following:
1. From this display, use option 8 (Data groups) to check which system
is the current source for the data group.
2. Determine which node has the role of current primary for the
application group. From the Work with Application Groups display,
use option 12 (Node entries), then check the current node role.
If the current primary node is correct and a data group with an incorrect
source system is active, end the data group and contact CustomerCare.
If the data groups in question have the correct source system but the
primary node for the application group is not correct, you need to
change the recovery domain for the application group to make the
correct node become primary. Use “Changing the sequence of backup
nodes” on page 91.
After you have ensured that the data groups have the correct
source system, resolve any error conditions reported on the Work
with Data Groups display.
Note: Not all data groups should necessarily be active. Only the data groups
currently being used for data replication should be active. You will need
to look at the current node roles and data providers for the node entries
to determine which data groups should be active.

*INACTIVE All replication in the data resource group is inactive. This may be normal
if replication was ended to perform certain activities. Use option 8 (Data
groups) to access the Work with Data Groups display.

89
Verifying the sequence of the recovery domain
Ensuring that sequence of the current backup nodes is set properly is critical to a
successful and predictable switch process. The current sequence of backup nodes
should match your recovery guidelines.
Do the following to confirm the sequence of the current backup nodes before
performing a switch and before removing or restoring a backup node from the cluster.
1. From the MIMIX Intermediate Main Menu, type 5 (Work with application groups)
and press Enter.
2. From the Work with Application Groups display, type 12 (Node entries) next to the
application group you want and press Enter.
3. The Work with Node Entries display appears, showing current information for the
nodes. Confirm that the current backup nodes have the sequence order that you
expect.
Note: It is important that you are viewing current information on the status view
of the display. Figure 6 shows an example of how the resulting Work with
Node Entries display appears with current status information. If you see
configured information instead, press F10 (View status).
4. If you need to change the sequence of current backup nodes, use “Changing the
sequence of backup nodes” on page 91.

Figure 6. Example of displaying the current sequence information for backup nodes

Work with Node Entries


System: NODED
Application group . . . . . : APP1

Type options, press Enter.


1=Add 2=Change 4=Remove 5=Display 6=Print 9=Start 10=End

-------------Current------------- Manager
Opt Node Role Sequence Data Provider Status
__ ________
__ NODEA *PRIMARY *NONE *ACTIVE
__ NODEB *BACKUP 1 NODEA *ACTIVE
__ NODEC *BACKUP 2 NODEA *ACTIVE
__ NODED *BACKUP 3 NODEA *ACTIVE

Bottom
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F6=Add F7=Systems F9=Retrieve
F10=View config F11=Sort by node F12=Cancel F18=Subset F24=More keys

90
Changing the sequence of backup nodes

Changing the sequence of backup nodes


Use this procedure if you need to change the sequence of the current backup nodes.
This procedure may change the configured sequence for multiple nodes so that you
can achieve the desired sequence for backup nodes. The changes are not effective
until Step 5 is performed.
Do the following from an active application group:
1. From the Work with Application Groups display, type 12 (Node entries) next to the
application group you want and press Enter.
2. The Work with Node Entries display appears. Using F10 to toggle between
configuration view and status views, confirm that the node with the configured role
of *PRIMARY is the same node that is shown as the current *PRIMARY role.
• If the same node is identified as *PRIMARY for the current role and the
configured role, skip to Step 4.
• If the configured *PRIMARY node does not match the current *PRIMARY
node, perform Step 3 to correct this situation before making any changes to
the configured sequence of backup nodes.
Figure 7 is an example of how configuration information appears on the Work with
Node Entries display.
3. Perform this step only if you need to correct the configured primary node to match
the current primary node. This step will demote the configured primary node to a
backup, then promote the correct node to become the configured primary node.
Do the following:
a. From the configuration view of the Work with Node Entries display, type 2
(Change) next to the configured primary node and press Enter.
b. On the Change Node Entry (CHGNODE) display, specify *BACKUP for Role
and press Enter. Then specify *FIRST for List position and press Enter.
c. On the Work with Node Entries display, press F5 (Refresh) to view changes.
All nodes in the configured view should have *BACKUP roles.
d. If necessary toggle to the status view to confirm which node is the current
primary node. Type 2 (Change) next to the current primary node and press
Enter.
e. On the Change Node Entry (CHGNODE) display, specify *PRIMARY for Role
and press Enter. Then press Enter two more times.
f. On the Work with Node Entries display, press F5 (Refresh) to view changes.
You should see the correct node as the configured primary node.
Note: The numbering for the backup sequence may not update; however, the
relative order for the configured backup sequence remains unchanged.
Gaps in configured sequence numbers are ignored when switching to a
backup. As long as the relative order is correct, it is not necessary to
change the configured sequence of backup nodes just to remove gaps
in numbering.

91
g. If the configured backup sequence is what you expect, skip to Step 5 to make
the change effective.
4. To change the sequence of backup nodes, do the following:
a. From the configured view of the Work with Node Entries display, type 2
(Change) next to the backup node whose sequence you want to change.
b. On the Change Node Entry (CHGNODE) display, specify *BACKUP for Role
and press Enter. Then specify either *FIRST or a number for List position and
press Enter.
Note: If you specify a number, it cannot already be used in the configured
sequence list.
c. On the Work with Node Entries display, press F5 (Refresh) to view changes.
d. Repeat Step 4 until the correct sequence is shown on the configuration view.
Note: Gaps in configured sequence numbers are ignored when switching to a
backup. For example, in a configuration with two backup nodes, there is
no operational difference between a backup sequence of 1, 2 and a
backup sequence of 2, 5 as long as the same nodes are specified in the
same relative order.
5. To make the changes to the backup order effective, do the following:
a. Press F12 (Cancel) to return to the Work with Application Groups display.
b. Type 9 (Start) next to the application group you want and press F4 (Prompt).
c. On the Start Application Group (STRAG) display, specify *CONFIG for Current
node roles and press Enter.
d. The Procedure prompt appears. If needed, specify a different value and then
press Enter.
6. Confirm that the node entries have changed. Type 12 (Node entries) next to the
application group and press Enter. If necessary, use F10 to access the status
view. The current backup nodes should be in the new order.

92
Changing the sequence of backup nodes

Figure 7. Example of displaying the configured sequence information for backup nodes:

Work with Node Entries


System: NODED
Application group . . . . . : APP1

Type options, press Enter.


1=Add 2=Change 4=Remove 5=Display 6=Print 9=Start 10=End

-----------Configured------------
Opt Node Role Sequence Data Provider
__ ________
__ NODEA *PRIMARY *PRIMARY
__ NODEB *BACKUP 1 *PRIMARY
__ NODEC *BACKUP 4 *PRIMARY
__ NODED *BACKUP 5 *PRIMARY

Bottom
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F6=Add F7=Systems F9=Retrieve
F10=View status F11=Sort by node F12=Cancel F18=Subset F24=More keys

Examples of changing the backup sequence


The following examples illustrate problems with the current backup sequence and
how to correct them.
Example 1 - Changing the backup sequence when primary node is ok
Table 18 shows a four-node environment where the current backup sequence does
not reflect the desired behavior in the event of a switch. Also, the relative order of the

93
configured backup sequence does not match the relative order of either the current
sequence or the desired sequence.

Table 18. Example 1, showing discrepancies in backup sequences

Desired Order Initial Values, Example 1


Work with Node Entries

Status View

-----------Current---------------
Opt Node Role Sequence Data Provider
__ ________
__ NODEA *PRIMARY *PRIMARY
__ NODEB *BACKUP 1 *PRIMARY
__ NODED *BACKUP 2 *PRIMARY
__ NODEC *BACKUP 3 *PRIMARY

Configured View

-----------Configured------------
Opt Node Role Sequence Data Provider
__ ________
__ NODEA *PRIMARY *PRIMARY
__ NODEC *BACKUP 1 *PRIMARY
__ NODEB *BACKUP 2 *PRIMARY
__ NODED *BACKUP 3 *PRIMARY

Each row in Table 19 shows a change to be made to the nodes on the configured
view of the Work with Node Entries display. The rows are in the order that the
changes need to occur to correct this example configuration to the desired order.

Table 19. Order in which to change nodes to achieve the desired configuration for example 1

Node to Change To Effect on Configured Order, Example 1 Notes


Change

NODEB Role = *BACKUP Configured View Intermediate step


Position = *FIRST
-----------Configured------------
Opt Node Role Sequence Data Provider
__ ________
__ NODEA *PRIMARY *PRIMARY
__ NODEB *BACKUP 1 *PRIMARY
__ NODEC *BACKUP 2 *PRIMARY
__ NODED *BACKUP 3 *PRIMARY

94
Changing the sequence of backup nodes

Table 19. Order in which to change nodes to achieve the desired configuration for example 1

Node to Change To Effect on Configured Order, Example 1 Notes


Change

NODED Role = *BACKUP Configured View Desired


Position = *FIRST configuration but it
-----------Configured------------ is not effective
Opt Node Role Sequence Data Provider until STRAG
__ ________
__ NODEA *PRIMARY *PRIMARY
ROLE (*CONFIG)
__ NODED *BACKUP 1 *PRIMARY is performed.
__ NODEB *BACKUP 2 *PRIMARY
__ NODEC *BACKUP 3 *PRIMARY

Example 2 - Correcting the configured primary node and changing the backup
sequence
Table 20 shows a four-node environment where the current backup sequence does
not reflect the desired behavior in the event of a switch. Also, the current and
configured primary node do not match. The configured primary node must be
corrected first, before attempting to correct any backup node sequence problems.

Table 20. Example 2, showing discrepancies in primary node and backup sequences

Desired Order Initial Values, Example 2


Work with Node Entries

Status View

-----------Current---------------
Opt Node Role Sequence Data Provider
__ ________
__ NODEA *PRIMARY *PRIMARY
__ NODEB *BACKUP 1 *PRIMARY
__ NODED *BACKUP 2 *PRIMARY
__ NODEC *BACKUP 3 *PRIMARY

Configured View

-----------Configured------------
Opt Node Role Sequence Data Provider
__ ________
__ NODEB *PRIMARY *PRIMARY
__ NODEC *BACKUP 1 *PRIMARY
__ NODEA *BACKUP 2 *PRIMARY
__ NODED *BACKUP 3 *PRIMARY

95
Each row in Table 21 shows a change to be made to the nodes on the configured
view of the Work with Node Entries display. The rows are in the order that the
changes need to occur to correct this example configuration to the desired order.

Table 21. Order in which to change nodes to achieve the desired configuration for example 2.

Node to Change To Effect on Configured Order, Example 2 Notes


Change

NODEB Role = *BACKUP Configured View Intermediate step


Position = *FIRST
-----------Configured------------
Opt Node Role Sequence Data Provider
__ ________
__ NODEB *BACKUP 1 *PRIMARY
__ NODEC *BACKUP 2 *PRIMARY
__ NODEA *BACKUP 3 *PRIMARY
__ NODED *BACKUP 4 *PRIMARY

NODEA Role = Configured View Intermediate step,


*PRIMARY corrects configured
-----------Configured------------ *PRIMARY.
Opt Node Role Sequence Data Provider
__ ________
The sequence
__ NODEA *PRIMARY *PRIMARY number for Backup
__ NODEB *BACKUP 1 *PRIMARY 3 may appear as 4.
__ NODEC *BACKUP 2 *PRIMARY The relative order is
__ NODED *BACKUP 3 *PRIMARY equivalent.

NODED Role = *BACKUP Configured View Desired


Position = *FIRST configuration but it
-----------Configured------------ is not effective until
Opt Node Role Sequence Data Provider STRAG ROLE
__ ________
__ NODEA *PRIMARY *PRIMARY
(*CONFIG) is
__ NODED *BACKUP 1 *PRIMARY performed.
__ NODEB *BACKUP 2 *PRIMARY
__ NODEC *BACKUP 3 *PRIMARY

96
CHAPTER 6 Working with procedures and steps

This chapter describes how to work with procedures and steps. Procedures are used
to perform operations associated with an application group or node. A set of default
procedures is shipped with MIMIX which provides the ability to start, end, perform pre-
check activity for switching, switch application groups, and the ability to run data
protection reports for systems. Each procedure consists of a sequence of steps and
may use multiple jobs which perform specific sub-tasks of the larger operation. Steps
also identify runtime attributes that determine how the procedure will start processing
the step and how it will respond if the step ends in error.
It is important to understand how multiple jobs are used to process steps for a
procedure. A procedure uses multiple asynchronous jobs to run the programs
identified within its steps. Starting a procedure starts one job for the application group
or node. For procedures associated with application groups, an additional job is
started for each of its data resource groups. These jobs operate independently and
persist until the procedure ends. Each persistent job evaluates each step in sequence
for work to be performed within the scope of the step program type. When a job for a
data resource group encounters a step that acts on data groups, it spawns an
additional job for each of its associated data groups. Each spawned data group job
performs the work for that data group and then ends.
This chapter contains the following topics:
• “Displaying status of procedures” on page 98 describes how to display the status
of the most recent runs of procedures for an application group, the conditions
which cause each procedure status value, and the actions required to resolve
problem statuses.
• “Responding to a procedure in *MSGW status” on page 101 describes how to
resolve procedure inquiry messages.
• “Resolving a *FAILED or *CANCELED procedure status” on page 102 describes
how to address procedures with these error statuses.
• “Displaying status of steps within a procedure run” on page 103 describes how to
display status of steps within a procedure as well as the differences between the
collapsed and expanded views of the Work with Step Status display.
• “Resolving problems with step status” on page 106 describes the conditions which
cause each step status value and the actions required to resolve problem
statuses. This includes how to resolve step inquiry messages and failed or
canceled steps.
• “Acknowledging a procedure” on page 110 describes how to manually change a
procedure with a status of *CANCELED, *FAILED, or *COMPERR to an
acknowledged status.
• “Running a procedure” on page 111 describes how to start a user or node
procedure and the parameter that controls the step at which the procedure
begins.

97
Working with procedures and steps

• “Canceling a procedure” on page 114 describes how to cancel an active


procedure.
• “Displaying scheduling information for automatic reports” on page 114 describes
how to access a view which shows the next scheduled run of data protection
reports on a specific system.
• “Displaying available status history of procedure runs” on page 115 describes how
to display status for previously completed procedure runs that have been kept for
their status history.

Displaying status of procedures


You can view the status of runs of procedures from the Work with Procedure Status
display. The term “the last run” of a procedure refers to the most recently started run
of a procedure, which may be in progress or may have completed.
To display the status of the most recent run of all procedures, do the following:
1. From the MIMIX Basic Main Menu, select option 1 (Work with application groups).
2. From the Work with Application Groups display, press F20 (Procedure Status).
The Work with Procedure Status display appears.
The most recent run of all procedures are listed on the Work with Procedure
Status display. Procedures of type *NODE are listed for each node in the instance
and have a status of *NEW if they have not run. Other procedure types that have
never been requested to run are not listed.
3. Press F18 (Subset) if you want to display only procedures of a specific type.
4. Locate the procedure you want and check the value of the Status column.
Possible status values and the actions to take to resolve problems are listed in
Table 22.

98
Displaying status of procedures

Figure 8 shows an example of the Work with Procedure Status display subsetted to
show only runs of a specific procedure and application group. The list is in reverse
chronological order so that the most recently started procedures are at the top.

Figure 8. A subsetted view of the Work with Procedure Status display.

Work with Procedure Status


System: SYSTEMA
Type options, press Enter.
5=Display 6=Print 8=Step status 9=Run 11=Display message
12=Cancel 13=Change status 14=Resume 37=Change schedule

Opt Procedure Type App Group Status ---Start Time----


or Node
__ SWTPLAN *SWTPLAN SAMPLEAG *COMPLETED 03/01/10 11:25:05
__ SWTPLAN *SWTPLAN SAMPLEAG *COMPLETED 03/01/10 11:04:58

Bottom
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F9=Retrieve F11=Duration F12=Cancel
F13=Repeat F18=Subset F21=Print list

A second view, F11 (Duration), shows columns for the Duration of the procedure and
the Node on which the procedure was started. Procedures of type *NODE that have
not run are blank for these columns and have a status of *NEW.
From the second view, F11 (Schedule summary) shows columns for the Frequency
and Scheduled time for procedures of type *NODE that are data protection reports.
The frequency indicates how often the report is automatically submitted to be run by
MIMIX. The scheduled time identifies when MIMIX will submit the next scheduled run.
Only data protection report procedures can be scheduled.
Timestamps are in the local job time. If you have not already ensured that the
systems in your installation use coordinated universal time (UTC), see “Setting the
system time zone and time” on page 322.
Table 22 identifies the possible status values that can appear on the Work with
Procedure Status display and identifies the action to take to resolve reported
problems.

Table 22. Procedure status values with action required

Category Status Value Description and Action Required


Note: For problem resolution, a virtual switch procedure differs from other types of procedures in several ways. For
more information, see “Performing a virtual switch” on page 323.

99
Working with procedures and steps

Table 22. Procedure status values with action required

Category Status Value Description and Action Required

Active *ACTIVE The procedure is currently running. No steps require attention.

*ATTN The procedure requires attention. Either there is a step with a status of
*MSGW, or there is an active step and one or more steps with step
status values of *ATTN, *CANCEL, *FAILED, or *IGNERR.
Action Required: Determine the status of each step and the action
required to correct that status. See “Resolving problems with step
status” on page 106.

*MSGW A step within the procedure is waiting for a response to an inquiry


message. The procedure cannot process the step or any subsequent
steps without a reply to the message.
Action Required: Display and respond to the inquiry message using
“Responding to a procedure in *MSGW status” on page 101.

*PENDCNL A request to cancel the procedure is in progress. When the activity for
the steps in progress at the time of the cancel request ends, the
procedure status changes to *CANCELED.

*QUEUED A request to run the procedure is currently waiting on the job queue.
When the procedure becomes an active job, the procedure status
changes to *ACTIVE.

Resumable *CANCELED Either the procedure was canceled and did not complete, or steps within
the procedure were canceled as a response to inquiry messages from
the steps. The procedure was partially performed.
Action Required: Use “Resolving a *FAILED or *CANCELED
procedure status” on page 102 to determine the state of your
environment and whether to resume the procedure or to acknowledge
its status.

*FAILED The procedure failed. Jobs for one or more steps had errors. Those
steps were configured to end if they failed. The procedure was partially
performed.
Action Required: Use “Resolving a *FAILED or *CANCELED
procedure status” on page 102 to determine the state of your
environment and whether to resume the procedure or to acknowledge
its status.

Acknowledged *ACKCANCEL The procedure was canceled and a user action acknowledged the
cancellation so that the procedure can no longer be resumed.

*ACKFAILED The procedure failed and a user action acknowledged the failure so that
the procedure can no longer be resumed.

*ACKERR The procedure completed with errors and a user action acknowledged
the procedure. It is assumed that the user reviewed the steps with
errors. A status of completed with errors is only possible when the steps
with errors had been configured (within the procedure) to ignore errors
or a user’s response to a step in message wait status was to ignore the
error and continue running the procedure. After the step is
acknowledged, the procedure status changes to *ACKERR.

100
Responding to a procedure in *MSGW status

Table 22. Procedure status values with action required

Category Status Value Description and Action Required

Completed *COMPERR The procedure completed with errors. One or more steps had errors and
were configured to continue processing after an error.
Action Recommended: Investigate the cause of the error and assess
its implications.

*COMPLETED The procedure completed successfully.

New *NEW The *NODE type procedure is new and has not yet been run for that
node.
Action: Use option 9 (Run) to run the procedure.

Responding to a procedure in *MSGW status


A procedure in *MSGW status is effectively paused at a known point in its processing
as a result of a runtime attribute on one of its steps. The procedure sent an inquiry
message because a step specified *MSGW for its Action before step (BEFOREACT)
attribute. All jobs for the procedure have completed processing all previous steps and
are waiting to run the step’s program. An operator response is required.
A virtual switch procedure (type *SWTVRT) is intended to remain in *MSGW status for
step “Start recovery processing of virtual switch” (MXVRTRCY) the entire time that
application tests are being performed on the designated test node. Do not respond to
the message for this step until testing is complete and you are ready to start the
recovery phase of the virtual switch. If the virtual switch procedure has a *MSGW
status for any other step, handle the message like you would for any other procedure.
To respond to a procedure in *MSGW status, do the following from the Work with
Procedure Status display:
1. To see which step is waiting, type 8 (Step status) next to the procedure and press
Enter.
2. The Work with Step Status display appears. The information on this display can be
used to determine which step is waiting to start. You will see steps with values of
*COMP, *IGNERR, or *DSBLD followed by no status for all remaining steps. The
first step with no status is the step that is waiting to start. Based on that step,
determine how to respond to the message and whether you are ready to respond.
3. You cannot display or respond to the procedure message from the Work with Step
Status display. Press F12 to return to the Work with Procedure Status display.
4. Type 11 (Display message) next to the procedure in *MSGW status and press
Enter.
5. You will see the message “Procedure name for application group name requires
response. (G C).”
Note: Procedures of type *NODE will have *NONE for name.

101
Working with procedures and steps

Do one of the following:


• A response of G (Go) is required to start processing the step. Type G and press
Enter.
• A response of C (Cancel) will cancel the procedure. Type C and press Enter.

Resolving a *FAILED or *CANCELED procedure status


When a procedure fails or is canceled, subsequent attempts to run the same
procedure will fail until user action is taken. You need to determine the best course of
action for your environment based on the implications of the partially performed
procedure. This topic will assist you in evaluating the cause of the failure or
cancellation, as well as the state of other steps within the procedure.
Important! Steps with failed or canceled jobs need to be resolved. Other
asynchronous jobs may have successfully processed the same step and
continued on to process other subsequent steps before the procedure ended. The
actions taken by those steps as well as by completed steps which preceded the
problem are not reversed. Some steps may not have been processed at all.
Do the following from the Work with Procedure Status display:
1. Type 8 (Step status) next to the *FAILED or *CANCELED run of the procedure
and press Enter.
2. The Work with Step Status display appears. Look for steps with a status of
*CANCEL, *FAILED, or *ATTN. Also, for procedures other than type *NODE, use
F7 (Expand) to see detailed status for the jobs which processed the steps.
A procedure with *FAILED status did not complete due to errors. In the collapsed
status view, one or more steps will have a status *ATTN or *FAILED. Other jobs
may have processed subsequent steps before the procedure ended. In the
expanded view, look for one or more jobs with a status of *FAILED. For detailed
information use “Resolving problems with step status” on page 106.
A procedure with *CANCELED status did not complete due to user action. Any of
the following may have occurred:
• A user canceled an inquiry message sent by the procedure because a step
was configured to wait for a reply before starting. This scenario is identified by
the absence of steps with status values of *FAILED, *CANCEL, or *ATTN.
Instead, you will see steps with values of *COMP, *IGNERR, or *DSBLD
followed by no status for all remaining steps. The first step with no status is the
step that waited to start. Continue with step Step 3.
• A user canceled an inquiry message sent by a step which had a job that ended
in error. At least one step in the collapsed view will have a status of *ATTN or
*CANCEL. One or more steps will have job with a status of *CANCEL in the
expanded view. Other jobs may have processed subsequent steps before the
procedure ended. For detailed information use “Resolving problems with step
status” on page 106.
• A user canceled the procedure by using option 12 (Cancel) from the Work with
Procedure Status display or by using the Cancel Procedure (CNLPROC)

102
Displaying status of steps within a procedure run

command. Steps in the collapsed view could have any status except *ACTIVE
or *MSGW. Determine if there are any jobs with status values of *FAILED or
*CANCEL in the expanded view. Other jobs may have processed subsequent
steps before the procedure ended. For detailed information use “Resolving
problems with step status” on page 106.
3. After you have completed your evaluation and have taken any needed corrective
action to resolve why jobs failed or were canceled, determine how to best
complete the procedure. Choices are:
• Resume the procedure. If you resume a failed procedure, processing will begin
with the step that failed. If you resume a canceled procedure, processing will
begin with steps following the canceled step. Optionally, if you were unable to
resolve a problem for a step in error, you can override the attributes of that step
for when the procedure is resumed. See “Resuming a procedure” on page 112.
Note: If a virtual switch procedure fails or is canceled at any step within the
range performed while replication status for the application group
indicates virtual switch activity (starting, testing, or recovering), the only
choice after taking corrective action is to resume the procedure. In this
scenario, you will not be able to run any other procedure to start, stop,
or switch the affected application group until the virtual switch procedure
is resumed and allowed to run to completion.
• Acknowledge the procedure status. Procedures with a status of *CANCELED
or *FAILED can be acknowledged (set to *ACKCANCEL or *ACKFAILED,
respectively) to indicated you have investigated the problem steps and want to
run the procedure again starting at its first step. This option should only be
used after you have evaluated the effect of activity performed by the
procedure. See “Acknowledging a procedure” on page 110.

Displaying status of steps within a procedure run


The Work with Step Status display provides access to detailed information about
status of steps for a specific run of a procedure for an application group.
Timestamps are in the local job time. If you have not already ensured that the systems
in your installation use coordinated universal time, see the MIMIX Administrator
Reference book for the setting system time topic.
To display step status for a procedure run, do the following:
1. Use one of the following to access the run of the procedure you want:
• “Displaying status of procedures” on page 98
• “Displaying available status history of procedure runs” on page 115
2. From the Work with Procedure Status display, type 8 (Step status) next to the run
of the procedure you want and press Enter.
3. For procedures other than type *NODE, press F7 (Expand) to view status of the
individual jobs used to process each step.

103
Working with procedures and steps

The steps listed on the Work with Step Status display appear in sequence number
order as defined by steps in the procedure. If the procedure is in progress, the display
shows status for the steps that have run, the start time and status of the step that is in
progress, and blank status and start time for steps that have not yet run.
Collapsed view - There are multiple views of the Work with Step Status display. The
view shown depends on the type of procedure.
Procedures other than *NODE - Figure 9 shows the collapsed view of for
procedures other than type *NODE. Each step of the procedure is shown as a
single row and step status represents the summary of all jobs used by the step. F7
(Expand) provides a view that shows details for additional jobs used to process
each step.
*NODE procedures - Figure 10 shows the display for *NODE procedures, which
only shows summary rows for steps. All steps of a node procedure run on the
node where the procedure started (Started on node field).

Figure 9. Collapsed view of Work with Step Status display for procedures other than
*NODE.

Work with Step Status


System: SYSTEMA
Procedure: SWTPLAN App. group: SAMPLEAG Type: *SWTPLAN
Procedure status: *COMPLETED Start time: 03/01/14 11:04:58

Type options, press Enter.


5=Display 6=Print 8=Work with job 11=Display message
13=Override step

Step Node Start Jobs


Opt Program Type Type Time Duration Status Pend
__ MXCHKCOM *AGDFN *LOCAL 11:05:00 00:00:01 *COMP *NO
__ MXCHKCFG *DGDFN *NEWPRIM 11:05:00 00:00:01 *COMP *NO
__ ENDUSRAPP *AGDFN *PRIMARY 11:05:00 00:00:03 *COMP *NO
__ MXENDDG *DGDFN *NEWPRIM 11:05:01 00:00:05 *COMP *NO
__ MXENDRJLNK *DGDFN *NEWPRIM 11:05:16 00:00:01 *COMP *NO
__ MXAUDACT *DGDFN *NEWPRIM 11:05:18 00:00:01 *COMP *NO
__ MXAUDCMPLY *DGDFN *NEWPRIM 11:05:19 00:00:01 *COMP *NO
__ MXAUDDIFF *DGDFN *NEWPRIM 11:06:10 00:00:54 *COMP *NO
More...
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F7=Expand F9=Retrieve F12=Cancel
F13=Repeat F14=Resume proc. F15=Cancel proc. F18=Subset F24=More keys

104
Displaying status of steps within a procedure run

Figure 10. Work with Step Status display for *NODE procedures.

Work with Step Status


System: SYSTEMA
Procedure: CRTDPRLIB Type: *NODE Started on node: SYSTEMA
Procedure status: *COMPLETED Start time: 03/01/14 11:04:58

Type options, press Enter.


5=Display 6=Print 8=Work with job 11=Display message
13=Override step
Step Start Jobs
Opt Program Time Duration Status Pend
__ STEP1 11:05:00 00:00:01 *COMP *NO
__ STEP2 11:06:00 00:01:03 *ATTN *NO
__ STEP3 11:07:01 00:10:05 *COMP *NO
__ STEP4 *DSBLD
__ STEP5 11:09:18 00:05:01 *COMP *NO

More...
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F9=Retrieve F12=Cancel
F13=Repeat F14=Resume proc. F15=Cancel proc. F18=Subset F24=More keys

Expanded view - Figure 11 shows an example of the expanded view available for
procedures other than type *NODE. In the expanded view, step programs of type
*AGDFN will have one row for each node on which the step runs. Steps which run
step programs at the level of the data resource group or data group are expanded to
have multiple rows so that the status of the step for each data resource group or data
group is visible. For step programs of type *DTARSCGRP, there will be a summary
row for the application group followed by a row for each data resource group within
the application group. For step programs of type *DGDFN, there will be a summary
row for the application group, then for each data resource group, there is a summary
row for the data resource group followed by a row for each of its data groups.
Summary rows are identified by a dash (-) in the columns that are being summarized.

105
Working with procedures and steps

Also, for step programs of type *AGDFN, the Data Rsc. Grp. column and the Data
Group column will always be blank. For step programs of type *DTARSCGRP, the
Data Group column will always be blank.

Figure 11. Expanded view of the Work with Step Status display.

Work with Step Status


System: SYSTEMA
Procedure: SWTPLAN App. group: SAMPLEAG Type: *SWTPLAN
Procedure status: *COMPLETED Start time: 03/01/14 11:04:58

Type options, press Enter.


5=Display 6=Print 8=Work with job 11=Display message
13=Override step

Step Data Data Start


Opt Program Rsc. Grp. Group Node Time Duration Status
__ MXCHKCOM LTIAS01 11:05:00 00:00:01 *COMP
__ MXCHKCFG - - LTIAS02 11:05:00 00:00:01 *COMP
__ MXCHKCFG DRG1 - LTIAS02 11:05:00 00:00:01 *COMP
__ MXCHKCFG DRG1 DG1A LTIAS02 11:05:00 00:00:01 *COMP
__ MXCHKCFG DRG1 DG1B LTIAS02 11:05:00 00:00:01 *COMP
__ MXCHKCFG DRG1 DG1C LTIAS02 11:05:00 00:00:01 *COMP
__ MXCHKCFG DRG2 - LTIAS02 11:05:00 00:00:01 *COMP
__ MXCHKCFG DRG2 DG2A LTIAS02 11:05:00 00:00:01 *COMP
More...
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F7=Expand F9=Retrieve F12=Cancel
F13=Repeat F15=Cancel proc. F18=Subset F24=More keys

Resolving problems with step status


When working with step status, it is important that you understand how multiple jobs
are used to process the steps in a procedure. For procedures other than type *NODE,
at any given time, job activity may be in progress for multiple steps. Or, one job may
have failed processing a step while other jobs may have already processed that step
and continued beyond it. Node procedures only have one job.
Important! Before you take action to resolve a problem with status for a step, be
sure you understand the current state of your environment as a result of
completed steps and steps in progress, as well as the effect of any action you
take.
Table 23 identifies the possible status values that can appear on the Work with Step
Status display and the action to take to resolve reported problems.

Table 23. Step status values with action required

Status Description and Action Required


Value

blank The procedure has started but processing has not yet started for the step.

106
Resolving problems with step status

Table 23. Step status values with action required

Status Description and Action Required


Value

*ATTN The step requires attention. The value *ATTN can only appear in the
collapsed view or on a summary row in the expanded view. If the procedure
status is considered active, at least one job submitted by this step has a
status of *FAILED, *CANCEL or *MSGW. If the procedure status is *FAILED
or *CANCELED, this step has at least one job that has not started or has a
status of *CANCEL or *FAILED.
Action Required: Use F7 to see the expanded view. Determine the specific
data resource group or data group for which the problem status exists. Then
address the status indicated for that job.

*ACTIVE The step is currently running.

*COMP The step has successfully completed.

*DSBLD The step has been disabled and did not run.

*CANCEL One or more jobs used by the step ended in error. In the expanded view of
or status, the job is identified as *CANCEL or *FAILED. The status is due to the
*FAILED error action specified for the step.
• For *CANCEL status, user action canceled the step. The step ran, ended
in error, and issued an inquiry message. The user’s response to the
message was Cancel.
• For *FAILED status, the step ran, one or more jobs ended in error. The
Action on error attribute specified to quit the job.
The type of step program used by the step determines what happens to other
jobs for the step and whether subsequent steps are prevented from starting,
as follows:
• If the step program is of type *DGDFN, jobs that are processing other data
groups within the same data resource group continue. When they
complete, the data resource group job ends. Subsequent steps that apply
to that data resource group or its data groups will not be started. However,
subsequent steps will still be processed for other data resource groups
and their data groups.
• If the step program is of type *DTARSCGRP, subsequent steps that apply
to that data resource group or its data groups will not be started. Jobs for
other data resource groups may still be running and will process
subsequent steps that apply to their data resource groups and data
groups.
• If the step program is of type *AGDFN, subsequent steps that apply to the
application group will not be started. Jobs for data resource group or data
group steps may still be running and will process subsequent steps that
apply to their data resource groups and data groups.
• If the step program is of type *NODE, subsequent steps will not be started.
When all asynchronous jobs for the procedure finish, the procedure status is
set to *CANCELED or *FAILED, accordingly. If both canceled and failed
steps exist when the procedure ends, the procedure status will be *FAILED.
Action Required: Determine the cause of the problem using “Resolving
*CANCEL or *FAILED step statuses” on page 109.

107
Working with procedures and steps

Table 23. Step status values with action required

Status Description and Action Required


Value

*IGNERR The step ran and an error occurred, but processing ignored the error and
continued.
Action Recommended: Use option 8 (Work with job) to determine the cause
of the failure. Consider whether any changes are needed for the procedure,
step, or operating environment to prevent this error from occurring again.

*MSGW The step ran and issued a message that is waiting to be answered. One or
more jobs for the step ended in error. The step attribute requires that an
operator respond to the message.
Action Required: Determine which job issued the message, investigate the
problem, and then respond to the inquiry message using “Responding to a
step with a *MSGW status” on page 108.

Responding to a step with a *MSGW status


When a step or a job for step has a status of *MSGW, it is the result of an error
condition. An inquiry message was sent because the step specified *MSGW for its
Action on error attribute. An operator response is required before any additional
processing for the job can occur.
To respond to a step in *MSGW status, do the following from the Work with Step
Status display:
1. Locate the step whose job is waiting (*MSGW). If needed, use F7 (Expand) to
view the Expanded view.
2. To view information about what caused the job to end in error, type 8 (Work with
job) next to job with *MSGW status and press Enter.
3. On the Work with Job display, type 10 (Display job log, if active, on job queue, or
pending) and press Enter.
4. The job log is displayed. Use F1 to view details of any of the messages. Find the
error that caused the job to end. You will see the inquiry message in the job log;
however you cannot respond to it from here.
5. Press F12 twice to return to the Work with Step Status display.
6. Type 11 (Display message) next to the step job in *MSGW status and press Enter.
7. You will see the message “Error in step at sequence number number in procedure
name. (R C I).” Do one of the following:
• A response of R (Retry) will retry processing the step program within the same
job. Type R and press Enter.
• A response of C (Cancel) will set the job status to *CANCEL as indicated in the
expanded view of step status. Subsequent steps are handled in the same
manner as if the Action on error has specified the value *QUIT. Type C and
press Enter.

108
Resolving problems with step status

• A response of I (Ignore) will set the job status to *IGNERR as indicated in the
expanded view of step status, and processing continues as if the job had not
ended in error. Type I and press Enter.

Resolving *CANCEL or *FAILED step statuses


Evaluate the cause of the failure or cancellation, as well as the state of other steps
within the procedure. All steps with failed or canceled jobs need to be resolved.
Important! For any step which ended in error, other asynchronous jobs may have
successfully processed the same step and continued on to process other
subsequent steps. The actions taken by those steps as well as by completed
steps which preceded the problem cannot be reversed.
Do the following from the Work with Step Status display:
1. Locate the step whose job has a step status of *CANCEL or *FAILED. If needed,
use F7 (Expand) to view the Expanded view.
2. All steps which have a job that has a step status of *CANCEL or *FAILED must be
evaluated and the cause of the problem must be resolved. To view information
about why a job had an error processing a step, do the following:
a. Type 8 (Work with job) next to the job you want and press Enter.
b. On the Work with Job display, type 4 (Work with spooled file) and press Enter.
c. Display the spooled file for the job and check for the cause of the error.
d. Evaluate whether any immediate action is needed due to the condition which
caused the error. Consider the nature and severity of the error.
3. If the procedure is still active and you need to take corrective action or perform
additional investigation, cancel the procedure using F15 (Cancel proc.). Any steps
that are currently running will complete, then the procedure status is set to
*CANCELED.
4. Check which steps have completed, failed, were canceled, or have not yet started.
Then evaluate the current state of your environment as a result. If needed, take
corrective action that is appropriate for the extent of the errors and the extent to
which steps completed.
Note: It is strongly recommended that you cancel the procedure, if it is active,
before attempting any corrective action.
5. Determine how to best complete the procedure in the current state of your
environment. When the procedure is *FAILED or *CANCELED, your choices are:
• Resume the procedure from the point where the procedure ended. If you
resume a failed procedure, processing will begin with the step that failed. If you
resume a canceled procedure, processing will begin with steps following the
canceled step. Optionally, if you were unable to resolve a problem for a step in
error, you can override the attributes of that step for when the procedure is
resumed. See “Resuming a procedure” on page 112.
Note: If a virtual switch procedure fails or is canceled at any step within the
range performed while replication status for the application group

109
Working with procedures and steps

indicates virtual switch activity (starting, testing, or recovering), the only


choice after taking corrective action is to resume the procedure. In this
scenario, you will not be able to run any other procedure to start, stop,
or switch the affected application group until the virtual switch procedure
is resumed and allowed to run to completion.
• Acknowledge the procedure status allowing the procedure for *CANCELED or
*FAILED to be resumed starting with the first step. This choice indicates you
have investigated the problem steps and want to run the procedure again
starting at its first step. This option should only be used after you have
evaluated the effect of activity performed by the procedure. See
“Acknowledging a procedure” on page 110.

Acknowledging a procedure
Acknowledging a procedure allows you to manually change the status of procedures
that either failed or have errors in order to control where the next attempt to run the
procedure will start. Procedures with a status of *CANCELED, *FAILED, or
*COMPERR can be acknowledged (set to *ACKCANCEL, *ACKFAILED, or
*ACKERR, respectively) to indicated you have investigated the problem steps and
understand the implications.
There is one exception to this for a virtual switch procedure. If a problem occurs in any
step within the range performed while the replication status for the application group
indicates virtual switch activity (starting, testing, or recovering) and the virtual switch
procedure fails or is canceled, you cannot acknowledge the failed or canceled
procedure. The procedure must be resumed.
A procedure of *CANCELED or *FAILED allows you to rerun the procedure from its
first step. Once acknowledged, a procedure with either of these statuses cannot be
resumed from the point where the procedure ended. This is appropriate when you
have determined that your environment will not be harmed if the next attempt to run
starts at the first step.
A *COMPERR procedure that is acknowledged (*ACKERR) can never be resumed
because the procedure completed. By acknowledging a procedure with this status,
you are confirming the problems have been reviewed.
The last run of a procedure with a status of *ACKCANCEL or *ACKFAILED and the
last run of the set of start/end/switch procedures can be returned to their previous
status (*CANCELED or *FAILED, respectively). The next attempt to run the procedure
will resume at the failed or canceled step or at the first step that has not been started.
Note: Acknowledging the last run of a failed or canceled procedure will acknowledge
all previous failed or canceled runs of the procedure.
Important! Before changing status of a procedure, it is important that you evaluate
and understand the effect of the partially performed procedure on your environment.
Changing procedure status does not reverse the actions taken by preceding steps
that completed or the actions performed by other asynchronous jobs which did
complete the same step and then processed subsequent steps. It may not be
appropriate for the next run of the procedure to begin with the first step, for example,

110
Running a procedure

if the failure occurred in a step which synchronizes data or changes states of MIMIX
processes. Likewise, it may not be appropriate to return to the previous status to
resume a procedure run was not recently run.
To change the status of a procedure, do the following:
1. From the Work with Procedure Status display type 13 (Change status) next to the
failed or canceled procedure you want and press Enter.
2. The Change Procedure Status (CHGPROCSTS) display appears. Specify the
value you want for the Status prompt and press Enter.
3. If you specified *ACK in Step 2, the Start time prompt appears with *ALL specified
for Start time (STARTTIME). To acknowledge all previously failed or canceled
runs of the selected procedure, press Enter.

Running a procedure
The procedure type determines what command to use to run the procedure.
Procedure types *USER and *NODE can be run from either the Work with Procedures
or Work with Procedure Status displays using option 9 (Run). *NODE procedures can
only run on the node where the procedure is started. All other procedure types must
be invoked by the application group command associated with the procedure type.
For example a procedure of type *START can only be invoked by the Start Application
Group (STRAG) command.
Multiple uniquely named procedures of type *USER for the same application group
can run at the same time. For procedures of type *END, *START, *SWTPLAN,
*SWTUNPLAN, or *SWTVRT, only one procedure for the application group can run at
a time.
Multiple uniquely named procedures of type *NODE can be run on a specific system
at the same time. Procedures of type *NODE which have the same name can run on
different systems at the same time.
Where should the procedure begin? The value specified for the Begin at step
(STEP) parameter on the request to run the procedure determines the step at which
the procedure will start. The status of the last run of the procedure determines which
values are valid.
The default value, *FIRST, will start the specified procedure at its first step. This value
can be used when the procedure has never been run, when its previous run
completed (*COMPLETED or *COMPERR), or when a user acknowledged the status
of its previous run which failed, was canceled, or completed with errors
(*ACKFAILED, *ACKCANCEL, or *ACKERR respectively).
Other values are for resolving problems with a failed or canceled procedure. When a
procedure fails or is canceled, subsequent attempts to run the same procedure will
fail until user action is taken. You will need to determine the best course of action for
your environment based on the implications of the canceled or failed steps and any
steps which completed.
The value *RESUME will start the last run of the procedure beginning with the step at
which it failed, the step that was canceled in response to an error, or the step

111
Working with procedures and steps

following where the procedure was canceled. The value *RESUME may be
appropriate after you have investigated and resolved the problem which caused the
procedure to end. Optionally, if the problem cannot be resolved and you want to
resume the procedure anyway, you can override the attributes of a step before
resuming the procedure.
The value *OVERRIDE will override the status of all runs of the specified procedure
that did not complete. The *FAILED or *CANCELED status of these procedures are
changed to acknowledged (*ACKFAILED or *ACKCANCEL) and a new run of the
procedure begins at the first step.
.

To run a procedure of type *USER, do the following:


1. From the Work with Procedures or Work with Procedure Status display type 9
(Run) next to the user procedure you want and press F4 (Prompt).
2. Specify the value you want for Begin at step and press Enter.
To run a procedure of type *NODE, do the following from the node where you want
the procedure to run:
1. From the Work with Procedures or Work with Procedure Status display type 9
(Run) next to the node procedure you want and press F4 (Prompt).
2. Specify the value you want for Begin at step and press Enter.
To run a procedure type other than *USER or *NODE, do the following:
From a command line, enter the application group command associated with the
procedure type. For example a procedure of type *START can only be invoked by
the Start Application Group (STRAG) command.
To resume a procedure with a status of *CANCELED or *FAILED, see “Resuming
a procedure” on page 112.

Resuming a procedure
To resume a procedure with a status of *CANCELED or *FAILED, do the following
from the Work with Step Status display:
1. Investigate and resolve problems for steps with errors. See “Resolving problems
with step status” on page 106.
2. Optional: If the problem cannot be resolved, and you want to resume the
procedure anyway, use option 13 (Override step) to change the configured value
of the step for when the procedure is resumed. See “Overriding the attributes of a
step” on page 112 for considerations and more information.
3. For procedures of type *USER or *NODE, use F14 (Resume proc.). For all other
procedure types, from a command line, enter the appropriate application group
command and specify *RESUME as the value for Begin at step (STEP).

Overriding the attributes of a step


The attributes of a step can be overridden to change the configured value of the step
for the current run of the procedure by using the Override Step (OVRSTEP)

112
Running a procedure

command. The attributes determine whether the step is run or actions if the step
errors for the current run of the procedure when it is resumed.
The OVRSTEP command can be used for a procedure that has a status of active
(*ACTIVE, *ATTN, *MSGW, *PENDCNL, or *QUEUED), *CANCELED or *FAILED
and steps that have a status of *CANCEL or *FAILED. The overridden values apply
only for the current run of the procedure when it is resumed.
Note: Regardless of procedure status, attributes cannot be overridden for a required
MIMIX step or any step with a step status of *COMP or *IGNERR.
A procedure with a status of *CANCELED or *FAILED requires user action to resolve
a problem. If the problem cannot be resolved and you want to resume the procedure
anyway, you can use the OVRSTEP command to disable the step in error or specify
the error action to occur when the step is retried.
Important! Overriding the attributes of a step should only be done after you have
considered how rerunning the step impacts your environment. It is important that
you understand the implications for steps which preceded the cancellation or
failure in the last run of the procedure. Processing for steps that completed is not
reversed.
The changes made when using the OVRSTEP command will only apply to the current
run of the procedure. The attributes that can be changed will vary depending on the
statuses of the specified procedure and step. Consider the following:
• When the specified procedure has a status of *NEW, *ACKCANCEL,
*ACKFAILED, *ACKERR, *COMPLETED, or *COMPERR, no attributes can be
overridden on any step in the procedure.
• When the specified procedure has a status that is considered active (*ACTIVE,
*ATTN, *MSGW, *PENDCNL, or *QUEUED), only the Action on error
(ERRACT) can be overridden.
• When the specified procedure has a status that can be resumed (*CANCELED
or *FAILED), the Action before step (BEFOREACT), Action on error
(ERRACT), or State (STATE) can be overridden only on steps that have not yet
run, that failed, or that were canceled.
Do the following from the Work with Step Status display:
1. Locate the step whose attribute you want to override. If needed, use F7 (Expand)
to view the Expanded view.
2. Type 13 (Override step) next to the step you want and press Enter.
3. On the Override Step (OVRSTEP) display, specify the values you want and press
Enter.
4. For procedures of type *USER or *NODE, use F14 (Resume proc.). For all other
procedure types, from a command line, enter the appropriate application group
command and specify *RESUME as the value for Begin at step (STEP). See
“Resuming a procedure” on page 112.

113
Working with procedures and steps

Canceling a procedure
Use this procedure to cancel a procedure with a status that is considered active. This
includes procedure statuses of: *ACTIVE, *ATTN, *MSGW, *PENDCNL, and
*QUEUED.
Important! Use this command with caution. Processing ends without reversing
any actions performed by completed steps, which may leave your environment in
an undesirable state. For example, ending a switch procedure could result in
partially switched data.
If you cancel a virtual switch procedure when it is at any step within the range
performed while the application group’s replication status indicates a virtual switch
is starting, testing, or recovering, the procedure must be resumed. You will not be
able to run any other procedure to start, stop, or switch the affected application
group until the virtual switch procedure is resumed and allowed to run to
completion.
The status of the procedure will be changed to *PENDCNL. If there are any inquiry
messages waiting for an operator response, they are processed as if the response
was Cancel. When all activity for currently running steps end, the status of the
procedure will be automatically changed to *CANCELED.
To cancel an active procedure, do one of the following:
Note: These steps must be done from the node on which the procedure was started.
• From the Work with Procedure Status display, type 12 (Cancel) next to the
procedure you want and press Enter.
• From the Work with Step Status display, press F15 (Cancel proc.).
A procedure that has been canceled can be resumed later, as long as its status has
not been changed to *ACKCANCEL. When a canceled procedure is resumed,
processing begins immediately after the point where it was ended.

Displaying scheduling information for automatic reports


The procedures of type *NODE with the names CRTDPRLIB, CRTDPRFLR, and
CRTDPRDIR are data protection reports that can be submitted to run automatically by
MIMIX. The Schedule summary view of the Work with Procedure Status display
allows you to see scheduling information for data protection reports.
To view when the next scheduled run of a procedure for a data protection report can
occur on a system, do the following:
1. From the MIMIX Intermediate Main Menu, select option 2 (Work with systems)
and press Enter.
2. From the Work with Systems display, select option 21 (Procedure status) next to
the system you want and press Enter.
3. The Work with Procedure Status display appears showing all procedures that can
run on the selected system. Press F11 two (2) times to access the view which
displays the frequency and scheduled time of the next run.

114
Displaying available status history of procedure runs

Only procedures for data protection reports can be scheduled for automatic
submission by MIMIX.

Figure 12. Schedule summary view of the Work with Procedure Status display, showing pro-
cedures for a single node.

Work with Procedure Status


System: SYSTEMA
Type options, press Enter.
5=Display 6=Print 8=Step status 9=Run 11=Display message
12=Cancel 13=Change status 14=Resume 37=Change schedule

App Group
Opt Procedure Type or Node Frequency -Scheduled Time--
__ CRTDPRLIB *NODE SYSTEMA *WEEKLY 02/02/14 03:00:00
__ CRTDPRFLR *NODE SYSTEMA *WEEKLY 02/02/14 03:00:00
__ CRTDPRDIR *NODE SYSTEMA *WEEKLY 02/02/14 03:00:00

Bottom
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F9=Retrieve F11=Start time F12=Cancel
F13=Repeat F18=Subset F21=Print list

Displaying available status history of procedure runs


The status of other previously performed runs of procedures may be available on the
Work with Procedure Status display, subject to the current settings of the Procedure
history retention policy.
To display status of all available runs of a selected procedure, do the following:
1. From the MIMIX Intermediate Main Menu, select the following:
• For *NODE procedures, type option 2 (Work with systems).
• For procedures other than type *NODE, type option 5 (Work with application
groups).
2. Type option 20 (Procedures) and press Enter.
3. The Work with Procedures display appears. Type 14 (Procedure status) next to
the procedure you want and press Enter.
Note: All available runs for the selected procedure are listed on the Work with
Procedure Status display. The last run procedure is at the top of the list.
For *NODE procedures, all nodes are included so the *NODE procedure at
the top of the list may not be the one for the node you want.
4. Locate the run of the procedure you want and check the value of the Status

115
Working with procedures and steps

column.
Note: To view status of all runs of all procedures, you can either press F20
(Procedure status) from the Work with Application Groups display, press F14
(Procedure status) from the Work with Procedures display, or enter the
command: WRKPROCSTS.

116
CHAPTER 7 Working with data group status

This chapter describes common MIMIX operations that help keep your MIMIX
environment running. In order for MIMIX to provide a hot backup of your critical
information, all processes associated with replication must be active at all times.
Supporting service jobs must also be active. MIMIX allows you to display and monitor
the statuses of these processes.
The topics included in this chapter are:
• “Displaying data group status” on page 118 describes how to access the Work
with Data Groups display and identifies where to find information to resolve errors
reported in each column of the display.
• “Displaying detailed status for a data group” on page 126 describes how to access
detailed status for a data group and describes what is shown on the multiple views
of the Data Group Status display.
• “Identifying replication processes with backlogs” on page 137 describes what
fields to check for detailed status of a data group.
• “Data group status in environments with journal cache or journal state” on
page 139 describes the additional information reported within data group status
configured to use MIMIX support for IBM’s High Availability Journal Performance
IBM i option 42, Journal Standby feature and Journal caching. This includes the
topic “Resolving a problem with journal cache or journal state” on page 141.

117
Displaying data group status
From the Work with Data Groups display you can start and end replication, track
replication status, perform a data group switch, as well as work with files, objects, and
tracking entries in error and access displays for data group entries and tracking
entries.
To access the Work with Data Groups display, do one of the following:
• From the MIMIX Basic Main menu, select option 6 (Work with data groups) and
press Enter.
• From the MIMIX Intermediate Main menu, select option 1 (Work with data groups)
and press Enter.

Figure 13. Sample Work with Data Groups display. The display uses letters and colored
highlighting to call your attention to warning and problem conditions. This example shows
items in color which would appear with color highlighting on the display. If you are viewing this
page in printed form, the color may not be shown.

Work with Data Groups CHICAGO


11:02:05
Type options, press Enter. Audits/Recov./Notif.: 001 / 002 / 003
5=Display definition 8=Display status 9=Start DG
10=End DG 12=Files needing attention 13=Objects in error
14=Active objects 15=Planned switch 16=Unplanned switch ...
---------Source--------- --------Target-------- -Errors-
Opt Data Group System Mgr DB Obj DA System Mgr DB Obj DB Obj
__ APP1 LONDON A I CHICAGO A I
__ APP2 LONDON A A A CHICAGO A A A
__ APP3 LONDON A I CHICAGO A I 2
__ CRITICALAP LONDON A R A A CHICAGO A A A 1 4
__ RJAPP4 LONDON A L CHICAGO A I

Bottom
F3=Exit F5=Refresh F7=Audits F8=Recoveries F9=Automatic refresh
F10=Legend F13=Repeat F16=DG definitions F23=More options F23=More keys

For each data group listed, you can see the current source system and target system
processes, and the number of errors reported.
The following fields and columns are available.
Audit/Recov./Notif. -These three fields, located in the upper right corner of the Work
with Data Groups display respectively identify the number of audits that require action
or attention, the number of new or in-progress recoveries, and the number of new
notifications that require action or attention. If more than 999 items exist in any field,
the field will display +++. When highlighted in reverse red, a problem exists that
requires action. When highlighted in reverse yellow, your attention may be needed to
prevent a problem from occurring.

#
Displaying data group status

For audits, the Audit severity policy in effect when audits run determines whether
certain audit statuses are reported as errors (red) or warnings (yellow).
For recoveries, it is important that there are no new or in-progress recoveries when
certain activity, such as ending MIMIX, is performed. You can view a list of the objects
that have new or in-progress recoveries from the Object Correction Activity window
for a data group within Vision Solutions Portal.
Data group - When a data group name is highlighted, a problem exists.
Source - The following columns provide summaries of processes that run on the
source system.
Mgr - Represents a summation of the system manager and the journal manager
processes on the source system of the data group. These processes must be
active (A) for replication to occur.
DB - Represents the status of the remote journal link. It is possible to have an
active status in this column even though the data group has not been started.
When the RJ link is active, database changes will continue to be sent to the target
system. MIMIX can read and apply these changes once the data group is started.
For data groups configured for source-send replication, this represents the status
of the database send process.
Obj - Represents a summation of the object processes that run on the source
system. These include the object send, object retrieve and container send
processes.
DA - This column represents the status of the data area polling process when the
data group replicates data areas through the data area poller. This column does
not contain data when data areas are replicated through the user journal with
advanced journaling or through the system journal.
Target - The following columns provide summaries of processes that run on the target
system.
Mgr - Represents a summation of the system manager, journal manager, and
target journal inspection processes on the target system of the data group. Target
journal inspection status includes status of inspection jobs for both target journals
(user and system) for the data group. Manager processes must be active (A) for
replication to occur.
DB - Represents the summation of status for the database reader process, the
database apply process, and access path maintenance jobs1. For data groups
configured for source-send replication, this column represents the summation of
the status of database apply processes and access path maintenance jobs.
Obj - Represents the object apply processes.
Errors - When any errors are indicated in the following columns (DB and Object), they
are highlighted in red.
DB - Represents the sum of the number for database files, IFS objects, *DTAARA
and *DTAQ objects that are on hold due to errors plus the number of logical (LF)

1. Access path maintenance jobs run only if the access path maintenance (APMNT) policy is
enabled.

104
and physical (PF) files that have access path maintenance1 failures for the data
group. To work with a subsetted list of file errors and access path errors, use
option 12 (Files needing attention). For a subsetted list of IFS object errors, use
option 51 (IFS tracking entries not active), For a subsetted list of *DTAARA and
*DTAQ errors, use option 53 (Object tracking entries not active).
Obj - Represents a count of the number of objects for which at least one activity
entry is in a failed state. To work with a subsetted list, use option 13 (objects in
error).
Use the following topics to resolve problems reported on this display:
• “Resolving auditing problems” on page 162.
• “Displaying recoveries” on page 211.
• “Displaying notifications” on page 205.
• “Identifying and resolving problems reflected in the Data Group column” on
page 120
• For Source and Target columns, see topics:
– “Displaying status of system-level processes” on page 180
– “Identifying and resolving replication problems reflected in the Source and
Target columns” on page 123
• For Error columns, see topics:
– “Working with files needing attention (replication and access path errors)” on
page 263,
– “Working with tracking entries” on page 272,
– “Working with objects in error” on page 278.

Identifying and resolving problems reflected in the Data Group column


When a data group name is highlighted in color, journaling problems exist that affect
replication of one or more types of data.

Table 24. Conditions which highlight the data group name in color.

Color Possible Problems

Red One of the following conditions exists:


• Files, IFS tracking entries, or object tracking entries defined to the data group
are not journaled or are journaled to a different journal than the configured
journal on the source system.
• The source side journal is in standby or inactive state.

#
Displaying data group status

Table 24. Conditions which highlight the data group name in color.

Color Possible Problems

Yellow One of the following conditions exists:


• Files, IFS tracking entries, or object tracking entries defined to the data group
are not journaled or are journaled to a different journal than the configured
journal on the target system. This is only enforced if the data group is set up
to journal on the target system as defined in the data group definition.
• Data group file entries, IFS tracking entries, or object tracking entries are on
hold for reasons other than an error.
• The journal cache value for the source journal does not match the configured
value in the journal definition.
• The journal cache value for the target journal does not match the expected
cache value and the database apply session is active. If another data group
is using the journal definition as a source journal, the actual journal cache
value may be different than the configured value.
• The target journal state value for the target journal does not does not match
the expected state value and the database apply session is active. If another
data group is using the journal definition as a source journal, the actual state
may be different than the configured value.
Note: In a cooperative processing environment, files, IFS tracking entries, or object
tracking entries being added dynamically to the configuration for user journal
replication may reflect an intermediate state of not journaled until they have been
synchronized and become active to MIMIX.

Resolving problems highlighted in the Data Group column


In most environments, the most likely causes indicated in Table 24 are problems with
journaling. Problems associated with journal state or journal cache are only reported
in data groups which are configured to use those high availability journal performance
enhancements.
Entries on hold for other reasons: If the data group name is highlighted in yellow,
do the following to check for entries that are on hold for reasons other than errors:
1. From the Work with Data Groups display, use option 8 (Display status).
2. From the Data Group Status display, press F8 (Database).
3. Check the value of the Held for other reasons field in the upper right corner of the
display. If the value is greater than zero (0), continue with the next step.
4. Press F11 three (3) times to access the File and Tracking Entry Status display.
This view separates the counts of entries held for other reasons by entry type.
5. Do the following to resolve each type of entry that has objects held for other
reasons:
a. Press F12 to return to the Work with Data Groups display.
b. Use options 17 (File entries), 50 (IFS trk entries), and 52 (Obj trk entries) as
needed to access the held entries.
c. Investigate the reason why the entry is held and take appropriate action to
resolve the problem.

104
For additional information about possible actions, see the tables in the following
topics:
• For file entries “Working with user journal replication errors” on page 263
• For IFS and Object tracking entries, see “Working with tracking entries” on
page 272 for additional details.
Journaling problems: If the data group name is highlighted in red or yellow, either
the objects are not being journaled, or they are being journaled to a different journal
than the configured journal.
Do the following to check for and resolve journaling problems:
1. From the Work with Data Groups display, use option 8 (Display status).
2. From the Data Group Status display, press F8 (Database).
3. Check the following:
a. The Jrn State and Cache Src and Tgt fields. These fields are located in the
upper left corner of the Data Group Database Status display. For each system
(Src or Tgt) status of the journal state is shown first, followed by the status of
the journal cache. The example below shows ‘v’ for value in all for status
positions. If any of these fields are highlighted, there is a problem. Use
“Resolving a problem with journal cache or journal state” on page 141.
Jrn State and Cache Src: v v Tgt: v v

b. The Not journaled Src and Tgt fields. These fields are located in the upper right
corner of the Data Group Database Status display. For each system (Src or
Tgt) value, if there is a number other than zero, there is a problem. Either the
objects are not being journaled, or they are being journaled to a different
journal than the configured journal. Continue with these steps to determine the
cause of the problem.
6. Press F11 (View) multiple times to switch the view until the display shows a list of
the objects with journaling problems. Note which objects have problems.
7. Press F12 (Cancel) multiple times to return to the Work with Data Groups display.
8. Access the journaled view to check the journaling status for objects with
problems:
• For files, use option 17 (File entries) for the data group. Then press F10
(journaled view) to see journaling status.
• For IFS tracking entries, use option 50 (IFS tracking entries) for the data
group. Then press F10 (journaled view) to see journaling status.
• For object tracking entries, use option 52 (object tracking entries) for the data
group. Then press F10 (journaled view) to see journaling status.
9. From the journaled view, locate the entries with a journal status (JRN1STS or
JRN2STS) value of *NO or *DIFFJRN.
a. If the journal status value is *NO, continue with these steps.

#
Displaying data group status

b. If the journal status value is *DIFFJRN, further investigation is needed to


determine the reason for journaling to a journal that is different than the
configured journal. This may require input from your system administrator. Use
“Resolving a problem for a journal status of *DIFFJRN” on page 143.
10. Press F12 to return to the Work with Data Groups display.
11. Type a 9 (Start journaling). The appropriate Start Journaling display appears. Do
one of the following:
• To start journaling using the command defaults, press Enter.
• To modify command defaults, press F4 (Prompt). Specify the values you
want and then press Enter to start journaling with those values.
12. Press F5 (Refresh) to update and view the current journaling status.

Identifying and resolving replication problems reflected in the Source


and Target columns
The status of each process is represented by a status letter and the color of the box
surrounding the letter. Table 25 describes the letters and colors used for status of the
replication process summaries shown in the Source and Target columns.

Table 25. Possible status values for source and target process summaries

I Inactive (highlighted red) – The process is currently not active.


When all database and object replication processes are inactive, the data group is
considered active. Use option 9 (Start DG) to start the data group.

L Inactive RJ link (highlighted red) – The RJ link is currently not active. This status is
only displayed in the database source column when a data group uses MIMIX RJ
support.
Check the status of the database and object replication processes.
• If replication processes are active, use option 44 (RJ links) to start the RJ link.
• If replication processes are all inactive, use option 12 (Start DG) to start the data
group.
• If replication processes have other statuses, correct those problems. Then use
option 44 (RJ links) to start the RJ link.

A Active (highlighted blue) – The process is currently active. For the database source
column, this value indicates that the send/receive processes are active.

C RJ Catch-up mode (highlighted blue) – The remote journal is currently in catch-up


mode. This status can only be displayed in the database source column for data
groups that use remote journaling. Catch-up mode indicates that the operating
system is transferring journal entries from the source system journal to the remote
journal as quickly as possible. When the database reader process is active, MIMIX
processes the journal entries as they reach the target system.

R Active RJ link (highlighted blue) – The RJ link is currently active. This status is only
displayed in the database source column when a data group uses MIMIX RJ
support.

104
Table 25. Possible status values for source and target process summaries

U Unknown (highlighted white) – The status of the process cannot be determined


possibly because of an error or communications problem.
To verify communications, do the following:
1. Press the F16 (DG definitions) key.
2. From the Work with Data Group Definitions display, use option 14 (Transfer
definitions) next to the data group you want.
3. From the Work with Transfer Definitions display, use option 11 (Verify
communications link) to verify the communication links available to the data
group.

J RJ Link in Threshold (highlighted turquoise) – The RJ link has fallen behind its
configured threshold.
Use option 8 (Display status) to view detailed status and determine the extent of
the backlog. For more information about which detailed status fields to check, see
“Identifying replication processes with backlogs” on page 137.

T Threshold reached (highlighted turquoise) – A process has fallen behind a


configured threshold. View detailed status to determine which process has
exceeded its backlog threshold and to determine the extent of the backlog.
Use option 8 (Display status) to view detailed status and determine which
replication processes have backlogs. For more information about which detailed
status fields to check, see “Identifying replication processes with backlogs” on
page 137.

V Virtual switch (highlighted yellow) – The apply processes and access path
maintenance for the data group are ended because the data group is participating
in a virtual switch of an application group. Other processes are active.
The apply processes and access path maintenance will be started automatically
during the recovery phase of the virtual switch procedure. User action is required to
initiate the recovery phase. See “Working with procedures and steps” on page 97.

X Switch mode (highlighted red) – The data group is in the middle of switching
(planned or unplanned) the data source system and status may not be retrievable
or accurate.
Wait for the switch to complete. To check for additional status, do one of the
following:
• In environments that use application groups, check the status of the switch
procedure using the WRKPROCSTS command.
• In environments with only data groups that switch using an implementation of
MIMIX Model Switch Framework, use the command CHKSWTFWK
SWTFWK(name) to check status. You can also go to the MIMIX Basic Main
Menu, select option 5 (Start or complete switch using Switch Asst.).

#
Displaying data group status

Table 25. Possible status values for source and target process summaries

P Partially active (highlighted red) - At least one subprocess is active (or is ended for
a virtual switch), but one or more subprocesses are not active. This status is only
displayed in process columns that represent multiple processes. The data group
name may also be shown in a highlighted field of red.
In the Target DB column, partial status is also possible when all other processes,
including database apply, are active but access path maintenance1 is enabled and
does not have at least one active job.
Use option 8 (Display status) to view detailed status and determine which
replication process is not active.
Then from the Work with Data Groups display, use option 9 (Start DG) to start the
data group.

D Disabled – The process is currently not active and the data group is disabled.
Note: The status value for a disabled data group is the letter D displayed in standard
format. No colored blocks are used.
For more information about disabled data groups, see “Disabling and enabling data
groups” on page 334.

W Waiting at a recovery point (highlighted red) - The process is currently suspended


at a recovery point.
Investigation and analysis of the recovery point data may be necessary. Ending the
data group will clear the recovery point, and the next start operation will attempt to
process the available entries, including any with corruptions.
1. Access path maintenance is available only on installations running 7.1.15.00 or higher.

Note: Use F10 (Legend) to view a pop-up window that displays the status values
and colors. To remove the pop-up window, press Enter or F12 (Cancel).

Setting the automatic refresh interval


You can control how frequently the data shown on the Work with Data Groups display
is refreshed by doing the following:
1. Press F9 (Automatic refresh).
2. The Automatic Refresh Value pop-up appears. Specify how long you want the
system to wait before refreshing the information and press Enter.
The status displayed will automatically refresh when the specified interval passes. To
end the automatic refresh process, press Enter.

104
Displaying detailed status for a data group
Detailed status is available for one data group at a time.
The Data Group Status display (DSPDGSTS command) uses multiple views to
present status of a single data group. The views identify and provide status for each
of the processes used by the data group. Error conditions for the data group as well
as process statistics and information about the last entry processed by each
replication process are included. Some fields are repeated on more than one view.
Do the following to access detailed status for a data group:
1. Use one of the following to locate the data group you want:
• To select a data group from a list of all data groups in the installation, select
option 6 (Work with data groups) on the MIMIX Basic Main Menu and press
Enter.
• To select a data group from a subsetted list for an application group, from the
Work with Application Groups display use option 13 (Data resource groups) to
select a resource group. On the resulting display use option 8 (Data groups).
2. The Work with Data Groups display appears. Type an 8 (Display status) next to
the data group you want and press Enter.
3. The Data Group Status display shows a merged view of data group activity on the
source and target systems. (See Figure 14.)
Only fields for the type of information replicated by the data group are displayed.
For example, if the data group replicates only objects from the system journal, you
will only see fields for system journal replication. If the data group replicates from
both the system journal and the user journal, you will see fields for both. To see
additional status information for object processes or database processes, do the
following:
• If the data group contains object information, press F7 (Object) to view
additional object status displays. The Data Group Object Status display
appears.
• If the data group contains database information, press F8 (Database) to view
additional database status displays. The Data Group Database Status display
appears. Tracking entry information for advanced journaling is also available.
4. For object information, there are three views. For database information, there are
four views available. Use F11 to change between views.
Note: If the data group contains both database and object information, you can
toggle between object details and database details by using the F7 and F8
keys.

Merged view
The initial view displayed is the merged view. This view summarizes status for the
replication paths configured for the data group. The status of each process is
represented by the color of the box surrounding the process and a status letter. Table
26 shows possible status values.

126
Displaying detailed status for a data group

Figure 14 shows a sample of the merged view of the Data Group Status display. The
data group in this view is configured for user journal replication using remote
journaling and for system journal replication. Also, access path maintenance is
enabled.

Figure 14. Merged view of data group status. The inverse highlighted blocks are not shown
in this example.

Data Group Status 17:39:36


Data group . . . . : CRITICALAP Database error/rcy. : 1 / 0
Elapsed time . . . : 00:52:51 Object in error/active . : 4 / 0
Transfer definition: PRIMARY-A State . . : *ASYNCPEND
--------------------------- Source Statistics ---------------------------
System: LONDON-A Jrn Mgr-A RJLNK Mon-A
Receiver Sequence # Date Time Trans/Hour
Database Source Jrn. LONDN0002 >0,000,002,591 4/20/08 11:02:35
Link-A RJ Tgt Jrn. LONDN0002 >0,000,002,591 4/20/08 11:02:35
Last Read . LONDN0002 >0,000,002,591 4/20/08 11:02:35
Entries not read: 0 Est. time to read:
Object Current . . AUDRCV0108 22,314,732 4/22/08 17:37:13 748
Send-I Last Read . AUDRCV0103 22,175,464 4/21/08 11:05:56
*SHARED Entries not read : 139,268 Est. time to read:
--------------------------- Target Statistics ---------------------------
System: CHICAGO-A Jrn Mgr-A DB Rdr- A AP Maint-A RJLNK Mon-A
Sys Jrn Insp -A Last Received Unprocessed Entry Count Est Time
User Jrn Insp-A Sequence # Entry Count Trans/Hour To Apply
DB Apply-A >0,000,002,590
Obj Apply-A 22,023,868 4

F3=Exit F5=Refresh F7=Object view F8=Database F9=Automatic refresh


F10=Restart statistics F12=Cancel F14=Start DG F24=More keys

Note: Journal sequence numbers shown in the Source Statistics and Target
Statistics areas may be truncated if the journal supports *MAXOPT3 for the
receiver size and the journal sequence number value exceeds the available
display field. When truncation is necessary, the most significant digits (left-
most) are omitted. Truncated journal sequence numbers are prefixed by '>'.
This is shown in Figure 14.

Table 26. Possible values for detailed status. Not all statuses are used by each process.

Color and Description


Status

Red When displayed on the Data group, Database errors, or Objects in


error fields, a problem exists that requires action.

Red - I The process is inactive.

Red - W The process is suspended at a recovery point. This status is only


available for apply processes.

Yellow When displayed on the Data group field, a problem exists that may
require attention.

127
Table 26. Possible values for detailed status. Not all statuses are used by each process.

Color and Description


Status

Yellow - P One or more of the processes is active but others are inactive. On the
merged view, this status is possible for the Object Send field.
When displayed for the journal manager field on the merged view and
object views 1 and 2, certain types of recoveries are delayed due to the
ASP threshold for recoveries policy while other recoveries are active.

Yellow - V The process is ended for a virtual switch. This status is available only
for apply processes and access path maintenance.

Turquoise - T The process has a backlog which exceeds its configured threshold. On
fields which summarize status for multiple processes, use F7 and F8 to
view the specific threshold. The -T is not shown in statistical fields. If a
threshold condition persists over time, refer to the MIMIX Administrator
Reference book for information about possible resolutions.

White - U The status of the process is unknown.

Blue - A The process is active.

Blue - C The RJ Link is in catch-up mode.This status is only possible for the
Database Link process in the merged view and the RJ link field in
some database views.

Green - D The data group is disabled. This also means the data group is currently
inactive.

Top left corner: The top left corner of the Data Group Status display identifies the
data group, the elapsed time, and the status of the transfer definition in use. The
elapsed time is the amount of time that has elapsed since you accessed this display
or used the F10 (Restart statistics) key.
Top right corner: The top right corner of the display contains fields that identify the
number of errors that MIMIX identified but could not correct as well as the files and
objects with detected problems that MIMIX is currently attempting or will attempt to
automatically correct. If the workstation supports colors, the number files and objects
in error will be displayed in red.
• The Database errors/rcy. fields. The first value identifies the number of errors in
user journal replication processes. This includes all file entries, IFS tracking
entries, and object tracking entries in error (*HLDERR status). When access path
maintenance1 is enabled, this also includes the number of logical and physical
files that have access path maintenance failures for the data group. The second
value identifies the number of recoveries that exist for objects replicated through
user journals.
• The Object in error/active fields indicate the number of objects that are failed and

1. Access path maintenance is available only on installations running MIMIX 7.1.15.00 or


higher.

128
Displaying detailed status for a data group

the number of objects with pending activity entries. The first number in these fields
indicates the number of objects defined to the data group that have a status of
*FAILED. The second number indicates the number of objects with active
(pending) activity entries or that have new or in-progress recoveries.
• The State field identifies the state of the remote journal link. The values for the
state field are the same as those which appear on the Work with RJ Links display.
This field is not shown if the data group uses source-send processes for user
journal replication.
More detailed breakouts of error and recovery counts are available on the object
views (F7) and Database views (F8). When you use the MIMIX portal application in
Vision Solutions Portal, you can see the actual objects affected and details about the
recovery actions by using the Correction Activity action from the Data Groups portlet.
Source statistics: The middle of the display shows status and summarized statistics
for the journals being used for replication and the processes that read from them. The
following process fields are possible:
System - Identifies the current source system definition. The status value is an
indication of the success in communicating with that system.
Jrn Mgr - Displays the status of the journal manager process for the source
system.
DA Poll - Displays the status of the data area poller. This field is present only if the
data group replicates data areas using this process.
RJLNK Mon - Displays status of the RJLNK monitor on the source system. This
field is present only for data groups that use remote journaling.
Database (Link or Send) - Identifies the status of the process which transfers user
journal entries from the source system to the target system.
Link - Displayed when the data group is configured for remote journaling. The
status is that of the of the RJ link.
Send -Displayed when the data group id configured for MIMIX source-send
processes. The status is that of the database send process.
Object Send - Displays a summation of status from the object send, object
retrieve, and container send processes. The highest priority status from each
process determines the status displayed. Use F7 (Object view) to see the
individual processes. When the data group uses a shared object send job, either
the value *SHARED or a three-character job prefix is displayed below the Send
process status, The value *SHARED indicates that the data group uses the MIMIX
generated shared object send prefix for this source system. A three-character
prefix indicates this data group uses a shared object send job on this system that
is shared only with other data groups which specify the same prefix.
For the Database and Object processes, additional fields identify current journal
information, the last entry that has been read by the process, and statistics related to
arrival rate, entries not read, and estimating the time to read.
Current - For the Database Send and Object Send processes, this identifies the
last entry in the currently attached journal receiver. This information is used to
show the arrival rate of entries to the journals.

129
Note: If the data group uses remote journaling, current information is displayed
in two rows, Source jrn and RJ tgt jrn. The source journal sequence
number refers to the last sequence number in the local journal on the
source system. The remote journaling target journal sequence number
refers to the last sequence number in the associated remote journal on the
target system.
Transactions per hour - For current journal information, this is based on the
number of entries to arrive on the journal over the elapsed time the statistics have
been gathered. For last read information, this is based on the actual number of
entries that have been read over the elapsed time the statistics have been
gathered.
Last Read - Identifies the journal entry that was last read and processed by the
object send, database send, or database reader.
Transactions per hour - For current journal fields, this is based on the number of
entries to arrive on the journal over the elapsed time the statistics have been
gathered. For last read fields, this is based on the actual number of entries that
have been read over the elapsed time the statistics have been gathered and will
change due to elapsed time and the rate at which entries arrive in the journal.
Entries not read - This a calculation of the number of journal entries between the
last read sequence number and the sequence number of the last entry in the
current receiver for the source journal. An asterisk (*) preceding this field indicates
that the journal receiver sequence numbers have been reset between the last
entry in the current receiver and the last read entry.
Estimated time to read - This is a calculation using the entries not read and the
transactions per hour rate. This calculation is intended to provide an estimate of
the length of time it may take the process (database reader, database send, or
object send) to complete reading the journal entries.
Target statistics: The lower part of the display shows status and summarized
statistics for all target system processing. The following process fields are possible:
System - Identifies the current target system definition. The status value is an
indication of the success in communicating with that system.
Jrn Mgr - Displays the status of the journal manager process for the target system.
DB Rdr - Displays status of the database reader. This field is present only for data
groups that use remote journaling.
AP Maint - Displays status of the access path maintenance1 processes. This field
is only present when optimized access path maintenance has been enabled.
RJLNK Mon - Displays status of the RJLNK monitor on the target system. This
field is present only for data groups that use remote journaling.
Sys Jrn Insp - Displays the status of target journal inspection for the system
journal (QAUDJRN) on the target system of the data group. This field is displayed

1. Access path maintenance is available only on installations running MIMIX 7.1.15.00 or


higher. In earlier levels of MIMIX, if parallel access path maintenance is enabled, its status is
displayed in the Prl AP Mnt field that appears in this location.

130
Displaying detailed status for a data group

when the journal definition for the system journal on the current target system
permits target journal inspection and the data group is enabled and has been
started at least once.
User Jrn Insp - Displays the status of target journal inspection for the user journal
on the target system of the data group. This field is displayed when the journal
definition for the user journal on the current target system permits target journal
inspection and the data group is enabled, performs user journal replication,
permits journaling on target, and has been started at least once.
DB Apply and Obj Apply - Each field displays the combined status for the apply
jobs in use by the process. For each process, additional fields show statistics for
the last received journal sequence number, number of unprocessed entries,
approximate number of transactions per hour being processed, and the
approximate amount of time needed to apply the unprocessed transactions for all
database or object apply sessions.

Object detailed status views


Figure 15, Figure 16, and Figure 17 show samples of the information available when
you use F7 (Object) to view the detailed object information. Use F11 to move between
the three views of detailed object status. On each view, you can use the F1 (Help) key
to see a description of that view’s contents.
In all object views, journal sequence numbers may be truncated if the journal supports
*MAXOPT3 for the receiver size and the journal sequence number value exceeds the
available display field. When truncation is necessary, the most significant digits (left-
most) are omitted. Truncated journal sequence numbers are prefixed by '>'.
Problem counts: The upper right corner of Figure 15 and Figure 16 shows these
counts:
• Objects in error - The number of objects in the data group that currently have a
status of *FAILED.
• Recoveries - The number of new or in-progress recovery actions for objects in the
data group that are replicated using the system journal (QAUDJRN)
Status values: In addition to the possible status values for processes indicated in
Table 26, Object views 1 and 2 (Figure 15 and Figure 16) also display these
additional status values that are unique to system journal replication processes in the
fields for the Retrievers, Senders, and Applies processes: The Min, Act, and Max
fields indicate the minimum, active, and maximum number of jobs for each process.
The number of active jobs varies based on the work load. The active count is
highlighted with color for the following conditions:
Red - The number of active jobs is zero (0).
Yellow - The number of active jobs is greater than zero (0) but less than the
minimum number of processes, or the apply process is ended for virtual switch
testing.
Turquoise - The process has a backlog that exceeds its configured threshold.
When this occurs, the backlog field for the process is also highlighted in the color
turquoise.

131
Blue - The number of active jobs is equal to or greater than the minimum number
of processes.
Figure 15 and Figure 18 show the active count highlighted.

Figure 15. Data group detail status, object view 1.

Data Group Object Status System: CHICAGO


17:50:00
Data group . . . . : CRITICALAP Objects in error . . . : 4
Elapsed time . . . : 00:52:51 Recoveries . , . . . . : 0

Send Process -I *SHARED


Jrn Manager -A
Receiver Sequence # Date Time Trans/Hour
Current . . AUDRCV0108 10,022,314,732 4/22/08 17:37:13 748
Last Read . AUDRCV0103 10,022,175,464 4/21/08 11:05:56
Entries not read: 139,268 Est. time to read:
--------------------- Object Retrieve/Container Send ----------------------
Retrievers Retrieve Senders Send Containers Containers
Min Act Max Backlog Min Act Max Backlog Sent Per Hour
1 0 5 1 0 5 1,145

------------------------------- Object Apply -------------------------------


Applies Apply Active Entries Entries
Min Act Max Backlog Objects Sequence # Applied Per Hour
1 1 5 4 >0,022,023,871 1,133

F3=Exit F5=Refresh F7=Merged view F8=Database view


F9=Automatic refresh F11=View 2 F12=Cancel F24=More keys

Figure 16. Data group detail status, object view 2.

Data Group Object Status System: CHICAGO


17:57:31
Data group . . . . : CRITICALAP Objects in error . . 4
Elapsed time . . . : 00:52:51 Recoveries . , . . . . : 0

Send Process -I *SHARED


Jrn Manager -A
Receiver Sequence # Date Time Trans/Hour
Current . . AUDRCV0108 10,022,314,732 4/22/08 17:37:13 748
Last Read . AUDRCV0103 10,022,175,464 4/21/08 11:05:56
Entries not read: 139,268 Est. time to read:
--------------------- Object Retrieve/Container Send ----------------------
Retrievers Retrieve Senders Send Containers Containers
Min Act Max Backlog Min Act Max Backlog Sent Per Hour
1 0 5 1 0 5 1,145

------------------------------- Object Apply -------------------------------


Applies Apply ------------- Last Applied -------------
Min Act Max Backlog Sequence # Type Object
1 1 5 0 >0,022,023,871 *DOC BVT#I/PBBDOCXX.002

F3=Exit F5=Refresh F7=Merged view F8=Database view


F9=Automatic refresh F11=View 3 F12=Cancel F24=More keys

132
Displaying detailed status for a data group

Figure 17. Data group detail status, object view 3.

DG Object Journal Entry Detail System: CHICAGO


18:01:20
Data group . . . . : CRITICALAP

Source system: LONDON-A


Entry Sequence # Receiver Date Time
Current entry TSF 10,022,314,732 AUDRCV0108 4/22/08 17:37:13
Last read entry - 10,022,175,464 AUDRCV0103 4/21/08 11:05:56
Last received -

Target system: CHICAGO-A


------------------------------- Object Send -------------------------------
Entry Sequence # Date Time Type Object
Active TCO >0,022,023,868 4/20/08 13:59:23 *DOC BVT#I/PBBDOCXX.002
Processed TCO >0,022,023,868 4/20/08 13:59:23 *DOC BVT#I/PBBDOCXX.002

------------------------------- Object Apply -------------------------------


Entry Sequence # Date Time Type Object
Processed TCA >0,022,023,871 4/20/08 13:59:23 *DOC BVT#I/PBBDOCXX.002

F3=Exit F5=Refresh F7=Merged view F8=Database view


F9=Automatic refresh F11=View 1 F12=Cancel F24=More keys

Database detailed status views


Figure 18, Figure 19, Figure 20, and Figure 21 show samples of the information
available when you use F8 (Database) to view the detailed database information. On
each view, you can use the F1 (Help) key to see a description of that view’s contents.
In database views that include sequence numbers, the journal sequence numbers
may be truncated if the journal supports *MAXOPT3 for the receiver size and the
journal sequence number value exceeds the available display field. When truncation
is necessary, the most significant digits (left-most) are omitted. Truncated journal
sequence numbers are prefixed by '>'.
Status values: Most fields that display status of a process have some or all of the
possible values indicated in Table 26. Possible values for the Jrn State and Cache
(Src and Tgt) fields are indicated in Table 28 (journal state) and Table 29 (journal
cache).
The data group configuration determines whether the Send process field is replaced
by the RJ Link field. When remote journaling is configured, the RJ Link and DB Rdr
fields are shown.
The AP Maint field is displayed on views 1 and 2 (Figure 18 and Figure 19) only when
the access path maintenance1 policy is enabled. When present, this field displays the
status of the access path maintenance job that persists while the database apply
process is active.

1. Access path maintenance is available only on installations running 7.1.15.00 or higher.

133
Problem counts: In the top right corner of database views 1 and 2 (Figure 18 and
Figure 19), these fields display combined counts of replicated entries, new or in-
progress recoveries, and errors for file entries, IFS tracking entries, and object
tracking entries:
• File and Tracking entries - The number of database files, journaled IFS objects,
and journaled data areas and data queues being replicated for the data group.
• Not journaled Src Tgt - The number of files and objects identified by file entries
and tracking entries that are currently not journaled or that are journaled to a
journal that is different than the configured journal on the source and target
systems. If the number of not journaled errors on either system exceeds 99,999,
that system’s field displays +++++.
• Held error and Rcy - The number of files and objects identified by file entries and
tracking entries that are held due to an error (*HLDERR) status, and the number
that have new or in-progress recovery actions for problems detected by MIMIX.
• Access path maint. errors - The number of database files with access path
maintenance errors.
• Held for other reasons - The number of files and objects identified by file entries
and tracking entries that are held or inactive for other reasons.
Database view 4 (Figure 21) separates this information into columns for file entries,
IFS tracking entries, and object tracking entries.
Apply sessions: If a data group has multiple database apply sessions you will see
an entry for each session in the Apply Status column on database views 1, 2, and 3
(Figure 18, Figure 19, and Figure 20). Each session has its own status value. In these
sample figures there is only one apply session (A) which is active (-A).
In database view 1, the Open Commit column displays the value *DLK when there is
an open commit cycle that has caused a deadlock condition for the apply process
which MIMIX could not automatically resolve. User action is necessary to resolve the
commit cycle.

Figure 18. Data group detail status—database view 1.


In this example, the Link status of -A and the presence of the Reader status indicate that the
data group uses remote journaling and access path maintenance. The display also shows that
journal standby state is active and journal caching is not active. The unprocessed entry count
indicates that the final journal entry has not been applied. The > character preceding
sequence numbers for the apply session indicate truncated sequence numbers that are asso-

134
Displaying detailed status for a data group

ciated with *MAXOPT3 support.

Data Group Database Status System:


CHICAGO
18:07:02
Data group . . . . : CRITICALAP File and Tracking entries : 12
Elapsed time . . . : 00:52:51 Not journaled Src: 1 Tgt: 1
Jrn State and Cache Src: A N Tgt: A N Held error: 1 Rcy: 0
RJ Link-A AP Maint-A Access path maint. errors : 1
Jrn Mgr-A DB Rdr- A Held for other reasons . : 0
Receiver Sequence # Date Time Trans/Hour
Source Jrn. LONDN0002 12,345,678,900,000,002,591 4/20/08 11:02:35
Rj Tgt Jrn. LONDN0002 12,345,678,900,000,002,591 4/20/08 11:02:35
Last Read . LONDN0002 12,345,678,900,000,002,591 4/20/08 11:02:35
Entries not read: 0 Est. time to read:
------------------------------- Database Apply ---------------------------
Apply Received Processed Unprocessed Entry Count Est Time Open
Status Sequence # Sequence # Entry Count Trans/Hour To Apply Commit
A-A >0,000,002,593 >0,000,002,592 1 *NO

F3=Exit F5=Refresh F7=Object view F8=Merged view


F9=Automatic refresh F11=View 2 F12=Cancel F24=More keys

135
Figure 19. Data group database status—view 2.
In this example, the Link status of A and the presence of the Reader status indicates that the
data group uses remote journaling. The display also shows that access path maintenance is
used and active, and that journal standby state is active and journal caching is not active.

Data Group Database Status System:


CHICAGO
16:07:03
Data group . . . . : CRITICALAP File and Tracking entries. : 12
Elapsed time . . . : 00:52:51 Not journaled Src: 1 Tgt: 1
Jrn State and Cache Src: A N Tgt: A N Held error: 1 Rcy: 0
RJ Link-A AP Maint-A Access path maint. errors : 1
Jrn Mgr-A DB Rdr- A Held for other reasons . : 0
Receiver Sequence # Date Time Trans/Hour
Source Jrn. LONDN0002 12,345,678,900,000,002,591 4/20/08 11:02:35
Rj Tgt Jrn. LONDN0002 12,345,678,900,000,002,591 4/20/08 11:02:35
Last Read . LONDN0002 12,345,678,900,000,002,591 4/20/08 11:02:35
Entries not read: 0 Est. time to read:
------------------------------- Database Apply ---------------------------
Apply Received Apply point Clock Time Hold MIMIX Log Open
Status Sequence # Sequence # Difference Sequence # Commit Id
A-A >0,000,002,590 >0,000,002,590

F3=Exit F5=Refresh F7=Object view F8=Merged view


F9=Automatic refresh F11=View 3 F12=Cancel F24=More keys

Figure 20. Data group database status, view 3.

DG Database Jrn Entry Detail System: CHICAGO


18:16:04
Data group . . . . : CRITICALAP

Source system: LONDON-A


Entry Sequence # Receiver Date Time
Current entry UMX 12,345,678,900,000,002,591 LONDN0002 4/20/08 11:02:35
RJ target entry UMX 12,345,678,900,000,002,591 LONDN0002 4/20/08 11:02:35
Last read entry UMX 12,345,678,900,000,002,591 LONDN0002 4/20/08 11:02:35
Last received - 12,345,678,900,000,002,590 - 4/20/08 11:01:04

Target system: CHICAGO-A


------------------------------- Database Apply -----------------------------
Apply Entry Sequence # Date Time Object Library Member
A-A UMX >0,000,002,590 4/20/08 11:01:04

F3=Exit F5=Refresh F7=Object view F8=Merged view


F9=Automatic refresh F11=View 1 F12=Cancel F24=More keys

136
Identifying replication processes with backlogs

Figure 21. Data group detail status—database view 4. In this example, the combined number
of file and tracking entries and number of recoveries (from database views 1 and 2) are sepa-
rated into columns for file entries, IFS tracking entries, and object tracking entries.

File and Tracking Entry Status System: CHICAGO


16:07:03
Data group . . . . : CRITICALAP

File IFS Trk Obj Trk


Entries Entries Entries
Number of entries . . . . : 7 3 2
Not journaled on source . . . : 1 0 0
Not journaled on target . . . : 0 1 0
Held due to error . . . . . . : 0 1 1
Access path maint. errors . . : 1 - -
Held for other reasons . . . : 0 0 0
New or in-progress recoveries : 0 0 0

F3=Exit F5=Refresh F7=Object view F8=Merged view


F9=Automatic refresh F11=View 1 F12=Cancel F24=More keys

Identifying replication processes with backlogs


If replication processes are active and have no reported error conditions, a replication
process that has exceeded its backlog threshold will have a status that reflects this
condition. However, if a replication process is inactive or has an error condition with a
higher priority status, the threshold condition will not be visible in the process status
until the process is started or the problem is resolved. Also, a backlog may exist but
not be large enough to exceed the threshold setting, or the threshold warning setting
may have been disabled (set to *NONE).
Do the following to check for a backlog condition:
1. To access the details for a data group, use the procedure in “Displaying detailed
status for a data group” on page 126.
2. Use F7 or F8 on the Data Group Status display to locate the appropriate view for
the process you want to check. Table 27 identifies this information and the

137
appropriate fields for each process.

Table 27. Location of fields which identify backlogs and threshold conditions for replication processes

Process Description

View Fields to Check for Backlog Fields Highlighted


When Threshold
Exceeded

RJ Link The backlog is the quantity of source journal entries that have not been transferred from the
local journal on the source system to the remote journal on the target system. The time
difference between the last entry in each journal can also be an indication of a backlog.

Merged view, Differences between journal entries • RJ tgt jrn Sequence # 1


Database views 1 identified by Source Jrn and RJ Tgt jrn • RJ tgt jrn Date and
and 2 for the database link. Time 2

DB Reader The backlog is the quantity of journal entries that are waiting to be read by the process. The
or DB Send time difference between the last entry that was read by the process and the last entry in the
journal on the source system can also be an indication of a backlog. This may be a temporary
condition due to maximized log space capacity. If the log space capacity was reached, the
database reader job will be idle until the database apply job is able to catch up. If the
condition is unable to resolve itself, action may be required.

Merged view, For remote journaling configurations, • Entries not read


Database views 1 differences between journal entries Sequence # 1
and 2 identified by Source Jrn and Last read. • Last Read Date and
For MIMIX source-send configurations, Time 2
differences between journal entries
identified by Current and Last Read.

DB Apply The backlog is the number of entries waiting to be applied to the target system. Each apply
session is listed as a separate entry with its own backlog. If a virtual switch procedure is in
progress, a backlog may occur during the testing and recovery phases of a virtual switch.

Database views 1, 2, Unprocessed Entry Count • Apply Status


and 3 • Unprocessed Entry
Count

Object The backlog is the quantity of journal entries that have not been read from the system journal.
Send The time difference between the last entry that was read by the process and the last entry in
the system journal can also be an indication of a backlog.
Multiple data groups sharing the object send job is one possible cause of a persistent
backlog.

Merged view, Differences between transactions • Entries not read


Object views 1, 2, identified for Object Current and Last Sequence # 1
and 3 Read • Last Read Date and
Time 2

Object The backlog is the number of entries for which MIMIX is waiting to retrieve objects.
Retrieve
Object views 1 and 2 Retrieve Backlog • Retrievers, Act column
• Retrieve Backlog

138
Data group status in environments with journal cache or journal state

Table 27. Location of fields which identify backlogs and threshold conditions for replication processes

Process Description

View Fields to Check for Backlog Fields Highlighted


When Threshold
Exceeded

Container The backlog is the number of packaged objects for entries that are waiting to be sent to the
Send target system.

Object views 1 and 2 Container Send Backlog • Senders, Act column


• Container Send
Backlog

Object The backlog is the number of entries waiting to be applied to the target system. If a virtual
Apply switch procedure is in progress, a backlog may occur during the testing and recovery phases
of a virtual switch.

Object views 1 and 2 Apply Backlog • Applies, Act column


• Apply Backlog

Notes:
1. When highlighted, the threshold journal entry quantity criterion is exceeded.
2. When highlighted the threshold time criterion is exceeded.

If you repeatedly have issues with processes in thresholds, the settings which control
them may need adjustment. For more information, see “Fine-tuning backlog warning
thresholds for a data group” in the MIMIX Administrator Reference book.

Data group status in environments with journal cache or


journal state
Additional information is reported within data group status configured to use MIMIX
support for IBM’s High Availability Journal Performance IBM i option 42, Journal
Standby feature and Journal caching. When these high availability journal
performance enhancements are in use, conditions that require action or attention are
reflected in these locations:
• The data group name is highlighted on the Work with Data Groups display. The
possible problems associated with journal cache or journal state are identified
Table 24 in topic “Identifying and resolving problems reflected in the Data Group
column” on page 120.
• Jrn State and Cache (Src and Tgt) fields within the data group detailed status are
highlighted. These fields are on the database views 1 and 2 of the Data Group
Database Status display (Figure 18, Figure 19 respectively, shown in “Database
detailed status views” on page 133). The possible values for the Jrn State and
Cache (Src and Tgt) fields are indicated in Table 28 (journal state) and Table 29
(journal cache).

139
The Jrn State and Cache (Src and Tgt) fields reflect journal standby state and journal
caching actual values for the journals when the IBM high availability performance
enhancements are installed on the systems defined to the data group. These fields
appear on database views 1 and 2 (Figure 18 and Figure 19). The target journal state
and cache values are set on the journal when the database apply session is started.
Journal State - The status values indicate the actual state value for the source and
target journals. Table 28 shows the possible values for each field.
Journal Cache - The status indicate the actual cache value for the source and
target journals. Table 29 shows the possible values for each field.
For each system (Src or Tgt) status of the journal state is shown first, followed by the
status of the journal cache. If a problem exists with journal state or journal cache, the
data group name is also highlighted with the same color. For information about
resolving journal cache or journal state problems, see “Resolving a problem with
journal cache or journal state” on page 141.

Table 28. Possible status values for Journal State fields

Field Color and Description


Status

Either system White U Unknown. MIMIX was not able to retrieve values, possibly
because the journal environment has not yet been built.

No color A Journal state is active

No color X The required IBM feature, IBM i option 42 - High


Availability Journal Performance, is not installed on this
system

No color S Journal is in standby state as expected

Source Red S Source journal is in standby state but that state is not
expected.

Red I Source journal in inactive state but that state is not


expected.

Target Yellow S Target journal state or cache is not as expected and the
database apply session is active

Yellow I Target journal state is inactive but that state is not


expected.

blank blank The IBM feature is installed but the data group is
configured to not journal on the target system.

140
Data group status in environments with journal cache or journal state

Table 29. Possible status values for Journal Cache fields

Field Color and Description


Status

White U Unknown. MIMIX was not able to retrieve values, possibly


because the journal environment has not yet been built

Either No color X The required IBM feature, IBM i option 42 - High


System Availability Journal Performance, is not installed on this
system

No color Y Caching is active

No color N Caching is not active.

Source Yellow Y Source journal cache value is not as expected.

Yellow N Source journal cache value is not as expected.

Target Yellow Y Target journal cache value not as expected and the
database apply session is active.

Yellow N Target journal cache value not as expected and the


database apply session is active.

blank blank The IBM feature is installed but the data group is
configured to not journal on the target system.

Resolving a problem with journal cache or journal state


Problems with journal state or journal cache can cause the name of a data group to
be highlighted on the Work with Data Groups display. If the data group name is
highlighted in red or yellow, do the following to check for and resolve problems:
1. From the Work with Data Groups display, use option 8 (Display status).
2. From the Data Group Status display, press F8 (Database).
3. The Jrn State and Cache Src and Tgt fields are located In the upper left corner of
the Data Group Database Status display. For each system (Src or Tgt) status of
the journal state is shown first, followed by the status of the journal cache. The
example below shows v for value in all for status positions. Based on the status
displayed in these fields, you can take the actions described in the following steps
to correct the problem:
Jrn State and Cache Src: v v Tgt: v v

4. Source system journal state (first Src: value) - If the source system state is red
and the value for the journal state is standby (S) or inactive (I), the journal state
must be changed and all data replicated through the user journal must be
synchronized. Do the following:

141
a. Press F12 (Cancel) to return to the Work with Data Groups display. Note which
system is specified as the source system for the data group.
b. Use option 45 (Journal Definitions) to view the journal definitions used for the
data group in error.
c. On the Work with Journal Definitions display, determine the journal name and
library specified for the system that is the source system for the data group.
d. Specify the name and library of the source system journal in the following
command:
CHJRN CHGJRN JRN(library/name) JRNSTATE(*ACTIVE)
e. All data replicated through the user journal must be synchronized. For detailed
information about synchronizing a data group, refer to your Runbook or to the
MIMIX Administrator Reference book.
5. Source system journal cache (second Src: value) - If the source system cache
is yellow, the actual status does not match the configured value in the journal
definition used on the source system. Do the following:
a. Press F12 (Cancel) to return to the Work with Data Groups display. Note which
system is specified as the source system for the data group.
b. Use option 45 (Journal Definitions) to view the journal definitions used for the
data group in error.
c. On the Work with Journal Definitions display, use option 5 (Display next to the
journal definition listed for the source system.
d. Check the value of the Journal caching (JRNCACHE) parameter.
e. Determine which value is appropriate for journal cache, the configured value or
the actual status value. Once you have determined this, either change the
journal definition value or change the journal cache (CHGJRN command) so
that the values match.
6. Target system state (first Tgt: value) or Target system cache (second Tgt:
value) - If the target system state or cache is yellow, the actual value for state or
cache does not match the configured value. Do the following:
a. Press F12 (Cancel) to return to the Work with Data Groups display. Note which
system is specified as the target system for the data group.
b. Use option 45 (Journal Definitions) to view the journal definitions used for the
data group in error.
c. On the Work with Journal Definitions display, use option 5 (Display next to the
journal definition listed for the target system.
d. Check the value of the following parameters, as needed:
• Target journal state (TGTSTATE)
• Journal caching (JRNCACHE)
e. Determine why the actual status of the journal state or journal cache does not
match the configured value of the journal definition used on the target system.

142
Data group status in environments with journal cache or journal state

f. Determine which values are appropriate for journal state and journal cache, the
configured value or the actual status value. Once you have determined this,
either change the journal definition value or change the journal state or cache
(CHGJRN command) so that the values match.

Resolving a problem for a journal status of *DIFFJRN


Problems with journaling can cause the name of a data group to be highlighted on the
Work with Data Groups display. If the cause of the problem is a journal status
(JRN1STS or JRN2STS) value of *DIFFJRN, the journal being used is different than
the configured journal in the data group definition. In this case, you should investigate
the reason for using the current journal and determine whether journaling can be
changed to use the configured journal. It may be necessary to involve your
administrator in making this determination. Consider the following:
• If the journal status is *DIFFJRN on the source system, determine the following:
• Does the object need to be replicated?
• If yes, does the object need to be journaled to the current journal because of
an application requirement? If yes, do one of the following:
• Reconfigure the object to an existing data group which uses the current
journal for replication.
• Configure a new data group to use the current journal and change the
configuration of the object to be replicated by that data group.
• If the journal status is *DIFFJRN on the target system for a switchable data group,
ensure it is correctly journaled before switching.
To check the journaling status and specify the journal to use, do the following:
1. From the Work with Data Groups display, use option 8 (Display status).
2. Check the journaling status for objects with problems:
• For files, use option 17 (File entries) for the data group. Then press F10
(journaled view) to see journaling status.
• For IFS tracking entries, use option 50 (IFS tracking entries) for the data
group. Then press F10 (journaled view) to see journaling status.
• For object tracking entries, use option 52 (object tracking entries) for the data
group. Then press F10 (journaled view) to see journaling status.
3. From the journaled view, locate the items with a Journal status (JRN1STS or
JRN2STS) value of *DIFFJRN.
4. Use option 9 (Start journaling) to view the appropriate Start Journaling display.
5. Press F10 (Additional parameters) to display the Force to configured journal
(FORCE) prompt.
6. Change the FORCE prompt to *YES to end journaling to a journal that is different
than the configured journal and start journaling using the configured journal. This
value will also attempt to start journaling for objects not currently journaled.
Journaling will not be ended for objects already journaled to the configured

143
journal.
7. Press Enter to start journaling.
8. Press F5 (Refresh) to update and view the current journaling status.

Resolving a deadlock condition for apply processes


If your environment uses commit cycle processing, it is possible for commit cycles
that are open for an extended time to cause a deadlock condition for database apply
processes. If MIMIX cannot automatically resolve the deadlock, the condition is
reported. The value *DLK is displayed in the Open Commit column on view 1 of the
Data Group Database Status display (Figure 18) and a message in the database
reader (or send) process job log identifies the job that started the long commit cycle.
The condition also triggers email notification for users who have a subscription
configured within Vision Solutions Portal that includes this event.
To resolve the deadlock condition, do the following:
1. From the target system, use the WRKMGLOG command to display the message
log.
2. Look for message LVE311D. The message identifies the job started the long
commit cycle.
3. Take the appropriate action for the job holding the commit cycle open to either
commit or rollback the commit cycle.
Once the commit cycle has been resolved, MIMIX will be able to automatically clear
the deadlock condition.

144
CHAPTER 8 Working with audits

Audits are defined by and invoked through rules and influenced by policies. Aspects
of audits include schedules, status, reported results, and their compliance status.
MIMIX is shipped so that auditing can occur automatically. For day-to-day operations,
auditing requires minimal interaction to monitor audit status and results. MIMIX user
interfaces separate audit runtime status, compliance status, and scheduling
information onto different views to simplify working with audits. Compliance errors and
runtime errors require different actions to correct problems.
This chapter provides information and procedures to support day-to-day operations
as well as to change aspects of the auditing environment. The following topics are
included.
• “Auditing overview” on page 146 describes concepts associated with auditing and
describes the differences between automatic priority audits and automatic
scheduled audits.
• “Guidelines and considerations for auditing” on page 151 identifies considerations
for specific audits, auditing best practices, and recommendations for checking the
audit results.
• “Displaying scheduling information for automatic audits” on page 155 describes
how to access views of the Work with Audits display where you can display when
prioritized audits will run and when scheduled audits will run.
• “Displaying audit runtime status and compliance status” on page 157 describes
how to access audit status and shows examples of additional views of the Work
with Audits display for status and compliance information.
• “Resolving auditing problems” on page 162 describes how to check for the
presence of auditing problems, resolve problems with runtime status, check the
job log of an audit, and resolve audit compliance problems.
• “Running an audit immediately” on page 169 describes how to perform an audit
manually.
• “Ending an audit” on page 170 describes how to end an audit that is in progress or
that has been requested and is waiting to run.
• “Displaying audit history” on page 171 describes how to display history for specific
audits of a data group.
• “Working with audited objects” on page 173 describes how to display a list of
objects compared by one or more audits.
• “Working with audited object history” on page 176 describes how to access the
audit history for a specific object.

145
Auditing overview
All businesses run under rules and guidelines that may vary in the degree and in the
methods by which they are enforced. In a MIMIX environment, auditing provides rules
and enforcement of practices that help maintain availability and switch-readiness at
all times.
Not using or limiting audit use does little to confirm the integrity of your data. These
approaches can mean lost time and issues with data integrity when you can least
afford them.
In reality, successful auditing means finding the right balance somewhere between
these approaches:
• Audit your entire replication environment every day. The benefit of this approach
is knowing that your data integrity exposure is limited to data that changed since
the last audit. The trade-off with this approach can be time and resources to
perform audits.
• Audit only replicated data that “needs” auditing. This approach can be faster and
use fewer resources because each audit typically has fewer objects to check. The
trade-offs are determining what needs auditing and knowing when objects were
last audited.
MIMIX makes auditing easy by automatically auditing all objects periodically and
auditing a subset of objects every day. MIMIX also provides the ability to fine-tune
aspects of auditing behavior and their automatic submission and the ability to
manually invoke an audit at any time.
For more information about how automatic audits are scheduled and how to change
schedules, see “Scheduling audits and reports” on page 60.

Components of an audit
Together, three components identify a unique audit. Each component must exist to
allow an audit to run.
Rule - A program by which an audit is defined and invoked. Each rule shipped with
MIMIX pre-defines a compare command to be invoked and the possible actions
that can be initiated, if needed, to correct detected problems. When invoked, each
rule can check only the class of objects associated with its compare command.
Names of rules shipped with MIMIX begin with the pound sign (#) character.
Data group - A data group provides the context of what to check and how results
are reported. Multiple audits (rules) exist for each data group.
Note: Audits are not allowed to run against disabled data groups.
Schedule - Each unique combination of audit rule and data group has its own
schedule, by which it is automatically submitted to run. MIMIX ships default
scheduling information associated with each shipped rule. Scheduling can be
adjusted for individual audits. A manually invoked audit can be thought of as an
immediate override of scheduling information.

146
Auditing overview

Although people use the terms “audit” and “rule” interchangeably, a rule is a
component of an audit. The process of auditing runs a rule program.

Phases of audit processing


The process of auditing consists of a compare phase and a recovery phase.
In the compare phase of an audit, the identified audit rule initiates a specific compare
command against the data group. The Audit level policy determines if an audit is
allowed to run and how aggressively an audit checks your environment during its
compare phase. If a shipped audit rule provides more than one audit level, each level
provides increasingly more checking capability.
If there are detected differences when the compare phase completes, the audit enters
its recovery phase to start automatic recovery actions as needed. MIMIX attempts to
correct the differences and sends generated reports, called recoveries, to the user
interface. MIMIX removes these generated reports when the recovery action
completes successfully. If the recovery job fails to correct the problem, MIMIX
removes the recovery and sends an error notification to the user interface.
Most audit rules support a recovery phase. MIMIX is shipped with defaults that enable
audits to enter the recovery phase automatically when needed. The recovery phase
can be optionally disabled in the Automatic audit recovery policy.

Object selection methods for automatic audits


MIMIX provides multiple approaches to performing audits automatically. The biggest
difference between these approaches is how objects are selected to be audited. The
other significant difference is when each type of audit is allowed to run.
• In scheduled object auditing, an audit run selects all objects that are configured
for the data group and within the class of objects checked by the audit. MIMIX
automatically runs an audit according to its specified scheduling criteria. Each
time a scheduled audit runs, all eligible configured objects are selected.
• In prioritized object auditing,1 an audit run selects replicated objects according
to their internally assigned priority category and an auditing frequency assigned to
the category. The result is often a subset of the objects replicated by the data
group. Each time a prioritized audit runs, its subset of objects selected to check
may be unique. MIMIX automatically runs a prioritized audit periodically within its
specified time range every day. It may run approximately once per hour or more
often during its time range.
An audit that is manually invoked from the Work with Audits display in the native user
interface is an immediate run of a scheduled audit. Priority audits cannot be manually
invoked from this display. From Vision Solutions Portal (VSP), you have the ability to
perform an immediate run of either method of auditing. From VSP, you also have the
ability to manually invoke an audit that selects only the objects that had errors or

1. In MIMIX DR installations, prioritized auditing is supported when running version 8.1.02.00 or


higher. Upgrades from earlier levels of MIMIX DR must enable automatic scheduling for pri-
oritized auditing

147
differences after the compare phase of the previous audit run (this is called
differences object auditing and is only available from VSP).
Prioritized auditing can reduce the impact of auditing on resources and performance.
This benefits customers who cannot complete IFS audits, cannot audit every day, or
do not audit at all because of time or resource issues.
When both types of auditing are used, you can achieve a balance between verifying
data integrity and resources. Either or both types of automatic auditing can be
disabled, although that is not recommended.

How priority auditing determines what objects to select


MIMIX determines the auditing priority of each replicated object based on its most
recent change, most recent audit, and the frequency specified for auditing priority
categories. At any time, every replicated object falls within one of several
predetermined categories. Objects in each category are eligible for selection
according to the frequency assigned to their category.
Every day, MIMIX automatically submits requests to run a prioritized audit periodically
within the time range specified in its schedule. Each time a prioritized audit starts, it
selects only the objects eligible in each category.

Table 30. Priority auditing categories

Category Description Eligibility Frequency

Objects not equal Objects for which MIMIX detected Objects in this category
differences since the start of the most have the highest priority
recent audit are in this category. This and are always
includes: selected.
• Objects that had any value other than
*EQ in their most recent audit, including
those with differences that were
automatically resolved.
• Objects for which target system
differences were detected by target
journal inspection or by virtual switch
processing.
• Objects with problems detected by
database replication that are being
tracked for automatic recovery.
• When MIMIX is configured for *DLY
commit mode, objects for which commit
rollback entries still exist in the journal
when those objects were part of a
commit cycle that MIMIX automatically
processed (by a commit operation) to
clear a deadlock condition.

148
Auditing overview

Table 30. Priority auditing categories

Category Description Eligibility Frequency

New objects A new object is one that has not been Objects in these
audited since it was created. categories are eligible
for selection according
Changed objects A changed object is one that has been to the category
modified since the last time it was audited. frequency specified in
Unchanged An unchanged object is one that has not the Priority audit policy.
objects been modified since the last time it was
audited.

Audited with no An object with no differences is one that


differences has not been modified since the last time it
was audited and has been successfully
audited with no changes on at least three
consecutive audit runs. Objects remain in
this category until a change occurs.

The #FILDTA audit always selects all members of a file for which auditing is less than 100
percent complete. The occurs in all of the above object selection categories.

Initially, the objects selected by a prioritized audit may be nearly the same as those
selected by a scheduled audit. However, over time the number of objects selected by
a prioritized object stabilizes to a subset of those selected by a scheduled audit.
When both scheduled and priority audits are allowed for the same rule and data
group, MIMIX may not start a prioritized audit if the scheduled audit will start in the
near future.

Audit status and results


Summarized audit status “bubbles up” to higher level user interfaces within the MIMIX
installation. The Work with Data Groups display (WRKDG command) identifies the
number of audits within the MIMIX installation that require action or attention. In
environments with application groups, the summarized replication status on the Work
with Application Groups display (WRKAG command) also includes summarized
status of audits within the application group.
Details about audit runtime status and compliance status values are reported on the
Work with Audits display (WRKAUD command). The information available about each
audit identifies the status of actions performed by its rule, how the audit selected
objects for comparison, the audit’s compliance status, policy values which affect the
actions of each phase, and scheduling information. When a phase completes, its
timestamps and statistics are also available. You can also access output files
containing detailed results for each audit.
You can control whether a subset of the audit runtime status values are considered
errors, warnings, or informational. This is controlled by the audit severity policy and
affects whether those statuses are included when auditing status is rolled up to other
status values within a MIMIX instance.

149
When audit recoveries are enabled, you can control the severity level of the
notifications that are returned when the rule ends in error with the Notification severity
policy.
You can also view job logs associated with notifications and recoveries. Job logs are
accessible from the system on which the audit comparison or recovery job ran.

Audit compliance
Compliance is an indication of whether an audit ran within the time frame of the
compliance thresholds set in auditing policies. The Audits included for compliance
policy1 determines which audits are checked for compliance.
When an audit that is being checked for compliance reaches one of the compliance
thresholds, its audit compliance status is reported as a *ATTN or *ACTREQ on the
Compliance summary view of the Work with Audits display. Audit compliance status is
included in the auditing status that is rolled up to other status values within a MIMIX
instance.
Compliance is checked based on the last completed compare date. Compliance
determines whether the date of the last completed compare completed by an audit is
within the range set by policies. The Audit warning threshold policy and the Audit
action threshold policy define when to indicate that an audit is approaching or
exceeding that range.
For audits configured for scheduled object auditing or both scheduled and prioritized
object auditing, compliance status is based on the last run of a scheduled audit or a
user-invoked audit. For audits configured for only prioritized object auditing,
compliance status is based on the last run, which may have been a prioritized audit or
a user-invoked audit. A user-invoked audit or a scheduled audit checked all objects
that are configured for the data group and within the class of objects checked by the
audit whereas a prioritized audit may have checked only a subset of those objects.
Audit runs that only selected objects that were previously different (*DIF) are never
considered in assessing compliance.

1. The Audits in compliance policy (AUDINCMPL) is available on the SETMMXPCY command


in installations running MIMIX service pack 8.0.07.00 or higher. For instances running earlier
software levels, all audits are checked for compliance.

150
Guidelines and considerations for auditing

Guidelines and considerations for auditing


Auditing is most effective when it is performed regularly and you take action to
investigate and resolve any reported differences that cannot be automatically
corrected.

Auditing best practices


Regular auditing helps you detect problems in a timely manner and can help you to
address detected problems during normal operations instead of during a crisis. MIMIX
is shipped with default values for auditing policies and automatic scheduling criteria
that Vision Solutions recommends as best practice. These values will automatically
be used in new installations and in new data groups. If you determine that the default
policies or scheduling do not meet your auditing needs, you can customize policy
values and scheduling criteria. Auditing best practices include:
Automatically auditing: MIMIX is shipped so that auditing occurs automatically.
• Allow both priority audits and scheduled audits to run automatically. This provides
a balance between checking all objects periodically and checking a subset of
objects every day. You can adjust scheduling criteria for when both types of audits
are automatically submitted to meet the needs of your environment. Scheduling
can be changed with the Set MIMIX Schedules (SETMMXSCD) command.
• Allow audits to perform the most extensive comparison possible. The shipped
value (level 30) for the Audit level policy enables this. If you choose to run audits
at a lower audit level, be aware of the risks, especially when switching.
• Allow audits to perform automatic recovery actions. This provides automatic
correction of detected problems. Recovery is possible when the Automatic audit
recovery policy is enabled.
• Allow MIMIX to run all audits even if you do not replicate certain object types (such
as DLOs). This ensures that if you add new objects in the future, you will be
automatically auditing them. Audits that do not have any objects to check
complete quickly with little use of system resources.
Manually auditing: In addition, manually invoke audits in these conditions:
• Before switching, run all audits at audit level 30. Click this link to see additional
information about the audit level policy.
• If you make configuration changes, run the #DGFE audit to check actual
configuration data against what is defined to your configuration. Click this link to
see additional information about when to run the #DGFE audit.
Where to run audits: Run audits from a management system. For most
environments, the management system is also the target system. If you cannot run
rules from the management system due to physical constraints or because of
complex configurations, you can change the Run rule on system policy to meet your
needs. Click this link to see additional information about the Run rule on system
policy.

151
Considerations for specific audits
#DGFE audit - This audit is not eligible for prioritized auditing because it checks
configuration data, not objects. As a result, configuration problems for a data group
can only be detected when a scheduled audit or a manually invoked audit runs.
Run the #DGFE audit during periods of minimal MIMIX activity to ensure that
replication is caught up and that added or deleted objects are reflected correctly in the
journal. If the command is run during peak activity, it may contain errors or indicate
that files are in transition.
In addition to regularly scheduled audits, check your configuration using the #DGFE
audits for your data groups whenever you make configuration changes, such as
adding an application or creating a library. Running the audit prior to audits that
compare attributes ensures that those audits will compare the objects and attributes
you expect to be present in your environment.
#DLOATR audit - This audit supports multiple levels of comparisons. The level used
is controlled by the value of the Audit level policy in effect when the audit runs. The
#DLOATR audit compares attributes as well as data for objects defined to a data
group when audit level 20 or 30 is used. Audit level 10 compares only attributes.
When data is compared the audit may take longer to run and may affect performance.
#FILDTA audit - This audit supports multiple levels of comparisons. The level used is
is controlled by the value of the Audit level policy in effect when the audit runs. The
#FILDTA audit compares all file member data defined for file members defined to a
data group only when audit level 30 is used. Level 10 and level 20 compare 5 percent
and 20 percent of data, respectively. Lower audit levels may take days or weeks to
completely audit file data. New files created during that time may not be audited.
Regardless of the audit level you use for regular auditing, Vision Solutions strongly
recommends running a level 30 audit before switching.
#IFSATR audit - This audit supports multiple levels of comparisons. The level used is
controlled by the value of the Audit level policy in effect when the audit runs. The
#IFSATR audit compares data when audit level 20 or 30 is used. At level 10, only
attributes are compared. Regardless of the audit level you use for regular auditing,
Vision Solutions strongly recommends running a level 30 audit before switching.
For journaled IFS objects (*DIR, *STMF, *SYMLNK) identified by IFS tracking entries,
a prioritized audit run of the #IFSATR audit will select the eligible objects from each
priority category.
For non-journaled IFS directory (*DIR) objects identified for system journal replication,
a prioritized audit run of the #IFSATR audit will select directory objects based on their
eligibility for each priority category and may include stream files (*STMF) and
symbolic links (*SYMLNK) within the directory under these conditions:
• When the *STMF or *SYMLNK object within the selected directory is new or
changed since the last audit run, or had differences in the last audit run.
• When the directory has been added to the configuration or moved into the
configured name space since the last audit run.
• When the directory meets selection criteria for either the unchanged objects or
audited with no differences priority category. This ensures that the objects within

152
Guidelines and considerations for auditing

the directory are periodically checked according to the frequency specified in the
priority audit schedule.
#MBRRCDCNT audit - This audit compares the number of current records
(*CURRDS) and the number of deleted records (*NBRDLTRCDS) for physical files
that are defined to an active data group. Equal record counts suggest but do not
guarantee that files are synchronized.
The #MBRRCDCNT audit does not have a recovery phase. Differences detected by
this audit appear as not recovered in the Audit Summary.
In some environments using commitment control, the #MBRRCDCNT audit may be
long-running. Refer to the MIMIX Administrator Reference book for information about
improve performance of this audit.

Recommendations when checking audit results


Consider these recommendations when you check results of audits:
• Always review the results of the audits. Audit results reflect only what was
actually compared. Some objects may not have been compared due to object
activity or due to the audit level policy value in effect, even when no differences
(*NODIFF) are reported. You may need to take actions other than running an
audit to correct detected issues. For example, you may need to change a
procedure so that target system objects are only updated by replication
processes.
• Be aware of priority auditing behavior. Priority audits differ from other audits in
how they select objects to audit and in the number of objects selected. Be aware
of the implications of those differences when checking audit results. Priority audits
select replicated objects based on their auditing eligibility. As a result, priority
audits cannot check newly created source objects until after their create
transactions have been replicated. Priority audits can return results indicating that
zero (0) objects were selected. This occurs when no objects were eligible for
selection by an audit.
• Deleted objects reported as not found. Audits can report not found conditions
for objects that have been deleted. A not found condition is reported when a
delete transaction is in progress for an object eligible for selection when the audit
runs.This is more likely to occur when there are replication errors or backlogs at
the time the audit runs.
• Fixing one error may expose another. It may take multiple iterations of running
audits with recoveries before the results are clean. Recovering from one error
may result in a different error surfacing the next time the audit is performed. For
example, a recovery that adds data group file entries may result in detecting a
database relationship difference (*DBRIND) error the next time the audit is
performed, where the root problem is that a library of logical files is not identified
for replication.
• Watch for trends in the audit results. Trends may indicate situations that need
further investigation. For example, objects that are being recovered for the same
reason every time you run an audit can be an indication that something in your
environment is affecting the objects between audits. In this case, investigating the

153
environment for the cause may determine that a change is needed in the
environment, in the MIMIX configuration, or in both. Trends may also indicate a
MIMIX problem, such as reporting an object as being recovered when it was not.
Report these scenarios to MIMIX CustomerCare. You can do this by creating a
new case using the Case Management page in Support Central.

154
Displaying scheduling information for automatic audits

Displaying scheduling information for automatic audits


An audit can be configured to run by schedule, by priority, by schedule and priority, or
not at all. The schedule summary views of the Work with Audits display allow you to
see scheduling information for each audit.
Do the following to view when an audit can occur for a specific audit and data group
combination:
1. From the MIMIX Intermediate Main Menu, type 6 (Work with audits) and press
Enter.
2. The Work with Audits display appears, showing either the audit summary or
compliance summary view. Press F10 as needed to access the Schedule
summary view.
3. The initial view of the Schedule summary is displayed (Figure 22). This view
shows the date and time of the next scheduled audit run. You cannot view the
exact time of when the next prioritized audit will run.
Use F11 to toggle between additional variations of audit schedule views.
• To view current scheduled auditing settings, press F11 (Figure 23).
• To view current priority auditing settings, press F11 twice (Figure 24).
Prioritized audit runs are allowed to start every day only during the specified
time range. Multiple runs of an audit may occur during that time.
The list is initially sorted by rule and data group name.To sort the list by scheduled
time, use F17.

Figure 22. Audit Schedule Summary, view - next scheduled time.

Work with Audits


System: AS01
Type options, press Enter.
5=Display 6=Print 7=History 8=Recoveries 9=Run rule 10=End
14=Audited objects 36=Change DG policies 37=Change audit schedule

Audit ---------Definition--------- -Scheduled Time--


Opt Rule DG Name System 1 System 2 Frequency Date Time
__ #DGFE EMP AS01 AS02 *WEEKLY 09/25/08 02:00:00
__ #DLOATR EMP AS01 AS02 *WEEKLY 09/25/08 02:25:00
__ #FILATR EMP AS01 AS02 *WEEKLY 09/25/08 02:10:00
__ #FILATRMBR EMP AS01 AS02 *WEEKLY 09/25/08 02:20:00
__ #FILDTA EMP AS01 AS02 *WEEKLY 09/25/08 02:35:00
__ #IFSATR EMP AS01 AS02 *WEEKLY 09/25/08 02:15:00
__ #MBRRCDCNT EMP AS01 AS02 *WEEKLY 09/25/08 02:30:00
__ #OBJATR EMP AS01 AS02 *WEEKLY 09/25/08 02:05:00

Bottom
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F10=Audit summary F11=Schedule settings
F14=Audited objects F17=Sort sched. time F24=More keys

155
Figure 23. Audit Schedule Summary, view - settings for scheduled (full) audits.

Work with Audits


System: AS01
Type options, press Enter.
5=Display 6=Print 7=History 8=Recoveries 9=Run rule 10=End
14=Audited objects 36=Change DG policies 37=Change audit schedule

Audit Weekday Rel.Day


Opt Rule DG Name Frequency Date SMTWTFS 12345L Time
__ #DGFE EMP *WEEKLY *NONE SMTWTFS 02:00:00
__ #DLOATR EMP *WEEKLY *NONE SMTWTFS 02:25:00
__ #FILATR EMP *WEEKLY *NONE SMTWTFS 02:10:00
__ #FILATRMBR EMP *WEEKLY *NONE SMTWTFS 02:20:00
__ #FILDTA EMP *WEEKLY *NONE SMTWTFS 02:35:00
__ #IFSATR EMP *WEEKLY *NONE SMTWTFS 02:15:00
__ #MBRRCDCNT EMP *WEEKLY *NONE SMTWTFS 02:30:00
__ #OBJATR EMP *WEEKLY *NONE SMTWTFS 02:05:00

Bottom
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F10=Audit summary F11=Priority settings
F14=Audited objects F17=Sort sched. time F24=More keys

Figure 24. Audit Schedule Summary, view - settings for priority audits.

Work with Audits


System: AS01
Type options, press Enter.
5=Display 6=Print 7=History 8=Recoveries 9=Run rule 10=End
14=Audited objects 36=Change DG policies 37=Change audit schedule

Audit -Start Range- ----Priority Objects Selected----


Opt Rule DG Name After Until New Chg Unchg No Diff
__ #DGFE EMP *NONE
__ #DLOATR EMP 03:00 08:00 *DAILY *DAILY *WEEKLY *MONTHLY
__ #FILATR EMP 03:00 08:00 *DAILY *DAILY *WEEKLY *MONTHLY
__ #FILATRMBR EMP 03:00 08:00 *DAILY *DAILY *WEEKLY *MONTHLY
__ #FILDTA EMP 03:00 08:00 *DAILY *DAILY *WEEKLY *MONTHLY
__ #IFSATR EMP 03:00 08:00 *DAILY *DAILY *WEEKLY *MONTHLY
__ #MBRRCDCNT EMP 03:00 08:00 *DAILY *DAILY *WEEKLY *MONTHLY
__ #OBJATR EMP 03:00 08:00 *DAILY *DAILY *WEEKLY *MONTHLY

Bottom
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F10=Audit summary F11=Priority settings
F14=Audited objects F17=Sort sched. time F24=More keys

156
Displaying audit runtime status and compliance status

Displaying audit runtime status and compliance status


Audit status and audit compliance status appear on different views of the Work with
Audits display.
Do the following to access audit status information:
1. Do one of the following:
• From the MIMIX Intermediate Main Menu, select option 6 (Work with audits)
and press Enter. Then use F10 as needed to access the Audit summary view.
• Enter the command: installation-library/WRKAUD
2. The Work with Audits display opens to the view that includes the highest severity
errors, which may be runtime or compliance.
• Use F10 to change between views for Audit summary, Compliance summary,
and Schedule summary.
• Use F11 from any view to display alternate columns of information.

Audits - Summary view and alternate columns


The Summary view of the Work with Audits display shows the Audit Status column
and the full name of the data group checked by the audit (Figure 25). This view also
includes the following information:
• The Object Diff column identifies the number of audited objects with differences
remaining after the audit completed.
• The Objects Selected column indicates how objects were selected for auditing in
the most recent run of the audit.
The list is sorted so that the audit with the highest severity status is at the top of the
list. Depending on the audit severity policy in effect at the time an audit ran, the status
value *NOTRUN may be considered an error, a warning, or informational, and the

157
status value *IGNOBJ may be considered a warning or informational. These
severities affect the sort order.

Figure 25. Audit Summary, view - data group definition columns

Work with Audits


System: AS01
Type options, press Enter.
5=Display 6=Print 7=History 8=Recoveries 9=Run rule 10=End
14=Audited objects 46=Mark recovered ...

Audit Audit ---------Definition--------- Object Objects


Opt Status Rule DG Name System 1 System 2 Diff Selected
__ *NOTRUN #OBJATR EMP AS01 AS02 0 *PTY
__ *CMPACT #DLOATR EMP AS01 AS02 0 *PTY
__ *CMPACT #FILATR EMP AS01 AS02 0 *PTY
__ *CMPACT #FILATRMBR EMP AS01 AS02 0 *PTY
__ *RCYACT #FILDTA EMP AS01 AS02 0 *PTY
__ *QUEUED #IFSATR EMP AS01 AS02 0 *PTY
__ *QUEUED #MBRRCDCNT EMP AS01 AS02 0 *PTY
__ *NODIFF #DGFE EMP AS01 AS02 0 *ALL

Bottom
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F10=Compliance summary F11=Last run
F14=Audited objects F16=Inst. policies F23=More options F24=More key

F11 changes the columns displayed so you can see the policies in effect when the
audit was last run (Figure 26).

Figure 26. Audit Summary view - last run columns.

Work with Audits


System: AS01
Type options, press Enter.
5=Display 6=Print 7=History 8=Recoveries 9=Run rule 10=End
14=Audited objects 46=Mark recovered ...

Audit Audit ------Last Run------- Object


Opt Status Rule DG Name Recovery Level Diff
__ *NOTRUN #OBJATR EMP *ENABLED *LEVEL30 0
__ *CMPACT #DLOATR EMP *ENABLED *LEVEL30 0
__ *CMPACT #FILATR EMP *ENABLED *LEVEL30 0
__ *CMPACT #FILATRMBR EMP *ENABLED *LEVEL30 0
__ *RCYACT #FILDTA EMP *ENABLED *LEVEL30 0
__ *QUEUED #IFSATR EMP *ENABLED *LEVEL30 0
__ *QUEUED #MBRRCDCNT EMP *ENABLED *LEVEL30 0
__ *NODIFF #DGFE EMP *ENABLED *LEVEL30 0

Bottom
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F10=Compliance summary F11=DG definition
F14=Audited objects F16=Inst. policies F23=More options F24=More key

The Last Run columns show the values of policies in effect at the time the audit was
last run through its compare phase.

158
Displaying audit runtime status and compliance status

Recovery identifies the value of the automatic audit recovery policy. When this
policy is enabled, after the comparison completes, MIMIX automatically starts
recovery actions to correct differences detected by the audit. Recovery may also
indicate a value of *DISABLED if a condition checked by the Action for running
audits (RUNAUDIT) policy existed and the policy value for that condition specified
*CMP, preventing audit recoveries from running.
Level identifies the value of the audit level policy. The audit level determines the
level of checking performed during the compare phase of the audit. If an audit was
never run, the value *NONE is displayed in both columns.

159
Audits - Compliance summary view and alternate columns
The Compliance summary view of the Work with Audits display shows the
Compliance column and the full name of the data group checked by the audit (Figure
27). This view also shows the timestamp of when the compare phase ended in the
Compare End column.
The list is initially sorted by compliance status. To sort the list by scheduled time, use
F17.
Audits with a compliance status of *OK includes audits that are in compliance, audits
that have been excluded from compliance checks by the Audits included for
compliance policy, and audits that have never been run because the associated data
group is new or has been recently enabled.

Figure 27. Audit Compliance, view - data group definition columns.

Work with Audits


System: AS01
Type options, press Enter.
5=Display 6=Print 7=History 8=Recoveries 9=Run rule 10=End
14=Audited objects 36=Change DG policies 37=Change audit schedule

Audit ---------Definition--------- ---Compare End---


Opt Compliance Rule DB Name System 1 System 2 Date Time
__ *OK #DGFE EMP AS01 AS02 09/25/08 12:15:34
__ *OK #DLOATR EMP AS01 AS02 09/25/08 12:15:34
__ *OK #FILATR EMP AS01 AS02 09/25/08 12:15:34
__ *OK #FILATRMBR EMP AS01 AS02 09/25/08 12:15:35
__ *OK #FILDTA EMP AS01 AS02 09/25/08 12:15:38
__ *OK #IFSATR EMP AS01 AS02 09/25/08 12:15:36
__ *OK #MBRRCDCNT EMP AS01 AS02 09/25/08 12:15:38
__ *OK #OBJATR EMP AS01 AS02 09/25/08 12:15:37

Bottom
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F10=Schedule summary F11=Next scheduled
F14=Audited objects F17=Sort sched. time F24=More keys

160
F11 changes the columns displayed so you can see when the scheduled audit run will
occur (Figure 28). The scheduled date and time in this view do not apply to prioritized
audit runs.

Figure 28. Audit Compliance, view 2 - next scheduled time columns.

Work with Audits


System: AS01
Type options, press Enter.
5=Display 6=Print 7=History 8=Recoveries 9=Run rule 10=End
14=Audited objects 36=Change DG policies 37=Change audit schedule

Audit -Scheduled Time-- ---Compare End---


Opt Compliance Rule DG Name Date Time Date Time
__ *OK #DGFE EMP 09/26/08 02:00:00 09/25/08 12:15:34
__ *OK #DLOATR EMP 09/26/08 02:25:00 09/25/08 12:15:34
__ *OK #FILATR EMP 09/26/08 02:10:00 09/25/08 12:15:34
__ *OK #FILATRMBR EMP 09/26/08 02:20:00 09/25/08 12:15:35
__ *OK #FILDTA EMP 09/26/08 02:35:00 09/25/08 12:15:38
__ *OK #IFSATR EMP 09/26/08 02:15:00 09/25/08 12:15:36
__ *OK #MBRRCDCNT EMP 09/26/08 02:30:00 09/25/08 12:15:38
__ *OK #OBJATR EMP 09/26/08 02:05:00 09/25/08 12:15:37

Bottom
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F10=Schedule summary F11=DG definition
F14=Audited objects F17=Sort sched. time F24=More keys

161
Resolving auditing problems
Audits run for individual data groups. To check whether auditing problems exist for
data groups in an installation, do the following:
1. To access the Work with Data Groups display, either select option 6 from the
MIMIX Basic Main menu or select option 1 from the MIMIX Intermediate Main
menu and press Enter.
2. The Audits field located in the upper right of the display identifies the total number
of audits in the installation that require action to correct a problem or to prevent a
situation from becoming a problem.
Work with Data Groups CHICAGO
11:02:05
Type options, press Enter. Audits/Recov./Notif.: 079 / 002 / 003
5=Display definition 8=Display status 9=Start DG
10=End DG 12=Files needing attention 13=Objects in error
14=Active objects 15=Planned switch 16=Unplanned switch ...
---------Source--------- --------Target-------- -Errors-
Opt Data Group System Mgr DB Obj DA System Mgr DB Obj DB Obj

Highlighting on the Audits field indicates the severity of the errors.


Red — At least one audit either failed, ended, has unresolved
differences, did not run and *NOTRUN status was considered an error
by policies in effect at runtime, or is out of compliance.
Yellow — At least one audit either did not run and *NOTRUN status
was considered a warning by policies in effect at runtime, ignored some
objects and the *IGNOBJ status was considered a warning by policies in effect at
runtime, is out of compliance and currently active, or is approaching out of
compliance.
3. Press F7 (Audits) to access the Work with Audits display.
4. The display opens to the view that includes the highest severity errors, which may
be runtime or compliance.
• If the Audit Status column is displayed, there is a problem with the runtime
status of an audit. Continue with “Resolving audit runtime status problems” on
page 162.
• If the Compliance column is displayed, there is a problem with compliance for
an audit. Continue with “Resolving audit compliance status problems” on
page 167.

Resolving audit runtime status problems


Audit runtime status is displayed on the Audit summary view of the Work with Audits
display. Audits with potential problems are at the top of the list.

162
Resolving auditing problems

You may also need to view the output file or the job log, which are only available from
the system where the audits ran. In most cases, this is the management system.
Work with Audits
System: AS01
Type options, press Enter.
5=Display 6=Print 7=History 8=Recoveries 9=Run rule 10=End
14=Audited objects 46=Mark recovered ...

Audit Audit ---------Definition--------- Object Objects


Opt Status Rule DG Name System 1 System 2 Diff Selected
__ *NOTRUN #OBJATR EMP AS01 AS02 0 *PTY

Do the following from the management system:


1. If necessary, do one of the following to access the Work with Audits display.
• From the MIMIX Intermediate Main Menu, select option 6 (Work with audits)
and press Enter. Then use F10 as needed to access the Audit summary view.
• From a command line, enter WRKAUD VIEW(*AUDSTS)
2. Check the Audit Status column for values shown in Table 31 and take the
indicated action.

Table 31. Addressing problems with audit runtime status

Status Action

*FAILED If the failed audit selected objects by priority and its timeframe for starting has not passed,
the audit will automatically attempt to run again.
The rule called by the audit failed or ended abnormally.
• To run the rule for the audit again, select option 9 (Run rule). This will check all objects
regardless of how the failed audit selected objects to audit.
• To check the job log, see “Checking the job log of an audit” on page 166.

163
Table 31. Addressing problems with audit runtime status

Status Action

*ENDED The #FILDTA audit or the #MBRRCDCNT audit ended either because of a policy in effect
or the data group status. The recovery varies according to why the audit ended.
To determine why the audit ended:
1. Select option 5 (Display) for the audit and press Enter.
2. Check the value indicated in the Reason for ending field. Then perform the appropriate
recovery, below.
If the reason for ending is *DGINACT, the data group status became inactive while the
audit was in progress.
1. From the command line, type WRKDG and press Enter.
• If all processes for the data group are active, skip to Step 2.
• If processes for the data group show a red I, L, or P in the Source and Target columns,
use option 9 (Start DG). Note that if the data group was inactive for some time, it may
have a threshold condition after being started. Wait for the threshold condition to clear
before continuing with Step 2.
2. When the data group is active and does not have a threshold condition, return to the
Work with Audits display and use option 9 (Run rule) to run the audit. This will check all
objects regardless of how the ended audit had selected objects to audit.
If the reason for ending is *MAXRUNTIM, the Maximum rule runtime (MAXRULERUN)
policy in effect was exceeded. Do one of the following:
• Wait for the next priority-based audit to run according to its timeframe for starting.
• Change the value specified for the MAXRULERUN policy using “Setting policies -
general” on page 31. Then, run the audit again, either by using option 9 (Run rule) or by
waiting for the next scheduled audit or priority-based audit to run.
If the reason for ending is *THRESHOLD, the DB apply threshold action (DBAPYTACT)
policy in effect caused the audit to end.
1. Determine if the data group still has a threshold condition. From the command line, type
WRKDG and press Enter.
2. If the data group shows a turquoise T in the Target DB column, the threshold exceeded
condition is still present. Wait for the threshold to resolve. If the threshold persists for an
extended time, you may need to contact your MIMIX administrator.
3. When the data group no longer has a threshold condition, return to the Work with Audits
display and use option 9 (Run rule) to run the audit. (This will check all objects.)

164
Resolving auditing problems

Table 31. Addressing problems with audit runtime status

Status Action

*DIFFNORCY The comparison performed by the audit detected differences. No recovery actions were
attempted because of a policy in effect when the audit ran. Either the Automatic audit
recovery policy is disabled or the Action for running audits policy prevented recovery
actions while the data group was inactive or had a replication process which exceeded its
threshold.
If policy values were not changed since the audit ran, checking the current settings will
indicate which policy was the cause. Use option 36 to check data group level policies and
F16 to check installation level policies.
• If the Automatic audit recovery policy was disabled, the differences must be manually
resolved.
• If the Action for running audits policy was the cause, either manually resolve the
differences or correct any problems with the data group status. You may need to start
the data group and wait for threshold conditions to clear. Then run the audit again.
To manually resolve differences do the following:
1. Type 7 (History) next to the audit with *DIFFNORCY status and press Enter.
2. The Work with Audit History display appears with the most recent run of the audit at the
top of the list. Type 8 (Display difference details) next to an audit to see its results in the
output file.
3. Check the Difference Indicator column. All differences shown for an audit with
*DIFFNORCY status need to be manually resolved. For more information about the
possible values, see “Interpreting audit results - supporting information” on page 360.
To have MIMIX always attempt to recover differences on subsequent audits, change the
value of the automatic audit recovery policy.

*NOTRCVD The comparison performed by the audit detected differences. Either some attempts to
recover differences failed, or the audit job ended before recoveries could be attempted on
all differences.
Note: For audits using the #MBRRCDCNT rule, automatic recovery is not possible. Other audits,
such as #FILDTA, may correct the detected differences.
Do the following:
1. Type 7 (History) next to the audit with *NOTRCVD status and press Enter.
2. The Work with Audit History display appears with the most recent run of the audit at the
top of the list. Type 8 (Display difference details) next to an audit to see its results in the
output file.
3. Check the Difference Indicator column. Any objects with a value of *RCYFAILED must
be manually resolved. For any objects with values other than *RCYFAILED or
*RECOVERED, run the audit again. For more information about the possible values,
see “Interpreting audit results - supporting information” on page 360.

165
Table 31. Addressing problems with audit runtime status

Status Action

*NOTRUN The audit request was submitted but did not run. Either the audit was prevented from
running by the Action for running audits policy in effect, or the data group was inactive
when a #FILDTA or #MBRRCDCNT audit was requested.
Note: An audit with *NOTRUN status may or may not be considered a problem in your environment.
The value of the Audit severity policy in effect determines whether the audit status *NOTRUN
has been assigned an error, warning, or informational severity. The severity determines
whether the *NOTRUN status rolls up into the overall status of MIMIX and affects the order in
which audits are displayed in interfaces.
A status of *NOTRUN may be expected during periods of peak activity or when data group
processes have been ended intentionally. However, if the audit is frequently not run due to
the Action for running audits policy, action may be needed to resolve the cause of the
problem.
To resolve this status:
1. From the command line, type WRKDG and press Enter.
2. Check the data group status for inactive or partially active processes and for processes
with a threshold condition.
3. When the data group no longer has a threshold condition and all processes are active,
return to the Work with Audits display and use option 9 (Run rule) to run the audit. (This
will check all objects.)

*IGNOBJ The audit ignored one or more objects because they were considered active or could not
be compared because of locks or authorizations that prevented access. All other selected
objects were compared and any detected differences were recovered.
Note: An audit with *IGNOBJ status may or may not be considered a problem in your environment.
The value of the Audit severity policy in effect determines whether the audit status *IGNOBJ
has been assigned a warning or informational severity. The severity determines whether the
*IGNOBJ status rolls up into the overall status of MIMIX, determines whether the ignored
objects are counted as differences, and affects the order in which audits are displayed in
interfaces.
To resolve this status:
1. From the Work with Audits display on the target system, type 7 (History) next to the audit
with *IGNOBJ status and press Enter.
2. The Work with Audit History display appears with the most recent run of the audit at the
top of the list. Type 8 (Display difference details) next to an audit to see its results in the
output file.
3. Check the Difference Indicator column to identify the objects with a status of *UN or *UA.
4. When the locks are released or replication activity for the object completes, do one of
the following:
• Wait for the next prioritized audit to run.
• Run the audit manually using option 9 (Run rule) from the Work with Audits display,

For more information about the values displayed in the audit results, see “Interpreting
audit results - supporting information” on page 360.

Checking the job log of an audit


An audit’s job log can provide more information about why an audit failed. If it still
exists, the job log is available on the system where the audit ran. Typically, this is the
management system.

166
Resolving auditing problems

You must display the notifications from an audit in order to view the job log. Do the
following:
1. From the Work with Audits display, type 7 (History) next to the audit and press
Enter.
2. The Work with Audit History display appears with the most recent run of the audit
at the top of the list.
3. Use option 12 (Display job) next to the audit you want and press Enter.
4. The Display Job menu opens. Select option 4 (Display spooled files). Then use
option 5 (Display) from the Display Job Spooled Files display.
5. Look for messages from the job log for the audit in question. Usually the most
recent messages are at the bottom of the display.
Message LVE3197 is issued when errors remain after an audit completed.
Message LVE3358 is issued when an audit failed. Check for following
messages in the job log that indicate a communications problem (LVE3D5E,
LVE3D5F, or LVE3D60) or a problem with data group status (LVI3D5E,
LVI3D5F, or LVI3D60).

Resolving audit compliance status problems


Audit compliance status is displayed on the Compliance summary view of the Work
with Audits display. Audits with potential problems are at the top of the list.
Compliance is determined for each individual audit based on the date when the audit
last completed its compare phase.
Work with Audits
System: AS01
Type options, press Enter.
5=Display 6=Print 7=History 8=Recoveries 9=Run rule 10=End
14=Audited objects 36=Change DG policies 37=Change audit schedule

Audit ---------Definition--------- ---Compare End---


Opt Compliance Rule DB Name System 1 System 2 Date Time
__ *ATTN #DGFE EMP AS01 AS02 09/25/08 12:15:34

Do the following from the management system:


1. If necessary, do one of the following to access the Work with Audits display.
• From the MIMIX Intermediate Main Menu, select option 6 (Work with audits)
and press Enter. Then use F10 as needed to access the Compliance view
• Enter the command: installation-library/WRKAUD VIEW(*COMPLY)
2. Check the Compliance column for values of *ATTN and *ACTREQ.
*ATTN -The audit is approaching an out of compliance state as determined by the
Audit warning threshold policy. Attention is required to prevent the audit from
becoming out of compliance.
*ACTREQ - The audit is out of compliance with the Audit action threshold policy.
Action is required. Perform an audit of the data group.

167
3. To resolve a problem with audit compliance, the audit in question must be run and
complete its compare phase.
• To see when the scheduled run of the audit will occur, press F11.
• To see when both scheduled and prioritized audits will run, press F10 to
access the Audit summary view, then use F11 to toggle between views.
• To run the audit now, select option 9 (Run rule) and press Enter. This action will
select all replicated objects associated with the class of the audit. For more
Information, see “Running an audit immediately” on page 169.

168
Running an audit immediately

Running an audit immediately


You always have the option of running an audit immediately. You can do this by
running the MIMIX rule associated with the audit. From the native user interface,
audits invoked in this manner always select all replication-eligible objects associated
with the class of object for the audit. When running an audit immediately from Vision
Solutions Portal, you have the ability to select whether the audit will select all
replication-eligible objects or only prioritized objects.
In most cases, you want to run the audit from the management system. Policies
determine whether a request to run an audit can be performed on the requesting
system.
Most users should perform this procedure form the management system.
To run a rule immediately, do the following:
1. From the MIMIX Intermediate Main Menu, select option 6 (Work with audits) and
press Enter.
2. Type option 9 (Run rule) next to the audit you want and press Enter.
Note: Audits are not allowed to run against disabled data groups.

169
Ending an audit
Only active or queued audits can be ended. This includes audits with the following
statuses: Currently comparing (*CMPACT), Currently recovering (*RCYACT), or
Currently waiting to run (*QUEUED).
You must end active or queued audits from the system that originated the audit. You
can end active or queued audits from any view of the Work with Audits display. This
procedure uses the Status view.
To end an active or queued audit, do the following:
1. From the MIMIX Intermediate Main Menu, select option 6 (Work with audits) and
press Enter. Then use F10 as needed to access the Audit summary view.
2. Check the value shown in the Audit Status column. Press F1 (Help) for a
description of status values.
3. Type option 10 (End) next to the active or queued audit you want to end and
press Enter.
4. Audits in *CMPACT or *QUEUED status are set back to their previous status
values. Audits in *RCYACT status are set according to the completed comparison
result as well as the results of any completed recovery actions.

170
Displaying audit history

Displaying audit history


The Work with Audit History display lists the available history for completed runs of a
specific combination of audit rule and data group. Each item listed is a history of a
completed audit run, shown in reverse chronological order so that the completed audit
with the most recent start time is at the top of the list. Audits that are new or that have
an active status are not included in this list.
Do the following to access retained history for a specific audit and data group
combination:
1. From the MIMIX Intermediate Main Menu, type 6 (Work with audits) and press
Enter.
2. From the Work with Audits display, type 7 (History) next to the audit and data
group you want and press Enter.
The amount of history information available is determined by how frequently an audit
runs and the settings of the Audit history retention policy. Having retained audit history
enables you to look for trends across multiple runs of an audit that may be an
indication of a configuration problem or some other issue with an object. For example,
the Work with Audit History display makes it easy to notice that particular audit of one
data group always has a similar number of recovered objects or always has
differences that cannot be recovered automatically.
The initial view shows (Figure 29) the final audit status and recovery phase statistics.

Figure 29. Work with Audit History - view of recovery results.

Work with Audit History


SYSTEM: AS01
Audit rule . . . . . . : #FILATR
Data group definition . : EMP AS01 AS02

Type options, press Enter.


5=Display 6=Print 8=View difference details 12=Display job
14=Audited objects 46=Mark recovered
------------------Objects-----------------
Audit Total Not Not
Opt Compare Start Status Selected Compared Recovered Recovered
__ 01/02/10 15:25:31 *NODIFF 91 0 0 0
__ 12/31/09 09:06:04 *NODIFF 0 0 0 0
__ 12/30/09 08:50:29 *AUTORCVD 4 0 0 3

BOTTOM
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F9=Retrieve F11=Summary results
F12=Cancel F13=Repeat F14=Audited objects F21=Print list

F11 toggles between this view and additional views.

171
The summary results view (Figure 30) shows the total number of objects selected by
the audit and whether the objects selected were the result of a priority audit or a
scheduled audit.

Figure 30. Work with Audit History - view of summary results.

Work with Audit History


SYSTEM: AS01
Audit rule . . . . . . : #FILATR
Data group definition . : EMP AS01 AS02

Type options, press Enter.


5=Display 6=Print 8=View difference details 12=Display job
14=Audited objects 46=Mark recovered

Audit Total Objects


Opt Compare Start Status Selected Selected
__ 01/02/10 15:25:31 *NODIFF 91 *ALL
__ 12/31/09 09:06:04 *NODIFF 0 *PTY
__ 12/30/09 08:50:29 *AUTORCVD 4 *PTY

BOTTOM
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F9=Retrieve F11=Compare results
F12=Cancel F13=Repeat F14=Audited objects F21=Print list

The compare results view (Figure 31) shows the duration of the audit as well as
statistics for the compare phase of the audit.

Figure 31. Work with Audit History - view of compare results.

Work with Audit History


SYSTEM: AS01
Audit rule . . . . . . : #FILATR
Data group definition . : EMP AS01 AS02

Type options, press Enter.


5=Display 6=Print 8=View difference details 12=Display job
14=Audited objects 46=Mark recovered
------------Objects-------------
Audit Audit Not Detected
Opt Compare Start Status Duration Compared Compared Not Equal
__ 01/02/10 15:25:31 *NODIFF 00:00:04 91 0 0
__ 12/31/09 09:06:04 *NODIFF 00:00:01 0 0 0
__ 12/30/09 08:50:29 *AUTORCVD 00:00:01 4 0 3

BOTTOM
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F9=Retrieve F11=Recovery results
F12=Cancel F13=Repeat F14=Audited objects F21=Print list

172
Working with audited objects

When viewing the Work with Audit History display from the system on which the audit
request originated, you can use options to view the object difference details detected
by the audit (option 8), the job log for the audit (option 9), and a list of objects that
were audited (option 14).

Prioritized audit runs with no selected objects


On the Work with Audit History display, it is possible to see repeated audit runs that
have zero (0) objects selected during the time frame that prioritized audits are allowed
to run each day. Zero objects selected means that no objects matched the frequency
specified for criteria for selecting objects at the time when the prioritized audit ran.
Consider this example of how prioritized audits operate. Audit #FILATR is set to run
priority audits using its shipped default values for priority auditing. This means the
audit will run approximately once per hour between 3 and 8 a.m. every day. Each
audit run will select the following:
• Any replicated objects that were not equal in their last audit.
• Any new replicated objects had never been audited.
• Any replicated objects that changed in the past 24 hours.
• Any replicated objects that did not change since they were audited a week ago.
• Any replicated objects that did not change since their last audit a month (30 days)
ago and have a history of repeated consecutive successful audits.
For the first run (between 3 and 4 a.m.) of a normal work day, it is likely the audit
selected objects in the new and changed in the past day categories, and may have
selected some objects in other categories as well. The second run is likely to have
selected fewer objects, and may have selected only objects that had differences from
the earlier run. If those differences were resolved, then the subsequent runs that day
are likely to have selected no objects because none were eligible. While such a daily
pattern may repeat, it is also subject to replication and other auditing activity within
your environment.

Working with audited objects


The Work with Audited Objects display shows a list of objects compared by one or
more audits. This information is available only on the originating system for audits
performed when the Audit history retention (AUDHST) policy in effect specified to
keep details relevant to the type of audit and those audits have not exceeded the
current policy's retention criteria.
The list of objects is sorted by severity of their final audit status (the status after
comparisons and any recovery actions complete), with the most severe status first.
Because this display lists audited object history, the #DGFE rule, which compares
configuration data, is not included.
When the objects listed are for only one audit, the display appears as shown in Figure
32. This layout is used when the display is invoked by option 14 (Audited objects) on
the Work with Audits display or the Work with Audit History display. Note that the

173
Audit start field is located at the top of the display in this case. If the selected audit is
the audit run with the latest start date, (*LAST) will also appear in the Audit start field.

Figure 32. Work with Audited Objects display for a single audit.

Work with Audited Objects


SYSTEM: AS01
Data group: EMP AS01 AS02
Audit rule: #FILATR Audit start: 06/17/09 15:01:34 (*LAST)

Type options, press Enter.


5=Display 6=Print 9=Object history

Audited Object
Opt Status Type Name
_ *NE *FILE L00SAMPLEA/RJFILE1
_ *EQ *FILE L00SAMPLEA/RJFILE2
_ *EQ *FILE L00SAMPLEA/RJFILE3
_ *RCVD *FILE L00SAMPLEA/RJFILE4

BOTTOM
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F9=Retrieve F12=Cancel F13=Repeat
F18=Subset F21=Print list F22=Display entire field

When the list includes objects from multiple audits, the display appears as shown in
Figure 33, with the specific audit rule and start time displayed in columns. A > symbol
next to an object name indicates a long object path name exists which can be viewed
with F22.
File member information is not automatically displayed. However, you can use F18 to
change subsetting criteria to include members. When member information is
displayed, the name is in the format: library/file(member). Also, the information
displayed for file members may not be from the most recently performed audit.
Because members can be compared by several audits, the most recent run of each of
those audits is evaluated. The evaluated audit run with most severe status is
displayed, even if it is not the most recently performed audit of the evaluated audit

174
Working with audited objects

runs. For all other objects, the information displayed is from the most recent audit run
that compared the object.

Figure 33. Work with Audited Objects display with all audits displayed.

Work with Audited Objects


System: AS01
Data group: EMP AS01 AS02
Audit rule: *ALL

Type options, press Enter.


5=Display 6=Print 9=Object history

Audited Object -----------Audit------------


Opt Status Type Name Rule Date Time
_ *NE *DTAARA L00SAMPLEA/AJDTAARA1 #OBJATR 12/11/09 09:40:27
_ *NE *DTAARA L00SAMPLEA/AJDTAARA2 #OBJATR 12/11/09 09:40:27
_ *NE *STMF /L00DIR/ALPHA.STM #IFSATR 12/11/09 09:47:57
_ *EQ *DTAARA L00SAMPLEA/DTAARA1 #OBJATR 12/11/09 09:40:27
_ *EQ *DTAARA L00SAMPLEA/DTAARA2 #OBJATR 12/11/09 09:40:27
_ *EQ *FILE L00SAMPLEA/RJFILE1 #FILATR 12/11/09 09:43:13
_ *EQ *FILE L00SAMPLEA/RJFILE2 #FILATR 12/11/09 09:43:13
_ *EQ *FILE L00SAMPLEA/RJFILE3 #FILATR 12/11/09 09:43:13

More...
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F9=Retrieve F12=Cancel F13=Repeat
F18=Subset F21=Print list F22=Display entire field

You can select option 5 to view the details of the audit in which the object was
compared, such as the audit compare and recovery timestamps, and option 9 to view
auditing history for a specific object.

Displaying audited objects from a specific audit run


Use this procedure to display the list of objects compared by a specific audit run.
For prioritized audits, not every object is audited in every audit run.
From Work with Audits display or the Work with Audit History display, do the following:
1. Ensure that you are on the system where the audit originated. The originating
system is included in the audit details, which you can view using option 5
(Display).
2. Type 14 (Audited objects) next to audit run that you want and press Enter.
3. If necessary, press F18 (Subset) to specify criteria for filtering the list by object
type, name, or audited status.

Displaying a customized list of audited objects


Use this procedure to list all objects compared by a data group or to specify filtering
criteria such as object type, name, or audited status.
From Work with Audits display or the Work with Audit History display, do the following:
1. Ensure that you are on the system where the audit originated. The originating

175
system is included in the audit details, which you can view using option 5
(Display).
2. Press F14 (Audited objects). The Work with Audited Objects (WRKAUDOBJ)
command appears.
3. Specify the Data group definition for which you want to see audited objects.
4. Specify the value you want for Object type and press Enter.
5. Additional fields appear based on the value specified in Step 5. Specify values to
define the criteria for selecting the objects to be displayed.
Note: The value specified for Member (MBR) determines whether member-level
objects are selected for their object history. The members selected are not
automatically displayed in the list. To include any selected members,
press F10 (Additional parameters), then specify *YES for Include member
(INCMBR).
6. Press Enter to display the list of objects from the retained history details.

Working with audited object history


The Work with Audited Obj. History display lists the available audit history for a single
object compared by the indicated audit rules within the indicated data group. This
capability provides the ability to check for trends for a specific object such as repeated
automatic recovery of a difference.
The audit history for an object is available only on the originating system for audits
performed when the Audit history retention (AUDHST) policy in effect specified to
keep details relevant to the type of audit and those audits have not exceeded the
current policy's retention criteria.
The list is sorted in reverse chronological order so that the audit history having the
most recent start date is at the top of the list.
When the displayed object history is for a file member, the member is represented as
object type *FILE with its name formatted as library/file(member). The Audit Rule
column appears in the list to identify which audit rule compared the member in the
audit run, as shown in Figure 34. When the audit history for any other object type is

176
Working with audited object history

displayed, there is only one possible audit rule so the Audit rule field is located at the
upper right of the display.

Figure 34. Work with Audited Obj. History display showing audit history for a file member

Work with Audited Obj. History


System: AS01
Data group: EMP AS01 AS02
Type: *FILE Name: ABCLIB/PF1(MBR1)

Type options, press Enter.


5=Display 6=Print 8=View difference details

Audit ----Compare Information--- -------Recovery Information------


Opt Rule Date Time Status Date Time Status
_ #FILDTA 06/17/09 15:22:48 *EQ
_ #MBRRCDCNT 06/17/09 15:01:34 *EQ
_ #FILATRMBR 06/16/09 15:01:27 *NE 06/16/09 15:04:25 *RECOVERED

Bottom
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F9=Retrieve F12=Cancel F13=Repeat
F21=Print list F22=Display entire field

From this display you can use option 5 to view details of the audit in which the object
was compared, such as its audit compare and recovery timestamps, and option 8 to
view object difference details that were detected by the audit.

Displaying the audit history for a specific object


Use this procedure to display the retained audit histories for a specific object.
From Work with Audits display or the Work with Audit History display, do the following:
1. Display a list of objects audited for a data group using either of the following
procedures:
• “Displaying audited objects from a specific audit run” on page 175
• “Displaying a customized list of audited objects” on page 175
2. From the Work with Audited Objects display, type 9 (Object history) next to the
object you want and press Enter.

177
Working with system-level processes

CHAPTER 9 Working with system-level processes

MIMIX uses several processes that run at the system level to support the replication
environment and provide additional functionality.
• System-level processes reported on the Work with Systems display (WRKSYS
command) include the system manager, journal manager, target journal
inspection, collector services, and if needed, cluster services.
• Also, single-system processes that monitor specific functions are reported on the
Work with Monitors display (WRKMON command) on each system in the
installation.
Typically, these processes are automatically started and ended when MIMIX is started
or ended. However, you may need to start or end individual processes when resolving
problems.
The following topics are included in this chapter to help you resolve problems with
system level processes:
• “Displaying status of system-level processes” on page 180 describes how to
check for expected status values on the Work with Systems display and identifies
where to find instructions for resolving unexpected status values.
• “Resolving system manager problems and backlogs” on page 181 describes how
to identify and resolve problems with system manager processes.
• “Starting system managers or a journal manager” on page 184 describes how to
start system manager or journal manager processes. In these instructions, MIMIX
will start both if necessary.
• “Ending system managers” on page 184 describes how to end system manager
processes associated with a specified system and identifies when the system
manager RJ links must also be ended. These instructions include optionally
ending the RJ links.
• “Ending a journal manager” on page 185 describes how to end a journal manager
on a selected system.
• “Starting collector services” on page 185 describes how to start collector services.
• “Ending collector services” on page 186 describes how to end collector services.
• “Starting target journal inspection processes” on page 186 describes how to start
target journal inspection on a selected system.
• “Ending target journal inspection processes” on page 187 describes how to end
target journal inspection on a selected system.
• “Displaying status of target journal inspection” on page 188 describes how to
display the status of a single inspection job on a system and how to resolve
problems with its status.
• “Displaying results of target journal inspection” on page 189 describes where to
find information about the objects identified by target journal inspection.

178
• “Identifying the last entry inspected on the target system” on page 191 describes
how to determine the last entry in the target journal and the last entry processed
by target journal inspection.
• “Displaying status of monitors on the local system” on page 192 describes how to
check status of monitors and which monitors should run on the local system.
• “Starting a monitor” on page 194 describes how to start a monitor on the local
system.
• “Ending a monitor” on page 195 describes how to stop a monitor on the local
system.
• “Starting and ending the master monitors” on page 195 describes which
processes will automatically start or end the master monitors and how to end them
manually when necessary.
• “Enabling or disabling a monitor” on page 197 describes how to enable a monitor
so it can be started or disable a monitor so that it cannot run.
• “Displaying status of a remote journal link” on page 199 describes how to display
the status of remote journal links, including those used by data group replication
processes and those used by system manager processes. This topic also
describes the delivery method and state values possible for RJ links.
• “Starting a remote journal link independently” on page 202 describes how to start
an RJ link for a data group or a system manager process independently from
other MIMIX processes.
• “Ending a remote journal link independently” on page 202 describes how to end
an RJ link for a data group or a system manager process independently from
other MIMIX processes.

179
Displaying status of system-level processes
Status of processes that run at the system level can be viewed from the Work with
Systems display.
1. Do one of the following:
• From MIMIX Intermediate Main Menu, select option 2 (Work with Systems) and
press Enter.
• From the Work with Application Groups display, use option 12 (Node entries).
On the resulting Work with Node Entries display, press F7 (Systems).
The Work with Systems display appears, showing the local system at the top of
the list.
2. Check the status values displayed.
• If one or more processes are *UNKNOWN, use “Verifying all communications
links” on page 344.
• If one or more processes are *INACTIVE, use option 9 (Start) to start them on
a system. Any processes that are not active on the system are preselected1 on
the resulting Start MIMIX Managers display.
• The expected status values for normal operation in a two-system environment
are shown below. If the values are not as expected, additional investigation is
needed.
Expected Status Values:
System Managers - *ACTIVE.
Investigate error statuses of *ACTREQ and *INACTRJ using “Resolving
system manager problems and backlogs” on page 181.
Journal Managers - *ACTIVE2.
Investigate warning statuses of *INACTASP and *RCYASP using “Resolving
ASP 1 storage exceeded threshold problems” on page 183.
Target Journal Inspection
For a system that is currently the target system for replication, the expected
status is *ACTIVE.
For other systems, the expected status is *NOTTGT.
Investigate any other status using “Displaying status of target journal
inspection” on page 188.
Collector Services - *ACTIVE.
Cluster Services
For most environments, the expected status is *NONE.
In environments licensed for MIMIX for PowerHA the expected status for a
system that participates in an IBM i cluster is *ACTIVE.

1. Except cluster services. To start cluster services, MIMIX for PowerHA users must specify
*YES for the Start cluster services prompt.
2. In environments with a MIMIX for PowerHA license and you are only controlling a switchable
independent ASP without MIMIX logical replication capability, the expected status is *NONE.

180
Resolving system manager problems and backlogs

Procedure Status - *COMP or *NONE.


Investigate any other status value, see “Displaying status of procedures” on
page 98.
For more information about operation for MIMIX for PowerHA, see the MIMIX
Operations with IBM PowerHA for i book.

Figure 35. Expected status on the Work with Systems display for a two-system environment.

Work with Systems


OSCAR
Local system definition . . : OSCAR
Cluster . . . . . . . . . . : *NONE

Type option, press Enter


4=Remove cluster node 5=Display 6=Print 7=System manager status
8=Work with data groups 9=Start 10=End 11=Jrn inspection status
System ----- Managers ----- Journal ----- Services ------
Opt Definition Type System Journal Inspect. Collector Cluster
___ OSCAR *MGT *ACTIVE *ACTIVE *ACTIVE *ACTIVE *NONE
___ HENRY *NET *ACTIVE *ACTIVE *NOTTGT *ACTIVE *NONE

Bottom
F3=Exit F5=Refresh F9=Automatic refresh F10=Legend F12=Cancel
F13=Repeat F16=System definitions

Resolving system manager problems and backlogs


System manager processes are required for MIMIX to operate. These instructions
describe how to check status for individual system manager processes, how to check
for a backlog, and how to resolve problems.
Each system manager process operates in one direction between a pair of systems.
In the simplest environment with one pair of systems (one management and one
network), each system participates in two system manager processes.
In environments with three or more systems, the number of system manager
processes in which each system participates depends on how the systems are
configured. Each management system communicates with all other management
systems and with each network system unless limited by the configuration of the
network systems.
To check system manager processes, do the following:
1. Do one of the following:
• From MIMIX Intermediate Main Menu, select option 2 (Work with Systems) and
press Enter.

181
• From the Work with Application Groups display, use option 12 (Node entries).
On the resulting Work with Node Entries display, press F7 (Systems).
The Work with Systems display appears. The status shown in the System
Managers column is a summary of all system manager processes through which
a system communicates, and includes status of their system manager RJ links
and processing jobs.
2. You can attempt to resolve system manager statuses of *ACTREQ, *INACTIVE,
or *INACTRJ from this display by doing the following:
a. Type a 9 (Start) next to the system definition you want and press Enter.
b. When the Start MIMIX Managers display appears press Enter.
c. Press F5 (Refresh).
To check specific system manager processes or check for a backlog, or if the
problem persists after attempting a start request continue with these instructions.
3. Type a 7 (System manager status) next to the system definition you want and
press Enter.
The Work with System Pair Status display appears, showing a list that represents
all of the system manager processes associated with the selected system
definition. Each row identifies a pair of systems between which a system manager
process exists and the direction (source to target) in which it transfers data.
4. Check each represented system manager process for statuses of *INACTIVE or
*INACTRJ and do the following to resolve them:
a. To start the system manager processes for the identified source system of a
pair of systems, type a 9 (Start) next to the pair and press Enter.
b. The Start MIMIX Managers display appears. The value *SYS is preselected for
the Manager prompt. Check the values of other prompts and change them as
needed to ensure that the values are appropriate for the system specified.
c. Press Enter.
The start request will start all system managers associated with the source
system of the pair for which the Start option was selected. In a two-node
environment, this affects both system manager processes between the two
systems. In environments with three or more systems, this affects system
manager processes between the selected system and all other systems with
which it communicates.
d. Press F5 (Refresh).
• If a status of *INACTRJ persists, use “Starting a remote journal link
independently” on page 202.
• If a status of *INACTIVE persists, contact your CustomerCare
representative.
5. For every process with a status of *ACTIVE, check the unprocessed entries count.
Unprocessed entries means that a backlog exists and action may be required.
The count identifies the number of entries from the system manager source that

182
Resolving ASP 1 storage exceeded threshold problems

have not been processed by the target’s processing job, and includes entries that
remote journaling has not sent to the target system.
a. Press F5 (Refresh) to ensure that the data displayed is current.
b. For process that have *ACTIVE status, evaluate the timestamp shown for
unprocessed entries. If the timestamp does not change when data is refreshed
(F5), contact CustomerCare.

Resolving ASP 1 storage exceeded threshold problems


When the journal manager status on the Work with Systems display (Figure 35)
displays either the *INACTASP or *RCYASP values, there is a potential problem with
the amount of storage being used by the system auxiliary storage pool (ASP1) on the
source system. The storage used exceeds the threshold for recoveries set in the ASP
threshold for recoveries (RCYASPTHLD) policy. As a result, the replication manager
(MXREPMGR) has delayed processing new recoveries for member data of *FILE
objects to avoid adding potentially large amounts of additional journal entries
associated with these recoveries into the user journal on a system that is already
storage-constrained. This condition can occur when the journal manager is active
(*RCYASP) or inactive (*INACTASP).
You can do one or more of the following to address this situation:
• Reduce the amount of storage used on ASP 1.
• Add more storage to ASP 1.
• Change the value specified for the RCYASPTHLD policy. See “Changing the
system ASP percent used threshold for recoveries” on page 34.
When the percentage of storage used by ASP 1 falls below the threshold policy,
MIMIX will automatically resume processing the recoveries that have been delayed.

183
Starting system managers or a journal manager
To start the system manager processes in which a system participates or to start a
journal manager for a system, do the following:
1. Do one of the following:
• From MIMIX Intermediate Main Menu, select option 2 (Work with Systems) and
press Enter.
• From the Work with Application Groups display, use option 12 (Node entries).
On the resulting Work with Node Entries display, press F7 (Systems).
2. The Work with Systems display appears. Type a 9 (Start) next to the system
definition you want and press Enter.
3. The Start MIMIX Managers display appears. By default, any manager that is not
running will be selected to start. Specify the value for the type of manager you
want to start at the Manager prompt and press Enter.
A start request also starts the RJ links for the system manager processes when
necessary.

Ending system managers


When ending system manager processes, it is usually sufficient to end only the
system manager processing jobs. These instructions address optionally ending the
system manager RJ links.
Important!
The system manager RJ links can remain active unless there is a need to prevent
MIMIX data from being sent to the target node. Scenarios when you should end
the system manager RJ links include before performing an IPL and before
performing a hardware upgrade.
To end the system manager processes in which a system participates, do the
following:
1. Do one of the following:
• From MIMIX Intermediate Main Menu, select option 2 (Work with Systems) and
press Enter.
• From the Work with Application Groups display, use option 12 (Node entries).
On the resulting Work with Node Entries display, press F7 (Systems).
2. The Work with Systems display appears with a list of the system definitions
defined for the MIMIX installation. Type a 10 (End) next to the system definition
you want and press Enter.
3. The End MIMIX Managers display appears. MIMIX preselects values for the
Manager, Target journal inspection, and Collector services prompts based on
whether the status of those processes on the specified system allow them to be
ended. Verify that the values shown reflect what you want and change them if
necessary.

184
Ending a journal manager

Note: For the Manager prompt, the value *ALL indicates that the journal
manager will also be ended. To end any system manager processes,
either *ALL or *SYS must be specified. By default, system manager RJ
links are not ended.
4. Do one of the following:
• To end only the system manager processing jobs, press Enter.
• To end system manager RJ links as well as the processing jobs, specify *YES
for the End system manager RJ links prompt. Then press Enter.

Ending a journal manager


To end a journal manager on a system, do the following:
1. Do one of the following:
• From MIMIX Intermediate Main Menu, select option 2 (Work with Systems) and
press Enter.
• From the Work with Application Groups display, use option 12 (Node entries).
On the resulting Work with Node Entries display, press F7 (Systems).
2. The Work with Systems display appears with a list of the system definitions
defined for the MIMIX installation. Type a 10 (End) next to the system definition
you want and press Enter.
3. The End MIMIX Managers display appears. Specify the value for the type of
manager you want to end at the Manager prompt and press Enter. The selected
managers are ended.

Starting collector services


To start collector services for a system, do the following
1. Do one of the following:
• From MIMIX Intermediate Main Menu, select option 2 (Work with Systems) and
press Enter.
• From the Work with Application Groups display, use option 12 (Node entries).
On the resulting Work with Node Entries display, press F7 (Systems).
2. The Work with Systems display appears. Type a 9 (Start) next to the system
definition you want and press Enter.
3. The Start MIMIX Managers display appears. At the Collector services prompt,
verify the value is *YES and press Enter.

185
Ending collector services
To end collector services for a system, do the following
1. Do one of the following:
• From MIMIX Intermediate Main Menu, select option 2 (Work with Systems) and
press Enter.
• From the Work with Application Groups display, use option 12 (Node entries).
On the resulting Work with Node Entries display, press F7 (Systems).
2. The Work with Systems display appears. Type a 10 (End) next to the system
definition you want and press Enter.
3. The End MIMIX Managers display appears. At the Collector services prompt, type
*YES and press Enter.

Starting target journal inspection processes


These instructions will start target journal inspection processes on a selected system.
If the system is the target system of one or more data groups whose journal
definitions are configured for target journal inspection, a journal inspection job is
started for the system journal and for each user journal on the system.
If the system is the target system for replication, an inspection job is started for the
system journal and for each user journal on the system that is identified within data
groups replicating to the system.
Target journal inspection processes start at the last sequence number in the currently
attached journal receiver in the following cases:
• When it is the first time a target journal inspection process is started
• When starting after being ended and the last processed receiver is no longer
available
• When starting after enabling target journal inspection in a journal definition where
it was previously disabled
When starting target journal inspection after it was previously ended, processing
begins with the next sequence number after the last processed sequence number.
To start target journal inspection processes for a system, do the following
1. Do one of the following:
• From MIMIX Intermediate Main Menu, select option 2 (Work with Systems) and
press Enter.
• From the Work with Application Groups display, use option 12 (Node entries).
On the resulting Work with Node Entries display, press F7 (Systems).
2. The Work with Systems display appears. Type a 9 (Start) next to the system
definition you want and press Enter.
3. The Start MIMIX Managers display appears. At the Target journal inspection

186
Ending target journal inspection processes

prompt, verify the value is *YES and press Enter.

Ending target journal inspection processes


These instructions will end target journal inspection processes on a selected system.
If the system is the target system for replication, the inspection process for the system
journal is ended and all inspection processes are ended for the user journals
identified as the target journal in data groups replicating to the system.
To end target journal inspection processes for a system, do the following
1. Do one of the following:
• From MIMIX Intermediate Main Menu, select option 2 (Work with Systems) and
press Enter.
• From the Work with Application Groups display, use option 12 (Node entries).
On the resulting Work with Node Entries display, press F7 (Systems).
2. The Work with Systems display appears. Type a 10 (End) next to the system
definition you want and press Enter.
3. The End MIMIX Managers display appears. At the Target journal inspection
prompt, verify the value is *YES and press Enter.

187
Displaying status of target journal inspection
Target journal inspection consists of a set of jobs that read journals on the target
system to check for people or processes other than MIMIX that have modified
replicated objects on the target system. Best practice is to allow target journal
inspection for all systems in your replication environment.
Each target journal inspection process runs on a system only when that system is the
target system for replication. The number of inspection processes depends on how
many journals are used by data groups replicating to that system. On a target system,
there is one inspection job for the system journal and one job for each target user
journal identified in data groups replicating to that system.
Because target journal inspection processes run at the system-level, the best location
to begin checking status is from the Work with Systems display.
1. Do one of the following:
• From MIMIX Intermediate Main Menu, select option 2 (Work with Systems) and
press Enter.
• From the Work with Application Groups display, use option 12 (Node entries).
On the resulting Work with Node Entries display, press F7 (Systems).
2. The Work with Systems display appears. The Journal Inspect. column shows the
summarized status of all journal inspection processes on a system.
• Expected values are either *ACTIVE or *NOTTGT.
• For all other status values, type 11 (Jrn inspection status) next to the system
you want and press Enter.
3. The Work with Journal Inspection Status display appears, listing the subset of
journal definitions for the selected system. The status displayed is for target
journal inspection for the journal associated with a journal definition.
Note: Journal definitions whose journals are not eligible for target journal
inspection are not displayed. This includes journal definitions that identify
the remote journal used in RJ configurations (whose names typically end
with @R) as well as the MXCFGJRN journal definition, which are for
internal use.
Table 32 identifies the status for the inspection job associated with a journal and
how to resolve problems.

Table 32. Status values for a single target journal inspection process.

Journal Description and Action


Inspection
Status

*INACTIVE Journal inspection is not active.


(inverse red) Use option 9 (Start) to start all eligible target journal inspection
processes on the system identified in the selected journal definition.

188
Displaying results of target journal inspection

Table 32. Status values for a single target journal inspection process.

Journal Description and Action


Inspection
Status

*UNKNOWN The status of the process on the system cannot be determined, possibly
(inverse white) because of an error or communications problem.
Use the procedure in “Verifying all communications links” on page 344.

*ACTIVE Target journal inspection is active for the journal identified in the journal
(inverse blue) definition.

*NEWDG Target journal inspection has not run because all enabled data groups
that use the journal definition as a target journal have never been
started.
The inspection process will start when one or more of the data groups
are started.

*NOTCFG Either the journal definition does not allow target journal inspection or all
enabled data groups that use the journal definition (user journal) prevent
journaling on the target system. Target journal inspection is not
performed for the journal.
For instructions for configuring target journal inspection, see topics
“Determining which data groups use a journal definition” and “Enabling
target journal inspection” in the MIMIX Administrator Reference book.

*NOTTGT The journal definition is not used as a target journal definition by any
enabled data group. Target journal inspection is not performed for the
journal.
This is the expected status when the journal definition is properly
configured for target journal inspection but the system is currently a
source system for all data groups using this journal definition.

Displaying results of target journal inspection


Target journal inspection sends a warning notification for each user other than the
MIMIXOWN user profile who changed objects on the target system since the
inspection job started. The notification identifies only the first object changed by that
user. Assuming that inspection jobs remain active for 24 hours, target journal
inspection sends a maximum of one notification a day for each user who changed
target system objects. If the target system configuration specifies a job restart time
(default is 24 hours), the inspection jobs restart with other system-level jobs and
target journal inspection resets the notification-per-user tracking based on the
configured restart time. If the system’s configured job restart time does not allow
restarting system-level jobs, the notification-per-user tracking is reset shortly after
midnight every day based on the system clock.
Note: The MIMIX portal application for Vision Solutions Portal provides enhanced
capabilities for displaying target journal inspection results. Notifications from
target journal inspection processes are identified as originating from
TGTJRNINSP in the Notifications portlet on the Summary page. Actions

189
available allow you to display details of the notification and to view all objects
changed on the target by the user identified in the notification as well as any
automatic recovery actions for those objects.

Displaying details associated with target journal inspection notifications


This procedure displays notifications sent by target journal inspection and describes
how to display related information from the native user interface.
To check for notifications for target journal inspection, do the following:
1. Do one of the following:
• On the Work with Application Groups display, the Notifications field indicates
whether any warning notifications exist. Press F15 (Notifications).
• On the Work with Data Groups display, the third number in the
Audits/Recov./Notif. field displays the number of new notifications. Press F8
(Recoveries), then press F10 (Work with Notifications).
• From a command line, enter: WRKNFY.
2. The Work with Notifications display appears. Notifications from target journal
inspection are identified by the name TGTJRNINSP in the Source column.
3. Type a 5 (Display) to view the notifications details.
4. On the Display Notification Details display, check these fields:
• The Originating system field on the Display Notification Details display
identifies the system on which target journal inspection ran and sent the
notification.
• The Notification details field identifies the user or program that made the
change, the first object changed, the location it was found in the inspected
journal, and a command string to run to see journal entries generated by the
user.
Note: If the text of the Notification details field is truncated, you can view the full
text of the message associated with the notification from the MIMIX
message log. Use “Displaying messages for TGTJRNINSP notifications”
on page 190.
5. Investigate why the identified user changed objects on the target system. Objects
may need to be repaired.

Displaying messages for TGTJRNINSP notifications


The text of notifications by target journal inspection vary slightly with the object type of
the reported object. When a notification is sent, an associated message is sent to the
MIMIX message log.
You can use the following commands to view the full text of notification messages,
Use the name of the originating system (Step 4 in previous procedure) as the name of
the originating system (ORGSYS) in these commands:
• For library-based objects, Enter:

190
Identifying the last entry inspected on the target system

WRKMSGLOG MSGID(LVE3902) PRC(TGTJRNINSP) ORGSYS(name)


• For IFS objects, Enter:
WRKMSGLOG MSGID(LVE3903) PRC(TGTJRNINSP) ORGSYS(name)
• For DLO objects, Enter:
WRKMSGLOG MSGID(LVE3904) PRC(TGTJRNINSP) ORGSYS(name)

Identifying the last entry inspected on the target system


For each target journal inspection process, you can view details that identify the last
journal entry inspected and identify the last entry in the current journal receiver.
Do the following:
1. From MIMIX Intermediate Main Menu, select option 2 (Work with Systems) and
press Enter.
2. The Work with Systems display appears. Type 11 (Jrn inspection status) next to
the system you want and press Enter.
3. The Work with Journal Inspection Status display appears. Type 5 (Display) next to
the journal definition on the system you want and press Enter.
4. The Display Journal Inspection Status Details display appears.
• The following fields identify the currently attached journal receiver and the last
entry in the current receiver: Journal, Journal receiver, Last journal entry
sequence. and Last journal entry time.
• The Target journal inspection fields identify the last entry processed by target
journal inspection.

191
Displaying status of monitors on the local system
Each node or system in the product configuration has MIMIX monitors which run on
that system to check for specific potential problems. The Work with Monitors display
indicates the status and other details about monitors on the local system.
Monitor status must be checked from each system. To check status on a system, do
the following:
Note: Environments that do not have application groups can start with Step 3.
1. Select option 1 (Work with application groups) from the MIMIX Basic Main Menu
and press Enter.
2. The Monitors field located in the upper right corner of the Work with Application
Groups display summarizes the status of the MIMIX monitors on the local system
with one of the following values.
*ACTIVE - All enabled monitors on the local system are active.
*ATTN - Either one or more monitors on the local system failed or there are both
active and inactive monitors on the local system. Investigation is necessary.
*INACTIVE - All enabled monitors on the local system are inactive.

Figure 36. Monitors field on the Work with Application Groups display

Work with Application Groups


System: SYSA
Monitors . . . . . : *ACTIVE
Notifications . . : *NONE
Type options, press Enter.
1=Create 2=Change 4=Delete 5=Display 6=Print 9=Start
10=End 12=Node entries 13=Data resource groups 15=Switch

3. Do one of the following to access the Work with Monitors display:


• From with Work with Application Groups display, press F14 (Monitors).
• Select option 12 (Work with monitors) from the MIMIX Basic Main Menu.
• Enter the command: WRKMON
The Work with Monitors display lists the shipped monitors and any user-created
monitors on the local system.
4. Verify which monitors are needed on the local system.Table 33 identifies monitors
that are part of MIMIX.
5. Check the Status column for status values of FAILED, FAILED/ACT, and

192
Displaying status of monitors on the local system

INACTIVE. These values require investigation or additional action.

Table 33. Possible monitors and the nodes on which they should be active

Monitor When and Where Needed

journal-name - remote journal link monitor Primary node and the


Checks the journal message queue for indications of current Backup node of
problems with the remote journal link. A monitor exists for application groups which
both the local and remote system of the RJ link. perform logical replication.

MMIASPMON - independent ASP threshold monitor On all nodes which control


Checks the QSYSOPR message queue for indications that an independent ASP.
the independent ASP threshold has been exceeded. This
monitor improves the ability to detect overflow conditions
that put your high availability solution at risk due to
insufficient storage.
Note: This monitor is shipped disabled. If you want to use this
monitor, user action is required to enable it on all systems
in the MIMIX installation that are using independent ASPs.

MMNFYNEWE - monitor for new object notification entries Primary node when the
Monitors the source system for the newly created libraries, application group is
folders, or directories that are not already included or configured for logical
excluded for replication by a data group configuration. replication.
Note: This monitor is shipped disabled. User action is required to
enable this monitor on the source system within your
MIMIX environment. Once enabled, the monitor will
automatically start with the master monitor.

MQRCDIMG - MQ record image monitor Primary node


This monitor is optional for environments that use MIMIX for
MQ. The monitor runs at a scheduled time and invokes the
IBM WebSphere MQ command RCDMQMIMG for each
queue manager on the system.
Note: This monitor is shipped disabled. User action is required to
enable this monitor on the source system within your
MIMIX environment. Once enabled, the monitor will
automatically start with the master monitor. This monitor is
available in MIMIX 8.1.17.00 and higher.

Table 34. Possible status values for monitors

Status Value Description


(native user inter-
face)

FAILED/ACT The resulting actions of the monitor failed but the monitor remains
active. If the monitor belongs to a group it will participate in the
decision to trigger the group.

193
Table 34. Possible status values for monitors

Status Value Description


(native user inter-
face)

FAILED The monitor has failed. The event program for the monitor requested
that the monitor be ended and the status set to failed, an unexpected
error occurred in the monitor, or the maximum number of restarts
have been attempted. If the monitor belongs to a group, it will
participate in the decision to trigger the group.

ENDING The monitor is ending normally. Once the monitor completes its
shutdown activities, the status will be changed to either INACTIVE or
FAILED depending on the return value from the last call to the event
program. User action may be required if the state persists.

INACTIVE The monitor has not been started. The monitor is not active.
User action is necessary to start the monitor.

HELD The monitor has been started, but its activity is suspended until a
release is requested.

DISABLED The monitor has been disabled and is not active. If the monitor
belongs to a group, it will not participate in the starting or triggering of
the group. A disabled monitor is not affected by master monitor
actions. User action is necessary to enable the monitor.

ACTIVE The monitor has been started and is active.

SCHEDULED The monitor is scheduled. This value is only valid for time monitors.

Starting a monitor
Only monitors with status values of Inactive, Failed, or Disabled can be started.
If a group monitor is started, unless you specify otherwise, the request is extended to
all monitors defined to the selected group monitor that have a status that allows them
to be started. The status of a monitor may prevent it from participating in the start
request. Monitors with a disabled status are never started with a group monitor.
Do the following to start a monitor:
1. From the MIMIX Basic Main Menu, select option 12 (Work with monitors) and
press Enter.
2. From the Work with Monitors display, type a 9 (Start) next to the monitor you want
and press F4 (Prompt).
3. The Start Monitor (STRMON) display appears. If the monitor you are starting is a
journal monitor or an extended journal monitor, you will see prompts for Journal
Receiver and Starting large sequence number. Specify values that identify the
starting journal receiver and the first journal entry to be considered for monitoring.

194
Ending a monitor

4. If the selected monitor is a group monitor and you want to start only the group
monitor itself, press F10 (Additional parameters). Then specify *NAME at the
Scope of request prompt.
5. Press Enter.

Ending a monitor
Only monitors with status values of Active, Scheduled, Held, or Failed/Act can be
ended.
If a group monitor is ended, unless you specify otherwise, the request is extended to
all monitors defined to the selected group monitor that have a status that allows them
to be ended. The status of a monitor may prevent it from participating in the end
request.
Do the following to end a monitor:
1. From the MIMIX Basic Main Menu, select option 12 (Work with monitors) and
press Enter.
2. From the Work with Monitors display, type a 10 (End) next to the monitor you
want and press F4 (Prompt).
3. The End Monitor (ENDMON) display appears.
4. If the selected monitor is a group monitor and you want to end only the group
monitor itself, press F10 (Additional parameters). Then specify *NAME at the
Scope of request prompt.
5. Press Enter.

Starting and ending the master monitors


A master monitor job runs in the MIMIXSBS subsystem on each system and controls
all other monitors and the MIMIX scheduler job on the same system. The master
monitor does not appear in the list shown on the Work with Monitors display.
The autostart job entry created in the subsystem during installation automatically
starts the master monitor job on that system. During normal operations, the Start
MIMIX (STRMMX) and End MIMIX (ENDMMX) commands address the master
monitor jobs and the monitors each job controls on all systems in the MIMIX
installation.
• The STRMMX command will start the master monitors on each system in a MIMIX
installation. In turn, the master monitor on each system will start the MIMIX
scheduler job and all monitors that are not disabled and which are configured to
automatically start with the master monitor.
• The ENDMMX command will end all monitors that are in an active state. Then, the
MIMIX scheduler job and the master monitor on each system are ended.
You may also need to start or end a master monitor independently of other MIMIX
jobs.

195
Starting the master monitor on the local system
The master monitor will attempt to restart itself a limited number of times when it ends
because of an error or if the job is abnormally ended. If an error condition prevents the
master monitor from restarting itself, the master monitor will end. This causes the
MIMIX scheduler job and all other active monitors to also end. When the master
monitor is started, other monitors that are configured to start with the master monitor
and which are not disabled are started.
Use the following command, specifying the name of the MIMIX installation library:
installation-library/STRMSTMON
The master monitor job name will be the installation library name when it is started.

Ending the master monitor on the local system


You may need to end the master monitor when cleaning up the runtime environment,
or when a CustomerCare representative directs you to end the master monitor.
Ending the master monitor also causes all other active monitors to end. Those
monitors will not function correctly without the master monitor. The MIMIX scheduler
job is also ended.
Use the following command, specifying the name of the MIMIX installation library:
installation-library/ENDMSTMON

196
Enabling or disabling a monitor

Enabling or disabling a monitor


A monitor must be enabled before it can be started. Disabled monitors cannot be
started and will not run. These states determine what other operations can be
performed on a monitor.
For monitors defined to a group monitor, status determines whether the monitors
participate in group decisions. The ability to change status is useful for preventing or
ensuring participation by members of a group. A monitor is a member of a group when
the value specified on its Monitor group parameter (MONGRP) matches the name of
an existing group monitor (event class of *GROUP).

Enabling a monitor
Enabling a monitor changes its status from DISABLED to INACTIVE.
Once enabled, a monitor can be started and run its normal operations. Once enabled,
a monitor can be started by the master monitor if it is configured to do so. If the
monitor is defined to a group, it will participate in requests to start all monitors in the
group and in decisions for the group monitor.
Do the following to enable a monitor:
1. From the MIMIX Basic Main Menu, select option 12 (Work with monitors) and
press Enter.
2. From the Work with Monitors display, type a 20 (Enable) next to the monitor you
want and press F4 (Prompt).
3. The Change Monitor Status (CHGMONSTS) display appears. The value
*ENABLED should be preselected for the Status prompt.
4. If the selected monitor is a group monitor and you want to enable only the group
monitor itself, press F10 (Additional parameters). Then specify *NAME at the
Scope of request prompt.
5. Press Enter.

Disabling a monitor
A monitor must have a status of Inactive or Failed before it can be disabled.
A monitor with a DISABLED status cannot run its normal operations and is prevented
from starting with the master monitor if configured to do so. If it is defined to a group, it
will not participate in requests to start all monitors in the group and will not participate
in decisions for the group monitor.
DISABLED monitors can be started individually. Once they are started, they will
participate in the triggering decision for any group to which they belong.
Do the following to disable a monitor:
1. From the MIMIX Basic Main Menu, select option 12 (Work with monitors) and
press Enter.
2. From the Work with Monitors display, type a 21 (Disable) next to the monitor you

197
want and press F4 (Prompt).
3. The Change Monitor Status (CHGMONSTS) display appears. The value
*DISABLED should be preselected for the Status prompt.
4. If the selected monitor is a group monitor and you want to disable only the group
monitor itself, press F10 (Additional parameters). Then specify *NAME at the
Scope of request prompt.
5. Press Enter.

198
Displaying status of a remote journal link

Displaying status of a remote journal link


Remote journal links (RJ links) are used by replication processes for data groups and
by system manager processes.
To check the status of an RJ link used by a data group, do the following:
From the Work with Data Groups display (WRKDG command), type 44 (RJ links)
next to the data group you want and press Enter.

To check the status of an RJ link used for by a system manager process, do the
following:
1. From the Work with Systems display (WRKSYS command), type 7 (System
manager status) next to the system you want and press Enter.
2. On the resulting Work with System Pairs display, type 12 (Work with RJ links) and
press Enter.

To check the status of all RJ links in a MIMIX installation, type the command:
installation-library/WRKRJLNK PROCESS(*ALL)

Figure 37. Example showing all RJ links in an environment with two systems and one data
group. The first two RJ links listed are for the data group. The last two are for system manager
processes between the systems.

Work with RJ Links


System: CHICAGO
Type options, press Enter.
1=Add 2=Change 4=Remove 5=Display 6=Print 9=Start 10=End
14=Build 15=Remove RJ connection 17=Work with jrn attributes
24=Delete target jrn environment

---Source Jrn Def--- ---Target Jrn Def---


Opt Name System Name System Priority Dlvry State
__ __________ ________ __________ ________
__ PAYABLES CHICAGO PAYABLES@R NEWYORK *SYSDFT *ASYNC *INACTIVE
__ PAYABLES NEWYORK PAYABLES@R CHICAGO *SYSDFT *ASYNC *INACTIVE
__ MXSYSMGR CHICAGO MXSYS_01 NEWYORK *SYSDFT *ASYNC *ASYNC
__ MXSYSMGR NEWYORK MXSYS_02 CHICAGO *SYSDFT *ASYNC *ASYNC

Bottom
Parameters or command
===> _________________________________________________________________________
F3=Exit F4=Prompt F5=Refresh F6=Add F9=Retrieve F11=View 2
F12=Cancel F13=Repeat F16=Jrn Definitions F18=Subset F21=Print list

You can view details about an RJ link by using option 5 (Display).

199
The Dlvry column indicates configured value for how the IBM i remote journal function
sends the journal entries from the source journal to the target journal. The possible
values for delivery are asynchronous (*ASYNC) and synchronous (*SYNC).
*ASYNC - Journal entries are replicated asynchronously, independent of the
applications that create the journal entries. The applications continue processing
while an independent system task delivers the journal entries. If a failure occurs
on the source system, journal entries on the source system may become trapped
because they have not been delivered to the target system.
*SYNC - Journal entries are replicated synchronously. The applications do not
continue processing until after the journal entries are sent to the target journal. If a
failure occurs on the source system, the target system contains the journal entries
that have been generated by the applications.
The State column represents the composite view of the state of the remote journal
link. Because the RJ link has both source and a target component, the state shown is
that of the component which has the most severe state. Table 35 shows the possible
states of an RJ link, listed in order from most severe to least severe.

Table 35. Possible states for RJ links, shown in order starting with most severe.

State Description

The following states are considered to be inactive:

*UNKNOWN The state of the link cannot be checked. Both journals defined to the
remote journal link do not reside on the local system and a transfer
definition does not exist between the local system and the source
system of the remote journal link.

*NOTAVAIL The ASP where the journal is located is varied off.

*NOTBUILT The remote journal link is defined to MIMIX but one of the associated
journal environments has not been built.

*SRCNOTBLT The remote journal link is defined to MIMIX but the associated source
journal environment has not been built.

*TGTNOTBLT The remote journal link is defined to MIMIX but the associated target
journal environment has not been built.

*FAILED The remote journal cannot receive journal entries from the source
journal due to an error condition.

*CTLINACT The remote journal link is processing a request for a controlled end.

*INACTIVE The remote journal link is not active.

The following states are considered to be active:

*INACTPEND An active remote journal link is in the process of becoming inactive. For
asynchronous delivery, this is a transient state that will resolve
automatically. For synchronous delivery, one system is inactive while the
other system is inactive with pending unconfirmed entries.

200
Displaying status of a remote journal link

Table 35. Possible states for RJ links, shown in order starting with most severe.

State Description

*SYNCPEND An active remote journal link is connected using synchronous delivery


and is running in catch-up mode. The state will become *SYNC when
catch-up mode ends.

*ASYNCPEND An active remote journal link is connected using asynchronous delivery


and is running in catch-up mode. The state will become *ASYNC when
catch-up mode ends.

*SYNC An active remote journal link is connected using synchronous delivery


mode.

*ASYNC An active remote journal link is connected using asynchronous delivery


mode.

201
Starting a remote journal link independently
If they are not already active, the remote journal (RJ) links used by replication and
system manager processes are started by operations that start those processes, such
as the STRMMX, STRDG, or STRMMXMGR commands.
To start a remote journal link separately from other MIMIX processes, do the
following:
1. Do one of the following to access a subset of RJ links for the type of process by
which they are used:
• To access the RJ links for a data group on the Work with Data Groups display,
type 44 (RJ links) next to the data group you want and press Enter.
• To access the RJ links for system manager processes on the Work with
Systems display, type 7 (System manager status) next to the system you want
and press Enter. On the resulting Work with System Pairs display, type 12
(Work with RJ links) next to the pair of systems for the system manager
process you want and press Enter.
2. From the Work with Remote Journal Links display, type a 9 (Start) next to the link
in the list that you want to start and press Enter.
3. The Start Remote Journal Link (STRRJLNK) display appears. Specify the value
you want for the Starting journal receiver prompt.
4. To start remote journaling for the specified link, press Enter.

Ending a remote journal link independently


Remote journal (RJ) links used by replication and system manager processes are
normally not ended by operations that end those processes. However, commands
which end those operations have the ability to optionally end the RJ links (ENDMMX,
ENDDG, or ENDMMXMGR commands).
This topic describes how to end an RJ link independently of other MIMIX processes.
Important!
The default value *IMMED for the End remote journal link (ENDOPT) parameter of
the ENDRJLNK command used in these instructions will perform an immediate
end for the specified RJ link. The following instructions allow you to specify
whether to perform an immediate or controlled end of the specified RJ link.
Regardless of the value specified for ENDOPT, an immediate end is always
performed in the following situations:
• The remote journal function is running in synchronous mode
(DELIVERY(*SYNC)).
• The remote journal function is performing catch-up processing.

To end a remote journal link separately from other MIMIX processes, do the following:
1. Do one of the following to access a subset of RJ links for the type of process by

202
Ending a remote journal link independently

which they are used:


• To access the RJ links for a data group on the Work with Data Groups display,
type 44 (RJ links) next to the data group you want and press Enter.
• To access the RJ links for system manager processes on the Work with
Systems display, type 7 (System manager status) next to the system you want
and press Enter. On the resulting Work with System Pairs display, type 12
(Work with RJ links) next to the pair of systems for the system manager
process you want and press Enter.
2. From the Work with Remote Journal Links display, type a 10 (End) next to the link
in the list that you want to end.
3. Do one of the following:
• To perform an immediate end from the source system, press Enter. This
completes the procedure for an immediate end.
• To perform a controlled end or to end from the target system, press F4
(Prompt), then continue with the next step.
4. The End Remote Journal Link (ENDRJLNK) display appears. Press F10
(Additional parameters).
5. To perform a controlled end, specify *CNTRLD at the End remote journal link
prompt. If you need to end from the target system, specify *TGT at the End RJ link
on system prompt. To process the request, press Enter.

203
Working with notifications and recoveries

CHAPTER 10Working with notifications and


recoveries

This topic describes what notifications and recoveries are and how to work with them.
This chapter includes the following topics:
• “What are notifications and recoveries” on page 204 defines terms used for
discussing notifications and recoveries and identifies the sources that create
them.
• “Displaying notifications” on page 205 identifies where notifications are viewed in
the user interfaces and how to work with them.
• “Notifications for newly created objects” on page 210 describes the MIMIX®
AutoNotify™ feature which can be used to monitor for newly created libraries,
folders, or directories.
• “Displaying recoveries” on page 211 identifies where recoveries in progress are
viewed in the user interfaces and how to work with them.

What are notifications and recoveries


A notification is the resulting automatic report associated with an event that has
already occurred. The severity of a notification is reflected in the overall status of the
installation.
Notifications can be generated in a variety of ways:
• Target journal inspection processes generate notifications when users or
programs other than MIMIX have changed objects on the target node.
• Recovery processing generates a notification when an object within the replication
namespace of a data group cannot be recovered because the object exists only
on the target node.
• The recovery phase of a virtual switch procedure generates one notification per
affected data group when recovery actions for objects changed on the designated
test node cannot be performed because the objects do not exist on the source
node.
• Rules that are not associated with the audits provided by MIMIX also generate
notifications to indicate that rule processing either ended in error or, if requested,
completed successfully.
• Shipped monitors, such as the MMNFYNEWE monitor for the MIMIX®
AutoNotify™ feature, generate notifications.
• Software install or upgrade processes may generate notifications for specific
scenarios within your configuration of which you should be aware.
• Custom automation may initiate user-generated notifications when user-defined

204
Displaying notifications

events are detected. User-generated notifications can be set to indicate a failure,


a warning, or a successful operation.
• Audits generate notifications as a secondary mechanism for reporting when the
activities performed by an audit complete or end in error. These notifications are
automatically marked as acknowledged. (The primary mechanism is to report
errors through replication processes and the audit summary.) Policies provide
considerable control over notifications generated by audits.
Because the manner in which notifications are generated can vary, it is important to
note that notifications can represent both real-time events as well as events that
occurred in the past but, due to scheduling, are being reported in the present. For
example, the ownership of a file is changed on the target system at 8:00 PM. If your
audit (CMPFILA) is scheduled to run at 1:00 AM, MIMIX will detect the change and
push a notification to the user interface when the audit completes. Previously,
detection of the change was contingent upon you viewing a report after the audit
completed and noticing the difference.
Recoveries - The term recovery is used in two ways. The most common use refers to
the recovery action taken by audits or replication processes to correct a detected
difference when automatic recovery polices are enabled. The second use refers to a
temporary report that provides details about a recovery action in progress. The report
is automatically created when the recovery action starts and is removed when it
completes. While it exists, the report identifies what originated the action and what is
being acted upon, and may include access to an associated output file (outfile) and
the job log for the associated job. The action which generated a report may also
generate a notification when the recovery action ends.
Note: Details for certain types of recoveries can only be viewed from Vision
Solutions Portal using the MIMIX portal application.

Displaying notifications
In environments with application groups, the Notifications field located in the upper
right corner of the Work with Application Groups display indicates the highest severity
level of new notifications for the MIMIX installation.
You can see a count of how many new notifications require action or attention for a
MIMIX installation from the Work with Data Groups display.
Do the following:
1. Optional. To access the Work with Data Groups display, either select option 6
from the MIMIX Basic Main menu or select option 1 from the MIMIX Intermediate
Main menu and press Enter.
The Notif. field is located in the upper right of the display. The third number

205
Working with notifications and recoveries

identifies the number of new notifications that require action or attention.


Work with Data Groups CHICAGO
11:02:05
Type options, press Enter. Audits/Recov./Notif.: 079 / 002 / 003
5=Display definition 8=Display status 9=Start DG
10=End DG 12=Files needing attention 13=Objects in error
14=Active objects 15=Planned switch 16=Unplanned switch ...
---------Source--------- --------Target-------- -Errors-
Opt Data Group System Mgr DB Obj DA System Mgr DB Obj DB Obj

Highlighting on the Notif. field indicates the severity of the errors.


Red — At least one notification with a severity of *ERROR exists.
Action is required to resolve the error.
Yellow — At least one notification with a severity of *WARNING exists.
Investigation may be necessary. An operation completed but an error
may exist.
2. Do one of the following to access the Work with Notifications display:
• From the Work with Application Groups display, press F15 (Notifications).
• From the Work with Data Groups display, press F8 (Recoveries), followed by
F10 (Work with Notifications).
• From a command line, enter the command WRKNFY.
The list is sorted so that new notifications appear at the top. The list includes
notifications from all sources.
3. To see details for a notification, type a 5 (Display) next to the notification you want
and press Enter.
Note: Notifications from audits are automatically set to a status of
acknowledged. Audit status and results should be checked from the Work
with Audits (WRKAUD) display.
4. Once you have resolved the problem identified by a new notification, you should
use option 46 to acknowledge that the reported problem has been resolved. Once
acknowledged, its severity is no longer reported in status information.

What information is available for notifications


The following information is available for notifications listed on the Work with
Notifications display. The F11 key toggles between views of status, timestamp, and
text of the notification.
Additional details are available for each notification through the Display Notification
Details display.
Status - The Work with Notifications display lists notifications grouped by their status.
*NEW - New notifications have not been acknowledged or removed and their
status is reflected in higher level status.
*ACK - Acknowledged notifications are archived as viewed and their status is no
longer reflected in higher level status.

206
Displaying notifications

Severity - Identifies the severity level of the notification.


*ERROR - An error occurred that requires immediate action.
*WARNING - Investigation may be necessary. An operation completed but an
error may exist. For example, the MIMIX AutoNotify feature issues notifications
with this severity that identify newly created objects that are not identified for
replication.
*INFO - No user intervention is required.
Notification - Displays the notification text sent by audits, automatic recoveries,
target journal inspection, monitors, user-defined or MIMIX rules, or a user-generated
notification. To view the full text, use option 5 to display the notification details.
Data group - Identifies the data group associated with the notification. User-defined
notifications and notifications from monitors or user-defined rules may indicate that
there is no associated data group.
Note: On the Status view of the Work with Notifications display, the F7 key toggles
between the Source column and the Data Group column. The full three-part
name is available in the Timestamp view (F11).
Date - Indicates the date the notification was sent.
Time - Indicates the time the notification was sent.
Source - Identifies the process, program, or command that generated the recovery.
Names that begin with the character # are generated by automatic recovery actions
for audits.1 Names that begin with the characters ## are generated by automatic
recovery actions for object replication that used the third delay retry processing.
From System - Identifies the name of the system on which the notification was
generated. The name From System is used on the Timestamp view (F11) of the Work
with Notifications display. When you display the notification details from the native
user interface, this is called the Originating system.

Detailed information
When you display a notification, you see its description, status, severity, data group,
source, and sender as described above. You also have access to the following
information:
Details - When the source of the notification is a rule, this identifies the command that
was initiated by the rule. When the source of the notification is user-generated, this
indicates the notification detail text specified when the notification entry was added.
When the source of the notification is a monitor, this describes the events that
resulted in the notification.
Output File - If available, this identifies an associated output file. Output file
information associated with a notification is only available from the sender system.
For user-generated notifications, output file information is available only if it was
specified when the notification was added.

1. In environments configured for use with IBM i clustering, several other processes besides
audits can generate recoveries with a source name beginning with the character #.

207
Working with notifications and recoveries

Job - If available, this identifies the job that generated the notification. Job information
associated with a notification is only available from the sender system. For user-
generated notifications, this information is available only if it was specified when the
notification was added.

Options for working with notifications


Table 36 identifies the possible actions you can take for a notification. From the
Notifications window, the Actions list for each notification contains only the actions
possible for the selected notification.

Table 36. Options available for notifications

Option Description

4=Remove Deletes the notification. You are prompted to confirm this choice.
For a notification generated by an audit or a MIMIX rule, the
associated job and output files are also deleted.
This must be performed from the system on which the notification
originated.

5=Display Displays available additional information associated with the


notification.
For notifications generated by rules, this includes the details of the
rule that generated the notification, including the substitution
variables for the command the rule initiated

6=Print Prints the information associated with the notification.

8=View results When the information is available, this provides the Name and
Library of the output file (outfile) associated with the notification.
This option is only available from the system on which the
notification originated.1

12=Display job Displays the job log for the job which generated the notification, if
it is available. This option is only available from the system on
which the notification originated

46=Acknowledge Sets the selected notification status to *ACK (Acknowledged).

47=Mark as new Sets the selected notification status to *NEW (New).


1. MIMIX manages an output file associated with a notification from an automatically recovery
action or a MIMIX rule when the output file exists in a specific library. The format of the library
name for such an output file is MIMIX-installation-library_0.

Displaying IFS entries associated with notifications from INSTALL


When a MIMIX instance is upgraded from an earlier software level to MIMIX 8.1.12.00
or higher, you may receive notifications whose source is “INSTALL”. These
notifications are sent because the upgrade process detected at least one data group
IFS entry within the configuration that has one or more asterisk (*) characters

208
Displaying notifications

specified in positions other than the last (trailing) character of the specified name. The
names in these IFS entries are considered advanced generic names.
To check for notifications about advanced generic names in IFS entries and locate the
affected IFS entries:
1. Do one of the following:
• On the Work with Application Groups display, the Notifications field indicates
whether any warning notifications exist. Press F15 (Notifications).
• On the Work with Data Groups display, the third number in the
Audits/Recov./Notif. field displays the number of new notifications. Press F8
(Recoveries), then press F10 (Work with Notifications).
• From a command line, enter: WRKNFY.
2. The Work with Notifications display appears. Notifications about advanced
generic names are identified by the name INSTALL in the Source column.
3. Press F11 (View Timestamp) to see the name of the affected data group.
4. From the command line, specify the data group in the following command:
WRKDGIFSE DGDFN(name system1 system2)
5. On the Work with DG IFS Entries display, check for entries that display either
*INCLD(A) or *EXCLD(A) in the Process Type column.
These entries have embedded asterisks and are considered advanced generic
names, which are important because:
• An advanced generic name in an *INCLD IFS entry will have any asterisk
characters embedded within the name interpreted as literal characters, not
wildcard characters. These entries cannot be changed or copied. They should
be evaluated, and if possible removed to avoid confusion with how asterisks
are interpreted.
• An advanced generic name in an *EXCLD IFS entry will be processed
differently after upgrading to 8.1.12.00 or higher. Asterisks embedded within its
name are processed as wildcard characters which act as a filter to exclude
objects that would otherwise be included by an existing entry with a basic
generic name.
For more information about advanced generic names, see the topic “Identifying IFS
objects for replication” in the MIMIX Administrator Reference book.

209
Notifications for newly created objects
The MIMIX® AutoNotify™ feature can be used to monitor for newly created libraries,
folders, or directories. The AutoNotify feature uses a shipped journal monitor called
MMNFYNEWE to monitor for new objects in an installation that are not already
included or excluded for replication by a data group. The AutoNotify feature monitors
the security audit journal (QAUDJRN), and when new objects are detected, issues a
warning notification.
The MMNFYNEWE monitor is shipped in a disabled state. In order to use this feature,
the MMNFYNEWE monitor must be enabled on the source system within your MIMIX
environment. Once enabled, this monitor will automatically start with the master
monitor.
Notifications will be sent when newly created objects meet the following conditions:
• The installation must have a data group configured whose source system is
the system the monitor is running on.
• The journal entry must be a create object (T-CO) or object management
change (T-OM).
• If the journal entry is a create object (T-CO), then the type must be new (N).
• The journal entry must be for a library, folder, or directory.
• If the journal entry is for a library, it cannot be a MIMIX generated library
since MIMIX generated libraries are not replicated by MIMIX.
• If the journal entry is for a directory, it cannot be the /LAKEVIEWTECH
directory, or any directory under /LAKEVIEWTECH.
• If the journal entry is for a directory, it must be a directory that is supported
for replication by MIMIX.
• The object is not already known (included or excluded) in the installation.
From the native user interface, notifications can be viewed from the Work with
Notifications (WRKNFY) display. The notification message will indicate required
actions. Additionally, when you view notifications from the MMNFYNEWE monitor
from Vision Solutions Portal, you have immediate access to actions to create
configuration entries for it.

210
Displaying recoveries

Displaying recoveries
Active recoveries are an indication of detected problems that are currently being
corrected by automatic recovery actions. Before certain activity, such as ending
MIMIX, it is important that no recoveries are in progress in the installation.
You can see a count of how many recoveries are in progress for a MIMIX installation
from the Work with Data Groups display.
Do the following:
1. To access the Work with Data Groups display, either select option 6 from the
MIMIX Basic Main menu or select option 1 from the MIMIX Intermediate Main
menu and press Enter.
2. The Recov. field is located in the upper right of the display. The middle number
identifies the number of new or in-progress recoveries. This includes recoveries
resulting from all sources.
Work with Data Groups CHICAGO
11:02:05
Type options, press Enter. Audits/Recov./Notif.: 079 / 002 / 003
5=Display definition 8=Display status 9=Start DG
10=End DG 12=Files needing attention 13=Objects in error
14=Active objects 15=Planned switch 16=Unplanned switch ...
---------Source--------- --------Target-------- -Errors-
Opt Data Group System Mgr DB Obj DA System Mgr DB Obj DB Obj

Note: Only recoveries initiated by audits or by object replication’s third delay


retry processing can be viewed from the native user interface. Recoveries
from all sources can be viewed from Vision Solutions Portal on the Object
Correction Activity window.
3. To view recoveries available in the native user interface, do the following:
a. Press F8 (Recoveries).
The Work with Recoveries display lists new, in-progress, and held recoveries
initiated by audits and object third delay retry processing.
b. To see details for a recovery, type a 5 (Display) next to the recovery you want
and press Enter.
4. To view all recoveries for the installation, do the following:
a. Log in to Vision Solutions Portal.
b. If necessary, create portal connections and add an instance to access the
MIMIX installation.
c. Select the Correction Activity action from the global menu on either the
Application Groups portlet or the Data Groups portlet.
d. For more information select the “i” icon to access online help.

211
What information is available for recoveries in the native user interface
The following information is available for the recoveries listed on the Work with
Recoveries display. The F11 key toggles between views of status, timestamp, and
text of the recoveries. Additional details are available for each recovery through the
Display Recovery Details display.
Each recovery provides a brief description of the recovery process taking place as
well as its current status.
Status - Shows the status of the recovery action.
*ACTIVE - The job associated with the recovery is active.
*ENDING - The job associated with the recovery is ending.
*HELD - The job associated with the recovery is held. A recovery whose source is
a replication process cannot be held.
Data group - Identifies the data group associated with the recovery.
Note: On the Status view of the Work with Recoveries display, the F7 key toggles
between the Source column and the Data Group column. The full three-part
name is available in the Timestamp view (F11).
Date - Indicates the date the recovery process started.
Time - Indicates the time the recovery process started.
Source - Identifies the process, program, or command that generated the recovery.
Names that begin with the character # are generated by automatic recovery actions
for audits.1 Names that begin with the characters ## are generated by automatic
recovery actions for object replication that used the third delay retry processing.
Sender or From System - Identifies the system from which the recovery originated.

Detailed information
When you display a recovery, you see its description, status, data group, source, and
sender as described above. You also have access to the following information.
Details - When the source of the recovery is a rule, this identifies the command run by
the rule in an attempt to recover from the detected error.
Output File - If available, this identifies an associated output file that lists the
detected errors the recovery is attempting to correct. Output file information
associated with a recovery is only available from the sender system.
Job - If available, this identifies the job that is performing the recovery action. Job
information associated with a recovery is only available from the sender system.

1. In environments configured for use with IBM i clustering, several other processes besides
audits can generate recoveries with a source name beginning with the character #.

212
Displaying recoveries

Options for working with recoveries in the native user interface


Table 37 identifies the possible actions you can take for a recovery. From the
Recoveries window, the Actions list for each recovery contains only the actions
possible for the selected recovery.

Table 37. Options available for recoveries

WRKRCY Description
Option

4=Remove Removes the specified recovery, if it is not held or active. A


confirmation panel is displayed after pressing Enter. Use this option to
remove orphaned recoveries whose associated recovery job ended.
This option is only available from the system on which the recovery job
ran.

5=Display Displays available additional information associated with the recovery.

6=Print Prints the information associated with the recovery.

8=View progress Displays a filtered view of the output file associated with the recovery.
MIMIX updates the output file while the recovery is in progress,
identifying the detected errors it is attempting to correct and marking
corrected errors as being recovered. This option is only available from
the system on which the recovery job is running.

10=End job Ends an active recovery job. This action is valid for recoveries with
names that begin with # and is only available from the system on
which the recovery job is running.

12=Display job Displays the job log for the recovery job associated in progress. This
option is only available from the system on which the recovery job is
running.

13=Hold job Places an active recovery job on hold. This action is valid for
recoveries with names that begin with # and is only available from the
system on which the recovery job is running.

14=Release job Releases a held recovery job. This action is valid for recoveries with
names that begin with # and is only available from the system on
which the recovery job is held.

Orphaned recoveries in the native user interface


There are times when recoveries exist but are no longer associated with a job. The
following conditions could cause recoveries to become orphaned:
• An unplanned switch has occurred
• The MIMIX subsystem was ended unexpectedly
• A recovery job was ended unexpectedly
When automatic audit recovery is enabled, orphaned recoveries are converted to
error notifications during system cleanup. If the orphaned recovery is older than the
cleanup time specified in the system definition, it is deleted.

213
When automatic database recovery or automatic object recovery is enabled,
orphaned recoveries are deleted, when possible.
Because recoveries are displayed on both systems, but jobs associated with them are
only accessible from the originating system, you need to verify that the recovery is
orphaned before removing it.

Determining whether a recovery is orphaned


Do the following to determine whether a recovery is orphaned:
1. From a command line, type WRKRCY and press Enter.
2. Press F11 to display the Timestamp view. This view allows you to see the From
System column which lists the system from which the recovery originated.
3. Ensure you are operating from the originating system. Then type a 12 next to the
recovery.
4. Do one of the following:
• If an error message is displayed indicating that the job associated with the
recovery is not found, follow the steps in “Removing an orphaned recovery” on
page 214.
• When the Display Job display appears, type a 10 in the Selection field and
press Enter. The status of the job is displayed. If the job associated with the
recovery is no longer valid, follow the steps in “Removing an orphaned
recovery” on page 214.

Removing an orphaned recovery


These procedures assume that you have already confirmed that the recovery is
orphaned using the procedures in “Determining whether a recovery is orphaned” on
page 214.
Do the following to remove an orphaned recovery:
1. From the originating system, type WRKRCY on the command line and press Enter.
2. After you have ensured that the recovery is orphaned, type a 4 next to the
orphaned recovery you wish to remove and press Enter.
3. Press Enter to confirm your request to remove the recovery.

214
CHAPTER 11 Starting and ending replication

MIMIX uses a number of processes to perform replication. These processes, along


with a number of supporting processes must be active to enable MIMIX to function.
These pairs of commands will start and end replication:
• The Start MIMIX (STRMMX) and End MIMIX (ENDMMX) commands will start or
stop replication processes as well as all supporting processes for the products in a
MIMIX installation library in a single operation. These commands are the preferred
method for starting and ending MIMIX.
• The Start Application Group (STRAG) and End Application Group (ENDAG)
commands will start or stop replication processes in environments configured with
application groups. Each command calls a default procedure with steps to perform
its operations and can be customized.
• The Start Data Group (STRDG) and End Data Group (ENDDG) commands will
start or stop data group replication processes. These commands are the basis for
controlling replication processes and are invoked programmatically by the
previously identified commands.
This chapter provides information about and procedures for use each set of
commands. The following topics are included:
• “Before starting replication” on page 217 applies to all methods of starting
replication.
• “Commands for starting replication” on page 217 describes the STRMMX,
STRAG, and STRDG commands and considerations for their use.
• “What occurs when a data group is started” on page 221 describes what the
STRDG command does in addition to starting replication, choices for specifying a
journal starting point, and options for clearing pending and error entries.
• “Starting MIMIX” on page 226 provides a procedure for using the STRMMX
command.
• “Starting an application group” on page 227 provides a procedure for using the
STRAG command.
• “Starting selected data group processes” on page 228 provides a procedure for
using the STRDG command and identifies when the start request should include
clearing pending entries.
• “Starting replication when open commit cycles exist” on page 230 describes when
MIMIX cannot start replication due to open commit cycles and how to resolve
them and start replication.
• “Before ending replication” on page 232 to all methods of ending replication.
• “Commands for ending replication” on page 232 describes the ENDMMX,
ENDAG, and ENDDG commands and considerations for their use, such as when
to perform a controlled end or when to end the RJ link.

215
Starting and ending replication

• “What occurs when a data group is ended” on page 238 describes the behavior of
the ENDDG command.
• “Ending MIMIX” on page 240 provides procedures for using the ENDMMX
command and describes when you may also need to end the MIMIX subsystem.
• “Ending an application group” on page 243 provides a procedure for using the
ENDAG command.
• “Ending a data group in a controlled manner” on page 244 provides procedures
for preparing to end, ending, and confirming that the end completed without
problems.
• “Ending selected data group processes” on page 247 provides a procedure using
the ENDDG command.
• “Enabling MIMIX-disabled target constraints” on page 248 describes the
requirements for enabling foreign key constraints that were previously disabled by
MIMIX and provides procedures for enabling constraints when the data group is
ended or while performing a controlled end.
• “What replication processes are started by the STRDG command” on page 251
describes which replication processes are started with each possible value of the
Start processes (PRC) parameter. Both data groups configured for remote
journaling and data groups configured for MIMIX source-send processing are
addressed.
• “What replication processes are ended by the ENDDG command” on page 255
describes what replication processes are ended with each possible value for the
End Options (PRC) parameter. Both data groups configured for remote journaling
and data groups configured for MIMIX source-send processing are addressed.

216
Before starting replication

Before starting replication


Consider the following:
• Before starting replication, the database files and objects to be replicated by a
data group must be synchronized between the systems defined to the data group.
For more information about performing the initial synchronization, see the MIMIX
Administrator Reference book.
• If you are using the MIMIX for MQ function, you must use the procedures in the
MIMIX for IBM WebSphere MQ book for initial synchronization and initial start of
data groups that replicate data for IBM WebSphere MQ.
• Data groups that are in a disabled state are not started. Only data groups that
have been enabled can be started.

Commands for starting replication


These commands start replication processes. The significant differences between
these commands are:
Start MIMIX (STRMMX) – The STRMMX command will start all MIMIX processes
in a MIMIX installation, including those used for replication, in a single operation
regardless of how replication is configured. This is the preferred method of
starting MIMIX. Optionally, this command can be used to start replication and
supporting processes associated with the local system.
Start Application Group (STRAG) – The STRAG command will start replication
processes for data groups that are part of an application group. This is the
preferred method of starting replication in application groups. The command
invokes a procedure which performs the operations to start replication for the
participating data groups.
Start Data Group (STRDG) – The STRDG command will start replication
processes for a data group and the remote journal link, if necessary. This
command is the basis for all other methods of starting replication. Optionally, this
command can specify a starting point in the journals, clear any pending or error
entries, set object auditing levels, and start a subset of the replication processes.

What is started with the STRMMX command


The STRMMX command is shipped with default values that will start all MIMIX
replication and supporting processes on all systems in the installation. Processes are
started in the following order:
MIMIX managers and services - All jobs for the system managers, journal
managers, target journal inspection, and collector services are started on the
specified systems. Remote journal links for system managers are started. If you
are using MIMIX with IBM i clustering, Cluster Services are started for all specified
systems that are configured for clustering.
Data groups - For enabled data groups, starts the replication processes, remote

217
journal links, and automatic recovery processes on the specified systems. Each
data group starts from the journal receiver in use when the data group ended and
with the sequence number following the last sequence number processed.
Master monitor - Starts the master monitor on each of the specified systems.
Monitors - On each of the specified systems, the master monitor starts the MIMIX
scheduler job and monitors that are not disabled and which are configured to start
with the master monitor.
Application groups - If all systems are specified, all application groups and any
associated data resource groups are started. If IBM i clustering is used, default
processing will start the IBM application CRG.
Note: The STRMMX command does not start promoter group activity. Start promoter
group activity using procedures in the Using MIMIX Promoter book.
Optionally, the command can be used to start MIMIX processes associated with the
local system. When the command request specifies *LOCAL for the System definition
(SYSDFN) parameter, single system processes on the local system are started and
multiple-system processes between the local system and another system are started.
Single system processes include the master monitor, MIMIX scheduler, monitors,
journal manager, target journal inspection, collector services, and cluster services.
Multiple-system processes include system managers, RJ links for system managers
and replication processes, data group replication processes and their supporting
processes. Application groups are not started.

STRMMX and ENDMMX messages


Once you have run the STRMMX or ENDMMX command, one of the following
messages is displayed:
• Completion LVI0902 – This message indicates that all MIMIX products were
started or ended successfully.
• Escape LVE0902 – This message indicates one or more MIMIX products failed to
start or end.

What is started by the default START procedure for an application group


The Start Application Group (STRAG) command automatically uses an application
group’s default START procedure unless you specify a different procedure.
When an application group is created, a default procedure named START is created
for it from a shipped default procedure. The procedure starts replication processes for
enabled data groups within a single application group. Steps in the shipped default
START procedure are described in the MIMIX Administrator Reference book.

Choices when starting or ending an application group


For the purpose of describing their use, the Start Application Group (STRAG) and End
Application Group (ENDAG) commands are quite similar. This topic describes their

218
Commands for starting replication

behavior for application groups that do not participate in a cluster controlled by the
IBM i operating system (*NONCLU application groups).
What is the scope of the request? The following parameters identify the scope of
the requested operation:
Application group definition (AGDFN) - Specifies the requested application group.
You can either specify a name or the value *ALL.
Resource groups (TYPE) - Specifies the types of resource groups to be
processed for the requested application group.
Data resource group entry (DTARSCGRP) - Specifies the data resource groups to
include in the request. The default is *ALL or you can specify a name. This
parameter is ignored when TYPE is *ALL or *APP.
What is the requested behavior? The following parameters, when available, define
the expected behavior:
Current node roles (ROLE) - Only available on the STRAG command, this
parameter is ignored for non-cluster application groups.
What procedure will be used? The following parameters identify the procedure to
use and its starting point:
Begin at step (STEP) - Specifies where the request will start within the specified
procedure. This parameter is described in detail below.
Procedure (PROC) - Specifies the name of the procedure to run to perform the
requested operation when starting from its first step. The value *DFT will use the
procedure designated as the default for the application group. The value
*LASTRUN uses the same procedure used for the previous run of the command.
You can also specify the name of a procedure that is valid the specified
application group and type of request.
Where should the procedure begin? The value specified for the Begin at step
(STEP) parameter on the request to run the procedure determines the step at which
the procedure will start. The status of the last run of the procedure determines which
values are valid.
The default value, *FIRST, will start the specified procedure at its first step. This value
can be used when the procedure has never been run, when its previous run
completed (*COMPLETED or *COMPERR), or when a user acknowledged the status
of its previous run which failed, was canceled, or completed with errors
(*ACKFAILED, *ACKCANCEL, or *ACKERR respectively).
Other values are for resolving problems with a failed or canceled procedure. When a
procedure fails or is canceled, subsequent attempts to run the same procedure will
fail until user action is taken. You will need to determine the best course of action for
your environment based on the implications of the canceled or failed steps and any
steps which completed.
The value *RESUME will start the last run of the procedure beginning with the step at
which it failed, the step that was canceled in response to an error, or the step
following where the procedure was canceled. The value *RESUME may be
appropriate after you have investigated and resolved the problem which caused the

219
procedure to end. Optionally, if the problem cannot be resolved and you want to
resume the procedure anyway, you can override the attributes of a step before
resuming the procedure.
The value *OVERRIDE will override the status of all runs of the specified procedure
that did not complete. The *FAILED or *CANCELED status of these procedures are
changed to acknowledged (*ACKFAILED or *ACKCANCEL) and a new run of the
procedure begins at the first step.

220
What occurs when a data group is started

What occurs when a data group is started


The Start Data Group (STRDG) command will start the replication processes for the
specified data group.
The STRDG command can be used interactively or programatically. Default values for
the command are used when it is invoked by the STRMMX command or by the
STRAG command running the default START procedure.
When a STRDG request is processed, MIMIX may take a few minutes while it does
the following for each specified data group:
• Determines whether the RJ links for replication and system managers processes
are active and whether all required system managers and journal managers on
each system are started. If necessary, the managers and the remote journal
functions defined by the RJ links are started.
• Determines the starting point for replication (database, object, or both, as
configured).
• Locates the starting point in the appropriate journal receiver. This will be the
starting point for send processes.
• If necessary, changes the object audit level of existing objects identified for
replication. This occurs when starting following a switch or a configuration change
to any data group object, IFS, or DLO entry. This ensures that all replicated
objects identified by all entries of each entry type are set with an object audit level
suitable for replication. The processing order for data group entries can affect the
auditing value of IFS objects. For examples and for information about manually
specifying the audit level of objects, see the MIMIX Administrator Reference book.
• Submits the appropriate start requests for the processes specified on the start
request.
• Makes configuration changes for the data group become effective. If a
configuration change affects the set of objects to be replicated, the start request
also automatically deploys the configuration changes to an internal list used by
other functions. This may cause the start request to take longer.
• If there are pending configuration entries that specified *YES for Synchronize on
start (SYNCONSTR) and which added objects into the replication name space,
requests to synchronize those objects using system journal replication processes
are submitted.
• Attempts to recover any existing access path maintenance errors for the data
group, if the Access path maintenance (APMNT) policy is enabled.
• If specified on the start request, clears all pending entries for apply processes and
clears all error entries identified in replication processing for the data group. There
are times when it is necessary to clear pending entries, error entries, or both, to
establish a new synchronization point for the data group.
Starting a data group may take longer if the remote journal function is operating in
catchup mode.

221
If a data group is participating in a virtual switch when you initiate a start data group
request, the data group’s apply processes will not be started. The apply processes
remain ended and will only be started by the virtual switch procedure during its
recovery phase.

Journal starting point identified on the STRDG request


On the STRDG command, you can optionally specify the point at which to start
replication in the journal receivers. The parameters for database and object journal
receivers and sequence numbers provide this capability. You may need to use these
parameters when starting data groups for the first time.
• For user journal replication, the IBM i remote journal function controls where
processing starts in the source journal receiver. The values specified for the
Database journal receiver (DBJRNRCV) and Database large sequence number
(DBSEQNBR2) identify the starting location for the database reader process and
the database apply process.
• For system journal replication, the value specified for Object journal receiver
(OBJJRNRCV) and Object large sequence number (OBJSEQNBR2) identify the
starting location for the object send process and the object apply process.
Note: The parameters Database sequence number (DBSEQNBR) and Object
sequence number (OBJSEQNBR) continue to be valid for journal
definitions which specify *MAXOPT2 for the Receiver size option
(RCVSIZEOPT) and for values that do not exceed 10 digits. To ensure
continued compatibility, the use of parameters DBSEQNBR2 and
OBJSEQNBR2 is recommended.

Journal starting point when the object send process is shared


When starting data groups that share an object send process, the first data group to
start will start the shared job at that data group’s starting point in the system journal
(QAUDJRN). As additional data groups start, each recognizes that the shared object
send job is active. The object send job determines whether the starting point for that
data group is earlier or later than the sequence number being read. If the data group’s
starting point is later, replication will begin when the shared job reaches the data
group's starting point. If the data group’s starting point is earlier, the shared job
completes its current block of entries, then returns to the earliest point for any of the
data groups being started. The shared job reads the earlier entries and routes the
transactions to the data group being started. When the shared job reaches the last
entry it read at the time of the STRDG request, it resumes routing transactions to all
active data groups using the shared job.
If the starting data group has a significant object send backlog, the other data groups
sharing the job will not receive transactions to replicate while the backlog for the
starting data group is being addressed. Therefore, when a significant backlog exists, it
is recommended that you change the data group configuration to use a dedicated job
(*DGDFN for object send prefix), start the data group, and allow it to catch up to the
current location of the shared job. Then end the data group, change its configuration
to use the desired shared job, and restart the data group.

222
What occurs when a data group is started

Clear pending and clear error processing


The Clear pending and Clear error prompts on the STRDG command provide
flexibility when starting a data group by allowing you to optionally reset error status
conditions on data group file entries and discard pending journal entries that are
stored in the journal log space. Clear pending resets the starting point for all data
group file entries and object entries. Clear error clears the hold log spaces.
When clearing pending entries, you can optionally specify which system to use for
determining database file network relationships when distributing files among
database apply sessions. The System for DB file relations (DBRSYS) prompt
identifies which system is used to assign data group file entries to apply sessions
when the start request specifies to clear pending entries in all apply sessions.
Table 38 shows the processing that occurs based on the selection made for the Clear
pending (CLRPND) and Clear error (CLRERR) prompts. The Clear pending and Clear
error prompts work independently. For example, when CLRPND(*NO) is selected, no
clear pending processing occurs.

Table 38. CLRPND and CLRERR processing

CLRPND CLRERR Processing Description Notes

*NO *NO Data groups start with regular processing: When CLRPND is *NO,
or • Data group file entry status remains unchanged. the value *CLRPND is
*CLRPND • Hold logs remain unchanged. the same as *NO for
CLRERR.

*NO *YES • Data group file entries in *HLDERR, *HLDRGZ, See File entry states
*HLDRNM, *HLDPRM, and *HLDRLTD status are See Log spaces
cleared.
• Tracking entries in *HLDERR status are cleared.
• Hold log space is deleted.
• Data group object activity entries in any failed
status are changed to CC (Completed by clear
request).

223
Table 38. CLRPND and CLRERR processing (Continued)

CLRPND CLRERR Processing Description Notes

*YES *NO Note: CLRPND(*YES) will not start a data group when See File entry apply
there are open commit cycles on files defined to the session assignment
data group.
See Single apply session
• Data group file entries in *HLDRGZ, *HLDRNM, processing
and *HLDPRM status are cleared and reset to
See Log spaces
active.
• Data group tracking entries in *HLDRNM are
cleared and reset to active.
• Data group file entries and tracking entries in
*HLDERR status remain unchanged.
• Data group object activity entries in any active
status are changed to CC (Completed by clear
request).
• If there is a requested status at the time of starting,
it is cleared.
• Journal, hold, tracking entry hold, and apply
history log spaces are deleted.
• The apply session to which data group file entries
are assigned may change.

*YES *YES Note: CLRPND(*YES) will not start a data group when See File entry states
or there are open commit cycles on files defined to the See File entry apply
data group.
*CLRPND session assignment
• Data group file entries in *HLDERR, *HLDRGZ, See Single apply session
*HLDRNM, *HLDPRM, and *HLDRLTD status are processing
cleared and reset to active.
See Log spaces
• Data group file entries in *HLDRTY status remain
When CLRPND is *YES,
unchanged.
the value *CLRPND is
• Data group object activity entries in any failed or the same as *YES for
active status are changed to CC (Completed by CLRERR.
clear request).
• Tracking entries in *HLDERR and *HLDRNM
status are cleared and reset to active.
• Tracking entries are primed if any configuration
changes occurred for data group object entries or
data group IFS entries.
• If there is a requested status at the time of starting,
it is cleared.
• Journal, hold, and apply history log spaces are
deleted.
• The apply session to which data group file entries
are assigned may change.

224
What occurs when a data group is started

Table 38. CLRPND and CLRERR processing (Continued)

CLRPND CLRERR Processing Description Notes

File entry states: Files in specific states will not reset to active when you specify *YES on the Clear Error
prompt. If you have set data group file entries to any of these states, the following process exception
applies:
Note: The only states that can be set using the Set Data Group File Entry (SETDGFE) command are *HLD,
*RLSWAIT, *ACTIVE, and *HLDIGN. All other states are the result of internal processing.
• *HLD - Journal entries cached before *YES is specified are discarded. If *ALL or *ALLSRC is specified
on the Start processes prompt, all subsequent entries from the specified starting point will be cached
again.
• *RLSWAIT - Journal entries are discarded as they wait for the synchronization point to arrive in the
journal stream. This occurs regardless of the value specified for Clear Error or Clear Pending.
• *HLDIGN - Journal entries are discarded until the file status is changed to something else.
• *HLDSYNC - Journal entries are ignored since an external process is actively synchronizing the file.
When that event completes normally, the file is set to *RLSWAIT.

File entry apply session assignment: Clear pending processing attempts to load balance the data group
file entries among the defined apply sessions. If the requested apply session in the data group file entry
definition is *ANY, or if it is *DGDFT and the requested apply session for the data group definition is *ANY,
then the apply session to which the data group file entry is assigned may be changed when processing
occurs. For data groups configured to replicate through the user journal and which use basic support for
constraints, the requested apply session may be ignored to ensure that related files are handled by the
same apply session. For data groups configured to allow target constraint management, configured files
with foreign key constraints may be automatically assigned to different apply sessions.
The value specified for System for DB file relations (DBRSYS) determines the system used to determine
database file relationships while assigning files to apply sessions. This parameter is evaluated only when
the start request specifies to clear pending entries for all database apply sessions. The default value, *TGT,
uses the target system to determine the file relationships.

Single apply session processing: In most situations, you will perform clear pending processing on all
apply sessions belonging to a data group by specifying *ALL or *DBALL on the Start processes (PRC)
prompt. MIMIX also supports the ability to perform clear pending processing on a single apply session,
which is useful for recovery purposes in certain error situations. The System for DB file relations (DBRSYS)
parameter is ignored when the start request specifies a specific apply session. To perform clear pending
processing on a single apply session, specify PRC(*DBAPY) and the specific apply session (APYSSN).

Log spaces: Because they have not been applied, journal entries that exist in the journal log space are
considered pending. Journal entries that exist in the hold log space, however, are considered in error. The
Clear pending and Clear error prompts affect which log spaces are deleted (and recreated) when a data
group is started.

225
Starting MIMIX
To start all MIMIX products within an installation library, do the following:
1. If you are starting MIMIX for the first time or starting MIMIX after a system IPL, do
the following:
a. Use the command WRKSBSJOB SBS(MIMIXSBS)to verify that the MIMIX
subsystem is running. If the MIMIXSBS is not already active, start the
subsystem using the STRSBS SBSD(MIMIXQGPL/MIMIXSBS)command.
b. If MIMIX uses TCP/IP for system communication, the TCP/IP servers must be
running. If TCP/IP is not already active, start TCP/IP using the port number
defined in the transfer definitions and the procedures described in “Starting the
TCP/IP server” on page 328.
2. Do one of the following:
• From the MIMIX Basic Main Menu, select option 2 (Start MIMIX) and press
Enter.
• From a command line type STRMMX and press Enter.
3. The Start MIMIX (STRMMX) display appears. Accept the default value for the
System definition prompt and press Enter.
4. If you see a confirmation display, press Enter to start MIMIX.

226
Starting an application group

Starting an application group


For an application group, a procedure for only one operation (start, end, or switch)
can run at a time. For information about parameters and shipped procedures, see
“What is started by the default START procedure for an application group” on
page 218 and “Choices when starting or ending an application group” on page 218.
To start an application group, do the following:
1. From the Work with Application Groups display, type 9 (Start) next to the
application group you want and press F4 (Prompt).
2. Verify that the values you want are specified for Resource groups and Data
resource group entry.
3. If you are starting after addressing problems with the previous start request,
specify the value you want for Begin at step. Be certain that you understand the
effect the value you specify will have on your environment.
4. Press Enter.
5. The Procedure prompt appears. Do one of the following:
• To use the default start procedure, press Enter.
• To use a different start procedure for the application group, specify its name.
Then press Enter.

227
Starting selected data group processes
This procedure can be used to do any of the following:
• Start all or selected processes for a data group, or start a specific database apply
process
• Specify a starting point for journal receivers when starting a data group
• Clear pending and error entries when starting a data group
Data groups that are in an application group: The preferred method of starting
data groups that are part of an application group is to use the Start Application Group
(STRAG) command. The default behavior of the STRDG command helps to enforce
this best practice when necessary by not allowing the command to run when the data
group is participating in a resource group with three or more nodes. (A data resource
group provides the association between one or more data groups and an application
group.) The STRDG request will run when the data group is participating in a resource
group with two nodes. You can also change the value specified for Override if in data
rsc. group (DTACRG) parameter to allow a data group to be started even if it is part of
a resource group.
In application group environments with three or more nodes, it is particularly important
to treat all members of an application group as one entity. For example, a
configuration change that is made effective by starting and ending a single data group
would not be propagated to the other data groups in the same resource group.
However, the same change would be propagated to the other data groups if it is made
effective by ending and starting the parent application group.
When to clear pending entries and entries in error: Table 39 identifies when it is
necessary to clear pending entries for apply processes and clear logs of entries
indicating files in error to establish a new synchronization point when starting a data
group. The reason for starting the data group determines whether you need to clear
only pending entries for transactions waiting to be applied, clear only errors, or both.
Before clearing pending entries, determine if there are any file entries on hold. These
are the transactions that will be lost by clearing pending entries.
When clearing pending entries, most environments can accept the default value for
the System for DB file relations prompt. If necessary, you can specify a value when
directed to by your MIMIX administrator.

228
Starting selected data group processes

Table 39. When to clear pending entries and entries in error when starting a data group

If starting the data group in any of these conditions: Specify these values on the
STRDG command:

After enabling a previously disabled data group Clear pending entries.


Specify *YES for the Clear pending
After changing the Number of DB apply sessions (NBRDBAPY)
prompt
parameter on the data group definition

After changing the value specified for the Target constraint


management element of the Database apply processing
(DBAPYPRC) parameter on the data group definition

After synchronizing database files and objects between two systems Clear pending entries and entries in
Note: This assumes that you have synchronized the objects and database error.
files and have changed the journal receivers using TYPE(*ALL) on • Specify *YES for the Clear
the CHGDGRCV command. pending prompt
• Specify *CLRPND or *YES for
After switching the direction of the data group, when starting
the Clear error prompt
replication on the system that now becomes the source system

For additional information about the STRDG command, refer to the following topics:
• “What occurs when a data group is started” on page 221
• “What replication processes are started by the STRDG command” on page 251

To start a data group, do the following:


1. From the Work with Data Groups display, type a 9 (Start DG) next to the data
group that you want to start and press Enter.
The Start Data Group (STRDG) display appears.
2. At the Start processes prompt, specify the value for the processes you want to
start. If you are starting the data group for the first time specify *ALL. To see a list
of values, press F4 (Prompt).
3. Press Enter.
4. Additional prompts appear. For most situations, you should accept the default
values. If necessary, specify the following:
• At the Database journal receiver and Database large sequence number
prompts, identify where the database reader and apply processes begin.
• At the Object journal receiver and Object large sequence number prompts,
identify where the object send and apply processes begin.
• If you are starting the data group for any of the reasons listed in Table 39,
specify the indicated values for that reason in the Clear pending and Clear
error prompts.
• If you are submitting this command for batch processing, you should specify
*NO for the Show confirmation screen prompt.

229
5. To start the data group, press Enter.

Starting replication when open commit cycles exist


Open commit cycles may be present when a data group ends, or if a system event or
failure occurred. In most conditions, an open commit cycle present at the time that a
data group ended will not prevent a request to start replication from running.
However, certain process, such as performing a hardware upgrade with a disk image
change, converting to MIMIX Dynamic Apply for cooperative processing, changing the
configured value of target constraint management, or enabling a disabled data group,
require a clear pending start. Open commit cycles must be resolved before
performing these operations.
MIMIX will prevent the data group from starting with open commit cycles when any of
these conditions exist:
• When the start request specifies to clear pending entries. Certain procedures may
require a clear pending start. Message LVE387F is issued with reason code AP.
• When the commit mode specified for the database apply process changed.
Changing the commit mode is not a common occurrence. Message LVEC0B3 is
issued.
• When you are changing the target constraint management value specified for the
database apply process.
When these conditions exist, the open commit cycles must be resolved.

Checking for open commit cycles


Do the following to check for open commit cycles:
1. From the MIMIX Basic Main Menu, type a 6 (Work with data groups) and press
Enter.
2. The Work with Data Group display appears. Type an 8 (Display status) next to the
data group you ended and press Enter.
3. Press F8 (Database) to view the Data Group Detail Status display.
4. For each apply session listed, check the value shown in the Open Commit column
at the right side of the display. If the value is *YES, open commit cycles exist for
the data group.

Resolving open commit cycles


This procedure assumes that the data group is ended and that you have confirmed
the presence of open commit cycles.
1. Start the data group, specifying *NO for the Clear pending prompt.
2. You must take action to resolve the open commit cycles, such as ending or
quiescing the application or closing the commit cycle. MIMIX will process the open

230
Starting replication when open commit cycles exist

commit cycles when they are resolved.


3. Perform a controlled end of the data group.
4. When the data group is ended, check for open commit cycles again.
You may need to repeat this procedure until all open commit cycles have been
resolved.

231
Before ending replication
Consider the following:
• If the next time you start the data groups requires that you clear pending entries,
or if you will be performing a switch, you should verify that no activity is still in
progress before you perform these activities. Use the command WRKDGACTE
STATUS (*ACTIVE) to ensure all activity entries completed.
• Data groups that are in a disabled state are not ended. Only data groups that
have been enabled and have been started can be ended.

Commands for ending replication


These commands end replication processes. The significant differences between
these commands are:
• End MIMIX (ENDMMX) - The ENDMMX command will end all MIMIX processes in
a MIMIX installation, including those used for replication, in a single operation.
Optionally, this command can be used to end replication and supporting
processes associated with the local system, including single-system processes
and processes that run between the local system and another system.
• End Application Group (ENDAG) - The ENDAG command will end replication
processes for data groups that are part of an application group. This is the
preferred method of ending replication in application groups. The command
invokes a procedure which performs the operations to end replication for the
participating data groups.
• End Data Group (ENDDG) - The ENDDG command will end the specified
replication processes for the data group either immediately or in a controlled
manner. This command is the basis for all other methods of ending replication,
and is also called by commands that perform switch operations. Optionally, this
command can end a subset of replication processes or a selected database apply
process, specify a wait time and end option for controlled ends, and end the
remote journal link.

Command choice by reason for ending replication


Table 40 lists common reasons for ending MIMIX activity and the appropriate
command to use. Depending on why you are ending replication, you may need to
choose values other than the defaults.

Table 40. Choosing the appropriate command to end replication

Reason for Ending Replication Use Command Additional Information

Ending communications for any ENDMMX


reason

Performing a full save and restore ENDMMX


of data that is defined to MIMIX

232
Commands for ending replication

Table 40. Choosing the appropriate command to end replication

Reason for Ending Replication Use Command Additional Information

Performing a save from the ENDAG or ENDDG When application groups are used, use
source system the ENDAG command with its default
END procedure. See “What is ended
by the default END procedure for an
application group” on page 237.

For ENDDG, specify *ALL for the Process


(PRC) parameter. See “What replication
processes are ended by the ENDDG
command” on page 255.

The save request may not be able to save


all the files or objects if they are opened
or locked by MIMIX.

Performing a save from the target ENDAG For details about the RUNBACKUP or
system PROC(RUNBACKUP) ENDTGT procedures, see the MIMIX
(preferred) Administrator Reference book.
or
You may be able to end only selected
ENDAG PROC(ENDTGT)
processes on the target system. See
or “Ending selected data group processes”
ENDDG PRC(*ALLTGT) on page 247.

The save request may not be able to save


all the files or objects if they are opened
or locked by MIMIX.

Preparing to update MIMIX ENDMMX See controlled end information in “Ending


software immediately or controlled” on page 234.

Performing an IPL of either ENDMMX Also end the RJ link


system

Upgrading the operating system ENDMMX Also end the RJ link


release on either system

Performing a switch in --- Let your switching mechanism end


preparation for performing replication (switch procedure for
maintenance on either system application group or (MIMIX Switch
Assistant or MIMIX Model Switch
Framework)

Ending only a selected replication ENDDG See “Ending selected data group
process processes” on page 247.

Changing configuration, such as ENDAG or When application groups are used, use
adding or changing data group ENDDG the ENDAG command.
entries The changes are not available to active
replication processes until the data group
processes are ended and restarted.

233
Additional considerations when ending replication
The following questions will help you determine additional options you may need
when ending replication. All methods of ending replication can accomplish these
activities, but in some, the action is not default or may require additional
programming.
• Do processes need to end in a controlled manner or can they be ended
immediately? Both commands support these options. For more information, see
“Ending immediately or controlled” on page 234
• Do you need to end only a subset of the replication processes? Only ENDDG
supports ending selected processes. For more information see “Ending all or
selected processes” on page 235.
• Do RJ links also need to end? Both replication and system manager processes
use RJ links. In most cases, RJ links for either purpose can remain active. For
scenarios when you should end the RJ links, see “When to end RJ links” on
page 236.

Ending immediately or controlled


Both ENDMMX and ENDDG commands provide the ability to choose whether
replication processes end immediately or in a controlled manner through the End
process (ENDOPT) parameter.
For the ENDAG command, the specified end procedure determines whether
replication processes end immediately or in a controlled manner. If the procedure
specifies a controlled end, the procedure also determines wait time and time out
options.
When you perform an immediate end, the processes end independently of each
other. For example, it is possible for the apply process to end before the send or
receive process. Each replication process verifies that its processing is at a point that
will permit ending, then ends. The amount of time it takes for an immediate end varies
depending on the delay values set for each manager and what each process is doing
at the time. An immediate end does not ensure that all journal entries generated are
sent to or applied on the target system.
If an incomplete IFS or object tracking entry for a data group is being processed
during an immediate end, the entire entry may not be applied. When the data group is
restarted, the entire incomplete entry is rewritten to ensure the integrity of the object.
When you perform a controlled end, MIMIX creates either a journal entry or log
space entry. This entry proceeds through the replication path. The date and time of
the entry are compared to the date and time of when the process being considered
was started. If the entry is earlier than the process start time, the end request is
ignored. If the entry is later than when the process being considered was started, the
process is ended.
A controlled end ensures that processes end in order and that each process
completes any queued or in-progress transactions before the next process is

234
Commands for ending replication

permitted to end. This ensures that you have a known point in each journal at which
you can restart replication.
If any processes have a backlog of entries, it may take some time for the entry
created by the request to be processed through the replication path. Any entries that
precede the entry requesting to end are processed first.
A data group that is ended in a controlled manner is prepared for a more effective and
safer start when the start request specifies to clear pending entries. The existence of
commit cycles implies that there is application activity on the source system that
should not be interrupted; replication should be allowed to continue through the end of
the commit cycle. It is preferable to ensure that commit cycles are resolved or
removed before ending a data group. There are conditions in which a data group will
not start if open commit cycles exist. For more information, see “Starting replication
when open commit cycles exist” on page 230.
If the request to perform a controlled end also includes ending the RJ link, the RJ link
is ended after all requested processes end.
Either type of end request may be ignored if the request is submitted just before the
time that MIMIX jobs are restarted daily. For more information about restarting jobs,
see ‘Configuring restart times for MIMIX jobs’ in the MIMIX Administrator Reference
book.

Controlling how long to wait for a controlled end to complete


On the ENDMMX or ENDDG command, when you request a controlled end you can
determine how long to wait for all specified data group processes to end. The Wait
time (seconds) (WAIT) parameter specifies how long to wait for all of the specified
data group processes to end. MIMIX will attempt to resolve all pending activity entries
before ending the data groups. If a numeric value was specified, and the selected
processes do not end within the specified time, the action specified for the Timeout
option (TIMOUTOPT) will occur.
The WAIT parameter also supports special values of *SBMRQS and *NOMAX. When
these values are used, the TIMOUTOPT parameter is ignored.
Note: If *ALL is specified for any part of the data group definition, the Wait time value
must be *SBMRQS (submit request).

Ending all or selected processes


MIMIX determines which data group replication processes to end based on the
command specified and options on the command.
The ENDMMX command ends all replication processes for all data groups on the
systems specified on the end request.
The default END procedure for the ENDAG command uses the default settings of the
ENDDG command. MIMIX also ships an ENDTGT procedure that, when specified on
the ENDAG command, will end only processes on the target system.
Only the ENDDG command supports the ability to end selected replication processes
through its Process (PRC) parameter. The default value is to end all replication
processes for the specified data groups. The configuration of each data group

235
determines which processes end with each possible value for the PRC parameter. If
you choose to use this parameter, be sure that you understand what processes will
end. See “What replication processes are ended by the ENDDG command” on
page 255.

When to end RJ links


The default behavior for RJ links used by replication processes is to remain active
when ending replication using the ENDMMX or ENDDG command or the shipped
default END procedure for the ENDAG command.
The default behavior for RJ links used by system manager processes is also to
remain active when ending replication using the ENDMMX command.
RJ links can normally remain active unless you have a need to prevent data from
being sent to the target system. Some situations where you need to end RJ links
include:
• Following a switch, to prevent data from returning to the system on which it
originated (round-tripping), and to reduce communications and DASD usage
• Before performing an IPL on either the source system or target system
• Before upgrading the IBM i release on either the source system or the target
system
• Before performing a hardware upgrade
You can optionally end replication-related RJ links when ending replication with the
ENDMMX or ENDDG command by changing the value of the End remote journaling
(ENDRJLNK) parameter and when ending replication for an application group by
adding the shipped step program MXENDRJLNK into the END procedure for that
application group.
You can optionally end system manager RJ links when ending MIMIX by changing the
value of the End system manager RJ links (ENDSMRJ) parameter on the ENDMMX
command.

What is ended by the ENDMMX command


The ENDMMX command will end all MIMIX processes on the specified systems in the
installation. When you use this command in environments with application groups, all
procedures must have a completed or acknowledged status. This command does not
run the default End procedures associated with application groups or explicitly end
application groups. Any processes for user applications or IBM cluster resource
groups must be ended separately. When you use this command, the following occurs:
Data groups - The end process specified is used to end all enabled data groups
and their supporting processes, including automatic recovery, on the specified
systems. This includes data groups associated with data resource groups. Default
values end data groups in a controlled manner.
Remote journal links for replication - If you selected to end remote journaling, all
remote journal links associated with replication processes for the specified
systems are ended.

236
Commands for ending replication

MIMIX managers and services - Ends the system managers, journal managers,
target journal inspection, and collector services on the specified systems.
Monitors - Ends all individual monitors currently active in the installation library on
the specified systems.
Master monitor - Ends the MIMIX scheduler job and the master monitor on each of
the specified systems.
MIMIX Promoter - Ends promoter group activity on the specified systems.
Audits and Recoveries - All queued audits, all audits in progress, and all
recoveries in progress that are associated with the specified systems are ended.
This includes jobs with locks on the installation library. Queued audits are set to
*NOTRUN and audits in comparison phase are set to *FAILED. Audits in recovery
phase reflect their state of processing at the time of the end request, which may
be *NOTRCVD.
Remote journal links for system managers - If you selected to end system
manager RJ links, all remote journal links associated with system manager
processes for the specified systems are ended.
Note: Cluster services is not ended when MIMIX managers end because cluster
services may be necessary for other applications.
Optionally, the command can be used to end MIMIX processes associated with the
local system. When the command request specifies *LOCAL for the System definition
(SYSDFN) parameter, single system processes on the local system are ended and
multiple-system processes between the local system and another system are ended.
Single system processes include the master monitor, MIMIX scheduler, monitors,
journal manager, target journal inspection, and collector services. Multiple-system
processes include system managers, RJ links for system managers and replication
processes (when requested), data group replication processes and their supporting
processes, promoter group activity, and audits. Application groups and cluster
services are not ended.

What is ended by the default END procedure for an application group


The End Application Group (ENDAG) command automatically uses an application
group’s default END procedure unless you specify a different procedure.
When an application group is created, a default procedure named END is created for
it from a shipped default procedure. The procedure ends replication processes for
enabled data groups within a single application group. Steps in the shipped default
END procedure, as well as steps in additional shipped procedures that end
application groups, are described in the MIMIX Administrator Reference book.

237
What occurs when a data group is ended
The End Data Group (ENDDG) command will end replication processes for the
specified data group.
The ENDDG command can be used interactively or programmatically. This command
is invoked by the ENDMMX command and by the ENDAG command running the
default END procedure, using values other than default for some parameters.
When an ENDDG request is processed, MIMIX may take a few minutes while it does
the following for each specified data group:
• The value you specify for the Process (PRC) parameter determines which data
group replication processes will end. The default value ends all data group
replication processes.
• When ending data groups that use a shared object send job, the job is ended by
the last data group to end.
• When ending data groups that perform access path maintenance, the database
apply process signals the access path maintenance job and then ends. The
access path maintenance job uses additional jobs, if needed, to change the
access path maintenance attribute to immediate on all files that MIMIX had
previously changed to delayed. Any files that could not be changed are identified
as having an access path maintenance error before the maintenance jobs end.
• The specified replication processes end in the manner specified for the End
process (ENDOPT) parameter. The command defaults to processing the end
request immediately (*IMMED). When invoked by the ENDMMX command, the
default value specified on ENDMMX is *CNTRLD, which takes precedence. When
invoked by a procedure specified on the ENDAG command, the procedure
determines whether ENDDG is passed parameter values or uses the command
defaults.
• if a controlled end is requested, it uses the specified Wait time and Timeout
options.
• If requested, the RJ link is ended. The RJ link is not automatically ended. In most
cases, the default value *NO for the End remote journaling (ENDRJLNK)
parameter is appropriate. Keeping the RJ link active allows database changes to
continue to be sent to the target system even though the data group is not active.
• The Enable constraints (ENBCST) parameter determines whether the database
apply processes will attempt to enable constraints on target system files that were
previously disabled by MIMIX. The default value, *DFT, will enable constraints if
appropriate configuration conditions exist and the end request specifies all apply
sessions and includes database processes that run on the target system.
• If you have used the MIMIX CDP feature to set a recovery point in a data group
and then end the data group, the recovery point will be cleared. When the data
group is started again, the apply processes will process any available
transactions, including those which may have had corruptions. (Recovery points
are set with the Set DG Recovery Point (SETDGRCYP) command.) If a recovery
window is configured for the data group, its configured duration is not affected by

238
What occurs when a data group is ended

requests to end or start the data group.


The ENDDG command does not end the system manager, journal manager, or other
processes that run at the node level. To end those processes, either use the
ENDMMX command or use the End MIMIX Managers (ENDMMXMGR) command
after replication processes have ended.
If a data group is participating in a virtual switch test when you initiate an end data
group request, the data group’s apply processes are already ended and are not
affected by the request. The apply processes remain ended and will only be started
by the virtual switch procedure during its recovery phase.

239
Ending MIMIX
For most configurations, It is recommended that you end MIMIX products from the
management system, which is usually the backup system. If your installation is
configured so that the backup system is a network system, you should end MIMIX
from the network system.
Notes:
• If you are ending MIMIX for a software upgrade or to install a service pack, use
the procedures in the software’s ReadMe document.
• The ENDMMX command cannot run when application groups are configured and
there are any active, failed, or canceled procedures.
• Determine whether you also need to end RJ links. See “When to end RJ links” on
page 236.
To end MIMIX, use the following procedures:
1. Use one of the following procedures:
• “Ending with default values” on page 240
• “Ending by prompting the ENDMMX command” on page 240
2. Complete any needed follow-up actions using the information and procedures in
“After you end MIMIX products” on page 241.

Ending with default values


Use these instructions to end all MIMIX processes in an installation library. RJ links
will remain active.
1. From the MIMIX Basic Main Menu, select option 3 (End MIMIX) and press Enter.
You will see a confirmation display.
2. From the confirmation display, you can press F1 (Help) to see a description of the
default values that will be used. To end MIMIX, press Enter.

Ending by prompting the ENDMMX command


These instructions end all MIMIX processes for the specified systems and optionally
end RJ links. In most cases RJ links can remain active. See “When to end RJ links”
on page 236 for situations when you should end RJ links.
Do the following:
1. From a command line, type ENDMMX and press F4 (Prompt).
2. The End MIMIX display appears. At the End process prompt, specify *CNTRLD
for a controlled end or *IMMED for an immediate end. This parameter applies to
the application group and data group processes only.
Note: When ENDMMX ends data groups, it waits for each data group to end
before attempting to end the next MIMIX function.
3. If you need to end RJ links used by replication processes, specify *YES for the

240
Ending MIMIX

End remote journaling prompt.


Note: If you specify *YES, all data groups using the remote journal link in the
installation library will be affected. If data groups in other installations are
using the same remote journal link, you should specify *NO.
4. If you specified *CNTRLD for Step 2, ensure that the values for the Wait time
(seconds) and Timeout option prompts are what you want for the controlled end.
5. At the System definition prompt, indicate the scope of the request by specifying
either *ALL or *LOCAL. This determines the systems on which to end MIMIX
processes.
6. If you need to end RJ links used by system manager processes, specify *YES for
the End system manager RJ links prompt.
7. To end MIMIX processes, press Enter.

After you end MIMIX products


Some pending transactions may not be handled before the end process completes.
You may need to ensure that all activity entries are complete before you issue
additional commands. Examples of scenarios where it is important to check whether
all pending transactions are completed include:
• Switching a data group (SWTDG command)
• Starting a data group with clear pending entries (STRDG CLRPND(*YES)).
To check for active entries, use the command WRKDGACTE STATUS(*ACTIVE).

When to also end the MIMIX subsystem - You will also need to end the MIMIX
subsystem when you need to IPL the system, when upgrading MIMIX software, and
when installing a MIMIX software service pack. The MIMIX subsystem must be ended
from the native user interface. To end the subsystem, do the following:
1. If you are using Vision Solutions Portal (VSP) end the VSP server on the systems
where it runs. This prevents object locking issues. Do one of the following:
• If the VSP server runs on an IBM i platform, use the following command:
VSI001LIB/ENDVSISVR
• If the VSP server runs on a Windows platform, from the Windows Start menu,
select
All Programs > Vision Solutions Portal > Stop Server and click Stop Server.
2. Enter the command WRKSBS. The Work with Subsystems display appears.
3. Type an 8 (Work with subsystem jobs) next to subsystem MIMIXSBS and press
Enter.
4. End any remaining jobs in a controlled manner. Type a 4 (End) next to the job and
press F4 (Prompt). The How to end (OPTION) parameter should have a value of
*YES. Press Enter. If you see a confirmation display, press Enter to continue.
5. Press F12 (Cancel) to return to the Work with Subsystems display.
6. Type a 4 (End subsystem) next to subsystem MIMIXSBS and press Enter.

241
242
Ending an application group

Ending an application group


For an application group, a procedure for only one operation (start, end, or switch)
can run at a time. For information about parameters and shipped procedures, see
“What is ended by the default END procedure for an application group” on page 237
and “Choices when starting or ending an application group” on page 218.
To end an application group, do the following:
1. From the Work with Application Groups display, type 10 (End) next to the
application group you want and press F4 (Prompt).
2. Verify that the values you want are specified for Resource groups and Data
resource group entry.
3. If you are starting the procedure after addressing problems with the previous end
request, specify the value you want for Begin at step. Be certain that you
understand the effect the value you specify will have on your environment.
4. Press Enter.
5. The Procedure prompt appears. Do one of the following:
• To use the default end procedure, press Enter.
• To use a different end procedure for the application group, specify its name.
Then press Enter.

243
Ending a data group in a controlled manner
The following procedures describe how to check for errors before requesting a
controlled end of a data group, how to perform the controlled end request, and how to
confirm that the end completed. Held files must be released and the apply process
must complete operations for journal entries stored in log spaces before you end data
group activity.
Data groups that are in an application group: The preferred method of ending data
groups that are part of an application group is to use the End Application Group
(ENDAG) command.

Preparing for a controlled end of a data group


It is good practice to ensure that errors are resolved before requesting a controlled
end of a data group.
Do the following:
1. From the Work with Data Groups display, type an 8 (Display status) next to the
data group you want to end and press Enter.
2. The Data Group Status display appears. In the upper right of the display, you
should see either one or both of the following fields. A non-zero value in these
fields will not prevent the end request from completing.
• Database errors identifies the number of items replicated through the user
(database) journal that have a status of *HLDERR. This number should be 0
before you end the data group.
• Object in error/active identifies two key statistics associated with objects
replicated through the system journal. The first number identifies the number of
objects that have a status of *FAILED and the second number identifies the
number of objects with active (pending) activity entries. Both numbers should
be 0 before you end the data group.
Note: Only information for the type of information replicated by the data group
appears on the status displays. For example, if the data group does not
contain database files, you will only see fields for object information.
3. For data groups which replicate from the user journal, you also need to check for
any files that are held for other reasons. Press F8 (Database). The Held for other
reasons field In the upper right of the Data Group Database Status display should
also be 0 before you end the data group.
A non-zero value may or may not prevent the end request from completing. For
more information, see topics “Working with files needing attention (replication and
access path errors)” on page 263.

Performing the controlled end


1. From the Work with Data Groups display, type a 10 (End DG) next to the data
group you want to end and press Enter.
2. The End Data Group (ENDDG) display appears. Specify *CNTRLD for the End

244
Ending a data group in a controlled manner

processes prompt.
3. If the data group uses remote journaling, verify that the value of the End remote
journaling prompt is what you want.
4. For the Wait Time (WAIT) parameter, specify how long MIMIX should try to end
the selected processes in a controlled manner. Use F1 (Help) to see additional
information about the possible options.
• Specify *SBMRQS to submit a request to end the data groups. The appropriate
actions are issued to end the specified processes, and control is returned to
the caller immediately. When you specify this value, the TIMOUTOPT
parameter is ignored so you can skip Step 5.
• Specify *NOMAX. When you specify this value, MIMIX will wait until all
specified MIMIX processes are ended.
• Specify a numeric value (number-of-seconds). MIMIX waits the specified time
for a controlled end to complete before using the option specified in the
TIMOUTOPT parameter.
5. If you specified a numeric value for the WAIT parameter in Step 4, you can also
use the Timeout Option (TIMOUTOPT) parameter. You can specify what action
you want the ENDDG command to perform if the time specified in the WAIT
parameter is reached:
• The current process should quit and return control to the caller (*QUIT).
• A new request should be issued to end all processes immediately
(*ENDIMMED). When this value is specified, pending activity entries may still
exist after the data group processes are ended.
• An inquiry message should be sent to the operator notifying of a possible error
condition (*NOTIFY). If you specify this value, the command must be run from
the target system.
6. If the data group you are ending is configured to allow target constraint
management, specify whether you want to enable constraints on target files which
were previously disabled by MIMIX. If you specified *SBMRQS for WAIT, the
database apply processes must be already ended by a previous controlled end. If
you specified a number for WAIT, the replication processes must end in the
specified time. The value *DFT will enable constraints for *ACTIVE file entries
after replication processes end.
7. Press Enter to process the command.

Confirming the end request completed without problems


After you request a controlled end of a data group, the Work with Data Group display
appears. Do the following:
1. From the Work with Data Group display appears. Type an 8 (Display status) next
to the data group you ended and press Enter.
2. The Data Group Status display appears. In the Target Statistics section near the
middle of the display, the Unprocessed Entry Count column should be blank for
any database apply processes and any object apply processes. If unprocessed

245
entries exist when you end the data group and perform a switch, you may lose
these entries when the data group is started following the switch.
Note: To ensure that you are aware of any possible pending or delayed activity
entries, enter the WRKDGACTE STATUS(*ACTIVE) command. Any
activities that are still in progress will be listed. Ensure that all activities are
completed.
3. Ensure that there are no open commit cycles.The next attempt to start the data
group will fail if open commit cycles exist and either the start request specified to
clear pending entries (CLRPND(*YES)) or the commit mode specified in the data
group definition changed. (Certain process, such as performing a hardware
upgrade with a disk image change, converting to MIMIX Dynamic Apply, changing
target constraint management configuration, or enabling a disabled data group,
require a clear pending start.) To verify commit cycles, do the following:
a. Press F8 (Database) to view the Data Group Detail Status display.
b. For each apply session listed, verify that the value shown in the Open Commit
column at the right side of the display is *NO.
c. If open commit cycles exist, restart the data group. You must take action to
resolve the open commit cycles, such as ending or quiescing the application or
closing the commit cycle. Then repeat the controlled end again.

246
Ending selected data group processes

Ending selected data group processes


This procedure can be used to end all or selected processes for a data group, or end
a specific database apply process.
Data groups that are in an application group: The preferred method of ending data
groups that are part of an application group is to use the End Application Group
(ENDAG) command. The default behavior of the ENDDG command helps to enforce
this best practice when necessary by not allowing the command to run when the data
group is participating in a resource group with three or more nodes. (A data resource
group provides the association between one or more data groups and an application
group.) The ENDDG request will run when the data group is participating in a
resource group with two nodes. You can also change the value specified for Override
if in data rsc. group (DTACRG) parameter to allow a data group to be ended even if it
is part of a resource group.
In application group environments with three or more nodes, it is particularly important
to treat all members of an application group as one entity. For example, a
configuration change that is made effective by starting and ending a single data group
would not be propagated to the other data groups in the same resource group.
However, the same change would be propagated to the other data groups if it is made
effective by ending and starting the parent application group.
For additional information about the ENDDG command, refer to the following topics:
• “What occurs when a data group is ended” on page 238
• “What replication processes are ended by the ENDDG command” on page 255

To selectively end processes for a data group, do the following:


1. From the Work with Data Groups display, type a 10 (End DG) next to the data
group that you want to end and press Enter.
2. The End Data Group (ENDDG) display appears. At the Process prompt, specify
the value for the processes you want to end. To see a list of values, press F4
(Prompt).
3. At the End process prompt, specify the value you want.
4. If the data group uses remote journaling, verify that the value of the End remote
journaling prompt is what you want.
5. If you want to end only a selected apply session, press F10 (Additional
parameters). Then specify the value for the session you want to end at the Apply
session prompt.
6. To end the selected processes, press Enter.

247
Enabling MIMIX-disabled target constraints
MIMIX supports t several ways of enabling constraints on target node files that were
previously disabled by MIMIX.
Default shipped procedures for switching application groups automatically enable
MIMIX-disabled constraints before performing the requested switch. However, you
may need to enable constraints outside the scope of a switch procedure, such as
before manually switching a data group with the Switch Data Group command or
before performing other operations for which you would want constraints enabled on
the target node. Even when target constraint management is not configured, you may
need to enable any remaining constraints that failed to be enabled on the target node.
You can manually request MIMIX to enable constraints in these ways:
• When ending an application group, the shipped procedures END and ENDTGT
include disabled steps MXECSTB and MXECSTR. When enabled, these steps
enable constraints previously disabled by MIMIX. To allow these steps to be
performed by either procedure, you must take action to enable the steps from
within the procedure. Enabling steps must be performed from the native user
interface.
• When ending replication processes for a specific data group using the Stop Data
Group dialog from Vision Solutions Portal (VSP) or the End Data Group (ENDDG)
command from the native user interface. Constraint processing can only occur
when the end request specifies all required values and other configuration
requirements are met.
• For an ended data group. When replication processes have already been ended
by a controlled end request and there are no open commit cycles for the data
group, you can enable constraints using the ENDDG command in the native user
interface or the Enable Constraints action for a data group from VSP.
Only MIMIX-disabled constraints on target node files for which there are existing data
group file entries with a status of *ACTIVE1 can be enabled.
When the following conditions exist and these values are specified on an end data
group request, MIMIX will enable constraints on the target node files after the
specified replication processes end and before the requested command completes:
• Process (PRC) must specify processes that run on the target node (*ALL,
*ALLTGT, *DBALL, *DBTGT, or *DBAPY)
• Configuration requirements for the specified End process (ENDOPT) must be
met.
– When specifying *IMMED for an immediate end, or when specifying *CNTRLD

1. For users of Vision Solutions Portal (VSP), data group file entries are known as file selection
rules. File selection rules are assigned an initial status when they are created, which is not
displayed in VSP. When replication processes start, file selection rules correspond to file
activities and their status become Active unless replication activity for the identified file or
member was previously held or in error. In VSP, only the file activities that are experiencing
database replication problems are displayed on the File Activity tab of the Data Group
Details and Activities portlet.

248
Enabling MIMIX-disabled target constraints

for a controlled end and a wait time (WAIT) value of *SBMRQS, database
apply processes must have already been ended by a previous controlled end
request.
– When specifying *CNTRLD for a controlled end and a number for wait time
(WAIT), the data group must be configured for target constraint management
and its replication processes must end within the specified wait time and have
no open commit cycles.
• Apply sessions (APYSSN) must specify *ALL.
• Enable constraints (ENBCST) must specify or resolve to *YES. The shipped
default value, *DFT, resolves to *YES when all other conditions are met.
If the requirements are not met, the constraints are not enabled and no error message
is issued.
MIMIX-disabled constraints that failed to be enabled may still exist on the target node
after the end data group request completes or after the configuration is changed to no
longer allow target constraint management.
When target constraint management is configured, specifying the value *NO for
ENBCST allows any MIMIX-disabled constraints to remain disabled, improving
performance of this end operation and the next start operation.

Enabling constraints for an *INACTIVE data group


To enable constraints for target node files in a data group whose replication
processes have already ended controlled and that have no open commits, do the
following:
1. From the Work with Data Groups display, type a 10 (End DG) next to the data
group that you want to end and press Enter.
2. The End Data Group (ENDDG) display appears. At the Process prompt, specify
*ALL, *ALLTGT, *DBALL, *DBTGT, or *DBAPY.
3. Press F10 (Additional parameters).
4. Confirm that the value *ALL is specified for the Apply session prompt.
5. Specify *YES for the Enable constraints prompt.
6. Press Enter.

Enabling constraints when ending a data group


Enabling constraints for target node files in a data group whose replication processes
are active requires performing a controlled end. Replication processes must end in
the specified wait time, and open commit cycles cannot exist when processes have
ended. Constraint processing occurs after the specified replication processes end
and before the command completes.
Do the following:
1. From the Work with Data Groups display, type a 10 (End DG) next to the data
group that you want to end and press Enter.

249
2. The End Data Group (ENDDG) display appears. At the Process prompt, specify
*ALL, *ALLTGT, *DBALL, *DBTGT, or *DBAPY.
3. Specify *CNTRLD for the End process prompt.
4. Press F9 (All parameters).
5. Specify either a number or the value *NOMAX for the Wait time (seconds) prompt.
6. Confirm that the value *ALL is specified for the Apply session prompt.
7. Specify *YES for the Enable constraints prompt.
8. Press Enter.

250
What replication processes are started by the STRDG command

What replication processes are started by the STRDG command


MIMIX determines how each data group is configured and starts the appropriate replication processes based on the value you specify for
the Start processes (PRC parameter). Default configuration values create data groups that use MIMIX Remote Journal support (MIMIX RJ
support) for database replication and source-send technology for object replication.
Table 41 identifies the processes that are started when MIMIX RJ support is used for database replication for each of the possible values
on the PRC parameter. An RJ link identifies the IBM i remote journal function, which transfers data to the target system. On the target
system, the data is processed by the MIMIX database reader (DBRDR) before the database apply process (DBAPY) completes
replication.
For data groups that use MIMIX RJ support, it is standard practice to leave the RJ link active when the data groups are ended. If the RJ
link is not already active when starting data groups, MIMIX starts the RJ link when the value specified for the PRC parameter includes
database source system processes or all processes. The RJ Link column in Table 41 shows the result of each process when the RJ link is
not active while the Notes column identifies behavior that may not be anticipated when the RJ link is already active.

Table 41. Processes started by data groups configured for MIMIX Remote Journal support. This assumes that all replication processes are inactive when the STRDG
request is made.

Value for Notes Source Processes Target Processes


PRC
DB replication Object replication DB replication Object replication

RJ Link 1 OBJSND OBJRTV CNRSND STSRCV DBRDR DBAPY2 OBJRCV CNRRCV STSSND OBJAPY

*ALL E Starts1 Starts Starts Starts Starts Starts Starts Starts Starts Starts Starts

*ALLSRC A, E Starts1 Starts Starts Starts Starts Inactive Inactive Starts Starts Starts Inactive

*ALLTGT A, B Inactive1 Inactive Inactive Inactive Inactive Starts Starts Inactive Inactive Inactive Starts

*DBALL A, E Starts1 Inactive3 Inactive3 Inactive3 Inactive3 Starts Starts Inactive3 Inactive3 Inactive3 Inactive3

Notes:
A. Data groups which use cooperative processing should have both database and object processes started to prevent objects and data on the target system from
becoming not fully synchronized.
B. When the RJ link is already active, database replication becomes operational.
C. When the RJ link is already active, database journal entries continue to transfer to the target system over the RJ link
D. When the RJ link is already active, database journal entries continue to transfer to the target system over the RJ link, where they will be processed by the DBRDR.
E. If data group data area entries are configured, the data area polling process also starts when values which start database source processes are selected.

251
What replication processes are started by the STRDG command

Table 41. Processes started by data groups configured for MIMIX Remote Journal support. This assumes that all replication processes are inactive when the STRDG
request is made.

Value for Notes Source Processes Target Processes


PRC
DB replication Object replication DB replication Object replication

RJ Link 1 OBJSND OBJRTV CNRSND STSRCV DBRDR DBAPY2 OBJRCV CNRRCV STSSND OBJAPY

*OBJALL A, C Inactive1 Starts Starts Starts Starts Inactive4 Inactive 4 Starts Starts Starts Starts

*DBSRC A, C, Starts1 Inactive3 Inactive3 Inactive3 Inactive3 Inactive Inactive Inactive3 Inactive3 Inactive3 Inactive3
E

*DBTGT A, B Inactive1 Inactive3 Inactive3 Inactive3 Inactive3 Starts Starts Inactive3 Inactive3 Inactive3 Inactive3

*OBJSRC A, C Inactive1 Starts Starts Starts Starts Inactive4 Inactive4 Starts Starts Starts Inactive

*OBJTGT A, C Inactive1 Inactive Inactive Inactive Inactive Inactive4 Inactive4 Inactive Inactive Inactive Starts

*DBRDR A, D Inactive1 Inactive3 Inactive3 Inactive3 Inactive3 Starts Inactive Inactive3 Inactive3 Inactive3 Inactive3

*DBAPY A, C Inactive1 Inactive3 Inactive3 Inactive3 Inactive3 Inactive4 Starts4 Inactive3 Inactive3 Inactive3 Inactive3

Notes:
A. Data groups which use cooperative processing should have both database and object processes started to prevent objects and data on the target system from
becoming not fully synchronized.
B. When the RJ link is already active, database replication becomes operational.
C. When the RJ link is already active, database journal entries continue to transfer to the target system over the RJ link
D. When the RJ link is already active, database journal entries continue to transfer to the target system over the RJ link, where they will be processed by the DBRDR.
E. If data group data area entries are configured, the data area polling process also starts when values which start database source processes are selected.
1. This column shows the effect of the specified value on the RJ link when the RJ link is not active. See the Notes for the effect of values when the RJ Link is already active, which is
default behavior.
2. If the access path maintenance (APMNT) policy has been enabled at the installation or data group level, an access path maintenance job is also started. Access path maintenance is
available on installations running 7.1.15.00 or higher.
3. These object replication processes are not available in data groups configured for database-only replication.
4. These database replication processes are not available in data groups configured for object-only replication.

Optionally, data groups can use source-send technology instead of remote journaling for database replication. Data groups created on
earlier levels of MIMIX may still be configured this way.

252
What replication processes are started by the STRDG command

Table 42 identifies the processes that are started by each value for Start processes when source-send technology is used for database
replication. The MIMIX database send (DBSND) process and database receive (DBRCV) process replace the IBM i remote journal
function and the DBRDR process, respectively.

Table 42. Processes started by data groups configured for Source Send replication This assumes that all replication processes are inactive when the STRDG request
is made.

Value for Notes Source Processes Target Processes


PRC
DB replication Object replication DB replication Object replication

DBSND 1 OBJSND OBJRTV CNRSND STSRCV DBRCV DBAPY2 OBJRCV CNRRCV STSSND OBJAPY

*ALL — Starts 1 Starts Starts Starts Starts Starts Starts Starts Starts Starts Starts

*ALLSRC A Starts 1 Starts Starts Starts Starts Starts Inactive Starts Starts Starts Inactive

*ALLTGT A Inactive Inactive Inactive Inactive Inactive Inactive Starts Inactive Inactive Inactive Starts

*DBALL A Starts 1 Inactive3 Inactive3 Inactive3 Inactive3 Starts Starts Inactive3 Inactive3 Inactive3 Inactive3

*OBJALL A Inactive 4 Starts Starts Starts Starts Inactive4 Inactive4 Starts Starts Starts Starts

*DBSRC A Starts 1 Inactive3 Inactive3 Inactive3 Inactive3 Starts Inactive Inactive3 Inactive3 Inactive3 Inactive3

*DBTGT A Inactive Inactive3 Inactive3 Inactive3 Inactive3 Inactive Starts Inactive3 Inactive3 Inactive3 Inactive3

*OBJSRC A Inactive4 Starts Starts Starts Starts Inactive4 Inactive4 Starts Starts Starts Inactive

*OBJTGT A Inactive4 Inactive Inactive Inactive Inactive Inactive4 Inactive4 Inactive Inactive Inactive Starts

*DBRDR 5 — — Inactive3 Inactive3 Inactive3 Inactive3 — — Inactive3 Inactive3 Inactive3 Inactive3

*DBAPY A Inactive4 Inactive3 Inactive3 Inactive3 Inactive3 Inactive4 Starts 4 Inactive3 Inactive3 Inactive3 Inactive3

Notes:
A. Data groups which use cooperative processing should have both database and object processes started to prevent objects and data on the target system from
becoming not fully synchronized.
1. When the database send (DBSND) process starts, the data area polling process also starts.
2. If the access path maintenance (APMNT) policy has been enabled at the installation or data group level, an access path maintenance job is also started. Access path maintenance is
available on installations running 7.1.15.00 or higher.
3. These object replication processes are not available in data groups configured for database-only replication.
4. These database replication processes are not available in data groups configured for object-only replication

253
What replication processes are started by the STRDG command

5. The database reader (*DBRDR) process is not used by data groups configured for source-send replication.

254
What replication processes are ended by the ENDDG command

What replication processes are ended by the ENDDG command


MIMIX determines how each data group is configured and ends the appropriate replication processes based on the value you specify for
the Process (PRC parameter). Default configuration values create data groups that use MIMIX Remote Journal support (MIMIX RJ
support) for database replication and source-send technology for object replication.
Table 43 identifies the processes that are ended by each value for PRC when MIMIX RJ support is used for database replication. An RJ
link identifies the IBM i remote journal function, which transfers data to the target system. On the target system, the data is processed by
the MIMIX database reader (DBRDR) before the database apply process (DBAPY) completes replication.
The communications defined by the RJ link remains active and is not affected by any value for PRC. In most cases, leaving the RJ link
active is preferable. If necessary, you can end the RJ link by changing value for End remote journaling (ENDRJLNK parameter). “Before
ending replication” on page 232 describes when you need to end the RJ link.

Table 43. Processes ended by data groups configured for MIMIX Remote Journal support. This assumes that all replication processes are active when the ENDDG
request is made and that the request does not specify to end the RJ link.

Value for Notes Source Processes Target Processes


PRC
DB replication Object replication DB replication Object replication

RJ Link 1 OBJSND OBJRTV CNRSND STSRCV DBRDR DBAPY2 OBJRCV CNRRCV STSSND OBJAPY

*ALL E Active1 Ends Ends Ends Ends Ends Ends Ends Ends Ends Ends

*ALLSRC A, E Active1 Ends Ends Ends Ends Active Active Ends Ends Ends Active

*ALLTGT — Active1 Active Active Active Active Ends Ends Active Active Active Ends

*DBALL B, E Active1 Active 3 Active 3 Active 3 Active 3 Ends Ends Active 3 Active 3 Active 3 Active 3

Notes:
A. Has no effect on database-only replication. New database journal entries continue to transfer to the target system over the RJ link, where they will be processed.
B. Data groups that use cooperative processing may be affected by the result of this value. Ending database processes while object processes remain active may
result in object activity entries being placed on hold. Similarly, ending object processes while database processes remain active may result in files being placed on
hold due to error.
C. New database journal entries continue to transfer to the target system over the RJ link. Existing entries stored in the log space on the target system before the end
request was processed will be applied.
D. New database journal entries continue to transfer to the target system over the RJ link, where they will be processed by the DBRDR.
E. The data area polling process ends when values which end database source processes are specified.

255
What replication processes are ended by the ENDDG command

Table 43. Processes ended by data groups configured for MIMIX Remote Journal support. This assumes that all replication processes are active when the ENDDG
request is made and that the request does not specify to end the RJ link.

Value for Notes Source Processes Target Processes


PRC
DB replication Object replication DB replication Object replication

RJ Link 1 OBJSND OBJRTV CNRSND STSRCV DBRDR DBAPY2 OBJRCV CNRRCV STSSND OBJAPY

*OBJALL A, B Active1 Ends Ends Ends Ends Active 4 Active 4 Ends Ends Ends Ends

*DBSRC A, B, Active1 Active 3 Active 3 Active 3 Active 3 Active Active Active 3 Active 3 Active 3 Active 3
E

*DBTGT B Active1 Active 3 Active 3 Active 3 Active 3 Ends Ends Active 3 Active 3 Active 3 Active 3

*OBJSRC A, B Active1 Ends Ends Ends Ends Active 4 Active 4 Ends Ends Ends Active

*OBJTGT A, B Active1 Active Active Active Active Active 4 Active 4 Active Active Active Ends

*DBRDR B, C Active1 Active 3 Active 3 Active 3 Active 3 Ends Active Active 3 Active 3 Active 3 Active 3

*DBAPY B, D Active1 Active 3 Active 3 Active 3 Active 3 Active Ends Active 3 Active 3 Active 3 Active 3

Notes:
A. Has no effect on database-only replication. New database journal entries continue to transfer to the target system over the RJ link, where they will be processed.
B. Data groups that use cooperative processing may be affected by the result of this value. Ending database processes while object processes remain active may
result in object activity entries being placed on hold. Similarly, ending object processes while database processes remain active may result in files being placed on
hold due to error.
C. New database journal entries continue to transfer to the target system over the RJ link. Existing entries stored in the log space on the target system before the end
request was processed will be applied.
D. New database journal entries continue to transfer to the target system over the RJ link, where they will be processed by the DBRDR.
E. The data area polling process ends when values which end database source processes are specified.
1. The RJ link is not ended by the End options (PRC) parameter. New database journal entries continue to transfer to the target system over the RJ link. See the Notes column for addi-
tional details.
2. On installations running 7.1.15.00 or higher, if access path maintenance is enabled, the database apply process signals the access path maintenance job and then ends. The access
path maintenance job uses additional jobs, if needed, to change the access path maintenance attribute to immediate on all files that MIMIX had previously changed to delayed. Any
files that could not be changed are identified as having an access path maintenance error before the maintenance jobs end.
On installations running software earlier than 7.1.15.00, if parallel access path maintenance function is enabled, the associated monitors are also ended when ENDDG command
specifies *DFT for End parallel AP maintenance (PRLAPMNT) and *ALL for the Apply session (APYSSN). When *YES is specified for PRLAPMNT, the function is always ended
regardless of the values specified for PRC or APYSSN.
3. These object replication processes are not available in data groups configured for database-only replication.

256
What replication processes are ended by the ENDDG command

4. These database replication processes are not available in data groups configured for object-only replication.

Optionally, data groups can use source-send technology instead of remote journaling for database replication. Data groups created on
earlier levels of MIMIX may still be configured this way. Table 44 identifies the processes that are ended by each value for End options
when source-send technology is used for database replication. The MIMIX database send (DBSND) process and database receive
(DBRCV) process are replaced by the IBM i remote journal function and the DBRDR process, respectively.

Table 44. Processes ended by data groups configured for Source Send replication This assumes that all replication processes are active when the ENDDG request is
made.

Value for Notes Source Processes Target Processes


PRC
DB replication Object replication DB replication Object replication

DBSND 1 OBJSND OBJRTV CNRSND STSRCV DBRCV DBAPY2 OBJRCV CNRRCV STSSND OBJAPY

*ALL — Ends 1 Ends Ends Ends Ends Ends Ends Ends Ends Ends Ends

*ALLSRC — Ends 1 Ends Ends Ends Ends Ends Active Ends Ends Ends Active

*ALLTGT — Active Active Active Active Active Active Ends Active Active Active Ends

*DBALL A Ends 1 Active 3 Active 2 Active 2 Active 2 Ends Ends Active 2 Active 2 Active 2 Active 2

*OBJALL A Active 4 Ends Ends Ends Ends Active 3 Active 3 Ends Ends Ends Ends

*DBSRC A Ends 1 Active 2 Active 2 Active 2 Active 2 Ends Active Active 2 Active 2 Active 2 Active 2

*DBTGT A Active Active 2 Active 2 Active 2 Active 2 Active Ends Active 2 Active 2 Active 2 Active 2

*OBJSRC A Active 3 Ends Ends Ends Ends Active 3 Active 3 Ends Ends Ends Active

*OBJTGT A Active 3 Active Active Active Active Active 3 Active 3 Active Active Active Ends

*DBRDR 5 — — Active 2 Active 2 Active 2 Active 2 — — Active 2 Active 2 Active 2 Active 2

*DBAPY A Active 3 Active 2 Active 2 Active 2 Active 2 Active 3 Ends 3 Active 2 Active 2 Active 2 Active 2

Notes:
A. Data groups that use cooperative processing may be affected by the result of this value. Ending database processes while object processes remain active may
result in object activity entries being placed on hold. Similarly, ending object processes while database processes remain active may result in files being placed on
hold due to error.
1. When the database send (DBSND) process ends, the data area polling process also ends.

257
What replication processes are ended by the ENDDG command

2. On installations running 7.1.15.00 or higher, if access path maintenance is enabled, the database apply process signals the access path maintenance job and then ends. The access
path maintenance job uses additional jobs, if needed, to change the access path maintenance attribute to immediate on all files that MIMIX had previously changed to delayed. Any
files that could not be changed are identified as having an access path maintenance error before the maintenance jobs end.
On installations running software earlier than 7.1.15.00, if parallel access path maintenance function is enabled, the associated monitors are also ended when ENDDG command
specifies *DFT for End parallel AP maintenance (PRLAPMNT) and *ALL for the Apply session (APYSSN). When *YES is specified for PRLAPMNT, the function is always ended
regardless of the values specified for PRC or APYSSN.
3. These object replication processes are not available in data groups configured for database-only replication.
4. These database replication processes are not available in data groups configured for object-only replication
5. The database reader (*DBRDR) process is not used by data groups configured for source-send replication.

258
CHAPTER 12 Resolving common replication
problems

Occasionally, a journaled transaction for a file or object may fail to replicate. User
intervention is required to correct the problem. This chapter provides procedures to
help you resolve problems that can occur during replication processing.
The following topics are included in this chapter:
• “Replication problems that MIMIX attempts to recover” on page 260 identifies the
common replication errors that MIMIX will automatically correct when the
database automatic recovery and object automatic recovery policies are enabled.
• “Working with user journal replication errors” on page 263 includes topics for how
to resolve a file that is held due to an error. It also includes topics about options for
placing a file on hold and releasing held files.
• “Working with tracking entries” on page 272 describes how to use tracking entries
to resolve replication errors for IFS objects, data areas, or data queues that are
replicated cooperatively with the user journal. It also includes topics about options
for placing a tracking entry on hold and releasing held tracking entries.
• “Working with objects in error” on page 278 describes how to resolve objects in
error by working with the data group activities used for system journal replication.
This topic includes information about how to retry failed activity entries and how to
determine whether MIMIX is automatically attempting to retry an activity.
• “Removing data group activity history entries” on page 285 describes how to
manually remove completed entries for system journal replication activity. This
may be necessary if you need to conserve disk space.
• “Working with message queues” on page 283 describes how to use the MIMIX
primary and secondary message queues from the native user interface.
• “Working with the message log” on page 284 describes how to access the MIMIX
message log from either user interface.

259
Replication problems that MIMIX attempts to recover
When the Automatic system journal recovery (OBJRCY) and Automatic user journal
recovery (DBRCY) policies are enabled, MIMIX replication processes automatically
attempt to correct detected replication errors. When an error is detected, the
replication process initiates a recovery action to perform the corrective activity.
Replication status remains unchanged for the file entry, tracking entry, or activity entry
associated with the affected file or object while the recovery action is being processed
or is waiting to be processed. If the recovery action is successful, the file or object is
tracked for prioritized auditing. If the recovery action cannot be performed or fails,
then a replication error is reported for the affected file or object.
From the native user interface, counts of new or in-progress recoveries are available
on the Work with Data Groups display and in data group details, but only a limited
subset of replication recoveries can be viewed from the Work with Recoveries display.
When a recovery fails and an error is reported, you can use options available from the
Work with Data Groups display to access details and to manually resolve the
problem.
If you use the MIMIX portal application in Vision Solutions Portal, you can use the
Object Correction Activity window to view new or in-progress recoveries for replication
activity as well as recoveries initiated by other processes, including target journal
inspection, audits, and virtual switch procedures.
The following topics identify the replication errors that MIMIX can detect and attempts
to automatically correct:
• “Replication errors handled by automatic database recovery” on page 260
• “Replication errors handled by automatic object recovery” on page 261
The Automatic system journal recovery (OBJRCY) and Automatic user journal
recovery (DBRCY) policies also enable the ability to automatically correct changes
detected by target journal inspection processes. The types of changes detected and
eligible for automatic correction are described in the MIMIX Administrator Reference
book.

Replication errors handled by automatic database recovery


MIMIX can detect and correct the most common file error situations that occur during
database replication. When the Automatic database recovery policy is enabled and
database replication processes detect an error listed in Table 45, a recovery action is
submitted. MIMIX periodically checks for and processes any recovery actions. The file
entry or tracking entry associated with the file or object in which the error was
detected is only placed in hold error (*HLDERR) status when the recovery action fails
or cannot be performed. Recovery actions that end in an error do not generate a
separate error notification because the error is already reflected in MIMIX status.

260
Replication problems that MIMIX attempts to recover

Table 45. Database replication errors detected and corrected when automatic database recovery is enabled.

Error Description

File level errors Typically invoked when there is a missing library, file, or member. Also invoked
- and - when an attempt to write a record to a file results in a unique key violation.
Unique-key record level Without database automatic recovery, these conditions result in the file being
error placed in *HLDERR status.

Record level errors Invoked when the database apply process detects a data-level issue while
processing record-level transactions.
Without database automatic recovery, any configured collision resolution
methods may attempt to correct the error. Otherwise, these conditions result in
the file being placed in *HLDERR status.

Errors on IFS objects Invoked during the priming of IFS tracking entries when replicated IFS objects
configured for user are determined to be missing from the target system. Priming of tracking entries
journal replication occurs when a data group is started after a configuration change or when Deploy
Data Grp. Configuration (DPYDGCFG) is invoked.

Errors on data area and Invoked during the priming of object tracking entries when replicated data area
data queue objects and data queue objects are determined to be missing from the target system.
configured for user Priming of tracking entries occurs when a data group is started after a
journal replication configuration change or when the Deploy Data Grp. Configuration (DPYDGCFG)
is invoked.

Errors when DBAPY Invoked when a temporary lock condition or an operating system condition exists
cannot open the file or that prevents the database apply process (DBAPY) from opening the file or
apply transactions to the applying transactions to the file. Without database automatic recovery, users
file typically have to release the file so the database apply process (DBAPY) can
continue without error.

Replication errors handled by automatic object recovery


MIMIX can detect and correct the most common object error situations that occur
during replication. When the Automatic object recovery policy is enabled and object
replication processes detect an error listed in Table 46, MIMIX automatically attempts
to correct the error by starting a job to perform an appropriate recovery action.
Unless the object is explicitly excluded from replication for a data group, the automatic
recovery action will synchronize the object to ensure that it is on the target system.
Note: Object automatic recovery does not detect or correct the following problems:
• Missing spooled files on the target system.
• Files and objects that are cooperatively processed. Although the files and
objects are not addressed, problems with authorities for cooperatively
processed files and objects are addressed.
• Activity entries that are “stuck” in a perpetual pending status (PR, PS, PA, or
PB).

261
Only when all recovery options are exhausted without success is an activity entry
placed in error status. Recovery actions that end in an error do not generate a
separate error notification because the error is already reflected in MIMIX status.

Table 46. Object replication errors detected and corrected when automatic object recovery is enabled.

Error Description

Missing objects An object (library-based, IFS, or DLO) exists on the source system and is within the name
on target space for replication, but MIMIX detects that the object does not exist on the target
system1 system. Without object automatic recovery, this results in a failed activity entry.
Notes:
• Missing spooled files are not addressed.
• Missing objects that are configured for cooperative processing are not synchronized.
However, any problems with authorities (*AUTL or *USRPRF) for the missing objects
are addressed.

Missing parent Any operation against an object whose parent object is missing on the target system.
objects on Without object automatic recovery, this condition results in a failed activity entry due to
target system1 the missing parent object.

Missing Any operation that requires a user profile object (*USRPRF) that does not exist on the
*USRPRF target system. Without object automatic recovery, this results in authority or object owner
objects on issues that cause replication errors.
target system1

Missing *AUTL Any operation that requires a authority list (*AUTL) that does not exist on the target
objects on system.Without object automatic recovery, this results in authority issues that cause
target system1 replication errors.

In-use Applications which hold persistent locks on objects can result in object replication errors if
condition the configured values for delay/retry intervals are exceeded. Default values in the data
group definition provide approximately 15 minutes during which MIMIX attempts to
access the object for replication. If the object cannot be accessed during this time, the
result is activity entries with errors of Failed Retrieve (for locked objects on the source
system) and Failed Apply (for locked objects on the target system) and a reason code of
*INUSE.
Notes:
1. The Number of third delay/retries policy and the Third retry interval policy determine
whether automatic recovery is attempted for this error. These policies are in effect
when the object automatic recovery policy is enabled.
2. Automatic recovery for this error is not attempted when the objects are configured for
cooperative processing.
1. The synchronize command used to automatically recover this problem during replication will correct this error any time
the command is used.

262
Working with user journal replication errors

Working with user journal replication errors


MIMIX reports user journal replication errors for files as status on the associated data
group file entry. This status is also reported at the data group level in a consolidated
form.
File replication problems are categorized as follows:
Held due to error - If a journal transaction is not replicated successfully, the file
entry is placed in *HLDERR status. This indicates a problem that must be
resolved.
Held for other reasons - File entries can also be placed in a variety of other held
statuses by user action or by MIMIX. Generally, these statuses are also
considered problems; some are transitional conditions that resolve automatically
while others require user action. To determine if there are files on hold for other
reasons, use the procedure in “Displaying detailed status for a data group” on
page 126.
For information about resolving problems with IFS objects and library-based objects
that are replicated by user journal, see “Working with tracking entries” on page 272.

Working with files needing attention (replication and access path errors)
The DB Errors column on the Work with Data Groups display identifies the number of
errors for user journal replication. Specifically, this column identifies the sum of the
number of database files, IFS, *DTAARA, and *DTAQ objects on hold due to errors
(*HLDERR) plus the number of LF and PF files that have access path maintenance1
failures for a data group. Data group file entries and tracking entries should not be left
in *HLDERR state for any extended time. Access path maintenance errors occur
when MIMIX could not change a file’s access path maintenance attribute back to
immediate.
To access a list of files in error for a data group, do the following:
1. From the MIMIX Basic Main Menu select option 6 (Work with data groups) and
press Enter.
2. The Work with Data Groups display appears. Type 12 (Files needing attention)
next to the data group you want which has errors identified in the DB Errors
column and press Enter.
3. The Work with DG File Entries display appears with a list of file entries for the data
group that have replication errors, access path maintenance2 errors, or both. Do
the following:
a. The initial view shows the current replication status of file entries. Any entry
with a status of *HLD, *HLDERR, *HLDIGN or *HLDRLTD indicates that action
is required. Use Table 47 to identify choices based on the file entry status.

1. Errors for the access path maintenance function are included on installations running MIMIX
7.1.15.00 or higher.
2. Access path maintenance errors can only be reported on data group file entries in installa-
tions running MIMIX 7.1.15.00 or higher.

263
Note: MIMIX retains log spaces for file entries with these statuses so that the
journal entries that are being held can be released and applied to the
target system. File entries should not be left in these states for an
extended period.
b. Use Table 47 to identify choices based on the file entry status and Table 48 to
identify available options from this display.
c. If necessary, take action to prevent the error from happening again. Refer to
the following topics:
• “Correcting file-level errors” on page 269
• “Correcting record-level errors” on page 270
4. Press F10 as needed on the Work with DG File entries display until you see the
access path maintenance view. The AP Maint. Status column identifies any AP
maintenance errors for a file with the value *FAILED and failures for logical files
associated with a file as *FAILEDLF.
Immediate action may not be necessary because MIMIX will attempt to retry
access path maintenance when the data group ends and when it is restarted. To
attempt an immediate retry, use option 40 (Retry AP maintenance).

Table 47. Possible actions based on replication status of a file entry

Status Preferred Action1

*ACTIVE Unless an error has occurred, no action is necessary. Entries in the user
journal for the file are replicated and applied. If necessary, any of the
options to hold journal entries can be used.

*HLD User action is required to release the file entry (option 26) so that held
journal entries from the user journal can be applied to the target system.

*HLDERR User action is required. Attempt to resolve the error by synchronizing the
file (option 16).
Note: Transactions and hold logs are discarded for file entries with a status of
*HLDERR and an error code of IG. Such a file must be synchronized.

*HLDIGN User action is required to either synchronize the file (option 16) or to
change the configuration if you no longer want to replicate the file.
Journal entries for the file are discarded. Replication is not occurring and
the file may not be synchronized.
Depending on the circumstances, Release may also be an option.

*HLDRGZ These are transitional states that should resolve to *ACTIVE. If these
*HLDRNM status persist, check the journaling status for the entry. MIMIX retains log
*HLDPRM spaces for the held journal entries for the duration of these temporary
hold requests.
*HLDSYNC

264
Working with user journal replication errors

Table 47. Possible actions based on replication status of a file entry

Status Preferred Action1

*HLDRTY The file entry is held because an entry could not be applied due to a
condition which required waiting on some other condition (such as in-
use). After a short delay, the database apply job will automatically
attempt to process this entry again. The preferred action is to allow
MIMIX to periodically retry the file entry. By default, the database apply
job will automatically attempt to process the entry every 5 minutes for up
to 1 hour.
Manually releasing the file entry will cause MIMIX to attempt to process
the entry immediately.
Note: From Vision Solutions Portal, you can view information about locks on the
object on both systems. From the File Activity tab on the Data Group
Details and Activities portlet, select the Compare Details action to view
the Locks tab of the Compare Object Details from Nodes window.

*HLDRLTD User action is required for a file in the same network. View the related
files (option 35). A file that is related due to a dependency, such as a
constraint or a materialized query table, is held. Resolving the problem
for the related held file will resolve this status.

*RLSWAIT The file is waiting to be released by the DB apply process and will be
changed to *ACTIVE. If the status does not change to *ACTIVE, check
the journaling status. If this status persists, you may need to synchronize
(option 16).

*CMPACT These are transitional states that should resolve automatically. The file
*CMPRLS entry represents a member that is being processed cooperatively
*CMPRPR between the CMPFILDTA command and the database apply process.

1. Evaluate the cause of the problem before taking any action.

Table 48. Options for working with file entries from the Work with DG FIle Entries display

Option Additional Information

9=Start journaling See “Starting journaling for physical files” on page 291.

10=End journaling See “Ending journaling for physical files” on page 292.

11=Verify journaling See “Verifying journaling for physical files” on page 293.

16=Sync DG file See topic ‘Synchronizing database files’ in the MIMIX


entry Administrator Reference book.

20=Work with file See topic “Working with journal transactions for files in error” on
error entries page 266.

23=Hold file See topic “Placing a file on hold” on page 267.

24=Ignore file See topic “Ignoring a held file” on page 267.

25=Release wait See topic “Releasing a held file at a synchronization point” on


page 268.

265
Table 48. Options for working with file entries from the Work with DG FIle Entries display

Option Additional Information

26=Release See topic “Releasing a held file” on page 268.

27=Release clear See topic “Releasing a held file and clearing entries” on page 269.

31=Repair member Available for entries with a status of *HLDERR that identify a
data member. See topic ‘Comparing and repairing file data - members
on hold (*HLDERR)’ in the MIMIX Administrator Reference book.

35=Work with related Displays file entries that are related to the selected file by
files constraints or by other dependencies such as materialized query
tables

40=Retry AP Retries access path maintenance operations on the target system


maintenance for the selected file. This option is only valid on data group file
entries that have an access path maintenance status of *FAILED
or *FAILEDLF.

Working with journal transactions for files in error


When resolving problems for a file that is in *HLDERR state, a MIMIX administrator
may find it useful to examine the journal entries that are being held by MIMIX.
Although you can determine why a file is in error from either the source or target
system, to view the actual journal entries, you must be on the target system. If you
attempt to view the journal entries from the source system, MIMIX will indicate that
you are on the incorrect system to view the information.
Do the following:
1. From the subsetted list of files in error for a data group on the Work with DG File
Entries display, type 20 (Work with file error entries) next to the file entry you want
and press Enter.
2. The Work with DG FE on Hold display appears. A variety of information about the
transaction appears on the display.
Note: The values shown in the Sequence number column may be truncated if
the journal supports *MAXOPT3 for the receiver size and the journal
sequence number value exceeds the available display field. When
truncation is necessary, the most significant digits (left-most) are omitted.
Truncated journal sequence numbers are prefixed by '>'. The First journal
sequence number field displays the full sequence number of the first item
displayed in the list.
a. Locate the transaction that caused the file to be placed on hold. Use the
Position to field to position the list to a specific sequence number.

266
Working with user journal replication errors

b. Select the option (Table 49) you want to use on the journal transaction:

Table 49. Options available from the Work with DG FE on Hold display.

2=Change You can change the contents or characteristics of the journal entry.
Use this option with caution. Any changes can affect the validity of
data in the journal entry.

4=Delete You can delete the journal entry.

5=Display You can display details for the specified journal entry associated with
the data group file entry in question.

9=Immediate You can immediately apply a transaction that has caused a file to go
apply on hold. The entry you selected is immediately applied to the file
outside of the apply process. If the apply is successful, the error/hold
entry that was applied is removed from the error/hold log. However, if
the apply fails, a message is issued and the entry remains in the
error/hold log. This process does not release the file; it only applies the
selected entry.

Placing a file on hold


Use this procedure to hold any journal entries for a file identified by a data group file
entry. Avoid leaving a file entry on hold for any extended period.
File entries with a status of *ACTIVE, *HLDRGZ, *HLDRNM, *HLDPRM, *HLDSYNC,
*HLDRLTD, and *RLSWAIT can be placed on hold.
The request changes the file entry status to *HLD. Any journal entries for the
associated file are replicated but not applied. If the file is being processed by an active
apply session, suspending of the update process can take a short time to complete.
You will receive a message when the file is held. MIMIX retains log spaces containing
any replicated journal entries in anticipation that the file entry will be released. When
the file is released, the accumulated journal entries will be applied. The *HLD status
remains until additional action is taken.
Do the following:
1. From the MIMIX Basic Main Menu select option 6 (Work with Data Groups) and
press Enter.
2. The Work with Data Groups display appears. Type 17 (File entries) next to the
data group you want and press Enter.
3. The Work with DG File Entries display appears. Type 23 (Hold file) next to the
entry you want and press Enter.

Ignoring a held file


Use this procedure to ignore any journal entries for an file identified by a data group
file entry. The request changes the file entry status to *HLDIGN. Any journal entries
for the associated file, including any hold logs, are discarded. The *HLDIGN status
remains until additional action is taken.

267
Note: Be certain that you want to use the ignore feature. Any ignored transactions
cannot be retrieved. You must replace the object on the target system with a
current version from the source system.
If a file has been on hold for a long time or you expect that it will be, the amount of
storage used by the error/hold log space can be quite large. If you anticipate that you
will need to save and restore the file or replace it for any other reason, it may be best
to just ignore all current transactions.
Do the following:
1. From the MIMIX Basic Main Menu select option 6 (Work with Data Groups) and
press Enter.
2. The Work with Data Groups display appears. Type 17 (File entries) next to the
data group you want and press Enter.
3. The Work with DG File Entries display appears. Type 24 (Ignore file) next to the
entry you want and press Enter.
The status of the file is changed to *HLDIGN. The file entry is ignored. Journal
entries for the file entry, including any hold logs, are discarded.

Releasing a held file at a synchronization point


Use this procedure to wait for a synchronization point to release any held journal
entries for file identified by a data group file entry, then resume replication.
The request changes the file entry status to *RLSWAIT. Any journal entries for the
associated file are discarded until a File member saved (F-MS) journal entry or a Start
of save of a physical file member using save-while-active function (F-SS) is
encountered. This is the synchronization point. The file entry status is then changed
to *ACTIVE and all journal entries that were held after the synchronization point are
applied.
If the F-MS or F-SS journal entry is not in the log space, the file entry remains in
*RLSWAIT status. If you are unsure as to how many save requests might accumulate
for an object, you can synchronize the file associated with the file entry. The entry
status will become *ACTIVE.
To wait for a synchronization point before releasing a held file, do the following:
1. From the MIMIX Basic Main Menu select option 6 (Work with Data Groups) and
press Enter.
2. The Work with Data Groups display appears. Type 17 (File entries) next to the
data group you want and press Enter.
3. The Work with DG File Entries display appears. Type 25 (Release wait) next to
the entry you want and press Enter.

Releasing a held file


Use this procedure to immediately release any held journal entries for file identified by
a data group file entry with a status of *HLD and resume replication.

268
Working with user journal replication errors

The request changes the file entry status to *ACTIVE. Any held journal entries for the
associated file are applied. Normal replication of the file resumes.
While a file is being released, the appropriate apply session suspends its operations
on other files. This allows the released file to catch up to the current level of
processing. If a file or member has been on hold for a long time, this can be lengthy.
Do the following to immediately release a held file or file member:
1. From the MIMIX Basic Main Menu select option 6 (Work with Data Groups) and
press Enter.
2. The Work with Data Groups display appears. Type 17 (File entries) next to the
data group you want and press Enter.
3. The Work with DG File Entries display appears. Type 26 (Release) next to the
entry you want and press Enter

Releasing a held file and clearing entries


Use this procedure to clear any held journal entries for a file identified by a data group
file entry, then resume replication.
The request changes the file entry status to *ACTIVE. Any held journal entries for the
associated file are discarded. Journal entries received after the file entry status
became *ACTIVE are applied, resuming normal replication.
If a file entry is on hold and its associated file has been synchronized in such a way
that the held entries already exist in the restored file, this procedure will ensure that
those entries are not re-applied. This procedure will not work if the file is being
actively updated on the source system.
Do the following to release a held file and clear any journal entries that were
replicated but not applied:
1. From the MIMIX Basic Main Menu select option 6 (Work with Data Groups) and
press Enter.
2. The Work with Data Groups display appears. Type 17 (File entries) next to the
data group you want and press Enter.
3. The Work with DG File Entries display appears. Type 27 (Release clear) next to
the entry you want and press Enter.

Correcting file-level errors


Typically, file-level errors can be categorized as one of the following:
• A problem with the configuration of files defined for replication.
• A discrepancy in the file descriptions between the management and network
systems
• An operational error.
This topic identifies the most common file-level errors and measures that you can
take to prevent the problem from recurring. See also “Correcting record-level errors”
on page 270.

269
Once you diagnose and correct a file-level error, the problem rarely manifests itself
again. Some of the most common file-level errors are:
• Authority: The MIMIXOWN user profile defined in the MIMIX job description does
not have authority to perform a function on the target system. You can prevent this
problem by ensuring that the MIMIXOWN user profile has all object authority
(*ALLOBJ). This guarantees that the user profile has all the necessary authority to
run IBM i commands and has the ability to access the library and files on the
management system. If you are attempting to replicate files that have protected
rows or columns, or system-period temporal tables and their associated history
tables, the MIMIXOWN user profile may not have been granted the necessary
function usage. For more information about requirements for replicating files that
use permissions or masks for row and column access control (RCAC), see the
MIMIX Administrator Reference book. For more information about the MIMIXOWN
user profile and authority, see the Using License Manager book.
• Objects existence or corruption: MIMIX cannot run a function against a file on
the target system because the file or a supporting object (such as logical files)
does not exist or has become damaged. System security is the only way to
prevent an object from being accidentally deleted from the target system. Make
sure that only the correct personnel have the ability to remove objects from the
target system where replicated data is applied. Also, ensure that application
programs do not delete files on the target system when there are no apply
sessions running.
• MIMIX subsystem ended: If the MIMIX subsystem is ended in an immediate
mode while MIMIX processes are still active, files may be placed in a “Held”
status. This is a result of MIMIX being unable to complete a transaction normally.
After MIMIX is restarted, you only need to release the affected files.

Correcting record-level errors


Record-level errors occur when MIMIX updates or attempts to update a file and the
feedback from the update process indicates a discrepancy between the files on the
management and network system. Record-level errors can usually be traced back to
problems with one of the following:
• The system
• Unique application environments, such as System 36 code running in native IBM
i.
• Operational errors.

Record written in error


This section describes the most common record-level errors. MIMIX DB Replicator
was able to write the record on the target system; however, it wrote to the wrong
relative record number. In most situations, the IBM i database function writes a new
record to the end of a file. MIMIX did so, but it did not match the relative record
number of the sending system. Usually this error occurs when transactions (journal
entries) are skipped on the send system. Common reasons why records are written in
error include the following:

270
Working with user journal replication errors

• Journaling was ended: When journaling is ended, transaction images are not
being collected. If users update the files while journaling is not running, no journal
entries are created and MIMIX DB Replicator has no way of replicating the
missing transactions. The best way to prevent this error is to restrict the use of the
Start Journaling Physical File (STRJRNPF) and End Journaling Physical File
(ENDJRNPF) commands.
• User journal replication was restarted at the wrong point: When you change
the starting point of replication for a data group, it is imperative that transactions
are not skipped.
• Apply session restarted after a system failure: This is caused when the target
system experiences a hard failure. MIMIX always updates its user spaces with the
last updated and sent information. When a system fails, some information may not
be forced to disk storage. The data group definition parameter for database apply
processing determines how frequently to force data to disk storage. When the
apply sessions are restarted, MIMIX may attempt to rewrite records to the target
system database.
• Unable to write/update a record: This error is caused when MIMIX cannot
access a record in a file. This is usually caused when there are problems with the
logical files associated with the file or when the record does not exist. The best
way to prevent this error is to make sure that replication is started in the correct
position. This error can also be due to one of the problems listed in topic
“Correcting file-level errors” on page 269.
• Unable to delete a record: This is caused when MIMIX is trying to delete a
record that does not exist or has a corrupted logical file associated with the
physical file. This error can also be due to one of the problems listed in topic
“Correcting file-level errors” on page 269.

271
Working with tracking entries
Tracking entries identify library-based objects (data areas and data queues) and IFS
objects configured for cooperative processing (advanced journaling).
You can access the following displays to work with tracking entries in any status:
• Work with DG IFS Trk. Entries display (WRKDGIFSTE command)
• Work with DG Obj. Trk. Entries display (WRKDGOBJTE command)
These displays provide access for viewing status and working with common problems
that can occur while replicating objects identified by IFS and object tracking entries.
Held tracking entries: Status for the replicated objects is reported on the associated
tracking entries. If a journal transaction is not replicated successfully, the tracking
entry is placed in *HLDERR status. This indicates a problem that must be resolved.
Tracking entries can also be placed in *HLD, *HLDIGN statuses by user action These
statuses are reported as ‘held for other reasons’ and also require user action.
When a tracking entry has a status of *HLD or *HLDERR, MIMIX retains log spaces
so that journal entries that are being held can be released and applied to the target
system. Tracking entries should not be left in these states for an extended period.
Additional information: To determine if a data group has any IFS objects, data
areas, or data queues configured for advanced journaling, see “Determining if non-file
objects are configured for user journal replication” on page 336.
When working with tracking entries, especially for IFS objects, you should be aware of
the information provided in “Displaying long object names” on page 330.

Accessing the appropriate tracking entry display


To access IFS tracking entry or object tracking entry displays for a data group, do the
following:
1. From the MIMIX Basic Main Menu select option 6 (Work with Data Groups) and
press Enter.
2. The Work with Data Groups display appears. Type 17 (File entries) next to the
data group you want and press Enter.
3. Next to the data group you want, type the number for the option you want and
press Enter. Table 50 shows the options for tracking entries.

Table 50. Tracking entry options on the Work with Data Groups display

Select Option Result

50=IFS trk entries Lists all IFS tracking entries for the selected data group on the
Work with DG IFS Trk. Entries display.

51=IFS trk entries Lists IFS tracking entries for the selected data group with inactive
not active status values (*HLD, *HLDERR, *HLDIGN, *HLDRNM, and
*RLSWAIT) on the Work with DG IFS Trk. Entries display.

272
Working with tracking entries

Table 50. Tracking entry options on the Work with Data Groups display

Select Option Result

52=Obj trk entries Lists all object tracking entries for the selected data group on the
Work with DG Obj. Trk. Entries display.

53=Obj trk entries Lists object tracking entries for the selected data group with
not active inactive status values (*HLD, *HLDERR, *HLDIGN, and
*RLSWAIT) on the Work with DG Obj. Trk. Entries display.

4. The tracking entry display you selected appears. Significant capability is available
for addressing common replication problems and journaling problems. Do the
following:
a. Use F10 to toggle between views showing status, journaling status, and the
database apply session in use.
b. Any entry with a status of *HLD, *HLDERR or *HLDIGN indicates that action is
required. The identified object remains in this state until action is taken.
Statuses of *HLD and *HLDERR result in journal entries being held but not
applied. Use Table 51 to identify choices based on the tracking entry status.
c. Use options identified in Table 52 to address journaling problems or replication
problems.

Table 51. Possible actions based on replication status of a tracking entry

Status Preferred Action1

*ACTIVE Unless an error has occurred, no action is necessary. Entries in the user
journal for the IFS object are replicated and applied. If necessary, any of
the options to hold journal entries can be used.

*HLD User action is required to release the entry (option 26) so that held
journal entries from user journal can be applied to the target system.

*HLDERR User action is required. Attempt to resolve the error by synchronizing the
file (option 16).

*HLDIGN User action is required to either synchronize the object (option 16) or to
change the configuration if you no longer want to replicate the object.
Journal entries for the object are discarded. Replication is not occurring
and the object may not be synchronized.
Depending on the circumstances, Release may also be an option.

*HLDRNM This is a transitional state for IFS tracking entries that should resolve to
*ACTIVE. If this status persists, check the journaling status for the entry.
Object tracking entries cannot have this status.

*RLSWAIT If the status does not change to *ACTIVE, you may need to synchronize
(option 16)
1. Evaluate the cause of the problem before taking any action.

273
Table 52. Options for working with tracking entries

Option Additional Information

4=Remove See “Removing a tracking entry” on page 276

5=Display Identifies an object, its replication status, journaling status, and the
database apply session used.

6=Print Creates a spooled file which can be printed

9=Start journaling See “Starting journaling for IFS objects” on page 294 and “Starting
journaling for data areas and data queues” on page 298.

10=End journaling See “Ending journaling for IFS objects” on page 295 and “Ending
journaling for data areas and data queues” on page 299.

11=Verify journaling See “Verifying journaling for IFS objects” on page 296 and
“Verifying journaling for data areas and data queues” on page 300.

16=Synchronize Synchronizes the contents, attributes, and authorities of the object


represented by the tracking entry between the source and target
systems.
For more information, see topic ‘Synchronizing tracking entries’ in
the MIMIX Administrator Reference book.

23=Hold See “Holding journal entries associated with a tracking entry” on


page 274.

24=Ignore See “Ignoring journal entries associated with a tracking entry” on


page 275.

25=Release wait See “Waiting to synchronize and release held journal entries for a
tracking entry” on page 275.

26=Release See “Releasing held journal entries for a tracking entry” on


page 276.

27=Release clear See “Releasing and clearing held journal entries for a tracking
entry” on page 276.

35=Work with related Displays IFS tracking entries that have the same file identifier
objects (FID) as the specified tracking entry. See “Related tracking entries”
on page 277.

Holding journal entries associated with a tracking entry


Use this procedure to hold any journal entries for an object identified by a tracking
entry. Avoid leaving a tracking entry on hold for any extended period.
The request changes the tracking entry status to *HLD. Any journal entries for the
associated IFS object, data area, or data queue are replicated but not applied. MIMIX
retains log spaces containing any replicated journal entries in anticipation that the
tracking entry will be released. When the tracking entry is released, the accumulated

274
Working with tracking entries

journal entries will be applied. The *HLD status remains until additional action is
taken.
Do the following:
1. Access the IFS or object tracking entry display as described in “Accessing the
appropriate tracking entry display” on page 272.
2. Type 23 (Hold) next to the tracking entry for the object you want and press Enter.

Ignoring journal entries associated with a tracking entry


Use this procedure to ignore any journal entries for an object identified by a tracking
entry. The request changes the tracking entry status to *HLDIGN. Any journal entries
for the associated IFS object, data area, or data queue, including any hold logs, are
discarded. The *HLDIGN status remains until additional action is taken.
Note: Be certain that you want to use the ignore feature. Any ignored transactions
cannot be retrieved. You must replace the object on the target system with a
current version from the source system.
If a tracking entry has been on hold for a long time or you expect that it will be, the
amount of storage used by the error/hold log space can be quite large. If you
anticipate that you will need to save and restore the object or replace it for any other
reason, it may be best to just ignore all current transactions.
Do the following:
1. Access the IFS or object tracking entry display as described in “Accessing the
appropriate tracking entry display” on page 272.
2. Type 24 (Ignore) next to the tracking entry for the object you want and press
Enter.

Waiting to synchronize and release held journal entries for a tracking


entry
Use this procedure to wait for a synchronization point to release any held journal
entries for an object identified by a tracking entry, then resume replication.
The request changes the tracking entry status to *RLSWAIT. Any journal entries for
the associated IFS object, data area, or data queue are discarded until an object
saved journal entry is encountered. This is the synchronization point. The tracking
entry status is then changed to *ACTIVE and all journal entries that were held after
the synchronization point are applied.
If the object saved journal entry is not in the log space, the tracking entry remains in
*RLSWAIT status. If you are unsure as to how many save requests might accumulate
for an object, you can synchronize the object associated with the tracking entry. The
tracking entry status will become *ACTIVE.
Do the following:
1. Access the IFS or object tracking entry display as described in “Accessing the
appropriate tracking entry display” on page 272.
2. Type 25 (Release wait) next to the tracking entry for the object you want and

275
press Enter.

Releasing held journal entries for a tracking entry


Use this procedure to immediately release any held journal entries for an object
identified by a tracking entry with a status of *HLD or *HLDERR and resume
replication.
The request changes the tracking entry status to *ACTIVE. Any held journal entries
for the associated IFS object, data area, or data queue are applied. Normal replication
of the object resumes.
Do the following:
1. Access the IFS or object tracking entry display as described in “Accessing the
appropriate tracking entry display” on page 272.
2. Type 26 (Release) next to the tracking entry for the object you want and press
Enter

Releasing and clearing held journal entries for a tracking entry


Use this procedure to clear any held journal entries for an object identified by a
tracking entry, then resume replication.
The request changes the tracking entry status to *ACTIVE. Any held journal entries
for the associated IFS object, data area, or data queue are discarded. Journal entries
received after the tracking entry status became *ACTIVE are applied, resuming
normal replication.
If a tracking entry is on hold and its associated object has been synchronized in such
a way that the held entries already exist in the restored object, this procedure will
ensure that those entries are not re-applied.
Do the following:
1. Access the IFS or object tracking entry display as described in “Accessing the
appropriate tracking entry display” on page 272.
2. Type 27 (Release clear) next to the tracking entry for the object you want and
press Enter.

Removing a tracking entry


Use this procedure to remove a duplicate tracking entry for an IFS object, data area,
or data queue. A tracking entry with a status of *HLDERR cannot be removed.
Note: Do not use this procedure to prevent user journal replication of an object
represented by a tracking entry. If you need to exclude the object from
replication or have it replicated through the system journal instead of the user
journal, change or create the appropriate data group IFS entry or object entry.
Do the following:
1. Access the IFS or object tracking entry display as described in “Accessing the
appropriate tracking entry display” on page 272.

276
Working with tracking entries

2. Type 4 (Remove) next to the tracking entry you want to remove and press Enter.
Note: Related tracking entries (entries with the same file identifier) will also be
removed when a tracking entry is removed.
3. You will see a confirmation display. To remove the tracking entry, press Enter.

Related tracking entries


Use this procedure to view related IFS tracking entries. Related IFS tracking entries
are tracking entries that have the same file identifier (FID).
Do the following:
1. Access the IFS or object tracking entry display as described in “Accessing the
appropriate tracking entry display” on page 272.
2. Type 35 (Work with related objects) next to the tracking entry whose related IFS
tracking entries you want to display and press Enter.

277
Working with objects in error
Use this topic to work with replication errors for objects replicated through the system
journal.
To access a list of objects in error for a data group, do the following:
1. From the MIMIX Basic Main Menu select option 6 (Work with Data Groups) and
press Enter.
2. The Work with Data Groups display appears. Type 13 (Objects in error) next to
the data group you want which has values shown in the Obj Errors column and
press Enter.
3. The Work with Data Group Activity display appears with a list of the objects in
error for the data group you selected. You can do any of the following:
• Use F10 (Error view) to see the reason why the object is in error.
• Use F11 to change between views for objects, DLOs, IFS objects, and spooled
files.
• Use the options identified in Table 53 to resolve the errors. Type the number of
the option you want next to the object and press Enter

Table 53. Options on the Work with Data Group Activity display for working with objects in
error.

4=Remove Use this option to remove an entry with a *COMPLETED or


*FAILED status from the list. For entries with *FAILED status, this
option removes only the failed entry. Prompting is available for
extended capability. You may need to take action to synchronize
the object associated with the entry.
Note: If an entry with a status of *FAILED has related entries in
*DELAYED status, you can remove both the failed and the
delayed entries in one operation by using option 14 (Remove
related).
For more information, see “Removing data group activity history
entries” on page 285.

7=Display message Use this option to display any error message that is associated
with the entry.

8=Retry Use this option to retry the data group activity. MIMIX changes the
entry status to pending and attempts the failed operation again.
Note: It is possible to schedule the request for a time when the retry is
more likely to be successful. For more information about retrying
failed entries, see “Retrying data group activity entries” on
page 281.

278
Working with objects in error

Table 53. Options on the Work with Data Group Activity display for working with objects in
error.

12=Work with entries Use this option to access the Work with DG Activity Entries
display. From the display you can display additional information
about replicated journal transactions for the object, including the
journal entry type and access type (if available), as well as see
whether the object is undergoing delay retry processing. You can
also take options to display related entries, view error messages
for a failure, and synchronize the object. For more information,
see “Using the Work with DG Activity Entries display” on
page 279.

14=Remove related Use this option to remove an entry with a status of *FAILED and
any related entries that have a status of *DELAYED. You may
need to take action to synchronize the object associated with the
entry.

Using the Work with DG Activity Entries display


From the Work with DG Activity Entries display, you can display information about and
take actions on activity entries for a replicated object. To access the display, select
option 12 (Work with entries) from the Work with Data Group Activity display.
Table 54 lists the available options,

Table 54. Options available on the Work with DG Activity Entries display.

4=Remove Use this option to remove an individual entry with a


*COMPLETED or *FAILED status from the list. For entries with
*FAILED status, this option removes only the failed entry. You may
need to take action to synchronize the object associated with the
entry.
Note: No prompting is available when using this option from this display.
To prompt for additional capability, use the option to remove from
the Work with Data Group Activity display. For more information,
see “Removing data group activity history entries” on page 285.

5=Display Use this option to display details about the individual entry. The
information available about the object includes whether the object
is undergoing delay retry processing, and journal entry
information, including access type information for T-SF, T-YC, and
T-ZC journal entry types. The subtype is shown for U-MX journal
entries. For more information, see “Determining whether an
activity entry is in a delay/retry cycle” on page 282

6=Print Use this option to print the entry.

7=Display message Use this option to display the error message associated with the
processing failure for the entry.

8=Retry Use this option to retry the data group activity entry. MIMIX
changes the entry status to pending and attempts the failed
operation again as soon as possible.

279
Table 54. Options available on the Work with DG Activity Entries display.

9=Display related Displays entries related to the specified object. For example, use
this option to see entries associated with a move or rename
operation for the object.

12=Display job Displays the job that was processing the object when the error
occurred, if the still job information exists and is on this system.

16=Synchronize Use this option to synchronize objects defined to MIMIX for


system journal replication (objects that are not configured for
cooperative processing). Activity entries with *ACTIVE or
*COMPLETED status can be synchronized, as well as entries
with a *FAILED status and with the following journal types: T-CO,
T-CP, T-OR, T-SE, T-ZC (see notes), T-YC, and T-SF (see notes).
A confirmation display allows you to confirm your choices before
the request is processed. Entries are placed in a ‘pending
synchronization’ status. When the data group is active, the
contents of the object, its attributes, and its authorities are
synchronized between the source and target systems. The status
of the activity entry is set to ‘completed by synchronization.’
Notes:
• To synchronize files defined for cooperative processing, use
the Synchronize DG File Entry (SYNCDGFE) command.
• Spooled files (T-SF journal entries) with the following access
types can be synchronized: C = spooled file created; U =
spooled file changed.
• Changed objects (T-ZC journal entries) with the following
access types can be synchronized: 1 (Add); 7 (Change); 25
(Initialize); 29 (Merge); 30 (Open); 34 (Receive); 36
(Reorganize); 50 (Set); and 51 (Send).

280
Retrying data group activity entries
Data group activity entries that did not successfully complete replication have a status
of *FAILED. These failed data group activity entries are also called error entries. You
can request to retry processing for these activity entries.
Activity entries with a status of *ACTIVE can also be retried in some circumstances.
For example, you may want to retry an entry that is delayed but which has no
preceding pending activity entry. Or, you may want to retry a pending entry that is
undergoing processing in a delay retry cycle.
The retry request places the activity entry in the queue for processing by the system
journal replication process where the failure or delay occurred. Activity entries with a
status of *FAILED or *DELAYED are set to *PENDING until they are processed.

Retrying a failed data group activity entry


You can manually request that MIMIX retry processing for a data group activity entry
that has a status of *FAILED. The retry can be requested from either the Work with
Data Group Activity display or from the Work with DG Activity Entries display.
Note: Only the Work with Data Group Activity supports the ability to schedule the
retry request for a time in the future when the request is more likely to be
successful.
To retry failed (error) activity entries, do the following:
1. From the Work with Data Groups display, type a 13 (Objects in error) next to the
data group you want that has values shown in the Obj Errors column and press
Enter.
2. The Work with Data Group Activity display appears with a list of the objects in
error for the data group selected. Type an 8 (Retry) next to the entry you want and
do one of the following:
• To submit the retry request for immediate processing, press Enter. Then skip
to Step 4.
• To schedule the retry request for a time at which it is more likely to be
successful, press F4 (Prompt).
3. On the Retry DG Activity Entries (RTYDGACTE) display, specify a value for the
Time of day to retry prompt. Then press Enter.
You can specify a specific time within 24 hours. The scheduled time is based on
the time on the system from which the request is submitted regardless of the
system on which the activity to retry occurs. When you submit a retry request for a
scheduled time, MIMIX will make the entry active and will wait until the specified
time before retrying the request. The scheduled time is the earliest the request will
be processed. Be sure to consider any time zone differences between systems as
you determine a scheduled time. For additional information and examples, press
F1 (Help).
4. The Confirm Retry of DG Activity display appears. Press Enter.
If failed activity entries occur frequently, consider using the third delay retry cycle.
When the Automatic object recovery policy is enabled, a third retry cycle is performed

281
using the settings in effect from the Number of third delay/retries and Third retry
interval (min.) policies. These policies can be set for the installation or for a specific
data group.

Determining whether an activity entry is in a delay/retry cycle


This procedure allows you to check the status of an activity entry to determine
whether MIMIX is attempting automatic delay retry cycles for the object.
1. From the Work with Data Groups display, type a 14 (Active objects) next to the
data group you want and press Enter.
2. The Work with Data Group Activity display appears with a list of the objects that
are actively being replicated.
3. Type a 12 (Work with Entries) next to the list entry for the object you want and
press Enter. The Work with DG Activity Entries display appears with the list of
activity entries for the object you selected.
4. To view additional details for an entry, type a 5 (Display) next to the activity entry
you want and press Enter. The Display DG Activity Details display appears.
5. Check the value listed in the Waiting for retry field.
The value *YES is displayed when the activity entry is undergoing automatic
delay/retry processing. Delayed or failed activity entries and pending activity
entries that are not in a delay retry cycle will always have a value of *NO.
6. When the value of the Waiting for retry field is *YES, the Delay/Retry Processing
Information fields are also available and provide the following information:
• The Retries attempted field identifies the number of times that MIMIX has
attempted to process the activity entry.
• The Retries remaining field identifies the remaining number of times that
MIMIX can automatically attempt to retry the activity entry. MIMIX uses only as
many of the remaining retry attempts as necessary to achieve a successful
attempt.
• The Delay interval (seconds) field identifies the number of seconds between
the previous attempt and the next retry attempt.
• The Timestamp of next attempt field identifies the approximate date and time
that MIMIX will make the next attempt to process the activity entry. If object
replication processes are busy processing other entries, there may be a delay
between this time and when processing of this entry is actually attempted. The
value *PENDING indicates that the time of the next attempt has passed and
processing for the entry is waiting while other entries are being processed.
This field is displayed only on the system of the process that is in delay/retry.

282
Working with message queues

Working with message queues


You can access the MIMIX primary and secondary message queues to display
messages or manage the list of messages.
Do the following to access a MIMIX message queue:
1. Type the command DSPMMXMSGQ and press F4 (Prompt).
2. Specify either *PRI or *SEC to access the message queue you want and press
Enter.
3. The Display MIMIX Message Queue display appears listing all of the current
messages. To view all of the information for a message, place the cursor on the
message you want and press Enter.
You can also use the function keys on this display to perform several message-
related tasks. Refer to the help text (F1 key) for information about these function keys.
Note: The MIMIX primary and secondary message queues are defined for each
system definition. You can control the severity and type of messages to be
sent to each message queue through parameters on the system definition.

283
Working with the message log
The MIMIX message log provides a common location for you to see all messages
related to MIMIX products. A consolidated list of messages for all systems in the
installation library is available on the management system.
Note: The target system only shows messages that occurred on the target system.
LVI messages are informational messages and LVE messages are error or diagnostic
messages. CPF messages are generated by an underlying operating function and
may be passed up to the MIMIX product.
Do the following to access the MIMIX message log:
1. Do one of the following to access the message log display:
• From the MIMIX Basic Main Menu, select option 13 (Work with messages) and
press Enter.
• From the MIMIX Intermediate Main Menu, select option 3 (Work with
messages) and press Enter.
2. The Work with Message Log appears with a list of the current messages. The
initial view shows the message ID and text.
3. Press F11 to see additional views showing the message type, severity, the
product and process from which it originated, whether it is associated with a group
(for MIMIX, a data group), and the system on which it originated.
4. You can subset the messages shown on the display. A variety of subsetting
options are available that allow you to manage the message log more efficiently.
5. To work with a message, type the number of the option you want and press Enter.
The following options are available:
• 4=Remove - Use this option if you want to delete a message. When you select
this option, a confirmation display appears. Verify that you want to delete the
messages shown and press Enter. The message is deleted only from the local
system.
• 5=Display message - Use this option to view the full text of the first level
message and gain access to the second level text.
• 6=Print - Use this option to print the information for the message.
• 8=Display details - Use this option to display details for a message log entry
including its from and to program information, job information, group
information, product, process, originating system, and call stack information.
• 9=Related messages - Use this option to display a list of messages that relate
to the selected message. Related messages include a summary and any detail
messages immediately preceding it. This can be helpful when you have a large
message log list and you want to show the messages for a certain job.
• 12=Display job - If job information exists on the system, you can use this option
to access job information for a message log entry. The Work with Jobs display
appears from which you can select options for displaying specific information
about the job.

284
Removing data group activity history entries

Removing data group activity history entries


MIMIX maintains history of successfully completed distribution requests to provide a
record of all object, DLO, and IFS replication activity completed by system journal
replication processes. While MIMIX efficiently uses disk space and removes
completed requests according to the value specified in the Keep data group history
parameter of the system definition, you may occasionally need to manually remove
completed activity entries. One reason to manually remove completed entries may be
to conserve disk space, while another may be to clean up entries for an object that
has been removed from replication as a result of a configuration change.
Note: Your business policies and procedures may require that you archive
completed activity entries to tape before you delete them.
To remove completed activity entries, do the following:
1. From the Work with Data Groups display, type 28 (Completed objects) next to the
data group you want and press Enter. The Work with Data Group Activity display
appears with a list of objects with completed entries.
2. Type a 4 (Remove) next to the entry you want and do one of the following:
• To remove all available completed entries for the selected object, press Enter.
Then continue with Step 4.
• To change the selection criteria to include entries for additional objects or to
limit the entries based on a time range, press F4 (Prompt). The Remove DG
Activity Entries (RMVDGACTE) display appears.
3. To change the selection criteria, do the following as needed:
• To remove a subset of completed entries for the selected object based on the
timestamp of the replicated journal entries, specify values for Starting date and
time and Ending date and time prompts.
• To expand the set of objects for which completed entries will be removed,
change the values of the following prompts as needed:
For an expanded set of object types, use the Object type prompt.
For a library based object, use the Object and Library prompts.
For a DLO, use the Document and Folder prompts.
For an IFS object use the IFS object prompt.
For a spooled file, use the Spooled file name, Output queue, and Library
prompts.
4. A confirmation display appears. Press Enter.

285
Starting, ending, and verifying journaling

CHAPTER 13Starting, ending, and verifying


journaling

This chapter describes procedures for starting and ending journaling. Journaling must
be active on all files, IFS objects, data areas and data queues that you want to
replicate through a user journal. Normally, journaling is started during configuration.
However, there are times when you may need to start or end journaling on items
identified to a data group.
The topics in this chapter include:
• “What objects need to be journaled” on page 287 describes, for supported
configuration scenarios, what types of objects must have journaling started before
replication can occur. It also describes when journaling is started implicitly, as well
as the authority requirements necessary for user profiles that create the objects to
be journaled when they are created.
• “What objects need to be journaled” on page 287 describes, for supported
configuration scenarios, what types of objects must have journaling started before
replication can occur. It also describes when journaling is started implicitly, as well
as the authority requirements necessary for user profiles that create the objects to
be journaled when they are created.
• “MIMIX commands for starting journaling” on page 289 identifies the MIMIX
commands available for starting journaling and describes the checking performed
by the commands. It also includes information for specifying journaling to the
configured journal.
• “Journaling for physical files” on page 291 includes procedures for displaying
journaling status, starting journaling, ending journaling, and verifying journaling for
physical files identified by data group file entries.
• “Journaling for IFS objects” on page 294 includes procedures for displaying
journaling status, starting journaling, ending journaling, and verifying journaling for
IFS objects replicated cooperatively (advanced journaling). IFS tracking entries
are used in these procedures.
• “Journaling for data areas and data queues” on page 298 includes procedures for
displaying journaling status, starting journaling, ending journaling, and verifying
journaling for data area and data queue objects replicated cooperatively
(advanced journaling). IFS tracking entries are used in these procedures.

286
What objects need to be journaled

What objects need to be journaled


A data group can be configured in a variety of ways that involve a user journal in the
replication of files, data areas, data queues and IFS objects. Journaling must be
started for any object to be replicated through a user journal or to be replicated by
cooperative processing between a user journal and the system journal.
Requirements for system journal replication - System journal replication
processes use a special journal, the security audit (QAUDJRN) journal. Events are
logged in this journal to create a security audit trail. When data group object entries,
IFS entries, and DLO entries are configured, each entry specifies an object auditing
value that determines the type of activity on the objects to be logged in the journal.
Object auditing is automatically set for all objects defined to a data group when the
data group is first started, or any time a change is made to the object entries, IFS
entries, or DLO entries for the data group. Because security auditing logs the object
changes in the system journal, no special action is need.
Requirements for user journal replication - User journal replication processes
require that journaling is started for the objects identified by data group file entries,
object tracking entries, and IFS tracking entries. Starting journaling ensures that
changes to the objects are recorded in the user journal, and are available for MIMIX to
replicate.
During initial configuration, the configuration checklists direct you when to start
journaling for objects identified by data group file entries, IFS tracking entries, and
object tracking entries. The MIMIX commands STRJRNFE, STRJRNIFSE, and
STRJRNOBJE simplify the process of starting journaling. If the journal the objects are
currently to is different than the journal defined in the data group definition, these
MIMIX commands provide a prompt to change journaling to the configured journal.
For more information about these commands, see “MIMIX commands for starting
journaling” on page 289.
Although MIMIX commands for starting journaling are preferred, you can also use
IBM commands (STRJRNPF, STRJRN, STRJRNOBJ) to start journaling if you have
the appropriate authority for starting journaling.
Requirements for implicit starting of journaling - Journaling can be automatically
started for newly created database files, data areas, data queues, or IFS objects
when certain requirements are met.
The user ID creating the new objects must have the required authority to start
journaling and the following requirements must be met:
• IFS objects - A new IFS object is automatically journaled if the directory in which it
is created is journaled as a result of a request that enabled journaling inheritance
for new objects. Typically, if MIMIX started journaling on the parent directory,
inheritance is enabled. If you manually start journaling on the parent directory
using the IBM command STRJRN, specify INHERIT(*YES) and SUBTREE
(*ALL). This will cause IFS objects created within the journaled directory to inherit
journal options and the journal state of the parent directory and also cause all
objects within the directory's subtree to be journaled.
• Database files created by SQL statements - A new file created by a CREATE

287
TABLE statement is automatically journaled if the library in which it is created
contains a journal named QSQJRN or if the library is journaled with appropriate
inherit rules.
• New *FILE, *DTAARA, *DTAQ objects - The default value (*DFT) for the Journal at
creation (JRNATCRT) parameter in the data group definition enables MIMIX to
automatically start journaling for physical files, data areas, and data queues when
they are created.
– On systems running IBM i 6.1 or higher releases, MIMIX uses the support
provided by the IBM i command Start Journal Library (STRJRNLIB).
Customers are advised not to re-create the QDFTJRN data area on systems
running IBM i 6.1 or higher.
When configuration requirements are met, MIMIX will start library journaling for
the appropriate libraries as well as enable automatic journaling for the configured
cooperatively processed object types. When journal at creation configuration
requirements are met, all new objects of that type are journaled, not just those
which are eligible for replication.
When the data group is started, MIMIX evaluates all data group object entries for
each object type. (Entries for *FILE objects are only evaluated when the data
group specifies COOPJRN(*USRJRN).) Entries properly configured to allow
cooperative processing of the object type determine whether MIMIX will enforce
library journaling. MIMIX uses the data group entry with the most specific match to
the object type and library that also specifies *ALL for its System 1 object (OBJ1)
and Attribute (OBJATR).
Note: MIMIX prevents library journaling from starting in the following libraries:
QSYS*, QRECOVERY, QRCY*, QUSR*, QSPL*, QRPL*, QRCL*, QRPL*,
QGPL, QTEMP and SYSIB*.
For example, if MIMIX finds only the following data group object entries for library
MYLIB, it would use the first entry when determining whether to enforce library
journaling because it is the most specific entry that also meets the OBJ1(*ALL)
and OBJATR(*ALL) requirements. The second entry is not considered in the
determination because its OBJ1 and OBJATR values do not meet these
requirements.
LIB1(MYLIB) OBJ1(*ALL) OBJTYPE(*FILE) OBJATR(*ALL) COOPDB(*YES)
PRCTYPE(*INCLD)
LIB1(MYLIB) OBJ1(MYAPP) OBJTYPE(*FILE) OBJATR(DSPF) COOPDB(*YES)
PRCTYPE(*INCLD)

Authority requirements for starting journaling


Normal MIMIX processes run under the MIMIXOWN user profile, which ships with
*ALLOBJ special authority. Therefore, it is not necessary for other users to account
for journaling authority requirements when using MIMIX commands (STRJRNFE,
STRJRNIFSE, STRJRNOBJE) to start journaling.
When the MIMIX journal managers are started, or when the Build Journaling
Environment (BLDJRNENV) command is used, MIMIX checks the public authority
(*PUBLIC) for the journal. If necessary, MIMIX changes public authority so the user ID
in use has the appropriate authority to start journaling.

288
MIMIX commands for starting journaling

Authority requirements must be met to enable the automatic journaling of newly


created objects and if you use IBM commands to start journaling instead of MIMIX
commands.
• If you create database files, data areas, or data queues for which you expect
automatic journaling at creation, the user ID creating these objects must have the
required authority to start journaling.
• If you use the IBM commands (STRJRNPF, STRJRN, STRJRNOBJ) to start
journaling, the user ID that performs the start journaling request must have the
appropriate authority requirements.
For journaling to be successfully started on an object, one of the following authority
requirements must be satisfied:
• The user profile of the user attempting to start journaling for an object must have
*ALLOBJ special authority.
• The user profile of the user attempting to start journaling for an object must have
explicit *ALL object authority for the journal to which the object is to be journaled.
• Public authority (*PUBLIC) must have *OBJALTER, *OBJMGT, and *OBJOPR
object authorities for the journal to which the object is to be journaled.

MIMIX commands for starting journaling


Before you use any of the MIMIX commands for starting journaling, the data group file
entries, IFS tracking entries, or object tracking entries associated with the command’s
object class must be loaded.
The MIMIX commands for starting journaling are:
• Start Journaling File Entries (STRJRNFE) - This command starts journaling for
files identified by data group file entries.
• Start Journaling IFS Entries (STRJRNIFSE) - This command starts journaling of
IFS objects configured for user journal replication. Data group IFS entries must be
configured and IFS tracking entries be loaded (LODDGIFSTE command) before
running the STRJRNIFSE command to start journaling.
• Start Journaling Obj Entries (STRJRNOBJE) - This command starts journaling of
data area and data queue objects configured for user journal replication. Data
group object entries must be configured and object tracking entries be loaded
(LODDGOBJTE command) before running the STRJRNOBJE command to start
journaling.
If you attempt to start journaling for files or objects that are already journaled, MIMIX
checks that the physical file, IFS object, data area, or data queue is journaled to the
journal configured in the data group definition. If the file or object is journaled to the
configured journal, the journaling status of the data group file entry, IFS tracking or
object tracking entry is changed to *YES. If the file or object is journaled using a
different journal than the configured journal, the journaling status is changed to
*DIFFJRN. If the attempt to start journaling fails for any other reason, the journaling
status is changed to *NO.

289
Forcing objects to use the configured journal
Journaled objects must use the journal defined in the data group definition in order for
replication to occur. Objects that are journaled to a journal that is different than the
journal defined in the data group definition can result in data integrity issues. MIMIX
identifies these objects with a journaling status of *DIFFJRN. A journal status of
*DIFFJRN should be investigated to determine the reason for using the different
journal. See “Resolving a problem for a journal status of *DIFFJRN” on page 143.
The MIMIX commands STRJRNFE, STRJRNIFSE, and STRJRNOBJE provide the
ability to specify the Force to configured journal (FORCE) prompt which determines
whether to end journaling for the selected objects that are currently journaled to a
different journal than the configured journal (*DIFFJRN), and then start journaling to
the configured journal.
To force journaled objects to use the journal configured in the data group definition,
specify *YES for the FORCE prompt in the MIMIX commands for starting journaling.
FORCE(*NO) is the command default for the STRJRNFE, STRJRNIFSE and
STRJRNOBJE commands when run from the native interface. See “MIMIX
commands for starting journaling” on page 289.

290
Journaling for physical files

Journaling for physical files


Data group file entries identify physical and logical files to be replicated. When data
group file entries are added to a configuration, they may have an initial status of
*ACTIVE and a journaling status of *NO even though the file may be journaled. For
the journaling status to be updated and accurately reflect journaling of the file, you
must verify journaling. In order for replication to occur, journaling must be started for
the files on the source system using the journal defined in the data group definition.
This topic includes procedures to display journaling status, and to start, end, or verify
journaling for physical files.

Displaying journaling status for physical files


Use this procedure to display journaling status for physical files identified by data
group file entries. Do the following:
1. From the MIMIX Intermediate Main Menu, type 1 and press Enter to access the
Work with Data Groups display.
2. On the Work with Data Groups display, type 17 (File entries) next to the data
group you want and press Enter.
3. The Work with DG File Entries display appears. The initial view shows the current
and requested status of the data group file entry. Press F10 (Journaled view).
At the right side of the display, the Journaled System 1 and System 2 columns
indicate whether the physical file associated with the file entry is journaled to the
configured journal (*YES), is not journaled (*NO), is journaled to a journal that is
different than the configured journal (*DIFFJRN), or whether journaling is not
permitted or not required for replication (*NA). When the column displays the
value *DIFFJRN, further investigation is required to determine the reason for
journaling to a different journal than the configured journal. See “Resolving a
problem for a journal status of *DIFFJRN” on page 143.
Note: Logical files will have a status of *NA. Data group file entries exist for
logical files only in data groups configured for MIMIX Dynamic Apply.

Starting journaling for physical files


Use this procedure to start journaling for physical files identified by data group file
entries. In order for replication to occur, journaling must be started for the file on the
source system using the journal configured in the data group definition.
This procedure invokes the Start Journaling File Entries (STRJRNFE) command. The
command can also be entered from a command line.
Do the following:
1. Access the journaled view of the Work with DG File Entries display as described
in “Displaying journaling status for physical files” on page 291.
2. From the Work with DG File Entries display, type a 9 (Start journaling) next to the
file entries you want. Then do one of the following:

291
• To start journaling using the command defaults, press Enter.
• To modify command defaults, press F4 (Prompt) then continue with the next
step.
3. The Start Journaling File Entries (STRJRNFE) display appears. The Data group
definition and the System 1 file identify your selection.
4. Specify the value you want for the Start journaling on system prompt. Press F4 to
see a list of valid values.
If journaling is started on the source system, a journal entry will be generated into
the user journal. As a result, replication processes will start journaling for these
objects on the target system if the data group definition specifies *YES for Journal
on target (JRNTGT).
5. If you want to use batch processing, specify *YES for the Submit to batch prompt.
6. Optional: If the file is journaled to a journal that is different than the configured
journal and you have determined that it is ok to force journaling to the configured
journal, press F10 (Additional parameters) to display the Force to configured
journal (FORCE) prompt.
Change the FORCE prompt to *YES to end journaling to a journal that is different
than the configured journal and then start journaling using the configured journal.
This value will also attempt to start journaling for objects not currently journaled.
Journaling will not be ended for objects already journaled to the configured
journal. For more information, see “Forcing objects to use the configured journal”
on page 290.
7. To start journaling for the physical file associated with the selected data group,
press Enter.
The system returns a message to confirm the operation was successful.

Ending journaling for physical files


Use this procedure to end journaling for files defined to a data group. Once journaling
for a file is ended, any changes to that file are not captured and are not replicated.
You may need to end journaling if a file no longer needs to be replicated or to correct
an error.
This procedure invokes the End Journaling File Entries (ENDJRNFE) command. The
command can also be entered from a command line.
To end journaling, do the following:
1. Access the journaled view of the Work with DG File Entries display as described
in “Displaying journaling status for physical files” on page 291.
2. From the Work with DG File Entries display, type a 10 (End journaling) next to the
file entry you want and do one of the following:
• To end journaling using command defaults, press Enter. Journaling is ended.
• To modify additional prompts for the command, press F4 (Prompt) and
continue with the next step.

292
Journaling for physical files

3. The End Journaling File Entries (ENDJRNFE) display appears. If you want to end
journaling for all files in the library, specify *ALL for the System 1 file prompt.
4. Specify the value you want for the End journaling on system prompt. Press F4 to
see a list of valid values.
If journaling is ended on the source system, a journal entry will be generated into
the user journal. As a result, replication processes will end journaling for these
objects on the target system if the data group definition specifies *YES for Journal
on target (JRNTGT).
5. If you want to use batch processing, specify *YES for the Submit to batch prompt.
6. To end journaling, press Enter.

Verifying journaling for physical files


Use this procedure to verify that files defined by a data group file entry are journaled
to the journal defined in the data group definition. When these conditions are met, the
journal status on the Work with DG File Entries display is set to *YES.
This procedure invokes the Verify Journaling File Entry (VFYJRNFE) command. The
command can also be entered from a command line.
To verify journaling for a physical file, do the following:
1. Access the journaled view of the Work with DG File Entries display as described
in “Displaying journaling status for physical files” on page 291.
2. From the Work with DG File Entries display, type a 11 (Verify journaling) next to
the file entry you want and do one of the following:
• To verify journaling using command defaults, press Enter.
• To modify additional parameters for the command, press F4 (Prompt) and
continue with the next step.
3. The Verify Journaling File Entry (VFYJRNFE) display appears. The Data group
definition prompt and the System 1 file prompt identify your selection.
4. Specify the value you want for the Verify journaling on system prompt. When
*DGDFN is specified, MIMIX considers whether the data group is configured for
journaling on the target system (JRNTGT) when determining where to verify
journaling.
5. If you want to use batch processing, specify *YES for the Submit to batch prompt.
6. To verify journaling, press Enter.
7. Press F5 (Refresh) to update and view the current journaling status.

293
Journaling for IFS objects
IFS tracking entries are loaded for a data group after they are configured for
replication through the user journal and the data group has been started. However,
loading IFS tracking entries does not automatically start journaling on the IFS objects
they identify. In order for replication to occur, journaling must be started on the source
system for the IFS objects identified by IFS tracking entries using the journal defined
in the data group definition.
This topic includes procedures to display journaling status, and to start, end, or verify
journaling for IFS objects identified for replication through the user journal.
You should be aware of the information in “Considerations for working with long IFS
path names” on page 330.

Displaying journaling status for IFS objects


Use this procedure to display journaling status for IFS objects identified by IFS
tracking entries. Do the following:
1. From the MIMIX Intermediate Main Menu, type 1 and press Enter to access the
Work with Data Groups display.
2. On the Work with Data Groups display, type 50 (IFS trk entries) next to the data
group you want and press Enter.
3. The Work with DG IFS Trk. Entries display appears. The initial view shows the
object type and status at the right of the display. Press F10 (Journaled view).
At the right side of the display, the Journaled System 1 and System 2 columns
indicate whether the IFS object identified by the tracking is journaled on each
system.
At the right side of the display, the Journaled System 1 and System 2 columns
indicate whether the objects associated with the file entry are journaled to the
configured journal (*YES), are not journaled (*NO), or are journaled to a journal
that is different than the configured journal (*DIFFJRN). When the column
displays the value *DIFFJRN, further investigation is required to determine the
reason for journaling to a different journal than the configured journal. See
“Resolving a problem for a journal status of *DIFFJRN” on page 143.

Starting journaling for IFS objects


Use this procedure to start journaling for IFS objects identified by IFS tracking entries.
This procedure invokes the Start Journaling IFS Entries (STRJRNIFSE) command.
The command can also be entered from a command line.
To start journaling for IFS objects, do the following:
1. If you have not already done so, load the IFS tracking entries for the data group.
For more information see the MIMIX Administrator Reference book.
2. Access the journaled view of the Work with DG IFS Trk. Entries display as

294
Journaling for IFS objects

described in “Displaying journaling status for IFS objects” on page 294.


3. From the Work with DG IFS Trk. Entries display, type a 9 (Start journaling) next to
the IFS tracking entries you want. Then do one of the following:
• To start journaling using the command defaults, press Enter.
• To modify the command defaults, press F4 (Prompt) and continue with the next
step.
4. The Start Journaling IFS Entries (STRJRNIFSE) display appears. The Data group
definition and IFS objects prompts identify the IFS object associated with the
tracking entry you selected. You cannot change the values shown for the IFS
objects prompts1.
5. Specify the value you want for the Start journaling on system prompt. Press F4 to
see a list of valid values.
If journaling is started on the source system, a journal entry will be generated into
the user journal. As a result, replication processes will start journaling for these
objects on the target system if the data group definition specifies *YES for Journal
on target (JRNTGT).
6. To use batch processing, specify *YES for the Submit to batch prompt and press
Enter. Additional prompts for Job description and Job name appear. Either accept
the default values or specify other values.
7. Optional: If the file is journaled to a journal that is different than the configured
journal and you have determined that it is ok to force journaling to the configured
journal, press F10 (Additional parameters) to display the Force to configured
journal (FORCE) prompt.
Change the FORCE prompt to *YES to end journaling to a journal that is different
than the configured journal and then start journaling using the configured journal.
This value will also attempt to start journaling for objects not currently journaled.
Journaling will not be ended for objects already journaled to the configured
journal. For more information, see “Forcing objects to use the configured journal”
on page 290.
8. The System 1 file identifier and System 2 file identifier prompts identify the file
identifier (FID) of the IFS object on each system. You cannot change the values2.
9. To start journaling on the IFS objects specified, press Enter.

Ending journaling for IFS objects


Use this procedure to end journaling for IFS objects identified by IFS tracking entries.
This procedure invokes the End Journaling IFS Entries (ENDJRNIFSE) command.
The command can also be entered from a command line.

1. When the command is invoked from a command line, you can change values specified for
the IFS objects prompts. Also, you can specify as many as 300 object selectors by using the
+ for more values prompt.
2. When the command is invoked from a command line, use F10 to see the FID prompts. Then
you can optionally specify the unique FID for the IFS object on either system. The FID values
can be used alone or in combination with the IFS object path name.

295
To end journaling for IFS objects, do the following:
1. Access the journaled view of the Work with DG IFS Trk. Entries display as
described in “Displaying journaling status for IFS objects” on page 294.
2. From the Work with DG IFS Trk. Entries display, type a 10 (End journaling) next to
the IFS tracking entries you want. Then do one of the following:
• To end journaling using the command defaults, press Enter.
• To modify the command defaults, press F4 (Prompt) and continue with the
next step.
3. The End Journaling IFS Entries (ENDJRNIFSE) display appears. The Data group
definition and IFS objects prompts identify the IFS object associated with the
tracking entry you selected. You cannot change the values shown for the IFS
objects prompts1.
4. Specify the value you want for the End journaling on system prompt. Press F4 to
see a list of valid values.
If journaling is ended on the source system, a journal entry will be generated into
the user journal. As a result, replication processes will end journaling for these
objects on the target system if the data group definition specifies *YES for Journal
on target (JRNTGT).
5. To use batch processing, specify *YES for the Submit to batch prompt and press
Enter. Additional prompts for Job description and Job name appear. Either accept
the default values or specify other values.
6. The System 1 file identifier and System 2 file identifier identify the file identifier
(FID) of the IFS object on each system. You cannot change the values shown2.
7. To end journaling on the IFS objects specified, press Enter.

Verifying journaling for IFS objects


Use this procedure to verify if an IFS object identified by an IFS tracking entry is
journaled correctly. This procedure invokes the Verify Journaling IFS Entries
(VFYJRNIFSE) command to determine whether the IFS object is journaled, whether it
is journaled to the journal defined in the data group definition, and whether it is
journaled with the attributes defined in the data group definition. The command can
also be entered from a command line.
To verify journaling for IFS objects, do the following:
1. Access the journaled view of the Work with DG IFS Trk. Entries display as
described in “Displaying journaling status for IFS objects” on page 294.
2. From the Work with DG IFS Trk. Entries display, type a 11 (Verify journaling) next
to the IFS tracking entries you want. Then do one of the following:
• To verify journaling using the command defaults, press Enter.
• To modify the command defaults, press F4 (Prompt) and continue with the next
step.
3. The Verify Journaling IFS Entries (VFYJRNIFSE) display appears. The Data

296
Journaling for IFS objects

group definition and IFS objects prompts identify the IFS object associated with
the tracking entry you selected. You cannot change the values shown for the IFS
objects prompts1.
4. Specify the value you want for the Verify journaling on system prompt. Press F4 to
see a list of valid values.
When *DGDFN is specified, MIMIX considers whether the data group is
configured for journaling on the target system (JRNTGT) and verifies journaling on
the appropriate systems as required.
5. To use batch processing, specify *YES for the Submit to batch prompt and press
Enter. Additional prompts for Job description and Job name appear. Either accept
the default values or specify other values.
6. The System 1 file identifier and System 2 file identifier identify the file identifier
(FID) of the IFS object on each system. You cannot change the values shown2.
7. To verify journaling on the IFS objects specified, press Enter.
“Using file identifiers (FIDs) for IFS objects” on page 338.

297
Journaling for data areas and data queues
Object tracking entries are loaded for a data group after they are configured for
replication through the user journal and the data group has been started. However,
loading object tracking entries does not automatically start journaling on the objects
they identify. In order for replication to occur, journaling must be started on the source
system for objects identified by tracking entries that use the journal defined in the data
group definition.
This topic includes procedures to display journaling status, and to start, end, or verify
journaling for data areas and data queues identified for replication through the user
journal.

Displaying journaling status for data areas and data queues


To check journaling status for data areas and data queues identified by object tracking
entries. Do the following:
1. From the MIMIX Intermediate Main Menu, type 1 and press Enter to access the
Work with Data Groups display.
2. On the Work with Data Groups display, type 52 (Obj trk entries) next to the data
group you want and press Enter.
3. The Work with DG Obj. Trk. Entries display appears. The initial view shows the
object type and status at the right of the display. Press F10 (Journaled view).
At the right side of the display, the Journaled System 1 and System 2 columns
indicate whether the objects associated with the tracking entry are journaled to the
configured journal (*YES), are not journaled (*NO), or are journaled to a journal
that is different than the configured journal (*DIFFJRN). When the column
displays the value *DIFFJRN, further investigation is required to determine the
reason for journaling to a different journal than the configured journal. See
“Resolving a problem for a journal status of *DIFFJRN” on page 143.

Starting journaling for data areas and data queues


Use this procedure to start journaling for data areas and data queues identified by
object tracking entries.
This procedure invokes the Start Journaling Obj Entries (STRJRNOBJE) command.
The command can also be entered from a command line.
To start journaling for data areas and data queues, do the following:
1. If you have not already done so, load the object tracking entries for the data
group. For more information see the MIMIX Administrator Reference book.
2. Access the journaled view of the Work with DG Obj. Trk. Entries display as
described in “Displaying journaling status for data areas and data queues” on
page 298.
3. From the Work with DG Obj. Trk. Entries display, type a 9 (Start journaling) next to
the object tracking entries you want. Then do one of the following:

298
Journaling for data areas and data queues

• To start journaling using the command defaults, press Enter.


• To modify the command defaults, press F4 (Prompt) and continue with the next
step.
4. The Start Journaling Obj Entries (STRJRNOBJE) display appears. The Data
group definition and Objects prompts identify the object associated with the
tracking entry you selected. Although you can change the values shown for these
prompts, it is not recommended unless the command was invoked from a
command line.
5. Specify the value you want for the Start journaling on system prompt. Press F4 to
see a list of valid values.
If journaling is started on the source system, a journal entry will be generated into
the user journal. As a result, replication processes will start journaling for these
objects on the target system if the data group definition specifies *YES for Journal
on target (JRNTGT).
6. To use batch processing, specify *YES for the Submit to batch prompt and press
Enter. Additional prompts for Job description and Job name appear. Either accept
the default values or specify other values.
7. Optional: If the file is journaled to a journal that is different than the configured
journal and you have determined that it is ok to force journaling to the configured
journal, press F10 (Additional parameters) to display the Force to configured
journal (FORCE) prompt.
Change the FORCE prompt to *YES to end journaling to a journal that is different
than the configured journal and then start journaling using the configured journal.
This value will also attempt to start journaling for objects not currently journaled.
Journaling will not be ended for objects already journaled to the configured
journal. For more information, see “Forcing objects to use the configured journal”
on page 290.
8. To start journaling on the objects specified, press Enter.

Ending journaling for data areas and data queues


Use this procedure to end journaling for data areas and data queues identified by
object tracking entries.
This procedure invokes the End Journaling Obj Entries (ENDJRNOBJE) command.
The command can also be entered from a command line.
To end journaling for data areas and data queues, do the following:
1. Access the journaled view of the Work with DG Obj. Trk. Entries display as
described in “Displaying journaling status for data areas and data queues” on
page 298.
2. From the Work with DG Obj. Trk. Entries display, type a 10 (End journaling) next
to the object tracking entries you want. Then do one of the following:
• To verify journaling using the command defaults, press Enter.

299
• To modify the command defaults, press F4 (Prompt) and continue with the next
step.
3. The End Journaling Obj Entries (ENDJRNOBJE) display appears. The Data group
definition and IFS objects prompts identify the object associated with the tracking
entry you selected. Although you can change the values shown for these prompts,
it is not recommended unless the command was invoked from a command line.
4. Specify the value you want for the End journaling on system prompt. Press F4 to
see a list of valid values.
If journaling is ended on the source system, a journal entry will be generated into
the user journal. As a result, replication processes will end journaling for these
objects on the target system if the data group definition specifies *YES for Journal
on target (JRNTGT).
5. To use batch processing, specify *YES for the Submit to batch prompt and press
Enter. Additional prompts for Job description and Job name appear. Either accept
the default values or specify other values.
6. To end journaling on the objects specified, press Enter.

Verifying journaling for data areas and data queues


Use this procedure to verify if an object identified by an object tracking entry is
journaled correctly. This procedure invokes the Verify Journaling Obj Entries
(VFYJRNOBJE) command to determine whether the object is journaled, whether it is
journaled to the journal defined in the data group definition, and whether it is journaled
with the attributes defined in the data group definition. The command can also be
entered from a command line.
To verify journaling for objects, do the following:
1. Access the journaled view of the Work with DG Obj. Trk. Entries display as
described in “Displaying journaling status for data areas and data queues” on
page 298.
2. From the Work with DG Obj. Trk. Entries display, type a 11 (Verify journaling) next
to the object tracking entries you want. Then do one of the following:
• To verify journaling using the command defaults, press Enter.
• To modify the command defaults, press F4 (Prompt) and continue with the next
step.
3. The Verify Journaling Obj Entries (VFYJRNOBJE) display appears. The Data
group definition and Objects prompts identify the object associated with the
tracking entry you selected. Although you can change the values shown for these
prompts, it is not recommended unless the command was invoked from a
command line.
4. Specify the value you want for the Verify journaling on system prompt. Press F4 to
see a list of valid values.
When *DGDFN is specified, MIMIX considers whether the data group is
configured for journaling on the target system (JRNTGT) and verifies journaling on
the appropriate systems as required.

300
Journaling for data areas and data queues

5. To use batch processing, specify *YES for the Submit to batch prompt and press
Enter. Additional prompts for Job description and Job name appear. Either accept
the default values or specify other values.
6. To verify journaling on the objects specified, press Enter.

301
Switching

CHAPTER 14 Switching

Switching is when you temporarily change the roles of the systems. The original
source system (production) becomes the temporary target system and the original
target system (backup) becomes the temporary source system. When the scenario
that required you to switch directions is resolved, you typically switch again to return
the systems to their original roles.
This chapter provides information and procedures to support switching. The following
topics are included:
• “About switching” on page 303 provides information about switching with MIMIX
including best practice and reasons why a switch should be performed. Subtopics
describe:
– What is a planned switch and requirements for a planned switch
– What is an unplanned switch and actions to be completed after the failed
source system is recovered
– The role of procedures for switching environments that use application groups
– The role of MIMIX Model Switch Framework for switching environments that do
not use application groups
• “Switching an application group” on page 308 describes how to run a procedure to
switch an application group.
• “Switching a data group-only environment” on page 309. describes how to switch
from the native user interface.
• “Determining when the last switch was performed” on page 311 describes how to
check the Last switch field which indicates the switch compliance status and
provides the date when the last switch was performed.
• “Performing a data group switch” on page 312 describes how to switch a single
data group using the SWTDG command.
• “Switch Data Group (SWTDG) command” on page 314 provides background
information about the SWTDG command, which is used in all switch interfaces.
In environments licensed for MIMIX DR, only virtual switching1 is supported.
Virtual switching does not reverse the roles of the systems and is not covered in this
chapter. For detailed information about virtual switching, see “Support for virtual
switching” on page 316.

1. Virtual switching (virtual failover) is supported in MIMIX DR installations running version


8.1.02.00 or higher.

302
About switching

About switching
Replication environments rarely remain static. Therefore, best practice is to perform
regular switches to ensure that you are prepared should you need to perform one
during an emergency.
MIMIX supports two methods for switching the direction in which replication occurs for
a data group. These methods are known as a planned switch and an unplanned
switch.
You may need to perform a switch for any of the following reasons:
• The production system becomes unavailable due to an unplanned outage. A
switch in this scenario is unplanned.
• You need to perform hardware or software maintenance on the production
system. Typically, you can schedule this in advance so the switch is planned.
• You need to test your recovery plan. This activity is also a planned switch. Virtual
switching can also be used to test applications prior to performing a planned
switch test of your recovery plan. Virtual switching is described in “Support for
virtual switching” on page 316.
Historically, the concept of switching consists of three phases: switch to the backup
system, synchronize the systems when the production is ready to use, and switch
back to the production system. This round-trip view of switching assumes your goal is
to return to your original production system as quickly as possible. However, some
customers may have an extended time pass between phase one and the other
phases, or may even view a switch as a one-way trip. MIMIX supports both
conceptual views of switching.
Switching data groups is only a part of performing a switch. MIMIX provides robust
support for customizing switching activity include all the needs of your environment.
Best practice for switching includes performing regular switches. Best practice also
includes performing all audits with the audit level set at level 30 immediately prior to a
planned switch to the backup system and before switching back to the production
system. Best practice is to use Vision Solutions Portal to perform switch operations. If
you use the native user interface, best practice is:
• For application groups, use option 4 (Switch all application groups) from the
MIMIX Basic Main Menu.
• For data group-only environments, use option 5 (start or complete switch using
Switch Asst.) from the MIMIX Basic Main Menu.

Planned switch - overview


You can start a planned switch from either system. In a planned switch, MIMIX
initiates a controlled shutdown of the data group. Both systems and the
communications between them must be active.
Preparing for a planned switch: Before you start a planned switch of a data group,
you should ensure that the following actions have been completed. Your enterprise
may have additional requirements.

303
Switching

• Ensure configuration changes are deployed. You can have MIMIX automatically
deploy configuration changes by ending and starting data group replication, or
you can manually deploy data group configuration within MIMIX. For more
information, see the MIMIX Administrator Reference book.
• After you have ensured configuration changes are deployed, perform a full set of
audits with the audit level policy set to level 30.
• Shut down any applications that use database files or objects defined to the data
group. If any users or other non-MIMIX processes remain active while the switch
is being performed, the data can become not synchronized between systems and
orphaned data may result.
• Ensure that there are no jobs other than MIMIX currently active on the source
system. This may require ending all interactive and batch subsystems other than
MIMIX and ending communications.
• Users should be prevented from accessing either system until after the switch is
complete and the data group is restarted.
• If you use user journal replication processes, you should address any files, IFS
tracking entries, or object tracking entries in error for your critical database files. If
you use system journal replication processes, you should address any object
errors.
You are not required to run journal analysis after a planned switch. MIMIX retains
information about where activity ended so that when you restart the data group, it is
started at the correct point.
When the data group is started, the temporary target system (the production system)
is now being updated with user changes that are being replicated from the temporary
source system (the backup system). Do not allow users onto the production system
until after the production system is caught up with these transactions and you run the
switch process again to revert to the normal roles.

Unplanned switch - overview


In an unplanned switch, the source system is assumed to be unavailable. An
unplanned switch is generally required when the source system fails and, in order to
continue normal operations, you must switch users to a backup system. (Typically
MIMIX is configured so that the target for replication is your backup system.)
You must run an unplanned switch from the target system. MIMIX performs a
controlled shutdown of replication processes on the target system. The controlled
shutdown allows all apply processing to catch up before the apply processes are
ended.
In an unplanned switch of a data group that uses remote journaling, the default
behavior is to end the RJ link.
Following an unplanned switch: Once the failed source system is recovered, the
following actions should be completed:
• You should perform journal analysis on that system before restarting the data
group or user applications. Journal analysis helps identify any possible loss of
data that may have occurred when the source system failed. Journal analysis

304
About switching

relies on status information on the source system about the last entry that was
applied. This information will be cleared when the data group is restarted.
• Communication between the systems must be active before you restart the data
group. The switch process is complete when you restart the data group. When
the data group is restarted, MIMIX notifies the source system that it is now the
temporary target system.
• New transactions are created on the temporary source system (the backup
system) while the production system (the temporary target system) is unavailable
for replication. After you have completed journal analysis, you can send these
new transactions to the production system to synchronize the databases. Once
the databases are synchronized, you must run the switch process again to revert
to the normal roles before allowing users onto the production system.
When the data group is started after a switch, any pending transactions are cleared.
The journal receiver is already changed by the switch process and the new journal
receiver and first sequence number are used.

Switching application group environments with procedures


Application groups can only be switched using procedures that are initiated by an
application group command for a specific operation. Procedures are a highly
customizable means of performing operations for application groups. Each
application group has a set of default procedures to perform operations such as
starting, ending, or switching replication, performing pre-checks prior to switching,
and performing virtual switch testing. Each operation is performed by a procedure that
consists of a sequence of steps and multiple jobs. Each step calls a predetermined
step program to perform a specific sub-task of the larger operation.
The following paragraphs describe the behavior of the switch (SWTAG) command for
application groups of type *NONCLU1.
What is the scope of the request? The following parameters identify the scope of
the requested operation:
Application group definition (AGDFN) - Specifies the requested application group.
You can either specify a name or the value *ALL.
Resource groups (TYPE) - Specifies the types of resource groups to be
processed for the requested application group.
Data resource group entry (DTARSCGRP) - Specifies the data resource groups to
include in the request. The default is *ALL or you can specify a name. This
parameter is ignored when TYPE is *ALL or *APP.
What is the requested switch behavior? The following parameters on the SWTAG
command define the expected behavior of the requested switch operation:
Switch type (SWTTYP) - This specifies the reason the application group is being
switched. The procedure called to perform the switch and the actions performed

1. Procedures are used for switching *CLU and *NONCLU application groups. Application
groups of type *NONCLU do not participate in a cluster controlled by the IBM i operating sys-
tem.

305
Switching

during the switch differ based on whether the current primary node (data source)
is available at the start of the switch procedure. The default value, *PLANNED,
indicates that the primary node is still available and the switch is being performed
for normal business processes (such as to perform maintenance on the current
source system or as part of a standard switch procedure). The value
*UNPLANNED indicates that the switch is an unplanned activity and the data
source system may not be available. The value *VIRTUAL indicates that the
application group environment will be modified to allow application testing to be
performed on a backup node in a process known as a virtual switch.
Node roles (ROLE) - This specifies which set of node roles will be used during the
switch. For a planned or unplanned switch, the specified value determines the
node that becomes the new primary node as a result of the switch. The default
value *CURRENT uses the current order of node roles. If the application group
participates in a cluster, the current roles defined within the CRGs will be used. If
*CONFIG is specified, the configured primary node will become the new primary
node and the new role of other nodes in the recovery domain will be determined
from their current roles. If you specify a name of a node within the recovery
domain for the application group, the node will be made the new primary node and
the new role of other nodes in the recovery domain will be determined from their
current roles. For a virtual switch, the value of this parameter determines the node
to be used for testing applications and does not cause changes to recovery
domain roles. In a virtual switch, the first backup node in the recovery domain is
used unless you specify the name of a backup node.
What procedure will be used? The following parameters identify the procedure to
use and its starting point:
Begin at step (STEP) - Specifies where the request will start within the specified
procedure. This parameter is described in detail below.
Procedure (PROC) - Specifies the name of the procedure to run to perform the
requested operation when starting from its first step. The value *DFT will use the
procedure designated as the default for the application group. The value
*LASTRUN uses the same procedure used for the previous run of the command.
You can also specify the name of a procedure that is valid the specified
application group and type of request.

Where should the procedure begin? The value specified for the Begin at step
(STEP) parameter on the request to run the procedure determines the step at which
the procedure will start. The status of the last run of the procedure determines which
values are valid.
The default value, *FIRST, will start the specified procedure at its first step. This value
can be used when the procedure has never been run, when its previous run
completed (*COMPLETED or *COMPERR), or when a user acknowledged the status
of its previous run which failed, was canceled, or completed with errors
(*ACKFAILED, *ACKCANCEL, or *ACKERR respectively).
Other values are for resolving problems with a failed or canceled procedure. When a
procedure fails or is canceled, subsequent attempts to run the same procedure will
fail until user action is taken. You will need to determine the best course of action for

306
About switching

your environment based on the implications of the canceled or failed steps and any
steps which completed.
The value *RESUME will start the last run of the procedure beginning with the step at
which it failed, the step that was canceled in response to an error, or the step
following where the procedure was canceled. The value *RESUME may be
appropriate after you have investigated and resolved the problem which caused the
procedure to end. Optionally, if the problem cannot be resolved and you want to
resume the procedure anyway, you can override the attributes of a step before
resuming the procedure.
The value *OVERRIDE will override the status of all runs of the specified procedure
that did not complete. The *FAILED or *CANCELED status of these procedures are
changed to acknowledged (*ACKFAILED or *ACKCANCEL) and a new run of the
procedure begins at the first step.
For more information about starting a procedure with the step at which it failed, see
“Resuming a procedure” on page 112.
For more information about customizing procedures, see the MIMIX Administrator
Reference book.

Switching data group environments with MIMIX Model Switch Framework


Note: MIMIX Model Switch Framework does not support switching application
groups. Only data groups that are not associated with application groups
should be switched with MIMIX Model Switch Framework.
MIMIX provides a customized implementation of MIMIX Model Switch Framework to
perform a switch. MIMIX Model Switch Framework is ideally suited for customizing a
switching solution that detects the need for an unplanned switch, switches the
direction of data group replication, and switches users to the backup system.
Typically, if you have a Runbook, it will direct you when to use your MIMIX Model
Switch Framework implementation for both planned and unplanned switches.
The MIMIX Model Switch Framework calls the Switch Data Group (SWTDG)
command. The SWTDG command only switches the direction in which replication
occurs for a single data group; it does not switch users or any other facets of your
normal operating environment to the backup system. However, MIMIX Model Switch
Framework can be configured to address these additional facets of your environment
for multiple data groups. If you choose to use the SWTDG command either by
invoking it from a command line or by using the options for switching on the Work with
Data Groups display, you must take action to switch users to the backup system and
address other requirements for operating there.
The switching option from the MIMIX Basic Main menu are implementations of MIMIX
Model Switch Framework. The implementation is identified within policies.
Instructions for switching using MIMIX Model Switch Framework are described in
“Switching a data group-only environment” on page 309.
For additional information see the chapter “Using the MIMIX Model Switch
Framework” in the Using MIMIX Monitor book.

307
Switching an application group
These instructions can be used to perform a planned or unplanned switch of an
application group. An application group can have only one procedure that performs a
start, end, or switch operation running at a time.
Before you begin, review topic “About switching” on page 303 for details about
preparing for a planned switch, actions following an unplanned switch, and
implications of the parameters available on the Switch Application Group (SWTAG)
command.
To switch an application group, do the following:
1. From the Work with Application Groups display, type 15 (Switch) next to the
application group you want and press Enter.
The Switch Application Group (SWTAG) display appears.
2. Verify that the values you want are specified for Resource groups and Data
resource group entry.
3. At the Switch type prompt, specify the type of switch to perform.
• For a planned switch, specify *PLANNED.
• For an unplanned switch, specify *UNPLANNED.
4. Verify that the default value *CURRENT for Node roles prompt is valid for the
switch you need to perform. If necessary, specify a different value.
5. If you are starting the procedure after addressing problems with the previous
switch request, specify the value you want for Begin at step. Be certain that you
understand the effect the value you specify will have on your environment.
6. Press Enter.
7. The Procedure prompt appears. Do one of the following:
• To use the default switch procedure for the specified switch type, press Enter.
• To use a different switch procedure for the application group, specify its name.
Then press Enter.
8. A switch confirmation panel appears. To perform the switch, press F16.

308
Switching a data group-only environment

Switching a data group-only environment


In environments that do not use application groups, option 5 (Start or complete switch
using Switch Asst.) on the MIMIX Basic Main Menu is designed to simplify switching
by using a default MIMIX Model Switch Framework implementation. When you use
this option, MIMIX keeps track of which phase of the switch process you are in. You
will see a confirmation display that is appropriate for each phase. Each phase will
prompt the Run Switch Framework command (RUNSWTFWK) with your default
switch framework and appropriate values for the phase.
To change the default switch framework to a different implementation, see “Policies
for switching with model switch framework” on page 48.

Switching to the backup system


This procedure switches operations to the backup system.
Before using this procedure, consult your runbook for any additional procedures that
must be performed when switching to the backup system.
1. If this is a planned switch, Vision Solutions strongly recommends that you perform
a full set of audits with the audit level policy set to level 30. Running the #FILDTA
audit at this audit level checks 100 percent of file member data for the data group
for synchronization between source and target systems.
2. Shut down all active applications that are reading or updating replicated objects
from the production and backup systems.
Do the following from the backup system:
3. Ensure that all transactions have been applied to the backup system by doing the
following:
a. Select option 6 (Work with data groups) from the MIMIX Basic Main Menu and
press Enter.
b. For each data group, select option 8 (Display status) and ensure that the
Unprocessed entry counts for both database and object apply have no values.
4. From the MIMIX Basic Main Menu, select option 5 (Start or complete switch using
Switch Asst.).
5. You will see the Confirm Switch to Backup confirmation display. Press F16 to
confirm your choice to switch MIMIX and specify switching options.
6. The Run Switch Framework (RUNSWTFWK) command appears. The default
Switch framework and the value *BCKUP for the Switch framework process are
preselected and cannot be changed. Do the following:
a. You must specify the type of switch to perform, *PLANNED or *UNPLANNED,
at the Switch type prompt.
b. You can change values for other parameters as needed.
c. To start the switch, press Enter.
7. Consult your runbook to determine if any additional steps are needed.

309
After you complete this phase of the switch you must wait until the original production
system is available again. Then perform the steps in “Synchronizing data and starting
MIMIX on the original production system” on page 310.

Synchronizing data and starting MIMIX on the original production system


This procedure synchronizes data and starts replication from the backup system to
the original production system. Synchronizing the data ensures that the data on both
systems is equivalent before replication is started.
Before using this procedure, consult your runbook for any additional procedures that
must be performed when synchronizing and starting replication from the backup
system to the original production system
Do the following from the backup system:
1. Ensure the original production system is available again.
2. From the MIMIX Basic Main Menu, select option 5 (Start or complete switch using
Switch Asst.).
3. You will see the Confirm Synchronize and Start confirmation display. Press F16 to
confirm your choice and specify switching options.
4. The Run Switch Framework (RUNSWTFWK) command appears. The default
Switch framework and the value *SYNC for the Switch framework process are
preselected and cannot be changed. Do the following:
a. Optionally, you can change the value of the Set object auditing level prompt.
b. To synchronize and start, press Enter.
5. Once replication has caught up, Vision Solutions strongly recommends that you
perform a full set of audits with the audit level policy set to level 30. Running the
#FILDTA audit at this audit level checks 100 percent of file member data for the
data group for synchronization between source and target systems.
6. Consult your runbook to determine if any additional steps are needed.
When you are ready to switch back to the original production system, use “Switching
to the production system” on page 310.

Switching to the production system


This procedure returns operations to the original production system.
Before using this procedure, consult your runbook for any additional procedures that
must be performed when switching to the production system
1. Shut down all active applications that are reading or updating replicated objects
from the production and backup systems.
Do the following from the original production system:
2. Ensure that all transactions have been applied by doing the following:
a. Select option 6 (Work with data groups) from the MIMIX Basic Main Menu and
press Enter.

310
Determining when the last switch was performed

b. For each data group, select option 8 (Display status) and ensure that the
Unprocessed entry counts for both database and object apply have no values.
3. From the MIMIX Basic Main Menu, select option 5 (Start or complete switch using
Switch Asst.).
4. You will see the Confirm Switch to Production confirmation display. Press F16 to
confirm your choice to switch MIMIX and specify switching options.
5. The Run Switch Framework (RUNSWTFWK) command appears. The default
Switch framework and the value *PROD for the Switch framework process are
preselected and cannot be changed. Do the following:
a. You can change values for other parameters as needed.
b. To start the switch, press Enter.
6. Consult your runbook to determine if any additional steps are needed.

Determining when the last switch was performed


Replication environments rarely remain static. Therefore, best practice is to perform
regular switches to ensure that you are prepared should you need to perform one
during an emergency.
The Last switch field indicates compliance with best practices. The status of the field
is highlighted to indicate the following:
Yellow - The number of days since the last switch is at the limit of what is
considered to be best practice. This threshold is determined by the Switch
warning threshold policy.
Red - The number of days since the last switch is beyond what is considered to be
best practice. This threshold is determined by the Switch action threshold policy.

311
Performing a data group switch
Performing a data group switch changes the direction of replication for a data group
through the Switch Data Group (SWTDG) command. Only replication for the selected
data group is switched. You may want to perform a data group switch if you are
having problems with an application that only affects a specific data group or if you
need to manually load balance because of heavily used applications.
Note: You cannot switch a disabled data group. For more information, see
“Disabling and enabling data groups” on page 334.
To perform a data group switch, do the following:
1. If you will be performing a planned switch, do the following:
a. Shut down any applications that have database file or objects defined to the
data group.
b. Ensure that you have addressed any critical database files that are held due to
error or held for other reasons.
c. Ensure there are no pending object activity entries by entering: WRKDGACTE
STATUS(*ACTIVE)
2. From the Work with Data Groups display, type the option for the type of switch you
want next to the data group you want to switch and press Enter.
• Use option 15 for a planned switch
• Use option 16 for an unplanned switch
3. Some of the parameter values that you may want to consider when the Switch
Data Group display appears are:
• If you specified Switch type of *PLANNED and have specified a number for the
Wait time (seconds) parameter, you can specify a value for the Timeout Option
parameter to specify what action you want the SWTDG command to perform if
the time specified in the Wait time (seconds) parameter is exceeded. When
you are performing a planned switch you may want to specify the number of
seconds to wait before all the active data group processes end. If you specify
*NOMAX the switch process will wait until all data group processes are ended.
This could delay the switch process.
• You can use the Conditions that end switch parameter to specify the types of
errors that you want to end the switch operation. For a planned switch, the
default value, *DFT, checks all possible values except *RCY and converts any
new or in-progress recoveries found to replication errors. For an unplanned
switch, *DFT will prevent the switch only when database apply backlogs exist.
• Verify that the value for the Start journaling on new source prompt is what you
want. If necessary, change the value.
4. After the confirmation screen, press F16 to continue.
5. Press Enter. Messages appear indicating the status of the switch request. When
you see a message indicating that the switch is complete, users can begin
processing as usual on the temporary source system.

312
Performing a data group switch

6. If you performed an unplanned switch, perform journal analysis on the original


source system as soon as it is available, to determine if any transactions were
missed. Use topic “Performing journal analysis” on page 356.
7. Start the data group, clearing pending entries, using the procedure in “Starting
selected data group processes” on page 228. This starts replication in the new
temporary direction.

313
Switch Data Group (SWTDG) command
The Switch Data Group (SWTDG) command provides the following parameters to
control how you want your switch operation handled:
• The Wait time (seconds) parameter (WAIT) is used to specify the number of
seconds to wait for all of the active data group processes to end. The function of
the default value *DFT is different for planned switches than it is for unplanned
switches. For a planned switch, the value *DFT is equivalent to the value
*NOMAX. For an unplanned switch, the value *DFT is set to wait 300 seconds (5
minutes) for all of the active data group processes to end.
• If you specify a value for the WAIT parameter you can use the Timeout option
parameter (TIMOUTOPT) to specify what action to take when the wait time you
specified is reached. The function of the default value *DFT is different for planned
switches than it is for unplanned switches. For a planned switch, the value *DFT is
equivalent to the value *QUIT. When the value specified for the WAIT parameter is
reached, the current process quits and returns control to the caller. For an
unplanned switch, the value *DFT is equivalent to the value *NOTIFY. When the
value specified for the WAIT parameter is reached, an inquiry message is sent to
notify the operator of a possible error condition.
• The Conditions that end switch (ENDSWT) parameter is used to specify which
conditions should end the switch process. The function of the default value *DFT
is different for planned switches than it is for unplanned switches.
– For a planned switch, the value *DFT checks all conditions except *RCY. If any
new or in-progress recoveries exist when *DFT is used, they are converted to
held file entries, held tracking entries, or failed object activity entries before
ending the process. The value *ALL provides the most comprehensive
checking for conditions that are not compatible with best practices for
switching. Additionally, the value *ALL ensures that your programs will
automatically include any future ENDSWT parameter values that may be
added to maintain a conservative approach to the switching operation.
– For an unplanned switch, the value *DFT ends the process if there are any
backlogs for the database apply process. However, backlogs on other user
journal processes are not checked and switch processing is not ended even
though conditions may exist which are not compatible with best practices for
switching and may result in the loss of data.
• The Start journaling on new source (STRJRNSRC) parameter is used to specify
whether you want to start journaling for the data group on the new source system.
• The End journaling on new target (ENDJRNTGT) parameter is used to specify
whether you want to end journaling of the data group on the new target system.
• The End remote journaling (ENDRJLNK) parameter is used in a planned switch of
a data group that uses remote journaling. This parameter specifies whether you
want to end remote journaling for the data group. The default behavior is to leave
the RJ link running. You need to consider whether to keep the RJ link active after
a planned switch of a data group. For more information, see “Before ending
replication” on page 232.

314
Switch Data Group (SWTDG) command

• The Change user journal receiver (CHGUSRRCV) parameter is used to specify


whether or not you want MIMIX to create and attach a new user (database)
journal receiver during the switch operation. If you have applications that are
dependent on the receiver name for recovery purposes, It is recommended that
you choose CHGUSRRCV(*NO) to prevent a new journal receiver from being
created during a data group switch.
• The Change system journal receiver (CHGSYSRCV) parameter is used to specify
whether or not you want MIMIX to create and attach a new journal receiver to the
system (audit) journal (QAUDJRN) during the switch operation. If you have
applications that are dependent on the receiver name for recovery purposes, it is
recommended that you choose CHGSYSRCV(*NO) to prevent a new journal
receiver from being created during a data group switch.
• The End if database errors (ENDDBERR) parameter has been obsoleted by the
Conditions that end switch (ENDSWT) parameter. Previously, the ENDDBERR
parameter was used to specify whether to switch the data group when data
replication errors exist. Use the ENDSWT parameter and specify *DBERR to
produce the equivalent of ENDDBERR(*YES), or *NONE to produce the
equivalent of ENDDBERR(*NO).
• The Confirm (CONFIRM) parameter is used to specify if a confirmation panel is
displayed. The default is *NO (the confirmation panel is not displayed). Note that
options for switching on the Work with Data Groups display call the SWTDG
command with *YES specified so that the confirmation panel is automatically
displayed and the user must press F16 to continue.

315
Support for virtual switching

CHAPTER 15 Support for virtual switching

This chapter describes what is virtual switching, the purpose and activity that occur in
each phase of the virtual switch procedure, application testing considerations, as well
as requirements and considerations for your MIMIX environment.
• “What is virtual switching?” on page 316
• “MIMIX virtual switch overview” on page 317
• “MIMIX environment considerations for virtual switching” on page 319
• “Testing considerations for virtual switching” on page 321
• “Performing a precheck for virtual switching” on page 323
• “Performing a virtual switch” on page 323

What is virtual switching?


Virtual switching is the process of testing that your production environment
applications can access the objects they need and run on a backup node in a MIMIX
environment without shutting down the applications or stopping MIMIX.
The purpose and scope of a virtual switch test differ from a planned switch test in the
following ways:
• A virtual switch test does not switch anything. MIMIX support for virtual
switching sets up your MIMIX environment to allow a designated backup node to
be used for testing while the production environment and MIMIX remain active.
Only apply processes and access path maintenance (if configured) for data
groups that currently use the test node as their backup node are ended. Other
replication processes remain active, so production node transactions are
processed, sent to the target node, and remain queued for apply processes until
testing is complete. While you are testing that the applications can run on the
designated test node, MIMIX tracks any changes made to replicated objects on
that node. After you indicate that testing is complete, MIMIX reverts those
changes, restarts the apply processes, and returns the node to normal operation.
• A planned switch test is a complete planned switch for the purpose of verifying
that all aspects of the switch perform as expected. A planned switch test shuts
down the production environment and MIMIX replication, switches the production
environment to the backup node (which becomes the new primary node), and
restarts MIMIX replication with the new primary node as its source. A second
planned switch is required to return the production environment to the original
primary node.
Virtual switching is an excellent way to identify and correct problems that could
otherwise inhibit or prevent applications from starting on the new primary node during
a planned or unplanned switch. However, a virtual switch does not replace the need

316
MIMIX virtual switch overview

to perform a planned switch test. Some applications are not suitable for virtual
switching. Also, a virtual switch cannot test some aspects of your environment, such
as IP impersonation, and does not allow you to validate that the MIMIX configuration
operates as expected from the test node.

MIMIX virtual switch overview


A virtual switch procedure provides the framework in which you can perform
application testing on a backup node in a MIMIX environment. The SWTVRT
procedure is created by default as one of the procedures created automatically when
a new application group is created.1

Phases of the virtual switch procedure


The SWTVRT procedure runs multiple steps to perform these major functions, which
can be considered as virtual switch phases:
• Precheck. The first set of steps check the configuration to ensure that no
restrictions exist, that requirements are met, and check for conditions that would
otherwise end the virtual switch. These operations can also be performed in
advance by using the virtual switch precheck procedure, PRECHKVRT.
• Setup. These steps change the MIMIX environment to prepare for testing and
affect objects on the target node of data groups within the application group when
that node is the designated test node for the virtual switch. These operations
begin with step MXSETVRTS, which changes the replication status to Virtual
switch starting and clears any retained information about any objects changed in
the previous virtual switch of the application group. For the affected data groups,
active audits are ended and prevented from running, the database apply and
object apply processes are ended, and access path maintenance is ended (if
configured). Subsequent replicated transactions are queued for apply processes.
Any foreign key constraints previously disabled by MIMIX are enabled on the
target objects. Status of target journal inspection jobs on the test node are
checked and are set to not issue notifications while testing is in progress for the
affected data groups. For files on the target node that are eligible for replication,
journaling is changed, if necessary, to use both before and after journaling
images.
• Test. The step MXSETVRTT changes the replication status to Virtual switch
testing, which indicates that the environment is ready for you to begin testing
applications on the backup node that you identified as the test node. You are
responsible for starting, testing, and ending the application on the test node. While
you are testing, the procedure waits for a user reply to a waiting message for step
MXVRTRCY. When you have completed testing, respond to this message to
indicate that you are ready for recovery to begin.
• Recover. When step MXVRTRCY runs, it changes the replication status to Virtual

1. The PRECHKVRT and SWTVRT procedures are added to existing applications groups when
the instance is upgraded to MIMIX version 8.1 or higher.

317
Support for virtual switching

switch recovery. The remaining steps return the test node to its original purpose
as a backup node in the MIMIX production environment. During recovery, all
foreign key constraints are disabled on files on the test node that are eligible for
replication. Triggers are disabled on all target node files eligible for replication. If
necessary, files on the test node which had their journal image settings changed
during the setup phase are changed back to their original settings. Target node
processes are started beginning with the last processed entry for the receiver and
sequence number, and apply processes begin applying the queued transactions
that accumulated during testing. Constraints on the target node are enabled
unless the data group is configured to allow MIMIX to perform target constraint
management. Target journal inspection is changed to allow sending notifications
again. If there are any remaining objects that could not be recovered, they will be
reflected in the user interface as requiring manual correction through replication
activity entries, notifications, or both. Then replication statuses will no longer
reflect a virtual switch status. Finally, prioritized audits are submitted to verify the
objects changed on the target node during testing, these audits run regardless of
whether prioritized auditing is enabled.
For more information about steps within the SWTVRT procedure, see the MIMIX
Administrator Reference book.

Automatic error recovery during a virtual switch


Even though the apply processes are ended during a virtual switch, other replication
processes are active and continue to automatically check for replication problems.
The automatic recovery actions for detected problems are not processed until the
recovery phase of the virtual switch.
During the test phase of a virtual switch, any recovery actions that require
synchronizing objects or data between the source and target nodes will capture
source objects according to the “during virtual switch” setting of the Source capture
delay policy. The delay time must be met before the synchronization request for the
affected object or records is submitted into the journal stream. The recovery action is
not processed until the recovery phase of the virtual switch. The policy values are
configurable. For more information about selecting an appropriate source capture
delay value during a virtual switch, see online help topics or see the MIMIX
Operations book.
Source node objects with successful recovery actions are flagged for prioritized
auditing. A prioritized auditing request is submitted at the end of the virtual switch
procedure regardless of whether you have enabled prioritized auditing.
If a recovery action is not successful, the activity entry for the affected object or file is
changed to a replication error status For objects not within the scope of replication
that are related to replicated objects, an error notification will be sent.

Viewing recovery actions for target objects changed during a virtual switch
When using the MIMIX portal application in Vision Solutions Portal, you can use
actions available in the Application Groups portlet and Data Groups portlet to view the
following information:

318
MIMIX environment considerations for virtual switching

• Object Correction Activity window - lists objects for which recovery actions exist
and shows status of the recovery actions. The window can be filtered to those
only for a virtual switch. There may be multiple problems or changes for an object
but the window lists only one row per object per data group.
• From the Application Groups portlet, you can use the Virtual Switch Activity action
for an application group to view the list objects changed on the test node and their
recovery actions. After the recovery phase of the virtual switch begins, you will be
able to see any objects that recovery actions were unable to recover.
The ability tor view the objects changed during a virtual switch and the recovery
actions are not available in the native user interface.

MIMIX environment considerations for virtual switching


Virtual switching is supported at the application group level within instances licensed
for MIMIX® Availability™. The instance cannot be configured as an INTRA
environment.
Configuration requirements: The following configuration requirements must be met:
• The application group must be of type non-clustered.
• Target journal inspection must be configured and active on the node used for
testing. Target journal inspection must be active for the system journal on the test
node and for all user journals on the test node that are currently the target node of
data groups within the application group.
• All data groups associated with the application group must be switchable, must
allow journaling on their target node, and must allow cooperative processing
through the user journal.
• Keyed replication is not supported and cannot be configured in any of the data
groups or selection rules for libraries and files (object entries and file entries)
associated with the application group.
• Objects replicated by the data groups associated with the application group must
be journaled on the target node.
You can use the procedure PRECHKVRT to check for configuration requirements in
advance. The virtual switch procedure SWTVRT also checks configuration
requirements and will not continue if a requirement is not met.
Operational considerations: There is exposure to your environment for the duration
of a virtual switch and differences in some MIMIX behavior. Be aware of the following:
• During a virtual switch, replication activity for your production environment is not
complete because apply processes are stopped. Journal entries for replicated
transactions accumulate in queues for apply processes on the target node. When
the recovery phase of the virtual switch begins and the apply processes are
started, their backlog may be at or exceed threshold conditions. It may take a
significant amount of time to process a large backlog of transactions. The queued
backlogs and retained IBM journals for the amount of replication activity and

319
Support for virtual switching

virtual switch activity that occurs during a virtual switch may cause DASD issues.
• For the duration of the virtual switch procedure, the virtual switch controls when
the apply processes and access path maintenance are ended and restarted. You
can start or stop other replication processes as needed on the production node.
• Virtual switching uses target journal inspection processes to identify objects that
have changed on the test node. Starting and stopping MIMIX also starts and stops
target journal inspection. Target journal inspection processes need to remain
active during virtual switch testing to avoid having a period of time where changes
made on the test node are not detected. Data groups affected by the virtual switch
can be started and stopped.
• Virtual switch status is triggered by specific step programs in the virtual switch
procedure and is reflected in the replication status of the application group and its
resource groups and data groups. At the application group level, replication status
will indicate statuses of virtual switch starting, virtual switch testing in progress,
and virtual switch recovery in progress when no other conditions that require
attention exist. While a virtual switch status is displayed, other less severe
conditions can exist and you will need to drill into more detailed interfaces to see
those problems. When using Vision Solutions Portal, an important difference in
the overall status column in the Application Groups portlet is that any virtual switch
status takes precedence over all other possible statuses so that you can easily
identify that a virtual switch is in progress. Detailed columns in that portlet will
reflect any other issues that need your attention.
• A failed or canceled virtual switch procedure can only be acknowledged if the
failure or cancel request occurred before step MXSETVRTS. If the procedure fails
or is canceled after the virtual switch starting status is displayed for replication,
you will not be able to start, stop, or switch the application group until after the
virtual switch procedure is resumed and runs to completion.
• In interfaces for starting a virtual switch, the default value for the test node is the
node identified as the first backup in the current node roles of the recovery
domain. If the application group has three or more nodes in its recovery domain,
you can select to use a different backup node. Replicate nodes cannot be
selected as the test node.
• In cascading environment, do not use the middle node of the cascade as the test
node for a virtual switch.
• Audits are not allowed to run during a virtual switch.
• Target journal inspection will not issue notifications while virtual switch testing is in
progress.
• If you subscribe to MIMIX events through Vision Solutions Portal, be aware that
some events will not be triggered during a virtual switch or when certain virtual
switch states occur. There are unique events that are triggered by virtual switch
status changes.For more information, see the list of MIMIX subscription events
available in online help.

320
Testing considerations for virtual switching

Testing considerations for virtual switching


Before you run a virtual switch procedure, evaluate the applications you intend to test
and make a plan so that testing can be as useful and the least disruptive to your
environment as possible.
General planning: In general, do not leave an application group in virtual switch test
status for an extended period of time. Have a plan for what needs to be tested, what
resources you need, the people needed to perform application testing, and how long
the test is expected to take. Consider the following:
• Determine how much time you need for application setup and testing. If you have
multiple applications to test, consider the total amount of time needed and total
number of objects affected. It may be useful to test a limited number of
applications during one run of the virtual switch procedure. You may not want to
allow an application group to remain in its testing phase overnight. Or you may
need to plan to run the procedure and test during off-peak production hours.
• If a virtual switch is in progress for an extended time, both source and target
nodes may be affected. Journal receivers can accumulate on the both source and
target nodes of the affected data groups, and the queued work for apply sessions
accumulate on the target node. On the target node, there are multiple
considerations that affect the recovery phase, including, the quantity of
accumulated production transactions that need to be applied, the number of
changed target objects that need to be recovered, and manual cleanup.
• All changes made to target node objects need to be reverted. The recovery phase
of a virtual switch only reverts (recovers) changed target objects that are identified
as eligible for replication by the configuration of the data groups affected by the
virtual switch. Manual cleanup is required for the following types of changes:
– Any spool files created, changed, or deleted on the test node are not
automatically recovered by MIMIX. Those changes remain on the target even
after the virtual switch has completed.
– MIMIX does not track or automatically revert any changes made to objects that
are not eligible for replication by the affected data groups. This includes objects
on test node as well as object on other connected nodes or the cloud.
• If you use row and column access control (RCAC) on a replicated file and change
permissions or masks on the target node during virtual switch testing, be aware
that virtual switch recovery will perform a full save/restore of the file.
Evaluate the application requirements: Consider the characteristics of the
applications you intend to test. Some applications may not be suitable or may require
additional effort to avoid affecting your larger environment.
• Enable user profiles needed for the application on the target node at the
beginning of virtual switch testing, before starting the application. Optionally, you
can add this activity to step STRUSRAPPV with other customized application
startup instructions.
• In general, if the application needs certain user profiles, authorities, libraries,
objects, or other items in order to run on the test node, consider adding

321
Support for virtual switching

configuration selection rules (data group entries) for those items so they are
replicated to the target node.
• Determine if the application needs communications access to the test node. This
may need to be set up at the start of the testing phase.
• If an application changes objects or creates data on other connected servers, the
cloud, or in other applications, you may need to temporarily disable those
capabilities to prevent generating data to those locations. If you cannot disable
this activity, any changes outside of the designated test node may cause
damaged objects and will require manual effort to revert. If disabling is not
possible, the application may not be a good candidate for virtual switch testing.
• If an application hard codes an IP address or system name, the application will not
function properly and will not find the address or system when started on the
virtual switch test node. These scenarios are issues in any switch and need to be
resolved. You may need to work with the application provider to resolve them.
• If your environment uses IP impersonation to change network names outside of
MIMIX as part of your normal planned switch processing, carefully evaluate how
applications interact with these changes to determine if virtual switching testing is
viable for the affected applications.
• Identify what actions you need to perform to start and end the application.
Optionally, you can enable and customize steps STRUSRAPPV and
ENDUSRAPPV with instructions to start and end the application on the test node
as part of the virtual switch procedure.
Testing in MIMIX for MQ environments: If your environment includes the MIMIX for
MQ feature and you perform a virtual switch of an application group that replicates an
IBM WebSphere MQ queue manager, the AMQAJRN journal should not be allowed to
change its journal receiver while in the application testing phase of the virtual switch.
Do not intentionally change this journal receiver during virtual switch testing. If the
receiver for this journal changes during virtual switch testing, a manual recovery
process is required after the virtual switch procedure has completed. You will need to
check for MQ-related topics in the Knowledgebase on Support Central and may need
to contact CustomerCare.

322
Performing a precheck for virtual switching

Performing a precheck for virtual switching


These instructions use the PRECHKVRT procedure1 for the selected application
group to validate that an application group environment meets configuration
requirements for virtual switching.

To run the virtual switch precheck, do the following:


1. From the Work with Application Groups display, type 20 (Procedures) next to the
application group you want and press Enter.
2. Available procedures for the application group are displayed on the Work with
Procedures display. Type 9 (Run) next to the PRECHKVRT procedure and press
Enter.
You can check status of the procedure and its steps from the Work with Procedure
Status display.

Performing a virtual switch


A virtual switch of an application group allows you to use a target node in your MIMIX
production environment to test applications. If you have not already done so, it is
strongly recommended that you take time to become familiar with how virtual
switching affects your MIMIX environment and the testing considerations for
applications. See “Support for virtual switching” on page 316 for details.

These instructions use the default Virtual Switch (DFT) procedure2 for a selected
application group from the native user interface. From this interface, you can start the
procedure and monitor its status, but you cannot view a list of objects changed by
testing on the target node or the status of recovery actions for those objects. Those
capabilities are available only from Vision Solutions Portal, which is recommended.
Starting the virtual switch procedure:
To start the virtual switch, do the following:
1. From the Work with Application Groups display, type 15 (Switch) next to the
application group you want and press Enter.
The Switch Application Group (SWTAG) display appears.
2. Specify *VIRTUAL for the Switch type prompt.
3. At the Node roles prompt, specify the node to be used for testing in the virtual
switch. (No changes to the recovery domain roles will occur.) The default value
*CURRENT will use the first backup node in the current node roles. This value is
appropriate for typical 2-node application groups. To use a different backup node

1. The PRECHKVRT procedure is created when an application group is created in an instance


running MIMIX version 8.1 or higher, or when an instance is upgraded to MIMIX version 8.1
or higher,
2. The SWTVRT procedure is created from a shipped default when an application group of type
Non-cluster (*NONCLU) is created in an instance running MIMIX version 8.1 or higher, or
when an instance is upgraded to MIMIX version 8.1 or higher

323
in an application groups with multiple backup nodes, specify its name.
4. Press Enter.
5. The Procedure prompt appears. Do one of the following:
• To use the default virtual switch procedure (SWTVRT), press Enter.
• To use a different virtual switch procedure for the application group, specify its
name.
6. Press Enter.
7. You can check status of the virtual switch procedure from the Work with
Procedures display. Refer to instructions in “Working with procedures and steps”
on page 97.
Important!
For problem resolution, a virtual switch procedure differs from other types of
procedures in these ways:
• A virtual switch procedure is intended to stay in *MSGW state at step
MXVRTRCY for the entire time that application tests are being performed on
the designated test node. Do not respond to the message for this step until
testing is complete and you are ready to start the recovery phase of the virtual
switch. For all other steps, handle a *MSGW procedure status like that for any
other procedure.
• If a problem occurs in any step within the range performed while the
application group has a replication status of *VRTSWTSTR, *VRTSWTTST or
*VRTSWTRCY and the virtual switch procedure fails or is canceled, you
cannot acknowledge the failed or canceled procedure. The virtual switch
procedure must be resumed after you have investigated the cause of the
problem for the affected step. Furthermore, you will not be able to run any
other procedure to start, stop, or switch the affected application group until the
virtual switch procedure is resumed and allowed to run to completion. The
range of steps to which this behavior applies begins with step MXSETVRTS
and applies through step MXFAILRCYV.
• If a problem with the source node occurs that requires an unplanned switch
while a virtual switch is in progress, the virtual switch procedure must be
allowed to complete. This is necessary even in environments with more than
one backup node in the application group’s recovery domain. If the virtual
switch is in its testing phase, end the application being tested and respond to
the message waiting at step MXVRTRCY. Allow the virtual switch procedure to
complete before attempting to start the unplanned switch. Any recovery
actions that require capturing data from the source node will fail and will have a
Not Recovered status in the Virtual Switch Activity window in Vision Solutions
Portal.
Testing phase: The MIMIX environment on the designated test node is ready for
application testing when the virtual switch procedure successfully processes step
MXSETVRTT. As a result of this step, the Work with Application Groups display
indicates that virtual switch testing is in progress with the value *VRTSWTTST in the

324
Performing a virtual switch

Repl. Status column if replication processes are active. (Apply processes and access
path maintenance remain ended while in the testing phase.)
Important!
While you are testing, the procedure is designed to wait for a user reply that
indicates testing is complete before it processes step MXVRTRCY. You will see
the status *MSGW on the Work with Procedure Status display for the duration of
the testing phase of the virtual switch. Do not respond to this message until
you have completed testing.
8. If you have not enabled and customized step STRUSRAPPV, you must manually
start the application to be tested on the designated test node.
9. Test the application.
While testing, it is good practice to keep a record of your test node activity that
may require manual cleanup when testing is complete. This includes any setup
that will need to be removed, any spool files created, and any changed objects
that are outside of the scope of replication for the participating data groups.
10. To see the objects eligible for replication that have been changed on the test
node, use the Virtual Switch Activity action for the application group from the
Application Groups portlet in Vision Solutions Portal.
11. For the duration of virtual switch testing, monitor the application group for potential
problems from the Work with Application Groups display. Check for any indication
of problems that require action and take action as needed.
12. When testing is complete, end the application on the test node.
Note: If you enabled step ENDUSRAPPV and customized it with actions to end
the application, you can skip this step. When enabled, the customized step
runs after you respond to the message waiting in Step 14.
13. Perform any necessary manual clean up on the test node, such as deleting any
unnecessary spooled files created during testing.
Recovery phase:
14. To indicate that testing is complete and you are ready for the virtual switch
procedure to start its recovery phase, you must respond to the waiting message.
Do the following:
a. From the Work with Application Groups display, type 21 (Procedure status)
next to the application group you want and press Enter.
b. From the Work with Procedure Status display. type 11 (Display message) next
to the virtual switch procedure (SWTVRT) and press Enter.
c. You will see the message “Procedure name for application group name
requires response. (G C).” To start the recovery phase of the virtual switch,
type G and press Enter.
Note: A response of C (Cancel) will cancel the procedure. If you cancel the
procedure, you must resume the procedure and allow it to run to
completion.
This starts the recovery phase of the virtual switch and changes the replication

325
status for the application group to indicate that virtual switch is recovering
(*VRTSWTRCY on the Work with Application Groups display). Apply processes
will be started and will begin processing the accumulated transactions and
recovery actions.
15. You can check the progress of recovery actions from the Virtual Switch Activity
window within Vision Solutions Portal.
16. When the procedure completes, check the Virtual Switch Activity window for any
objects with failed recoveries. You will need to take action to address these
objects.
Additional information about using the Virtual Switch Activity window to check
recovery status and start manual resolution of failed recoveries is available in online
help for a MIMIX instance from within Vision Solutions Portal.

326
CHAPTER 16 Less common operations

This chapter describes how to perform infrequently used operations that help keep
your MIMIX environment running. The following topics are included:
• “Starting the TCP/IP server” on page 328 contains the procedure for starting the
TCP/IP server.
• “Ending the TCP/IP server” on page 329 contains the procedure for ending the
TCP/IP server.
• “Working with objects” on page 330 contains tips for working with long object and
IFS path names.
• “Viewing status for active file operations” on page 331 describes how to check
status when replicating database files that you are reorganizing or copying with
MIMIX Promoter.
• “Identifying data groups that use an RJ link” on page 332 describes how to
determine which data groups use a remote journal link.
• “Identifying journal definitions used with RJ” on page 333 describes how to
determine whether a journal definition is defined to one or more remote journal
links.
• “Disabling and enabling data groups” on page 334 describes when it can be
beneficial to disable and enable data groups. Procedures for these processes are
included in this topic.
• “Determining if non-file objects are configured for user journal replication” on
page 336 provides procedures for determining whether configured for IFS objects,
data areas, and data queues are configured to be cooperatively processed
through the user journal.
• “Using file identifiers (FIDs) for IFS objects” on page 338 describes file identifiers
(FIDs) which are used by commands to uniquely identify the correct IFS tracking
entries to process.

327
Starting the TCP/IP server
Use this procedure if you need to manually start the TCP/IP server.
Once the TCP communication connections have been defined in a transfer definition,
the TCP server must be started on each of the systems identified by the transfer
definition.
You can also start the TCP/IP server automatically through an autostart job entry.
Either you can change the transfer definition to allow MIMIX to create and manage
the autostart job entry for the TCP/IP server, or you can add your own autostart job
entry. MIMIX only manages entries for the server when they are created by transfer
definitions.
When configuring a new installation, transfer definitions and MIMIX-added autostart
job entries do not exist on other systems until after the first time the MIMIX managers
are started. Therefore, during initial configuration you may need to manually start the
TCP server on the other systems using the STRSVR command.
Note: Use the host name and port number (or port alias) defined in the transfer
definition for the system on which you are running this command.
Do the following on the system on which you want to start the TCP server:
1. From the MIMIX Intermediate Main Menu, select option 13 (Utilities menu) and
press Enter.
2. The Utilities Menu appears. Select option 51 (Start TCP server) and press Enter.
3. The Start Lakeview TCP Server display appears. At the Host name or address
prompt, specify the host name or address for the local system as defined in the
transfer definition.
4. At the Port number or alias prompt, specify the port number or alias as defined in
the transfer definition for the local system.
Note: If you specify an alias, you must have an entry in the service table on this
system that equates the alias to the port number.
5. Press Enter.
6. Verify that the server job is running under the MIMIX subsystem on that system.
You can use the Work with Active Jobs (WRKACTJOB) command to look for a job
under the MIMIXSBS subsystem with a function of PGM-LVSERVER.

328
Ending the TCP/IP server

Ending the TCP/IP server


To end the TCP server, do the following on both systems defined by the transfer
definition. One example of why you might end the TCP server is when you are
preparing to upgrade the MIMIX products in a product library.
Note: Use the host name and port number (or port alias) defined in the transfer
definition for the system you on which you are running this command
To end the TCP server on a system, do the following:
1. From the MIMIX Intermediate Main Menu, select option 13 (Utilities menu) and
press Enter.
2. The Utilities Menu appears. Select option 52 (End TCP server) and press Enter.
3. The End Lakeview TCP Server display appears. At the Host name or address
prompt, specify the host name for the local system as specified in the transfer
definition.
4. At the Port number or alias prompt, verify that the value shown is what you want. If
necessary change the value.
Note: If the configuration uses port aliases, specify the alias for local system.
Otherwise, specify the port number for the local system.
5. Press Enter.

329
Working with objects
When working with objects, these tips may be helpful.

Displaying long object names


The names of some IFS entries cannot be fully displayed in the limited space on a
“Work with” display. These entries are shown with a ‘>’ character in the right-most
column of the Object field.
You can display long object names from the following displays:
• Work with Data Group IFS Entries display
• Work with Data Group Activity
• Work with Data Group Activity Entries
To display the entire object name from any of these displays, position the cursor on
an entry which indicates a long name and press F22 (Display entire field).

Considerations for working with long IFS path names


MIMIX currently replicates IFS path names of 512 characters. However, any MIMIX
command that takes an IFS path name as input may be susceptible to a 506
character limit. This character limit may be reduced even further if the IFS path name
contains embedded apostrophes (') or trailing blanks. In this case, the supported IFS
path name length is reduced by four characters for every apostrophe the path name
contains.
For information about IFS path name naming conventions, refer to the IBM book,
Integrated File System.

Displaying data group spooled file information


If spooled files are created as a result of MIMIX replication, you can access the
spooled file and the associated data group entry from the Work with Data Group
Activity display.
To access the spooled file information, do the following:
1. From the MIMIX Basic Main Menu, select option 6 (Work with data groups) and
press Enter.
2. The Work with Data Groups display appears. Select option 14 (Active objects) for
the data group you want to view and press Enter. The Work with Data Group
Activity display appears.
3. From this display, press F16 (Spooled Files) to access the Display Data Group
Spooled Files display. This display lists all of the current spooled files and shows
the mapping of their names between the source and target systems.

330
Viewing status for active file operations

Viewing status for active file operations


If you are replicating database files that you are reorganizing or copying with MIMIX
Promoter, you can check on the status of these operations. Do the following:
1. From the MIMIX Basic Main Menu, use F21 (Assistance level) to access the
intermediate menu.
2. From the MIMIX Intermediate Main Menu, select option 13 (Utilities menu) and
press Enter.
3. From the MIMIX Utilities Menu, select option 63 (Work with copy status) and
press Enter.
4. The Work with Copy Status display appears. From this display you can track the
status of active copy or reorganize operations, including the replication of physical
file data as specified by METHOD(*DATA) on the Synchronize Data Group File
Entry (SYNCDGFE) command.
Note: You can only see status for the system on which you are working.

331
Identifying data groups that use an RJ link
Use this procedure to determine which data groups use a remote journal link before
you end a remote journal link or remove a remote journaling environment.
1. Enter the command WRKRJLNK and press Enter.
2. Make a note of the name indicated in the Source Jrn Def column for the RJ Link
you want.
3. From the command line, type WRKDGDFN and press Enter.
4. For all data groups listed on the Work with DG Definitions display, check the
Journal Definition column for the name of the source journal definition you
recorded in Step 2.
• If you do not find the name from Step 2, the RJ link is not used by any data
group. The RJ link can be safely ended or can have its remote journaling
environment removed without affecting existing data groups.
• If you find the name from Step 2 associated with any data groups, those data
groups may be adversely affected if you end the RJ link. A request to remove
the remote journaling environment removes configuration elements and
system objects that need to be created again before the data group can be
used. Continue with the next step.
5. Press F10 (View RJ links). Consider the following and contact your MIMIX
administrator before taking action that will end the RJ link or remove the remote
journaling environment.
• When *NO appears in the Use RJ Link column, the data group will not be
affected by a request to end the RJ link or to end the remote journaling
environment.
Note: If you allow applications other than MIMIX to use the RJ link, they will be
affected if you end the RJ link or remove the remote journaling
environment.
• When *YES appears in the Use RJ Link column, the data group may be
affected by a request to end the RJ link. If you use the procedure for ending a
remote journal link independently in topic “Ending a remote journal link
independently” on page 202, ensure that any data groups that use the RJ link
are inactive before ending the RJ link.

332
Identifying journal definitions used with RJ

Identifying journal definitions used with RJ


To see whether a journal definition is defined to one or more remote journal links, do
the following:
1. From the MIMIX Basic Main Menu, select option 11 (Configuration menu) and
press Enter.
2. The MIMIX Configuration menu appears. Select option 3 (Work with journal
definitions) and press Enter.
The Work with Journal Definitions display appears. By default, the display shows
journal definitions used by replication processes.
Optional: to see journal definitions used by other internal processes, press
F18 (Subset), specify *INTERNAL at the Used by process prompt, and press
Enter.
The RJ Link column indicates whether a journal definition is used by a remote
journal link. Possible value are:
*SOURCE - The journal definition is a source journal definition in a remote journal
link.
*TARGET - The journal definition is the target journal definition in a remote journal
environment.
*BOTH - The journal definition is the source journal definition for one remote
journal link and is also a target journal definition for another remote journal link in
a cascading environment.
*NONE - The journal definition is not configured in a remote journal link.
3. To see the remote journal links associated with a journal definition, type 12 (Work
with RJ Links) and press Enter.

333
Disabling and enabling data groups
Data groups are in an enabled state when they are created. An enabled data group
may have active or inactive replication processes. MIMIX supports the concept of a
disabled state for data groups in a replication environment. When a data group is
disabled, it cannot be started, ended, switched, or audited. However, its configuration
is retained so that the data group can be enabled and used again when needed.
In environments where multiple data groups exist within a single resource group,
changes to data group configuration entries that identify objects to replicate are
propagated to the data groups within a resource group entry as follows:
• If the configuration entry changes are made from an enabled data group, they are
propagated to all data groups within the resource group entry, including disabled
data groups.
• If configuration entry changes are made from a disabled data group, they are not
propagated to the other data groups in the resource group entry.
Data groups associated with an application group may be enabled and disabled
automatically as part of a switch procedure (SWTAG command). If an application
group includes a data group that you do not want to participate in switching
operations, you can change its data group definition to specify *NO for its Allow to be
switched (ALWSWT) parameter.
Data groups can also be manually enabled or disabled with the Change Data Group
(CHGDG) command.
The ability to disable and enable a data group can be beneficial in a variety of
scenarios.
• Disabling a data group is useful for test environments. If you create a data group
for testing purposes, when testing is complete you can simply end and disable the
data group until it is needed again instead of having to delete the data group to
clean up your environment. This provides the benefit of retaining your configured
data group object, file, IFS, and DLO entries while the data group is not needed.
Additionally, the journal manager does not retain journal receivers that have not
been processed by a disabled data group, which allows you to save storage
space on your system.
• In environments that do not use application groups and have non-switchable data
groups configured to replicate in different directions, the ability to disable data
groups can simplify starting replication. You can avoid having to start each data
group individually by disabling the data groups whose direction is not what you
want. Then you can simply start the remaining data groups and the MIMIX
environment using the Start MIMIX (STRMMX) command.
By default, only enabled data groups are shown on the Work with Data Groups
(WRKDG) display. You can display disabled data groups by specifying *ALL or
*DISABLED on the STATE parameter. Disabled data groups are indicated by a status
of D (in green).
Before a data group can be disabled, its replication processes must be ended
(inactive status) and any constraints on target node objects that were previously

334
Disabling and enabling data groups

disabled by MIMIX must be enabled. In environments that use the MIMIX® CDP™
feature, data group processes cannot be suspended at a recovery point before
ending replication because the request to end the data group will clear any recovery
point.
When a disabled data group is enabled, any pending entries must be cleared when
the data group is started. Specify CLRPND(*YES) on the Start Data Group command.

Disabling a data group


These instructions change the state of a data group to *DISABLED. System manager
processes for the data group must be active when you perform these instructions.
To disable an individual data group, do the following:
1. End the data group and enable any constraints previously disabled by MIMIX
using the instructions in “Enabling constraints when ending a data group” on
page 249. Wait for the data group status to become *INACTIVE before continuing.
2. On a command line, type CHGDG and press Enter.
The Change Data Group display appears.
3. At the Data group definition prompts, fill in the values you want or press F4 for a
valid list.
4. At the State prompt, specify *DISABLE.
5. Press Enter to confirm your changes.

Enabling a data group


These instructions enable a data group by changing its state from DISABLED to
*INACTIVE. System manager processes for the data group must be active when you
perform these instructions.
To enable an individual data group, do the following:
1. On a command line, type CHGDG and press Enter.
The Change Data Group display appears.
2. At the Data group definition prompts, fill in the values you want or press F4 for a
valid list.
3. At the State prompt, specify *ENABLE.
4. Press Enter to confirm your changes.
Note: To start an enabled data group, you must specify *YES for the Clear pending
entries prompt on the Start Data Group (STRDG) command. See “Starting
selected data group processes” on page 228.

335
Determining if non-file objects are configured for user
journal replication
MIMIX can take advantage of IBM i journaling functions that provide change-level
details in journal entries in a user journal for object types other than files (*FILE).
When properly configured, MIMIX can cooperatively process IFS stream files, data
areas, and data queues between system journal and user journal replication
processes. This enables changes to data or attributes to be replicated through the
user journal instead of replicating the entire object through the system journal every
time a change occurs.

Determining how IFS objects are configured


In order for IFS objects to be replicated from the user journal, one or more data group
IFS entries must be configured to process cooperatively with the user journal. Also,
IFS tracking entries must exist for the object identified by the data group IFS entries.
To determine if a data group has any IFS objects that are configured for user journal
replication and has any corresponding IFS tracking entries, do the following:
1. From the MIMIX Basic Main Menu select option 6 (Work with data groups) and
press Enter.
2. The Work with Data Groups display appears. Type 22 (IFS entries) next to the
data group you want and press Enter.
The Work with DG IFS Entries display appears, showing the IFS entries
configured for the data group.
3. Press F10 twice to access the CPD view.
4. The values shown in the Coop with DB column indicate how objects identified by
the data group IFS entries will be replicated.
• Entries with the value *YES are configured for user journal replication.
Continue with the next step to ensure that IFS tracking entries exist for the IFS
objects. Replication cannot occur without tracking entries.
• Entries the value *NO are configured for system journal replication.
To view additional information for a data group IFS entry, type 5 (Display) next to
the entry and press Enter.
5. Press F12 (Cancel) to return to the Work with Data Groups display. Then type 50
(IFS trk entries) next to the data group you want and press Enter.
6. The Work with DG IFS Trk. Entries display appears with a list of tracking entries
for the IFS objects identified for replication by the data group. If there are no
tracking entries listed but Step 4 indicates that properly configured data group IFS
entries exist, the tracking entries must be loaded. For more information about
loading tracking entries, see the MIMIX Administrator Reference book.

336
Determining if non-file objects are configured for user journal replication

Determining how data areas or data queues are configured


In order for data area and data queue objects to be replicated from the user journal,
one or more data group object entries must be configured to process cooperatively
with the user journal. Also, object tracking entries must exist for the object identified
by the data group object entries.
To determine if a data group has any data area or data queue objects that are
configured for user journal replication and has any corresponding object tracking
entries, do the following:
1. From the MIMIX Basic Main Menu select option 6 (Work with data groups) and
press Enter.
2. The Work with Data Groups display appears. Type 20 (Object entries) next to the
data group you want and press Enter.
The Work with DG Object Entries display appears, showing the object entries
configured for the data group.
3. For each entry in the list, do the following:
a. Type a 5 (Display) next to the entry and press Enter.
b. The object entry must have the following values specified in the fields
indicated:
• The Object type field must be *ALL, *DTAARA, or *DTAQ
• The Cooperate with database field must be *YES
• The Cooperating object types field must specify *DTAARA to replicate data
areas and *DTAQ to replicate data queues.
4. Press F12 (Cancel) to return to the Work with Data Groups display. Then type 52
(Obj trk entries) next to the data group you want and press Enter.
5. The Work with DG Obj. Trk. Entries display appears with a list of tracking entries
for the data area and data queue objects identified for replication by the data
group. If there are no tracking entries listed but Step 3 indicates that properly
configured data group object entries exist, the tracking entries must be loaded.
For more information about loading tracking entries, see the MIMIX Administrator
Reference book.

337
Using file identifiers (FIDs) for IFS objects
Commands used for user journal replication of IFS objects use file identifiers (FIDs) to
uniquely identify the correct IFS tracking entries to process. The System 1 file
identifier and System 2 file identifier prompts ensure that IFS tracking entries are
accurately identified during processing. These prompts can be used alone or in
combination with the System 1 object prompt.
These prompts enable the following combinations:
• Processing by object path: A value is specified for the System 1 object prompt
and no value is specified for the System 1 file identifier or System 2 file identifier
prompts.
When processing by object path, a tracking entry is required for all commands
with the exception of the SYNCIFS command. If no tracking entry exists, the
command cannot continue processing. If a tracking entry exists, a query is
performed using the specified object path name.
• Processing by object path and FIDs: A value is specified for the System 1
object prompt and a value is specified for either or both of the System 1 file
identifier or System 2 file identifier prompts.
When processing by object path and FIDs, a tracking entry is required for all
commands. If no tracking entry exists, the command cannot continue processing.
If a tracking entry exists, a query is performed using the specified FID values. If
the specified object path name does not match the object path name in the
tracking entry, the command cannot continue processing.
• Processing by FIDs: A value is specified for either or both of the System 1 file
identifier or System 2 file identifier prompts and, with the exception of the
SYNCIFS command, no value is specified for the System 1 object prompt. In the
case of SYNCIFS, the default value *ALL is specified for the System 1 object
prompt.
When processing by FIDs, a tracking entry is required for all commands. If no
tracking entry exists, the command cannot continue processing. If a tracking entry
exists, a query is performed using the specified FID values.

338
CHAPTER 17 Troubleshooting - where to start

Occasionally, a situation may occur that requires user intervention. This section
provides information to help you troubleshoot problems that can occur in a MIMIX
environment.
You can also consult our website at www.mimix.com for the latest information and
updates for MIMIX products.
The following topics are included in this chapter:
• “Reducing contention between MIMIX and user applications” on page 341
describes a processing timing issue that may be resolved by specifying an Object
retrieval delay value on the commands for creating or changing data group
entries.
• “Data groups cannot be ended” on page 342 describes possible causes for a data
group that is taking too long to end.
• “Verifying a communications link for system definitions” on page 343 describes
the process to verify that the communications link defined for each system
definition is operational.
• “Verifying the communications link for a data group” on page 344 includes a
process to use before synchronizing data to ensure that the communications link
for the data group is active.
• “Checking file entry configuration manually” on page 345 includes the process for
checking that correct data group file entries exist with respect to the data group
object entries. This process uses the Check DG File Entries (CHKDGFE)
command.
• “Data groups cannot be started” on page 347 describes some common reasons
why a data group may not be starting.
• “Cannot start or end an RJ link” on page 348 describes possible reasons that can
prevent you from starting or ending an RJ link. This topic includes a procedure for
removing unconfirmed entries to free an RJ link.
• “RJ link active but data not transferring” on page 349 describes why an RJ link
may not be transferring data and how to resolve this problem.
• “Errors using target journal defined by RJ link” on page 350 describes why errors
when using a target journal defined by an RJ link can occur and how to resolve
them.
• “Verifying key attributes” on page 351 includes a procedure for verifying key
attributes using the VFYKEYATR (Verify Key Attributes) command.
• “Working with data group timestamps” on page 352 describes timestamps and
includes information for creating, deleting, displaying, and printing them.
• “Removing journaled changes” on page 355 describes the configuration
conditions that must be met using the Remove Journaled Changes
(RMVJRNCHG) journal entry.

339
Troubleshooting - where to start

• “Performing journal analysis” on page 356 describes and includes the procedure
for performing journal analysis of the source system.

340
Reducing contention between MIMIX and user applications

Reducing contention between MIMIX and user applica-


tions
If your applications are failing in an unexpected manner, it may be caused by MIMIX
locking your objects for object retrieval processing while your applications are trying to
access the object. This is a processing timing issue and can be significantly reduced,
or eliminated, by specifying an appropriate delay value for the Object retrieval delay
element under the Object processing (OBJPRC) parameter on the change or create
data group definition commands.
Although you can specify this value at the data group level, you can override the data
group value at the object level by specifying an Object retrieval delay value on the
commands for creating or changing data group entries.
For more information see “Selecting an object retrieval delay” in the MIMIX
Administrator Reference book.
You should use care when choosing the object retrieval delay. A long delay may
impact the ability of MIMIX system journal replication processes to move data from a
system in a timely manner. Too short a delay may allow MIMIX to retrieve an object
before an application is finished with it. You should make the value large enough to
reduce or eliminate contention between MIMIX and applications, but small enough to
allow MIMIX to maintain a suitable high availability environment.

341
Data groups cannot be ended
A controlled end for a data group may take some time if there is a backlog of files to
process or if there are a number of errors that MIMIX is attempting to resolve before
ending.
If you think that a data group is taking too long to end, check the following for possible
causes:
• Check to see how many transactions are backlogged for the apply process. Use
option 8 (Display status) on the Work with Data Groups display to access the
detailed status. A number in the Unprocessed Entry Count column indicates a
backlog. Use F7 and F8 to see additional information.
• Determine which replication process is not ending. Use the command
WRKSBSJOB SBS(MIMIXSBS) to see the jobs in the MIMIXSBS subsystem.
Look for jobs for replication processes that have not changed to a status of END.
For example, abc_OBJRTV, where abc is a 3-character prefix.
• Check the QSYSOPR message log to see if there is message that requires a
reply.
• You can use the WRKDGACTE STATUS(*ACTIVE) command to ensure all data
group activity entries are completed. If a controlled end was issued, all activity
entries must be processed before the object processes are ended.

342
Verifying a communications link for system definitions

Verifying a communications link for system definitions


Do the following to verify that the communications link defined for each system
definition is operational:
1. From the MIMIX Basic Main Menu, type an 11 (Configuration menu) and press
Enter.
2. From the MIMIX Configuration Menu, type a 1 (Work with system definitions) and
press Enter.
3. From the Work with System Definitions display, type an 11 (Verify
communications link) next to the system definition you want and press Enter. You
should see a message indicating the link has been verified.
Note: If the system manager is not active, this process will only verify that
communications to the remote system is successful. You will also see a
message in the job log indicating that “communications link failed after 1
request.” This indicates that the remote system could not return
communications to the local system.
4. Repeat this procedure for all system definitions. If the communications link
defined for a system definition uses SNA protocol, do not check the link from the
local system.
Note: If your transfer definition uses the *TCP communications protocol, then
MIMIX uses the Verify Communications Link command to validate the
information that has been specified for the Relational database (RDB)
parameter. MIMIX also uses VFYCMNLNK to verify that the System 1 and
System 2 relational database names exist and are available on each
system.

343
Verifying the communications link for a data group
Before you synchronize data between systems, ensure that the communications link
for the data group is active. This procedure verifies the primary transfer definition
used by the data group. If your configuration requires multiple data groups, be sure to
check communications for each data group definition.
Do the following:
1. From the MIMIX Basic Main Menu, type an 11 (Configuration menu) and press
Enter.
2. From the MIMIX Configuration Menu, type a 4 (Work with data group definitions)
and press Enter.
3. From the Work with Data Group Definitions display, type an 11 (Verify
communications link) next to the data group you want and press F4.
4. The Verify Communications Link display appears. Ensure that the values shown
for the prompts are what you want.
5. To start the check, press Enter.
6. You should see a message "VFYCMNLNK command completed successfully."
If your data group definition specifies a secondary transfer definition, use the following
procedure to check all communications links.

Verifying all communications links


The Verify Communications Link (VFYCMNLNK) command requires specific system
names to verify communications between systems. When the command is called from
option 11 on the Work with System Definitions display or option 11 on the Work with
Data Groups display, MIMIX identifies the specific system names.
For transfer definitions using TCP protocol: MIMIX uses the Verify
Communications Link (VFYCMNLNK) command to validate the values specified for
the Relational database (RDB) parameter. MIMIX also uses VFYCMNLNK to verify
that the System 1 and System 2 relational database names exist and are available on
each system.
When the command is called from option 11 on the Work with Transfer Definitions
display or when entered from a command line, you will receive an error message if
the transfer definition specifies the value *ANY for either system 1 or system 2.
1. From the Work with Transfer Definitions display, type an 11 (Verify
communications link) next to all transfer definitions and press Enter.
2. The Verify Communications Link display appears. If you are checking a Transfer
definition with the value of *ALL, you need to specify a value for the System 1 or
System 2 prompt. Ensure that the values shown for the prompts are what you
want and then press Enter.
You will see the Verify Communications Link display for each transfer definition
you selected.
3. You should see a message "VFYCMNLNK command completed successfully."

344
Checking file entry configuration manually

Checking file entry configuration manually


The Check DG File Entries (CHKDGFE) command provides a means to detect
whether the correct data group file entries exist with respect to the data group object
entries configured for a specified data group in your MIMIX configuration. When file
entries and object entries are not properly matched, your replication results can be
affected.
Note: The preferred method of checking is to use automatic scheduling for the
#DGFE audit, which calls the CHKDGFE command and can automatically
correct detected problems. For additional information, see “Interpreting results
for configuration data - #DGFE audit” on page 361.
To check your file entry configuration manually, do the following:
1. On a command line, type CHKDGFE and press Enter. The Check Data Group File
Entries (CHKDGFE) command appears.
2. At the Data group definition prompts, select *ALL to check all data groups or
specify the three-part name of the data group.
3. At the Options prompt, you can specify that the command be run with special
options. The default, *NONE, uses no special options. If you do not want an error
to be reported if a file specified in a data group file entry does not exist, specify
*NOFILECHK.
4. At the Output prompt, specify where the output from the command should be
sent—to print, to an outfile, or to both. See Step 6.
5. At the User data prompt, you can assign your own 10-character name to the
spooled file or choose not to assign a name to the spooled file. The default, *CMD,
uses the CHKDGFE command name to identify the spooled file.
6. At the File to receive output prompts, you can direct the output of the command to
the name and library of a specific database file. If the database file does not exist,
it will be created in the specified library with the name MXCDGFE.
7. At the Output member options prompts, you can direct the output of the command
to the name of a specific database file member. You can also specify how to
handle new records if the member already exists. Do the following:
a. At the Member to receive output prompt, accept the default *FIRST to direct
the output to the first member in the file. If it does not exist, a new member is
created with the name of the file specified in Step 6. Otherwise, specify a
member name.
b. At the Replace or add records prompt, accept the default *REPLACE if you
want to clear the existing records in the file member before adding new
records. To add new records to the end of existing records in the file member,
specify *ADD.
8. At the Submit to batch prompt, do one of the following:
• If you do not want to submit the job for batch processing, specify *NO and
press Enter to check data group file entries.

345
• To submit the job for batch processing, accept *YES. Press Enter and continue
with the next step.
9. At the Job description prompts, specify the name and library of the job description
used to submit the batch request. Accept MXAUDIT to submit the request using
the default job description, MXAUDIT.
10. At the Job name prompt, accept *CMD to use the command name to identify the
job or specify a simple name.
11. To start the data group file entry check, press Enter.

346
Data groups cannot be started

Data groups cannot be started


Two common reasons why a data group cannot be started are as follows:
• The communications link between systems defined to the data group is not active.
Use the procedure “Verifying a communications link for system definitions” on
page 343.
• The journaling environment for the data group has not been built. Verify that
journaling environment defined in the journal definition exists. If necessary, use
the appropriate procedure in the MIMIX Administrator Reference book.
• The journal receiver has been deleted from the system. You can use WRKJRNA
to determine if the journal receiver exists on the source system.

347
Cannot start or end an RJ link
In normal operations, unconfirmed entries are automatically handled by the RJ link
monitors.
However, there is a scenario where you may end up with a backlog of unconfirmed
entries that can prevent you from starting or ending an RJ link. This problem can
occur when all of the following are true:
• The data group is not switchable or you do not want to switch it
• A link failure on an RJ link that is configured for synchronous delivery leaves
unconfirmed entries
• The RJ link monitors are not active, either because you are not using them or they
failed as a result of a bad link
To recover from this situation, you should run the Verify Communications Link
(VFYCMNLNK) command to assist you in determining what may by wrong and why
the RJ link will not start.
If you are using an independent ASP, check the transfer definition to ensure the
correct database name has been specified.
You also need to end the remote journal link from the target system. Ending the link
from the target system is a restriction of the IBM remote journal function.

Removing unconfirmed entries to free an RJ link


If you need to remove a backlog of unconfirmed entries, do the following:
1. Use the WRKRJLNK command to display the status of the RJ link. The status
shown on the Work with RJ Links display is the status of the link on the system
where you entered the command. (This system is identified at the upper right
corner of the display.) An RJ link with unconfirmed entries will have a state of
*INACTPEND.
Note: You may need to access this display from the other system defined by the
RJ link.
2. Ending the remote journal link on the system with unconfirmed entries will cause
them to be deleted. Do the following:
a. Type 10 (End) next to the link and press F4 (Prompt).
b. The End Remote Journal Link (ENDRJLNK) display appears. Default values
on this command ends the link from the source system. If there are
unconfirmed entries on the target system, press F10 (Additional parameters).
Then specify *TGT at the End RJ link on system prompt.
c. To process the request, press Enter.

348
RJ link active but data not transferring

RJ link active but data not transferring


Following an initial program load (IPL), the RJ link may appear to be active when data
cannot actually flow from the source system to the target system journal receiver. This
is an operating system restriction. MIMIX does not receive notification of a failure.
To recover, end the RJ lInk and restart it following an IPL. This can be included in
automation programs.

349
Errors using target journal defined by RJ link
If you receive errors when using a target journal defined by an RJ link, you may need
to change the journal definition and journaling environment. This situation is caused
when the target journal definition is created as a result of adding an RJ link based on
a source journal definition which specified QSYSOPR as the threshold message
queue.
If you receive errors when using the target journal, do the following:
1. On the Work with Journal Definitions display, locate the target journal definition
that is identified by the errors.
2. Type a 5 (Display) next to the target journal definition and press Enter.
3. Page down to see the value of the Threshold message queue.
• If the value is QSYSOPR, press F12 and continue with the next step.
• For any other value, the cause of the problem needs further isolation beyond
this procedure.
4. Type a 2 (Change) next to the target journal definition and press Enter.
5. Press F9 (All parameters), then page down to locate the Threshold message
queue and Library prompts.
6. Change the Threshold message queue prompt to *JRNDFN and the Library
prompt to *JRNLIB, or to other acceptable values.
7. To accept the change, press Enter.

350
Verifying data group file entries

Verifying data group file entries


The Verify Data Group File Entries (VFYDGFE) command allows you to verify files
from a specific library by verifying the current state of the file on the system identified
in the data group as the source of data.
This procedure generates a report in a spooled file named MXVFYDGFE. The
information in the report includes whether each member for the specified search
criteria is defined to MIMIX, the journal and library to which it is journaled, whether it
uses after-image journaling or before- and after-image journaling, the apply session
used. This information can help you verify that you have all the files you need from a
library properly defined to MIMIX DB Replicator.
To verify data group file entries, do the following:
1. On a command line, type VFYDGFE (Verify Data Group File Entries). The Verify
DG File Entries display appears.
2. Specify the name of the data group at the Data group definition prompt.
3. At the System 1 file and Library prompts, specify the value you want and the
library in which the files are located.
4. If you want to create a spooled file that can be printed, specify *PRINT at the
Output prompt. Then press Enter.

Verifying key attributes


Before you configure for keyed replication, verify that the file or files you for which you
want to use keyed replication are actually eligible.
Do the following to verify that the attributes of a file are appropriate for keyed
replication:
1. On a command line, type VFYKEYATR (Verify Key Attributes). The Verify Key
Attributes display appears.
2. Do one of the following:
• To verify a file in a library, specify a file name and a library.
• To verify all files in a library, specify *ALL and a library.
• To verify files associated with the file entries for a data group, specify
*MIMIXDFN for the File prompt and press Enter. Prompts for the Data group
definition appear. Specify the name of the data group that you want to check.
3. Press Enter.
4. A spooled file is created that indicates whether you can use keyed replication for
the files in the library or data group you specified. Display the spooled file
(WRKSPLF command) or use your standard process for printing. You can use
keyed replication for the file if *BOTH appears in the Replication Type Allowed
column. If a value appears in the Replication Type Defined column, the file is
already defined to the data group with the replication type shown.

351
Working with data group timestamps
Timestamps allow you to view the performance of the database send, receive, and
apply processes for a data group to identify potential problem areas, such as a slow
send process, inadequate communications capacity, or excessive overhead on the
target system. Although they can assist you in identifying problem areas, timestamps
are not intended as an accurate means of calculating the performance of MIMIX.
A timestamp is a single record that is passed between all replication processes. The
timestamp originates on the source system as a journal entry, is sent to the target
system, and then processed by the associated apply session. The timestamp record
is updated with the date and time at each of the following areas during the replication
process:
• Created - Date and time the journal entry is created
• Sent - Date and time when the journal entry is sent to the target system
• Received - Date and time when the journal entry is received
• Applied - Date and time when the journal entry is applied
Note: For data groups that use remote journaling, the created and sent timestamps
will be set to the same value. The received timestamp will be set to the time
when the record was read on the target system by the database reader
process.
After all four timestamps have been added, the journal entry is converted and placed
into a file for viewing or printing. You can view timestamps only from the management
system. The system manager must be active to return the timestamps to the
management system.

Automatically creating timestamps


The data group definition includes a parameter for automatically creating timestamps.
MIMIX automatically creates a timestamp after the number of journal entries specified
in the Timestamp interval (TSPITV) has passed. The timestamp entry created is
placed at the end of all current entries in the journal receiver. You specify this value
when you create or change a data group definition. You can change this value at any
time.
Note: Data groups configured for remote journaling will not automatically generate
timestamps. To generate timestamps in this case, refer to “Creating
timestamps for remote journaling processing” on page 353.

Creating additional timestamps


Note: By using the Create Data Group Timestamps (CRTDGTSP) command in a
batch job, you can use timestamps to monitor performance at critical times in
your daily processing.
To create one or more timestamps, do the following:
1. From the Work with Data Groups display, type 41 (Timestamps) next to the data
group you want and press Enter.

352
Working with data group timestamps

2. The Work with DG Timestamps display appears. Type a 1 (Create) next to the
blank line at the top of the display and press Enter.
3. The Create Data Group Timestamps display appears. Specify the name of the
data group and the number of timestamps you want to create and press Enter.
Note: You should generate multiple timestamps to receive a more accurate view
of replication process performance.

Creating timestamps for remote journaling processing


If you need to generate timestamps to monitor replication performance, you can set
up automation to create them for remote journaling (RJ) data groups that you wish to
monitor.
In this procedure, you will create an interval monitor using the Create Monitor Object
(CRTMONOBJ) command. This is accomplished by specifying *CMD for the interface
exit program on the monitor object, and then Create Data Group Timestamps
(CRTDGTSP) for the command (*CMD). You can also run CRTDGTSP manually or
schedule a job to run the command in batch. For more information, see “Creating an
interval monitor” in the MIMIX Monitor book.
Do the following to create an interval monitor:
1. From the Work with Monitors display, type a 1 (Create) in the Opt column next to
the blank line at the top of the list and press Enter.
2. The Create Monitor Object (CRTMONOBJ) display appears. Do the following:
a. At the Monitor prompt, provide a unique name for the monitor.
b. At the Event class prompt, specify *INTERVAL.
c. At the Interface exit program prompt, specify *CMD.
d. At the Time interval (sec.) prompt, specify how often the interval monitor
should run and press Enter. By default, this monitor runs every 15 seconds.
Use your data group time stamp interval (default is every 20,000 entries) to
estimate how many entries you process a day. From there, determine how
often you need to run the monitor in order to provide an adequate sample.
3. The Add Monitor Information (ADDMONINF) display appears. Do the following:
a. At the Command prompt, type CRTDGTSP.
b. At the Library prompt, type the name of your installation library and press F4
(Prompt).
4. The Create Data Group Timestamps (CRTDGTSP) display appears. Do the
following:
a. At the Data group definition prompts, specify the name of the RJ data group.
b. At the Number of stamps to create prompt, specify the number of timestamps
you want to create and press Enter.
5. From the Work with Monitors display, type a 9 (Start) next to the interval monitor
you created. This allows you to start generating timestamps. For information
about viewing timestamps, see “Displaying or printing timestamps” on page 354.

353
Repeat this procedure for each RJ data group for which you want to generate
timestamps.

Deleting timestamps
You can delete all timestamps or you can select a group of one or more timestamps to
delete.
To delete timestamps for a data group, do the following:
1. From the Work with Data Groups display, type 41 (Timestamps) next to the data
group you want and press Enter.
2. The Work with DG Timestamps display appears. Type a 4 (Delete) next to the
timestamps you want to delete and press Enter.
3. A confirmation screen appears. Press Enter.

To selectively delete a range of timestamps, do the following:


1. Type the command DLTDGTSP and press F4 (Prompt).
2. The Delete Data Group Timestamps display appears. Specify values you want for
the Data group definition prompt.
3. Specify the values you want for the Starting date and time prompt and for the
Ending date and time prompt, then press Enter.

Displaying or printing timestamps


To display or print data group timestamps, do the following:
1. From the Work with Data Groups display, type 41 (Timestamps) next to the data
group you want and press Enter.
2. The Work with DG Timestamps display appears. Do one of the following:
• To display the timestamp information, type a 5 (Display) next to the data group
you want.
• To print the timestamp information, type a 6 (Print) next to the data group you
want.
3. Press Enter.
4. If you selected to display, the Display Data Group Timestamps display appears. If
you selected to print, a spooled file is created that you can print using your
standard printing procedures.

354
Removing journaled changes

Removing journaled changes


If the necessary environment is available, MIMIX can support the Remove Journaled
Changes (RMVJRNCHG) journal entry by simulating the Remove Journaled Changes
process on the backup system.
Note: This is a long running procedure and will affect your existing journal changes.
Ensure that performing this procedure is appropriate for your environment.
In order to use the Remove Journaled Changes journal entry, you must meet the
following criteria:
• You must be configured for both before and after image journaling. This can be
defined as a default file entry option at the data group level or it can be defined for
individual data group file entries.
• You must be configured with *SEND as the value of the Before images element of
the DB journal entry processing (DBJRNPRC) parameter of the data group
definition. This permits the database apply process to roll back certain types of
journal entries.
• If you have large objects (LOBs), *YES must be the value for the Use remote
journal link (RJLNK) parameter of the data group definition.
• The target system (where replicated changes are applied) must have the log
spaces that contain the original transactions. To ensure that the appropriate log
spaces are retained, you can do one of the following:
– Calculate how many log spaces need to be retained using the log space size
and the size and number of the receivers containing the appropriate journal
transactions. Then, set elements of the database apply processing
(DBAPYPRC) parameter in the data group definition.
– Use the Hold Data Group Log (HLDDGLOG) command to place a hold on the
delete operation of all log spaces for all apply sessions defined to the specified
data group. The log spaces are held until a request to release them with
Release Data Group Log (RLSDGLOG) command is received.
If you are changing an existing data group to have these values, you must end and
restart the data group before you are able to use the RMVJRNCHG command.

355
Performing journal analysis
When a source system fails before MIMIX has sent all journal entries to the target
system, unprocessed transactions occur. Unprocessed transactions can also occur if
journal entries are in the communications buffer being sent to the target system when
the sending system fails.
Following an unplanned switch, unprocessed transactions on the original source
system must be addressed in order to prevent data loss before synchronizing data
and starting data groups.
The journal analysis process finds any missing transactions that were not sent to the
target system when the source system went down and an unplanned switch to the
backup was performed. Once unprocessed transactions are located, users must
analyze the journal entries and take appropriate actions to resolve them.
The time at which to perform journal analysis is when the original source system has
been brought back up and before performing the synchronization phase of the switch
(which synchronizes data and starts data groups). Analyze all data groups that were
not disabled at the time of the unplanned switch.
Note: The journal analysis tool is limited to database files replicated from a user
journal. The tool does not identify unprocessed transactions for data areas,
data queues, or IFS objects replicated through a user journal, or database files
configured for replication from the system journal.
From the original source system, do the following:
1. Ensure the following are started:
a. The port communications jobs (PORTxxxxx)
b. The MIMIX system managers using STRMMXMGR SYSDFN(*ALL)
MGR(*SYS) TGTJRNINSP(*NO)
IMPORTANT! Only the system managers should be started at this time. Do not
start journal managers. Also, do not start data groups at this time! Doing so will
delete the data required to perform the journal analysis.
2. From the Work with Data Groups display on the original source system, enter 43
(Journal analysis) next to the data group to be analyzed.
The Journal Analysis of Files display appears.
3. Check the following:
• If a pop-up window with the message “Journal analysis information not
collected” is displayed in the list area, press Enter to collect journal analysis
information, then go to Step 6.
• If there is no pop-up window and message LVI379A is displayed at the bottom
of the display, journal analysis determined that all journal entries have been
applied. There are no unprocessed entries to display. The sequence number of
the last applied journal entry is displayed in the Last applied field. No further
action is needed.

356
Performing journal analysis

• If there is no pop-up window and no message at the bottom of the display,


information about files from a previous run of journal analysis exists, go to
Step 4.
4. If you did not see a pop-up window in Step 3, and information about a previous
run exists, clear data from the previous run of journal analysis and collect new
information by doing the following:
a. If you want to keep information from a previous run, make a copy of file
DM6500P located in the installation library.
b. Press F9 (Update display) to clear the screen and collect the new information.
A pop-up confirmation window with the following message is displayed:
“WARNING! The journal analysis journal entries file will be cleared!”
c. Press Enter to submit the update request.
5. Press Enter to submit the update request.
6. The request to collect journal analysis information is submitted by job RTVFILANZ
using the job description MIMIXQGPL/MIMIXDFT. When the job completes,
“LVI3855 Retrieval of affected files for journal analysis completed normally”
appears in the message log. Press F5 (Refresh) to see the collected information.
It may take a short time to collect the information.
7. Retrieve journal entries. The journal entries for the files identified on the display
must be retrieved before you can use options to display or print statistics (5 and 6)
or display journal entries (11). Do one of the following:
• Press F14 (Retrieve all entries). A pop-up window stating “Confirm retrieval of
ALL analysis journal entries” appears. Press Enter. (The retrieved information
is placed in an internal file.This does not produce a spool file.)
• If there are a large number of files listed on the display, you may want to
retrieve entries for only a selected file at a time. Type option 9 (Retrieve journal
entries) next to the file to retrieve journal entries for and press Enter. The
retrieved journal entries are placed in a spool file named MXJEANZL.
Message: “LVI3856 Retrieval of journal entries for journal analysis completed
normally” appears in the message log.
8. Review the collected information using the following:
• Use option 11 (Display journal entries) to view the entries for each file.
• Use F21 (Print list) to print all entries for a file.
• You can use options 5 (Display statistics) and 6 (Print statistics) to see the
statistical breakdown of journal entries for a selected file member identified by
journal analysis. The statistics include the number of adds, deletes, and
updates, along with the related file transactions and dates of the first and last
journal entries.
Figure 38 shows an example of the information displayed by option 11 for one
journal entry.

357
Figure 38. Sample of one journal entry

Data group definition: <DGDFN> <SYS1> <SYS2>


Journal definition: <JRNDFN> <SYSDFN>

File identification
File . . . . : <FILE>
Library . : <LIB>
Member . . . : <MBR>

Journal header information


Journal code . . . . . : R Record-level information
Journal type . . . . . : DL Delete record
Generated date . . . . : 9/08/09
Generated time . . . . : 10:36:31
Job name . . . . . . . : <JOB NAME>
User name . . . . . . : <USER>
Job number . . . . . . : <JOB NBR>
Program name . . . . . : <PROGRAM>

Journal header information (continued)


Record length . . . . : 607
Record number . . . . : 838
Operation indicator . : 0
Commit cycle ID . . . : 0

Journal identification
Journal name . . . . . : <JOURNAL>
Library . . . . . . : <JRNLIB>

Receiver identification
Receiver name . . . . : <RCVR>
Library . . . . . . : <RCVRLIB>
Sequence number . . . : <JOURNAL SEQUENCE>

9. Determine what action you need to take for each unprocessed entry. For example:
• You may need to run the original job again on the current source system to
reproduce the entries.
• If a file has already been updated on the current source system (manually or
otherwise), you may need to merge data from both files. If this is the case, do
not synchronize the files.
• If there are write changes (R-PT entries), these changes should be made on
the current source system before running the synchronization phase of the
switch or starting data groups in order to maintain Relative Record Number
consistency within the file. If this is done after the data group has been started,
the relative record numbers could become unsynchronized between the two
systems.
Note: It is the customer’s responsibility to fix the files.

Removing journal analysis entries for a selected file


You can use option 4 (Remove journal entries) to remove all journal analysis journal
entries for a selected file member. A confirmation display appears to confirm your
choices. When you continue with the confirmation, the journal entries for the selected

358
Performing journal analysis

file member are immediately removed from the journal analysis information that is
displayed. It does not delete any other information contained in other MIMIX files.

359
Interpreting audit results - supporting information

Interpreting audit results -


APPENDIX A
supporting information

Audits use commands that compare and synchronize data. The results of the audits
are placed in output files associated with the commands. The following topics provide
supporting information for interpreting data returned in the output files.
• “When the difference is “not found”” on page 363 provides additional
considerations for interpreting result of not found in priority audits.
• “Interpreting results for configuration data - #DGFE audit” on page 361 describes
the #DGFE audit which verifies the configuration data defined to your
configuration using the Check Data Group File Entries (CHKDGFE) command.
• “Interpreting results of audits for record counts and file data” on page 364
describes the audits and commands that compare file data or record counts.
• “Interpreting results of audits that compare attributes” on page 367 describes the
Compare Attributes commands and their results.

360
Interpreting results for configuration data - #DGFE audit

Interpreting results for configuration data - #DGFE audit


The #DGFE audit verifies the configuration data that is defined for replication in your
configuration. This audit invokes the Check Data Group File Entries (CHKDGFE)
command for the audit’s comparison phase. The CHKDGFE command collects data
on the source system and generates a report in a spooled file or an outfile.
The report is available on the system where the #DGFE audit or CHKDGFE
command ran. In the report, the values in the Result column indicate detected
problems and the result of any automated audit recoveries.
Table 55 shows the possible Result values and corresponding recovery actions. If the
Automatic audit recovery policy is enabled at the time the audit runs, the audit will
attempt to perform the recovery action indicated and the status will become either
*RECOVERED or *RCYFAILED. If the recovery action could not be performed or if
the Automatic audit recovery policy is disabled, user action is needed to manually
perform the actions needed to correct the reported problem.

Table 55. CHKDGFE - possible results and actions for resolving errors

Result Description and Recovery Actions

*NODGFE No file entry exists. The object is identified by an object entry which
specifies COOPDB(*YES) but the file entry necessary for cooperative
processing is missing.
Create the DGFE or change the DGOBJE to COOPDB(*NO)
Note: Changing the object entry affects all objects using the object entry. If you
do not want all objects changed to this value, copy the existing DGOBJE
to a new, specific DGOBJE with the appropriate COOPDB value.

*EXTRADGFE An extra file entry exists. The object is identified by a file entry and an
object entry which specifies COOPDB(*NO). The file entry is extra when
cooperative processing is not used.
Delete the DGFE or change the DGOBJE to COOPDB(*YES)
Note: Changing the object entry affects all objects using the object entry. If you
do not want all objects changed to this value, copy the existing DGOBJE
to a new, specific DGOBJE with the appropriate COOPDB value.

*NOFILE No file exists for the existing file entry.


Delete the DGFE, re-create the missing file, or restore the missing file.

*NOMBR No file member exists for the existing file entry.


Delete the DGFE for the member or add the member to the file.

*RCYFAILED Automatic audit recovery actions were attempted but failed to correct the
detected error.
Run the audit again.

*RECOVERED Recovered by automatic recovery actions.


No action is needed.

*UA File entries are in transition and cannot be compared.


Run the audit again.

361
The Option column of the report provides supplemental information about the
comparison. Possible values are:
*NONE - No options were specified on the comparison request.
*NOFILECHK - The comparison request included an option that prevented an
error from being reported when a file specified in a data group file entry does not
exist.
*DGFESYNC - The data group file entry was not synchronized between the
source and target systems. This may have been resolved by automatic recovery
actions for the audit.
One possible reason why actual configuration data in your environment may not
match what is defined to your configuration is that a file was deleted but the
associated data group file entries were left intact. Another reason is that a data group
file entry was specified with a member name, but a member is no longer defined to
that file. If you use #DGFE audit with automatic scheduling and automatic audit
recovery enabled, these configuration problems can be automatically detected and
recovered for you. Table 56 provides examples of when various configuration errors
might occur.

Table 56. CHKDGFE - possible error conditions

Result File Member DGFE DGOBJE exists


exists exists exists

*NODGFE Yes Yes No COOPDB(*YES)

*EXTRADGFE Yes Yes Yes COOPDB(*NO)

*NOFILE No No Yes Exclude

*NOMBR Yes No Yes No entry

362
When the difference is “not found”

When the difference is “not found”


For audits that compare replicated data, a difference indicating the object was not
found requires additional explanation. This difference can be returned for these
audits:
• For the #FILDTA and #MBRRCDCNT audits, a value of *NF1 or *NF2 for the
difference indicator (DIFIND) indicates the object was not found on one of the
systems in the data group. The 1 and 2 in these values refer to the system as
identified in the three-part name of the data group.
• For the #FILATR, #FILATRMBR, #IFSATR, #OBJATR, and #DLOATR audits, a
not found condition is indicated by a value of *NOTFOUND in either the system 1
indicator (SYS1IND) or system 2 indicator (SYS2IND) fields. Typically, the DIFIND
field result is *NE.
Audits can report not found conditions for objects that have been deleted from the
source system. A not found condition is reported when a delete transaction is in
progress for an object eligible for selection when the audit runs. This is more likely to
occur when there are replication errors or backlogs, and when policy settings do not
prevent audits from comparing when a data group is inactive or in a threshold
condition.
A scheduled audit will not identify a not found condition for an object that does not
exist on either system because it selects existing objects based on whether they are
configured for replication by the data group. This is true regardless of whether the
audit is automatically submitted or run immediately.
Because a priority audit selects already replicated objects, it will not audit objects for
which a create transaction is in progress.
Prioritized audits will not identify a not found condition when the object is not found on
the target system because prioritized auditing selects objects based on the replicated
objects database. Only objects that have been replicated to the target system are
identified in the database.
Priority audits can be more likely to report not found conditions when replication errors
or backlogs exist.

363
Interpreting results of audits for record counts and file
data
The audits and commands that compare file data or record counts are as follows:
• #FILDTA audit or Compare File Data (CMPFILDTA) command
• #MBRRCDCNT audit or Compare Record Count (CMPRCDCNT) command
Each record in the output files for these audits or commands identifies a file member
that has been compared and indicates whether a difference was detected for that
member.

What differences were detected by #FILDTA


The Difference Indicator (DIFIND) field identifies the result of the comparison. Table
57 identifies values for the Compare File Data command that can appear in this field

Table 57. Possible values for Compare File Data (CMPFILDTA) output file field Difference
Indicator (DIFIND)

Values Description

*APY The database apply (DBAPY) job encountered a problem


processing a U-MX journal entry for this member.

*CMT Commit cycle activity on the source system prevents active


processing from comparing records or record counts in the
selected member.

*CO Unable to process selected member. Cannot open file.

*CO (LOB) Unable to process selected member containing a large object


(LOB). The file or the MIMIX-created SQL view cannot be opened.

*DT Unable to process selected member. The file uses an unsupported


data type.

*EQ Data matches. No differences were detected within the data


compared. Global difference indicator.

*EQ (DATE) Member excluded from comparison because it was not changed or
restored after the timestamp specified for the CHGDATE
parameter.

*EQ (OMIT) No difference was detected. However, fields with unsupported


types were omitted.

*FF The file feature is not supported for comparison. Examples of file
features include materialized query tables.

*FMC Matching entry not found in database apply table.

*FMT Unable to process selected member. File formats differ between


source and target files. Either the record length or the null
capability is different.

364
Interpreting results of audits for record counts and file data

Table 57. Possible values for Compare File Data (CMPFILDTA) output file field Difference
Indicator (DIFIND)

Values Description

*HLD Indicates that a member is held or an inactive state was detected.

*IOERR Unable to complete processing on selected member. Messages


preceding LVE0101 may be helpful.

*NE Indicates a difference was detected.

*NF1 Member not found on system 1.

*NF2 Member not found on system 2.

*REP The file member is being processed for repair by another job
running the Compare File Data (CMPFILDTA) command.

*SJ The file is not journaled on the source node.

*SP Unable to process selected member. See messages preceding


message LVE3D42 in job log.

*SW The source file is journaled but not to the journal specified in the
journal definition.

*SYNC The file or member is being processed by the Synchronize DG File


Entry (SYNCDGFE) command.

*UE Unable to process selected member. Reason unknown. Messages


preceding message LVE3D42 in job log may be helpful.

*UN Indicates that the member’s synchronization status is unknown.

See “When the difference is “not found”” on page 363 for additional information.

What differences were detected by #MBRRCDCNT


Table 58 identifies values for the Compare Record Count command that can appear
in the Difference Indicator (DIFIND) field.

Table 58. Possible values for Compare Record Count (CMPRCDCNT) output file field Dif-
ference Indicator (DIFIND)

Values Description

*APY The database apply (DBAPY) job encountered a problem


processing a U-MX journal entry for this member.

*CMT Commit cycle activity on the source system prevents active


processing from comparing records or record counts in the
selected member.

*EC The attribute compared is equal to configuration

365
Table 58. Possible values for Compare Record Count (CMPRCDCNT) output file field Dif-
ference Indicator (DIFIND)

Values Description

*EQ Record counts match. No difference was detected within the record
counts compared. Global difference indicator.

*FF The file feature is not supported for comparison. Examples of file
features include materialized query tables.

*FMC Matching entry not found in database apply table.

*HLD Indicates that a member is held or an inactive state was detected.

*LCK Lock prevented access to member.

*NE Indicates a difference was detected.

*NF1 Member not found on system 1.

*NF2 Member not found on system 2.

*SJ The file is not journaled on the source node.

*SW The source file is journaled but not to the journal specified in the
journal definition.

*UE Unable to process selected member. Reason unknown. Messages


preceding LVE3D42 in job log may be helpful.

*UN Indicates that the member’s synchronization status is unknown.

See “When the difference is “not found”” on page 363 for additional information.

366
Interpreting results of audits that compare attributes

Interpreting results of audits that compare attributes


Each audit that compares attributes does so by calling a Compare Attributes1
command and places the results in an output file. Each row in an output file for a
Compare Attributes command can contain either a summary record format or a
detailed record format. Each summary row identifies a compared object and includes
a prioritized object-level summary of whether differences were detected. Each detail
row identifies a specific attribute compared for an object and the comparison results.
For example, an authorization list can contain a variable number of entries. When
comparing authorization lists, the CMPOBJA command will first determine if both lists
have the same number of entries. If the same number of entries exist, it will then
determine whether both lists contain the same entries. If differences in the number of
entries are found or if the entries within the authorization list are not equal, the report
will indicate that differences are detected. The report will not provide the list of
entries—it will only indicate that they are not equal in terms of count or content.
You can see the full set of fields in the output file by viewing it from the native user
interface.

What attribute differences were detected


The Difference Indicator (DIFIND) field identifies the result of the comparison. Table
59 identifies values that can appear in this field. Not all values may be valid for every
Compare command.
When the output file is viewed from the native user interface, the summary row is the
first record for each compared object and is indicated by an asterisk (*) in the
Compared Attribute (CMPATR) field. The summary row’s Difference Indicator value is
the prioritized summary of the status of all attributes checked for the object. When
included, detail rows appear below the summary row for the object compared and
show the actual result for the attributes compared.
The Priority2 column in Table 59 indicates the order of precedence MIMIX uses when
determining the prioritized summary value for the compared object.

Table 59. Possible values for output file field Difference Indicator (DIFIND)

Values1 Description Summary


Record2 Priority

*EC The values are based on the MIMIX configuration settings. The 5
actual values may or may not be equal.

*EQ Record counts match. No differences were detected. Global 5


difference indicator.

*EM Established mapping for file identifier (FID). Attribute indicator only 6
for CMPIFSA *FID attribute.

1. The Compare Attribute commands are: Compare File Attributes (CMPFILA), Compare
Object Attributes (CMPOBJA), Compare IFS Attributes (CMPIFSA), and Compare DLO Attri-
butes (CMPDLOA).

367
Table 59. Possible values for output file field Difference Indicator (DIFIND)

Values1 Description Summary


Record2 Priority

*NA The values are not compared. The actual values may or may not 5
be equal.

*NC The values are not equal based on the MIMIX configuration 3
settings. The actual values may or may not be equal.

*NE Indicates differences were detected. 2

*NS Indicates that the attribute is not supported on one of the systems. 5
Will not cause a global not equal condition.

*NM Not mapping consistently for file identifier (FID). Attribute indicator 2
only for CMPIFSA *FID attribute.

*RCYSBM Indicates that MIMIX submitted an automatic audit recovery action


that must be processed through the user journal replication
processes. The database apply (DBAPY) will attempt the recovery
and send an *ERROR or *INFO notification to indicate the outcome
of the recovery attempt.

*RCYFAILED Used to indicate that automatic recovery attempts via automatically


submitted audits failed to recover the detected difference.

*RECOVERED Indicates that recovery for this object was successful. 13

*UA Object status is unknown due to object activity. If an object 2


difference is found and the comparison has a value specified on
the Maximum replication lag prompt, the difference is seen as
unknown due to object activity. This status is only displayed in the
summary record.
Note: The Maximum replication lag prompt is only valid when a data
group is specified on the command.

*UN Indicates that the object’s synchronization status is unknown. 4


1. Not all values may be possible for every Compare command.
2. Priorities are used to determine the value shown in output files for Compare Attribute commands.
3. The value *RECOVERED can only appear in an output file modified by a recovery action. The object was initially
found to be *NM, *NE or *NC but MIMIX automatic recovery actions recovered the object.

For most attributes, when the outfile is viewed from the native user interface, when a
detailed row contains blanks in either of the System 1 Indicator or System 2 Indicator
fields, MIMIX determines the value of the Difference Indicator field according to Table

368
Interpreting results of audits that compare attributes

60. For example, if the System 1 Indicator is *NOTFOUND and the System 2 Indicator
is blank (Object found), the resultant Difference Indicator is *NE.

Table 60. Difference Indicator values that are derived from System Indicator values.

Difference Indicator
System 1 Indicator
Object *NOTCMPD *NOTFOUND *NOTSPT *RTVFAILED *DAMAGED
Found (blank
value)
Object Found *EQ / *NE / *NA *NE *NS *UN *NE
(blank value) *UA / *EC /
*NC
System *NOTCMPD *NA *NA *NE *NS *UN *NE
2 *NOTFOUND *NE / *UA *NE / *UA *EQ *NE / *UA *NE / *UA *NE
Indicator
*NOTSPT *NS *NS *NE *NS *UN *NE
*RTVFAILED *UN *UN *NE *UN *UN *NE
*DAMAGED *NE *NE *NE *NE *NE *NE

When viewed through Vision Solutions Portal, data group directionality is


automatically resolved so that differences are viewed as Source and Target instead of
System1 and System2.
For a small number of specific attributes, the comparison is more complex. The
results returned vary according to parameters specified on the compare request and
MIMIX configuration values. For more information about comparison results for
journal status and other journal attributes, auxiliary storage pool ID (*ASP), user
profile status (*USRPRFSTS), and user profile password (*PRFPWDIND) see the see
the MIMIX Administrator Reference book.

Where was the difference detected


The System 1 Indicator (SYS1IND) and System 2 (SYS2IND) fields show the status
of the attribute on each system as determined by the compare request. Table 61
identifies the possible values. These fields are available in both summary and detail
rows in the output file.

Table 61. Possible values for output file fields SYS1IND and SYS2IND

Value Description Summary Record1


Priority

<blank> No special conditions exist for this object. 5

*DAMAGED Object damaged condition. 3

*MBRNOTFND Member not found. 2

*NOTCMPD Attribute not compared. Due to MIMIX configuration settings, this N/A2
attribute cannot be compared.

369
Table 61. Possible values for output file fields SYS1IND and SYS2IND

Value Description Summary Record1


Priority

*NOTFOUND Object not found. 1

*NOTSPT Attribute not supported. Not all attributes are supported on all IBM N/A2
i releases. This is the value that is used to indicate an
unsupported attribute has been specified.

*RTVFAILED Unable to retrieve the attributes of the object. Reason for failure 4
may be a lock condition.
1. The priority indicates the order of precedence MIMIX uses when setting the system indicators fields in the summary
record.
2. This value is not used in determining the priority of summary level records.

For comparisons which include a data group, the Data Source (DTASRC) field
identifies which system is configured as the source for replication.

What attributes were compared


In each detailed row, the Compared Attribute (CMPATR) field identifies a compared
attribute. For more information about identifying attributes that can be compared by
each command and the possible values returned, see the MIMIX Administrator
Reference book.

370
APPENDIX B IBM Power™ Systems operations
that affect MIMIX

The following topics describe how to protect the integrity of your MIMIX environment
when you perform operations such as IPLs and IBM i operating system upgrades.
Only basic procedures for a standard one-to-one MIMIX installation are covered. If
you are operating in a complex environment—if you have cluster, SAP R/3, IBM
WebSphere MQ, or other application considerations, for example—contact your
Certified MIMIX Consultant. Ultimately, you must tailor these procedures to suit the
needs of your particular environment.
These topics describe MIMIX-specific steps only. Refer to the user manuals that
correspond to any additional applications installed in your environment. For
instructions on performing IBM Power™ Systems operations, consult your IBM
manuals or the IBM Information Center at
http://publib.boulder.ibm.com/pubs/html/as400/infocenter.html.
The following topics are included:
• “MIMIX procedures when performing an initial program load (IPL)” on page 372
includes the MIMIX-specific steps for performing an initial program load (IPL) to
help ensure the integrity of your MIMIX environment is not compromised.
• “MIMIX procedures when performing an operating system upgrade” on page 374
describes when and how to perform recommended MIMIX-specific steps while
performing a standard upgrade of IBM i.
• “MIMIX procedures when upgrading hardware without a disk image change” on
page 382 describes MIMIX prerequisites and procedures for performing a
hardware upgrade without a disk image change.
• “MIMIX procedures when performing a hardware upgrade with a disk image
change” on page 385 describes prerequisites for saving and restoring MIMIX
software when upgrading from one system to another.
• “Handling MIMIX during a system restore” on page 391 includes prerequisites for
restoring MIMIX software within a MIMIX system pair, to one system from a save
of the other system when an environment meets the conditions specified.

371
IBM Power™ Systems operations that affect MIMIX

MIMIX procedures when performing an initial program


load (IPL)
An initial program load (IPL) loads the operating system and prepares the system for
user operations. Performing the recommended MIMIX-specific steps can help ensure
that objects are not damaged during the IPL and that the integrity of your MIMIX
environment is not compromised.
Note: This procedure describes an IPL performed under normal circumstances.
It does not address IPL considerations for system switching environments.
Before beginning this procedure, review your startup procedures to determine
whether subsystems will start after the IPL. This startup program is defined in the
QSTRUPPGM system value.
To perform an IPL in a MIMIX environment, do the following:
1. End MIMIX including the MIMIX managers and end the RJ links using the
following command:
ENDMMX ENDOPT(*IMMED) ENDRJLNK(*YES) SYSDFN(*LOCAL)
ENDSMRJ(*YES)
For more information about the ENDMMX command, see “Commands for ending
replication” on page 232.
2. If the VSP server is running on the system you are about to IPL, use the following
command to end the VSP server:
VSI001LIB/ENDVSISVR
3. Ensure that all MIMIX jobs are ended before performing this step. End the
MIMIX subsystem on the system you are about to IPL. Type the following on a
command line and press Enter:
ENDSBS SBS(MIMIXSBS) OPTION(*IMMED)
Note: If you are running VSP in the MIMIX subsystem for any product, be aware
that all VSP processes will end.
4. Perform the IPL.
5. If your subsystems do not start during the startup procedures defined in the
QSTRUPPGM system value, start the MIMIX subsystems on both the source and
target systems. On each system, type the following on a command line and press
Enter:
STRSBS SBSD(MIMIXQGPL/MIMIXSBS)
6. Verify the communication links start, using the Verify Communications Link
(VFYCMNLNK) command. For more information about the VFYCMNLNK
command, see “Verifying a communications link for system definitions” on
page 343.
If the communications link is not active, you may need to start the port job. On
a command line type the following and press Enter:
STRSVR HOST(host-name-or-address) PORT(port-number)

372
MIMIX procedures when performing an initial program load (IPL)

7. Start MIMIX from either the source or target system. The Start MIMIX (STRMMX)
command starts the MIMIX processes for the installation, including the MIMIX
managers and the data groups.
For more information about the STRMMX command, see “Starting MIMIX” on
page 226.
8. If you ended the VSP server, restart it using the following command:
VSI001LIB/STRVSISVR

373
IBM Power™ Systems operations that affect MIMIX

MIMIX procedures when performing an operating system


upgrade
This topic describes when and how to perform recommended MIMIX-specific steps
while performing a standard upgrade of the IBM i operating system (slip-install, where
the IBM i release is upgraded without a restore of the user libraries). Performing these
recommended steps can help ensure that MIMIX products start properly once the
operating system upgrade is complete.
Table 62 indicates which procedures are needed for different upgrade scenarios. Use
these instructions in conjunction with the instructions provided by IBM for upgrading
from one IBM i release to another IBM i release.

Table 62. IBM i operating system upgrade scenarios and recommended processes for handling MIMIX
during the upgrade

To upgrade Perform these procedures

Backup system only 1. Perform the preparation steps described in “Prerequisites for performing
an OS upgrade on either system” on page 375.
2. Follow the procedure in “MIMIX-specific steps for an OS upgrade on a
backup system” on page 375.

Production system only 1. Perform the preparation steps described in “Prerequisites for performing
an OS upgrade on either system” on page 375.
2. Perform one of the following procedures:
• If you need to maintain user access to production applications during the
upgrade, perform a planned switch as described in “MIMIX-specific steps
for an OS upgrade on a production system with switching” on page 377.
Your production operations will be temporarily running on the backup
system.
• If you have more flexibility with scheduling downtime, you can perform the
upgrade without switching as described in “MIMIX-specific steps for an
OS upgrade on the production system without switching” on page 379.

Both backup and production 1. Perform the preparation steps described in “Prerequisites for performing
systems an OS upgrade on either system” on page 375.
2. Upgrade the backup system first following the “MIMIX-specific steps for
an OS upgrade on a backup system” on page 375. By doing this first, you
can ensure that the backup system supports all the capabilities of the
production system and you can work through problems or custom
operations before affecting your production environment.
3. Once you have the verified that the backup system is upgraded and
operating as desired, perform one of the following procedures to upgrade
IBM i on the production system:
• If you need to maintain user access to production applications during the
upgrade, perform a planned switch as described in “MIMIX-specific steps
for an OS upgrade on a production system with switching” on page 377.
Your production operations will be temporarily running on the backup
system.
• If you have more flexibility with scheduling downtime, you can perform the
upgrade without switching as described in “MIMIX-specific steps for an
OS upgrade on the production system without switching” on page 379

374
MIMIX procedures when performing an operating system upgrade

Prerequisites for performing an OS upgrade on either system


Before you start an upgrade of the IBM i operating system on either system, do the
following:
1. Access Support information on the web as you perform the following steps to
ensure that the system is ready to upgrade:
a. Check the compatibility of the operating systems on the production and backup
systems, ensuring the systems will meet the requirements of a MIMIX-
supported environment once the IBM i operating system upgrade has
occurred.
b. Ensure the recommended IBM IBM i PTFs have been applied according to
your IBM i version.
c. Ensure the recommended MIMIX service packs have been applied according
to your MIMIX version. Review the Read Me document that corresponds to the
MIMIX service pack, and check the website for relevant Technical Alerts and
FAQs.
2. Review your startup procedures to understand how your environment is
configured to start after an IPL. This startup program is defined in the
QSTRUPPGM system value. An IBM i upgrade may include rebuilding access
paths, converting formats, or performing other operations that must be complete
before MIMIX or other applications are started. The upgrade may not complete
successfully if your QSTRUPPGM procedures start MIMIX or other applications
during an IPL. Ensure that these processes are disabled before continuing with
the IBM i upgrade.

MIMIX-specific steps for an OS upgrade on a backup system


Use this procedure to upgrade the operating system on the backup system.
Notes:
• If you plan to upgrade both the production and backup systems during the same
scheduled maintenance period, upgrade the backup system first.
• In the following steps, the terms production and backup always refer to the original
roles of the systems before upgrading the operating system on either system. The
icons at the beginning of some steps show the state of the systems and
replication as a result of the action in the step. The arrow in the icon indicates the
direction and state of replication for a classic production to backup environment.
When performing an operating system upgrade of a backup system in a MIMIX
environment, do the following:
1. Ensure that you have completed any prerequisite tasks for your upgrade scenario.
See Table 62 for a list of required tasks for different upgrade scenarios.
2. End all user applications, user interfaces, and operations actively running on the
backup system.
3. End MIMIX including the MIMIX managers and end the RJ links using the
following command:

375
IBM Power™ Systems operations that affect MIMIX

ENDMMX ENDOPT(*IMMED) ENDRJLNK(*YES) SYSDFN(*LOCAL)


ENDSMRJ(*YES)
For more information about the ENDMMX command, see “Commands for ending
replication” on page 232.
4. If the VSP server is running on the system you are about to upgrade, use the
following command to end the VSP server:
VSI001LIB/ENDVSISVR
5. Ensure that all MIMIX jobs are ended before performing this step. End the
MIMIX subsystem on the system you are about to upgrade. Type the following on
a command line and press Enter:
ENDSBS SBS(MIMIXSBS) OPTION(*IMMED)
Note: If you are running VSP in the MIMIX subsystem for any product, be aware
that all VSP processes will end.
6. Complete the operating system upgrade. Allow any upgrade conversions and
access path rebuilds to complete before continuing with the next step.
Note: During the IBM i upgrade, make sure you perform a system save on the
system being upgraded. This step will provide you with a backup of
existing data.
7. Ensure the names of the journal receivers match the journal definitions:
a. From the backup system, specify the command:
installation-name/WRKJRNDFN JRNDFN(QAUDJRN *LOCAL)
b. Next to the JRNDFN(QAUDJRN *LOCAL) journal definition, specify 14 (Build)
and press F4. Type *JRNDFN for the Source for values parameter and press
Enter.
8. Start the MIMIX subsystems on both the production and backup systems using
the following command from each system:
STRSBS SBSD(MIMIXQGPL/MIMIXSBS)
9. Verify the communication links start, using the Verify Communications Link
(VFYCMNLNK) command. For more information about the VFYCMNLNK
command, see “Verifying a communications link for system definitions” on
page 343.
If the communications link is not active, you may need to start the port job. On
a command line type the following and press Enter:
STRSVR HOST(host-name-or-address) PORT(port-number)
10. Perform a normal start of data groups and application groups from any system
using the STRMMX command. This step also starts the MIMIX managers.
11. Perform your normal process for validating the IBM i release upgrade.
Notes:
• At your convenience, schedule a switch to verify that your applications function on
the new operating system on the backup system.
• MIMIX supports replication for up to two version level differences. If running

376
MIMIX procedures when performing an operating system upgrade

different OS versions, Vision Solutions recommends that the backup node run the
higher OS. The following restrictions and limitations may also apply:
• All objects must be compiled using the Target Release parameter to specify
the IBM i version of the lower level operating system.
• Possible inherent restrictions include those specific to new functionality in
the OS. There may be features in a higher OS release that would not be
available in a lower one, such as new command parameters or APIs.
• Some errors may be encountered during or following a role swap due to the
different OS versions.
12. If you ended the VSP server, restart it using the following command:
VSI001LIB/STRVSISVR

MIMIX-specific steps for an OS upgrade on a production system with


switching
Use this procedure if you need to maintain user access to production applications
during the production system upgrade. This procedure temporarily switches
production activity to the backup system before the upgrade and switches back to
normal operations after the production system upgrade is complete.
Notes:
• In the following steps, the terms production and backup always refer to the original
roles of the systems before upgrading the operating system on either system. The
icons at the beginning of some steps show the state of the systems and
replication as a result of the action in the step. The arrow in the icon indicates the
direction and state of replication for a classic production to backup environment.
• MIMIX supports replication for up to two version level differences. If running
different OS versions, Vision Solutions recommends that the backup node run the
higher OS. The following restrictions and limitations may also apply:
• All objects must be compiled using the Target Release parameter to specify
the IBM i version of the lower level operating system.
• Possible inherent restrictions include those specific to new functionality in
the OS. There may be features in a higher OS release that would not be
available in a lower one, such as new command parameters or APIs.
• Some errors may be encountered during or following a role swap due to the
different OS versions.
To perform an operating system upgrade of the production system in a MIMIX
environment while maintaining availability, do the following:
1. Ensure that you have completed any prerequisite tasks for your upgrade scenario.
See Table 62 for a list of required tasks for different upgrade scenarios.
2. If applicable, disable auditing, including prioritized audits to avoid having them
start before ending MIMIX or immediately after re-starting MIMIX during the
upgrade. For instructions to disable audits, see “Preventing audits from running”
on page 46.

377
IBM Power™ Systems operations that affect MIMIX

3. Use the procedures in your Runbook to perform a planned switch to the


backup system.
Note: Do not perform steps to synchronize data and start replication from the
backup system to the original production system.
If you do not have a Runbook, you need to follow your processes for the following:
a. End all user applications, user interfaces, and operations actively running on
the production system.
b. Perform a planned switch to the backup system.
c. Start user applications on the backup system and allow users to access their
applications from the backup system.
4. End MIMIX, including the MIMIX managers, and end the RJ links using the
following command:
ENDMMX ENDOPT(*IMMED) ENDRJLNK(*YES) SYSDFN(*LOCAL)
ENDSMRJ(*YES)
For more information about the ENDMMX command, see “Commands for ending
replication” on page 232.
5. If the VSP server is running on the system you are about to upgrade, use the
following command to end the VSP server:
VSI001LIB/ENDVSISVR
6. Ensure that all MIMIX jobs are ended before performing this step. End the
MIMIX subsystem on the system you are about to upgrade. Type the following on
a command line and press Enter:
ENDSBS SBS(MIMIXSBS) OPTION(*IMMED)
Note: If you are running VSP in the MIMIX subsystem for any product, be aware
that all VSP processes will end.
7. On the production system, complete the operating system upgrade. Allow any
upgrade conversions and access path rebuilds to complete before continuing with
the next step.
Note: During the IBM i upgrade, make sure you perform a system save on the
system being upgraded. This step will provide you with a backup of
existing data.
8. Ensure the names of the journal receivers match the journal definitions:
a. From the original production system, specify the command:
installation-name/WRKJRNDFN JRNDFN(QAUDJRN *LOCAL)
b. Next to the JRNDFN(QAUDJRN *LOCAL) journal definition, specify 14 (Build)
and press F4. Type *JRNDFN for the Source for values parameter and press
Enter.
9. Start the MIMIX subsystem using the following command:
STRSBS SBSD(MIMIXQGPL/MIMIXSBS)
10. Verify the communication links start, using the Verify Communications Link
(VFYCMNLNK) command. For more information about the VFYCMNLNK

378
MIMIX procedures when performing an operating system upgrade

command, see “Verifying a communications link for system definitions” on


page 343.
If the communications link is not active, you may need to start the port job. On
a command line type the following and press Enter:
STRSVR HOST(host-name-or-address) PORT(port-number)
11. Start the MIMIX managers and collector services with the following command:
STRMMXMGR SYSDFN(*ALL) MGR(*SYS) COLSRV(*YES)
12. If applicable, start the VSP server using the command:
VSI001LIB/STRVSISVR
13. Follow your Runbook procedures to start replication (sync). If you do not
have a Runbook, follow your processes for starting data groups or application
groups.
14. Ensure that no backlog exists. See “Identifying replication processes with
backlogs” on page 137.
15. Re-enable all audits.
16. When you are ready to switch back to the production system and start
replication, follow your Runbook procedures. If you do not have a Runbook,
follow your processes to switch replication so that you return to your normal
replication environment.

MIMIX-specific steps for an OS upgrade on the production system with-


out switching
Use this procedure if you have more flexibility with scheduling downtime and can
perform the upgrade without switching.
Note: In the following steps, the terms production and backup always refer to the
original roles of the systems before upgrading the operating system on either
system. The icons at the beginning of some steps show the state of the
systems and replication as a result of the action in the step. The arrow in the
icon indicates the direction and state of replication for a classic production to
backup environment.
To perform an operating system upgrade of the production system in a MIMIX
environment without switching, do the following:
1. Ensure that you have completed any prerequisite tasks for your upgrade scenario.
See Table 62 for a list of required tasks for different upgrade scenarios.
2. End all user applications, user interfaces, and operations actively running on the
production system.
For more information, refer to your Runbook and your applications’ user manuals.
3. End the data groups from either system using the command:
ENDDG DGDFN(*ALL) ENDOPT(*CNTRLD)
For more information about ending data groups see “Commands for ending
replication” on page 232.

379
IBM Power™ Systems operations that affect MIMIX

4. Wait until the status of each data group becomes inactive (red) by monitoring the
status on the Work with Data Groups (WRKDG) display.
For more information about the WRKDG display, see “Displaying data group
status” on page 118.
5. If you have applications that use commitment control, ensure there are no open
commit cycles. For more information, see “Checking for open commit cycles” on
page 230.
If an open commit cycle exist, restart the data group and repeat Step 3, Step 4,
and Step 5 until there is no open commit cycle for any apply session.
6. Use the following command to end other MIMIX products in the installation library,
end the MIMIX managers, and end the RJ links:
ENDMMX ENDOPT(*CNTRLD) ENDRJLNK(*YES) ENDSMRJ(*YES)
7. End the MIMIX subsystems on the production system and on the backup system.
On each system, type the following on a command line and press Enter:
ENDSBS SBS(MIMIXSBS) OPTION(*IMMED)
8. Complete the operating system upgrade. Allow any upgrade conversions and
access path rebuilds to complete before continuing with the next step.
Note: During the IBM i upgrade, make sure you perform a system save on the
system being upgraded. This step will provide you with a backup of
existing data.
9. Start the MIMIX subsystems on the production system and the backup system as
you would during the synchronization phase of a switch. From each system, type
the following on a command line and press Enter:
STRSBS SBSD(MIMIXQGPL/MIMIXSBS)
10. Ensure the names of the journal receivers match the journal definitions:
a. From the production system, specify the command:
installation-name/WRKJRNDFN JRNDFN(QAUDJRN *LOCAL)
b. Next to the JRNDFN(QAUDJRN *LOCAL) journal definition, specify 14 (Build)
and press F4. Type *JRNDFN for the Source for values parameter and press
Enter.
c. Record the newly attached journal receiver name by placing the cursor on the
posted message and pressing F1 or Help.
11. Using the information you gathered in Step 10, start each data group as follows
(This step also starts the MIMIX managers.):
a. From the WRKDG display, type an 9 (Start DG) next to the data group and
press Enter. The Start Data Group display appears.
b. At the Object journal receiver prompt, specify the receiver name recorded in
Step 10c.
c. At the Object large sequence number prompt, specify *FIRST.
d. At the Clear pending prompt, specify *YES.

380
MIMIX procedures when performing an operating system upgrade

12. Start any applications that you disabled prior to completing the IBM i upgrade
according to your Runbook instructions. These applications are normally started
in the program defined in the QSTRUPPGM system value. Allow users back on
the production system.
Note: MIMIX supports replication for up to two version level differences. If running
different OS versions, Vision Solutions recommends that the backup node run
the higher OS. The following restrictions and limitations may also apply:
• All objects must be compiled using the Target Release parameter to specify
the IBM i version of the lower level operating system.
• Possible inherent restrictions include those specific to new functionality in
the OS. There may be features in a higher OS release that would not be
available in a lower one, such as new command parameters or APIs.
• Some errors may be encountered during or following a role swap due to the
different OS versions.

381
IBM Power™ Systems operations that affect MIMIX

MIMIX procedures when upgrading hardware without a


disk image change
This topic describes MIMIX prerequisites and procedures for a hardware upgrade
without a disk image change that will change a model, feature, or serial number and
require a new license key. Performing these steps can ensure that MIMIX products
start properly once the hardware upgrade is complete.

Considerations for performing a hardware system upgrade without a disk


image change
Before you start a hardware upgrade on either system, consider the following:
• Ensure the new system is compatible with and meets the requirements for a
MIMIX-supported environment. For more information, see the Supported
Environments Matrix in the Technical Documents section of Support Central.
• Apply the latest MIMIX fixes on both systems. The fixes are available by product
in the Downloads section of Support Central.
• Obtain new MIMIX product license keys. These codes are required for products
when a model, feature, or serial number changes. For more information, see
“Working with license keys” in the Using License Manager book.
• Determine whether a planned switch is required prior to the hardware upgrade.
For example, a switch would be necessary if the source system is being upgraded
and users need to continue working while the upgrade takes place. To perform a
switch, follow the steps in your runbook. For more information, see “Switching” on
page 302.
• Determine if the transfer definitions need to be changed. For example, transfer
definitions would need to be changed if the IP addresses or host names change.
For more information, see “Configuring transfer definitions” in the MIMIX
Administrator Reference book.

MIMIX-specific steps for a hardware upgrade without a disk image


change
Use this procedure to restart your MIMIX installation when updating your hardware. If
you have special considerations, contact your Certified MIMIX Consultant for
assistance. Before you begin, ensure that “Considerations for performing a hardware
system upgrade without a disk image change” on page 382 have been reviewed and
completed where applicable.

Hardware upgrade without a disk image change - preliminary steps


To perform this portion of the upgrade process, do the following on the system prior to
the upgrade:
1. Ensure MIMIX is operating normally before performing the upgrade. There should
be no files or objects in error and all transactions should be caught up. See
“Resolving common replication problems” on page 259 for more information about

382
MIMIX procedures when upgrading hardware without a disk image change

resolving problems.
2. If upgrading a production system, ensure users are logged off the system and
perform a controlled end of all MIMIX data groups. For more information, see
“Ending a data group in a controlled manner” on page 244.
3. Optional step: If upgrading the source system and performing a planned switch,
follow the steps in your runbook. For more information, see “Switching” on
page 302.
4. Use the following command to end all MIMIX products in the installation library,
end the MIMIX managers, and end the RJ links:
ENDMMX ENDOPT(*CNTRLD) ENDRJLNK(*YES) ENDSMRJ(*YES)
5. Record status information for each data group in case it is needed later. Do the
following:
WRKDG DGDFN(*ALL) OUTPUT(*OUTFILE) OUTFILE(MIMIXQGPL/SWITCH)
6. Optional step: Save the MIMIX software and Vision Solutions Portal by doing a
full system save or by saving the following MIMIX installation libraries and IFS
directories:
• LAKEVIEW
• MIMIXQGPL
• MIMIX-installation-library
• MIMIX-installation-library_0
• MIMIX-installation-library_1
• VSI001LIB
• /LakeviewTech (directory tree)
• /visionsolutions/http/vsisvr

Hardware upgrade without a disk image change - subsequent steps


To perform this portion of the upgrade process, do the following on the system after
the upgrade has been completed:
1. Optional step: Update any transfer definitions that require changes. For more
information, see “Configuring transfer definitions” in the MIMIX Administrator
Reference book.
2. Enter the new product license key on the system. Do the following:
a. From the MIMIX main menu select option 31 (Product Management Menu).
The License Manager Main Menu appears.
b. Select option 1 (Update license key). The Update License Keys (UPDLICKEY)
command appears. Follow the instructions displayed for obtaining license
keys. For more information, see “Obtaining license keys using UPDLICKEY
command” in the Using License Manager book.
3. If the installed version of License Manager is at service pack 8.1.08.00 or higher,
skip this step. For earlier versions of License Manager (8.1.01.00 through

383
IBM Power™ Systems operations that affect MIMIX

8.1.07.00), you must reinstall License Manager using the installation wizard or the
instructions in the “Installing products” topic in the Using License Manager book.
4. Confirm that communications work between the new system and other systems in
the MIMIX environment. For more information, see “Verifying a communications
link for system definitions” on page 343.
5. Optional step: If you need to keep users active, perform a planned switch to the
backup system by following the steps in your runbook. See “Considerations for
performing a hardware system upgrade without a disk image change” on
page 382 to determine if a switch is required.
If you do not have a Runbook, you need to follow your processes for the following:
a. End all user applications, user interfaces, and operations actively running on
the production system.
b. Perform a planned switch to the backup system.
c. Start user applications on the backup system and allow users to access their
applications from the backup system.
6. Start MIMIX from either the source or target system. The Start MIMIX (STRMMX)
command starts the MIMIX processes for the installation, including the MIMIX
managers, data groups, and application groups. For more information about the
STRMMX command, see “Starting MIMIX” on page 226.
7. Run your MIMIX audits to verify the systems are synchronized. See “Running an
audit immediately” on page 169 for more information about running audits.

384
MIMIX procedures when performing a hardware upgrade with a disk image change

MIMIX procedures when performing a hardware upgrade


with a disk image change
When a hardware upgrade is being performed on a system, MIMIX software may
need to be saved from the system being replaced and then restored to the system
that is its replacement. The saved MIMIX information must be restored on a system
with the same designated system type (management or network). For example, if the
network system is being replaced, MIMIX software must be saved from the network
system and restored on the new network system. A network system cannot be
restored to a new management system.
This topic describes steps to consider prior to saving and restoring MIMIX software
when upgrading from one system to another. Performing these steps can ensure that
MIMIX products start properly once the hardware upgrade is complete.
IMPORTANT! To ensure the integrity of your data, contact your Certified MIMIX
Consultant for assistance performing a hardware upgrade.

Considerations for performing a hardware system upgrade with a disk


image change
Before you start a hardware upgrade on either system, consider the following:
• Contact your contact your Certified MIMIX Consultant prior to performing the
upgrade for instructions that may be specific to your environment.
• Ensure the new system is compatible with and meets the requirements for a
MIMIX-supported environment. For more information, see the Supported
Environments Matrix in the Technical Documents section of Support Central.
• Apply the latest MIMIX fixes on both systems. The fixes are available by product
in the Downloads section of Support Central.
• Obtain new MIMIX product license keys. These codes are required for products
when a model, feature, or serial number changes. For more information, see
“Working with license keys” in the Using License Manager book.
• Determine whether a planned switch is required prior to the hardware upgrade.
For example, a switch would be necessary if the source system is being upgraded
and users need to continue working while the upgrade takes place. To perform a
switch, follow the steps in your runbook. For more information, see “Switching” on
page 302.
• Determine if the transfer definitions need to be changed. For example, transfer
definitions would need to be changed if the IP addresses or host names change.
For more information, see “Configuring transfer definitions” in the MIMIX
Administrator Reference book.
• Copy all automation for MIMIX to the new machine, including exit programs.
• Transfer any modifications of programs such as QSTARTUP to the new system.
Modifications may be needed to start the MIMIX subsystem after an IPL. Refer to
your Runbook for an overview of the required automation changes that need to be
performed on the system.

385
IBM Power™ Systems operations that affect MIMIX

MIMIX-specific steps for a hardware upgrade with a disk image change


Use this procedure to save and restore your MIMIX installation when updating your
hardware with a disk image change. If you have special considerations, contact your
Certified MIMIX Consultant for assistance. Before you begin, ensure that
“Considerations for performing a hardware system upgrade with a disk image
change” on page 385 have been reviewed and completed where applicable.

Hardware upgrade with a disk image change - preliminary steps


To perform the save portion of the upgrade process, do the following on the old
system prior to the upgrade:
1. Ensure MIMIX is operating normally before performing the upgrade. There should
be no files or objects in error and all transactions should be caught up. See
“Resolving common replication problems” on page 259 for more information about
resolving problems.
2. Optional step: Perform a switch by following the steps in your runbook. See
“Considerations for performing a hardware system upgrade with a disk image
change” on page 385 to determine if a switch is required.
3. Ensure users are logged off the system and all applications have ended. Perform
a controlled end of all MIMIX data groups. For more information, see “Ending a
data group in a controlled manner” on page 244.
4. Use the following command to end all MIMIX products in the installation library,
end the MIMIX managers, and end the RJ links:
ENDMMX ENDOPT(*CNTRLD) ENDRJLNK(*YES) ENDSMRJ(*YES)
5. Ensure there are no open commit cycles. For more information, see “Checking for
open commit cycles” on page 230.
If open commit cycles exist, restart the data group and repeat Step 4 to end all
MIMIX products.
6. Each data group’s Last Read journal receiver and sequence number for both
system and user journals will be needed later in these instructions. Print the status
information for each data group by doing the following:
a. From the Work with Data Groups display, type 8 (Display detail status) next to
each data group.
b. Press Enter.
c. Press F7 for object status and print the display. Keep the information for later
use.
d. Press F8 for database status and print the display. Keep the information for
later use.
7. Print the list of system values. You will need this list in the next section after the
disk image is changed. Type the following on a command line and press Enter:
WRKSYSVAL SYSVAL(*ALL) OUTPUT(*PRINT)
8. Optional step: Save the MIMIX software and Vision Solutions Portal from the old
system by doing a full system save or by saving the following MIMIX installation

386
MIMIX procedures when performing a hardware upgrade with a disk image change

libraries and IFS directories:


• LAKEVIEW
• MIMIXQGPL
• MIMIX-installation-library
• MIMIX-installation-library_0
• MIMIX-installation-library_1
• VSI001LIB
• /LakeviewTech (directory tree)
• /visionsolutions/http/vsisvr

Hardware upgrade with a disk image change - subsequent steps


To perform this portion of the upgrade process, do the following after you have
upgraded and restored all user data, including all MIMIX libraries:
Note: To ensure that journaling is properly started, restore journals and journal
receivers before restoring user data.
9. Ensure the following system values are set the same way on the new system as
they were on the old system: QAUDCTL, QAUDLVL, QALWOBJRST,
QALWUSRDMN, QLIBLCKLVL, QRETSVRSEC, QSECURITY, and QSYSLIBL.
For more information about system value settings, see the MIMIX Administrator
Reference book.
10. Type the following on a command line and press Enter:
LAKEVIEW/UPDINSPRD
11. Enter the new product license key on the system. Do the following:
a. From the MIMIX main menu select option 31 (Product Management Menu).
The License Manager Main Menu appears.
b. Select option 1 (Update license key). The Update License Key (UPDLICKEY)
command appears. Follow the instructions displayed for obtaining license
keys. For more information, see “Obtaining license keys using UPDLICKEY
command” in the Using License Manager book.
12. If the installed version of License Manager is at service pack 8.1.08.00 or higher,
skip this step. For earlier versions of License Manager (8.1.01.00 through
8.1.07.00), you must reinstall License Manager using the installation wizard or the
instructions in the “Installing products” topic in the Using License Manager book.
13. Do the following from a command line on each system in the replication
environment to register MIMIX exit points in the system registry:
a. Add the MIMIX installation library to your library list.
b. Type CALL MXXPREG and press Enter.
14. Update any transfer definitions that require changes. For more information, see
“Considerations for performing a hardware system upgrade with a disk image
change” on page 385.

387
IBM Power™ Systems operations that affect MIMIX

15. Ensure all automation, including MIMIX exit programs, for MIMIX is available and
configured on the new system.
16. Make any necessary modifications to the QSTARTUP program. This may need to
be modified to start the MIMIX subsystem after an IPL. For more information, see
“Considerations for performing a hardware system upgrade with a disk image
change” on page 385.
17. Start the MIMIX subsystem with the following command:
STRSBS SBSD(MIMIXQGPL/MIMIXSBS)
18. On each system, do the following to clean up journals associated with system
manager RJ links:
a. Use the following command to display a list of the appropriate journal
definitions:
WRKJRNDFN JRNDFN(*ALL *LOCAL) RJLNK(*TGT) PROCESS(*INTERNAL)
b. For each journal definition listed, do the following:
• Type option 17 (Work with jrn attributes) and press Enter.
• Press F15 (Work with receiver directory).
• Type option 4 (Delete) for all receivers in the list. If message CPA7025 is
issued, reply with an “I”.
19. Confirm that communications work between the new system and other systems in
the MIMIX environment. For more information, see “Verifying a communications
link for system definitions” on page 343.
20. Start the system manager with the following command:
STRMMXMGR SYSDFN(*ALL) MGR(*SYS)
21. Optional step: Perform a data group switch by following the steps in your
runbook, then skip to Step 23 for the target system. See “Considerations for
performing a hardware system upgrade with a disk image change” on page 385 to
determine if a switch is required.

Important!
The role of the upgraded system within your production replication environment
determines what step to perform next.
• If the upgraded system is both the source and target of data groups within
your environment, you must perform both Step 22 and Step 23.
• If the upgraded system is only a source system, perform Step 22.
• If the upgraded system is only a target system, perform Step 23.

22. If the source system was upgraded, use this step to ensure that journaling
on the new system is correct and start replication processes:
a. On the source system, type the following on a command line, and press Enter:
WRKJRNDFN JRNDFN(*ALL *LOCAL) PROCESS(*REP)

388
MIMIX procedures when performing a hardware upgrade with a disk image change

b. Press F10 to verify the Receiver Prefix, Library, and all other parameters
(option 5) are correct. Make any necessary changes from the MIMIX
management system before continuing.
c. From the Work with Journal Definitions display, do the following to build the
journaling environment and record the journal receiver names:
• Type option 14 next to a journal definition with a value of *SRC or *NONE in
the RJ Link column and press F4 (Prompt).
• Type *JRNDFN for the Source for values parameter and press Enter.
• Record the newly attached journal receiver name by placing the cursor on
the posted message and pressing F1 or Help. You will need this information
to start data groups.
d. Repeat Step 22c for each journal definition with an RJ Link value of *SRC or
*NONE.
e. For each data group that has the upgraded system as its source system, run
the following commands to ensure that files and objects are journaled to the
correct journal:
• To verify file entries are journaled, run:
VFYJRNFE DGDFN(dgname) FILE1(*ALL) JRNSYS(*SRC)
• To verify IFS entries are journaled, run:
VFYJRNIFSE DGDFN(dgname)JRNSYS(*SRC)
• To verify object entries are journaled, run:
VFYJRNOBJE DGDFN(dgname)JRNSYS(*SRC)

f. Start the VSP server using the command:


VSI001LIB/STRVSISVR

g. For each data group that has the upgraded system as its source system, start
the data group with a clear pending start from the receivers recorded in
Step 22c of this procedure:
STRDG DGDFN(data-group-name) DBJRNRCV(user-journal-receiver)
DBSEQNBR2(*FIRST) OBJJRNRCV(security-journal-receiver)
OBJSEQNBR2(*FIRST) CLRPND(*YES) DTACRG(*YES)
h. User and application activity can be resumed on the system.
23. If the target system was upgraded, use this step to ensure that journaling
on the new system is correct and start replication processes:
i. On the target system, type the following on a command line and press Enter:
WRKJRNDFN JRNDFN(*ALL *LOCAL) PROCESS(*REP)
j. Press F10 to verify the Receiver Prefix, Library, and all other parameters
(option 5) are correct. Make any necessary changes from the MIMIX
management system before continuing.
k. From the Work with Journal Definitions display, do the following to build the
journaling environment:

389
IBM Power™ Systems operations that affect MIMIX

• Type option 14 next to a journal definition with a value of *SRC or *NONE in


the RJ Link column and press F4 (Prompt).
• Type *JRNDFN for the Source for values parameter and press Enter.
l. Repeat Step 23k for each journal definition with an RJ Link value of *SRC or
*NONE.
m. On a command line, type the following and press Enter:
WRKJRNDFN JRNDFN(*ALL *LOCAL) RJLNK(*TGT)

n. For the journal definitions listed on the Work with Journal Definitions display,
do the following:
• Type option 17 (Work with jrn attributes) and press Enter.
• Press F15 (Work with receiver directory).
• Type option 4 (Delete) for all receivers in the list. If message CPA7025 is
issued, reply with an “I”.
o. Repeat Step 23n for each journal definition associated with the target system.
p. For each data group that has the upgraded system as its target system, run the
following commands to ensure that files and objects are journaled to the
correct journal:
• To verify file entries are journaled, run:
VFYJRNFE DGDFN(dgname) FILE1(*ALL) JRNSYS(*TGT)
• To verify IFS entries are journaled, run:
VFYJRNIFSE DGDFN(dgname)JRNSYS(*TGT)
To verify object entries are journaled, run:
VFYJRNOBJE DGDFN(dgname)JRNSYS(*TGT)

q. Start the VSP server using the command:


VSI001LIB/STRVSISVR

r. Use the following command and the printed information collected in Step 6 of
“Hardware upgrade with a disk image change - preliminary steps” on page 386
to start all of the data groups.
Note: For each data group, use the Last read receiver and sequence number
from its printed Object Status view (Step 6c) as the values for
OBJJRNRCV andOBJSEQNBR2, and the values from the printed
Database Status view (Step 6d) as the values for DBJRNRCV and
DBSEQNBR2 in the following command:
STRDG DGDFN(data-group-name) DBJRNRCV(last-read-database-
journal-receiver) DBSEQNBR2(last-read-database-sequence-number)
OBJJRNRCV(last-read-object-journal-receiver) OBJSEQNBR2(last-
read-object-sequence-number) CLRPND(*YES) DTACRG(*YES)
24. From both systems, delete any old receivers with different library or prefix names.
25. Run your MIMIX audits to verify the systems are synchronized. See “Running an
audit immediately” on page 169 for more information about running audits.

390
Handling MIMIX during a system restore

Handling MIMIX during a system restore


Occasionally, an entire system may need to be restored because of a system failure.
For example, if there is a processor or OS (operating system) failure. This topic
includes prerequisites for restoring MIMIX software within a MIMIX system pair (two
systems using the same MIMIX installation) to one system from a save of the other
system. A system restore may need to be performed when the when the following
conditions exist:
• The original production system, including the Licensed Internal Code and the
OS, has been recovered from the backup system by tape.
• The IBM installed release level is the same on each system.
IMPORTANT! To ensure the integrity of your data, contact your Certified MIMIX
Consultant for assistance performing a system restore.
For information about MIMIX-supported environments, see the Supported
Environments Matrix in the Technical Documents section of Support Central.

Prerequisites for performing a restore of MIMIX


Before you restore MIMIX on a system, consider the following steps which help
ensure that MIMIX products start properly once the restore is complete:
• Contact your Certified MIMIX Consultant prior to performing the restore for
instructions that may be specific to your environment.
• Locate your MIMIX product license keys. These codes may be required after the
restore. For more information, see “Working with license keys” in the Using
License Manager book.

391
Index
Symbols #DLOATR considerations 152
*ATTN #FILDTA considerations 152
application group 83 #IFSATR considerations 152
managers for node 87 #MBRRCDCNT considerations 153
monitors 86 after a configuration change 151
replication 89 authority level to run 24
*CANCEL, step status of 109 automatic starting of 60
*CANCELED, procedure status of 102, 113 before switching 151
*FAILED activity entry 278, 281 best practice 24, 151
*FAILED status bi-directional environment considerations 29
procedure 102, 113 change history retention criteria 43
step 109 change status severity 44
*HLD changing automatic schedule 68
file entry 263 check for compliance 45
tracking entry 272 check for reported problems 162
*HLDERR compare phase 147
file entry 263 comparison levels 53
tracking entry 272 compliance 160
*HLDRLTD file entry 263 compliance threshold 52, 53
*INACTIVE definition of 18
application group 85 differences, resolving 162
node managers 88 displaying compliance status 167
replication 89 displaying history 171
*MSGW status displaying status 157
procedure 101 displaying time of next scheduled run 155
step 108 displaying when prioritized runs are allowed
*UNKNOWN 85 155
enable or disable automatic 70
A ending 170
accessing history 171
MIMIX Main Menu 26 job log 166
activity entries, object last performed 167
confirm delay/retry cycle 282 last successful run 150
failed, resolving 278 no objects selected 173
remove history 285 objects compared 173
retrying 281 policies which affect 40
additional resources 14 policies, runtime behavior 40
application group policies, submitting automatically 61
resolving reported problems 82 prevent from running 46, 70
status of 80 priority selection example 173
application group definition 18, 21 priority, changing selection frequency 69
application node status 84 problems reported in installation 118
applications, reducing contention with 341 recovery phase 147
ASP 1 recovery threshold, storage used ex- results recommendations 153
ceeds 183 retain history of 55
audit rule name and description 65
running immediately 169
default settings for automatic 64 schedule for automatic runs 155
#DGFE considerations 152 status
runtime 162

392
status, compliance 160 switching 303
three or more node considerations 29 best practices
when not to audit 30 auditing 151
audit history bi-directional environment policy consider-
change criteria 43 ations 29
audit level
best practice 54 C
changing before switch 42 cancel
audit results 162 procedure 114
#DGFE rule 361 clear error entries
#FILDTA rule 364 processing 223
#MBRRCDCNT rule 364 when to 228
interpreting, attribute comparisons 367 clear pending entries
interpreting, file data comparisons 364 check for open commits 230
resolving problems 162, 361 open commit cycle prevents 230
troubleshooting 166 processing 223
auditing level, object resolving open commits before 230
set when starting a data group 221 when to 228
used for replication 287 cluster services 22
authority level cold start, replacement for 223
for product access 24 collector services 22
automatic error recovery ending 186
replication, policies for 35 starting 185
system journal replication 261 status 180
user journal replication 260 collector services status 87
automatic recovery command, by mnemonic
audits 50 SETMMXSCD 61
concept 18 command, by name
system journal replication 50 Set MIMIX Schedules 61
user journal replication 50 Work with Audit History 171
AutoNotify feature, MIMIX 210 commands, by mnemonic
auxiliary storage pool, system (ASP1) 34 CHGDG 334
CHGMONSTS 197
B CHGPROCSTS 110
backlog CHKDGFE 345, 361
starting shared object send job 222 CNLPROC 114
system manager 181 CRTDGTSP 352
backlog, identifying a 137 DLTDGTSP 354
backup node sequence DSPDGSTS 126
changing 91 DSPDGTSP 354
examples of changing 93 DSPMMXMSGQ 283
verifying 90 ENDAG 215
backup system 19 ENDDG 215, 232, 238
best practice ENDJRNFE 292
audit level 54, 151 ENDJRNIFSE 295
audit level before switch 42, 54, 151 ENDJRNOBJE 299
audit threshold 52, 53 ENDMMX 215, 232, 240
switch frequency 303, 311 ENDMON 195
switch threshold 57 ENDMSTMON 196

393
ENDRJLNK 202 Display Data Group Timestamps 354
ENDSVR 329 Display MIMIX Message Queue 283
HLDDGLOG 355 End Application Group 215
MIMIX 26 End Data Group 215, 232, 238
RLSDGLOG 355 End Journaling File Entries 292
RUNPROC 111 End Journaling IFS Entries 295
STRAG 215, 217 End Journaling Obj Entries 299
STRDG 215, 217, 221 End Lakeview TCP Server 329
STRJRNFE 291 End Master Monitor 196
STRJRNIFSE 294 End MIMIX 215, 232, 240
STRJRNOBJE 298 End Monitor 195
STRMMX 215, 217, 226 End RJ Link 202
STRMON 194 Hold Data Group Log 355
STRMSTMON 196 MIMIX 26
STRRJLNK 202 Release Data Group Log 355
STRSVR 328 Run Procedure 111
SWTDG 312, 314 Start Application Group 215, 217
VFYCMNLNK 343, 344 Start Data Group 215, 217, 221
VFYJRNFE 293 Start Journaling File Entries 291
VFYJRNIFSE 296 Start Journaling IFS Entries 294
VFYJRNOBJE 300 Start Journaling Obj Entries 298
VFYKEYATR 351 Start Lakeview TCP Server 328
WRKAG 80 Start Master Monitor 196
WRKAUDHST 171 Start MIMIX 215, 217, 226
WRKAUDOBJ 173 Start Monitor 194
WRKAUDOBJH 176 Start RJ Link 202
WRKCPYSTS 331 Switch Data Group 312, 314
WRKDG 118 Verify Communications Link 343, 344
WRKDGACT 278 Verify Journaling File Entry 293
WRKDGACTE 279 Verify Journaling IFS Entries 296
WRKDGFE 263 Verify Journaling Obj Entries 300
WRKDGIFSTE 272 Verify Key Attributes 351
WRKDGOBJTE 272 Work with Application Groups 80
WRKDGTSP 352 Work with Audited Obj. History 176
WRKDTARGE 88 Work with Audited Objects 173
WRKMSGLOG 284 Work with Copy Status 331
WRKNFY 206 Work with Data Group Activity 278
WRKNODE 87 Work with Data Groups 118
WRKPROCSTS 98 Work with Data Rsc. Grp. Ent. 88
WRKRJLNK 199, 332 Work with DG Activity Entries 279
WRKSTEPSTS 103 Work with DG File Entries 263
commands, by name Work with DG IFS Tracking Ent. 272
Cancel Procedure 114 Work with DG Obj Tracking Ent. 272
Change Data Group 334 Work with DG Timestamps 352
Change Monitor Status 197 Work with Message Log 284
Change Procedure Status 110 Work with Node Entries 87
Check Data Group File Entries 345, 361 Work with Notifications 206
Create Data Group Timestamps 352 Work with Procedure Status 98
Delete DG Timestamps 354 Work with RJ Links 199, 332
Display Data Group Status 126 Work with Step Status 103

394
commit cycles D
effect on audit comparison 364, 365 data areas and data queues
commit cycles, open determining configuration of 337
checking for 230 holding user journal entries for 274
checking for after a controlled end 245 resolving problems 273
preventing problems with 235 tracking entries 272
preventing STRDG request 230 verifying journaling 300
commit mode change data group 18
prevents starting with open commits 230 backlogs 137
communications controlled vs. immediate end 234
ending TCP sever 329 definition 21
starting TCP sever 328 determining if RJ link used 332
compare phase 147 disabling 335
compliance enabling 335
audit 160 ending considerations 238
change what audits are checked 45 ending controlled 244
concept 150 ending immediately 247
switch 311 ending selected processes 247
switch, policies for 49 indication of disabled state 334
compliance status recovery point cleared 238
switch 311 starting selected processes 228
concepts state, disabled or enabled 334
auditing 146 status, database view 133
MIMIX 18 status, detailed 126
configuration status, merged view 127
audit after changing 151 status, object view 131
determining data areas and data queues 337 status, summary 118
determining, IFS objects 336 switching 307, 312
results of #DGFE audit after changing 361 timestamps 352
configuration changes deployed 221 when to exclude from auditing 30
constraints data group entry
enabling for *INACTIVE data group 249 description 21
enabling MIMIX-disabled 248 data groups
enabling when ending replication 249 accessing 118
manually enable on target 248 data resource group 88
contacting Vision Solutions 15 replication status summary 89
contention with applications, reducing 341 database apply (DBAPY) status 134
controlled end database apply cache policy 51
confirm end 245 database error recovery, automatic 261
description 234 deadlock condition
procedure 244 resolving 144
wait time 235 definition
cooperative processing 21 application group 21
copying active files 331 data group 21
correcting journal 21
file-level errors 269 remote journal (RJ) link 21
record-level errors 270 system 21
CustomerCare 15 transfer 21
definitions

395
application group 18 included processes 236
delay/retry cycle, confirm object in a 282 using default values 240
delaying captures for replication recoveries 36 using specified values 240
disabled audit 70 when to end RJ links 236
disabled data group 334 ending replication 215
disabled report 73 choices 232
displaying controlled vs. immediate 234
data group spooled file information 330 ending RJ link
data group status details 126 independently 202
long IFS object names 330 when to end 184, 236
RJ link status 199 errors
documents, MIMIX 11 file level 269
record level 270
E system journal replicated objects 278
ending target journal of RJ link 350
audit 170 user journal replicated files 263
collector services 186 user journal replicated objects 272
journal managers 185 example
master monitor 196 priority audit object selection 173
MIMIX managers 184, 185 examples
MIMIXSBS subsystem 241 changing backup node sequence 93
monitor 195
system manager RJ links 184 F
system managers 184 file
target journal inspection 187 file-level errors 269
TCP server 329 hold journal entries 267
ending data group new 287
clears recovery point 238 not journaled 122
considerations when ending 238 record-level errors 270
controlled end 244 replicated 263
controlled end and enable constraints 249 file identifiers (FIDs) 338
controlled end wait time 235 file in error
controlled vs. immediate 234 examine held journal entries 266
how to confirm end 245 resolving 263
immediate end 247 file on hold
processes 235 release and apply held entries 268
processes, effect of 255 release and clear entries 269
processes, specifying selected 247 release at synchronization point 268
when to end RJ link 236 files with protected rows or columns 270
ending journaling
data areas and data queues 299 H
files 292 hardware upgrade
IFS objects 295 MIMIX-specific steps 382
IFS tracking entry 295 no disk image change 382
object tracking entry 299 prerequisites 382
ending MIMIX 240 with a disk image change 385
controlled vs. immediate 234 held error (*HLDERR)
end subsystem, when to also 241 file entry 263
follow up after 241 preferred action for entry 264, 273

396
tracking entry 272 requirements and restrictions 288
history journal cache or state
audited object 176 resolving problems 122, 141
completed audits 171 status 140
displaying audit 171 journal definition 21
history log, removing completed entries 285 defined to RJ Link 333
history of, retaining journal entry
procedures 33 description 20
hold (*HLD) unconfirmed 348
preferred action for held entry 264, 273 Journal manager
put file entry on hold 267 resolving problems 183
put tracking entry on hold 274 journal manager 22
release a held file entry 268 ending 185
release a held tracking entry 276 resolving problems 180
hold ignore (*HLDIGN) starting 184
preferred action for ignored entry 264, 273 status 180
put file entry on hold ignore 267 journal receiver 20
put tracking entry on hold ignore 275 journal status
hold related (*HLDRLTD) 264 resolving problems 143
hot backup 16 journaling 20
data group problem with 120
I ending for data areas and data queues 299
i5/OS upgrade 374 ending for files defined to a data group 292
IFS objects ending for IFS objects 295
determining configuration 336 implicitly started 287
file IDs (FIDs) 338 requirements for starting 287
hold user journal entries for 274 starting for data areas and data queues 298
path names 330 starting for IFS objects 294
resolving problems 273 starting for physical files 291
tracking entries for 272 starting, ending, and verifying 286
verifying journaling 296 verifying for data areas and data queues 300
immediate end verifying for IFS objects 296
description 234 verifying for physical files 293
incomplete tracking entry 234 journaling status
information and additional resources 14 data areas and data queues 298
inspection files 291
target journal 23 IFS objects 294
IPL 372
K
J keyed replication
job log verifying file attributes 351
for audit 166
jobs L
used by procedures 97 last audit performed 150
used by procedures, status of 104 last switch performed 311
journal 20 log space 23
inspection on target system 23 long IFS path names 330
journal at create
requirements 287

397
M N
management system 20 names, displaying long 330
manager network system 20
journal 22 new hardware upgrade 385
system 22 MIMIX-specific steps 386
manager status 87, 180 prerequisites 385
master monitor new objects
ending 195, 196 IFS object journal at create requirements 287
starting 195, 196 journal at create selection criteria 288
menu newly created objects, notification of 210
MIMIX Main 26 node entries 87
message queue, primary and secondary 283 node status
messages application group 84
ENDMMX 218 data resource group 89
STRMMX 218 nodes, policy considerations for multiple 29
MIMIX CDP feature not journaled
exclude from audit 30 resolving problems 122
recovery point cleared 238 notification status 205
MIMIX installation 18 notifications
MIMIX managers definition 19, 204
checking for a backlog 181 displaying 205
ending 185 new problems in installation 118
resolving problems 180, 181, 183 severity level 150, 207
starting 184 status 206
MIMIX Model Switch Framework 24, 307
policy default 57 O
MIMIX rules 146 object
MIMIX subsystem (MIMIXSBS) audited history 176
starting 226 object auditing
when to end 241 concept 20
MIMIX Switch Assistant 24 setting level with STRDG 221
setting default switch framework 48 used for replication 287
setting switch compliance policies 49 object error recovery, automatic 262
MMNFYNEWE monitor 210 object send process
monitor considerations for starting a shared 222
disabling 197 objects
enabling 197 audited object list 173
ending 195 configuration of non-file 336
master 195 displaying long IFS names 330
nodes where needed 193 displaying objects in error 128
starting 194, 195 displaying objects with active entries 129
status 192 in error, resolving 278
monitor for newly created objects 210 new 287
monitors reducing contention 341
nodes where needed 86 tracking entries for data areas and data
status of 85 queues 272
MXSCHED job 60 open commit cycles
audit results 364, 365
prevent problems with 235

398
resolving before starting replication 230 notification severity 51
shown in status 245 object only on target 51
when starting a data group 230 object recovery notify on success 50
operations prioritized audit in effect 155
common, where to start 117 procedure history retention 58
less common 327 run rule on system 54
orphaned recoveries 213 switch action threshold 57
output file fields switch warning threshold 57
Difference Indicator 364, 367 synchronize threshold size 56
System 1 Indicator field 369 third delay retry interval 56
System 2 Indicator field 369 third delay retry interval, number of 56
user journal apply threshold 51
P PPRC replication status 84, 89
path names, IFS 330 problems
planned switch 303 troubleshoot 339
policies 19 problems, journaling
audit, automatically submitting 61 data areas and data queues 298
audit, runtime behavior of 40 files 291
changing values 31 IFS objects 294
for auditing 40 problems, resolving
for replication 35 audit results 361
for switching 48 auditing 162
installation-level only 33 common errors 259
introduction 28 common system level errors 178
multi-node and bi-directional environment data group cannot end 342
considerations 29 data group cannot start 347
policy files in error 263
action for running audits 54 files not journaled 122
ASP threshold for recoveries 58 journal cache or state 122, 141
audit action threshold 53 journal status 143
audit history retention 55 not journaled 122
audit level 53 objects in error 278
audit notify on success 50 open commits when starting data group 230
audit rule 50 RJ link cannot end 348
audit severity 44 RJ link cannot start 348
audit warning threshold 52 system level processes 180
audits to include in compliance 45 tracking entries 272
automatic audit recovery 50 procedure
automatic database recovery 50 acknowledging failed or canceled 110
automatic object recovery 50 begin at step 111, 219, 306
changing source capture delay 36 canceling 114
CMPRCDCNT commit threshold 58 defined 23
data group definition 50 directory data protection report 58
database apply cache 51 displaying status 98
default model switch framework 57 history retention 58
directory depth 58 how to run 111
independent ASP library ratio 57 last run 98
journaling attribute difference action 51 multiple jobs 97
maximum rule runtime 52 multiple jobs, status of 104
next automatic report run 114

399
overriding step attributes 112 changing backup sequence 91
resolve problems 98, 99 verifying sequence 90
resuming canceled or failed 112 recovery phase 147
run type *NODE 112 recovery point
run type *USER 112 cleared by ENDDG 238
run type other than *USER 112 recovery, automatic 18
status 99 related IFS tracking entries 277
status history of a 115 release (*RLS)
step status 103 held file entry 268
procedure history held tracking entry 276
change criteria 33 release clear (*RLSCLR)
procedures 97 file entry 269
change directory depth for data protection re- tracking entry 276
ports 33 release wait (*RLSWAIT)
change history retention criteria 33 file entry 268
history retention 33 tracking entry 275
procedures and steps remote journal
differences in virtual switch 324 i5/OS function 20
processes remote journal (RJ) link 21
system level 178 remote journal environment
production system 19 processes ended by ENDDG 255
publications, IBM 14 processes started by STRDG 251
unconfirmed journal entry 348
Q removing
QSTRUPPGM system value 372, 375 activity history entries 285
duplicate tracking entries 276
R reorganizing, active files 331
recommendations replication
auditing 151 automatic error recovery 35
before planned switch 303 backlogs, identifying 137
checking audit results 153 before starting 217
policies in bi-directional environment 29 commands for ending 232
policies in three or more node environment commands for starting 217
29 direction of 19
starting shared object send 222 ending 215
recoveries policies that affect 35
active in installation 118 recovery source capture delay 37
definition 19, 205 resolve replication errors 259
detected database errors 261 starting 215
displaying count of in progress 211 status summary 84, 89
displaying details 211 supported paths 16
occurring in installation 211 switching 302
orphaned 213 system journal 16, 22
orphaned, identifying 214 user journal 16, 22
orphaned. removing 214 replication path 22
recoveries delayed by ASP threshold 183 replication, problems
recovery troubleshoot 339
source capture delay 36 where to start 259
recovery domain report
automatic starting of 60

400
changing automatic schedule 72 procedure 111
default schedule 67 running
enable or disable automatic 73 audits immediately 169
next scheduled run 114
requirements S
audits 151 schedule
enable constraints manually 248 automatic audit or report 62
journal at create 287 automatic audits, displaying 155
journaling 287 changeable criteria 61
resolving problems changing audit 68
application group 82 changing report 72
application group *ATTN status 83 priority audit 63
application group other problem status val- shipped defaults for auditing 64
ues 84 shipped defaults for reports 67
common replication errors 259 scheduler job, MXSCHED 60
data resource group status 88 servers
node entry status 87 ending TCP 329
system level jobs 180 starting TCP 328
troubleshooting 339 service
resource group, data 88 cluster 22
status 88 status collector 22
restore MIMIX services
prerequisites 391 collector, ending 186
restrictions collector, starting 185
journal at create 288 severity level, notification 207
QDFTJRN data area 288 source system 19
retry objects in error 281 spooled files, displaying MIMIX-created 330
retrying, data group activity entries 281 standby journaling
RJ link 21 IBM i5/OS option 42 139
ending independently 202 overview 139
errors for target journal of 350 starting
identifying data groups that use 332 collector services 185
journal definitions by an 333 master monitor 195, 196
operating independently 202 MIMIX managers 184
removing unconfirmed entries 348 monitor 194
status 199 procedure at step 111, 219, 306
when to end 236 RJ link independently 202
RJ link, ending system manager 184 system and journal managers 184
rule target journal inspection 186
#DGFE 65 TCP server 328
#DLOATR 66 virtual switch 323
#FILATR 66 starting data group
#FILATRMBR 66 at specified journal location 222
#FILDTA 66 deploy configuration 221
#IFSATR 66 prevented by open commit cycles 230
#OBJATR 66 procedure 228
descriptions 65 processes, effect of 251
rules set object auditing level 221
MIMIX 146 when to clear entries 228
run

401
starting journaling ended 270
data areas and data queues 298 ending 241
file entry 291 starting 226
files 291 switch
IFS objects 294 precheck for virtual 323
IFS tracking entry 294 Switch Assistant, MIMIX 24
object tracking entry 298 switch framework
starting MIMIX disable policy when not used 49
included processes 217 specify a default 48
procedure 226 switch testing, virtual 316
starting replication 215 switch, virtual 316
before 217 performing a 323
choices 217 problem resolution differences 324
status switching 302
active file operations 331 application group 308
application group 80 best practice 24, 303, 311
audit compliance 167 change audit level before 42, 54
audits 157 compliance 311
collector services 87 conditions that end 314
data group detail 126 description, planned switch 303
database apply (DBAPY) 134 description, unplanned switch 304
journal cache or state 140 journal analysis after unplanned switch 356
journaling data areas and data queues 298 last switch field 311
journaling files 291 phases of a 303
journaling IFS objects 294 policies for 48
journaling tracking entries 294, 298 reasons for 303
monitors 85, 192 setting switch compliance policies 49
node entries 87 setting switch framework policy 48
notification 206 switch framework vs. SWTDG command 307
notifications 205 SWTDG command details 314
procedures 98 unplanned, actions to complete an 304
recoveries active in installation 211 using option 6 on MIMIX Basic Main Menu
replication, application group level 84 309
replication, data resource group level 89 using STRDG command 312
replication, logical 89 synchronize
replication, PPRC 89 file entry 264
RJ link 199 objects, system journal replicated 280
steps in a procedure 103 tracking entries 274
switch compliance 311 system definition 21
switching 124 system journal replication 16, 22
system and journal managers 87 detailed status 131
system-level processes 180 errors automatically recovered 261
target journal inspection processes 188 journaling requirements 287
Work with Data Groups display 118 system level processes 178
step system manager 22
begin procedure at 111, 219, 306 backlog 181
defined 23 ending 184
resolve problems 106 resolving problems 180, 181
status 103, 106 starting 184
subsystem, MIMIXSBS status 180

402
when to end associated RJ link 184 performing journal analysis 356
system roles unprocessed entries 245
management or network 20 upgrade
production or backup 19 hardware, no disk image change 382
source or target 19 hardware, with a disk image change 385
new hardware 385
T OS/400 374
target constraint management 248 user journal replication 16, 22
target journal inspection 23 detailed status 133
last entry inspected 191 errors automatically recovered 260
results 189 journaling requirements 287
starting 186 non-file objects 336
status 180, 188 tracking entries 272
target system 19 tracking entry 21
threshold
ASP storage used 183 V
audit action 53 verifying
audit warning 52 communications link 343, 344
CMPRCDCNT open commit 58 journaling, IFS tracking entries 296
switch action 57 journaling, object tracking entries 300
switch warning 57 journaling, physical files 293
synchronize size 56 key attributes 351
System ASP percent used 58 viewing status, active file operations 331
system ASP percent used recovery 34 virtual switch
user journal apply 51 application testing considerations 321
timestamps 352 automatic error recovery 318
automatically created 352 check MIMIX prerequisites 323
creating additional 352 compared to test planned switch 316
deleting 354 configuration requirements 319
displaying 354 major phases of 317
printing 354 MIMIX considerations 319
tips operational considerations 319
displaying data group spooled files 330 performing a 323
displaying long IFS object names 330 procedure problem differences 324
removing journaled changes 355 source capture delay during testing 37
working with active file operations 331
tracking entry 21 W
file identifiers (FIDs) 338 wait time, data group controlled end 235
IFS 272 wait time, data group controlled end during
incomplete 234 switch 314
not journaled 122
object 272
related IFS entries 277
removing duplicate 276
transfer definition 21

U
unconfirmed journal entries, removing 348
unplanned switch 304

403

You might also like