Rule | Rules are listed within relevant categories, such as Cloud and Storage. |
Availability | If a particular type of collection is not licensed or collected, risk mitigation data will not be available, regardless of how a rule is configured. In some cases, a Portal may have the necessary license, but collection may not have been enabled and/or completed. • Data Protection rules require a Backup Manager license. • Storage requires a Capacity Manager license. |
Description | The full description of the Risk Mitigation rule can be viewed by placing your mouse over the description. |
Notes | Enter operational notes for future reference. |
Status | • Green indicates successful collection of risk mitigation historical data for enabled rules. • Red indicates failed historical data collection. It could be that collection is attempting to access data for a product module that is not in your Portal environment. Click the red icon to view the Database Error Aggregation report. • A non-colored icon indicates that the background process did not run, typically because the rule is not enabled. |
State | Indicates if the rule is Enabled or Disabled. |
Last Run | The date and time that the background process ran and evaluated the collected data against the rule’s configured parameters. |
Risk Mitigation Rule | Description |
Data Protection Rules | |
Backup Job Size Variance | Compares client’s average job size, which may help to identify backup issues. |
Client Consecutive Failure | Evaluates clients where consecutive backups have failed or no backups have occurred for consecutive days. This rule examines the past 14 days of history, providing insights to possible problematic clients. Best Practice: Schedule this rule to run every day at the end of the backup window. This rule works with any backup product. |
Client Overall Status Summary | Considers clients for which backup jobs were not successful to determine risk. This rule helps in finding such clients by providing Status Summary. Determining if client backups were successful is complicated, especially if there are multiple policies and schedules defined for that client and if there are multiple streams per backup set. Also there needs to be an established cutoff time to determine what to do if a client is still running or has not made all of its attempts. The following criteria is considered: 1. If a client fails all of its jobs it is failed. 2. If a client successfully completes all of its jobs it is successful. 3. If a client completes all of its jobs with status 1 (skipped files) it was partially successful and probably OK. 4. If a client has a mixture of successful jobs and failed jobs, it needs further examination to determine if the jobs were truly successful. Now, there is logic that can be applied to #4 in order to programmatically determine whether a client was successful or not, but that logic varies from customer to customer. |
Clients with No Recent Backups | Reviews details of clients that have not been backed up in a defined number of days to help determine if the clients are at risk. Specify the number of days for which backups have not occurred to determine the risk. |
Compliance RTO RPO | Considers RTO (Recovery Time Objectives) and RPO (Recovery Point Objectives) for backups by determining when/if the last full backup was performed. Then, add in the time it takes to apply any incremental backups. Assists in computing RTO (Recovery Time Objectives) and RPO (Recovery Point Objectives) for backups by determining when/if the last full backup was performed. Then, add in the time it takes to apply any incremental backups to determine if you are meeting your SLAs. |
NetBackup Disk Pool Forecast | Provides NetBackup disk pool statistics for the number of weeks in the selected period are examined to forecast the date when storage will run out within the next three years. If the prediction is beyond three years, a status is returned. |
Storage Rules | |
Hot Array Ports | Identifies overactive array ports, which may indicate a risk to application performance. Array port performance data is examined to identify spikes in data transferred. |
Hot LUNs by Read IO | Reveals spikes in Read I/O performance metrics, which may indicate an area of risk. This rule uses a unique yet simple algorithm to identify abnormal performance patterns. |
Hot LUNs by Read Response | Reveals spikes in Read Response Time metrics, which may indicate an area of risk. This rule uses a unique yet simple algorithm to identify abnormal performance patterns. |
Hot LUNs by Write IO | Reveals spikes in Write I/O activity, which may indicate an area of risk. This rule uses a unique yet simple algorithm to identify abnormal performance patterns. |
Hot LUNs by Write Response | Reveals spikes in Write Response Time metrics, which may indicate an area of risk. This rule uses a unique yet simple algorithm to identify abnormal performance patterns. |
Thin Pool Forecast | Uses multi-vendor and multi-metric pool capacity and forecast data to identify storage at risk. |