NetApp OnCommand Console Administration

You might also like

You are on page 1of 538

NetApp OnCommand Console

Administration Task Help


For Use with Core Package 5.0

NetApp, Inc.
495 East Java Drive
Sunnyvale, CA 94089 USA
Telephone: +1 (408) 822-6000
Fax: +1 (408) 822-4501
Support telephone: +1 (888) 4-NETAPP
Documentation comments: doccomments@netapp.com
Information Web: http://www.netapp.com
Part number: 215-05997_A0
July 2011

Table of Contents | 3

Contents
About this document .................................................................................. 13
Welcome to OnCommand console Help ................................................... 15
How to use OnCommand console Help .................................................................... 15
Bookmarking your favorite topics ................................................................. 15
Understanding how the OnCommand console works ............................................... 15
About the OnCommand console ................................................................... 15
Window layout and navigation ..................................................................... 15
Window layout customization ....................................................................... 16
How the OnCommand console works with the Operations Manager
console and NetApp Management Console ............................................ 17
Launching the Operations Manager console ................................................. 17
Installing NetApp Management Console ...................................................... 18
How the OnCommand console works with AutoSupport ............................. 19

Dashboard ................................................................................................... 21
Understanding the dashboard .................................................................................... 21
OnCommand console dashboard panels ....................................................... 21
Monitoring the dashboard ......................................................................................... 22
Monitoring dashboard panels ........................................................................ 22
Page descriptions ....................................................................................................... 22
Availability dashboard panel ......................................................................... 22
Events dashboard panel ................................................................................. 23
Full Soon Storage dashboard panel ............................................................... 24
Fastest Growing Storage dashboard panel .................................................... 24
Dataset Overall Status dashboard panel ........................................................ 25
Resource Pools dashboard panel ................................................................... 25
External Relationship Lags dashboard panel ................................................ 26
Unprotected Data dashboard panel ............................................................... 26

Events and alarms ...................................................................................... 29


Understanding events and alarms .............................................................................. 29
What events are ............................................................................................. 29
What alarms are ............................................................................................. 29
Guidelines for creating alarms ...................................................................... 30

4 | OnCommand Console Help


How to know when an event occurs .............................................................. 31
Description of event severity types ............................................................... 31
Alarm configuration ...................................................................................... 32
Configuring alarms .................................................................................................... 32
Creating alarms for events ............................................................................. 32
Creating alarms for a specific event .............................................................. 33
Managing events and alarms ..................................................................................... 34
Resolving events ........................................................................................... 34
Editing alarm properties ................................................................................ 35
Configuring the mail server for alarm notifications ...................................... 36
Monitoring events and alarms ................................................................................... 36
Viewing event details .................................................................................... 36
Viewing alarm details .................................................................................... 37
Page descriptions ....................................................................................................... 37
Events tab ...................................................................................................... 37
Alarms tab ..................................................................................................... 40
Create Alarm dialog box ............................................................................... 41
Edit Alarm dialog box ................................................................................... 43

Jobs .............................................................................................................. 45
Understanding jobs .................................................................................................... 45
Understanding jobs ........................................................................................ 45
Managing jobs ........................................................................................................... 45
Canceling jobs ............................................................................................... 45
Monitoring jobs ......................................................................................................... 46
Monitoring jobs ............................................................................................. 46
Page descriptions ....................................................................................................... 46
Jobs tab .......................................................................................................... 46

Servers ......................................................................................................... 53
Understanding virtual inventory ................................................................................ 53
How virtual objects are discovered ............................................................... 53
Monitoring virtual inventory ..................................................................................... 53
Monitoring VMware inventory ..................................................................... 53
Monitoring Hyper-V inventory ..................................................................... 57
Managing virtual inventory ....................................................................................... 58
Adding virtual objects to a group .................................................................. 59
Adding a virtual machine to inventory .......................................................... 59

Table of Contents | 5
Preparing a virtual object managed by the OnCommand console for
deletion from inventory ........................................................................... 60
Performing an on-demand backup of virtual objects .................................... 61
Restoring backups from the Server tab ......................................................... 64
Mounting and unmounting backups in a VMware environment ................... 67
Page descriptions ....................................................................................................... 71
VMware ......................................................................................................... 71
Hyper-V ......................................................................................................... 80

Storage ......................................................................................................... 85
Physical storage ......................................................................................................... 85
Understanding physical storage .................................................................... 85
Configuring physical storage ........................................................................ 87
Managing physical storage ............................................................................ 89
Monitoring physical storage .......................................................................... 94
Page descriptions ......................................................................................... 100
Virtual storage ......................................................................................................... 118
Understanding virtual storage ..................................................................... 118
Managing virtual storage ............................................................................. 120
Monitoring virtual storage ........................................................................... 123
Page descriptions ......................................................................................... 125
Logical storage ........................................................................................................ 134
Understanding logical storage ..................................................................... 134
Managing logical storage ............................................................................ 136
Monitoring logical storage .......................................................................... 141
Page descriptions ......................................................................................... 150

Policies ....................................................................................................... 167


Local policies .......................................................................................................... 167
Understanding local policies ....................................................................... 167
Configuring local policies ........................................................................... 170
Managing local policies .............................................................................. 172
Page descriptions ......................................................................................... 175

Datasets ...................................................................................................... 181


Understanding datasets ............................................................................................ 181
What a dataset is .......................................................................................... 181
Dataset concepts .......................................................................................... 181
Role of provisioning policies in dataset management ................................. 183

6 | OnCommand Console Help


Role of protection policies in dataset management ..................................... 183
What conformance monitoring and correction is ........................................ 184
Datasets of physical storage objects ............................................................ 184
Datasets of virtual objects ........................................................................... 185
Data ONTAP licenses used for protecting or provisioning data ................. 195
Descriptions of dataset protection status ..................................................... 197
Configuring datasets ................................................................................................ 199
Adding a dataset of physical storage objects .............................................. 199
Adding a dataset of virtual objects .............................................................. 203
Editing a dataset to add virtual object members ......................................... 209
Editing a dataset to assign storage service and remote protection of
virtual objects ........................................................................................ 210
Editing a dataset of virtual objects to configure local policy and local
backup .................................................................................................... 211
Editing a dataset containing virtual objects to reschedule or modify local
backup jobs ............................................................................................ 212
Editing a dataset to remove protection from a virtual object ...................... 213
Adding a dataset of physical storage objects with dataset-level custom
naming ................................................................................................... 214
Adding a dataset of virtual objects with dataset-level custom naming ....... 215
Editing a dataset of virtual objects for dataset-level custom naming .......... 217
Editing a dataset of physical storage objects for dataset-level custom
naming ................................................................................................... 218
Selecting virtual objects to create a new dataset ......................................... 219
Selecting virtual objects to add to an existing dataset ................................. 220
Configuring local backups for multiple datasets of virtual Hyper-V
objects .................................................................................................... 221
Managing datasets ................................................................................................... 222
Performing an on-demand backup of a dataset ........................................... 222
Deleting a dataset of virtual objects ............................................................ 226
Suspending dataset protection and conformance checking ......................... 226
Resuming protection and conformance checking on a suspended dataset .. 227
Changing a storage service on datasets of storage objects .......................... 228
Attaching a storage service to existing datasets of storage objects ............. 229
Restoring data backed up from a dataset of physical storage objects ......... 230
Repairing datasets that contain deleted virtual objects ............................... 231

Table of Contents | 7
Evaluating and resolving issues displayed in the Conformance Details
dialog box .............................................................................................. 232
Monitoring datasets ................................................................................................. 236
Overview of dataset status types ................................................................. 236
How to evaluate dataset conformance to policy .......................................... 239
Monitoring dataset status ............................................................................ 244
Monitoring backup and mirror relationships ............................................... 245
Listing nonconformant datasets and viewing details .................................. 246
Evaluating and resolving issues displayed in the Conformance Details
dialog box .............................................................................................. 246
Page descriptions ..................................................................................................... 251
Datasets tab ................................................................................................. 251
Create Dataset dialog box or Edit Dataset dialog box ................................ 258

Backups ..................................................................................................... 269


Understanding backups ........................................................................................... 269
Types of backups ......................................................................................... 269
Backup version management ...................................................................... 269
Backup scripting information ...................................................................... 270
Retention of job progress information ........................................................ 271
Guidelines for mounting or unmounting backups in a VMware
environment ........................................................................................... 271
How the Hyper-V plug-in uses VSS ........................................................... 272
How the Hyper-V plug-in handles saved-state backups ............................. 274
Overlapping policies and Hyper-V hosts .................................................... 274
Co-existence of SnapManager for Hyper-V with the Hyper-V plug-in ...... 275
How to manually transition SnapManager for Hyper-V dataset
information ............................................................................................ 275
Managing backups ................................................................................................... 276
Performing an on-demand backup of virtual objects .................................. 276
Performing an on-demand backup of a dataset ........................................... 280
On-demand backups using the command-line interface ............................. 283
Mounting or unmounting backups in a VMware environment ................... 284
Manually mounting or unmounting backups in a Hyper-V environment
using SnapDrive for Windows .............................................................. 288
Locating specific backups ........................................................................... 292
Deleting a dataset backup ............................................................................ 293

8 | OnCommand Console Help


Monitoring backups ................................................................................................. 294
Monitoring local backup progress ............................................................... 294
Page descriptions ..................................................................................................... 294
Backups tab ................................................................................................. 294

Restore ....................................................................................................... 297


Understanding restore ............................................................................................. 297
Restoring data from backups ....................................................................... 297
Restore scripting information ...................................................................... 298
Managing restore ..................................................................................................... 299
Restoring data from backups created by the OnCommand console ............ 299
Monitoring restore ................................................................................................... 302
Viewing restore job details .......................................................................... 302

Reports ...................................................................................................... 303


Understanding reports ............................................................................................. 303
Reports management ................................................................................... 303
Reports tab customization ........................................................................... 304
Types of object status .................................................................................. 305
What report scheduling is ............................................................................ 305
Managing reports ..................................................................................................... 306
Scheduling reports ....................................................................................... 306
Viewing the scheduled reports log .............................................................. 307
Sharing reports ............................................................................................ 307
Deleting a report .......................................................................................... 308
Page descriptions ..................................................................................................... 308
Reports tab ................................................................................................... 308
Schedule Report dialog box ........................................................................ 310
Share Report dialog box .............................................................................. 311
Events reports .......................................................................................................... 311
Understanding events reports ...................................................................... 312
Page descriptions ......................................................................................... 312
Inventory reports ..................................................................................................... 314
Understanding inventory reports ................................................................. 314
Page descriptions ......................................................................................... 314
Storage capacity reports .......................................................................................... 323
Understanding storage capacity reports ...................................................... 323
Monitoring storage capacity reports ............................................................ 328

Table of Contents | 9
Page descriptions ......................................................................................... 338
Database schema ..................................................................................................... 359
How to access DataFabric Manager server data ......................................... 359
Supported database views ........................................................................... 360
alarmView ................................................................................................... 361
cpuView ...................................................................................................... 362
designerReportView .................................................................................... 363
Database view datasetIOMetricView .......................................................... 363
Database view datasetSpaceMetricView .................................................... 364
Database view datasetUsageMetricCommentView .................................... 366
hbaInitiatorView .......................................................................................... 367
hbaView ...................................................................................................... 367
initiatorView ................................................................................................ 367
reportOutputView ........................................................................................ 367
sanhostlunview ............................................................................................ 368
usersView .................................................................................................... 368
volumeDedupeDetailsView ........................................................................ 369

Administration .......................................................................................... 371


Users and roles ........................................................................................................ 371
Understanding users and roles ..................................................................... 371
Configuring users and roles ......................................................................... 375
Groups ..................................................................................................................... 375
Understanding groups ................................................................................. 375
Configuring groups ..................................................................................... 377
Managing groups ......................................................................................... 379
Page descriptions ......................................................................................... 381
Alarms ..................................................................................................................... 385
Understanding alarms .................................................................................. 385
Configuring alarms ...................................................................................... 385
Managing events and alarms ....................................................................... 388
Page descriptions ......................................................................................... 390
Host services ........................................................................................................... 393
Understanding host services ........................................................................ 393
Configuring host services ............................................................................ 394
Managing host services ............................................................................... 400
Monitoring host services ............................................................................. 407

10 | OnCommand Console Help


Page descriptions ......................................................................................... 409
Storage systems users .............................................................................................. 411
Understanding storage system users ........................................................... 411
Configuring storage system users ............................................................... 411
Storage system configuration .................................................................................. 412
Understanding storage system configuration .............................................. 412
Configuring storage systems ....................................................................... 413
vFiler configuration ................................................................................................. 414
Understanding vFiler unit configuration ..................................................... 414
Configuring vFiler units .............................................................................. 415
Options .................................................................................................................... 415
Page descriptions ......................................................................................... 415
Backup setup options .................................................................................. 419
Global naming settings setup options .......................................................... 421
Costing setup options .................................................................................. 445
Database backup setup options ................................................................... 448
Default thresholds setup options ................................................................. 453
Discovery setup options .............................................................................. 462
File SRM setup options ............................................................................... 469
LDAP setup options .................................................................................... 473
Monitoring setup options ............................................................................ 478
Management setup options .......................................................................... 491
Systems setup options ................................................................................. 496

Security and access ................................................................................... 505


Understanding RBAC ............................................................................................. 505
What RBAC is ............................................................................................. 505
How RBAC is used ..................................................................................... 505
How roles relate to administrators .............................................................. 505
Example of how to use RBAC to control access ........................................ 505
Administrator roles and capabilities ............................................................ 506
Access permissions for the Virtual Infrastructure Administrator role ........ 508
Understanding authentication .................................................................................. 509
Authentication methods on the DataFabric Manager server ....................... 509
Authentication with LDAP .......................................................................... 509

Plug-ins ...................................................................................................... 511


Hyper-V troubleshooting ......................................................................................... 511

Table of Contents | 11
Error: Vss Requestor - Backup Components failed with partial writer
error. ...................................................................................................... 511
Error: Failed to start VM. Job returned error 32768 ................................... 512
Error: Failed to start VM. You might need to start the VM using HyperV Manager ............................................................................................. 512
Error: Vss Requestor - Backup Components failed. An expected disk did
not arrive in the system .......................................................................... 512
Error: Vss Requestor - Backup Components failed. Writer Microsoft
Hyper-V VSS Writer involved in backup or restore encountered a
retryable error ........................................................................................ 513
Hyper-V virtual objects taking too long to appear in OnCommand
console ................................................................................................... 514
Increasing SnapDrive operations timeout value in the Windows registry . . 514
MBR unsupported in the Hyper-V plug-in ................................................. 514
Some types of backup failures do not result in partial backup failure ........ 515
Space consumption when taking two snapshot copies for each backup ..... 515
Virtual machine snapshot file location change can cause the Hyper-V
plug-in backup to fail ............................................................................. 516
Virtual machine backups taking too long to complete ................................ 516
Virtual machine backups made while a restore operation is in progress
might be invalid ..................................................................................... 516
Volume Shadow Copy Service error: An internal inconsistency was
detected in trying to contact shadow copy service writers. ................... 517
Hyper-V VHDs do not appear in the OnCommand console ....................... 518

Copyright information ............................................................................. 519


Trademark information ........................................................................... 521
How to send your comments .................................................................... 523
Index ........................................................................................................... 525

12 | OnCommand Console Help

13

About this document


This document is a printable version of the OnCommand console Help. It is intended to be used for
offline searches when you do not have access to the Help on a management station. The Help
contains administrative tasks, as well as conceptual and reference material that can be useful in
understanding how to use the OnCommand console.

14 | OnCommand Console Help

15

Welcome to OnCommand console Help


How to use OnCommand console Help
This Help includes information for all features included in the OnCommand console.
By using the table of contents, the index, or the search tool, you can find information about features
and how to use them.
Help is available from each tab and from the menu bar of the OnCommand console, as follows:

To learn about a specific parameter, click


.
To view all the Help contents, click the Help menu and select Contents.
You can expand any portion of the Table of Contents in the navigation pane to find more detailed
information.

You can also print selected Help topics.


Note: The search tool does not work for partial terms, only whole words.

Bookmarking your favorite topics


In the Help Favorites tab, you can add bookmark links to Help topics that you use frequently. Help
bookmarks provide fast access to your favorite topics.
Steps

1. Navigate to the topic you want to add as a favorite.


2. Click the Favorites tab, then click Add.

Understanding how the OnCommand console works


About the OnCommand console
The OnCommand console provides a centralized Web interface from which you can flexibly and
efficiently manage your physical and virtual storage infrastructure.

Window layout and navigation


Most windows in the OnCommand console have the same general layout.
Not every window contains every element in the following diagram.

16 | OnCommand Console Help

Menu bar

File

View

Administration

Help

Groups:

Dashboard
panel tabs

Dashboard

List of
related
objects

List of views
Breadcrumb
trail

Command
buttons
List of
objects
Details
for the
selected
object

Tabs

Use a minimum display setting of 1280 by 1024 pixels.


By default, the Dashboard tab, Events tab, Storage tab, Server tab, Policies tab, Datasets tab, and
Reports tab are open when you first log in to the OnCommand console.

Window layout customization


The OnCommand console enables you to customize the window layout. By customizing the
windows, you can control which data is viewable or how it is displayed.
Sorting

You can click the column headings to sort the column entries in ascending order
and display the sort arrows (
and
specify the order in which entries appear.

Filtering

). You can then use the sort arrows to

You can use the filter icon ( ) to display only those entries that match the
conditions provided. You can use the character filter (?) or string filter (*) to
narrow your search. You can apply filters to one or more columns. The column
heading is highlighted if a filter is applied. For example, you can search for
alarms configured for a particular event type: Aggregate Overcommitted. In the
Alarms tab, you can use the filter in the Event column. You can use the string
filter to search for alarms configured for the event "Aggregate Overcommitted."
In the string filter, when you type *aggr, all events whose names start with "aggr"
are listed.

Welcome to OnCommand console Help | 17

Note: If an entry in the column contains "?" or "*", to use the character filter or

string filter, you must enclose "?" or "*" in square brackets.


Hiding or
redisplaying
the columns
Customizing
the layout

You can click the column display icon (


display.

) to select the columns you want to

You can drag the bottom of the "list of objects" area up or down to resize the
main areas of the window. You can also choose to display or hide the "list of
related objects" and "list of views" panels. You can drag vertical dividers to
resize the width of columns or other areas of the window.

How the OnCommand console works with the Operations Manager console
and NetApp Management Console
The OnCommand console provides centralized access to a variety of storage capabilities. While you
can perform most virtualization tasks directly in the OnCommand console graphical user interface,
many physical storage tasks require the Operations Manager console or NetApp Management
Console.
The OnCommand console automatically launches these other consoles when they are required to
complete a task. You must install NetApp Management Console separately. You can also access the
Operations Manager console from the OnCommand console File menu at any time.

Launching the Operations Manager console


You can launch the Operations Manager console from the OnCommand console to perform many of
your physical storage tasks.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

During this task, the OnCommand console launches the Operations Manager console. Depending on
your browser configuration, you can return to the OnCommand console by using the Alt-Tab key
combination or clicking the OnCommand console browser tab. After the completion of this task, you
can leave the Operations Manager console open, or you can close it to conserve bandwidth.
Step

1. Click the File menu, then click Operations Manager.


The Operations Manager console opens in a separate browser tab or window.

18 | OnCommand Console Help


Related references

Administrator roles and capabilities on page 506

Installing NetApp Management Console


You can download and install NetApp Management Console through the OnCommand console.
NetApp Management Console is required to perform many of your physical storage tasks.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

During this task, the OnCommand console launches the Operations Manager console. Depending on
your browser configuration, you can return to the OnCommand console by using the Alt-Tab key
combination or clicking the OnCommand console browser tab. After the completion of this task, you
can leave the Operations Manager console open, or you can close it to conserve bandwidth.
Steps

1. Log in to the OnCommand console if necessary.


2. Click the File menu, then click Download Management Console.
A separate browser tab or window opens to the Management Console Software page in the
Operations Manager console.
3. Click the download link for the Linux or Windows installation.
4. In the download dialog box, click Save File.
The executable file is downloaded to your local system, from the system on which the
OnCommand Core Package was installed.
5. From the download directory, run the nmconsole-setup-xxx.xxx executable file.
The NetApp Management Console installation wizard opens.
6. Follow the prompts to install NetApp Management Console.
Result

After installation, you can access NetApp Management Console from the following locations:

On Windows systems, the default installation path is C:\Program Files\NetApp\Management


Console.
You can launch the console from the NetApp directory on the Start menu.
On Linux systems, the default installation path is /usr/lib/ NetApp/management_console/.
You can launch the console from /usr/bin.

Welcome to OnCommand console Help | 19


Related references

Administrator roles and capabilities on page 506

How the OnCommand console works with AutoSupport


If you have AutoSupport enabled, weekly messages that contain information about your operating
environment are automatically sent to your internal support organization, NetApp support personnel,
or both.
If you have AutoSupport enabled for the OnCommand console, the weekly AutoSupport messages
contain accounting information stored by DataFabric Manager server in addition to any other
information that is sent by other applications.
The information sent from DataFabric Manager server includes, but is not limited to, the following
counts:

The total number of host services registered with DataFabric Manager server.
Of the total number of host services, the number of VMware host services and the number of
Hyper-V host services.
The total number of host services with pending authorization.
The total number of storage systems that have host services connected to them. This is the total
number of unique storage systems known to all host services registered with DataFabric Manager
server.
Of the total number of storage systems that have host services connected to them, the number of
each FAS system model.
The total number of vFilers that are connected to host services.
The total number of VMware virtual centers.
The total number of VMware datacenters.
The total number of virtual machines.
Of the total number of virtual machines, the number of VMware virtual machines and the number
of Hyper-V virtual machines.
The total number of hypervisors.
Of the total number of hypervisors, the number of VMware hypervisors.
The total number of Hyper-V parents.
The total number of datastores.
Of the total number of datastores, the number of SAN datastores and the number of NAS
datastores.
The maximum number of virtual machines in a datastore.
The minimum number of virtual machines in a datastore.
The average number of virtual machines per datastore.
The maximum number of virtual machines on an ESX server.
The minimum number of virtual machines on an ESX server.
The average number of virtual machines per ESX server.
The maximum number of virtual machines on a Hyper-V server.

20 | OnCommand Console Help

The minimum number of virtual machines on a Hyper-V server.


The average number of virtual machines per Hyper-V server.
The total number of virtual machines that span datastores.
The total number of some other types of objects, including Open Systems SnapVault (OSSV)
relationships, backup jobs, restore jobs, mount jobs, unmount jobs, and various types of datasets.

21

Dashboard
Understanding the dashboard
OnCommand console dashboard panels
The OnCommand console dashboard contains multiple panels that provide cumulative at-a-glance
information about your storage and virtualization environment. The dashboard provides various
aspects of your storage management environment, such as the availability of storage objects, events
generated for storage objects, resource pools, and dataset overall status.
The following panels are available in the OnCommand console dashboard:
Availability

Provides information about the availability of storage controllers (standalone controllers and HA pairs) and vFiler units that are discovered and
monitored. You can also view the number of controllers and vFiler units that
are either online or offline.

Events

Provides information about the status of the objects by listing the top five
events, based on their severity.

Resource Pools

Displays the resource pools that have existing or potential space shortages.

Full Soon Storage

Displays the top five aggregates and volumes that are likely reach a
configured threshold, based on the number of days before this threshold is
reached. You can also view the trend and space utilization of a particular
aggregate or volume.

Fastest Growing
Storage

Displays the top five aggregates and volumes whose space usage is rapidly
increasing. You can also view the growth rate, trend, and space utilization of
a particular aggregate or volume.

Dataset Overall
Status

Displays the number of datasets in one of the following statuses: Error,


Warning, or Normal.

External
Relationship Lags

Displays the relative percentages of external SnapVault, qtree SnapMirror,


and volume SnapMirror relationships with lag times in Error, Warning, and
Normal status.

Unprotected Data

Displays the number of unprotected storage and virtual objects that are being
monitored.

Get Started

Enables you to navigate to the Getting Started with NetApp Software page in
the NetApp University Web site.

22 | OnCommand Console Help


Related information

Getting Started with NetApp Software - http://communities.netapp.com/community/


netapp_university/getting_started

Monitoring the dashboard


Monitoring dashboard panels
You can use the dashboard panels to monitor your physical storage, logical storage, virtual storage
and non-storage objects.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Log in to the OnCommand console. By default, the Dashboard tab is displayed.


2. To view details about any of the information displayed in the dashboard panels, click the panel
heading to display the relevant OnCommand console tab.
You can also click any hypertext links in the individual dashboard panels to view detailed
information.
Related references

Administrator roles and capabilities on page 506

Page descriptions
Availability dashboard panel
This panel provides information about the availability of storage controllers (stand-alone controllers
and HA pairs) and vFiler units that are discovered and monitored by the OnCommand console.
Panel display
icon links you to the Storage tab where you can access details about your storage controllers
The
and vFiler units.
Controllers Displays the percentage of storage controllers that are either online or offline. For
example, when ten storage controllers are being monitored and all the controllers are

Dashboard | 23
online, the Controllers area displays 100% up. If five controllers are offline, the
Controllers area displays 50% down. You can also view the number of storage
controllers that are online.
You can view more information about the storage controllers by clicking in the
Controllers area.
vFiler Units Displays the percentage of vFiler units that are either online or offline. For example,
when ten vFiler units are being monitored and all the vFiler units are online, the
vFiler Units area displays 100% up. If five vFiler units are offline, the vFiler Units
area displays 50% down. You can also view the number of vFiler units that are
online.
You can view more information about the vFiler units by clicking in the vFiler Units
area.

Events dashboard panel


This panel provides information about the status of the managed objects and the management host by
listing the top five events that are generated by OnCommand console. The Events dashboard panel
displays only current events. The top five events are listed based on their severity levels. The highest
severity level is represented as Emergency.
Panel display
The

icon links you to the Events tab where you can view a list of events and their properties.

Events The events are displayed in the order of their severity as follows:
Emergency

The event source unexpectedly stopped performing and experienced


unrecoverable data loss. You must take corrective action immediately to
avoid extended downtime.

Critical

A problem occurred that might lead to service disruption if corrective


action is not taken immediately.

Error

The event source is still performing; however, corrective action is


required to avoid service disruption.

Warning

The event source experienced an occurrence that you should be aware of.
Events of this severity do not cause service disruption, and corrective
action might not be required.

Information The event may be of interest to the administrator. This severity does not
represent an abnormal operation.
By clicking the specific event, you can view more information about the event from the
Events tab.

24 | OnCommand Console Help

Full Soon Storage dashboard panel


This panel displays the top five aggregates and volumes that will soon reach the configured
threshold, based on the number of days before this threshold is reached. This enables you to take
corrective action to prevent the aggregates or volumes from reaching the threshold.
Panel display
icon links you to the Storage tab where you can view the aggregates and volumes
The
approaching the specified threshold.
Resource

Lists the volumes or aggregates that will soon reach the specified threshold. By
clicking the name of a resource, you can view more information about it in the
Volumes view or the Aggregates view, depending on the type of resource you
select.

Days to Full

Displays the estimated number of days remaining before the volume or


aggregate reaches the threshold.

Trend

Displays, as a trend line, information about the space used in the volume or the
aggregate for the past 30 days.

Space
Utilization

Displays the amount of space used within the volume or aggregate.

Fastest Growing Storage dashboard panel


This panel displays the top five aggregates and volumes that are using the most space. You can also
view the growth rate, trend, and space utilization.
Panel display
The
icon links you to the Storage tab where you can view the aggregates and volumes using the
most space.
Resource

Displays the five fastest-growing aggregates and volumes. By clicking the name
of a resource, you can view more information about it in the Volumes view or
the Aggregates view, depending on the type of resource you select.

Growth Rate
(%)

Displays the percent growth rate of the space used by the fastest-growing storage
systems. The growth rate is determined by dividing the daily growth rate by the
total amount of space in the storage system.

Trend

Displays, as a trend line, information about the space used in the aggregate or
volume for the past 30 days.

Space
Utilization

Displays the amount of space used within the volume or aggregate.

Dashboard | 25

Dataset Overall Status dashboard panel


This panel summarizes for you the number of datasets in overall Error, Warning, or Normal status.
Panel display
icon links you to the Datasets tab where you can examine the details of a datasets overall
The
status.
This panel displays the number of datasets in overall Error status, overall Warning status, or overall
Normal status.
Error

The number of datasets with an overall status of Error. A dataset is designated with
overall error status based upon the following status values:
DR status condition: Error
Protection status condition: Lag error or Baseline failed
Conformance status condition: Nonconformant
Space status condition: Error
Resource status condition: Emergency, Critical, or Error

Warning The number of datasets with an overall status of Warning. A dataset is designated with
overall warning status based upon the following status values:
DR status condition: Warning
Protection status condition: Job failure, Lag warning, Uninitialized, or No protection
policy for a non-empty dataset
Conformance status condition: NA
Space status condition: Warning
Resource status condition: Warning
Normal

The number of datasets with an overall status of Normal.

Resource Pools dashboard panel


This panel displays the total space allocated to and the percentage of space utilized by each resource
pool, listed by resource pool name.
Panel display
The
icon links you to the Resource Pools window in the NetApp Management Console where
you can access details about individual resource pools.
For each existing resource pool, this panel displays the following information:

26 | OnCommand Console Help


Name

The name of the resource pool

Total Size

The total size in kilobytes, megabytes, gigabytes, or terabytes of the resource


pool

Space Utilization The percentage of the resource pool's capacity that is being utilized
Items are sorted in decreasing order of available space.

External Relationship Lags dashboard panel


This panel summarizes for you the relative percentages of external SnapVault, Qtree SnapMirror, and
Volume SnapMirror relationships with lag times of Normal, Warning, and Error status.
Panel display
The

icon links you to the External Relationships window in the NetApp Management Console.

This panel uses colored bars to indicate the relative percentages of external SnapVault, Qtree
SnapMirror, and volume SnapMirror relationships with lag times in Error, Warning, and Normal
status.
The Warning and Error status percentages indicate the portion of external SnapVault, Qtree
SnapMirror, and Volume SnapMirror relationships whose current lag times have exceeded the time
specified in the global Warning and Error threshold settings for those relationships. Normal status
percentages indicate the portion of external relationships whose current lag times are still within
normal range.
External relationships are protection relationships that are monitored but not managed by the
OnCommand console. The lag is the time since the last successful data update associated with an
external protection relationship was completed.

Unprotected Data dashboard panel


This panel summarizes the total number of storage objects (volumes and qtrees) and virtual objects
(Hyper-V virtual machines, VMware virtual machines, VMware datastores, and VMware
datacenters) that are not protected.
Panel displays
Displays the number of unprotected objects being monitored.
Volumes

Displays the number of unprotected volumes in your domain. The hypertext links
you to the Unprotected Data window in the NetApp Management Console.

Qtrees

Displays the number of unprotected qtrees in your domain. The hypertext links you
to the Unprotected Data window in the NetApp Management Console.

Hyper-V VMs Displays the number of unprotected Hyper-V virtual machines in your domain.
The hypertext links you to the Hyper-V VMs view of the Server tab.

Dashboard | 27
VMware VMs Displays the number of unprotected VMware virtual machines in your domain.
The hypertext links you to the VMware VMs view of the Server tab.
Datastores

Displays the number of unprotected VMware datastores in your domain. The


hypertext links you to the VMware Datastores view of the Server tab.

Datacenters

Displays the number of unprotected VMWare datacenters in your domain. The


hypertext links you to the VMware Datacenters view of the Server tab.

Storage objects are unprotected if they do not belong to a dataset or if they belong to an unprotected
dataset. Datasets are unprotected if they do not have an assigned protection policy or if they have an
assigned protection policy but do not have an initial relationship created (the dataset has never
conformed to the protection policy).
Virtual objects are unprotected if they do not belong to a dataset that has been assigned a local
policy.

28 | OnCommand Console Help

29

Events and alarms


Understanding events and alarms
What events are
Events are generated automatically when a predefined condition occurs or when an object crosses a
threshold. Event messages inform you when specific events occur. All events are assigned a severity
type and are automatically logged in the Events tab.
You can configure alarms to send notification automatically when specific events or severity types
occur. If an application is not configured to trigger an alarm when an event is generated, you can find
out about the event by checking the Events tab.
It is important that you take immediate corrective action for events with severity level Error, Critical,
or Emergency. Ignoring such events can lead to poor performance and system unavailability.
Note: Event types are predetermined. Although you cannot add or delete event types, you can
manage notification of events. However, you can modify the event severity type from the
DataFabric Manager server command-line interface.

You can use the Events tab to acknowledge and resolve events, and also create alarms for specific
events.

What alarms are


Alarms are configured notifications that are sent whenever a specific event or an event of a specific
severity type occurs. You can create alarms for any defined group or the global group for which you
want automatic notification of events.
Alarms are not the events themselves, only the notification of events. Alarms are not the same as user
alerts.
Note: By default, the DataFabric Manager server sends user alerts (e-mail messages) to all users
who exceed their quota limits.

You can use the Alarms tab to create, edit, delete, test, and enable or disable alarms.
Related concepts

What a global group is on page 376

30 | OnCommand Console Help

Guidelines for creating alarms


You can create alarms to notify you whenever a specific event or an event of a specific severity type
occurs. You can create alarms based on the event type, event severity, or event class. You can use the
Alarms tab or the Events tab to create a new alarm.

Group
You can create alarms only at the group level. You must decide the group for which the alarm is
added. If you want to set an alarm for a specific object, you must first create a group with that
object as the only member. For example, if you want to closely monitor a single aggregate by
configuring an alarm, you must create a group, and add the aggregate into the group. You can
then configure an alarm for the newly created group.
Note: By default, there exists a global group, and all objects and groups belong to the global
group.

Event
If you add an alarm based on the type of event generated, you should decide which events require
an alarm.
Event severity
You should decide if any event of a specified severity type should trigger the alarm and, if so,
which severity type.
Event class
You can configure a single alarm for multiple events using event class. If you add an alarm based
on the event class, you should decide if an event in an event class should trigger the alarm and, if
so, which event class. For example, the expression userquota.*|qtree.* matches all user
quota or qtree events.
Note: You can view the list of event classes from the CLI, by using the following command:
dfm eventType list. You can view the list of events specific to an event class by using the
following command: dfm eventType list -Cevent-class

Modes of event notification


You should decide who or what needs to receive the event notification. You can specify one or
more of the following modes of event notification:

E-mails
You must provide the administrator user names or e-mail addresses of users other than the
administrator.
Pagers
You must provide the user names of the administrators or pager numbers of the
nonadministrator users.
Note: You must ensure that proper e-mail addresses and pager numbers of administrators
and nonadministrator users are configured.

SNMP listener traps

Events and alarms | 31

You must provide the SNMP traphost. Optionally, you should provide the SNMP community
name.
Script
You must provide the complete path of a script that is executed when an alarm occurs and the
user name that runs the script.
Effective time for repeat notification
You can configure an alarm to repeatedly send notification to the recipients for a specified time.
You should determine the time from which the event notification is active for the alarm. If you
want the event notification repeated until the event is acknowledged, you should determine how
often you want the notification to be repeated.

How to know when an event occurs


You can learn about event occurrences by viewing the events list or by configuring alarms to
automatically notify you when events occur.

Viewing the events list


You can use the Events tab to view a list of all events that occurred and to view detailed
information about any selected event.
Configuring alarms
You can use the Alarms tab to add an alarm that sends notifications automatically when an event
occurs.

Description of event severity types


Each event is associated with a severity type to help you prioritize the events that require immediate
corrective action.
Emergency The event source unexpectedly stopped performing and experienced unrecoverable
data loss. You must take corrective action immediately to avoid extended downtime.
Critical

A problem occurred that might lead to service disruption if corrective action is not
taken immediately.

Error

The event source is still performing; however, corrective action is required to avoid
service disruption.

Warning

The event source experienced an occurrence that you should be aware of. Events of
this severity do not cause service disruption, and corrective action might not be
required.

Information The event occurs when a new object is discovered, or when a user action is
performed. For example, when a group is created, an alarm is configured, or when a
storage system is added, the event with severity type Information is generated. No
action is required.
Normal

A previous abnormal condition for the event source returned to a normal state and the
event source is operating within the desired thresholds.

32 | OnCommand Console Help

Alarm configuration
DataFabric Manager server uses alarms to notify you when events occur. DataFabric Manager server
sends the alarm notification to one or more specified recipients in different formats, such as e-mail
notification, pager alert, an SNMP traphost, or a script you wrote (you should attach the script to the
alarm).
You should determine the events that cause alarms, whether the alarm repeats until it is
acknowledged, and how many recipients an alarm has. Not all events are severe enough to require
alarms, and not all alarms are important enough to require acknowledgment. Nevertheless, to avoid
multiple responses to the same event, you should configure DataFabric Manager server to repeat
notification until an event is acknowledged.
Note: DataFabric Manager server does not automatically send alarms for the events.

Configuring alarms
Creating alarms for events
The OnCommand console enables you to configure alarms for immediate notification of events. You
can also configure alarms even before a particular event occurs. You can add an alarm based on the
event, event severity type, or event class from the Create Alarm dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
You must have your mail server configured so that the DataFabric Manager server can send e-mails
to specified recipients when an event occurs.
You must have the following information available to add an alarm:

The group with which you want the alarm associated.


The event name, event class, or event severity type that triggers the alarm.
The recipients and the modes of event notifications.
The period during which the alarm is active.

You must have the following capabilities to perform this task:

DFM.Event.Write
DFM.Alarm.Write

About this task

Alarms you configure based on the event severity type are triggered when that event severity level
occurs.

Events and alarms | 33


Steps

1. Click the Administration menu, then click the Alarms option.


2. From the Alarms tab, click Create.
3. In the Create Alarm dialog box, specify the condition for which you want the alarm to be
triggered.
Note: An alarm is configured based on event type, event severity, or event class.

4. Specify either one or more means of alarm notification properties.


5. Click Create, then click Close.
Related concepts

Guidelines for creating alarms on page 30


Related tasks

Configuring the mail server for alarm notifications on page 36


Related references

Administrator roles and capabilities on page 506

Creating alarms for a specific event


The OnCommand console enables you to configure an alarm when you want immediate notification
for a specified event name or event class, or if events (of a specified severity level) occur. You can
add an alarm from the Create Alarm dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
You must have your mail server configured so that the DataFabric Manager server can send e-mails
to specified recipients when an event occurs.
You must have the following information available to add an alarm:

The group with which you want the alarm associated


The event name, event class, or severity type that triggers the alarm
The recipients and the modes of event notifications
The period during which the alarm is active

You must have the following capabilities to perform this task:

DFM.Event.Write
DFM.Alarm.Write

34 | OnCommand Console Help


About this task

Alarms you configure for a specific event are triggered when that event occurs.
Steps

1. Click the View menu, then click the Events option.


2. From the events list in the Events tab, analyze the events, and then determine the event for which
you want to create an alarm notification.
3. Select the event for which you want to create an alarm.
4. Click Create Alarm.
The Create Alarm dialog box opens, and the event is selected by default.
5. Specify either one or more means of alarm notification properties.
6. Click Create, then click Close.
Related concepts

Guidelines for creating alarms on page 30


Related tasks

Configuring the mail server for alarm notifications on page 36


Related references

Administrator roles and capabilities on page 506

Managing events and alarms


Resolving events
After you have taken corrective action of a particular event, you should mark the event as resolved to
avoid multiple event notifications. You can mark the events as resolved from the Events tab.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu, then click the Events option.


2. From the events list in the Events tab, select the event that you want to acknowledge.

Events and alarms | 35


3. Click Acknowledge.
When you do not acknowledge and mark an event resolved, you will receive multiple event
notifications for the same event.
4. Find the cause of the event and take corrective action.
5. Click Resolve to mark the event as resolved.
Related references

Administrator roles and capabilities on page 506

Editing alarm properties


You can edit the configuration of an existing alarm from the Edit Alarm dialog box. For example, if
you have created a script that is executed when there is an event notification, you can provide the
complete script path in Edit Alarm dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
You must have the following capabilities to perform this task:

DFM.Event.Write
DFM.Alarm.Write

Steps

1. Click the Administration menu, then click the Alarms option.


2. From the alarms list, select the alarm whose properties you want to modify.
3. From the Alarms tab, click Edit.
4. In the Edit Alarm dialog box, edit the properties of the alarm as required.
5. Click Edit.
Result

The new configuration is immediately activated and displayed in the alarms list.
Related references

Administrator roles and capabilities on page 506

36 | OnCommand Console Help

Configuring the mail server for alarm notifications


You must configure the mail server so that when an event occurs the DataFabric Manager server can
send e-mails to specified recipients.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

During this task, the OnCommand console launches the Operations Manager console. Depending on
your browser configuration, you can return to the OnCommand console by using the Alt-Tab key
combination or clicking the OnCommand console browser tab. After the completion of this task, you
can leave the Operations Manager console open, or you can close it to conserve bandwidth.
Steps

1. From the File menu, click Operations Manager.


2. In the Operations Manager console, click the Control Center tab.
3. Click Setup menu, and then click Options.
4. In Edit options, click the Events and Alerts option.
5. In the Events and Alerts Options page, specify the name of your mail server.
Related references

Administrator roles and capabilities on page 506

Monitoring events and alarms


Viewing event details
You can view the details of each event such as the source of the event, the event type, the condition
that triggered the event, the time and date on which the event was triggered, and event severity, from
the Events tab. You can also view common details of multiple events.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.

Events and alarms | 37


Steps

1. Click the View menu, then click the Events option.


2. From the Events tab, select the event to view the details.
You can view the event details in the Details area.
Related references

Administrator roles and capabilities on page 506

Viewing alarm details


You can view the list of alarms created for various events from the Alarms tab. You can also view
alarm properties such as the event name, severity of the event, and group associated with the alarm.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click the Alarms option.


2. From the Alarms tab, select the alarm to view the details.
You can view the alarm details in the Details area.
Related references

Administrator roles and capabilities on page 506

Page descriptions
Events tab
The Events tab provides a single location from which you can view a list of events and their
properties. You can perform various actions on these events such as navigating to the Alarms tab,
configuring alarms (by clicking the Manage alarms link), acknowledging, and resolving events.

Command buttons on page 38


Events list on page 38
Details area on page 39

38 | OnCommand Console Help


Command buttons
The command buttons enable you to perform the following management tasks for a selected event:
Acknowledge Acknowledges the selected events.
Your user name and the time are entered in the events list (Acknowledged By and
Acknowledged) for the selected events. When you acknowledge an event, you take
responsibility for managing that event.
Resolve

Resolves the selected events.


Your user name and the time are entered in the events list (Resolved By and
Resolved) for the selected events. After you have taken a corrective action for the
event, you must mark the event as resolved.
Note: You can view the resolved event in the Events All report.

Create Alarm Launches the Create Alarm dialog box in which you can create alarms for the
selected event.
Refresh

Refreshes the list of events.

Events list
The Events list displays a list of all the events that occurred. By default, the most recent events are
listed. The list of events is updated dynamically, as events occur. You can select an event to see the
details for that event.
ID

Displays the ID of the event. By default, this column is hidden.

Source ID

Displays the ID of the object with which the event is associated. By default,
this column is hidden.

Triggered

Displays the time and date the event was triggered.

Source

Displays the full name of the object with which the event is associated.

Event

Displays the event names. You can select an event to display the event details.

State

Displays the event state: New, Acknowledged, or Resolved.

Severity

Displays the severity type of the event. You can filter this column to show all
severity types. The event severity types are Normal, Information, Warning,
Error, Critical, and Emergency.

Acknowledged By Displays the name of the person who acknowledged the event. The field is
blank if the event is not acknowledged. By default, this column is hidden.
Acknowledged

Displays the date and time when the event was acknowledged. The field is
blank if the event is not acknowledged. By default, this column is hidden.

Events and alarms | 39


Resolved By

Displays the name of the person who resolved the event. This field is blank if
the event is not resolved. By default, this column is hidden.

Resolved

Displays the date and time at which the event was resolved. This field is blank
if the event is not resolved. By default, this column is hidden.

Current

Displays a "Yes" if the event is a current event, and displays a "No" if the
event is a history event.

Details area
Apart from the event details displayed in the events list, you can view other additional details of the
events in the area below the events list.
Event

Displays the event names. You can select an event to display the event details.

About

Brief description of the event.

Triggered

Displays the time and date the event was triggered.

State

Displays the event state: New, Acknowledged, or Resolved.

Severity

Displays the severity type of the event. You can filter this column to show all
severity types. The event severity types are Normal, Information, Warning, Error,
Critical, and Emergency.

Source

Displays the full name of the object with which the event is associated. By
clicking the source, you can view the details of the object from the corresponding
inventory view.

Type

The type of the source that triggered the event.

Condition

A description of the condition that triggered the event.

Notified

The date and time on which the event was notified.

Acknowledged Displays the date and time when the event was acknowledged. The field is blank if
the event is not acknowledged.
Resolved

Displays the date and time at which the event was resolved. This field is blank if
the event is not resolved.

Related references

Window layout customization on page 16

40 | OnCommand Console Help

Alarms tab
The Alarms tab provides a single location from which you can view a list of alarms configured based
on event, event severity type, and event class. You can also perform various actions from this
window, such as edit, delete, test, and enable or disable alarms.

Command buttons on page 40


Alarms list on page 40
Details area on page 41

Command buttons
The command buttons enable you to perform the following management tasks for a selected event:
Create

Launches the Create Alarm dialog box in which you can create an alarm based on event,
event severity type, and event class.

Edit

Launches the Edit Alarm dialog box in which you can modify alarm properties.

Delete

Deletes the selected alarm.

Test

Tests the selected alarm to check its configuration, after creating or editing the alarm.

Enable Enables an alarm to send notifications.


Disable Disables the selected alarm when you want to temporarily stop its functioning.
Refresh Refreshes the list of alarms.
Alarms list
The Alarms list displays a list of all the configured alarms. You can select an alarm to see the details
for that alarm.
Alarm ID

Displays the ID of the alarm.

Event

Displays the event name for which the alarm is created.

Event Severity

Displays the severity type of the event.

Group

Displays the group name with which the alarm is associated.

Enabled

Displays Yes if the selected alarm is enabled or No if the selected alarm is


disabled.

Start

Displays the time at which the selected alarm becomes active. By default, this
column is hidden.

End

Displays the time at which the selected alarm becomes inactive. By default, this
column is hidden.

Events and alarms | 41


Repeat Interval Displays the time period (in minutes) before the DataFabric Manager server
continues to send a repeated notification until the event is acknowledged or
(Minutes)
resolved. By default, this column is hidden.
Repeat Notify

Displays Yes if the selected alarm is enabled for repeated notification, or


displays No if the selected alarm is disabled for repeated notification. By
default, this column is hidden.

Event Class

Displays the class of event that is configured to trigger an alarm. By default, this
column is hidden.
You can configure a single alarm for multiple events using the event class. The
event class is a regular expression that contains rules, or pattern descriptions, that
typically use the word "matches" in the expression. For example, the
userquota.*|qtree.* expression matches all user quota or qtree events.

Details area
Apart from the alarm details displayed in the alarms list, you can view other additional properties of
the alarms in the area below the alarms list.
Effective Time Range

The time during which an alarm is active.

Administrators (Email
Address)

The e-mail address of the administrator, to which the alarm


notification is sent.

Administrators (Pager
Number)

The pager number of the administrator, to which the alarm


notification is sent.

Email Addresses (Others)

The e-mail addresses of nonadministrator users, to which the alarm


notification is sent.

Pager Numbers (Others)

The pager numbers of nonadministrator users, to which the alarm


notification is sent.

SNMP Trap Host

The SNMP traphost system that receives the alarm notification in the
form of SNMP traps.

Script Path

The name, along with the path of the script that is run when an alarm
is triggered.

Related references

Window layout customization on page 16

Create Alarm dialog box


The Create Alarm dialog box enables you to create alarms based on the event type, event severity, or
event class. You can create alarms for specific event or many events.

Event Options on page 42

42 | OnCommand Console Help

Notification Options on page 42


Command buttons on page 42

Event Options
You can create an alarm based on event name, event severity type, or event class:
Group

Displays the group that receives an alert when an event or event type triggers an
alarm.

Event

Displays the names of the events that triggers an alarm.

Event
Severity

Displays the severity types of the event that triggers an alarm. The event severity
types are Normal, Information, Warning, Error, Critical, and Emergency.

Event Class

Specifies the events classes that triggers an alarm.


The event class is a regular expression that contains rules, or pattern descriptions,
that typically use the word "matches" in the expression. For example, the
expression userquota.* matches all user quota events.

Notification Options
You can specify alarm notification properties by selecting one of the following check boxes:
SNMP Trap Host

Specifies the SNMP traphost that receives the notification.

E-mail Administrator
(Admin Name)

Specifies the name of the administrator who receives the e-mail


notification. You can specify multiple administrator names, separated
by commas.

Page Administrator
(Admin Name)

Specifies the administrator who receives the pager notification. You


can specify multiple administrator names, separated by commas.

E-mail Addresses
(Others)

Specifies the e-mail addresses of nonadministrator users who receive


the notification. You can specify multiple e-mail addresses, separated
by commas.

Pager Numbers (Others)

Specifies the pager numbers of other nonadministrator users who


receive the notification. You can specify multiple pager numbers,
separated by commas.

Script Path

Specifies the name of the script that is run when the alarm is triggered.

Repeat Interval
(Minutes)

Specifies whether an alarm notification is repeated until the event is


acknowledged, and, if so, how often the notification is repeated.

Effective Time Range

Specifies the time during which the alarm is active.

Command buttons
You can use command buttons to perform the following management tasks for a selected event:

Events and alarms | 43


Create

Creates an alarm based on the properties that you specify.

Cancel

Does not save the alarm configuration and closes the Create Alarm dialog box.

Edit Alarm dialog box


The Edit Alarm dialog box enables you to edit alarm properties such as group with which the alarm
is associated, event type, event severity, event class, and notification options.

Event Options on page 43


Notification Options on page 43
Command buttons on page 44

Event Options
You can edit alarm properties such as group with which the alarm is associated, event type, event
severity, or event class.
Group

Displays the group (and its subgroups) that receives an alert when an event or
event type triggers an alarm.

Event

Displays the name of the event that triggers an alarm.

Event
Severity

Displays the severity type of the event that triggers the alarm.

Event Class

Specifies the class of the event that triggers the alarm.


The event class is a regular expression that contains rules, or pattern descriptions,
that typically use the word "matches" in the expression. For example, the
expression userquota.* matches all user quota events.

Notification Options
You can edit alarm notification properties by selecting one of the following check boxes:
SNMP Trap Host

Specifies the SNMP traphost that receives the notification.

E-mail Administrator
(Admin Name)

Specifies the name of the administrator who receives the e-mail


notification of the event. You can specify multiple administrator
names, separated by commas.

Page Administrator
(Admin Name)

Specifies the administrator who receives the pager notification of the


event. You can specify multiple administrator names, separated by
commas.

E-mail Addresses
(Others)

Specifies the e-mail addresses of other users (other than the


administrators) who receive the notification. You can specify multiple
e-mail addresses separated by commas.

44 | OnCommand Console Help


Pager Numbers (Others) Specifies the pager numbers of other users (other than the
administrators) who receive the notification. You can specify multiple
pager numbers, separated by commas.
Script Path

Specifies the name of the script that is run when the alarm is triggered.

Repeat Interval
(Minutes)

Specifies whether an alarm notification is repeated until the event is


acknowledged, and how often the notification is repeated.

Effective Time Range

Specifies the time during which the alarm is active.

Command buttons
You can use command buttons to perform the following management tasks for a selected event:
Edit

Modifies the alarm properties that you specify.

Cancel Does not save the modification of alarm configuration, and closes the Edit Alarm dialog
box.

45

Jobs
Understanding jobs
Understanding jobs
A job is typically a long-running operation. The OnCommand console enables you to create, manage,
and monitor jobs. From the Jobs tab, you can view all jobs that are currently running as well as jobs
that have completed.
Following are three examples of a job:

A scheduled local backup of a dataset


A mirror transfer job
A mount or unmount of a VMware snapshot copy

Managing jobs
Canceling jobs
You can use the Jobs tab to cancel a job if it is taking too long to complete, encountering too many
errors, or is no longer needed. You can cancel a job only if its status and type allow it. You can
cancel any job that has the status Running or Running with failures.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu, then click the Jobs option.


2. From the list of jobs, select one or more jobs.
3. Click Cancel.
Note: The Cancel button is enabled only for jobs that are either Running or Running with

failures. If the Cancel button is not enabled, that job type cannot be canceled.
4. At the confirmation prompt, click Yes to cancel the selected job.

46 | OnCommand Console Help


Related references

Administrator roles and capabilities on page 506

Monitoring jobs
Monitoring jobs
You can monitor for job status and other job details using the Jobs tab. For example, you can view
the progress of an on-demand backup job and see whether there are any errors.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu, then click the Jobs option.


2. Select a job in the jobs list to see information about that job.
The Groups selection list in the toolbar enables you to display only the data that pertains to
objects in a selected group. This setting persists until you log out or choose a different group.
Related references

Administrator roles and capabilities on page 506

Page descriptions
Jobs tab
The Jobs tab enables you to view the current status and other information about all jobs that are
currently running as well as jobs that have completed. You can use this information to see which jobs
are still running and which jobs have succeeded. This tab displays up to 25,000 of the jobs in all
states.

Command buttons on page 47


View Jobs drop-down list on page 47
Jobs list on page 47
Completed job steps on page 51
Job details on page 52

Jobs | 47
Command buttons
Cancel Stops the selected jobs. You can select multiple jobs and cancel them simultaneously.
This button is enabled only for certain job types and when the selected jobs are running.
Refresh Updates the jobs list.
View Jobs drop-down list
Selecting these options displays all jobs that were started during the specified time range; these
displays do not include earlier but still in-process jobs. All ranges are based on a 24-hour day for
which 00:00 represents midnight.
1 Day

Displays all jobs that were started between midnight of the previous day and now. This
period can cover up to 47 hours and 59 minutes.
For example, if you click 1 Day at 15:00 on February 14 (on a 24-hour clock), the list
includes all jobs that were started from 00:00 (midnight) on February 13 to the current
time on February 14. This list covers the full day of February 13 plus the partial current
day of February 14.

1 Week

Displays all jobs that were started between midnight of the same day in the previous
week (seven days ago) and now. This period can cover up to seven days, 23 hours, and
59 minutes.
For example, if you click 1 Week at 15:00 on Thursday, February 14 (on a 24-hour
clock), the list includes all jobs that were started from 00:00 (midnight) the previous
Thursday (February 7) to the current time on February 14. This list covers seven full
days plus the partial current day.

1 Month Displays all jobs that were started between midnight of the same day in the previous
month and now. This period can cover from 28 through 32 days, depending on the
month.
For example, if you click 1 Month at 15:00 on Thursday, February 14 (on a 24-hour
clock), the list includes all jobs that were started from 00:00 (midnight) on January 14 to
the current time on February 14.
All

Displays all jobs.

Note: On very large or very busy systems, the Jobs tab might be unresponsive for long periods
while loading 1 Month or All data. If the application appears unresponsive for these large lists,
select a shorter time period (such as 1 Day).

Jobs list
Displays a list of the jobs that are in progress. You can customize the display by using the following
filtering and sorting options in the columns of the jobs list.
Note: You can display no more than 25,000 records simultaneously.

48 | OnCommand Console Help


Job

The identification number of the job. The default jobs list includes this column.
The job identification number is unique and is assigned by the server when it starts
the job. You can search for a particular job by entering the job identification
number in the text box provided by the column filter.

Job Type

The type of job, which is determined by the policy assigned to the dataset or by the
direct request initiated by a user. The default jobs list includes this column. The
job types are as follows:
Backup Deletion

A job that deletes backups of volumes of a dataset.

Backup Mount

An operation that mounts a selected backup to an ESX


server.

Backup Unmount

An operation that unmounts a backup from an ESX


server.

Failover

A dataset failover from a primary node to a disaster


recovery node. Applies only if the dataset is enabled
for disaster recovery.

Host Service
Resource Discovery

A job that discovers virtual servers and the mapping


between virtual and physical storage systems.

Host Service Settings


Configuration

A job that configures a host service.

Host Service Storage


Configuration Import

A job that imports the configuration of the storage


system from the host service to the Express edition of
the DataFabric Manager server.

LUN Destruction

A job that deletes a LUN.

LUN Resizing

A job that changes the LUN size.

Local Backup

A local scheduled backup protection operation based


on Snapshot technology.

Local Backup
Confirmation

A local scheduled backup protection operation based


on Snapshot technology. Applies if a dataset is an
application-generated dataset and if the application is
responsible for creating local backups.

Local Then Remote


Backup

A local backup protection operation on the primary


node of the dataset followed by a transfer of the
backups to remote nodes of the dataset.

Member Dedupe

A deduplication space savings operation initiated on


volumes of a dataset.

Jobs | 49

Migration (One-Step)

A job that begins migrating a dataset or vFiler unit to


a new storage system and automatically performs the
cutover operation.

Migration
Cancellation

A job that cancels a dataset or vFiler unit migration.

Migration Cleanup

A job that deletes the old storage after a dataset or


vFiler unit migration cutover.

Migration
Completion

A migration operation that switches the source of a


dataset or vFiler unit from the old storage system to a
new storage system.

Migration
Relinquishment

A job that transfers the migration capability of a


dataset.

Migration Repair

A CLI-based repair operation.

Migration Rollback

A job that reverses the migration and restores the


original source storage systems as the active
accessible systems.

Migration Start

An operation that begins migrating a dataset or vFiler


unit to a new storage system.

Migration Update

A job that updates the SnapMirror relationships that


were created as part of the migration start operation.

Mirror

A scheduled protection mirror operation based on


SnapMirror technology.

On-Demand
Protection

A backup or mirror operation that is initiated by


clicking the Backup Now button in the Datasets tab.
The types of tasks performed are determined by the
policy configured for the dataset.

Provisioning

A job that provisions containers into a dataset based


on the associated policy and dataset attributes.

Relationship Creation

A job that creates a protection relationship based on


protection technology.

Relationship
Destruction

A protection relationship delete operation based on


protection technology.

Remote Backup

A scheduled backup to secondary storage based on


SnapVault or qtree SnapMirror technology.

Restore

A protection data restore job that is initiated from the


Datasets tab or the Backups tab.

50 | OnCommand Console Help

Server Configuration

A job that begins the initial setup and configuration of


the Express edition of the DataFabric Manager server.

Snapshot Copies
Deletion

A job that deletes Snapshot copies of volumes of a


dataset containing physical storage objects.

Snapshot Copy
Deletion

A job that deletes a Snapshot copy from any volume,


not just members of a dataset.

Storage Deletion

A job that deletes a volume, qtree, or LUN from the


storage system.
Attention: This operation destroys the data in the

deleted volume, qtree, or LUN, and cannot be


reversed.
Storage Resizing

A job that changes the storage size or quota limit. If


the selected container is a volume, this job type
changes the size, Snapshot reserve, and maximum size
of the volume. If the selected container is a qtree, this
job type changes the quota limit of the qtree.

Volume Dedupe

A deduplication space savings operation initiated on


any volume, not just members of a dataset.

Volume Migration

A job of a volume migration in secondary or tertiary


storage.

Volume Resizing

A job that changes the storage size of any volume, not


just members of a dataset.

Volume Undedupe

A deduplicated volume that is converted to a normal


volume.

Object

The name of the object on which the job was started. The default jobs list includes
this column.

Object Type

The type of object on which the job was started. The default jobs list includes this
column. Examples of object types are Aggregate, Dataset, and vFiler unit.

Start

The date and time the job was started. The default jobs list includes this column.

Bytes
Transferred

The amount of data (in megabytes, gigabytes, or kilobytes) that was transferred
during the job. This column is not displayed in the jobs list by default.
Note: This number is an approximation and does not reflect an exact count; it is
always less than the actual number of bytes transferred. For jobs that take a
short time to complete, no data transfer size is reported.

Job Status

The running status of the job. The default jobs list includes this column. The
progress options are as follows:

Jobs | 51

Failed

All tasks in the job failed.

Partially Failed

One or more of the tasks in the job failed and one or more of
the tasks completed successfully.

Succeeded

All tasks completed successfully.

Running with
Failures

The job is currently running but one or more tasks in the job
failed.

Running

The job is currently running.

Queued

The job is not running yet. However, it is scheduled to run


after other provisioning or protection jobs on the same
dataset are completed.

Canceled

The job stopped because the Cancel button was clicked to


stop the job before it was completed.

Canceling

The Cancel button was clicked and the job is in the process
of being canceled.

End

The date and time the job ended. The default jobs list includes this column.

Policy

The name of the data protection policy associated with the job. This column is not
displayed in the jobs list by default.

Source Node

The name of the storage resource that contains the data being protected. This
column is not displayed in the jobs list by default.

Destination
Node

The name of the storage resource to which the data is transferred during the job.
This column is not displayed in the jobs list by default.

Submitted By The policy that automatically started the job or the user name of the person who
started the job. This column is not displayed in the jobs list by default.
Description

A description of the job taken from the policy configuration or the job description
entered when the job was manually started. This column is not displayed in the
jobs list by default.

Completed job steps


Displays detailed information about each task in the selected job. You can select a step to see its
details.
Time Stamp The date and time the step was completed.
Step

A description of the step: for example, Start, In progress, or End.

Result

The result of the step. Result options are as follows:


Error

The step failed.

52 | OnCommand Console Help

Warning

The step succeeded but with a possible problem.

Normal

The step succeeded.

Job details
Displays details for the currently highlighted job appear in the lower right window.
Dataset

The name of the dataset to which the job belongs.

Job Description

A description of the specified job.

Event
Description

A description of any events associated with the job.

Policy

The name of the data protection policy associated with the job. This column is
not displayed in the jobs list by default.

Job Type

The type of job, which is determined by the policy assigned to the dataset or by
the direct request initiated by a user.

Source

The name of the storage resource that contains the data being protected. This
column is not displayed in the jobs list by default.

Destination

The name of the storage resource to which the data is transferred during the job.
This column is not displayed in the jobs list by default.

Submitted By

The policy that automatically started the job or the user name of the person who
started the job. This column is not displayed in the jobs list by default.

Bytes
Transferred

The amount of data (in megabytes, gigabytes, or kilobytes) that was transferred
during the job. This column is not displayed in the jobs list by default.
Note: This number is an approximation and does not reflect an exact count; it
is always less than the actual number of bytes transferred. For jobs that take a
short time to complete, no data transfer size is reported.

Related references

Window layout customization on page 16

53

Servers
Understanding virtual inventory
How virtual objects are discovered
After you successfully install and register a host service with DataFabric Manager server, DataFabric
Manager server automatically begins a job to discover the virtual server inventory.
The storage credentials that you set when configuring the host service (and vCenter properties for
VMware) are pushed to the storage inventory, and DataFabric Manager server begins to map each
server to storage.
If this automatic discovery job is not successful, you can fix the errors noted in the event log and then
manually start a discovery job from the Host Services tab.
When you make changes to the virtual infrastructure, the results do not immediately appear in the
OnCommand console. To see the updated inventory, manually refresh the host service information
from the Host Services tab.
Related concepts

What a host service is on page 393


Configuring a new host service on page 395
Rediscovering virtual object inventory on page 408

Monitoring virtual inventory


Monitoring VMware inventory
Viewing the VMware virtual center inventory
You can view your inventory of VMware virtual centers, add a virtual center to a group, and view
related storage objects from the VMware Virtual Centers view.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.

54 | OnCommand Console Help


About this task

When you make changes to the virtual infrastructure, the results do not immediately appear in the
OnCommand console. To see the updated inventory, manually refresh the host service information
from the Host Services tab.
Steps

1. Click the View menu, then click the Server option.


2. In the Server tab, click the VMware folder, then click Virtual Centers.
Related tasks

Rediscovering virtual object inventory on page 408


Refreshing host service information on page 408
VMware Virtual Centers view on page 71
Related references

Administrator roles and capabilities on page 506


Viewing the VMware datacenter inventory
You can view your inventory of VMware datacenters, start a backup job, add a datacenter to a
dataset, add a datacenter to a group, and view related storage objects from the VMware Datacenters
view.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

When you make changes to the virtual infrastructure, the results do not immediately appear in the
OnCommand console. To see the updated inventory, manually refresh the host service information
from the Host Services tab.
Steps

1. Click the View menu, then click the Server option.


2. In the Server tab, click the VMware folder, then click Datacenters.
Related tasks

Rediscovering virtual object inventory on page 408


Refreshing host service information on page 408

Servers | 55

VMware Datacenters view on page 72


Related references

Administrator roles and capabilities on page 506


Viewing the VMware ESX server inventory
You can view your inventory of VMware ESX servers, add an ESX server to a group, and view
related storage objects from the VMware ESX Servers view.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

When you make changes to the virtual infrastructure, the results do not immediately appear in the
OnCommand console. To see the updated inventory, manually refresh the host service information
from the Host Services tab.
Note: If you move an ESX server from one vCenter to another, DataFabric Manager server still
shows the ESX server and its objects in the inventory for the original vCenter host service. In this
case, you must explicitly remove the ESX server from the original vCenter.
Steps

1. Click the View menu, then click the Server option.


2. In the Server tab, click the VMware folder, then click ESX Servers.
Related tasks

Rediscovering virtual object inventory on page 408


Refreshing host service information on page 408
VMware ESX Servers view on page 73
Related references

Administrator roles and capabilities on page 506

56 | OnCommand Console Help

Viewing the VMware virtual machine inventory


You can view and monitor your inventory of VMware virtual machines, start a backup job, add a
virtual machine to a dataset, start a restore job, mount or unmount a virtual machine, add a virtual
machine to a group, and view related storage objects from the VMware VMs view.
Before you begin

For the OnCommand console to show guest virtual machine properties such as DNS name and IP
address, VMware Tools must be installed and running on the guest virtual machine.
You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

When you make changes to the virtual infrastructure, the results do not immediately appear in the
OnCommand console. To see the updated inventory, manually refresh the host service information
from the Host Services tab.
Steps

1. Click the View menu, then click the Server option.


2. In the Server tab, click the VMware folder, then click VMware VMs.
Related tasks

Rediscovering virtual object inventory on page 408


Refreshing host service information on page 408
VMware VMs view on page 74
Related references

Administrator roles and capabilities on page 506


Viewing the VMware datastore inventory
You can view and monitor your inventory of VMware datastores, start a backup job, add a datastore
to a dataset, start a restore job, mount or unmount a datastore, add a datastore to a group, and view
related storage objects from the VMware Datastores view.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.

Servers | 57
About this task

When you make changes to the virtual infrastructure, the results do not immediately appear in the
OnCommand console. To see the updated inventory, manually refresh the host service information
from the Host Services tab.
Steps

1. Click the View menu, then click the Server option.


2. In the Server tab, click the VMware folder, then click Datastores.
Related tasks

Rediscovering virtual object inventory on page 408


Refreshing host service information on page 408
VMware Datastores view on page 77
Related references

Administrator roles and capabilities on page 506

Monitoring Hyper-V inventory


Viewing the Hyper-V server inventory
You can view your inventory of Hyper-V servers, add a server to a group, and view related storage
objects from the Hyper-V Servers view.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

When you make changes to the virtual infrastructure, the results do not immediately appear in the
OnCommand console. To see the updated inventory, manually refresh the host service information
from the Host Services tab.
Steps

1. Click the View menu, then click the Server option.


2. In the Server tab, click the Hyper-V folder, then click Hyper-V Servers.
Related tasks

Rediscovering virtual object inventory on page 408

58 | OnCommand Console Help

Refreshing host service information on page 408


Hyper-V Servers view on page 80
Related references

Administrator roles and capabilities on page 506


Viewing the Hyper-V virtual machine inventory
You can view and monitor your inventory of Hyper-V virtual machines, start a backup job, add a
virtual machine to a dataset, start a restore job, add a virtual machine to a group, and view related
storage objects from the Hyper-V VMs view.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

When you make changes to the virtual infrastructure, the results do not immediately appear in the
OnCommand console. To see the updated inventory, manually refresh the host service information
from the Host Services tab.
Note: The Related Objects pane does not show LUNs that were created on the virtual machine
using the Microsoft iSCSI software initiator.
Steps

1. Click the View menu, then click the Server option.


2. In the Server tab, click the Hyper-V folder, then click Hyper-V VMs.
Related tasks

Rediscovering virtual object inventory on page 408


Refreshing host service information on page 408
Hyper-V VMs view on page 81
Related references

Administrator roles and capabilities on page 506

Managing virtual inventory

Servers | 59

Adding virtual objects to a group


You can add a virtual object to an existing group. You can also create a new group and add the
virtual object to it.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu, then click the Server option.


2. In the Server tab, select the type of virtual object in the list of views.
3. Select the virtual object you want to add to the group.
4. Click Add to Group.
5. In the Add to Group dialog box, perform the appropriate action:
If you want to..

Then..

Add the virtual object to an existing group.

Select the appropriate group.

Add the virtual object to a new group.

Type the name of the new group in the New Group field.

6. Click Ok.
Result

The virtual object is added to the group.


Related references

Administrator roles and capabilities on page 506

Adding a virtual machine to inventory


If you want to recover data from a virtual machine that has been deleted from the OnCommand
console, you must manually add the virtual machine to the inventory.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the OnCommand Server tab.

60 | OnCommand Console Help


2. Select Datastores and select the listing for the virtual machine that you want to add.
3. Click Add to Group.

Preparing a virtual object managed by the OnCommand console for


deletion from inventory
Before you use a third-party virtual object management tool, such as vSphere, to delete from
inventory a virtual object, such as a virtual machine or a datastore, that is currently a managed by the
OnCommand console, you must first remove that object from any dataset to which it belongs.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

Failure to remove a virtual object from a dataset before deleting it from inventory causes backup
failures in the deleted object's former dataset.
Steps

1. After you decide to delete a specific virtual object from inventory, but before you actually delete
it, click the OnCommand console Server tab.
2. Find and select the listing for the virtual object that you want to delete, and note whether any
datasets are listed in that object's Dataset(s) column.
Datasets that are listed in the selected object's Dataset(s) column indicate that the virtual object is
a member of those datasets.
3. If no datasets are listed for the selected virtual object, use your usual tool to delete that object
from inventory.
4. If the selected object belongs to one or more datasets, click Datasets in the Related Objects pane
to display the dataset hyperlinks, then complete the following actions for each hyperlink:
a. Click the dataset hyperlink.
The OnCommand console displays the Datasets tab with the dataset in question selected.
b. Click Edit to display the Edit Dataset dialog box for the selected dataset.
c. Click Data to display the Data area
d. Remove the virtual object that you want to delete from its dataset.
e. Click OK to finalize the removal.
5. After you have removed the virtual object from all datasets, use your favorite tool to delete the
virtual object from inventory.

Servers | 61
Related references

Administrator roles and capabilities on page 506

Performing an on-demand backup of virtual objects


You can protect your virtual machines or datastores by adding them to an existing or new dataset and
performing an on-demand backup.
Before you begin

You must have reviewed the Guidelines for performing an on-demand backup on page 277
You must have reviewed the Requirements and restrictions when performing an on-demand
backup on page 279
You must have added the virtual objects to an existing dataset or have created a dataset and added
the virtual objects that you want to back up.
You must be authorized to perform all the steps of this task; your RBAC administrator can
confirm your authorization in advance.
You must have the following information available:

Dataset name
Retention duration
Backup settings
Backup script location
Backup description

About this task

If you perform a backup of a dataset containing Hyper-V virtual machines and you are currently
restoring those virtual machines, the backup might fail.
Steps

1. Click the View menu, then click the Server option.


2. In the Server tab, choose the virtual machine or datastore that you want to back up.
If you want to back up...

Then...

A virtual machine or datastore that does not


belong to an existing dataset

You must first add it to an existing dataset.

A virtual machine or datastore, but no datasets


currently exist

You must first create a dataset and then add to it


the required virtual machines or datastores.

3. Click Backup and select the Back Up Now option.


4. In the Back Up Now dialog box, select the dataset that you want to back up.

62 | OnCommand Console Help


If the virtual machine or datastore belongs to multiple datasets, you must select one dataset to
back up.
5. Specify the local protection settings, backup script path, and backup description for the ondemand backup.
If you have already established local policies for the dataset, that information automatically
appears for the local protection settings for the on-demand backup. If you change the local
protection settings, the new settings override only the existing application policies for this ondemand backup.
6. If you want a remote backup to begin after the local backup has finished, select the Start remote
backup after local backup check box.
7. Click Back Up Now.
After you finish

You can monitor the status of your backup from the Jobs tab.
Related references

Jobs tab on page 46


Administrator roles and capabilities on page 506
Guidelines for performing an on-demand backup
Before performing an on-demand backup of a dataset, you must decide how you want to assign
resources and assign protection settings.
General properties information
When performing an on-demand backup, you need to provide information about what objects you
want to back up, to assign protection and retention settings, and to specify script information that
runs before or after the backup operation.
Dataset
name

You must select the dataset that you want to back up.

Local
protection
settings

You can define the retention duration and the backup settings for your on-demand
backup, as needed.
Retention

You can choose to keep a backup until you manually delete it, or
you can assign a retention duration. By specifying a length of time
to keep the on-demand local backup, you can override the retention
duration in the local policy you assigned to the dataset for this
backup. The retention duration of a local backup defaults to a
retention type for the remote backup.

Servers | 63

A combination of both the remote backup retention type and storage


service is used to determine the remote backup retention duration.
For example, if you specify a local backup retention duration of two
days, the retention type of the remote backup is Daily. The dataset
storage service then verifies how long daily remote backups are kept
and applies this to the backup. This is the retention duration of the
remote backup.
The following table lists the local backup retention durations and the
equivalent remote backup retention type:

Backup
settings

Backup
script path

Local retention duration

Remote retention
type

Less than 24 hours

Hourly

1 day up to, but not including, 7 days

Daily

1 week up to, but not including, 31 days

Weekly

More than 31 days

Monthly

You can choose your on-demand backup settings based on the type
of virtual objects you want to back up.
Allow saved state
backup (Hyper-V
only)

You can choose to skip the backup if it


causes one or more of the virtual machines
to go offline. If you do not choose this
option, and your Hyper-V virtual
machines are offline, backup operations
fail.

Create VMware
snapshot
(VMware only)

You can choose to create a VMware


formatted in addition to the storage system
Snapshot copies created during local
backup operations.

Include
independent disks
(VMware only)

You can include independent disks.


VMDKs belong to VMware virtual
machines in the current dataset, but reside
on datastores that are not part of the
current dataset.

You can specify a script that is invoked before and after the local backup. The script
is invoked on the host service and the path is local to the host service. If you use a
PowerShell script, you should use the drive letter convention. For other types of

64 | OnCommand Console Help


scripts, you can use either the drive letter convention or the Universal Naming
Convention.
Backup
description

You can provide a description for the on-demand backup so you can easily find it
when you need it.

Clustered virtual machine considerations (Hyper-V only)


Dataset backups of clustered virtual machines take longer to complete when the virtual machines run
on different nodes of the cluster. When virtual machines run on different nodes, separate backup
operations are required for each node in the cluster. If all virtual machines run on the same node,
only one backup operation is required, resulting in a faster backup.
Requirements and restrictions when performing an on-demand backup
You must be aware of the requirements and restrictions when performing an on-demand backup.
Some requirements and restrictions apply to all types of objects and some are specific to Hyper-V or
VMware virtual objects.
Requirements

Virtual machines or datastores must first belong to a dataset before backing up.
You can add virtual objects to an existing dataset or create a new dataset and
add virtual objects to it.

Hyper-V specific Each virtual machine contained in the dataset that you want to back up must
requirements
contain at least 300 MB of free disk space. Each Windows volume in the
virtual machine (guest OS) must have at least 300 MB free disk space. This
includes the Windows volumes corresponding to VHDs, iSCSI LUNs, and
pass-through disks attached to the virtual machine.
Hyper-V virtual machine configuration files, snapshot copy files, and VHDs
must reside on Data ONTAP LUNs, otherwise backup operations fail.
VMware specific Backup operations of datasets containing empty VMware datacenters or
datastores will fail. All datacenters must contain datastores or virtual machines
requirements
to successfully perform a backup.
Virtual disks must be contained within folders in the datastore. If virtual disks
exist outside of folders on the datastore, and that data is backed up, restoring
the backup could fail.
NFS backups might take more time than VMFS backups. This is because it
takes more time for VMware to commit snapshots in a NFS environment.
Hyper-V specific Partial backups are not supported. If the Hyper-V VSS writer fails to back up
one of the virtual machines in the backup and the failure occurs at the Hyper-V
restrictions
parent host, the backup fails for all of the virtual machines in the backup.

Restoring backups from the Server tab

Servers | 65

Restoring a datastore using the Restore wizard


You can use OnCommand console to recover a datastore from a local or remote backup. By doing so,
you overwrite the existing content with the backup you select.
About this task

Once you start the restoration, you cannot stop the process.
Steps

1. Click the View menu, then click the Server option.


2. In the Server tab, select All Datastores to sort the backup table by datastores.
3. Select a virtual machine that has the backup you want.
4. Click Restore.
The Restore wizard opens.
5. Select the datastore that contains the backup from the list of backed-up entities.
6. Select the following restore options:
Option

Description

Start VM after restore

Restores the contents of your virtual machine from a Snapshot copy and restarts
the virtual machine after the operation completes.

Pre/Post Restore Script Runs a script that is stored on the host service server before or after the restore
operation.

The Restore Wizard displays the location of the virtual hard disk (.vhd) file.
7. From this wizard, click Restore to begin the restoration.
Restoring a VMware virtual machine using the Restore wizard
You can use OnCommand console to recover a VMware virtual machine from a local or remote
backup. By doing so, you overwrite the existing content with the backup you select.
About this task

The process for restoring a VMware virtual machine differs from restoring a Hyper-V virtual
machine in that you can restore an entire virtual machine or its disk files. Once you start the
restoration, you cannot stop the process, and you cannot restore from a backup of a virtual machine
after you delete the dataset the virtual machine belonged to.

66 | OnCommand Console Help


Steps

1. Click the View menu, then click the Server option.


2. In the Server tab, click All VMware VMs to sort the backup table by VMware virtual machines.
3. Select a virtual machine from the list of backed-up entities.
4. Click Restore.
The Restore wizard opens and lists the dataset which includes the backup of the virtual machine.
5. Select one of the following recovery options:
Option

Description

The entire virtual


machine

Restores the contents of your virtual machine from a Snapshot copy to its original
location. The Restart VM checkbox is enabled if you select this option and the
virtual machine is registered.

Particular virtual
disks

Restores the contents of the virtual disks on a virtual machine to a different


location. This option is enabled if you uncheck the entire virtual machine option.

6. In the ESX host name field, select the name of the ESX host. The ESX host is used to mount the
virtual machine components.
This option is available if you want to restore virtual disk files or the virtual machine is on a
VMFS datastore.
7. In the Pre/Post Restore Script field, type the name of the script that you want to run before or
after the restore operation.
8. Click Next.
9. From this wizard, review the summary of restore operations and click Restore to begin the
restoration.
Related tasks

Adding a virtual machine to inventory on page 59


Where to restore a backup on page 297
Restoring a Hyper-V virtual machine using the Restore wizard
You can use the OnCommand console to recover a Hyper-V virtual machine from a local or remote
backup. By doing so, you overwrite the existing content with the backup you select.
About this task

If you start a restore operation of a Hyper-V virtual machine, and another backup or restoration of the
same virtual machine is in process, it fails. Once you start the restoration, you cannot stop the
process.

Servers | 67
Steps

1. Click the View menu, then click the Server option.


2. In the Server tab, select All Hyper-V VMs to sort the backup table by Hyper-V virtual machines.
3. Select a virtual machine that has the backup you want.
4. Click Restore.
The Restore wizard opens.
5. Select the dataset that contains the backup from the list of backed-up entities.
6. Select the following restore options:
Option

Description

Start VM after restore

Restores the contents of your virtual machine from a Snapshot copy and restarts
the virtual machine after the operation completes.

Pre/Post Restore Script Runs a script that is stored on the host service server before or after the restore
operation.

The Restore Wizard displays the location of the virtual hard disk (.vhd) file.
7. From this wizard, click Restore to begin the restoration.

Mounting and unmounting backups in a VMware environment


Mounting backups in a VMware environment from the Server tab
You can mount existing backups onto an ESX server for backup verification prior to completing a
restore operation or to restore a virtual machine to an alternate location. All the datastores and the
virtual machines within the backup are mounted to the ESX server that you specify. Both the Mount
and Unmount buttons are disabled for Hyper-V backups.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

In the OnCommand console Backups tab, deleting a mirror source backup copy prevents you from
mounting its partner mirror destination backup copy. For a Mirror-generated destination backup copy
to be mountable, its associated mirror source backup copy must still exist on the source node.
Steps

1. Click the View menu, then click the Server option.

68 | OnCommand Console Help


2. In the Server tab, click the VMware option, then click VMware VMs or Datastores.
3. Select a virtual machine or datastore and click Mount.
You cannot mount Hyper-V backups using this button.
4. In the Mount Backup dialog box, select an unmounted backup that you want to mount.
You can only mount one backup each time and you cannot mount a mounted backup.
5. Select from the drop-down list the name of the ESX server to which you want to mount the
backup.
6. Click Mount.
A dialog box appears with a link to the mount job and when you click the link, the Jobs tab
appears.
After you finish

You can monitor the status of your mount and unmount jobs in the Jobs tab.
Related references

Administrator roles and capabilities on page 506


Guidelines for mounting or unmounting backups in a VMware environment on page 271
Unmounting backups in a VMware environment from the Server tab
After you are done using a mounted backup for verification or to restore a virtual machine to an
alternate location, you can unmount the mounted backup from the ESX server that it was mounted to.
When you unmount a backup, all the datastores in that backup are unmounted and can no longer be
seen from the ESX server that you specify. Both the Mount and Unmount buttons are disabled for
Hyper-V backups.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

If there are virtual objects in use from the previously mounted datastores of a backup, the unmount
operation fails. You must manually clean up the backup prior to mounting the backup again because
its state reverts to not mounted.
If all the datastores of the backup are in use, the unmount operation fails but this backup's state
changes to mounted. You can unmount the backup after determining the datastores are not in use.

Servers | 69
Steps

1. Click the View menu, then click the Server option.


2. In the Server tab, click the VMware option, then click VMware VMs or Datastores.
3. Select a virtual machine or datastore and click Unmount.
4. In the Unmount Backup dialog box, select a mounted backup to unmount.
5. Click Unmount.
6. At the confirmation prompt, click Yes.
A dialog box opens with a link to the unmount job displays and when you click the link, the Jobs
tab appears.
After you finish

If the ESX server becomes inactive or reboots during an unmount operation, the job is terminated and
the mount state remains mounted and the backup stays mounted on the ESX server.
You can monitor the status of your mount and unmount jobs in the Jobs tab.
Related references

Administrator roles and capabilities on page 506


Guidelines for mounting or unmounting backups in a VMware environment on page 271
Mounting backups in a VMware environment from the Backups tab
You can mount existing backups onto an ESX server for backup verification prior to completing a
restore operation or to restore a virtual machine to an alternate location. All the datastores and the
virtual machines within the backup are mounted to the ESX server that you specify. Both the Mount
and Unmount buttons are disabled for Hyper-V backups.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

In the OnCommand console Backups tab, deleting a mirror source backup copy prevents you from
mounting its partner mirror destination backup copy. For a Mirror-generated destination backup copy
to be mountable, its associated mirror source backup copy must still exist on the source node.
Steps

1. Click the View menu, then click the Backups option.

70 | OnCommand Console Help


2. In the Backups tab, select an unmounted backup that you want to mount.
3. Click Mount.
You cannot mount Hyper-V backups using this button.
4. In the Mount Backup dialog box, select from the drop-down list the name of the ESX server to
which you want to mount the backup.
You can only mount one backup each time and you cannot mount a mounted backup.
5. Click Mount.
A dialog box appears with a link to the mount job and when you click the link, the Jobs tab
appears.
After you finish

You can monitor the status of your mount and unmount jobs in the Jobs tab.
Related references

Administrator roles and capabilities on page 506


Guidelines for mounting or unmounting backups in a VMware environment on page 271
Unmounting backups in a VMware environment from the Backups tab
After you are done using a mounted backup for verification or to restore a virtual machine to an
alternate location, you can unmount the mounted backup from the ESX server that it was mounted to.
When you unmount a backup, all the datastores in that backup are unmounted and can no longer be
seen from the ESX server that you specify. Both the Mount and Unmount buttons are disabled for
Hyper-V backups.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

If there are virtual objects in use from the previously mounted datastores of a backup, the unmount
operation fails. You must manually clean up the backup prior to mounting the backup again because
its state reverts to not mounted.
If all the datastores of the backup are in use, the unmount operation fails but this backup's state
changes to mounted. You can unmount the backup after determining the datastores are not in use.
Steps

1. Click the View menu, then click the Backups option.

Servers | 71
2. In the Backups tab, select a mounted backup to unmount.
3. Click Unmount.
4. At the confirmation prompt, click Yes.
A dialog box opens with a link to the unmount job displays and when you click the link, the Jobs
tab appears.
After you finish

If the ESX server becomes inactive or restarts during an unmount operation, the job is terminated and
the mount state remains mounted and the backup stays mounted on the ESX server.
You can monitor the status of your mount and unmount jobs in the Jobs tab.
Related references

Administrator roles and capabilities on page 506


Guidelines for mounting or unmounting backups in a VMware environment on page 271

Page descriptions
VMware
VMware Virtual Centers view
The VMware Virtual Centers view lists the discovered virtual centers. You can access this view by
clicking View > Server > VMware > Virtual Centers.
From the VMware Virtual Centers view, you can add a virtual center to a group, and view objects
that are related to each virtual center.

Breadcrumb trail on page 71


Command buttons on page 72
Virtual centers list on page 72
Related Objects pane on page 72

Breadcrumb trail
The breadcrumb trail is created as you navigate from one list of storage objects to another. As you
navigate, each time you double-click certain items in these lists, another breadcrumb is added to
the trail, providing a string of hyperlinks that captures your navigation history. If you want to
revisit a previously viewed list, you can click the breadcrumb that links to it. You can also click the
icon in the breadcrumb to display a sub-navigation list of objects associated with the hyperlinked
object.

72 | OnCommand Console Help


Command buttons
Add to Group Opens the Add to Group dialog box that enables you to add the selected datacenter
to the destination group.
Refresh

Refreshes the list of virtual centers.

Virtual centers list


Displays information about the virtual centers that have been discovered by DataFabric Manager
server. You can double-click a virtual center to display the objects in that virtual center.
Virtual Center

Name of the virtual center.

Related Objects pane


Displays the storage controllers and vFiler units that are related to the selected virtual center.
Related references

Window layout customization on page 16


VMware Datacenters view
The VMware Datacenters view lists the discovered datacenters. You can access this window by
clicking View > Server > VMware > Datacenters.
From the VMware Datacenters view, you can start on-demand backup jobs, add a datacenter to a
group or dataset, and view objects that are related to each datacenter.

Breadcrumb trail on page 72


Command buttons on page 72
Datacenters list on page 73
Related Objects pane on page 73

Breadcrumb trail
The breadcrumb trail is created as you navigate from one list of storage objects to another. As you
navigate, each time you double-click certain items in these lists, another breadcrumb is added to
the trail, providing a string of hyperlinks that captures your navigation history. If you want to
revisit a previously viewed list, you can click the breadcrumb that links to it. You can also click the
icon in the breadcrumb to display a sub-navigation list of objects associated with the hyperlinked
object.
Command buttons
Back Up

Opens a list of backup commands.

Servers | 73

Using New
Dataset

Opens the Create Dataset dialog box to create a new dataset


and add the selected datacenter to that dataset.

Using Existing
Dataset

Opens the Add to Existing Dataset dialog box to add the


selected datacenter to an existing dataset.

Add to Group Opens the Add to Group dialog box that enables you to add the selected datacenter
to the destination group.
Refresh

Refreshes the list of datacenters.

Datacenters list
Displays information about the datacenters that have been discovered by DataFabric Manager server.
You can double-click a datacenter to display the objects in that datacenter.
Datacenter

Name of the datacenter.

Protected

Indicates whether the datacenter is protected. Valid values are "Yes" and "No."
A datacenter is protected if it is a member of a dataset that has a local policy
assigned to it.

Virtual Center Name of the virtual center with which the datacenter is associated.
Dataset

The names of the datasets of which the datacenter is a member.

Related Objects pane


Displays the storage controllers, vFiler units, and datasets that are related to the selected datacenter.
Related references

Window layout customization on page 16


VMware ESX Servers view
The VMware ESX Servers view lists the discovered VMware ESX servers. You can access this
window by clicking View > Server > VMware > ESX Servers.
From the VMware ESX Servers view, you can add a server to a group, and view objects that are
related to each server.

Breadcrumb trail on page 73


Command buttons on page 74
ESX servers list on page 74
Related Objects pane on page 74

Breadcrumb trail
The breadcrumb trail is created as you navigate from one list of storage objects to another. As you
navigate, each time you double-click certain items in these lists, another breadcrumb is added to

74 | OnCommand Console Help


the trail, providing a string of hyperlinks that captures your navigation history. If you want to
revisit a previously viewed list, you can click the breadcrumb that links to it. You can also click the
icon in the breadcrumb to display a sub-navigation list of objects associated with the hyperlinked
object.
Command buttons
Add to Group Opens the Add to Group dialog box that enables you to add the selected server to
the destination group.
Refresh

Refreshes the list of ESX servers.

ESX servers list


Displays information about the ESX servers that have been discovered by DataFabric Manager
server. You can double-click an ESX server to display the virtual machines that are mapped to that
ESX server.
ESX Server

Name of the VMware ESX server.

Datacenter

Name of the datacenter with which the ESX server is associated.

Virtual Center

Name of the virtual center with which the datacenter is associated.

Related Objects pane


Displays the storage controllers and vFiler units that are related to the selected ESX server.
Related references

Window layout customization on page 16


VMware VMs view
The VMware VMs view lists the discovered VMware virtual machines. You can access this window
by clicking View > Server > VMware > VMware VMs.
From the VMware VMs view, you can start on-demand backup and restore jobs, add virtual
machines to groups and datasets, mount or unmount virtual machines, and view objects that are
related to each virtual machine.

Breadcrumb trail on page 75


Command buttons on page 75
VMware VMs list on page 75
VDisks tab on page 76
Datasets tab on page 76
Related Objects pane on page 77

Servers | 75
Breadcrumb trail
The breadcrumb trail is created as you navigate from one list of storage objects to another. As you
navigate, each time you double-click certain items in these lists, another breadcrumb is added to
the trail, providing a string of hyperlinks that captures your navigation history. If you want to
revisit a previously viewed list, you can click the breadcrumb that links to it. You can also click the
icon in the breadcrumb to display a sub-navigation list of objects associated with the hyperlinked
object.
Command buttons
Back Up

Opens a list of backup commands.


Using New
Dataset

Opens the Create Dataset dialog box to create a new dataset


and add the selected VMware virtual machine to that
dataset.

Using Existing
Dataset

Opens the Add to Existing Dataset dialog box to add the


selected VMware virtual machine to an existing dataset.

Back Up Now

Opens the Back Up Now dialog box to start a backup job.

Restore

Starts the Restore wizard to begin a restore job.

Mount

Enables you to mount a selected backup to an ESX server if you want to verify its
content before restoring it.

Unmount

Enables you to unmount a backup after you mount it on an ESX server and verify
its contents.

Add to Group Opens the Add to Group dialog box that enables you to add the selected virtual
machine to the destination group.
Refresh

Refreshes the list of VMware virtual machines.

VMware VMs list


Displays information about the VMware virtual machines that have been discovered by DataFabric
Manager server.
Virtual
Machine

Name of the VMware virtual machine.

Protected

Indicates whether the data in the virtual machine is protected. Valid values are
"Yes" and "No."
A virtual machine is protected if any of the following conditions are true:

The virtual machine is a member of a dataset that has a local policy assigned
to it.

76 | OnCommand Console Help

The virtual machine is in a datastore that is a member of a dataset that has a


local policy assigned to it.
The virtual machine is in a datacenter that is a member of a dataset that has a
local policy assigned to it.

ESX Server

Name of the ESX server in which the virtual machine runs.

Datacenter

Name of the datacenter with which the VM is associated.

Virtual Center

Name of the virtual center with which the datacenter is associated.

State

The state of the virtual machine.


Powered Off

The virtual machine is down.

Powered On

The virtual machine is up.

Suspended

The operating system for the virtual machine is down.

DNS Name

The DNS name for the virtual machine.

IP Address

The IP address of the virtual machine. One or more IP addresses might be listed
for a virtual machine.

Dataset(s)

The names of the datasets of which the virtual machine is a member.

VDisks tab
Displays detailed information about the VDisks for the selected virtual machine.
VDisk

The name of the VDisk.

Disk Type The disk type of the VDisk. Possible values are "Raw Device Mapping" or "Regular."
Datastore The datastore to which the VDisk is mapped.
Datasets tab
Displays detailed information about the dataset of which the selected virtual machine is a member.
Dataset

The name of the datasets of which the virtual machine is a member.

Storage
Service

The storage service that is assigned to the dataset.

Local Policy

The local policy that is assigned to the dataset. This policy might be the default
policy associated with the dataset or it might be a local policy assigned by an
administrator as part of a dataset modification.

Servers | 77
Related Objects pane
Displays the datastores, ESX servers, storage controllers, vFiler units, volumes, LUNs, datasets, and
backups that are related to the selected VMware virtual machine.
Related references

Window layout customization on page 16


VMware Datastores view
The VMware Datastores view lists the discovered datastores. You can access this window by
clicking View > Server > VMware > Datastores.
From the VMware Datastores view, you can start on-demand backup and restore jobs, add datastores
to groups and datasets, mount or unmount datastores, and view objects that are related to each
datastore.

Breadcrumb trail on page 77


Command buttons on page 77
Datastores list on page 78
Storage Details tab on page 78
Related Objects pane on page 80

Breadcrumb trail
The breadcrumb trail is created as you navigate from one list of storage objects to another. As you
navigate, each time you double-click certain items in these lists, another breadcrumb is added to
the trail, providing a string of hyperlinks that captures your navigation history. If you want to
revisit a previously viewed list, you can click the breadcrumb that links to it. You can also click the
icon in the breadcrumb to display a sub-navigation list of objects associated with the hyperlinked
object.
Command buttons
Back Up

Opens a list of backup commands.


Using New
Dataset

Opens the Create Dataset dialog box to create a new


dataset and add the selected datastore to that dataset.

Using Existing
Dataset

Opens the Add to Existing Dataset dialog box to add the


selected datastore to an existing dataset.

Back Up Now

Opens the Back Up Now dialog box to start a backup job.

Restore

Starts the Restore wizard to begin a restore job.

Mount

Enables you to mount a selected backup to an ESX server if you want to verify its
content before restoring it.

78 | OnCommand Console Help


Unmount

Enables you to unmount a backup after you mount it on an ESX server and verify
its contents.

Add to Group Opens the Add to Group dialog box that enables you to add the selected datastore
to the destination group.
Refresh

Refreshes the list of datastores.

Datastores list
Displays information about the datastores that have been discovered by DataFabric Manager server.
You can double-click a datastore to display the objects in that datastore.
Datastore

The name of the datastore.

Protected

Indicates whether the data in the virtual machine is protected. Valid values
are "Yes" and "No."
A datastore is protected if any of the following conditions are true:

The datastore is a member of a dataset that has a local policy assigned to


it.
The datastore is in a datacenter that is a member of a dataset that has a
local policy assigned to it.

Type

The type of datastore. Valid values are NFS and VMFS.

Datacenter

The name of the datacenter with which the datastore is associated.

Virtual Center

The name of the virtual center with which the datastore is associated.

Capacity (GB)

The configured amount of space in the datastore.

Used Capacity
(GB)

The amount of space in the datastore that is used.

Dataset

The names of the datasets of which the datastore is a member.

Hosted on Data
ONTAP

Indicates whether the datastore is hosted on Data ONTAP. Valid values are
"Yes" and "No."

Storage Details tab


Displays detailed information about the selected datastore.
Details for NFS type datastores:
Overview Export Path

The path used by the datastore to export data.

Servers | 79

Capacity
(GB)

Volume Thin
Provisioning
Enabled

Indicates whether the datastore is configured for thin


provisioning. Valid values are "true" if the feature is configured
and "false" if not.

Dedupe

Indicates whether the datastore is configured for deduplication.


Valid values are "true" if the feature is configured and "false" if
not.

Autosize

Indicates whether the datastore is configured for automatic


storage sizing based on usage. Valid values are "true" if the
feature is configured and "false" if not.

Datastore Usage

The percentage of total capacity of the datastore that is used.

Volume Usage

The percentage of total capacity of the volume that is used.

Space Savings

The amount of space that was saved on the volume because


deduplication is enabled.

Aggregate Usage The percentage of total capacity of the aggregate that is in


use.
Details for VMFS type datastores:
LUN Paths

A list of the LUNs that are mapped to the datastore.

Overview IGroup

Capacity
(GB)

For SAN datastores, the VMware IGroup with which the


datastore is associated. This field is not displayed for NAS
datastores.

LUN Space
Reservation

Indicates whether the LUN is configured for space reservation.


Valid values are "true" if the feature is configured and "false" if
not.

Volume Thin
Provisioning
Enabled

Indicates whether the volume is configured for thin


provisioning. Valid values are "true" if the feature is configured
and "false" if not.

Dedupe

Indicates whether the volume is configured for deduplication.


Valid values are "true" if the feature is configured and "false" if
not.

Autosize

Indicates whether the volume is configured for automatic


storage sizing based on usage. Valid values are "true" if the
feature is configured and "false" if not.

Datastore Usage

The percentage of total capacity of the datastore that is in use.

80 | OnCommand Console Help

LUN Usage

The percentage of total capacity of the LUN that is in use.

Volume Usage

The percentage of total capacity of the volume that is in use.

Space Savings

The amount of space that was saved on the volume because


deduplication is enabled.

Aggregate Usage The percentage of total capacity of the aggregate that is in


use.
Related Objects pane
Displays the virtual machines, ESX servers, storage controllers, vFiler units, volumes, LUNs,
datasets, and backups that are related to the selected datastore.
Related references

Window layout customization on page 16

Hyper-V
Hyper-V Servers view
The Hyper-V Servers view lists the discovered Hyper-V servers. You can access this window by
clicking View > Server > Hyper-V > Hyper-V Servers.
From the Hyper-V Servers view, you can add a server to a group, and view objects that are related to
each server.

Breadcrumb trail on page 80


Command buttons on page 80
Hyper-V Servers list on page 81
Related Objects pane on page 81

Breadcrumb trail
The breadcrumb trail is created as you navigate from one list of storage objects to another. As you
navigate, each time you double-click certain items in these lists, another breadcrumb is added to
the trail, providing a string of hyperlinks that captures your navigation history. If you want to
revisit a previously viewed list, you can click the breadcrumb that links to it. You can also click the
icon in the breadcrumb to display a sub-navigation list of objects associated with the hyperlinked
object.
Command buttons
Add to Group Opens the Add to Group dialog box that enables you to add the selected server to
the destination group.

Servers | 81
Refresh

Refreshes the list of Hyper-V servers.

Hyper-V Servers list


Displays information about the Hyper-V servers that have been discovered by DataFabric Manager
server. You can double-click a server to display the children of that server.
Hyper-V Server

The name of the Hyper-V server.

Domain Name

The domain name that is used by the Hyper-V server.

Related Objects pane


Displays the storage controllers and vFiler units that are related to the selected Hyper-V server.
Related references

Window layout customization on page 16


Hyper-V VMs view
The Hyper-V VMs view lists the discovered Hyper-V virtual machines. You can access this window
by clicking View > Server > Hyper-V > Hyper-V VMs.
From the Hyper-V VMs view, you can start on-demand backup and restore jobs, add virtual
machines to groups and datasets, and view objects that are related to each virtual machine.

Breadcrumb trail on page 81


Command buttons on page 81
Hyper-V VMs list on page 82
VDisks tab on page 82
Datasets tab on page 82
Related Objects pane on page 83

Breadcrumb trail
The breadcrumb trail is created as you navigate from one list of storage objects to another. As you
navigate, each time you double-click certain items in these lists, another breadcrumb is added to
the trail, providing a string of hyperlinks that captures your navigation history. If you want to
revisit a previously viewed list, you can click the breadcrumb that links to it. You can also click the
icon in the breadcrumb to display a sub-navigation list of objects associated with the hyperlinked
object.
Command buttons
Back Up

Opens a list of backup commands.

82 | OnCommand Console Help

Using New
Dataset

Opens the Create Dataset dialog box to create a new dataset


and add the selected Hyper-V virtual machine to that
dataset.

Using Existing
Dataset

Opens the Add to Existing Dataset dialog box to add the


selected virtual machine to an existing dataset.

Back Up Now

Opens the Back Up Now dialog box to start a backup job.

Restore

Starts the Restore wizard to begin a restore job.

Add to
Group

Opens the Add to Group dialog box that enables you to add the selected virtual
machine to the destination group.

Refresh

Refreshes the list of Hyper-V virtual machines.

Hyper-V VMs list


Displays information about the Hyper-V virtual machines that have been discovered by DataFabric
Manager server.
Virtual Machine The name of the Hyper-V virtual machine.
Protected

Indicates whether the data in the virtual machine is protected. Valid values are
"Yes" and "No."
A virtual machine is protected if it is a member of a dataset that has a local
policy assigned to it.

Hypervisor

The name of the server that manages the virtual machine.

State

The state of the virtual machine.

DNS Name

The DNS name for the virtual machine.

Dataset(s)

The names of the datasets of which the virtual machine is a member.

VDisks tab
Displays detailed information about the VDisks for the selected virtual machine.
VDisk

The name of the VDisk for the selected virtual machine.

VHD Type The virtual hard disk type for the selected virtual machine. Possible values are "Boot
Disk," "Cluster Shared Volume," "Passthrough," or "Regular."
Datasets tab
Displays detailed information about the dataset of which the selected virtual machine is a member.
Dataset

The name of the dataset of which the Hyper-V virtual machine is a member.

Servers | 83
Storage
Service

The storage service that is assigned to the dataset.

Local Policy

The local policy that is assigned to the dataset. This policy might be the default
policy associated with the storage service or it might be a local policy assigned by
an administrator as part of a storage service modification.

Related Objects pane


Displays the Hyper-V servers, storage controllers, vFiler units, volumes, LUNs, datasets, and
backups that are related to the selected Hyper-V virtual machine.
Related references

Window layout customization on page 16

84 | OnCommand Console Help

85

Storage
Physical storage
Understanding physical storage
What physical storage objects are
You can monitor and manage physical storage objects such as clusters, storage systems, aggregates,
and disks by using the OnCommand console.
You can view detailed information about the physical storage objects that are discovered and
monitored by clicking the appropriate view option.
Cluster

A group of connected storage systems that share a global namespace that you can
manage as a single virtual server or multiple virtual servers, providing performance,
reliability, and scalability benefits. The Clusters view displays all the clusters that are
monitored by OnCommand console and all of the controllers that are part of the
cluster.

Storage
System

Also known as storage controller, is a hardware device running Data ONTAP that
receives data from and sends data to native disk shelves, third-party storage, or both.
The Storage Controllers view displays all the storage systems that are discovered and
monitored by the OnCommand console.

Aggregate

An aggregate contains a defined amount of RAID-protected physical storage that can


be expanded dynamically at any time.
To support the differing security, backup, performance, and data sharing needs of
your users, you should group the physical data storage resources on your storage
system into one or more aggregates. These aggregates provide storage to the volume
or volumes that they contain. The Aggregates view displays all the aggregates
belonging to the storage systems and clusters that are discovered and monitored by
the OnCommand console.

Disk

The basic unit of physical storage for a Data ONTAP system. Multiple disks are
contained by a disk shelf. A Data ONTAP node, can accommodate multiple disk
shelves; the number and capacity varies according to the node's specifications. Disk
shelves provide the physical storage on which logical objects such as aggregates and
volumes are located. The Disks view displays all the disks that are monitored by the
OnCommand console.

86 | OnCommand Console Help

What cluster-related objects are


OnCommand enables you to include cluster-related objects, such as controllers and virtual servers, in
a group. This enables you to easily monitor cluster-related objects that belong to a particular group.
The cluster-related objects are as follows:
Virtual server

A single file-system namespace. A virtual server has separate network access


and provides the same flexibility and control as a dedicated node. Each virtual
server has its own user domain and security domain. It can span multiple
physical nodes.
A virtual server has a root volume that constitutes the top level of the namespace
hierarchy; additional volumes are mounted to the root volume to extend the
namespace. A virtual server is associated with one or more logical interfaces
(LIFs) through which clients access the data on the storage server. Clients can
access the virtual server from any node in the cluster, but only through the
logical interfaces that are associated with the virtual server.

Namespace

Every virtual server has a namespace associated with it. All the volumes
associated with a virtual server are accessed under the virtual server's
namespace. A namespace provides a context for the interpretation of the
junctions that link together a collection of volumes.

Junction

A junction points from a directory in one volume to the root directory of another
volume. Junctions are transparent to NFS and CIFS clients.

Logical
interface

An IP address with associated characteristics, such as a home port, a list of ports


to fail over to, a firewall policy, a routing group, and so on. Each logical
interface is associated with a maximum of one virtual server to provide client
access to it.

Cluster

A group of connected storage systems that share a global namespace and that
you can manage as a single virtual server or multiple virtual servers, providing
performance, reliability, and scalability benefits.

Storage
controller

The component of a storage system that runs the Data ONTAP operating system
and controls its disk subsystem. Storage controllers are also sometimes called
controllers, storage appliances, appliances, storage engines, heads, CPU
modules, or controller modules.

Ports

A port represents a physical Ethernet connection. In a Data ONTAP cluster,


ports are classified into the following three types:

Data ports
Provide data access to NFS and CIFS clients.
Cluster ports
Provide communication paths for cluster nodes.
Management ports

Storage | 87

Provide data access to Data ONTAP management utility.


Data LIF

A logical network interface mainly used for data transfers and operations. A data
LIF is associated with a node or virtual server in a Data ONTAP cluster.

Node
management
LIF

A logical network interface mainly used for node management and maintenance
operations. A node management LIF is associated with a node and does not fail
over to a different node.

Cluster
management
LIF

A logical network interface used for cluster management operations. A cluster


management LIF is associated with a cluster and can fail over to a different node.

Interface group A single virtual network interface that is created by grouping together multiple
physical interfaces.
What deleted objects are
Deleted objects are the storage objects you have deleted from the OnCommand console. When you
delete a storage object, it is not removed from the OnCommand console database, it is only deleted
from the OnCommand console display and is no longer be monitored by the OnCommand console.
If you delete an object from the database, DataFabric Manager server also deletes all the child objects
it contains. For example, if you delete a storage system, all volumes and qtrees in the storage system
are deleted. Similarly, if a volume is deleted, all the qtrees in the volume are deleted. However, if you
delete a SnapMirror object, only the SnapMirror destination object (volume or qtree) is deleted from
the database.
What happens when storage objects are deleted
With the OnCommand console, you can stop monitoring a storage object (aggregate, volume, or
qtree) by deleting it from the Global group. When you delete an object, the DataFabric Manager
server stops collecting and reporting data about it. Data collection and reporting is resumed only
when the object is added back to the OnCommand console database.
Note: When you delete a storage object from any group other than Global, the object is deleted

only from that group; DataFabric Manager server continues to collect and report data about it. You
must delete the object from the Global group if you want the DataFabric Manager server to stop
monitoring it.

Configuring physical storage

88 | OnCommand Console Help

Adding clusters
You can add a new cluster and monitor it by using the Storage tab.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

During this task, the OnCommand console launches the Operations Manager console. Depending on
your browser configuration, you can return to the OnCommand console by using the Alt-Tab key
combination or clicking the OnCommand console browser tab. After the completion of this task, you
can leave the Operations Manager console open, or you can close it to conserve bandwidth.
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, select the Clusters view.
3. Click Add.
The Clusters, All page is displayed in the Operations Manager console.
4. In the (New Storage System) text box, type the fully qualified domain name of the cluster you
want to add, then click Add.
The cluster is added and displayed in the Global list of clusters.
Note: Identifying the cluster and determining its status might take a few minutes. The

DataFabric Manager server displays an Unknown status until it determines the identity and
status of the cluster.
Related references

Administrator roles and capabilities on page 506


Adding storage controllers
You can add a new storage controller and monitor it by using the Storage tab.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

During this task, the OnCommand console launches the Operations Manager console. Depending on
your browser configuration, you can return to the OnCommand console by using the Alt-Tab key

Storage | 89
combination or clicking the OnCommand console browser tab. After the completion of this task, you
can leave the Operations Manager console open, or you can close it to conserve bandwidth.
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, select the Storage Controllers view.
3. Click Add.
The Storage Systems, All page is displayed in the Operations Manager console.
4. In the (New Storage System) text box, type the fully qualified domain name of the storage
controller you want to add, then click Add.
The storage controller is added and displayed in the Global list of storage systems.
Note: Identifying the controller and determining its status might take a few minutes. The
DataFabric Manager server displays an Unknown status until it determines the identity and
status of the controller.
Related references

Administrator roles and capabilities on page 506

Managing physical storage


Editing storage controller settings
You can edit the following controller settings of the storage system from the Edit Storage Controller
Settings page: the primary IP address, remote platform management IP address, console terminal
server address, login details, preferred SNMP version, owner details, resource tag, maximum active
data transfers, and storage controller thresholds.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

During this task, the OnCommand console launches the Operations Manager console. Depending on
your browser configuration, you can return to the OnCommand console by using the Alt-Tab key
combination or clicking the OnCommand console browser tab. After the completion of this task, you
can leave the Operations Manager console open, or you can close it to conserve bandwidth.

90 | OnCommand Console Help


Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, select the Storage Controllers view.
3. Select the storage controller you want to modify.
4. Click Edit.
The Edit Storage Controller Settings page is displayed in the Operations Manager console.
5. Modify the properties of the storage controller.
6. Click Update.
Changes to the settings are updated in the DataFabric Manager server.
7. Press Alt-Tab or click the OnCommand console browser tab to return to the OnCommand
console.
Related references

Administrator roles and capabilities on page 506


Editing cluster settings
You can edit the following cluster settings from the Edit Cluster Settings page: the primary IP
address, owner e-mail address, owner name, resource tag, monitoring options, management options,
and threshold values.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

During this task, the OnCommand console launches the Operations Manager console. Depending on
your browser configuration, you can return to the OnCommand console by using the Alt-Tab key
combination or clicking the OnCommand console browser tab. After the completion of this task, you
can leave the Operations Manager console open, or you can close it to conserve bandwidth.
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, select the Clusters view.
3. Select the cluster you want to modify.
4. Click Edit.

Storage | 91
The Edit Cluster Settings page is displayed in the Operations Manager console.
5. Edit the cluster settings.
6. Click Update.
Changes to the settings are updated in the DataFabric Manager server.
7. Press Alt-Tab or click the OnCommand console browser tab to return to the OnCommand
console.
Related references

Administrator roles and capabilities on page 506


Editing aggregate settings
You can edit the following settings for an aggregate from the Edit Aggregate Settings page: owner email address, owner name, resource tag, threshold values, and alerts. When you edit the threshold
settings of a specific aggregate, the edited settings override the global settings.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

During this task, the OnCommand console launches the Operations Manager console. Depending on
your browser configuration, you can return to the OnCommand console by using the Alt-Tab key
combination or clicking the OnCommand console browser tab. After the completion of this task, you
can leave the Operations Manager console open, or you can close it to conserve bandwidth.
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, select the Aggregates view.
3. Select the aggregate you want to modify.
4. Click Edit.
The Edit Aggregate Settings page is displayed in the Operations Manager console.
5. Modify the storage capacity threshold settings for the selected aggregate.
6. Click Update.
Changes to the settings are updated in the DataFabric Manager server.
7. Press Alt-Tab or click the OnCommand console browser tab to return to the OnCommand
console.

92 | OnCommand Console Help


Related references

Administrator roles and capabilities on page 506


Adding storage systems to a group
You can add storage systems to an existing group. You can also create a new group and add the
storage systems to it. This enables you to easily monitor all the storage systems that belong to a
group.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, select the Storage Controllers view.
3. Select the storage systems you want to add to the group, and click Add to Group.
4. In the Add to Group dialog box, perform the appropriate action:
If you want to..

Then..

Add the storage systems to an existing group

Select the appropriate group.

Add the storage systems to a new group

Type the name of the new group in the New Group field.

5. Click OK.
Result

The storage systems are added to the group.


Related references

Administrator roles and capabilities on page 506


Adding clusters to a group
You can add clusters to an existing group. You can also create a new group and add the clusters to it.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.

Storage | 93
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, select the Clusters view.
3. Select the clusters you want to add to the group, and click Add to Group.
4. In the Add to Group dialog box, perform the appropriate action:
If you want to..

Then..

Add the clusters to an existing group

Select the appropriate group.

Add the clusters to a new group

Type the name of the new group in the New Group field.

5. Click OK.
Result

The clusters are added to the group.


Related references

Administrator roles and capabilities on page 506


Adding aggregates to a group
You can add aggregates to an existing group. You can also create a new group and add aggregates to
it.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, select the Aggregates view.
3. Select the aggregates you want to add to the group, and click Add to Group.
4. In the Add to Group dialog box, perform the appropriate action:
If you want to..

Then..

Add the aggregates to an existing group

Select the appropriate group.

Add the aggregates to a new group

Type the name of the new group in the New Group field.

5. Click OK.

94 | OnCommand Console Help


Result

The aggregates are added to the group.


Related references

Administrator roles and capabilities on page 506

Monitoring physical storage


Overview of the monitoring process
Monitoring involves several processes. The DataFabric Manager server discovers the objects
available on your network, and then periodically monitors the objects and data that it collects from
the discovered objects, such as CPU usage, interface statistics, free disk space, qtree usage, and
storage object ID.
The DataFabric Manager server generates events when it discovers a storage system, when the status
is abnormal, or when a predefined threshold is breached. You can configure the DataFabric Manager
server to send a notification to a recipient when an event triggers an alarm.
Aggregate capacity thresholds and their events
You can configure capacity thresholds for aggregates and events for these thresholds from
DataFabric Manager server. You can set alarms to monitor the capacity and committed space of an
aggregate. You can also take corrective actions based on the event generated.
You can configure alarms to send notification whenever an event related to the capacity of an
aggregate occurs. For the Aggregate Full threshold, you can also configure an alarm to send
notification only when the condition persists over a specified time.
By default, if you have configured an alarm to alert you to an event, the DataFabric Manager server
issues the alarm only once per event. You can configure the alarm to repeat until you receive an
acknowledgment.
Note: If you want to set an alarm for a specific aggregate, you must create a group with that
aggregate as the only member.

You can set the following aggregate capacity thresholds:


Aggregate Full
(%)

Description: Specifies the percentage at which an aggregate is full.


Note: To reduce the number of Aggregate Full Threshold events generated,
you can set an Aggregate Full Threshold Interval. This causes DataFabric
Manager server to generate an Aggregate Full event only if the condition
persists for the specified time.

Default value: 90 percent


Event generated: Aggregate Full

Storage | 95

Event severity: Error


Corrective Action
Perform one or more of the following actions:

To free disk space, ask your users to delete files that are no longer needed
from volumes contained in the aggregate that generated the event.
You must add one or more disks to the aggregate that generated the event.
Note: After you add a disk to an aggregate, you cannot remove it
without first destroying all flexible volumes present in the aggregate to
which the disk belongs. You must destroy the aggregate after all the
flexible volumes are removed from the aggregate.

Aggregate Nearly
Full (%)

You must temporarily reduce the Snapshot reserve.


By default, the reserve is 20 percent of disk space. If the reserve is not in
use, reducing the reserve can free disk space, giving you more time to add
a disk.
There is no way to prevent Snapshot copies from consuming disk space
greater than the amount reserved for them. It is, therefore, important to
maintain a large enough reserve for Snapshot copies so that the active file
system always has space available to create new files or modify existing
ones. For more information about the Snapshot reserve, see the Data
ONTAP Data Protection Online Backup and Recovery Guide.

Description: Specifies the percentage at which an aggregate is nearly full.


Default value: 80 percent
The value for this threshold must be lower than the value for Aggregate Full
Threshold for DataFabric Manager server to generate meaningful events.
Event generated: Aggregate Almost Full
Event severity: Warning
Corrective action
Perform one or more of the actions mentioned in Aggregate Full.

Aggregate
Overcommitted
(%)

Description: Specifies the percentage at which an aggregate is


overcommitted.
Default value: 100 percent
Event generated: Aggregate Overcommitted
Event severity: Error
Corrective action
You should perform one or more of the following actions:

96 | OnCommand Console Help

You must create new free blocks in the aggregate by adding one or more
disks to the aggregate that generated the event.
Note: You must add disks with caution. After you add a disk to an
aggregate, you cannot remove it without first destroying all flexible
volumes present in the aggregate to which the disk belongs. You must
destroy the aggregate after all the flexible volumes are destroyed.

You must temporarily free some already occupied blocks in the aggregate
by taking unused flexible volumes offline.
Note: When you take a flexible volume offline, it returns any space it
uses to the aggregate. However, when you bring the flexible volume
online again, it requires the space again.

Aggregate Nearly
Overcommitted
(%)

Permanently free some already occupied blocks in the aggregate by


deleting unnecessary files.

Description: Specifies the percentage at which an aggregate is nearly


overcommitted.
Default value: 95 percent
The value for this threshold must be lower than the value for Aggregate Full
Threshold for DataFabric Manager server to generate meaningful events.
Event generated: Aggregate Almost Overcommitted
Event severity: Warning
Corrective action
Perform one or more of the actions provided in Aggregate Overcommitted.

Aggregate
Snapshot Reserve
Nearly Full
Threshold (%)

Description: Specifies the percentage of the Snapshot copy reserve on an


aggregate that you can use before the system generates the Aggregate
Snapshots Nearly Full event.
Default value: 80 percent
Event generated: Aggregate Snapshot Reserve Almost Full
Event severity: Warning
Corrective action: None
There is no way to prevent Snapshot copies from consuming disk space
greater than the amount reserved for them. If you disable the aggregate
Snapshot autodelete option, it is important to maintain a large enough
reserve.
See the Operations Manager Help for instructions on how to identify
Snapshot copies you can delete. For more information about the Snapshot

Storage | 97

reserve, see the Data ONTAP Data Protection Online Backup and Recovery
Guide.
Aggregate
Snapshot Reserve
Full Threshold
(%)

Description: Specifies the percentage of the Snapshot copy reserve on an


aggregate that you can use before the system generates the Aggregate
Snapshots Full event.
Default value: 90 percent
Event generated: Aggregate Snapshot Reserve Full
Event severity: Warning
Corrective action: None
There is no way to prevent Snapshot copies from consuming disk space
greater than the amount reserved for them.

Note: A newly created traditional volume tightly couples with its containing aggregate so that the
capacity of the aggregate determines the capacity of the new traditional volume. Therefore, you
should synchronize the capacity thresholds of traditional volumes with the thresholds of their
containing aggregates.
Related information

Data ONTAP Data Protection Online Backup and Recovery Guide - now.netapp.com/NOW/
knowledge/docs/ontap/ontap_index.shtml
Viewing the cluster inventory
You can use the Clusters view to monitor your inventory of clusters and view information about
related storage objects, capacity graphs, and cluster hierarchy details.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, click Clusters.
Result

The list of clusters are displayed.

98 | OnCommand Console Help


Related references

Administrator roles and capabilities on page 506


Viewing the storage controller inventory
You can use the Storage Controllers view to monitor your inventory of storage controllers, and view
information about related storage objects, and capacity graphs.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, click Storage Controllers.
Result

The list of storage controllers are displayed.


Related references

Administrator roles and capabilities on page 506


Viewing the aggregate inventory
You can use the Aggregates view to monitor your inventory of aggregates, and view information
about related storage objects, space usage details, and capacity graphs.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, click Aggregates.
Result

The list of aggregates are displayed.

Storage | 99
Related references

Administrator roles and capabilities on page 506


Viewing the disk inventory
You can monitor your inventory of disks and view information about the disk properties and related
storage objects from the Disks view.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, click Disks.
Result

The list of disks are displayed.


Related references

Administrator roles and capabilities on page 506


Viewing the deleted object inventory
You can use the Deleted Objects view to view a deleted object's properties, such as the object type,
and gather information about the time at which the object was deleted, the user who deleted it, and
whether the parent object was deleted. You can also retrieve the deleted objects by using this view.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, click Deleted Objects.
Result

The list of deleted objects are displayed.

100 | OnCommand Console Help


Related references

Administrator roles and capabilities on page 506

Page descriptions
Clusters view
The Clusters view displays detailed information about the clusters you are monitoring, as well as
their related objects, and also enables you to perform tasks such as editing the cluster settings,
grouping the clusters, and refreshing the monitoring samples.

Breadcrumb trail on page 100


Command buttons on page 100
List view on page 101
Overview tab on page 101
Graph tab on page 102
Cluster Hierarchy tab on page 102
Related Objects pane on page 102

Breadcrumb trail
The breadcrumb trail is created as you navigate from one list of storage objects to another. As you
navigate, each time you double-click certain items in these lists, another breadcrumb is added to
the trail, providing a string of hyperlinks that captures your navigation history. If you want to
revisit a previously viewed list, you can click the breadcrumb that links to it. You can also click the
icon in the breadcrumb to display a sub-navigation list of objects associated with the hyperlinked
object.
Command buttons
The command buttons enable you to perform the following tasks for a selected cluster:
Add

Launches the Storage Systems, All page in the Operations Manager console.
You can add clusters from this page.

Edit

Launches the Edit Cluster Settings in the Operations Manager console. You
can modify the cluster settings from this page.

Delete

Deletes the selected cluster. Deleting a cluster does not also delete the cluster
from the OnCommand console database, but the deleted cluster is no longer be
monitored.

Add to Group

Displays the Add to Group dialog box, which enables you to add the selected
cluster to the destination group.

Storage | 101
Refresh
Monitoring
Samples

Refreshes the database sample of the selected cluster and enables you to view
the updated details.

More Actions

Refresh

Refreshes the list of clusters.

View Events
Displays the events associated with the cluster in the Events tab. You can
sort the information based on the event severity, source ID, date of event
trigger, state, and so on.

Note: You can modify cluster settings, delete a cluster, add clusters to a group, refresh monitoring
samples, and view events for a cluster by right-clicking the selected cluster.

List view
The List view displays, in tabular format, the properties of all the discovered clusters. You can
customize your view of the data by clicking the filters for the columns.
You can double-click a cluster to display its child objects. The breadcrumb trail is modified to
display the selected cluster.
ID

Displays the cluster ID. By default, this column is hidden.

Name

Displays the name of the cluster.

Serial Number

Displays the serial number of the cluster.

Controller Count

Displays the number of controllers in the cluster.

Vserver Count

Displays the number of Vservers created in the cluster.

Location

Displays the physical location of the cluster.

Aggregate Used
Capacity (GB)

Displays the amount of space used in the aggregate.

Aggregate Total
Capacity (GB)

Displays the total capacity of the aggregate.

Primary IP Address

Displays the IP address of the cluster.

Status

Displays the current status of the cluster, based on the events generated
for the cluster. The status can be Normal, Warning, Error, Critical,
Emergency, or Unknown.

Overview tab
The Overview tab displays information about the selected cluster, such as the list of LIFs and ports.
Contact Email

Displays the e-mail address of the administrator for the cluster.

102 | OnCommand Console Help


Uptime

Displays the duration for which the cluster is online.

LIFs

Displays the list of logical interfaces on the cluster.

Ports

Displays the list of physical ports on the controllers in the cluster.

Graph tab
The Graph tab visually represents the various statistics about the clusters, such as performance and
capacity. You can select the graph you want to view from the drop-down list.
You can view the graphs representing a selected time period, such as one day, one week, one month,
to export graph details, such as the space savings
three months, or one year. You can also click
trend, used capacity, total capacity, and space savings achieved through deduplication.
Cluster Hierarchy tab
The Cluster Hierarchy tab displays details about the cluster objects in the selected cluster, such as
LIFs, storage controllers, Vservers, and aggregates.
Related Objects pane
The Related Objects section enables you to view and navigate to the groups, storage controllers,
aggregates, volumes, and Vservers related to the cluster.
Groups

Displays the groups to which the selected cluster belongs.

Storage Controllers

Displays the storage controllers in the selected cluster.

Aggregates

Displays the aggregates in the selected cluster.

Volumes

Displays the volumes in the selected cluster.

Vservers

Displays the Vservers in the selected cluster.

Related references

Window layout customization on page 16


Storage Controllers view
The Storage Controllers view displays detailed information about the storage controllers that are
monitored, as well as their related objects, and also enables you to perform tasks such as editing the
storage controller settings, grouping the storage controllers, and refreshing the monitoring samples.

Breadcrumb trail on page 103


Command buttons on page 103
List view on page 104
Map view on page 104

Storage | 103

Overview tab on page 107


Capacity tab on page 108
More Info tab on page 108
Graph tab on page 108
Related Objects pane on page 109

Breadcrumb trail
The breadcrumb trail is created as you navigate from one list of storage objects to another. As you
navigate, each time you double-click certain items in these lists, another breadcrumb is added to
the trail, providing a string of hyperlinks that captures your navigation history. If you want to
revisit a previously viewed list, you can click the breadcrumb that links to it. You can also click the
icon in the breadcrumb to display a sub-navigation list of objects associated with the hyperlinked
object.
Command buttons
The command buttons enable you to perform the following tasks for a selected storage controller:
Add

Launches the Storage Systems, All page in the Operations Manager console.
You can add storage controllers from this page.

Edit

Launches the Edit Storage Controller Settings page in the Operations


Manager console. You can modify the storage controller settings from this
page.

Delete

Deletes the selected storage controller.

Add to Group

Displays the Add to Group dialog box, which enables you to add the selected
storage controller to the destination group.

Refresh
Monitoring
Samples

Refreshes the database sample of the selected storage controller and enables
you to view the updated details.

More Actions

Refresh

Refreshes the list of storage controllers.

Grid (

View Events
Displays the events associated with the storage controller in the Events
tab. You can sort the information based on the event severity, source ID,
date of event trigger, state, and so on.

Displays a list view of the storage controllers.


Note: You can modify storage controller settings, delete a storage

controller, add storage controllers to a group, refresh monitoring samples,


and view events for a storage controller by right-clicking the selected
storage controller.

104 | OnCommand Console Help

TreeMap (

Displays a map view of the storage controllers.

List view
The List view displays, in tabular format, the properties of all the discovered storage controllers. You
can customize your view of the data by clicking the column filters.
You can double-click a storage controller to display its child objects. The breadcrumb trail is
modified to display the selected storage controller.
ID

Displays the storage controller ID. By default, this column is hidden.

Name

Displays the name of the storage controller.

Type

Displays the type of storage controller. The controller can be a stand-alone


system, a clustered system, or a node in an HA pair.

Status

Displays the current status of the storage controller based on the events
generated. The status can be Normal, Warning, Error, Critical, Emergency, or
Unknown.

Cluster

Displays the name of the cluster to which the storage controller belongs.

Model

Displays the model number of the storage controller.

Serial Number

Displays the serial number of the storage controller (This number is also
provided on the chassis).

System ID

Displays the ID number of the storage controller.

Used Capacity
(GB)

Displays the amount of space used by the storage controller.

Total Capacity
(GB)

Displays the total space available in the storage controller.

IP Address

Displays the IP address of the storage controller. By default, this column is


hidden.

State

Displays the current state of the storage controller. The state can be Up, Down,
Error, or Unknown. By default, this column is hidden.

Map view
The Map view enables you to view the properties of the storage controllers which are displayed as
rectangles with different sizes and colors. The size and color of the rectangles are based on the
options you select for the Size and Color fields in the properties area.
Storage
Controller
Filter

Enables you to display capacity, CPU utilization, and status information about the
storage controllers in varying rectangle sizes and colors:

Storage | 105

Size

Specifies the size of the rectangle based on the option you select from the
drop-down list. You can choose one of the following options:

Used Capacity (default): The amount of physical space (in GB) used
by application or user data in the storage controller. The size of the
rectangle increases when the value for used capacity increases.
Available Capacity: The amount of physical space (in GB) that is
available in the storage controller. The size of the rectangle increases
when the value for available capacity increases.
Committed Capacity: The amount of physical space (in GB) allocated
to user and application data. The size of the rectangle increases when
the value for committed capacity increases.
Saved Capacity: The amount of space (in GB) saved in the storage
controller. The size of the rectangle increases when the value for saved
capacity increases.
Status: The current status of the storage controller based on the events
generated. The size of the rectangle varies from large to small in the
following order: Emergency, Critical, Error, Warning, Normal, and
Unknown. For example, a controller with an Emergency status is
displayed as a larger rectangle than a controller with a Critical status.
CPU Utilization: The CPU usage (in percentage) of the storage
controller. The size of the rectangle increases when the value for CPU
utilization increases.

Color Specifies the color of the rectangle based on the option you select from the
drop-down list. You can choose one of the following options:

Status (default): The current status of the storage controller based on


the events generated. Each status displays a specific color: Emergency
(

), Critical (

), Error (

), Warning (

), Normal (

), and

Unknown (
).
Available %: The percentage of space available in the storage
controller. The color varies based on the specified threshold values and
the space available in the controller. For example, in a storage
controller with a size of 100 GB, if the Volume Nearly Full Threshold
and Volume Full Threshold are set to default values of 80% and 90%,
respectively, the color of the rectangle depends on the following
conditions:

If the available space in the controller is more than 20 GB and less

than 100 GB, the color displayed is green (


). When the
available space reduces, the green color changes to a lighter shade.
If the available space in the controller is less than 20 GB but more
than 10 GB, the color displayed is orange (

). When the

106 | OnCommand Console Help

available space reduces, the orange color changes to a darker


shade.
If the available space in the controller is less than 10 GB, the color

displayed is red (
). When the available space reduces, the red
color changes to a darker shade.
Used %: The percentage of space used in the storage controller. The
color displayed varies based on the following conditions:

If the used space in the controller is less than the Volume Nearly
Full Threshold value of the controller, the color displayed is green
(
). When the used space reduces, the green color changes to a
darker shade.
If the used space in the controller exceeds the Volume Nearly Full
Threshold value but is less than the Volume Full Threshold value
of the controller, the color displayed is orange (
). When the
used space reduces, the orange color changes to a lighter shade.
If the used space in the controller exceeds the Volume Full

Threshold value of the controller, the color displayed is red (


).
When the used space reduces, the red color changes to a lighter
shade.
Committed Capacity: The amount of physical space (in GB) allocated

). When the
to application or user data. The color displayed is blue (
committed capacity reduces, the blue color changes to a lighter shade.
Saved Capacity: The amount of space (in GB) saved in the storage

controller. The color displayed is blue (


). When the saved capacity
reduces, the blue color changes to a lighter shade.
CPU Utilization: The CPU usage (in percentage) of the storage
controller. The color displayed varies based on the following
conditions:

If the CPU usage of the controller is less than the Host CPU Too

Busy Threshold value, the color displayed is green (


). When the
CPU usage reduces, the green color changes to a darker shade.
If the CPU usage of the controller exceeds the Host CPU Too Busy
Threshold value, the color displayed is red (
). When the CPU
usage reduces, the red color changes to a lighter shade.

General

Enables you to filter storage controllers based on the name, status, or both.

Storage | 107

Note: You can filter by entering regular expressions instead of the full name of

the controller. For example, xyz* lists all the controllers that begin with the name
xyz.
Capacity

Enables you to filter storage controllers based on the used capacity, available
capacity, committed capacity, and saved capacity. You can specify the capacity
range by dragging the sliders.

Performance Enables you to filter storage objects based on the CPU utilization of the storage
controller. You can specify the CPU utilization range by dragging the sliders.
Overview tab
The Overview tab displays information about the selected storage controller, such as the IP address,
network interface connection, status, and AutoSupport details.
Name

Displays the name of the storage controller.

Operating
System

Displays the version of the operating system the storage controller is running.

IP Address

Displays the IP address of the storage controller in IPv4 or IPv6 format.

Network
Interface count

Displays the number of network interfaces on the storage controller.

Network
Interfaces

Displays the names of the network interfaces on the storage controller.

Status

Displays the current status of the storage controller based on the events
generated. The status can be Normal, Warning, Error, Critical, Emergency, or
Unknown.

Up Time

Displays the duration for which the storage controller is online.

Remote Platform Displays the status of the Remote LAN Module (RLM) card that is installed on
the controller. The status can be one of the following:
Management
Online

This status is displayed when the DataFabric Manager server


is communicating with the RLM card using the IP address
you have set.

Unavailable This status is displayed when the DataFabric Manager server


is unable to communicate with the RLM card. This can be due
to the following reasons:

The IP address of the card is not set.


The IP address of the card is set, but the DataFabric
Manager server is not able to communicate with the card.
The storage controller does not support an RLM card.

108 | OnCommand Console Help

The storage controller supports an RLM card, but the card


is not accessible.
The RLM card is not functioning.

You can perform remote maintenance operations for the storage controller by
clicking the remote platform management link.
Contact

Displays the contact information of the administrator for the storage controller.

Location

Displays the location of the storage controller.

AutoSupport

Displays the status of AutoSupport : "Yes" if AutoSupport is enabled and


"Unknown" if not.

Capacity tab
The Capacity tab displays information about the capacity of storage objects and disks within the
storage controller.
Storage
Capacity

Displays the number of aggregates, volumes, qtrees, or LUNs, if any, that the
storage system contains, including the capacity that is currently in use. You can click
the number corresponding to the storage capacity for more information.

Physical
Space

Displays the number of data, spare, and parity disks and their data capacities on the
storage controller. The Total disks field under Physical space provides information
about the disks on the storage controller. You can click the number corresponding to
Total Disks to view more information from the Disks view.

More Info tab


The More Info tab displays the following information about the broken disks in a storage controller:
Failed Disks

Displays the number of failed disks on the storage controller.

Failed Disk Info Displays the location of the failed disk on the storage controller.
Initiators

Displays the number of LUN initiators available in the storage controller. You
can double-click the number corresponding to the initiator for more information.

Protocols

Displays the list of protocols that are supported by the storage controller, such as
NFS, CIFS, FCP, and iSCSI.

Graph tab
The Graph tab visually represents the various statistics about the storage controller, such as
performance and capacity. You can select the graph you want to view from the drop-down list.

Storage | 109
You can view the graphs representing a selected time period, such as one day, one week, one month,
three months, or one year. You can also click
to export graph details, such as the space savings
trend, used capacity, total capacity, and space savings achieved through deduplication.
Related Objects pane
The Related Objects section enables you to view and navigate to the groups, volumes, aggregates,
and vFiler units related to the storage controller.
Groups

Displays the groups to which the storage controller belongs.

Volumes

Displays the volumes in the selected storage controller.

Aggregates

Displays the aggregates in the selected storage controller.

vFiler Units

Displays the vFiler units in the selected storage controller.

Related references

Window layout customization on page 16


Aggregates view
The Aggregates view displays detailed information about the aggregates in storage systems that are
monitored, as well as their related objects, and also enables you to perform tasks such as editing the
aggregate settings, grouping the aggregates, and refreshing monitoring samples.

Breadcrumb trail on page 109


Command buttons on page 110
List view on page 110
Map view on page 111
Overview tab on page 114
Capacity tab on page 115
Space Breakout tab on page 115
Graph tab on page 115
Related Objects pane on page 115

Breadcrumb trail
The breadcrumb trail is created as you navigate from one list of storage objects to another. As you
navigate, each time you double-click certain items in these lists, another breadcrumb is added to
the trail, providing a string of hyperlinks that captures your navigation history. If you want to
revisit a previously viewed list, you can click the breadcrumb that links to it. You can also click the
icon in the breadcrumb to display a sub-navigation list of objects associated with the hyperlinked
object.

110 | OnCommand Console Help


Command buttons
The command buttons enable you to perform the following tasks for a selected aggregate:
Edit

Launches the Edit Aggregate Settings in the Operations Manager console.


You can modify the aggregate settings from this page.

Delete

Deletes the selected aggregate.

Add to Group

Displays the Add to Group dialog box, which enables you to add the
selected aggregate to the destination group.

Refresh Monitoring Refreshes the database sample of the selected aggregate and enables you to
view the updated details.
Samples
More Actions

Refresh

Refreshes the list of aggregates.

Grid (

View Events
Displays the events associated with the aggregate in the Events tab. You
can sort the information based on the event severity, source ID, date of
event trigger, and state.

Displays the list view of the aggregate.

Note: You can add an aggregate to a group, refresh monitoring samples,


modify aggregate settings, view events for an aggregate, and delete an
aggregate by right-clicking the selected aggregate.

TreeMap (

Displays the map view of the aggregate.

List view
The list view displays, in tabular format, the properties of all the discovered aggregates. You can
customize your view of the data by clicking the column filters.
You can double-click the name of an aggregate to display its child objects. The breadcrumb trail is
modified to display the selected aggregate.
ID

Displays the aggregate ID. By default, this column is hidden.

Name

Displays the name of the aggregate.

Storage System

Displays the name of the storage system that contains the aggregate.

Type

Displays the type of aggregate selected. By default, this column is hidden.

Block Type

Displays the block format of the aggregate, either 32-bit or 64-bit.

RAID Type

Displays the RAID protection scheme, if specified. By default, this column is


hidden.
The RAID protection scheme can be one of the following:

Storage | 111

RAID 0

All the raid groups in the aggregate are of type raid0.

RAID 4

All the raid groups in the aggregate are of type raid4.

RAID DP

All the raid groups in the aggregate are of type


raid_dp.

Mirrored RAID 0

All the raid groups in the mirrored aggregate are of


type raid0.

Mirrored RAID 4

All the raid groups in the mirrored aggregate are of


type raid4.

Mirrored RAID
DP

All the raid groups in the mirrored aggregate are of


type raid_dp.

State

Displays the current state of an aggregate. The state can be Online, Offline, or
Unknown.

Status

Displays the current status of an aggregate. The status can be Normal,


Warning, Error, Critical, Emergency, or Unknown.

Used Capacity
(GB)

Displays the amount of space used in the aggregate.

Available
Capacity (GB)

Displays the amount of space available in the aggregate.

Committed
Capacity (GB)

Displays the total space reserved for all flexible volumes on an aggregate.

Total Capacity
(GB)

Displays the total space of the aggregate.

Host ID

Displays the host ID to which the aggregate is related. By default, this column
is hidden.

Map view
The Map view enables you to view the properties of the aggregates which are displayed as rectangles
with different sizes and colors. The size and color of the rectangles are based on the options you
select for the Size and Color fields in the properties area.
Aggregate
Filter

Enables you to display capacity and status information about the aggregates in
varying rectangle sizes and colors:
Size

Specifies the size of the rectangle based on the option you select from the
drop-down list. You can select one of the following options:

Used % (default): The percentage of space used in the aggregate. The


size of the rectangle increases when the value for used space increases.

112 | OnCommand Console Help

Available %: The percentage of space available in the aggregate. The


size of the rectangle increases when the value for available space
increases.
Growth Rate: The rate at which data in the aggregate is increasing. The
size of the rectangle increases when the value for growth rate increases.
Unused Snapshot Reserve: The amount of unused Snapshot reserve
space (in GB). The size of the rectangle increases when the value for
unused snapshot reserve increases.
Days to Full: The number of days needed for the aggregate to reach full
capacity. The size of the rectangle increases when the value for days to
full increases.
Size: The total size (in GB) of the aggregate. The size of the rectangle
increases when the value for size increases.
Used Capacity: The amount of physical space (in GB) used by
application or user data in the aggregate. The size of the rectangle
increases when the value for used capacity increases.
Available Capacity: The amount of physical space (in GB) that is
available in the aggregate. The size of the rectangle increases when the
value for available capacity increases.
Committed Capacity: The amount of physical space (in GB) allocated
to user and application data. The size of the rectangle increases when
the value for committed capacity increases.
Saved Capacity: The amount of space (in GB) saved in the aggregate.
The size of the rectangle increases when the value for saved capacity
increases.
Status: The current status of the aggregate based on the events
generated. The size of the rectangle varies from large to small in the
following order: Emergency, Critical, Error, Warning, Normal, and
Unknown. For example, an aggregate with an Emergency status is
displayed as a larger rectangle than an aggregate with a Critical status.

Color Specifies the color of the rectangle based on the option you select from the
drop-down list. You can select one of the following options:

Status (default): The current status of the aggregate based on the events
generated. Each status displays a specific color: Emergency (
Critical (

), Error (

), Warning (

), Normal (

),

), and

Unknown (
).
Used %: The percentage of space used in the aggregate. The color
displayed varies based on the following conditions:

If the used space in the aggregate is less than the Nearly Full
Threshold value of the aggregate, the color displayed is green

Storage | 113

(
). When the used space reduces, the green color changes to a
darker shade.
If the used space in the aggregate exceeds the Nearly Full
Threshold value but is less than the Full Threshold value of the
aggregate, the color displayed is orange (
). When the used space
reduces, the orange color changes to a lighter shade.
If the used space in the aggregate exceeds the Full Threshold value

of the aggregate, the color displayed is red (


). When the used
space reduces, the red color changes to a lighter shade.
Aggregate with Snapshot Reserve: The size (in GB) of the Snapshot
reserve in an aggregate. If the aggregate has an allocated Snapshot
reserve, (

) is displayed. If the aggregate does not have an allocated

) is displayed.
Snapshot reserve, (
Unused Snapshot Reserve: The amount of unused Snapshot reserve
space (in GB) in the aggregate. The color varies based on the specified
threshold values and the unused Snapshot reserve space in the
controller. For example, in an aggregate with a size of 100 GB, if the
Aggregate Snapshot Reserve Nearly Full and Aggregate Snapshot
Reserve Full Threshold are set to default values of 80% and 90%,
respectively, the color of the rectangle depends on the following
conditions:

If the unused Snapshot reserve space in the aggregate is more than

20 GB but less than 100 GB, the color displayed is green (


).
When the unused Snapshot reserve space reduces, the green color
changes to a lighter shade.
If the unused Snapshot reserve space in the aggregate is less than 20

GB but more than 10 GB, the color displayed is orange (


).
When the unused Snapshot reserve space reduces, the orange color
changes to a darker shade.
If the unused Snapshot reserve space in the aggregate is less than 10

GB, the color displayed is red (


). When the unused Snapshot
reserve space reduces, the red color changes to a darker shade.
Raid-4/DP: The RAID protection scheme, if specified. The color
displayed varies based on the raid protection types: raid0 (
(

), raid_dp (

), and mixed_raid_type (

).

), raid4

114 | OnCommand Console Help

Size: The total size (in GB) of the aggregate. The color displayed is
blue (
). When the size of the aggregate reduces, the blue color
changes to a lighter shade.
Available %: The percentage of space available in the aggregate. The
color varies based on the specified threshold values and the space
available in the aggregate. For example, in an aggregate with a size of
100 GB, if the Aggregate Nearly Full Threshold and Aggregate Full
Threshold are set to default values of 80% and 90%, respectively, the
color of the rectangle depends on the following conditions:

If the available space in the aggregate is more than 20 GB but less

than 100 GB, the color displayed is green (


). When the available
space reduces, the green color changes to a lighter shade.
If the available space in the aggregate is less than 20 GB but more

than 10 GB, the color displayed is orange (


). When the available
space reduces, the orange color changes to a darker shade.
If the available space in the aggregate is less than 10 GB, the color

displayed is red (
). When the available space reduces, the red
color changes to a darker shade.
Committed Capacity: The amount of physical space (in GB) allocated

). When the
to application or user data. The color displayed is blue (
committed capacity reduces, the blue color changes to a lighter shade.
Saved Capacity: The amount of space (in GB) saved in the aggregate.
The color displayed is blue (
). When the saved capacity reduces,
the blue color changes to a lighter shade.

General

Enables you to filter aggregates based on the name, status, or both.


Note: You can filter by entering regular expressions instead of the full name of the

aggregate. For example, xyz* lists all the aggregates that begin with the name xyz.
Capacity

Enables you to filter aggregates based on the used %, growth rate, used capacity,
available capacity, saved capacity, and so on. You can specify the capacity range by
dragging the sliders.

Overview tab
The Overview tab displays details about the selected aggregate, such as the storage object name and
options to enable Snapshot copies.
Full Name

Displays the full name of the aggregate.

Storage | 115
Storage System

Displays the name of the storage system that contains the aggregate. You
can view more information about the storage system by clicking the link.

Snapshot Copies
Enabled

Specifies whether automatic Snapshot copies are enabled.

Snapshot Auto
Delete

Specifies whether a Snapshot copy will be deleted to free space when a write
to a volume fails due to lack of space in the aggregate.

Capacity tab
The Capacity tab displays information about the capacity of storage objects and disks within the
storage system.
Storage
Capacity

Displays the number of volumes and qtrees, if any, that the aggregate contains, if
any, including the capacity uses by each object. You can click the number to
display the volumes or qtrees contained in the volume.

Physical
Space

Displays the number and capacity of data disks and parity disks assigned to the
aggregate. You can click the number corresponding to Total Disks to view more
information from the Disks view.

Space Breakout tab


The Space Breakout tab displays information about the space used, and space available, for Snapshot
copies and data.
Graph tab
The Graph tab visually represents the performance characteristics of an aggregate. You can select the
graph you want to view from the drop-down list.
You can view the graphs representing a selected time period, such as one day, one week, one month,
three months, or one year. You can also click
trend, used capacity, and total capacity.

to export graph details, such as the used capacity

Related Objects pane


The Related Objects section enables you to view and navigate to the groups, storage controllers,
volumes, and disks related to the aggregate.
Groups

Displays the groups to which the aggregate belongs.

Storage Controllers

Displays the storage controllers that contain the selected aggregate.

Volumes

Displays the volumes in the selected aggregate.

Disks

Displays the disks in the selected aggregate.

116 | OnCommand Console Help


Related references

Window layout customization on page 16


Disks view
The Disks view displays detailed information about disks and its related objects.

Breadcrumb trail on page 116


Command button on page 116
List view on page 116
Overview tab on page 117
Related objects pane on page 117

Breadcrumb trail
The breadcrumb trail is created as you navigate from one list of storage objects to another. As you
navigate, each time you double-click certain items in these lists, another breadcrumb is added to
the trail, providing a string of hyperlinks that captures your navigation history. If you want to
revisit a previously viewed list, you can click the breadcrumb that links to it. You can also click the
icon in the breadcrumb to display a sub-navigation list of objects associated with the hyperlinked
object.
Command button
The command button enables you to perform the following task for a selected disk:
Refresh

Refreshes the list of disks.

List view
The List view displays, in tabular format, the properties of all the discovered disks. You can
customize your view of the data by clicking the column filters.
ID

Displays the disk ID. By default, this column is hidden.

Disk Name

Displays the name of the disk.

Controller

Displays the name of the storage controller that contains the disk.

Aggregate

Displays the name of the parent aggregate of the disk.

Aggregate ID Displays the ID of the aggregate to which the disk belongs. By default, this column
is hidden.
Type

Displays the type of the disk.

Size (GB)

Displays the size of the disk.

Shelf ID

Displays the ID of the shelf on which the disk is located.

Storage | 117
Bay ID

Displays the ID of the bay within the shelf on which the disk is located.

Plex ID

Displays the ID of the plex to which the disk is assigned.

Status

Displays the current status of the disk, such as Active, Reconstruction in Progress,
Scrubbing in Progress, Failed, Spare, or Offline.

Host ID

Displays the ID of the host to which the disk is related. By default, this column is
hidden.

Overview tab
The Overview tab displays the following information about the selected disk:
Firmware Revision Number

Displays the latest version of the firmware installed on the disk.

Vendor

Displays the name of the disk vendor.

Disk Model

Displays the model number of the disk.

Related objects pane


The Related Objects section displays the aggregates related to the disk.
Aggregates

Displays the aggregates in the selected disk.

Related references

Window layout customization on page 16


Deleted Objects view
The Deleted Objects view displays detailed information about the storage objects that you deleted.
Deleting a storage object does not also delete the object from the OnCommand console database, but
the deleted object is no longer monitored by OnCommand console. You can also restore deleted
storage objects from the Deleted Objects view.

Breadcrumb trail on page 117


Command buttons on page 118
List view on page 118

Breadcrumb trail
The breadcrumb trail is created as you navigate from one list of storage objects to another. As you
navigate, each time you double-click certain items in these lists, another breadcrumb is added to
the trail, providing a string of hyperlinks that captures your navigation history. If you want to
revisit a previously viewed list, you can click the breadcrumb that links to it. You can also click the
icon in the breadcrumb to display a sub-navigation list of objects associated with the hyperlinked
object.

118 | OnCommand Console Help


Command buttons
The command buttons enable you to perform the following tasks for a selected storage object:
Recover Enables you to recover a storage object that has been deleted. You cannot recover a
storage object that has been deleted by the DataFabric Manager server. Also, you cannot
recover a storage object whose parent object is deleted.
Data collection and reporting for a storage object resumes when you recover it.
Note: You can also recover a storage object by right-clicking the selected deleted
object.

Refresh Refreshes the list of deleted storage objects.


List view
The List view displays, in tabular format, the properties of the deleted storage objects. You can
customize your view of the data by clicking the column filters.
ID

Displays the storage object ID. By default, this column is hidden.

Name

Displays the name of the storage object.

Type

Displays the storage object type, such as volume, qtree, or aggregate.

Storage Server Displays the parent of the storage object, such as volume, vFiler unit, or Vserver.
Deleted

Displays the date and time that the storage object was deleted.

Deleted By

Displays the name of the user who deleted the storage object.

Parent Deleted Displays "Yes" if the parent object is deleted or displays "No" if the parent object
is not deleted.
Parent ID

Displays the ID of the parent object. By default, this column is hidden.

Parent Name

Displays the name of the parent object. By default, this column is hidden.

Related references

Window layout customization on page 16

Virtual storage
Understanding virtual storage

Storage | 119

What vFiler units are


A vFiler unit is a partition of a storage system and the associated network resources. Each vFiler
partition appears to the user as a separate storage system on the network and functions as a storage
system.
Access to vFiler units can be restricted so that an administrator can manage and view files only on an
assigned vFiler unit, not on other vFiler units that reside on the same storage system. In addition,
there is no data flow between vFiler units. When using vFiler units, you can be sure that no sensitive
information is exposed to other administrators or users who store data on the same storage system.
You can assign volumes or LUNs to vFiler units in NetApp Management Console. You can create up
to 65 vFiler units on a storage system.
To use vFiler units you must have the MultiStore software licensed on the storage system that is
hosting the vFiler units.
You can use vFiler templates to simplify creation of vFiler units. You create a template by selecting a
set of vFiler configuration settings, including CIFS, DNS, NIS, and administration host information.
You can configure as many vFiler templates as you need.
Discovery of vFiler units
The OnCommand console monitors the hosting storage systems to discover vFiler units. You must
set authentication credentials for the hosting storage system to ensure that the OnCommand console
discovers the vFiler units.
Attention: If you encounter a "timed out" error message when setting the authentication

credentials for the hosting storage system, you must set the credentials again.
The server monitors the hosting storage system once every hour to discover new vFiler units that you
configured on the storage system. The server deletes from the database the vFiler units that you
destroyed on the storage system.
You can change the default monitoring interval from the Monitoring setup options, or by using the
following CLI command:
dfm option set vFilerMonInterval=1hour

You can disable the vFiler discovery from the Discovery setup options, or by using the dfm option
set discovervfilers=no CLI command.
When the OnCommand console discovers a vFiler unit, it does not add the network to which the
vFiler unit belongs to its list of networks on which it runs host discovery. In addition, when you
delete a network, the server continues to monitor the vFiler units in that network.

120 | OnCommand Console Help

What Vservers are


A Vserver represents a single file-system namespace. It has separate network access and provides the
same flexibility and control as a dedicated node.Each Vserver has its own user domain and security
domain that can span multiple physical nodes.
A Vserver has a root volume that constitutes the top level of the namespace hierarchy; additional
volumes are mounted to the root volume to extend the namespace. A Vserver is associated with one
or more logical interfaces through which clients access the data on the storage system (or Vserver).
Clients can access the Vserver from any node in the cluster through the logical interfaces that are
associated with the Vserver.
Note: A namespace provides a context for determining the junctions that link together a collection

of volumes. All the volumes associated with a Vserver are accessed from the Vserver's namespace.

Managing virtual storage


Editing vFiler unit settings
You can edit the following vFiler unit settings from the Edit vFiler Settings page: primary IP address,
login details, password, login protocol, owner name and e-mail address, resource tag, Host.equiv
option, and vFiler unit thresholds.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

During this task, the OnCommand console launches the Operations Manager console. Depending on
your browser configuration, you can return to the OnCommand console by using the Alt-Tab key
combination or clicking the OnCommand console browser tab. After the completion of this task, you
can leave the Operations Manager console open, or you can close it to conserve bandwidth.
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, select the vFiler Units view.
3. Select the vFiler unit you want to modify.
4. Click Edit.
The Edit vFiler Settings page is displayed in the Operations Manager console.
5. Modify the properties of the vFiler unit, as required.

Storage | 121
6. Click Update.
7. Press Alt-Tab or click the OnCommand console browser tab to return to the OnCommand
console.
Result

The modified settings are updated in the DataFabric Manager server.


Related references

Administrator roles and capabilities on page 506


Editing Vserver settings
You can edit the following virtual server settings from the Edit Vserver Settings page: the primary IP
address, resource tag, and owner e-mail address and name.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

During this task, the OnCommand console launches the Operations Manager console. Depending on
your browser configuration, you can return to the OnCommand console by using the Alt-Tab key
combination or clicking the OnCommand console browser tab. After the completion of this task, you
can leave the Operations Manager console open, or you can close it to conserve bandwidth.
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, select the Vservers view.
3. Select the virtual server you want to modify.
4. Click Edit.
The Edit vserver Settings page is displayed in the Operations Manager console.
5. Modify the properties of the virtual server, as required.
6. Click Update.
7. Press Alt-Tab or click the OnCommand console browser tab to return to the OnCommand
console.

122 | OnCommand Console Help


Result

The modified settings are updated in the DataFabric Manager server.


Related references

Administrator roles and capabilities on page 506


Adding vFiler units to a group
You can add a vFiler unit to an existing group. You can also create a new group and add a vFiler unit
to it. This enables you to easily monitor all the vFiler units that belong to a group.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, select the vFiler Units view.
3. Select the vFiler unit you want to add to the group.
4. In the Add to Group dialog box, perform the appropriate action:
If you want to..

Then..

Add the vFiler unit to an existing group.

Select the appropriate group.

Add the vFiler unit to a new group.

Type the name of the new group in the New Group field.

5. Click OK.
Result

The vFiler unit is added to the group.


Related references

Administrator roles and capabilities on page 506

Storage | 123

Adding Vservers to a group


You can add a virtual server to an existing group. You can also create a new group and add a virtual
server to it. This enables you to easily monitor all the virtual servers that belong to a group.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, select the Vservers view.
3. Select the virtual server you want to add to the group.
4. In the Add to Group dialog box, perform the appropriate action:
If you want to..

Then..

Add the virtual server to an existing group.

Select the appropriate group.

Add the virtual server to a new group.

Type the name of the new group in the New Group field.

5. Click OK.
Result

The virtual server is added to the group.


Related references

Administrator roles and capabilities on page 506

Monitoring virtual storage


Prerequisites for monitoring vFiler units
Before you enable monitoring of vFiler units, you must ensure that the hosting storage system is
running a supported Data ONTAP release and it is part of the same routable network as the
DataFabric Manager server. NDMP discovery must also be enabled.
You must meet the following requirements before monitoring vFiler units:

Supported Data ONTAP release


The MultiStore monitoring feature supports hosting storage systems running Data ONTAP 6.5 or
later.

124 | OnCommand Console Help


Note: To run a command on a vFiler unit using a Secure Socket Shell (SSH) connection, you
must ensure that the hosting storage system is running Data ONTAP 7.2 or later.

Network connectivity
To monitor a vFiler unit, DataFabric Manager server and the hosting storage system must be part
of the same routable network that is not separated by firewalls.
Hosting storage system discovery and monitoring
You must first discover and monitor the hosting storage system before discovering and
monitoring the vFiler units.
NDMP discovery
DataFabric Manager server uses NDMP as the discovery method to manage SnapVault and
SnapMirror relationships between vFiler units. To use NDMP discovery, you must first enable
SNMP and HTTPS discovery.
Monitoring the default vFiler unit
When you enable your core license, which includes MultiStore, Data ONTAP automatically
creates a default vFiler unit on the hosting storage system unit called vfiler0. The OnCommand
console does not provide vfiler0 details.
Editing user quotas
To edit user quotas that are configured on vFiler units, ensure that the hosting storage systems are
running Data ONTAP 6.5.1 or later.
Monitoring backup relationships
For hosting storage systems that are backing up data to a secondary system, you must ensure that
the secondary system is added to the vFiler group. DataFabric Manager server collects details
about vFiler unit backup relationships from the hosting storage system. You can then view the
backup relationships if the secondary storage system is assigned to the vFiler group, even though
the primary system is not assigned to the same group.
Monitoring SnapMirror relationships
For hosting storage systems that are mirroring data to a secondary system, you must ensure that
the secondary system is added to the vFiler group. DataFabric Manager server collects details
about vFiler unit SnapMirror relationships from the hosting storage system. DataFabric Manager
server displays the relationships if the destination vFiler unit is assigned to the vFiler group, even
though the source vFiler unit is not assigned to the same group.

Viewing the vFiler unit inventory


You can use the vFiler Units view to monitor your inventory of vFiler units and view information
about their properties, related storage objects, and capacity graphs.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.

Storage | 125
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, click vFiler Units.
Result

The list of vFiler units are displayed.


Related references

Administrator roles and capabilities on page 506


Viewing the Vserver inventory
You can use the Vservers view to monitor your inventory of Vservers and view information about
their properties, related objects, capacity graphs, and namespace hierarchy details.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, click Vservers.
Result

The list of Vservers are displayed.


Related references

Administrator roles and capabilities on page 506

Page descriptions
vFiler Units view
The vFiler Units view displays detailed information about the vFiler units that are monitored, as well
as their related objects, and also enables you to perform tasks such as editing the vFiler unit settings,
grouping the vFiler units, and refreshing the monitoring samples.

Breadcrumb trail on page 126


Command buttons on page 126
List view on page 127

126 | OnCommand Console Help

Map view on page 127


Overview tab on page 129
Capacity tab on page 129
Graph tab on page 129
Related Objects pane on page 130

Breadcrumb trail
The breadcrumb trail is created as you navigate from one list of storage objects to another. As you
navigate, each time you double-click certain items in these lists, another breadcrumb is added to
the trail, providing a string of hyperlinks that captures your navigation history. If you want to
revisit a previously viewed list, you can click the breadcrumb that links to it. You can also click the
icon in the breadcrumb to display a sub-navigation list of objects associated with the hyperlinked
object.
Command buttons
The command buttons enable you to perform the following tasks for a selected vFiler unit:
Edit

Launches the Edit vFiler Settings page in the Operations Manager console.
You can modify the vFiler unit settings from this page.

Delete

Deletes the selected vFiler unit.

Add To Group

Displays the Add to Group dialog box, which enables you to add the selected
vFiler unit to a destination group.

Refresh
Monitoring
Samples

Refreshes the database sample of the selected vFiler unit and enables you to
view the updated details.

More Actions

Refresh

Refreshes the list of vFiler units.

Grid (

View Events
Displays the events associated with the vFiler unit in the Events tab. You
can sort the information based on the event severity, source ID, date of
event trigger, and state.

Displays the list view of the vFiler units.

Note: You can add a vFiler unit to a group, refresh monitoring samples,

modify the settings for a vFiler unit, view events for a vFiler unit, and
delete a vFiler unit by right-clicking the selected vFiler unit.
TreeMap (

Displays the map view of the vFiler units.

Storage | 127
List view
The list view displays, in tabular format, the properties of all the discovered vFiler units. You can
customize your view of the data by clicking the column filters.
ID

Displays the vFiler unit ID. By default, this column is hidden.

Name

Displays the name of the vFiler unit.

Hosting Storage
System

Displays the full name of the hosting storage system of the vFiler unit.

IP Space

Displays the IP space in which the vFiler unit is created and can subsequently
participate.

Primary IP
Address

Displays the primary IP address of the network interface.

Status

Displays the current status of a vFiler unit. The status can be Normal,
Warning, Error, Critical, Emergency, or Unknown.

State

Displays "Up" if the vFiler unit is online and, "Down" if not. By default, this
column is hidden.

Map view
The Map view enables you to view the properties of the vFiler units which are displayed as
rectangles with different sizes and colors. The size and color of the rectangles are based on the
options you select for the Size and Color fields in the properties area.
vFiler
Filter

Enables you to display capacity and status information about the vFiler units in
varying rectangle sizes and colors:
Size

Specifies the size of the rectangle based on the option you select from the
drop-down list. You can select one of the following options:

Used % (default): The percentage of space used in the vFiler unit. The
size of the rectangle increases when the value for used space increases.
Available %: The percentage of space available in the vFiler unit. The
size of the rectangle increases when the value for available space
increases.
Used Capacity: The amount of physical space (in GB) used by
application or user data in the vFiler unit. The size of the rectangle
increases when the value for used capacity increases.
Available Capacity: The amount of physical space (in GB) that is
available in the vFiler unit. The size of the rectangle increases when the
value for available capacity increases.
Status: The current status of the vFiler unit based on the events
generated. The size of the rectangle varies from large to small in the

128 | OnCommand Console Help

following order: Emergency, Critical, Error, Warning, Normal, and


Unknown. For example, a vFiler unit with an Emergency status is
displayed as a larger rectangle than a vFiler unit with a Critical status.
Color Specifies the color of the rectangle based on the option you select from the
drop-down list. You can select one of the following options:

Status (default): The current status of the vFiler unit based on the events
generated. Each status displays a specific color: Emergency (
Critical (

), Warning (

), Normal (

), and Unknown

(
).
Used %: The percentage of space used in the vFiler unit. The color
displayed varies based on the following conditions:

), Error (

),

If the used space in the vFiler unit is less than the Volume Nearly
Full Threshold value, the color displayed is green (
). When the
used space reduces, the green color changes to a darker shade.
If the used space in the vFiler unit exceeds the Volume Nearly Full
Threshold value but is less than the Volume Full Threshold value,
). When the used space reduces,
the color displayed is orange (
the orange color changes to a lighter shade.
If the used space in the vFiler unit exceeds the Volume Full

). When the used


Threshold value, the color displayed is red (
space reduces, the red color changes to a lighter shade.
Available %: The percentage of space available in the vFiler unit. The
color varies based on the following conditions:

If the available space in the vFiler unit exceeds the Volume Full
Threshold value, the color displayed is green (
). When the
available space reduces, the green color changes to a lighter shade.
If the available space in the vFiler unit is less than the Volume Full
Threshold value but exceeds the Volume Nearly Full Threshold
value, the color displayed is orange (
). When the available space
reduces, the orange color changes to a darker shade.
If the available space in the vFiler unit is less than the Volume
Nearly Full Threshold value, the color displayed is red (
). When
the available space reduces, the red color changes to a darker shade.

General

Enables you to filter vFiler units based on the name, status, or both.

Storage | 129

Note: You can filter by entering regular expressions instead of the full name of the

vFiler unit. For example, xyz* lists all the vFiler units that begin with the name xyz.
Capacity

Enables you to filter storage objects based on used capacity, available capacity, used
%, and available %. You can specify the capacity range by dragging the sliders.

Overview tab
The Overview tab displays details about the selected vFiler unit, such as the system ID, information
about protocols, ping status, and domains.
Name

Displays the name of the vFiler unit.

Hosting Storage System Displays the full name of the hosting storage system of the vFiler unit.
System ID

Displays the universal unique identifier (UUID) of the vFiler unit.

Ping timestamp

Displays the date and time that this vFiler unit was last queried.

Ping status

Displays the status of the ping request sent to the vFiler unit.

Protocols enabled

Displays the list of protocols enabled for the vFiler unit.

NFS service

Displays the service status (Up or Down) of the NFS protocol.

CIFS service

Displays the service status (Up or Down) of the CIFS protocol.

iSCSI service

Displays the service status (Up or Down) of the iSCSI protocol.

Capacity tab
The Capacity tab displays information about the capacity volumes and qtrees that were added at the
time of creating the vFiler unit.
Volume Displays the number of volumes the vFiler unit contains, including the used and total
capacity of the volumes. By clicking the number corresponding to the volumes, you can
view more information about these volumes from the Volumes view.
Qtree Displays the number of qtrees contained within the volumes in the vFiler unit, including the
used and total capacity of the qtrees. By clicking the number corresponding to the qtrees,
you can view more information about these qtrees from the Qtrees view.
Graph tab
The Graph tab visually represents the performance of a vFiler unit. The graphs display the volume
capacity used versus the total capacity in the vFiler unit, vFiler capacity used, volume capacity used,
and CPU usage (%). You can select the graphs from the drop-down list.

130 | OnCommand Console Help


You can view the graphs representing a selected time period, such as one day, one week, one month,
three months, or one year. You can also click
to export graph details, such as the space savings
trend, used capacity, total capacity, and space savings achieved through deduplication.
Related Objects pane
The Related Objects section enables you to view and navigate to the groups, volumes, and qtrees
related to the vFiler unit.
Groups

Displays the groups to which the vFiler unit belongs.

Volumes Displays the volumes in the selected vFiler unit. The volumes that were added after the
vFiler creation are displayed along with the volumes added during the vFiler creation.
Qtrees

Displays the qtrees in the selected vFiler unit. The qtrees that were added after the vFiler
creation are displayed along with the qtrees added during the vFiler creation.

Related references

Window layout customization on page 16


Vservers view
The Vservers view displays detailed information about the Vservers that are monitored, as well as
their related objects, and also enables you to perform tasks such as editing the Vserver settings,
grouping the Vservers, and refreshing monitoring samples.

Breadcrumb trail on page 130


Command buttons on page 131
List view on page 131
Map view on page 132
Overview tab on page 133
Graph tab on page 134
Namespace Hierarchy tab on page 134
Related Objects pane on page 134

Breadcrumb trail
The breadcrumb trail is created as you navigate from one list of storage objects to another. As you
navigate, each time you double-click certain items in these lists, another breadcrumb is added to
the trail, providing a string of hyperlinks that captures your navigation history. If you want to
revisit a previously viewed list, you can click the breadcrumb that links to it. You can also click the
icon in the breadcrumb to display a sub-navigation list of objects associated with the hyperlinked
object.

Storage | 131
Command buttons
Edit

Launches the Edit Vserver Settings page in the Operations Manager console.
You can modify the Vserver settings from this page.

Delete

Deletes the selected Vserver.

Add To Group

Displays the Add to Group dialog box, which enables you to add the selected
Vserver to a destination group.

Refresh Monitoring Refreshes the database sample of the selected Vserver and enables you to
view the updated details.
Samples
More Actions

Refresh

Refreshes the list of Vservers.

Grid (

View Events
Displays the events associated with the Vserver in the Events tab. You
can sort the information based on the event severity, source ID, date of
event trigger, and state.

Displays the list view of the Vserver.

Note: You can add a Vserver to a group, refresh monitoring samples,


modify the settings for a Vserver, view events for a Vserver, and delete a
Vserver by right-clicking the selected Vserver.

TreeMap (

Displays the map view of the Vserver.

List view
The list view displays, in tabular format, the properties of all the discovered Vservers. You can
customize your view of the data by clicking the column filters.
You can double-click a Vserver to display its child objects. The breadcrumb trail is modified to
display the selected Vserver.
ID

Displays the Vserver ID. By default, this column is hidden.

Name

Displays the name of the Vserver.

Cluster

Displays the name of the cluster to which the Vserver belongs.

Root Volume

Displays the name of the root volume of the Vserver.

Name Service
Switch

Displays the information type gathered from hosts.

NIS Domain

Displays the Network Information Service (NIS) domain name.

Status

Displays the current status of a Vserver. The status can be Critical, Error,
Warning, Normal, or Unknown.

132 | OnCommand Console Help


Map view
The Map view enables you to view the properties of the Vservers which are displayed as rectangles
with different sizes and colors. The size and color of the rectangles are based on the options you
select for the Size and Color fields in the properties area.
Vserver
Filter

Enables you to display capacity and status information about the Vservers in varying
rectangle sizes and colors:
Size

Specifies the size of the rectangle based on the option you select from the
drop-down list. You can select one of the following options:

Used % (default): The percentage of space used in the Vserver. The size
of the rectangle increases when the value for used space increases.
Available %: The percentage of space available in the Vserver. The size
of the rectangle increases when the value for available space increases.
Saved %: The percentage of space saved in the Vserver. The size of the
rectangle increases when the value for saved space increases.
Used Capacity: The amount of physical space (in GB) used by
application or user data in the Vserver. The size of the rectangle
increases when the value for used capacity increases.
Available Capacity: The amount of physical space (in GB) that is
available in the Vserver. The size of the rectangle increases when the
value for available capacity increases.
Saved Capacity: The amount of physical space (in GB) that is saved in
the Vserver. The size of the rectangle increases when the value for saved
capacity increases.
Status: The current status of the Vserver based on the events generated.
The size of the rectangle varies from large to small in the following
order: Emergency, Critical, Error, Warning, Normal, and Unknown. For
example, a Vserver with an Emergency status is displayed as a larger
rectangle than a Vserver with a Critical status.

Color Specifies the color of the rectangle based on the option you select from the
Color drop-down list. The options can be one of the following:

Status (default): The current status of the Vserver based on the events
generated. Each status displays a specific color: Emergency (
Critical (

), Error (

), Warning (

), Normal (

),

), and Unknown

(
).
Used %: The percentage of space used in the Vserver. The color
displayed varies based on the following conditions:

Storage | 133

Threshold value, the color displayed is green (


). When the used
space reduces, the green color changes to a darker shade.
If the used space in the Vserver exceeds the Volume Nearly Full
Threshold value but is less than the Volume Full Threshold value,
the color displayed is orange (
). When the used space reduces,
the orange color changes to a lighter shade.
If the used space in the Vserver exceeds the Volume Full Threshold

value, the color displayed is red (


). When the used space reduces,
the red color changes to a lighter shade.
Available %: The percentage of space available in the Vserver. The
color varies based on the following conditions:

If the used space in the Vserver is less than the Volume Nearly Full

If the available space in the Vserver exceeds the Volume Full


Threshold value, the color displayed is green (
). When the
available space reduces, the green color changes to a lighter shade.
If the available space in the Vserver is less than the Volume Full
Threshold value but exceeds the Volume Nearly Full Threshold
). When the available space
value, the color displayed is orange (
reduces, the orange color changes to a darker shade.
If the available space in the Vserver is less than the Volume Nearly

). When the
Full Threshold value, the color displayed is red (
available space reduces, the red color changes to a darker shade.
Saved Capacity: The amount of space (in GB) saved in the Vserver. The
color displayed is blue (
). When the saved capacity reduces, the blue
color changes to a lighter shade.

General

Enables you to filter Vservers based on the name, status, or both.


Note: You can filter by entering regular expressions instead of the full name of the
Vserver. For example, xyz* lists all the Vservers that begin with the name xyz.

Capacity

Enables you to filter storage objects based on used %, available %, saved %, used
capacity, and available capacity. You specify the capacity range by dragging the
sliders.

Overview tab
The Overview tab displays details about the selected Vserver, such as the primary IP address, LIF
information, and the capacity of the volume that the Vserver contains.

134 | OnCommand Console Help


Primary IP

Displays the primary IP address of the Vserver.

LIFs

Displays the number of LIFs that are associated with the Vserver.

Volume Capacity Displays the number of volumes, if any, that the Vserver contains and the
capacity of the volumes that are currently in use.
Graph tab
The Graph tab visually represents the performance of a Vserver. The graphs display the volume
capacity used versus the total capacity in the Vserver, logical interface traffic per second, and volume
capacity used. You can select the graphs from the drop-down list.
You can view the graphs representing a specified time period, such as one day, one week, one month,
three months, or one year. You can also click
trend, used capacity, and total capacity.

to export graph details, such as the used capacity

Namespace Hierarchy tab


The Namespace Hierarchy tab displays the hierarchical view of the Vserver's namespace, including
the junctions and volumes. You can browse through these volumes and browse to the corresponding
Volumes view.
Related Objects pane
The Related Objects section enables you to view and navigate to the groups, clusters, and volumes
related to the Vserver.
Groups

Displays the groups to which the Vserver belongs.

Clusters

Displays the clusters in the selected Vserver.

Volumes

Displays the volumes in the selected Vserver.

Related references

Window layout customization on page 16

Logical storage
Understanding logical storage

Storage | 135

What logical storage objects are


Logical storage includes file system objects, such as volumes, qtrees, LUNs, and Snapshot copies.
Volumes File systems that hold user data that is accessible through one or more of the access
protocols supported by Data ONTAP. The Volumes view displays all the volumes
discovered and monitored by OnCommand console.
Qtrees

Logically defined file system that can exist as a special subdirectory of the root directory
within either a traditional volume or a flexible volume. There is no maximum limit on
the number of qtrees you can create in storage systems. The Qtrees view displays all the
qtrees monitored by OnCommand console.

LUN

Logical unit of storage identified by a number. The LUNs view displays all the LUNs
monitored by the OnCommand console.

About quotas
Quotas provide a way to restrict or track the disk space and number of files used by a user, group, or
qtree. Quotas are applied to a specific volume or qtree.
Why you use quotas
You can use quotas to limit resource usage, to provide notification when resource usage reaches
specific levels, or simply to track resource usage.
You specify a quota for the following reasons:

To limit the amount of disk space or the number of files that can be used by a user or group, or
that can be contained by a qtree
To track the amount of disk space or the number of files used by a user, group, or qtree, without
imposing a limit
To warn users when their disk usage or file usage is high

Overview of the quota process


Quotas can cause Data ONTAP to send a notification (soft quota) or to prevent a write operation
from succeeding (hard quota) when quotas are exceeded.
When Data ONTAP receives a request to write to a volume, it checks to see whether quotas are
activated for that volume. If so, Data ONTAP determines whether any quota for that volume (and, if
the write is to a qtree, for that qtree) would be exceeded by performing the write operation. If any
hard quota would be exceeded, the write operation fails, and a quota notification is sent. If any soft
quota would be exceeded, the write operation succeeds, and a quota notification is sent.

136 | OnCommand Console Help

What Snapshot copies are


Snapshot copies are read-only images of a traditional volume, a FlexVol volume, or an aggregate that
captures the state of the file system at a point in time. Snapshot copies are your first line of defense to
back up and restore data.
Data ONTAP maintains a configurable Snapshot copy schedule that creates and deletes Snapshot
copies automatically for each volume.

Managing logical storage


Editing volume quota settings
You can modify threshold conditions for volumes, Snapshot copies, and user quotas, and set alerts to
receive a notification when a threshold is crossed.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

During this task, the OnCommand console launches the Operations Manager console. Depending on
your browser configuration, you can return to the OnCommand console by using the Alt-Tab key
combination or clicking the OnCommand console browser tab. After the completion of this task, you
can leave the Operations Manager console open, or you can close it to conserve bandwidth.
Steps

1. Click the View menu, then click the Storage tab.


2. In the Storage tab, select the Volumes view.
3. Select the volume you want to modify.
4. Click Edit.
The Edit Volume Quota Settings page is displayed in the Operations Manager console.
5. Modify the capacity threshold settings for the selected volume.
6. Click Update
Changes in the settings are updated in the DataFabric Manager server.
7. Press Alt-Tab or click the OnCommand console browser tab to return to the OnCommand
console.

Storage | 137
Related references

Administrator roles and capabilities on page 506


Editing qtree quota settings
You can modify threshold conditions for qtrees and set alerts to receive notification when a threshold
is crossed.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

During this task, the OnCommand console launches the Operations Manager console. Depending on
your browser configuration, you can return to the OnCommand console by using the Alt-Tab key
combination or clicking the OnCommand console browser tab. After the completion of this task, you
can leave the Operations Manager console open, or you can close it to conserve bandwidth.
Steps

1. Click the View menu, then click the Storage tab.


2. In the Storage tab, select the Qtrees view.
3. Select the qtree you want to modify.
4. Click Edit.
The Edit Qtree Quota Settings page is displayed in the Operations Manager console.
5. Modify the capacity threshold settings for the selected qtree.
6. Click Update.
7. Press Alt-Tab or click the OnCommand console browser tab to return to the OnCommand
console.
Related references

Administrator roles and capabilities on page 506

138 | OnCommand Console Help

Editing LUN path settings


You can edit the following settings for a selected LUN path: owner e-mail address, owner name,
resource tag, and the description of the LUN.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

During this task, the OnCommand console launches the Operations Manager console. Depending on
your browser configuration, you can return to the OnCommand console by using the Alt-Tab key
combination or clicking the OnCommand console browser tab. After the completion of this task, you
can leave the Operations Manager console open, or you can close it to conserve bandwidth.
Steps

1. Click the View menu, then click the Storage tab.


2. In the Storage tab, select the LUNs view.
3. Select the LUN path you want to modify.
4. Click Edit.
The Edit LUN Path Settings page is displayed in the Operations Manager console.
5. Modify the attributes of the selected LUN.
6. Click Update.
Changes to the settings are updated in the DataFabric Manager server.
7. Press Alt-Tab or click the OnCommand console browser tab to return to the OnCommand
console.
Related references

Administrator roles and capabilities on page 506

Storage | 139

Editing quota settings


You can edit the following storage capacity threshold settings for a selected user quota: disk space
used, disk space hard limit, disk space soft limit, disk space threshold, files used, files hard limit, and
files soft limit.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

During this task, the OnCommand console launches the Operations Manager console. Depending on
your browser configuration, you can return to the OnCommand console by using the Alt-Tab key
combination or clicking the OnCommand console browser tab. After the completion of this task, you
can leave the Operations Manager console open, or you can close it to conserve bandwidth.
Steps

1. Click the View menu, then click the Storage tab.


2. In the Storage tab, select the Quota Settings view.
3. Select the user quota you want to modify.
4. Click Edit Settings.
The Edit Quota Settings page is displayed in the Operations Manager console.
5. Modify the properties of the selected quota on the storage system.
6. Click Update.
Changes to the settings are updated in the DataFabric Manager server.
7. Press Alt-Tab or click the OnCommand console browser tab to return to the OnCommand
console.
Related references

Administrator roles and capabilities on page 506


Adding volumes to a group
You can add volumes to an existing group. You can also create a new group and add volumes to it.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.

140 | OnCommand Console Help


Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, select the Volumes view.
3. Select the volumes you want to add to the group and click Add to Group.
4. In the Add to Group dialog box, perform the appropriate action:
If you want to..

Then..

Add the volumes to an existing group

Select the appropriate group.

Add the volumes to a new group

Type the name of the new group in the New Group field.

5. Click OK.
Result

The volumes are added to the group.


Related references

Administrator roles and capabilities on page 506


Adding qtrees to a group
You can add qtrees to an existing group. You can also create a new group and add qtrees to it.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, select the Qtrees view.
3. Select the qtrees you want to add to the group and click Add to Group.
4. In the Add to Group dialog box, perform the appropriate action:
If you want to..

Then..

Add the qtrees to an existing group

Select the appropriate group.

Add the qtrees to a new group

Type the name of the new group in the New Group field.

5. Click OK.

Storage | 141
Result

The qtrees are added to the group.


Related references

Administrator roles and capabilities on page 506


Adding LUNs to a group
You can add LUNs to an existing group. You can also create a new group and add the LUNs to it.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, select the LUNs view.
3. Select the LUNs you want to add to the group and click Add to Group.
4. In the Add to Group dialog box, perform the appropriate action:
If you want to..

Then..

Add the LUNs to an existing group

Select the appropriate group.

Add the LUNs to a new group

Type the name of the new group in the New Group field.

5. Click OK.
Result

The LUNs are added to the group.


Related references

Administrator roles and capabilities on page 506

Monitoring logical storage


Volume capacity thresholds and events
DataFabric Manager server features thresholds to help you monitor the capacity of flexible and
traditional volumes. You can configure alarms to send notification whenever an event related to the
capacity of a volume occurs. You can also take corrective actions based on the event generated. For

142 | OnCommand Console Help


the Volume Full threshold, you can configure an alarm to send notification only when the condition
persists over a specified period.
By default, if you have configured an alarm to alert you to an event, the DataFabric Manager server
issues the alarm only once per event. You can configure the alarm to repeat until it is acknowledged.
Note: If you want to set an alarm for a specific volume, you must create a group with that volume
as the only member.

You can set the following volume capacity thresholds:


Volume Full
Threshold (%)

Description: Specifies the percentage at which a volume is considered full.


Note: To reduce the number of Volume Full Threshold events
generated, you can set the Volume Full Threshold Interval to a non-zero
value. By default, the Volume Full threshold Interval is set to zero.
Volume Full Threshold Interval specifies the time during which the
condition must persist before the event is triggered. Therefore, if the
condition persists for the specified time, DataFabric Manager server
generates a Volume Full event.

If the threshold interval is 0 seconds or a value less than the volume


monitoring interval, DataFabric Manager server generates the
Volume Full events.
If the threshold interval is greater than the volume monitoring
interval, DataFabric Manager server waits for the specified
threshold interval, which includes two or more monitoring intervals,
and generates a Volume Full event only if the condition persisted
throughout the threshold interval.

For example, if the monitoring cycle time is 60 seconds and the


threshold interval is 90 seconds, the threshold event is generated only if
the condition persists for two monitoring intervals.
Default value: 90
Event generated: Volume Full
Event severity: Error
Corrective action
Perform one or more of the following actions:

Ask your users to delete files that are no longer needed, to free disk
space.
For flexible volumes containing enough aggregate space, you can
increase the volume size.
For traditional volumes containing aggregates with limited space, you
can increase the size of the volume by adding one or more disks to the
aggregate.

Storage | 143

Note: Add disks with caution. After you add a disk to an aggregate,

you cannot remove it without destroying the volume and its


aggregate.

For traditional volumes, temporarily reduce the Snapshot copy reserve.


By default, the reserve is 20 percent of the disk space. If the reserve is
not in use, reduce the reserve free disk space, giving you more time to
add a disk. There is no way to prevent Snapshot copies from
consuming disk space greater than the amount reserved for them.
Therefore, it is important to maintain a large enough reserve for
Snapshot copies. By maintaining the reserve for Snapshot copies, the
active file system always has space available to create new files or
modify existing ones. For more information about the Snapshot copy
reserve, see the Data ONTAP Data Protection Online Backup and
Recovery Guide.

Volume Nearly Full Description: Specifies the percentage at which a volume is considered
nearly full.
Threshold (%)
Default value: 80. The value for this threshold must be lower than the
value for the Volume Full Threshold in order for DataFabric Manager
server to generate meaningful events.
Event generated: Volume Almost Full
Event severity: Warning
Corrective action
Perform one or more of the actions mentioned in Volume Full.
Volume Space
Reserve Nearly
Depleted Threshold
(%)

Description: Specifies the percentage at which a volume is considered to


have consumed most of its reserved blocks. This option applies to
volumes with LUNs, Snapshot copies, no free blocks, and a fractional
overwrite reserve of less than 100%. A volume that crosses this threshold
is getting close to having write failures.
Default value: 80
Event generated: Volume Space Reservation Nearly Depleted
Event severity: Warning

Volume Space
Reserve Depleted
Threshold (%)

Description: Specifies the percentage at which a volume is considered to


have consumed all its reserved blocks. This option applies to volumes
with LUNs, Snapshot copies, no free blocks, and a fractional overwrite
reserve of less than 100%. A volume that has crossed this threshold is
getting dangerously close to having write failures.
Default value: 90

144 | OnCommand Console Help

Event generated: Volume Space Reservation Depleted


Event severity: Error
When the status of a volume returns to normal after one of the preceding
events, events with severity 'Normal' are generated. Normal events do not
generate alarms or appear in default event lists, which display events of
warning or worse severity.
Volume Quota
Overcommitted
Threshold (%)

Description: Specifies the percentage at which a volume is considered to


have consumed the whole of the overcommitted space for that volume.
Default value: 100
Event generated: Volume Quota Overcommitted
Event severity: Error
Corrective action
Perform one or more of the following actions:

Volume Quota
Nearly
Overcommitted
Threshold (%)

Create new free blocks by increasing the size of the volume that
generated the event.
Permanently free some of the occupied blocks in the volume by
deleting unnecessary files.

Description: Specifies the percentage at which a volume is considered to


have consumed most of the overcommitted space for that volume.
Default Value: 95
Event generated: Volume Quota Almost Overcommitted
Event Severity: Warning
Corrective action:
Perform one or more of the actions mentioned in Volume Quota
Overcommitted.

Volume Growth
Event Minimum
Change (%)

Description: Specifies the minimum change in volume size (as a


percentage of total volume size) that is acceptable. If the change in volume
size is more than the specified value, and the growth is abnormal in
relation to the volume-growth history, DataFabric Manager server
generates a Volume Growth Abnormal event.
Default value: 1
Event generated: Volume Growth Abnormal

Storage | 145
Volume Snap
Reserve Full
Threshold (%)

Description: Specifies the value (percentage) at which the space that is


reserved for taking volume Snapshot copies is considered full.
Default value: 90
Event generated: Volume Snap Reserve Full
Event severity: Error
Corrective action: None
There is no way to prevent Snapshot copies from consuming disk space
greater than the amount reserved for them. If you disable the volume
Snapshot autodelete option, it is important to maintain a large
enough reserve. Disabling would ensure Snapshot copies that there is
always space available to create new files or modify present ones. For
instructions on how to identify Snapshot copies you can delete, see the
Operations Manager Help.

User Quota Full


Threshold (%)

Description: Specifies the value (percentage) at which a user is considered


to have consumed all the allocated space (disk space or files used) as
specified by the user quota. The user quota includes hard limit in
the /etc/quotas file. If this limit is exceeded, DataFabric Manager
server generates a User Disk Space Quota Full event or a User Files Quota
Full event.
Default value: 90
Event generated: User Quota Full

User Quota Nearly Description: Specifies the value (percentage) at which a user is considered
Full Threshold (%) to have consumed most of the allocated space (disk space or files used) as
specified by the user quota. The user quota includes hard limit in
the /etc/quotas file. If this limit is exceeded, DataFabric Manager
server generates a User Disk Space Quota Almost Full event or a User
Files Quota Almost Full event.
Default value: 80
Event generated: User Quota Almost Full
Volume No First
Snapshot
Threshold (%)

Description: Specifies the value (percentage) at which a volume is


considered to have consumed all the free space for its space reservation.
This is the space that the volume needs when the first Snapshot copy is
created.
This option applies to volumes that contain space-reserved files, no
Snapshot copies, a fraction of Snapshot copies overwrite reserve set to
greater than 0, and where the sum of the space reservations for all LUNs
in the volume is greater than the free space available to the volume.
Default value: 90

146 | OnCommand Console Help

Event generated: Volume No First Snapshot


Volume Nearly No
First Snapshot
Threshold (%)

Description: Specifies the value (percentage) at which a volume is


considered to have consumed most of the free space for its space
reservation. This is the space that the volume needs when the first
Snapshot copy is created.
This option applies to volumes that contain space-reserved files, no
Snapshot copies, a fractional overwrite reserve set to greater than 0, and
where the sum of the space reservations for all LUNs in the volume is
greater than the free space available to the volume.
Default value: 80
Event generated: Volume Almost No first Snapshot

Note: When a traditional volume is created, it is tightly coupled with its containing aggregate so

that its capacity is determined by the capacity of the aggregate. For this reason, you should
synchronize the capacity thresholds of traditional volumes with the thresholds of their containing
aggregates.
Related information

Data ONTAP Data Protection Online Backup and Recovery Guide - now.netapp.com/NOW/
knowledge/docs/ontap/ontap_index.shtml
Qtree capacity thresholds and events
The OnCommand console enables you to monitor qtree capacity and set alarms. You can also take
corrective actions based on the event generated.
DataFabric Manager server features thresholds to help you monitor the capacity of qtrees. Quotas
must be enabled on the storage systems. You can configure alarms to send notification whenever an
event related to the capacity of a qtree occurs.
By default, if you have configured an alarm to alert you to an event, the DataFabric Manager server
issues the alarm only once per event. You can configure the alarm to continue to alert you with
events until it is acknowledged. For the Qtree Full threshold, you can also configure an alarm to send
notification only when the condition persists over a specified period.
Note: If you want to set an alarm for a specific qtree, you must create a group with that qtree as the
only member.

You can set the following qtree capacity thresholds:


Qtree Full
(%)

Description: Specifies the percentage at which a qtree is considered full.


Note: To reduce the number of Qtree Full Threshold events generated, you can
set a Qtree Full Threshold Interval to a non-zero value. By default, the Qtree Full

Storage | 147

threshold Interval is set to zero. The Qtree Full Threshold Interval specifies the
time during which the condition must persist before the event is generated. If the
condition persists for the specified amount of time, DataFabric Manager server
generates a Qtree Full event.

If the threshold interval is 0 seconds or a value less than the volume


monitoring interval, DataFabric Manager server generates Qtree Full events.
If the threshold interval is greater than the volume monitoring interval,
DataFabric Manager server waits for the specified threshold interval, which
includes two or more monitoring intervals, and generates a Qtree Full event
only if the condition persisted throughout the threshold interval.

For example, if the monitoring cycle time is 60 seconds and the threshold
interval is 90 seconds, the threshold event is generated only if the condition
persists for two monitoring intervals.
Default value: 90 percent
Event generated: Qtree Full
Event severity: Error
Corrective action
Perform one or more of the following actions:

Ask users to delete files that are no longer needed, to free disk space.
Increase the hard disk space quota for the qtree.

Qtree Nearly Description: Specifies the percentage at which a qtree is considered nearly full.
Full
Default value: 80 percent
Threshold
Event severity: Warning
(%)
Corrective action
Perform one or more of the following actions:

Ask users to delete files that are no longer needed, to free disk space.
Increase the hard disk space quota for the qtree.

Related information

Data ONTAP Data Protection Online Backup and Recovery Guide - now.netapp.com/NOW/
knowledge/docs/ontap/ontap_index.shtml

148 | OnCommand Console Help

Why Snapshot copies are monitored


The Snapshot copy monitoring and space management help you monitor and generate reports on
Snapshot copies and how they influence your space management strategy.
By using DataFabric Manager server, you can determine the following information about Snapshot
copies:

How much aggregate and volume space is used for Snapshot copies?
Is there adequate space for the first Snapshot copy?
Which Snapshot copies can be deleted?
Which volumes have high Snapshot copy growth rates?
Which volumes have Snapshot copy reserves that are nearing capacity?

Viewing the volume inventory


You can use the Volumes view to monitor your inventory of volumes and view information about
volume properties, related storage objects, capacity graphs, space usage details, and protection
details.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, click Volumes.
Result

The list of volumes are displayed.


Related references

Administrator roles and capabilities on page 506


Viewing the LUN inventory
You can use the LUNs view to monitor your inventory of LUNs and view information about LUN
properties, related storage objects, and capacity graphs.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.

Storage | 149
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, click LUNs.
Result

The list of LUNs are displayed.


Related references

Administrator roles and capabilities on page 506


Viewing the qtree inventory
You can use the Qtrees view to monitor your inventory of qtrees and view information about qtree
properties, related storage objects, and capacity graphs.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, click Qtrees.
Result

The list of qtrees are displayed.


Related references

Administrator roles and capabilities on page 506


Viewing the quota settings inventory
You can use the Quota Settings view to monitor your inventory of quotas and properties of the
quotas.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.

150 | OnCommand Console Help


Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, click Quota Settings.
Result

The list of quotas are displayed.


Related references

Administrator roles and capabilities on page 506


Viewing the Snapshot copies inventory
You can use the Snapshot Copies view to monitor your inventory of Snapshot copies, and view
properties of Snapshot copies and information about related storage objects.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu, then click the Storage option.


2. In the Storage tab, click Snapshot Copies.
Result

The list of Snapshot copies are displayed.


Related references

Administrator roles and capabilities on page 506

Page descriptions
Volumes view
The Volumes view displays detailed information about the volumes in the storage systems that are
monitored, as well as their related objects, and also enables you to perform tasks such as editing the
volume settings, grouping the volumes, and refreshing monitoring samples.

Breadcrumb trail on page 151


Command buttons on page 151
List view on page 152

Storage | 151

Map view on page 153


Overview tab on page 156
Space Breakout tab on page 156
Capacity tab on page 156
Protection tab on page 156
Storage Server tab on page 156
Graph tab on page 156
Related Objects pane on page 157

Breadcrumb trail
The breadcrumb trail is created as you navigate from one list of storage objects to another. As you
navigate, each time you double-click certain items in these lists, another breadcrumb is added to
the trail, providing a string of hyperlinks that captures your navigation history. If you want to
revisit a previously viewed list, you can click the breadcrumb that links to it. You can also click the
icon in the breadcrumb to display a sub-navigation list of objects associated with the hyperlinked
object.
Command buttons
The command buttons enable you to perform the following tasks for a selected volume:
Edit

Launches the Edit Volume Quota Settings in the Operations Manager


console. You can modify the storage capacity threshold settings for a specific
volume from this page.

Delete

Deletes the selected volume

Add to Group

Displays the Add to Group dialog box, which enables you to add the selected
volume to the destination group.

Refresh Monitoring Refreshes the database sample of the selected volume and enables you to
view the updated details.
Samples
More Actions

Refresh

Refreshes the list of volumes.

Grid (

View Events
Displays the events associated with the storage system in the Events tab.
You can sort the information based on the event severity, source ID, date
of event trigger, and state.

Displays the list view of volumes.

Note: You can add a volume to a group, refresh monitoring samples,


modify volume settings, view events for a volume, and delete a volume by
right-clicking the selected volume.

TreeMap (

Displays the map view of volumes.

152 | OnCommand Console Help


List view
The List view displays, in tabular format, the properties of all the discovered volumes. You can
customize your view of the data by clicking the column filters.
You can double-click a volume to display its child objects. The breadcrumb trail is modified to
display the selected volume.
ID

Displays the volume ID. By default, this column is hidden.

Name

Displays the name of the volume.

Aggregate

Displays the name of the aggregate that contains the volume.

Storage Server

Displays the parent of a storage object such as a volume, a storage system


running Data ONTAP 7-Mode,a vFiler unit, or a Vserver.

Type

Displays the type of volume selected.

Block Type

Displays the block format of the volume as a 32-bit or 64-bit.

RAID

Displays the RAID protection scheme. The RAID protection scheme can be on
of the following:
RAID 0

All the raid groups in the volume are of type raid0.

RAID 4

All the raid groups in the volume are of type raid4.

RAID DP

All the raid groups in the volume are of type raid_dp.

Mirrored RAID 0

All the raid groups in the mirrored volume are of type


raid0.

Mirrored RAID 4

All the raid groups in the mirrored volume are of type


raid4.

Mirrored RAID DP All the raid groups in the mirrored volume are of type
raid_dp.
State

Displays the current state of a volume. The state can be Online, Offline,
Initializing, Failed, Restricted, Partial, or Unknown.

Status

Displays the current status of a volume. The status can be Normal, Warning,
Error, Critical, Emergency, or Unknown.

Used Capacity
(GB)

Displays the amount of space used in the volume.

Total Capacity
(GB)

Displays the total space of the volume.

Aggregate ID

Displays the ID of the aggregate. By default, this column is hidden.

Host ID

Displays the ID of the host to which the volume is related. By default, this
column is hidden.

Storage | 153
Map view
The Map view enables you to view the properties of the volumes which are displayed as rectangles
with different sizes and colors. The size and color of the rectangles are based on the options you
select for the Size and Color fields in the properties area.
Volume
Filter

Enables you to display capacity and status information about the volumes in varying
rectangle sizes and colors:
Size

Specifies the size of the rectangle based on the option you select from the
drop-down list. You can select one of the following options:

Used % (default): The percentage of space used in the volume. The size
of the rectangle increases when the value for used space increases.
Available %: The percentage of space available in the volume. The size
of the rectangle increases when the value for available space increases.
Growth Rate: The rate at which data in the volume is growing. The size
of the rectangle increases when the value for growth rate increases.
Near to Max Size: The threshold value specified to generate an alert
before the volume reaches maximum size. The size of the rectangle
increases when the value for near to maximum size increases.
Inode Used %: The percentage of inode space used in the volume. The
size of the rectangle increases when the value for inode % increases.
Days to Max Size: The number of days needed for the volume to reach
maximum size. The size of the rectangle increases as the number of days
to maximum size increases.
Snapshot Used %: The percentage of space used in the Snapshot copy.
The size of the rectangle increases when the value for Snapshot used %
increases.
Saved %: The percentage of space saved in the volume. The size of the
rectangle increases when the value for saved % increases.
Used Capacity: The amount of physical space (in GB) used by
application or user data in the volume. The size of the rectangle
increases when the value for used capacity increases.
Available Capacity: The amount of physical space (in GB) that is
available in the volume. The size of the rectangle increases when the
value for available capacity increases.
Saved Capacity: The amount of space (in GB) saved in the volume. The
size of the rectangle increases when the value for saved capacity
increases.
Available Snapshot Reserve: The amount of Snapshot reserve space (in
GB) available in the volume. The size of the rectangle increases when
the value for available snapshot reserve increases.
Status: The current status of the volume based on the events generated.
The size of the rectangle varies from large to small in the following

154 | OnCommand Console Help

order: Emergency, Critical, Error, Warning, Normal, and Unknown. For


example, a volume with an Emergency status is displayed as a larger
rectangle than a volume with a Critical status.
Color Specifies the color of the rectangle based on the option you select from the
drop-down list. You can select one of the following options:

Status (default): The current status of the volume based on the events
generated. Each status displays a specific color: Emergency (
Critical (

), Warning (

), Normal (

), and Unknown

(
).
Used %: The percentage of space used in the volume. The color
displayed varies based on the following conditions:

), Error (

),

If the used space in the volume is less than the Nearly Full Threshold
value of the volume, the color displayed is green (
). When the
used space reduces, the green color changes to a darker shade.
If the used space in the volume exceeds the Nearly Full Threshold
value but is less than the Full Threshold value of the volume, the
). When the used space reduces, the
color displayed is orange (
orange color changes to a lighter shade.
If the used space in the volume exceeds the Full Threshold value of

). When the used space


the volume, the color displayed is red (
reduces, the red color changes to a lighter shade.
Available %: The percentage of space available in the volume. The color
varies based on the following conditions:

If the available space in the volume exceeds the Full Threshold value
of the volume, the color displayed is green (
). When the available
space reduces, the green color changes to a lighter shade.
If the available space in the volume is less than the Full Threshold
value but exceeds the Nearly Full Threshold value of the volume, the
color displayed is orange (
). When the available space reduces,
the orange color changes to a darker shade.
If the available space in the volume is less than the Nearly Full
Threshold value of the volume, the color displayed is red (
).
When the available space reduces, the red color changes to a darker
shade.

Storage | 155

Saved Capacity: The amount of space (in GB) saved in the storage

controller. The color displayed is blue (


). When the saved capacity
reduces, the blue color changes to a lighter shade.
Available Snapshot Reserve: The amount of Snapshot reserve space (in

GB) available in the volume. The color displayed is blue (


). When
the available Snapshot reserve space reduces, the blue color changes to a
lighter shade.
Inode Used %: The percentage of inode space used in the volume. The
color varies based on the following conditions:

If the inode used space in the volume is less than the nearly full
threshold value, the color displayed is green (
). When the inode
used space reduces, the green color changes to a darker shade.
If the inode used space in the volume exceeds the nearly full
threshold value but is less than the full threshold value, the color
displayed is orange (
). When the inode used space reduces, the
orange color changes to a lighter shade.
If the inode used space in the volume exceeds the full threshold
). When the inode used space
value, the color displayed is red (
reduces, the red color changes to a lighter shade.
Note: The threshold values for inode used % are defined by the
DataFabric Manager server.

Auto Size: If the containing aggregate has sufficient space, the volume
can automatically increase to a maximum size. If auto size is enabled,
(
) is displayed. If auto size is disabled, (
) is displayed.
Snapshot Overflow: The amount of additional space (in GB) used by the
Snapshot copies apart from the allocated Snapshot reserve space. The
color displayed is blue (
). When the Snapshot overflow reduces, the
blue color changes to a lighter shade.

General

Enables you to filter volumes based on name, status, or both.


Note: You can filter by entering regular expressions instead of the full name of the
volume. For example, xyz* lists all the volumes that begin with the name xyz.

Capacity

Enables you to filter storage objects based on the growth rate, used capacity, available
capacity, saved capacity, and so on. You can specify the capacity range by dragging
the sliders.

156 | OnCommand Console Help


Overview tab
The Overview tab displays details about the selected volume, such as storage spaced used, and SRM
path of the volume.
Full Name

Displays the full name of the volume.

Storage Server

Displays the name of the storage server that contains the volume. You
can view more information about the storage server by clicking the link.

Total Capacity

Displays the total amount of space available in the volume to store data.

Quota Committed
Space

Displays the space reserved in the quotas.

SRM Path

Displays the SRM path to which the volume is mapped.

Space Breakout tab


The Space Breakout tab displays the percentage of space used for Snapshot copies and volume data.
Also, indicates the size of the Snapshot reserve as a proportion of total volume size.
Capacity tab
The Capacity tab displays the number of qtrees, LUNs, and Snapshot copies the volume contains,
including the amount of capacity currently in use.
Protection tab
Displays information about the SnapMirror relationships of the volume and whether scheduled
Snapshot copies are enabled.
Storage Server tab
Displays information about the total volume size and amount of space used in the volume.
Note: If multiple volumes belonging to different storage server types are selected, empty labels are
displayed.

Graph tab
The Graph tab visually represents the performance characteristics of a volume. You can select the
graph you want to view from the drop-down list.
You can view the graphs representing a selected time period, such as one day, one week, one month,
three months, or one year. You can also click
to export graph details, such as used capacity trend,
used capacity, total capacity, and space savings achieved through deduplication.

Storage | 157
Related Objects pane
The Related Objects section enables you to view and navigate to the groups, storage controllers,
aggregates, qtrees, LUNs, Snapshot copies, datasets, and datastores related to the volume.
Groups

Displays the groups to which the volumes belong.

Storage Controllers

Displays the storage controllers that contain the selected volume .

Aggregates

Displays the aggregates in the selected volume.

Qtrees

Displays the qtrees in the selected volume.

LUNs

Displays the LUNs in the selected volume.

Snapshot Copies

Displays the Snapshot copies in the volume.

Datasets

Displays the datasets in the volume.

Datastores

Displays the datastores in the volume.

Related references

Window layout customization on page 16


LUNs view
The LUNs view displays detailed information about the LUNs that are monitored, as well as their
related objects, and also enables you to perform tasks such as editing the LUN Path settings,
grouping the LUNs, and refreshing monitoring samples.

Breadcrumb trail on page 157


Command buttons on page 157
List view on page 158
Overview tab on page 158
Graph tab on page 159
Related Objects pane on page 159

Breadcrumb trail
The breadcrumb trail is created as you navigate from one list of storage objects to another. As you
navigate, each time you double-click certain items in these lists, another breadcrumb is added to
the trail, providing a string of hyperlinks that captures your navigation history. If you want to
revisit a previously viewed list, you can click the breadcrumb that links to it. You can also click the
icon in the breadcrumb to display a sub-navigation list of objects associated with the hyperlinked
object.
Command buttons
The command buttons enable you to perform the following tasks for a selected LUN:

158 | OnCommand Console Help


Edit

Launches the Edit LUN Path Settings in the Operations Manager console.
You can modify the attributes for a selected LUN from this page.

Delete

Deletes the selected LUN.

Add to Group

Displays the Add to Group dialog box, which enables you to add the selected
LUN to the destination group.

Refresh
Monitoring
Samples

Refreshes the database sample for the selected LUN and enables you to view
the updated details.

More Actions

Refresh

Refreshes the list of LUNs.

View Events
Displays the events associated with the LUN in the Events tab. You can
sort the information based on the event severity, source ID, date of event
trigger, and state.

Note: You can add a LUN to a group, refresh monitoring samples, modify LUN path settings,
view events for a LUN, and delete a LUN by right-clicking the selected LUN.

List view
The List view displays, in tabular format, the properties of all the discovered LUNs. You can
customize your view of the data by clicking the column filters.
ID

Displays the LUN ID. By default, this column is hidden.

LUN Path

Displays the path to the LUN including the volume and qtree name.

Initiator Group Specifies the initiator group (igroup) to which the LUN is mapped.
Description

Displays the description you provide when creating the LUN on your storage
system.

Size (GB)

Displays size of the LUN.

Storage Server Displays the name of the storage controller or vFiler unit on which the LUN
resides.
Status

Displays the current status of a LUN. The status can be Normal, Warning, Error,
Critical, Emergency, or Unknown.

File System ID

Displays the file system ID that contains the LUN. By default, this column is
hidden.

Overview tab
The Overview tab displays details about the selected LUN such as SRM path, size, connected ports,
and host details.

Storage | 159
Full Path

Displays the complete path to the LUN.

Size

Displays the size of the LUN.

Contained File
System

Displays the name of the file system (volume or qtree) on which this LUN
resides.

Mapped To

Displays the following information about the LUN that is exported:

The initiator group (igroup) to which the LUN is mapped.


The total number of LUNs exported from the storage system.

This is represented as igroup_name(n), where n is the total number of


LUNs exported.
Serial Number

Displays the serial number of the LUN.

SRM Path

Displays the SRM path to which the LUN is mapped. You can modify the
SRM path by clicking the link.

SAN Host

Displays the monitored host in a SAN that initiates requests to the storage
systems to perform tasks.

Space Reservation
Enabled

Displays whether space reservation is enabled.

HBA Port

Specifies the HBA ports that SAN hosts use to connect to each other in a
SAN environment.

Graph tab
The Graph tab visually represents the performance characteristics of a LUN. You can select the
graph you want to view from the drop-down list.
You can view the graphs representing a selected time period, such as one day, one week, one month,
three months, or one year. You can also click
second and LUN bytes written per second.

to export graph details, such as LUN bytes read per

Related Objects pane


The Related Objects section enables you to view and navigate to the groups and storage objects
related to the LUN.
Groups

Displays the groups to which the selected LUN belongs.

Volumes

Displays the volumes in the selected LUN.

Qtrees

Displays the qtrees in the selected LUN.

Datastores

Displays the datastores in the selected LUN.

160 | OnCommand Console Help


Related references

Window layout customization on page 16


Qtrees view
The Qtrees view displays detailed information about the qtrees that are monitored, as well as their
related objects, and also enables you to perform tasks, such as editing the qtree settings, grouping the
qtrees, and refreshing monitoring samples.

Breadcrumb trail on page 160


Command buttons on page 160
List view on page 161
Overview tab on page 161
Graph tab on page 162
Related Objects pane on page 162

Breadcrumb trail
The breadcrumb trail is created as you navigate from one list of storage objects to another. As you
navigate, each time you double-click certain items in these lists, another breadcrumb is added to
the trail, providing a string of hyperlinks that captures your navigation history. If you want to
revisit a previously viewed list, you can click the breadcrumb that links to it. You can also click the
icon in the breadcrumb to display a sub-navigation list of objects associated with the hyperlinked
object.
Command buttons
The command buttons enable you to perform the following tasks for a selected qtree:
Edit

Launches the Edit Qtree Settings in the Operations Manager console. You
can modify the capacity threshold settings for a selected qtree from this page.

Delete

Deletes the selected qtree.

Add to Group

Displays the Add to Group dialog box, which enables you to add the selected
qtree to the destination group.

Refresh
Monitoring
Samples

Refreshes the database samples of the qtree and enables you to view the
updated details.

More Actions

Refresh

Refreshes the list of qtrees.

View Events
Displays the events associated with the qtree in the Events tab. You can
sort the information based on the event severity, source ID, date of event
trigger, and state.

Storage | 161
Note: You can add a qtree to a group, refresh monitoring samples, modify qtree settings, view
events for a qtree, and delete a qtree by right-clicking the selected qtree.

List view
The List view displays, in tabular format, the properties of all the discovered qtrees. You can
customize your view of the data by clicking the column filters.
You can double-click a qtree to display its child objects. The breadcrumb trail is modified to display
the selected qtree.
ID

Displays the qtree ID. By default, this column is hidden.

Qtree Name

Displays the name of the qtree.

Storage Server

Displays the name of the storage controller or vFiler unit containing the qtree.

Volume

Displays the name of the volume containing the qtree.

Status

Displays the current status of a qtree. The status can be Normal, Warning,
Error, Critical, Emergency, or Unknown.

Quota Limit (MB) Displays the quota limit, if any.


Volume ID

Displays the ID of the volume which contains the qtree. By default, this
column is hidden.

Aggregate ID

Displays the ID of the aggregate that contains the volume which in turn
contains the qtree. By default, this column is hidden.

Storage Path Type Displays the direct or indirect storage path type of the qtree. By default, this
column is hidden.
Overview tab
The Overview tab displays details about the selected qtree, such as storage capacity, SnapMirror
relationships and the SRM path.
Full Path

Display the complete path of the qtree.

SnapMirror

Indicates whether the qtree is a source or destination in a SnapMirror


relationship.

Days to Full

Displays the estimated amount of time left before the storage space is
full.

Scheduled Snapshot
Copies

Displays information about the SnapMirror relationships of the qtree and


whether scheduled Snapshot copies are enabled.

SnapVault

Displays whether or not the qtree is backed up. If the qtree is backed up,
the SnapVault destination is displayed.

162 | OnCommand Console Help


Last SnapVaulted

Displays the date and time when the qtree was last backed up using
SnapVault.

SRM Path

Displays the SRM path to which the qtree is mapped.

Quota Limit

Displays the quota limit, if any.

Used Capacity

Displays the total capacity allocated and used by the qtree.

Daily Growth Rate

Displays the change in the disk space (number of bytes) used in the qtree
if the amount of change between the last two samples continues for 24
hours.

Growth Rate (%)

Displays the change in the amount of used space in the qtree reserve.

Graph tab
The Graph tab visually represents about the performance of a qtree. You can select the graphs you
want to view from the drop-down list.
You can view the graphs representing a selected time period, such as one day, one week, one month,
three months, or one year. You can also click
and used capacity.

to export graph details, such as used capacity trend

Related Objects pane


The Related Objects section enables you to view and navigate to the groups, volumes, LUNs,
datasets, and datastores related to the qtree.
Groups

Displays the groups to which the qtree belongs.

Volumes

Displays the volumes that contains the selected qtree.

LUNs

Displays the LUNs in the selected qtree.

Datasets

Displays the datasets in the selected qtree.

Datastores

Displays the datastores in the selected qtree.

Related references

Window layout customization on page 16


Quota Settings view
The Quota Settings view enables you to view detailed information about the user and user group
quota settings, and also enables you to perform tasks such as editing the quota settings and refreshing
the list of quotas.

Breadcrumb trail on page 163


Command buttons on page 163

Storage | 163

List view on page 163

Breadcrumb trail
The breadcrumb trail is created as you navigate from one list of storage objects to another. As you
navigate, each time you double-click certain items in these lists, another breadcrumb is added to
the trail, providing a string of hyperlinks that captures your navigation history. If you want to
revisit a previously viewed list, you can click the breadcrumb that links to it. You can also click the
icon in the breadcrumb to display a sub-navigation list of objects associated with the hyperlinked
object.
Command buttons
The command buttons enable you to perform the following tasks for a selected quota:
Edit

Launches the Edit Quota Settings page in the Operations Manager console. You can edit
the selected quota from this page.

Refresh Refreshes the list of user and user group quotas.


Note: You can also modify the settings for a quota by right-clicking the selected quota.

List view
The List view displays, in tabular format, the properties of all the discovered user and user group
quotas. You can customize your view of the data by clicking the column filters.
ID

Displays the ID of the quota. By default, this column is hidden.

User Name

Displays the name of the user or user group.

File System

Displays the name and path of the volume or qtree on which the user
quota resides.

Type

Displays the type of quota. The quota can be either a user quota or a
group quota.

Status

Displays the current status of the quotas. The status can be Normal,
Warning, Error, Critical, Emergency, or Unknown.

Disk Space Used (MB)

Displays the total amount of disk space used.

Disk Space Threshold


(MB)

Displays the disk space threshold as specified in the /etc/quotas file


of the storage system.

Disk Space Soft Limit


(MB)

Displays the soft limit on disk space as specified in the /etc/quotas


file of the storage system.

Disk Space Hard Limit


(MB)

Displays the hard limit on disk space as specified in the /etc/quotas


file of the storage system.

164 | OnCommand Console Help


Disk Space Used (%)

Displays the percentage of disk space used.

Files Used

Displays the total number of files used.

Files Soft Limit

Displays the soft limit on files as specified in the /etc/quotas file of


the storage system.

Files Hard Limit

Displays the hard limit on files as specified in the /etc/quotas file of


the storage system.

Files Used (%)

Displays the percentage of files used.

Related references

Window layout customization on page 16


Snapshot Copies view
The Snapshot Copies view enables you to view detailed information about Snapshot copies, related
objects, and storage servers to which the Snapshot copies are associated.

Command button on page 164


List view on page 164
Related objects pane on page 165

Command button
The command button enables you to perform the following task for a selected Snapshot copy:
Refresh

Refreshes the list of Snapshot copies.

List view
The List view displays, in tabular format, the properties of all the discovered Snapshot copies. You
can customize your view of the data by clicking the column filters.
ID

Displays the Snapshot copy ID. By default, this column is hidden.

Name

Displays the name of the Snapshot copy.

Volume

Displays the name of the volume that contains the Snapshot copy.

Aggregate

Displays the name of the aggregate that contains Snapshot copy.

Storage Server Displays the name of the storage controller, vFiler unit, or Vserver that contains
the Snapshot copy.
Access Time

Displays the time when the Snapshot copy was last accessed.

Dependency

Displays the names of the applications that are accessing the Snapshot copy (for
example, SnapMirror), if any.

Storage | 165
Related objects pane
The Related objects section enables you to view and navigate to the aggregates and volumes related
to the Snapshot copies.
Aggregates

Displays the aggregates that contains the Snapshot copy.

Volumes

Displays the volumes that contains the Snapshot copy.

Related references

Window layout customization on page 16

166 | OnCommand Console Help

167

Policies
Local policies
Understanding local policies
Local policy and backup of a dataset's virtual objects
A dataset's local policy in the OnCommand console enables you to specify the start times, stop times,
frequency, retention time, and warning and error event thresholds for local backups of its VMware or
Hyper-V virtual objects.
What local protection of virtual objects is
Local protection of a dataset's virtual objects consists of the OnCommand console making Snapshot
copies of the VMware virtual objects or Hyper-V virtual objects that reside as images on your
storage systems and saving those Snapshot copies as backup copies locally on the same storage
systems.
In case of data loss or corruption due to user or software error, you can restore the lost or damaged
virtual object data from saved local Snapshot copies as long as the primary storage systems on which
the virtual objects reside remain intact and operating.
What a local policy is
A local policy is a configured combination of Snapshot copy schedules, retention times, warning
threshold, and error threshold levels that you can assign to a dataset. After you assign a local policy
to a dataset, that policy applies to all the virtual objects that are included in that dataset.
The OnCommand console allows you to configure multiple local policies with different settings from
which you can select one to assign to a dataset.
You can also use policies supplied by the OnCommand console.
Local protection and remote protection of virtual objects
After Snapshot copies of a dataset's virtual objects are generated as local backup, remote protection
operations that are specified in an assigned storage service configuration can save these backup
copies to secondary storage.
Secondary or tertiary protection of virtual objects cannot be accomplished unless Snapshot copies
have been generated on the primary node by backup jobs carried out on demand or through local
policy.

168 | OnCommand Console Help

Local policies supplied by the OnCommand console


To simplify the task of providing local protection to virtual objects in a dataset the OnCommand
console provides a set of local policies (preconfigured combinations of local backup schedules, local
backup retention settings, lag warning and lag error thresholds, and optional backup scripts)
specifically to support local backup of certain types of data.
The preconfigured local policies include the following set:
This default policy enforces the following VMware environment-optimized settings
VMware
local backup related to local backup scheduling and retention. This policy can also be renamed
and modified, to accommodate different circumstances.
policy
template
Hourly backups without VMware snapshot (crash consistent) every hour
between 7 am and 7 pm every day including weekends
Daily backups with VMware snapshot at 10 PM every night every day including
weekends
Weekly backup with VMware snapshot every Sunday midnight
Retention settings: Hourly backups for 2 days, Daily backups for 1 week,
Weekly backups for 2 weeks
Issue a warning if there are no backups for: 1.5 days
Issue an error if there are no backups for: 2 days
Backup script path: empty
This default policy enforces the following Hyper-V environment-optimized settings
Hyper-V
local backup related to local backup scheduling and retention. This policy can also be renamed
and modified, to accommodate different circumstances.
policy
template
Hourly backups every hour between 7 am and 7 pm every day including
weekends
Daily backups at 10 PM every night every day including weekends
Weekly backup every Sunday midnight
Retention settings: Hourly backups for 2 days, Daily backups for 1 week,
Weekly backups for 2 weeks
Issue a warning if there are no backups for: 1.5 days
Issue an error if there are no backups for: 2 days
Skip backups that will cause virtual machines to go offline
Start remote backup after local backup
Backup script path: empty

Policies | 169

Guidelines for configuring a local policy


Before you use the Create Local Policy dialog box or Edit Local Policy dialog box to create or edit a
local policy, you must make some decisions about the policy to input in your command selection and
in the configuration page.
Virtual object type
To open the Create Local Policy dialog box you specify with the Create command whether the local
policy you want to create is for a dataset of VMware objects or for a dataset of Hyper-V objects.
Name
Name

Your company might have a naming convention for policies. When specifying a name
for a new policy, make sure you follow that convention.

Description Use a description that helps someone unfamiliar with the policy to understand its
purpose.
Schedule and Retention
Schedule

You can set up multiple schedules of multiple types (Hourly, Daily, Weekly) for your
local backups. Each schedule has a start time, a stop time and a frequency with which
its backups are executed.
If you intend to implement local backups on multiple datasets of Hyper-V objects that
are associated with the same Hyper-V server, you must configure separate local
policies with non-overlapping schedules to assign separately to each separate dataset.

Retention

You can specify a retention period to be associated with each type of backup schedule
(Hourly, Daily, or Weekly). A retention period specifies the minimum length of time
that a backup copy is maintained, before it is eligible to be purged. A retention period
assigned to one type of backup schedule applies to all backup copies of that type.

Backup
Options

Depending on the type of virtual objects the dataset contains, you can enable
additional operations to be performed on those objects during backup.

Create VMware Snapshots


Creates VMware quiesced snapshots before taking the storage system Snapshot
copies during local backup operations (displayed for datasets of VMware objects).
Include independent disks
Includes independent VMDKs in local backups associated with this schedule for
this dataset (displayed for datasets of VMware objects only).
Allow saved state backups
Allows the local backup of a dataset of virtual machines to proceed even if some
of the virtual machines in that dataset are in saved state or shut down. Virtual
machines in saved state or that are shut down, receive saved state or offline

170 | OnCommand Console Help

backup. Performing a saved state or offline backup can cause downtime (displayed
for datasets of Hyper-V objects).
If this option is not selected, encountering a virtual machine that is in saved state
or that is shutdown fails the dataset backup.
Start remote backup after local backup
Starts a remote backup of data to secondary storage after the local backup is
finished if a storage service that specifies a remote backup is also assigned to the
dataset.

Backup Settings
Issue a warning if
there is no backup
for:

Decide how long the OnCommand console waits before issuing a warning
event if no local backup has successfully finished during that time.

Issue an error if
there is no backup
for:

Decide how long the OnCommand console waits before issuing an error
event if no local backup has successfully finished during that time.

Backup script path: You can specify a path to a backup script (located on the system on which
the host service runs) to specify additional operations to be executed with
the local backup. If you use a PowerShell script, you should use the drive
letter convention. For other types of scripts, you can use either the drive
letter convention or the Universal Naming Convention.

Configuring local policies


Adding a local policy
You can create local policies to schedule local backup jobs and designate retention periods for the
local backup copies for datasets of virtual objects.
Before you begin

You must have reviewed the Guidelines for configuring a local policy on page 169.
You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

The Create Local Policy dialog box enables you to create a new local policy, or add a preconfigured
local policy supplied by the OnCommand console.

Policies | 171
Steps

1. Click the View menu and click the Policies option.


2. Click Create in the Policies tab and select the option for the appropriate local policy type.
3. In the Create Local Policy dialog box, select the Name option and enter the requested
information in the associated content area.
4. Select the Schedule and Retention option and make your selections in the associated content
area.
5. Select the Backup Settings option and make your selections in the associated content area.
You can also change this information for this policy at a later time.
6. After you enter your desired amount of information about this policy, click OK.
Result

The OnCommand console creates your new policy and lists it in the Policies tab.
Related concepts

Guidelines for configuring a local policy on page 169


Related references

Administrator roles and capabilities on page 506


Name and Description area on page 177
Schedule and Retention area on page 177
Backup Settings area on page 179
Editing a local policy
You can edit an existing local policy if you want to modify the name or description, schedule and
retention times, or the backup options assigned to the policy.
Before you begin

You must have reviewed the Guidelines for configuring a local policy on page 169.
You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

The Edit Local Policy dialog box enables you to modify an existing local policy.

172 | OnCommand Console Help


Steps

1. Click the View menu and click the Policies option.


2. From the Policies tab, select the row in the local policies list containing the policy you want to
edit, and then click Edit Policy.
3. In the Edit Local Policy dialog box click the selection button on one or more of the following
options to display and modify their related settings:

Name
Schedule and Retention
Backup Settings

4. After you have changed the desired information, click OK.


Result

The OnCommand console updates your policy and lists it in the Policies tab.
Related concepts

Guidelines for configuring a local policy on page 169


Related references

Administrator roles and capabilities on page 506


Name and Description area on page 177
Schedule and Retention area on page 177
Backup Settings area on page 179

Managing local policies


Editing a dataset of virtual objects to configure local policy and local backup
You can select, modify, or configure new local policies to automate local protection of datasets
containing virtual VMware objects or virtual Hyper-V objects.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu and click the Datasets option.

Policies | 173
2. In the Datasets tab, select the dataset on which you want to schedule and configure local backups
and click Edit.
3. In the Edit Dataset dialog box, select the Local Policy option and drop down list and complete
one of the following actions:

If you want to assign an existing local policy, select that policy from the Local Policy drop
down list.
If you want to assign an existing local policy with some modifications select that policy, make
your modifications in the content area, and click Save.
If you want to configure a new local policy to apply to this dataset, select the Create New
option, configure the policy in the content area, and click Create.

4. After you finish assigning a new or existing local policy to this dataset, if you want to test
whether your dataset's new configuration is in conformance with OnCommand console
requirements before you apply them, click Test Conformance to display the Dataset
Conformance Report.

If the Dataset Conformance Report displays no warning or error information, click Close and
continue.
If the Dataset Conformance Report displays warning or error information, read the Action and
Suggestion information resolve the conformance issue, then click Close and continue.

5. Click OK.
Any local policy assignment, modification, or creation that you completed will be applied to the
local protection of the virtual objects in the selected dataset.
Related references

Administrator roles and capabilities on page 506


Local Policy area on page 266
Administrator roles and capabilities on page 506
Administrator roles and capabilities on page 506
Copying a local policy
You can make a new local policy by copying an existing local policy and modifying it to meet your
particular requirements.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

Making multiple copies of a local policy, then configuring the copies with non-overlapping
schedules, and then assigning each copy to a different dataset is a good way to implement the local

174 | OnCommand Console Help


backup of multiple datasets of Hyper-V objects because a Hyper-V host does not allow simultaneous
or overlapping local backups on virtual machines associated with it.
Steps

1. Click the View menu and click the Policies option.


2. From the Policies tab, select the row in the local policies list containing the policy that you want
to copy and then click Copy.
The OnCommand console lists a copy of the policy (labeled as "Copy of.." ) in the Policies tab.
3. Select the copied policy and click Edit.
4. Make any required name changes and schedule modifications on the Edit Local Policy dialog
box and click OK.
The OnCommand console updates the copied policy with your changes and lists it in the Policies
tab.
Related references

Administrator roles and capabilities on page 506


Deleting a local policy
You can delete local policies that are not currently assigned to datasets of virtual objects.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu and click the Policies option.


2. From the Policies tab, select the row in the local policies list containing the policy you want to
delete.
3. Click the Dependencies tab to determine whether this local policy is currently assigned to one or
more datasets.
4. If one or more assigned datasets are listed on the Dependencies tab, either edit those datasets to
unassign this local policy or do not delete this local policy.
5. If no assigned datasets are listed on the Dependencies tab, click Delete, and click OK to confirm.
The OnCommand console removes this local policy de-lists it in the Policies tab.

Policies | 175
Related references

Administrator roles and capabilities on page 506

Page descriptions
Policies tab
The Policies tab enables you to add, edit, copy, delete, and list local protection policies. You can
assign the listed local policies to your virtual object datasets to configure local protection of their
virtual object members.

Command buttons on page 175


Local policies list on page 175
Overview tab on page 176
Details area on page 176
Dependencies tab on page 176

Command buttons
The command buttons enable you perform the following tasks related to local policies:
Create

Enables you to create a new local policy.

If you select the VMware Policy sub-option, starts the Create Local Policy dialog box
for adding a local protection policy specific to VMware objects.
If you select the Hyper-V Policy sub-option, starts the Create Local Policy dialog box
for adding a local protection policy specific to Hyper-V objects.

Edit

Enables you to edit the selected local policy.

Copy

Enables you to copy the selected local policy.

Delete

Enables you to delete the selected local policy if that policy is not currently attached to a
dataset.

Refresh Updates the information that is displayed for all local policies listed on the Policies tab.
Local policies list
Lists information about existing local application policies. Select a row in the list to display
information in the Details area.
Name

The name of the policy.

Type

The type of local policy: Hyper-V and VMware.

Description

Briefly describes the local policy

176 | OnCommand Console Help


Overview tab
Schedules Displays the scheduled local backup jobs by schedule type (Hourly, Daily, Weekly, or
Monthly) and time.
Retention Displays the period of time that the backup Snapshot copies associated with each
backup schedule type are retained in storage before becoming subject to automatic
purging.
Details area
The area below the Local policies list displays information about the selected local policy.
Issue a warning if there Displays the period of time after which the OnCommand console issues
a warning event if no local backup has successfully finished during that
are no backups for
time.
Issue an error if there
are no backups for

Displays the period of time after which the OnCommand console issues
an error event if no local backup has successfully finished during that
time.

Backup Script Path

Displays a path to an optional backup script (located on the system on


which the host service is installed) that can specify additional operations
to be executed in association with the local backup.

Dependencies tab
Displays information about datasets that are assigned this local policy.
Name

Displays the names of datasets assigned this local policy.

Protection Status

Displays whether or not the dataset is protected by a protection policy.

Related references

Window layout customization on page 16


Create Local Policy dialog box and Edit Local Policy dialog box
The Create Local Policy dialog box and the Edit Local Policy dialog box each enable you to
configure a new or existing local protection policy to apply to virtual objects in your datasets.
Options
The options enable you to perform the following types of configuration on a local policy:
Name

Enables you to edit the policy name and description.

Schedule and
Retention

Enables you to create, edit, and delete backup and retention schedules
associated with this local policy.

Policies | 177
Backup Settings

Enables you to specify no-backup warning and error thresholds, and a path
to an optional backup script

Policy Summary section


Displays the current local policy name, description, backup script path (if any), no-backup warning
and error times, and a list of the local backup schedules and their associated retention times.
Command buttons
Command buttons enable you to perform the following tasks for a local policy:
OK

Saves the latest changes that you have made to the data in the Create Local Policy dialog
box or Edit Local Policy dialog box as the latest configuration for this policy.

Cancel Cancels any changes you have made to the settings in the Create Local Policy dialog box
or Edit Local Policy dialog box since the last time you opened it.
Name and Description area
The Name and Description area of the Create Local Policy dialog box or Edit Local Policy dialog
box enables you to specify a name and description for a local policy.
Name

Enables you to change the name of the current policy.


Unless you change it, the OnCommand console displays a default policy name that it
has assigned.

Description Enables you to enter or modify a short description of the current policy.
Schedule and Retention area
The Schedule and Retention area of the Create Local Policy dialog box or Edit Local Policy dialog
box enables you to configure a schedule of local backup jobs that you can apply to members of a
virtual object dataset and also specify how long to retain the resulting Snapshot copies before their
deletion.
Schedule and
Retention

Either specifies the local backup schedule and retention settings assigned to the
current local policy, or enables you to create a new backup and retention schedule
for the current local policy.

Add

Enables you to add a schedule to be applied to the current local policy backup.

Delete

Enables you to delete the selected local policy backup schedule

Schedule Type Displays the type of backup schedule (Hourly, Daily, Weekly, or Monthly).
Start Time

Enables you to select the time of day the local backup starts.

178 | OnCommand Console Help


End Time

Enables you to select the time of day an hourly local backup ends (Applies to
Hourly type schedules only).

Recurrence

Enables you to select the frequency with which local policy backups occur for the
associated schedule. Recurrence settings vary by schedule type:

Retention

Hourly
You can specify recurrence by hours and minutes.
Daily
Recurrence is fixed at once a day.
Weekly
You can specify recurrence by days of the week.
Monthly
You can specify recurrence by days of the month.

Enables you to select the period of time that local backup copies generated by a
schedule remain on the storage system before becoming subject to purging.
You can use any valid number and either Minutes, Hours, Days, or Weeks to set
the backup retention time.
All schedules of one type use the same retention setting. For example, changing
the retention setting for one Hourly schedule configured for this policy changes
the retention setting for all the Hourly schedules configured for this policy.

Backup
Options

Enables you to view and select additional options to be implemented with your
local backups.

Create VMware Snapshots


Creates VMware quiesced snapshots before taking the storage system
Snapshot copies during local backup operations (displayed for datasets of
VMware objects).
Include independent disks
Includes independent VMDKs in local backups associated with this schedule
for this dataset (displayed for datasets of VMware objects only).
Allow saved state backups
Allows the backup of a dataset of virtual machines to proceed even if some of
the virtual machines in that dataset are in saved state or shut down. Virtual
machines in saved state or shut down, receive saved state or offline backup.
Performing a saved state or offline backup can cause downtime. (displayed for
datasets of Hyper-V objects only.)
If this option is not selected, encountering a virtual machine that is in saved
state or shutdown fails the dataset backup.
Start remote backup after local backup

Policies | 179

If remote backup of data to secondary storage is specified by a storage service


assigned to this dataset, whether to start that remote backup immediately after
the local backup is finished.
If this option is not selected, any remote backup to secondary storage starts
according the schedules configured for the protection policy specified for the
storage service.
Backup Settings area
The Backup Settings area of the Create Local Policy dialog box or the Edit Local Policy dialog box
enables you to specify lag threshold warning and error levels and an optional path to a backup script
if you need a script to specify additional operations to be carried out in association with backups in
this local policy.
Backup Settings

Either specifies the current lag warning threshold, lag error threshold, and
optional backup script path, or enables you to set the lag warning threshold,
lag error threshold, and optional backup script path for the current local
policy backup.

Issue a warning if
there are no
backups for

Specifies a period of time after which the OnCommand console issues a


warning event if no local backup has successfully finished.

Issue an error if
there are no
backups for

Specifies a period of time after which the OnCommand console issues an


error event if no local backup has successfully finished.

Backup Script Path Displays the existing backup script path, or, optionally, enables you to enter
a path to a backup script (located on the system upon which the host service
is installed) to specify additional operations to be executed with the backup.

180 | OnCommand Console Help

181

Datasets
Understanding datasets
What a dataset is
A dataset is a set of physical or virtual data containers that you can configure as a unit for the
purpose of group protection or group provisioning operations.

You can use the OnCommand console to configure datasets that contain virtual VMware objects
or virtual Hyper-V objects.
You can also use the OnCommand console and the associated NetApp Management Console to
configure datasets that contain physical storage systems with aggregates, volumes, qtrees, and
LUNs

During dataset configuration you can additionally configure or assign local protection or remote
protection arrangements and schedules that apply to all objects in that dataset. You can also start ondemand protection operations for all objects in that dataset with one command.

Dataset concepts
Associating data protection, disaster recovery, a provisioning policy, or storage service with a dataset
enables storage administrators automate tasks, such as assigning consistent policies to primary data,
propagating policy changes, and provisioning new volumes, qtrees, or LUNS on primary and
secondary dataset nodes.
Configuring a dataset combines the following objects:
Dataset of
physical
storage objects

For protection purposes, a collection of physical resources on a primary node,


such as volumes, flexible volumes, and qtrees, and the physical resources for
copies of backed-up data on secondary and tertiary nodes.
For provisioning purposes, a collection of physical resources, such as volumes,
flexible volumes, qtrees, and LUNs, that are assigned to a dataset node. If the
protection policy establishes a primary and one or more nonprimary nodes, each
node of the dataset is a collection of physical resources that might or might not
be provisioned from the same resource pool.

Dataset of
virtual objects

A collection of supported VMware virtual objects or Hyper-V virtual objects that


reside on storage systems. These virtual objects can also be backed up locally
and backed up or mirrored on secondary and tertiary nodes.

Resource pool

A collection of physical resources from which storage is provisioned. Resource


pools can be used to group storage systems and aggregates by attributes, such as
performance, cost, physical location, or availability. Resource pools can be

182 | OnCommand Console Help


assigned directly to the primary, secondary, or tertiary nodes of datasets of
physical storage objects. They can be assigned indirectly both to datasets of
virtual objects and to datasets of physical storage objects through a storage
service.
Data protection Defines how to protect primary data on primary, secondary or tertiary storage, as
well as when to create copies of data and how many copies to keep. Protection
policy
policies can be assigned directly to datasets of physical storage objects. They can
be assigned indirectly to both datasets of virtual objects and to datasets of
physical storage objects through a storage service.
Provisioning
policy

Defines how to provision storage for the primary or secondary dataset nodes, and
provides rules for monitoring and managing storage space and for allocating
storage space from available resource pools. Provisioning policies can be
assigned directly to the primary, secondary, or tertiary nodes of datasets of
physical storage objects. They can be assigned indirectly to both datasets of
virtual objects and datasets of physical storage objects through a storage service.

Storage service A single dataset configuration package that consists of a protection policy,
provisioning policies, resource pools, and an optional vFiler template (for vFiler
unit creation). You can assign a single uniform storage service to datasets with
common configuration requirements as an alternative to separately assigning the
same protection policy, provisioning policies, resource pools, and setting up
similar vFiler unit attachments to each of them.
The only way to configure a dataset of virtual objects with secondary or tertiary
backup and mirror protection and provisioning is by assignment of a storage
service. You cannot configure secondary storage vFiler attachments for datasets
of virtual objects.
Local policy

A policy that schedules local backup jobs and designates retention periods for the
local backup copies for datasets of virtual objects.

Related objects Are Snapshot copies, primary volumes, secondary volumes, or secondary qtrees
that are generated as a result of local policy or storage service protection jobs or
provisioning jobs.
The OnCommand console lists related objects for each dataset on the Datasets
tab.
Naming
settings

Are character strings and naming formats that are applied when naming related
objects that are generated as a result of local policy or storage service protection
jobs or provisioning jobs.

Datasets | 183

Role of provisioning policies in dataset management


Provisioning policies specify the resource pool location and the configuration requirements of the
storage systems and their container objects that can be used to supply the primary, secondary, and
tertiary storage needs of the storage objects or virtual objects in a dataset.
Direct or indirect assignment of a provisioning policy
Provisioning policies can be assigned to datasets directly or indirectly, through storage services.

You can assign provisioning policies directly to each node in a dataset that is configured to
include and manage physical storage objects as members.
You can also assign provisioning policies to storage services, which are preconfigured
combinations of protection policies, provisioning policies, and resource pools.
You can then assign storage services directly both to datasets configured for physical storage
objects and datasets configured for virtual objects.

Where you can create, modify, and assign provisioning policies


You can create and modify provisioning policies, and assign them to storage dataset nodes or storage
services using the associated program, NetApp Management Console. Information on any of those
tasks is in the NetApp Management Console help.

Role of protection policies in dataset management


A protection policy specifies a dataset's primary node, secondary node, and tertiary node data
protection topology, backup or mirror schedules, backup copy retention times, backup bandwidth
consumption, and other aspects related to backing up physical storage objects and virtual objects in a
dataset.
Direct or indirect assignment of a protection policy
Protection policies can be assigned to datasets directly or indirectly, through storage services.

You can assign protection policies directly to datasets that are configured to include and manage
physical storage objects as members.
You can also assign protection policies to storage services, which are preconfigured combinations
of protection policies, provisioning policies, and resource pools.
You can then assign storage services directly both to datasets configured for physical storage
objects and datasets configured for virtual objects.

Where you can create, modify, and assign protection policies


You can create and modify protection policies, and assign them to storage datasets or storage
services using the associated program, NetApp Management Console. Information on any of those
tasks is in the NetApp Management Console help.

184 | OnCommand Console Help

What conformance monitoring and correction is


The OnCommand console actively monitors a dataset's resources during remote protection
configuration and at hourly intervals thereafter to ensure that the resource configuration remains in
conformance with the protection policies and provisioning policies that are included in the storage
service assigned to it.
When the OnCommand console encounters a condition that is not in conformance with the
provisioning and protection policies, it first attempts to correct that condition automatically;
however, if the condition requires user consent to correct or if the condition must be corrected
manually, then the OnCommand console assigns nonconformant status to the dataset. If the dataset is
tagged with this status, operations associated with that dataset's assigned protection and provisioning
policies have a high likelihood of partial or full failure until the nonconformant condition is resolved.
If the nonconformant condition is encountered during initial storage service assignment or later on,
that condition is flagged for the administrator in the Datasets tab, from which the administrator can
manually display the Conformance Details dialog box.

Datasets of physical storage objects


The OnCommand console and the associated application, NetApp Management Console enable you
to group physical storage systems and the container objects (the aggregates, volumes, qtrees, or
LUNs) that reside on them into datasets for purposes of data protection.
You can set up and enhance automated protection and provisioning for a dataset of physical storage
objects by applying the following dataset configuration tasks:

assigning a protection policy


assigning a disaster recovery capable protection policy
assigning provisioning policies to the dataset's secondary and tertiary nodes
assigning a resource pool to the dataset's secondary and tertiary nodes
assigning a storage service (a preconfigured combination of a protection policy, provisioning
policies, and resource pools)
enabling an online migration capability

Objects that a dataset of physical storage objects can include


Datasets of physical storage objects can include containers and the physical storage systems on
which they are located.
You can include the following types of objects as members of a dataset of physical storage objects:

qtrees
volumes
aggregates
hosts
vFiler units

Datasets | 185

Open Systems SnapVault directories


Open Systems SnapVault hosts

Effect of time zone on scheduled protection operations in datasets of physical objects


The actual execution time of the protection jobs that are scheduled in the local policy and storage
service that you assign to a dataset of physical objects depends on the time zone that is specified for
the dataset, the time zones that are specified for the assigned resource pools, or (in the absence of
those settings) on the time that is set on the DataFabric Manager server.
Administrators can assign optional dataset or resource pool time zone settings to datasets of physical
objects in the NetApp Management Console interface. For additional information on the effect of
time zone settings on datasets of physical objects, see the NetApp Management Console help.
Effect of no time zone settings assigned to datasets of physical objects or to
resource pools
If a dataset of physical objects or its assigned physical resource pool elements are not assigned time
zone settings, then by default the NetApp Management Console interface executes the protection
schedule for its dataset members in accordance with the clock and time zone setting on the
DataFabric Manager server.
For example, without dataset-level time zone settings configured for them, datasets in Los Angeles
and London with a daily backup scheduled for 9 p.m. (Eastern Standard Time) from a NetApp
Management Console and a DataFabric Manager server in New York will, by default, execute
simultaneously at 6 p.m. (Pacific Standard Time) on a primary data node in Los Angeles, or at 2 a.m.
(GMT) at a primary data node in London.
Effect of time zone settings assigned to datasets
If those Los Angeles and London datasets are assigned Pacific Standard Time and GMT time zone
settings, respectively, then the NetApp Management Console adjusts the schedule to execute nonsimultaneous daily backups at 9 p.m. (Pacific Standard Time) in Los Angeles and at 9 p.m. (GMT) in
London.
Effect of time zone settings assigned to resource pools
If the secondary dataset nodes in the Los Angeles and London datasets are assigned resource pools
and those resource pools are assigned time zone settings, then any scheduled protection jobs from
secondary to tertiary storage execute according to the assigned schedule and the time zone settings
for those resource pools.

Datasets of virtual objects


The OnCommand console enables you to group VMware virtual objects or Hyper-V virtual objects
that reside as data on your storage systems into datasets for purposes of data protection.
You can set up and enhance automated protection and provisioning for a dataset of virtual objects by
configuring the following types of protection:

186 | OnCommand Console Help

You can assign a local policy to configure local backup job scheduling and local backup copy
retention of your virtual object data.
You can assign a storage service (a preconfigured combination of a protection policy,
provisioning policies, and resource pools) to configure secondary storage and tertiary storage
backup and mirroring of your virtual object data.

VMware objects that a dataset can include


Datasets that you configure to include and protect VMware objects are constrained by VMwarespecific restrictions.

A dataset designed for VMware virtual objects can include datacenter, virtual machine, and
datastore objects.
VMware datacenter objects that you include in a dataset cannot be empty.
They must contain virtual machine objects or datastore objects that also contain virtual machine
objects for successful backup.
VMware and Hyper-V objects can not coexist in one dataset.
VMware objects and storage system container objects (such as aggregates, volumes, and qtrees)
cannot coexist in one dataset.
If you add a datastore object to a dataset, all the virtual machine objects that are contained in that
datastore are protected by the dataset's assigned local backup policy or storage service.
If a virtual machine resides on more than one datastore, you can exclude one or more of those
datastores from the dataset.
No local or remote protection is configured for the excluded datastores.
You might want to exclude datastores that contain swap files that you want to exclude from
backup.
If a virtual machine is added to a dataset, all of its VMDKs are protected by default unless one of
the VMDKs is on a datastore that is in the "exclusion list" of that dataset.
VMDKs on a datastore object in a dataset must be contained within folders in that datastore. If
VMDKs exist outside of folders on the datastore, and that data is backed up, restoring the backup
could fail.

Hyper-V objects that a dataset can include


Datasets that you configure to include and protect Hyper-V objects are constrained by Hyper-V
specific restrictions.

A dataset with Hyper-V objects can include only virtual machines.


VMware and Hyper-V objects cannot co-exist in one dataset.
Hyper-V objects and storage system container objects (such as aggregates, volumes, and qtrees)
cannot coexist in one dataset.
Shared virtual machines and dedicated virtual machines cannot coexist in one dataset.

Datasets | 187

Best practices when adding or editing a dataset of virtual objects


When you create or edit a dataset of virtual objects, observing best practices specific to datasets of
virtual objects helps you to avoid some performance and space usage problems after configuration.
General best practices for datasets of virtual objects
The following configuration practices apply to both datasets of Hyper-V objects and datasets of
VMware objects:

To avoid conformance and local backup issues caused by primary volumes reaching their
Snapshot copy maximum of 255, best practice is to limit the number of virtual objects included in
a primary volume, and limit the number of datasets to which each primary volume is directly or
indirectly included as a member.
A primary volume that hosts virtual objects that are included in multiple datasets is subject to
retaining an additional Snapshot copy of itself for every local backup on any dataset that any of
its virtual object children are members of.
To avoid backup schedule inconsistencies, best practice is to include only virtual objects that are
located in the same time zone in one dataset.
The schedules for the local protection jobs and remote protection jobs specified in the local
policies and storage services that are assigned a dataset of virtual objects are carried out
according to the time in effect on the host systems that are associated with the dataset's virtual
objects.

Best practices specific to datasets of Hyper-V objects


The following configuration practices apply specifically to datasets containing Hyper-V objects:

To ensure faster dataset backup of virtual machines in a Hyper-V cluster, best practice is to run
all the virtual machines on one node of the Hyper-V cluster.
When virtual machines run on different Hyper-V cluster nodes, separate backup operations are
required for each node in the cluster. If all virtual machines run on the same node, only one
backup operation is required, resulting in a faster backup.

Best practices specific to datasets of VMware objects


The following configuration practices apply specifically to datasets containing VMware objects:

If a virtual machine resides on more than one datastore, you can exclude one or more of those
datastores from the dataset. No local or remote protection is configured for the excluded
datastores.
You might want to exclude datastores that contain swap files that you want to exclude from
backup.
To avoid an excessive amount of secondary space provisioned for backup, best practice when
creating volumes to host the VMware datastores whose virtual machines will be protected by the
OnCommand console backup is to size those volumes to be not much larger than the datastores
they host.
The reason for this practice is that when provisioning secondary storage space to back up virtual
machines that are members of datastores, the OnCommand console allocates secondary space that

188 | OnCommand Console Help


is equal to the total space of the volume or volumes in which those datastores are located. If the
host volumes are much larger than the datastores they hold, an excessive amount of provisioned
secondary space can result.
Storage services and remote backup of a dataset's virtual objects
You can specify remote backup of a dataset's VMware or Hyper-V virtual objects by assigning that
dataset a storage service. A storage service specifies the type of remote protection (backup or mirror)
to apply and the type of storage systems on which to store the secondary or tertiary data.
What remote protection is
Remote protection of a dataset's virtual objects consists of the OnCommand console taking the most
recent Snapshot copies of the VMware virtual objects or Hyper-V virtual objects that reside as
images on your storage systems and saving those Snapshot copies to remote storage systems at
secondary or tertiary locations.
In case of data loss or corruption at the primary site, or even damage or disabled storage systems at
the primary site, you can restore the lost or damaged virtual object data from Snapshot copies saved
to a remote location.
Remote protection and local protection
The Snapshot copies that are saved to secondary storage are those that have been generated by local
on demand backup jobs or by backup jobs scheduled through the dataset's assigned local policy.
What storage services are
Storage services are preconfigured combinations of protection policies, provisioning policies, and
resource pools that can be assigned as a package to a dataset.
Storage services are configured in NetApp Management Console, but the OnCommand console
allows you to view the properties of existing storage services and assign one of them to the current
dataset. When you are creating a dataset, you can select a storage service, review the storage service's
data protection topology (its primary, secondary, and if applicable, tertiary storage), review the
storage service's protection schedule, specify supplementary backup scripts, and assign that storage
service to the current dataset.
The OnCommand console supplies a set of storage services from which you can select that are
specifically configured for protection of virtual objects in a small to mid-sized data storage
management operation.
To support data management and protection of virtual objects at an enterprise level, you might need
to use the NetApp Management Console to configure custom storage services, and custom protection
and provisioning policies components.
Details on creating and customizing storage services are available in the NetApp Management
Console help.

Datasets | 189
Storage services enabled for disaster recovery support cannot be assigned to datasets of virtual
objects.
Preconfigured storage services supplied by the OnCommand console
To simplify the task of providing provisioning storage for virtual objects and remote protection to
virtual objects in a dataset, the OnCommand console provides a set of storage services, preconfigured
combinations of protection policies, provisioning policies, and resource pools specifically designed
to facilitate remote protection of virtual objects.
The listed storage services are optimal for use in storage facilities with five or fewer storage systems
in single resource pools. The preconfigured storage services are assigned with a Mirror protection
policy or none. You can copy, clone, or modify these storage services with other protection policies
using the NetApp Management Console. The preconfigured storage services include the following
provisioning and protection policies:
Thin Provisioned
Space for VMFS
Datastores with
Mirror

This storage service is configured to provide thin provisioning and mirror


protection for VMFS datastore objects.

Thin Provisioned
Space for NFS
Datastores with
Mirror

Provisioning policy: Thin Provisioned Space for VMFS Datastores


This policy enables deduplication and thin provisioning. It facilitates
maximized space savings. It is required for 50% guarantee program.
Protection policy: Mirror
Resource pools: Default storage service primary pool; Default storage
service mirror pool

This storage service is configured to provide thin provisioning and mirror


protection for NFS datastore objects.

Provisioning policy: Thin Provisioned Space for NFS Datastores


This policy enables deduplication and thin provisioning. It facilitates
maximized space savings. It is required for 50% guarantee program.
Protection policy: Mirror
Resource pools: Default storage service primary pool; Default storage
service mirror pool

This storage service is configured to provide thin provisioning and mirror


Thin Provisioned
Space for Hyper-V protection for backing storage LUNs in Hyper-V environments.
Storage with Mirror
Provisioning policy: Thin Provisioned Space for Hyper-V Storage
This policy is required for 50% guarantee program.
Protection policy: Mirror
Resource pools: Default storage service primary pool; Default storage
service mirror pool

190 | OnCommand Console Help


Thin Provisioned
Space for Hyper-V
Delegated Storage
with Mirror

This storage service is configured to provide thin provisioning and mirror


protection in Hyper-V environments. It supports a delegation model where
the volume is provisioned according to best practices. The LUNs will be
provisioned by SnapDrive for Windows on the Hyper-V parent.

Provisioning policy: Thin Provisioned Space for Hyper-V Delegated


Storage
This policy is required for 50% guarantee program.
Protection policy: Mirror
Resource pools: Default storage service primary pool; Default storage
service mirror pool

Reserved Data
Space for VMFS
Datastores with
Mirror

This storage service is configured to provide provisioning and mirror


protection for write guaranteed VMFS datastores.

Reserved Data
Space for NFS
Datastores with
Mirror

This storage service is configured to provide thick provisioning and mirror


protection for NFS datastores.

Provisioning policy: Reserved Data Space for VMFS Datastores


Protection policy: Mirror
Resource pools: Default storage service primary pool; Default storage
service mirror pool

Provisioning policy: Reserved Data Space for NFS Datastores


Protection policy: Mirror
Resource pools: Default storage service primary pool; Default storage
service mirror pool

This storage service is configured to provide thick provisioning and mirror


Reserved Data
Space for Hyper-V protection for backing LUNs in Hyper-V environments. The LUNs could be
Storage with Mirror backing store for Cluster Shared Volumes.

Reserved Data
Space for Hyper-V
Delegated Storage
with Mirror

Provisioning policy: Reserved Data Space for Hyper-V Storage


Protection policy: Mirror
Resource pools: Default storage service primary pool; Default storage
service mirror pool

This storage service is configured to provide thick provisioning and mirror


protection for backing storage for Hyper-V environments. It supports a
delegation model where the volume is provisioned according to best
practices. The LUNs are be provisioned using SnapDrive for Windows on a
Hyper-V parent.

Provisioning policy: Reserved Data Space for Hyper-V Delegated


Storage
Protection policy: Mirror

Datasets | 191

Resource pools: Default storage service primary pool; Default storage


service mirror pool

Thin Provisioned
Space for VMFS
Datastores

This storage service is configured to provide thin provisioning for VMFS


datastore objects.

Thin Provisioned
Space for NFS
Datastores

This storage service is configured to provide thin provisioning for NFS


datastore objects.

Thin Provisioned
Space for Hyper-V

This storage service is configured to provide thin provisioning for backing


storage LUNs in Hyper-V environments.

Thin Provisioned
Space for Hyper-V
Delegated Storage

Provisioning policy:Thin Provisioned Space for VMFS Datastores


Protection policy: None
Resource pools: Default storage service primary pool

Provisioning policy:Thin Provisioned Space for NFS Datastores


Protection policy: None
Resource pools: Default storage service primary pool

Provisioning policy: Thin Provisioned Space for Hyper-V Storage


Protection policy: Storage
Resource pools: Default storage service primary pool

This storage service is configured to provide thin provisioning in Hyper-V


environments. It supports a delegation model where the volume is
provisioned according to best practices. The LUNs will be provisioned by
SnapDrive for Windows on the Hyper-V parent.

Provisioning policy:Thin Provisioned Space for Hyper-V Delegated


Storage
Protection policy: None
Resource pools: Default storage service primary pool

Reserved Data
Space for VMFS
Datastores

This storage service is configured to provide provisioning for write


guaranteed VMFS datastores.

Reserved Data
Space for NFS
Datastores

This storage service is configured to provide thick provisioning for NFS


datastores.

Provisioning policy: Reserved Data Space for VMFS Datastores


Protection policy: None
Resource pools: Default storage service primary pool

Provisioning policy: Reserved Data Space for NFS Datastores


Protection policy: None

192 | OnCommand Console Help

Reserved Data
Space for Hyper-V
Storage

This storage service is configured to provide thick provisioning for backing


LUNs in Hyper-V environments. The LUNs could be backing store for
Cluster Shared Volumes.

Reserved Data
Space for Hyper-V
Delegated Storage

Resource pools: Default storage service primary pool

Provisioning policy: Reserved Data Space for Hyper-V Storage


Protection policy: None
Resource pools: Default storage service primary pool

This storage service is configured to provide thick provisioning for backing


storage for Hyper-V environments. It supports a delegation model where
the volume is provisioned according to best practices. The LUNs are be
provisioned using SnapDrive for Windows on a Hyper-V parent.

Provisioning policy: Reserved Data Space for Hyper-V Delegated


Storage
Protection policy: None
Resource pools: Default storage service primary pool

Local policy and backup of a dataset's virtual objects


A dataset's local policy in the OnCommand console enables you to specify the start times, stop times,
frequency, retention time, and warning and error event thresholds for local backups of its VMware or
Hyper-V virtual objects.
What local protection of virtual objects is
Local protection of a dataset's virtual objects consists of the OnCommand console making Snapshot
copies of the VMware virtual objects or Hyper-V virtual objects that reside as images on your
storage systems and saving those Snapshot copies as backup copies locally on the same storage
systems.
In case of data loss or corruption due to user or software error, you can restore the lost or damaged
virtual object data from saved local Snapshot copies as long as the primary storage systems on which
the virtual objects reside remain intact and operating.
What a local policy is
A local policy is a configured combination of Snapshot copy schedules, retention times, warning
threshold, and error threshold levels that you can assign to a dataset. After you assign a local policy
to a dataset, that policy applies to all the virtual objects that are included in that dataset.
The OnCommand console allows you to configure multiple local policies with different settings from
which you can select one to assign to a dataset.
You can also use policies supplied by the OnCommand console.

Datasets | 193
Local protection and remote protection of virtual objects
After Snapshot copies of a dataset's virtual objects are generated as local backup, remote protection
operations that are specified in an assigned storage service configuration can save these backup
copies to secondary storage.
Secondary or tertiary protection of virtual objects cannot be accomplished unless Snapshot copies
have been generated on the primary node by backup jobs carried out on demand or through local
policy.
Local policies supplied by the OnCommand console
To simplify the task of providing local protection to virtual objects in a dataset the OnCommand
console provides a set of local policies (preconfigured combinations of local backup schedules, local
backup retention settings, lag warning and lag error thresholds, and optional backup scripts)
specifically to support local backup of certain types of data.
The preconfigured local policies include the following set:
This default policy enforces the following VMware environment-optimized settings
VMware
local backup related to local backup scheduling and retention. This policy can also be renamed
and modified, to accommodate different circumstances.
policy
template
Hourly backups without VMware snapshot (crash consistent) every hour
between 7 am and 7 pm every day including weekends
Daily backups with VMware snapshot at 10 PM every night every day including
weekends
Weekly backup with VMware snapshot every Sunday midnight
Retention settings: Hourly backups for 2 days, Daily backups for 1 week,
Weekly backups for 2 weeks
Issue a warning if there are no backups for: 1.5 days
Issue an error if there are no backups for: 2 days
Backup script path: empty
This default policy enforces the following Hyper-V environment-optimized settings
Hyper-V
local backup related to local backup scheduling and retention. This policy can also be renamed
and modified, to accommodate different circumstances.
policy
template
Hourly backups every hour between 7 am and 7 pm every day including
weekends
Daily backups at 10 PM every night every day including weekends
Weekly backup every Sunday midnight
Retention settings: Hourly backups for 2 days, Daily backups for 1 week,
Weekly backups for 2 weeks
Issue a warning if there are no backups for: 1.5 days
Issue an error if there are no backups for: 2 days
Skip backups that will cause virtual machines to go offline

194 | OnCommand Console Help

Start remote backup after local backup


Backup script path: empty

Effect of time zone on scheduled protection operations in datasets of virtual objects


The actual execution time of the protection jobs that are scheduled in the local policy and storage
service that you assign to a dataset of virtual objects depends on the times that are set on the systems
that run the host services associated with that dataset's virtual object members.
Best practice to accommodate the time zone effect
The schedules for the local protection jobs and remote protection jobs specified in the local policies
and storage services that are assigned a dataset of virtual objects are carried out according to the time
in effect on the systems that run the host services that are associated with the dataset's virtual objects.
Therefore, when you create datasets of virtual objects, the best practice is to include as members in
any one dataset only those virtual objects that are associated with host services that are located in the
same time zone.
However, if you cannot avoid including virtual object members whose host services are located in
different time zones, you should be aware of how time zone differences can affect protection job
times on the different virtual object members.
Local backup execution in a dataset containing virtual objects from different time
zones
If a dataset contains a mix of virtual object members that are associated with different host services
in different time zones, a local backup job, scheduled at a specified time by that dataset's local
policies, executes for each individual virtual object member according to the time that is set on its
associated host system. The DataFabric Manager server records the two actions using the time
settings of the most lagging time zone.
For example, if virtual machines associated with a host service in California and virtual machines
associated with a host service in New York are added to the same dataset, and that dataset's local
policy schedule specifies a local backup at 9 a.m., the virtual machines in New York are locally
backed up at 9 a.m. Eastern time (or 6 a.m. Pacific time), and the virtual machines in California are
locally backed up at 9 a.m. Pacific time. The DataFabric Manager server records two backup
versions: one at 6 a.m. Pacific time, containing backup for the New York virtual machines only, and
one at 9 a.m. Pacific time, containing backup for Sunnyvale virtual machines only.
Remote backup execution in a dataset containing virtual objects from different time
zones
If a dataset contains a mix of virtual object members that are associated with different host services
in different time zones, a remote protection job, scheduled at a specified time by that dataset's storage

Datasets | 195
service settings, executes for every virtual object member at a single common time that is determined
by the host systems in the most lagging time zone.
For example, if virtual machines associated with a host service in California and virtual machines
associated with a host service in New York are added to the same dataset, and the protection policy
schedule in that dataset's storage service specifies a remote backup at 9 a.m., both the virtual
machines in California and the virtual machines in New York are backed up at 9 a.m. Pacific time (or
12 noon Eastern time).

Data ONTAP licenses used for protecting or provisioning data


There are several Data ONTAP licensed options that you can use to protect or provision your data.
After you have purchased the software licenses you need, you can assign these licenses to your
primary and secondary storage from the Storage Systems Hosts window.
When you purchase a Data ONTAP option license, you receive a code composed of a string of
characters, such as ABCDEFG, that is unique to a particular service. You receive license codes for
every protocol and option, or service, that you purchase.
Not all purchased license codes are installed on a storage system before it is shipped from the
factory. Some licenses are installed after the system is set up. You can purchase license codes to
enable additional services at any time. If you misplace a license code, you can contact NetApp
technical support or log in to the NetApp Support Site to obtain a copy.
You must enter a software license code on a storage system to enable the corresponding service. You
do not need to indicate which license the code enables. The code is matched automatically to the
appropriate service license.
Note: The Licenses area is visible only when the selected host is a single storage system running

Data ONTAP. If you plan to use Open Systems SnapVault to back up data on a host that is not
running Data ONTAP, you select the secondary storage system to license the necessary Data
ONTAP services.
The following licenses are available for use with DataFabric Manager server:
SnapMirror license You install a SnapMirror license on each of the source and destination
storage systems for the mirrored data. If the source and destination volumes
are on the same system, only one license is required.
SnapMirror replicates data to one or more networked storage systems.
SnapMirror updates the mirrored data to keep it current and available for
disaster recovery, offloading tape backup, read-only data distribution, testing
on nonproduction systems, online data migration, and so on. You can also
enable the SnapMirror license to use Qtree SnapMirror for backup.
To use SnapMirror software, you must update the
snapmirror.access option in Data ONTAP to specify the destination

systems that are allowed to access the primary data source system. For more
information about the snapmirror.access option, see the Data ONTAP
Data Protection Online Backup and Recovery Guide.

196 | OnCommand Console Help


SnapVault Data
ONTAP secondary
license

You install the SnapVault Secondary license on storage systems that host the
backups of protected data. SnapVault creates backups of data stored on
multiple primary storage systems and copies the backups to a secondary
storage system. If data loss or corruption occurs, backed-up data can be
restored to a primary or open storage system with little of the downtime and
uncertainty associated with conventional tape backup and restore operations.
For versions of Data ONTAP 7.3 or later, a single storage system can
contain a SnapVault Data ONTAP Primary license and a SnapVault
Secondary license.

SnapVault Data
ONTAP primary
license

You install the SnapVault Data ONTAP Primary license on storage systems
running Data ONTAP that contain host data to be backed up. For versions of
Data ONTAP 7.3 or later, a single storage system can contain a SnapVault
Data ONTAP Primary license and a SnapVault Secondary license.

SnapVault
Windows Primary
License

You install the SnapVault Windows Primary license on a secondary storage


system, in addition to the SnapVault Secondary license, to support a
Windows-based primary storage system running the Open Systems
SnapVault agent. A Windows-based primary storage system running the
Open Systems SnapVault agent does not require a SnapVault license.

You install the SnapVault Open File Manager license on a secondary storage
SnapVault
Windows Open File system to enable the backup of open files on Windows primary storage
systems running the Open Systems SnapVault agent.
Manager license
You must install the SnapVault Windows Primary license and the
SnapVault Data ONTAP Secondary license on the secondary storage system
before installing the SnapVault Open File Manager license.
SnapVault UNIX
primary license

You install the SnapVault UNIX Primary license on a secondary storage


system, in addition to the SnapVault Secondary license, to support a UNIXbased primary storage system (AIX, HP-UX, or Solaris) running the Open
Systems SnapVault agent. A UNIX-based primary storage system running
the Open Systems SnapVault agent does not require a SnapVault license.

SnapVault Linux
primary license

You install the SnapVault Linux Primary license on a secondary storage


system, in addition to the SnapVault Secondary license, to support a Linuxbased primary storage system running the Open Systems SnapVault agent. A
Linux-based primary storage system running the Open Systems SnapVault
agent does not require a SnapVault license.

NearStore Option
license

The NearStore license enables your storage system to use transfer resources
as conservatively as if it were optimized as a backup system. This approach
is useful when the storage system on which you want to store backed-up data
is not a system optimized for storing backups, and you want to minimize the
number of transfer resources the storage system requires.
Storage systems using the NearStore license must meet the following
criteria:

Datasets | 197

The storage system must be a FAS30xx, FAS31xx, or FAS60xx system.


The version of Data ONTAP must be 7.1 or later.
If you plan to use the SnapVault service, the storage system must have a
SnapVault secondary license enabled.

Deduplication
license

The deduplication license (a_sis) enables you to consolidate blocks of


duplicate data into single blocks to store more information using less storage
space.

SnapMirror Sync
license

The SnapMirror Sync license enables you to replicate data to the destination
as soon as it is written to the source volume. SnapMirror Sync is a feature of
SnapMirror.

MultiStore Option
license

The MultiStore Option license enables you to partition the storage and
network resources of a single storage system so that it appears as multiple
storage systems on the network. Each virtual "storage system" created as a
result of the partitioning is called a vFiler unit. A vFiler unit, using the
resources assigned, delivers file services to its clients as a storage system
does.
The storage resource assigned to a vFiler unit can be one or more qtrees or
volumes. The storage system on which you create vFiler units is called the
hosting storage system. The storage and network resources used by the
vFiler units exist on the hosting storage system.
Be sure the host on which you intend to install the MultiStore Option license
is running Data ONTAP 6.5 or later.

FlexClone license

The FlexClone license is necessary on storage systems that you intend to use
as resources for secondary nodes for datasets of virtual objects.

Single file restore


license

The Single file restore license is necessary on storage systems that you
intend to use as primary storage for datasets of virtual objects.

Related information

Data ONTAP Data Protection Tape Backup and Recovery Guide - http://now.netapp.com/NOW/
knowledge/docs/ontap/ontap_index.shtml

Descriptions of dataset protection status


A dataset is remotely protected only if the secondary storage system specified in the protection
relationship is successfully backing up data, and if the copies of the data can be restored. You can
monitor dataset status using the Datasets tab.
You should regularly monitor a dataset's protection, because the OnCommand console cannot
sufficiently protect the dataset under the following conditions:

198 | OnCommand Console Help

If a secondary storage system runs out of storage space necessary to meet the retention duration
required by the protection policy
If the lag thresholds specified by the policy are exceeded

The following list describes protection status values and their descriptions:
Baseline Failed

No initial baseline data transfers have registered a backup version.

Initializing

The dataset is conforming to the protection policy and the initial baseline data
transfer is in process.

Job Failure

The most recent protection job did not succeed.


This status might result for any of the following reasons:

Lag Error

A backup from a SnapVault or Qtree SnapMirror relationship failed or could


not be registered.
A mirror copy from a SnapMirror relationship failed or could not be
registered.
Local backups (Snapshot copies) failed on the primary node.

The dataset has reached or exceeded the lag error threshold specified in the
assigned protection policy. This value indicates that there has been no successful
backup or mirror copy of a node's data within a specified period of time.
This status might result for any of the following reasons:

Lag Warning

The most recent local backup (Snapshot copy) on the primary node is older
than the threshold setting permits.
The most recent backup (SnapVault or Qtree SnapMirror) is older than the
lag threshold setting or no backup jobs have completed since the dataset was
created.
The most recent mirror (SnapMirror) copy is older than the lag threshold
setting or no mirror jobs have completed since the dataset was created.

The dataset has reached or exceeded the lag warning threshold specified in the
assigned protection policy. This value indicates that there has been no successful
backup or mirror copy of a node's data within a specified period of time.
This status might result for any of the following reasons:

The most recent local backup (Snapshot copy) on the primary node is older
than the threshold setting permits.
The most recent backup (SnapVault or Qtree SnapMirror) is older than the
lag threshold setting or no backup jobs have completed since the dataset was
created.
The most recent mirror (SnapMirror) copy is older than the lag threshold
setting or no mirror jobs have completed since the dataset was created.

Datasets | 199
No Protection
Policy

The dataset is managed by the OnCommand console but no protection policy has
been assigned to the dataset.

Protected

The dataset has an assigned policy and it has conformed to that policy at least
once.

Protection
Suspended

An administrator has requested that scheduled backups be temporarily halted


until the administrator requests that they be resumed.

Uninitialized

This status might result for any of the following reasons:

The dataset has a protection policy that does not have any protection
operations scheduled.
The dataset does not contain any data to be protected.
The dataset does not contain storage for one or more destination nodes.
The single node dataset does not have any backup versions.
An application dataset requires at least one backup version associated with it.
The dataset does not contain any backup or mirror relationships.

No Local Policy A dataset of virtual objects has no local policy assigned.


Attention: Importing an external relationship into a dataset temporarily changes the dataset's

protection status to Unitialized. When the next scheduled backup or mirror backup job runs or
when you run an on-demand backup the protection status changes to reflect the results of the
protection job.

Configuring datasets
Adding a dataset of physical storage objects
You can use the Add Dataset wizard to add a dataset to manage protection for physical storage
objects sharing the same protection requirements, or to manage provisioning for the dataset members.
Before you begin

You must already be familiar with the Decisions to make before adding datasets of physical
storage objects (for protection) on page 200.
You must have NetApp Management Console installed.
You must have gathered the protection information that you need to complete this task:

Dataset properties
Dataset naming properties
Group membership

You must have gathered the provisioning information that you need to complete this task:

200 | OnCommand Console Help

Provisioning policy for primary node


Migration capability
vFiler unit assignment
Provisioning policy for nonprimary dataset nodes
You must be authorized to perform all the steps of this task; your RBAC administrator can
confirm your authorization in advance.

About this task

During this task, the OnCommand console launches NetApp Management Console. Depending on
your browser configuration, you can return to the OnCommand console by using the Alt-Tab key
combination or clicking the OnCommand console browser tab. After the completion of this task, you
can leave the NetApp Management Console open, or you can close it to conserve bandwidth.
Steps

1. Click the View menu and select the Datasets option.


2. In the Datasets tab, click Create and select the Dataset with storage objects option to start the
NetApp Management Console Add Dataset wizard.
3. Complete the steps in the Add Dataset wizard to create a dataset of physical storage objects.
After you complete the wizard, the new dataset is listed in the Datasets tab.
4. To provide data protection or disaster recovery protection for the new dataset, select that dataset
and click Protection Policy to start the Dataset Policy Change wizard.
5. When finished, press Alt-Tab or click the OnCommand console browser tab to return to the
OnCommand console.
Related references

Administrator roles and capabilities on page 506


Decisions to make before adding datasets of physical storage objects (for protection)
Before you use the Add Dataset wizard to create a new dataset of physical storage objects and the
Dataset Policy Change wizard to configure protection, you must decide how you want to protect the
data.
Dataset
properties

Is there a dataset naming convention that you can use to help administrators
easily locate and identify datasets?
Dataset names can include the following characters but cannot be only
numeric:
a to z
A to Z

Datasets | 201

0 to 9
. (period)
_ (underscore)
- (hyphen)
space

Dataset
naming
properties

Group
membership

Resources for
primary
storage

If you use any other characters when naming the dataset, they do not appear in
the name.
What is a good description of the dataset membership?
Use a description that helps someone unfamiliar with the dataset to understand
its purpose.
Who is the owner of the dataset?
If an event on the dataset triggers an alarm, who should be contacted?
You can specify one or more individual e-mail addresses or a distribution list
of people to be contacted.
What time zone do you want to use to schedule operations on the dataset be
scheduled according to the local time zone for the data?
You can specify a time zone in the wizard or use the default time zone, which
is the system time zone used by the DataFabric Manager server.
Do you want to use the actual dataset name or a custom label in your datasetlevel Snapshot copy, primary volume, secondary volume, or secondary qtree
naming?
For customizing the naming settings of object types, do you want the current
default naming format to apply to one or more object types that are generated
in this dataset?
If you want to customize the dataset-level naming formats for one or more
object types, in what order do you want to enter the naming attributes for
Snapshot copy, primary volume, secondary volume, or secondary qtree?
Do you need to create a collection of datasets and resource pools based on
common characteristics, such as location, project, or owning organization?
Is there an existing group to which you want to add this dataset?

Will you assign a resource pool or individual physical resources as destinations


for your primary storage?
If using a resource pool:

For the primary node in the dataset, which resource pool meets its provisioning
requirements?
If no resource pool meets the requirements of the primary node, you can create
a new resource pool for each node at the Resource Pools window.

202 | OnCommand Console Help

Verify that you have the appropriate software licenses on the storage you
intend to use.

If using individual resources:

Protection
policy

After you create a new dataset of physical objects, you protect it by running the
NetApp Management Console Dataset Policy Change wizard to assign a
protection policy.

Resources for
secondary or
tertiary
storage

If you prefer not to use resource pools for automatic provisioning, you can
select individual physical resources as members of your dataset.
Verify that you have the appropriate software licenses on the storage you
intend to use.

Which protection policy meets the requirements of the dataset?


Review the policies listed on the Protection Policies window to see if any are
suitable for your new dataset.
If no protection policy meets the requirements of your new dataset, is there a
protection policy that would be suitable with minor modifications?
If so, you can copy that protection policy to create a new policy you can
modify as needed for the new dataset. If not, you can run the Add Protection
Policy wizard to create a new policy for the dataset.

When you assign a protection policy, will you assign a resource pool or individual
physical resources as destinations for your backups and mirror copies?
You do not have to assign a resource pool or physical resources to a node to create
a new dataset. However, the dataset will be nonconformant with its policy until
resources are assigned to each node, because the NetApp Management Console
data protection capability cannot carry out the protection specified by the policy.
If using a resource pool:

For the secondary or third nodes in the dataset, which resource pool meets
their provisioning requirements?
For example, the resource pool you assign to a mirror node should contain
physical resources that would all be acceptable destinations for mirror copies
created of the dataset members.
If no resource pool meets the requirements of a node, you can create a new
resource pool for each node at the Resource Pools window.
Verify that you have the appropriate software licenses on the storage you
intend to use.

If using individual resources:

If you prefer not to use resource pools for automatic provisioning, you can
select individual physical resources as destinations for backups and mirror
copies of your dataset.

Datasets | 203

Verify that you have the appropriate software licenses on the storage you
intend to use.

Adding a dataset of virtual objects


The OnCommand console enables you to create datasets to manage the protection and provisioning
of virtual objects in VMware or Hyper-V environments.
Before you begin

Review the Guidelines for adding a dataset of virtual objects on page 204.
Review the Requirements and restrictions when adding a dataset of virtual objects on page 207
Review the Best practices when adding or editing a dataset of virtual objects on page 187

Have the protection information available that you need to complete this task:

The type of virtual objects, either VMware objects or Hyper-V objects, that you want to
include
The name you want to give this dataset
The user group to whom you want this dataset visible
Whether you want to specify dataset-level custom naming formats for the Snapshot copy,
volume, and qtree objects that are generated by local policy or storage service protection jobs
on the virtual objects in this dataset
You must be authorized to perform all the steps of this task; your RBAC administrator can
confirm your authorization in advance.

About this task

The Create Dataset dialog box allows you to create an empty dataset, a populated but unprotected
dataset, a populated, remotely protected dataset, a populated locally protected dataset, or a populated
partially-configured or fully-configured dataset. A minimally configured dataset is an empty dataset.
Datasets of virtual objects must have any secondary protection and provisioning configured through a
storage service that you assign to them using the OnCommand console.
Steps

1. Click the View menu and click the Datasets option.


2. In the Datasets tab, click Create and select the option for the appropriate virtual object dataset
type.

To create a dataset to manage VMware objects, select the Dataset with VMware objects
option.
To create a dataset to manage Hyper-V objects, select the Dataset with Hyper-V objects
option.

204 | OnCommand Console Help


3. In the Create Dataset dialog box, select the Name option and enter the requested information in
the sub-tabs of the associated content area.
a. Enter the dataset name and administrative contact information in the General Properties tab.
b. If you want to specify dataset-level naming formats to apply to the Snapshot copy, volume,
and qtree objects that are generated by local policy or storage service protection jobs that are
run on this dataset, select options in the Naming Properties tab.
4. If you want to specify, at this time, the virtual objects to include in this dataset, select the Data
option and make your selections in the associated content area.
You can also add or change this information for this dataset at a later time.
5. If you want to specify, at this time, a storage service that will execute remote protection for the
objects in this dataset, select the Storage service option and make your selection.
You can also add or change this information for this dataset at a later time; however, changing
this information later, requires more time.
6. If you want to specify or create and configure, at this time, a local policy that will execute local
protection for the objects in this dataset, select the Local Policy option and make your local
policy selection or configuration.
You can also add or change this information for this dataset at a later time.
7. After you specify your desired information, if you want to test whether your dataset's new
configuration is in conformance with the OnCommand console requirements before you apply
them, click Test Conformance to display the Dataset Conformance Report.

If the Dataset Conformance Report displays no warning or error information, click Close and
continue.
If the Dataset Conformance Report displays warning or error information, read the Action and
Suggestion information resolve the conformance issue, then click Close and continue.

8. Click OK.
The OnCommand console creates your new dataset and lists it in the Datasets tab.
Related references

Administrator roles and capabilities on page 506


Create Dataset dialog box or Edit Dataset dialog box on page 258
Name area on page 260
Data area on page 264
Storage Service area on page 265
Local Policy area on page 266
Guidelines for adding a dataset of virtual objects
Before you create a dataset of virtual objects, you must decide the type of virtual objects to include
and the name of the dataset. If you want to fully configure your dataset in the initial session, you

Datasets | 205
must also decide on the set of virtual objects to include, the local policy that you want to assign, and,
if appropriate, the storage service that you want to assign.
The virtual object types you want to include
The dataset that you create can include either VMware virtual objects or Hyper-V virtual objects. A
dataset configured for VMware virtual objects can include Datacenter, Virtual Machine, and
datastore objects. A dataset configured for Hyper-V virtual objects can include virtual machines.
VMware objects and Hyper-V objects cannot co-exist in one dataset.
The scope of your initial dataset configuration
If you are ready to do so, the Edit Dataset dialog box and its four content areas enable you to create a
fully populated, fully protected dataset in one session.
However, if you are not ready to create a fully configured dataset in the initial session, the Edit
Dataset dialog box also allows you to create an empty dataset, a dataset of unprotected virtual
objects, a dataset of just locally protected virtual objects, or a dataset of virtual objects that are both
locally protected and remotely protected.
You can later edit a partially configured dataset to fully configure its membership or data protection.
General properties information
When you create a dataset for virtual objects, the minimum information you need to provide is the
general property information. If you complete a dataset configuration specifying only this
information, your result is a named but empty dataset.
Name

Your company might have a naming convention for datasets. If so, then the best
practice is for the dataset name to follow those requirements.

Description A useful description is one that helps someone unfamiliar with the dataset to
understand its purpose.
Owner

The name of the person or organization that is responsible for maintaining the virtual
objects that are included in this dataset.

Contact

You can specify one or more individual e-mail addresses or a distribution list of
people to be contacted when an event on the dataset triggers an alarm.

Group

If appropriate, you can specify the group to which you assign this dataset.

Naming properties information


When you create a dataset of virtual objects, you can specify dataset-level character strings and
naming formats to be applied to Snapshot copy, primary volume, secondary volume, or secondary
qtree objects that are generated by local policy or storage service protection jobs on this dataset.
You can also accept the globally configured default naming formats for those objects.

206 | OnCommand Console Help


Custom Label

You can accept the default naming process of including the dataset name as part
of the entire name of related objects created for this dataset or you can specify a
custom character string to use instead.

Snapshot
Copy

You can accept the global default Snapshot copy naming format to be applied to
all local policy or storage service generated Snapshot copies for this dataset, or
you can specify an alternative custom dataset-level naming format.

Secondary
Volume

You can accept the global default secondary volume naming format to be applied
to all local policy or storage service generated secondary volumes for this dataset,
or you can specify an alternative custom dataset-level naming format.

Secondary
Qtrees

You can accept the global default secondary qtree naming format to be applied to
all local policy or storage service generated secondary qtrees for this dataset, or
you can specify an alternative custom dataset-level naming format.

Data information
If you want to populate your dataset with virtual objects in the same session that you create it, you
must be ready to specify the following information:
Group

The resource group from which you want to select virtual objects to
include in the dataset.

Resource type
(applies to datasets
configured for
VMware objects)

If you are configuring a dataset for VMware virtual objects, what class of
supported VMware object (Datacenter, Virtual Machine, or datastore) you
want to include.

Resources in the
dataset

What virtual objects that meet your group and type selection criteria you
want to include in the dataset.
Any VMware datacenter objects that you include in a dataset cannot be
empty. They must contain virtual machine objects or a datastore that
contains virtual machine objects for successful backup.

Spanned entities
(applies to datasets
configured for
VMware objects)

If one of the VMware virtual machine objects that you want to include in a
dataset spans two or more datastores, the option of whether to include or
exclude any of those datastores from that dataset.

Local policy information


If you want to set up local protection, you must decide whether to apply an existing local policy or to
customize an additional local policy. Local policies specify the backup times, local copy retention
times, no-backup warning and error times, and path to any additional backup script you require.

Datasets | 207
If you want to set up remote protection, you must still generate local backup copies in primary
storage that an assigned storage service can copy and store in secondary storage. The usual method of
generating backups for this purpose is by scheduling local backup copies in the local policy.
If you intend to implement local backups on multiple datasets of Hyper-V objects that are associated
with the same Hyper-V server, you must configure separate local policies with non-overlapping
schedules to assign separately to each dataset.
Storage services information
If you want to set up remote protection (backup or mirroring to secondary and possibly tertiary
storage locations), you must be ready to select a storage service that is configured with a protection
policy that supports this possibility and to specify the path to any additional backup script that you
require.
Requirements and restrictions when adding a dataset of virtual objects
You must be aware of the requirements and restrictions when creating or editing a dataset of virtual
objects. Some requirements and restrictions apply to datasets of all types of virtual objects and some
are specific to datasets of Hyper-V or datasets of VMware virtual objects.
General requirements and restrictions

OnCommand console administrators attempting to assign a storage service to a dataset, require


RBAC read permission (DFM.Policy.Read) to any protection policy included in that storage
service.
This is in addition to having RBAC permissions DFM.Policy.Read, DFM.StorageService.Read
and DFM.StorageService.Attach to the storage service itself.
Administrators of datasets of virtual objects who are attempting to attach storage services to those
datasets require RBAC DFM.Resource.Control permission if the storage service they are
attempting to attach assigns a provisioning policy to the dataset primary node.
Even though provisioning policy assignment does not actually apply to datasets of virtual objects,
the DFM.Resource.Control permission is necessary to allow access to the underlying storage on
which the virtual objects are located.
VMware objects and Hyper-V objects cannot coexist in the same dataset.
Virtual objects (VMware objects or Hyper-V objects) cannot coexist in the same dataset with and
storage system container objects (such as aggregates, volumes, and qtrees).
To provide an application dataset with application consistent backup protection, the OnCommand
console operator must assign to that application dataset a storage service that is configured with a
protection policy that uses a "Mirror then backup" type protection topology.

VMware specific restrictions

VMware datacenter objects that you include in a dataset cannot be empty.


They must contain datastore or virtual machine objects for successful backup.

208 | OnCommand Console Help

VMDKs on a datastore object in a dataset must be contained within folders in that datastore. If
VMDKs exist outside of folders on the datastore, and that data is backed up, restoring the backup
could fail.

Hyper-V specific restrictions

A dataset with Hyper-V objects can only include virtual machines.


Hyper-V objects and storage system container objects (such as aggregates, volumes, and qtrees)
cannot coexist in the same dataset.
Shared virtual machines and dedicated virtual machines cannot coexist in the same dataset.
Each virtual machine contained in the dataset that you want to back up must contain at least 300
MB of free disk space.
Each Windows volume in the virtual machine (guest OS) must have at least 300 MB free disk
space. This includes the Windows volumes corresponding to VHDs, iSCSI LUNs and passthrough disks attached to the virtual machine.

Best practices when adding or editing a dataset of virtual objects


When you create or edit a dataset of virtual objects, observing best practices specific to datasets of
virtual objects helps you to avoid some performance and space usage problems after configuration.
General best practices for datasets of virtual objects
The following configuration practices apply to both datasets of Hyper-V objects and datasets of
VMware objects:

To avoid conformance and local backup issues caused by primary volumes reaching their
Snapshot copy maximum of 255, best practice is to limit the number of virtual objects included in
a primary volume, and limit the number of datasets to which each primary volume is directly or
indirectly included as a member.
A primary volume that hosts virtual objects that are included in multiple datasets is subject to
retaining an additional Snapshot copy of itself for every local backup on any dataset that any of
its virtual object children are members of.
To avoid backup schedule inconsistencies, best practice is to include only virtual objects that are
located in the same time zone in one dataset.
The schedules for the local protection jobs and remote protection jobs specified in the local
policies and storage services that are assigned a dataset of virtual objects are carried out
according to the time in effect on the host systems that are associated with the dataset's virtual
objects.

Best practices specific to datasets of Hyper-V objects


The following configuration practices apply specifically to datasets containing Hyper-V objects:

To ensure faster dataset backup of virtual machines in a Hyper-V cluster, best practice is to run
all the virtual machines on one node of the Hyper-V cluster.

Datasets | 209
When virtual machines run on different Hyper-V cluster nodes, separate backup operations are
required for each node in the cluster. If all virtual machines run on the same node, only one
backup operation is required, resulting in a faster backup.
Best practices specific to datasets of VMware objects
The following configuration practices apply specifically to datasets containing VMware objects:

If a virtual machine resides on more than one datastore, you can exclude one or more of those
datastores from the dataset. No local or remote protection is configured for the excluded
datastores.
You might want to exclude datastores that contain swap files that you want to exclude from
backup.
To avoid an excessive amount of secondary space provisioned for backup, best practice when
creating volumes to host the VMware datastores whose virtual machines will be protected by the
OnCommand console backup is to size those volumes to be not much larger than the datastores
they host.
The reason for this practice is that when provisioning secondary storage space to back up virtual
machines that are members of datastores, the OnCommand console allocates secondary space that
is equal to the total space of the volume or volumes in which those datastores are located. If the
host volumes are much larger than the datastores they hold, an excessive amount of provisioned
secondary space can result.

Editing a dataset to add virtual object members


You can add virtual VMware or virtual Hyper-V objects to existing datasets that were created to
contain VMware objects or Hyper-V objects.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

Datasets created as VMware datasets can only contain VMware objects.


Datasets created as Hyper-V datasets can only contain Hyper-V objects. Hyper-V shared virtual
machines and dedicated virtual machines cannot be included in the same dataset.
VMware datacenter objects that you include in a dataset cannot be empty.
They must contain datastore or virtual machine objects for successful backup.

Steps

1. Click the View menu and click the Datasets option.


2. In the Datasets tab, select the dataset to which you want to add virtual objects, and click Edit.

210 | OnCommand Console Help


3. If the selected dataset currently has no members or local policy assigned, specify, when
prompted, whether the members that you want to add are VMware objects Hyper-V objects.
4. In the Edit Dataset dialog box, select the Data option.
The associated content area displays all the virtual objects of the type (VMware or Hyper-V) that
are compatible with the selected dataset type.
5. Select the virtual objects to include in this dataset.
6. After you finish adding the desired virtual objects to your dataset, click OK.
The set of virtual objects in this dataset is updated with your additions.
Related references

Administrator roles and capabilities on page 506


Data area on page 264

Editing a dataset to assign storage service and remote protection of virtual


objects
You can configure remote protection of a dataset's virtual objects by assigning a storage service to
that dataset.
Before you begin

Local backup copies of the dataset's primary data, generated either by on-demand backups or by
scheduled backups specified in the local policy, must exist on the primary node for transfer by the
assigned storage service to a secondary node.
You must have a storage service available for assignment that is configured to support the
dataset's protection and provisioning requirements.
You must be authorized to perform all the steps of this task; your RBAC administrator can
confirm your authorization in advance.

OnCommand console administrators attempting to assign a storage service to a dataset,


require RBAC read permission (DFM.Policy.Read) to any protection policy included in that
storage service.
This is in addition to having RBAC permissions DFM.Policy.Read,
DFM.StorageService.Read and DFM.StorageService.Attach to the storage service itself.
Administrators of datasets of virtual objects who are attempting to attach storage services to
those datasets require RBAC DFM.Resource.Control permission if the storage service they
are attempting to attach assigns a provisioning policy to the dataset primary node.
Even though provisioning policy assignment does not actually apply to datasets of virtual
objects, the DFM.Resource.Control permission is necessary to allow access to the underlying
storage on which the virtual objects are located.

Datasets | 211
About this task

Datasets of virtual objects must have any secondary protection and provisioning configured
through a storage service that you assign to them using the OnCommand console.
To provide an application dataset with application consistent backup protection, the OnCommand
console operator must assign to that application dataset a storage service that is configured with a
protection policy that uses a "Mirror then backup" type protection topology.

Steps

1. Click the View menu and click the Datasets option.


2. In the Datasets tab, select the dataset for whose virtual objects you want to configure remote
protection, and click Edit.
3. In the Edit Dataset dialog box use the Storage service option to view and assign a storage
service that will execute remote protection for the objects in this dataset.
As you make your selection, the associated content area displays the administrative, topological,
protection schedule, and backup retention information on the storage service that you select.
4. After you make your selection, if you want to test whether your dataset's settings and resources
conform to the requirements of its newly assigned storage service before you apply that storage
service, click Test Conformance to display the Dataset Conformance Report.

If the Dataset Conformance Report displays no warning or error information, click Close and
continue.
If the Dataset Conformance Report displays warning or error information, read the Action and
Suggestion information resolve the conformance issue, then click Close and continue.

5. When you are ready to confirm the storage service assignment, click OK.
The OnCommand console saves your dataset with its newly added members and lists your dataset
and its members in the Datasets tab.
Related references

Administrator roles and capabilities on page 506


Storage Service area on page 265

Editing a dataset of virtual objects to configure local policy and local


backup
You can select, modify, or configure new local policies to automate local protection of datasets
containing virtual VMware objects or virtual Hyper-V objects.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.

212 | OnCommand Console Help


Steps

1. Click the View menu and click the Datasets option.


2. In the Datasets tab, select the dataset on which you want to schedule and configure local backups
and click Edit.
3. In the Edit Dataset dialog box, select the Local Policy option and drop down list and complete
one of the following actions:

If you want to assign an existing local policy, select that policy from the Local Policy drop
down list.
If you want to assign an existing local policy with some modifications select that policy, make
your modifications in the content area, and click Save.
If you want to configure a new local policy to apply to this dataset, select the Create New
option, configure the policy in the content area, and click Create.

4. After you finish assigning a new or existing local policy to this dataset, if you want to test
whether your dataset's new configuration is in conformance with OnCommand console
requirements before you apply them, click Test Conformance to display the Dataset
Conformance Report.

If the Dataset Conformance Report displays no warning or error information, click Close and
continue.
If the Dataset Conformance Report displays warning or error information, read the Action and
Suggestion information resolve the conformance issue, then click Close and continue.

5. Click OK.
Any local policy assignment, modification, or creation that you completed will be applied to the
local protection of the virtual objects in the selected dataset.

Editing a dataset containing virtual objects to reschedule or modify local


backup jobs
You can modify the schedule of local backup jobs that are configured in the local policy assigned to
a dataset containing virtual VMware objects or virtual Hyper-V objects.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

If you need to reschedule or modify the local backup jobs associated with the local policy of a dataset
of virtual objects, you can edit the Local Policy settings of that dataset.

Datasets | 213
Steps

1. Click the View menu and click the Datasets option.


2. In the Datasets tab, select the dataset on which you want to schedule and configure local backups
and click Edit.
3. In the Edit Dataset dialog box, locate the Local settings option and click >.
4. Modify the local backup jobs as needed.
5. After you finish changing the schedule for the local policy to this dataset, click OK.
Result

Any local policy modification that you completed will be applied to the local protection of the virtual
objects in all datasets that use that local policy.
Related references

Local Policy area on page 266


Administrator roles and capabilities on page 506

Editing a dataset to remove protection from a virtual object


You can permanently remove dataset-based local or remote protection from a virtual object by
removing that object from the protected dataset.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu and click the Datasets option.


2. In the Datasets tab, select the dataset that contains the virtual object whose local or remote
protection you want to remove and click Edit.
3. In the Edit Dataset dialog box, select the Data option.
The associated content area displays all the current members of the selected dataset in the list box
that is labeled "Members in the dataset."
4. Select the virtual objects that you want to remove as dataset member and click the < button.
5. After you finish removing the virtual objects from your dataset, click OK.
The set of virtual objects in this dataset is updated with your removals.

214 | OnCommand Console Help


Result

The virtual objects continue to exist as objects, but not as members of the dataset from which they
are removed. Protection and provisioning jobs that are executed on the remaining objects in the
dataset are no longer executed on the removed objects.
Related references

Administrator roles and capabilities on page 506


Data area on page 264

Adding a dataset of physical storage objects with dataset-level custom


naming
When you create a new dataset of physical storage objects, to improve recognition and usability, you
can customize the naming formats that are applied to that dataset's related objects (the Snapshot
copies, primary volumes, secondary volumes, or secondary qtrees that are generated in that dataset
by protection jobs run on that dataset).
Before you begin

Have the custom naming information available to complete this task:

The related object types whose naming you want to customize


If you want to include a custom label for your dataset in your custom naming formats, the
character string that you want to use
If you want to customize naming settings by selecting and ordering NetApp Management
Console-supplied attributes in the Add Dataset wizard Naming Properties page, the naming
attributes that you want to include in the naming format
If you plan to assign a policy, you must to be assigned a role that enables you to view policies.
If you plan to assign a provisioning policy, you must be assigned a role that enables you to attach
the resource pools configured for the policy.
You must be authorized to perform all the steps of this task; your RBAC administrator can
confirm your authorization in advance.

About this task

During this task, the OnCommand console launches NetApp Management Console. Depending
on your browser configuration, you can return to the OnCommand console by using the Alt-Tab
key combination or clicking the OnCommand console browser tab. After the completion of this
task, you can leave the NetApp Management Console open, or you can close it to conserve
bandwidth.
Dataset-level naming properties, if customized for the related object types in a dataset, override
the global naming settings for those object types in that dataset.

Datasets | 215
Steps

1. Click the View menu and click the Datasets option.


2. In the Datasets tab, click Create and select Dataset with Storage entities to start the Add
Dataset wizard.
3. When the wizard displays the Naming Properties panel, specify how you want the dataset
reference displayed in the names for your object types:

If you want the OnCommand console to include the name of the dataset in the names of its
generated Snapshot copy, primary volume, secondary volume, or secondary qtree objects,
select Use dataset name.
If you want the OnCommand console to use a custom character string in place of the dataset
name in the names of its generated related object, select Use custom label and enter the
character strings that you want to use.

4. Customize the naming settings for your related object types:

If you want the current global naming format to apply to one or more object types that are
generated in this dataset, select the Use global naming format option for those object types.
If you want to customize the dataset-level naming formats for one or more object types that
are generated in this dataset, select the Use custom format option for those object types, and
type the naming attributes in the order that you want those attributes to appear.

5. When you complete your naming configuration, click Next and complete the dataset creation.
6. To return to the Datasets tab, press Alt-Tab.
Result

After dataset creation is complete, the OnCommand console applies the custom dataset-level naming
formats to all objects created by protection and provisioning jobs for that dataset.
Related references

Administrator roles and capabilities on page 506

Adding a dataset of virtual objects with dataset-level custom naming


When you create a dataset of virtual objects, to improve recognition and usability, you can customize
the naming formats that are applied to that dataset's related objects (Snapshot copies, secondary
volumes, or secondary qtrees that are generated by protection jobs run on that dataset).
Before you begin

Have the protection information available:

The name you want to give this dataset


The user group that you want this dataset to be visible to.

216 | OnCommand Console Help

The dataset-level custom naming formats that you want to specify for the related object types
that are generated by local policy or storage service protection jobs on the virtual objects in
this dataset.
You must be authorized to perform all the steps of this task; your RBAC administrator can
confirm your authorization in advance.

About this task

Dataset level naming properties customized for a related object type in a dataset override any
conflicting global naming settings that might be configured for that object type.
Steps

1. Click the View menu and click the Datasets option.


2. In the Datasets tab, click Create and select the option for the appropriate virtual object dataset
type.

To create a dataset to manage VMware objects, select the Dataset with VMware entities
option.
To create a dataset to manage Hyper-V objects, select the Dataset with Hyper-V entities
option.

3. In the Create Dataset dialog box, select the Name option and enter the requested information in
the sub-tabs of the associated content area.
a. In the General Properties tab, enter the dataset name and administrative contact information.
b. In the Naming Properties tab, specify dataset-level naming formats to apply to the object
types that are generated by protection jobs run on this dataset.
4. If you want to specify, at this time, the virtual objects to be included in this dataset, select the
Data option and make your selections in the associated content area.
You can also add or change this information for this dataset at a later time.
5. If you want to specify, at this time, a storage service that executes remote protection for the
objects in this dataset, select the Storage service option and make your selection.
You can also add or change this information for this dataset at a later time.
6. If you want to specify or create and configure, at this time, a local policy that executes local
protection for the objects in this dataset, select the Local Policy option and make your local
policy selection or configuration.
You can also add or change this information for this dataset at a later time.
7. After you specify your desired amount of information about this dataset, click OK.
The OnCommand console creates your new dataset and lists it in the Datasets tab.

Datasets | 217
Related references

Administrator roles and capabilities on page 506


Name area on page 260

Editing a dataset of virtual objects for dataset-level custom naming


To improve recognition and usability, you can edit an existing dataset of virtual objects to customize
the dataset-level naming format that is applied to that dataset's related objects (the Snapshot copies,
secondary volumes, or secondary qtrees that are generated by protection jobs run on that dataset).
Before you begin

Have the custom naming information available:

The name of the dataset for which you want to configure custom naming
The related object types whose naming you want to customize
If you want to include a custom label for your dataset in your custom naming format, the
character string that you want to use
If you want to customize the naming settings by entering attributes on the Naming Properties
tab, the naming attributes that you want to include in the naming format
If you want to customize the naming settings by naming a pre-authored naming script, the
name and location of that script.
If you plan to assign a policy, you must be assigned a role that enables you to view policies.
If you plan to assign a provisioning policy, you must be assigned a role that enables you to attach
the resource pools configured for the policy.
You must be authorized to perform all the steps of this task; your RBAC administrator can
confirm your authorization in advance.

About this task

Dataset-level naming properties customized for the related object types in a dataset override any
conflicting global naming settings that might be configured for those related object types.
Steps

1. Click the View menu and click the Datasets option.


2. In the Datasets tab, select the dataset for which you want to configure dataset-level custom
naming, and click Edit.
3. In the Edit Dataset dialog box, select the Name & Properties option.
4. Click the Naming Properties tab.
5. Select the options and enter the formats necessary to specify the particular dataset-level custom
naming formats that you want.
6. After you finish the naming customizations that you want for this dataset, click OK.

218 | OnCommand Console Help


Related references

Administrator roles and capabilities on page 506

Editing a dataset of physical storage objects for dataset-level custom


naming
To improve recognition and usability, you can edit an existing dataset to customize the naming
format that is applied to that dataset's protection-related objects (Snapshot copies, primary volumes,
secondary volumes, or secondary qtrees that are generated by protection jobs run on that dataset).
Before you begin

Have the custom naming information available to complete this task:

The name of the dataset for which you want to configure custom naming
The related object whose naming you want to customize
If you want to include a custom name for your dataset in your custom naming format, the
character string that you want to use
If you want to customize naming settings by selecting and ordering attributes from the
Naming Properties page, the naming attributes that you want to include in the naming format
If you plan to assign a policy, you must be assigned a role that enables you to view policies.
If you plan to assign a provisioning policy, you also need a role that enables you to attach the
resource pools configured for the policy.
You must be authorized to perform all the steps of this task; your RBAC administrator can
confirm your authorization in advance.

About this task

During this task, the OnCommand console launches NetApp Management Console. Depending
on your browser configuration, you can return to the OnCommand console by using the Alt-Tab
key combination or clicking the OnCommand console browser tab. After the completion of this
task, you can leave the NetApp Management Console open, or you can close it to conserve
bandwidth.
Dataset-level naming properties customized for the protection-related objects override any
conflicting global naming settings that might be configured.

Steps

1. Click the View menu and click the Datasets option.


2. In the Datasets tab, select the dataset for which you want to configure dataset-level custom
naming, and click Edit to display the Edit Dataset window.
3. Click the Naming Properties option and customize the naming settings for your selected object
type.
4. When you complete your naming configuration, click OK and complete the dataset edit.

Datasets | 219
5. Press Alt-Tab or click the OnCommand console browser tab to return to the OnCommand
console.
Result

After the dataset edit is complete, the OnCommand console applies the custom dataset-level naming
format that you specified to all future objects of that type that are generated by that dataset's
protection and provisioning jobs.
Related references

Administrator roles and capabilities on page 506

Selecting virtual objects to create a new dataset


You can select a set of virtual objects (of the same type) that you want to manage as one group and
combine them into a new dataset.
Before you begin

Have the protection information available that you need to complete this task:

The name you want to give this dataset


The user group that you want this dataset visible to.
Whether you want to use the dataset name or substitute name as a prefix to names assigned to
backup copies that are generated for this dataset.
You must be authorized to perform all the steps of this task; your RBAC administrator can
confirm your authorization in advance.

About this task

The objects that you select must all be of the same virtual object type.
The supported object types include VMware datacenter, virtual machine, or datastore objects or
Hyper-V virtual machine objects.
Any VMware datacenter objects that you include in a dataset cannot be empty.
They must contain datastore or virtual machine objects for successful backup.

Steps

1. Click the View menu and click the Servers option.


2. In the Server tab select the type of the virtual objects that you want to include in a new dataset.
Virtual objects of the selected type are displayed.
3. Select the objects that you want to include in the new dataset.
To select more than one object, hold down Ctrl while you make your selections.

220 | OnCommand Console Help


4. After you complete your selections, click Add to new dataset.
The Create Dataset dialog box appears, listing the selected objects as part of a new dataset.
5. In the Create Dataset dialog box, select the Name option and enter the requested information in
the associated content area.
6. If you want to specify, at this time, a storage service that will execute remote protection for the
objects in this dataset, select the Storage Service option and make your selection.
You can also add or change this information for this dataset at a later time.
7. If you want to specify or create and configure, at this time, a local policy that will execute local
protection for the objects in this dataset, select the Local Policy option and make your local
policy selection or configuration.
You can also add or change this information for this dataset at a later time.
8. After you specify your desired amount of information about this dataset, click OK.
9. When the OnCommand console displays a confirmation box that informs you that your new
dataset is created, complete one of the following actions:

Click Close to close the confirmation box and view the Server tab.
Click the linked dataset name to view the listing of the new dataset in the Datasets tab.

Related references

Administrator roles and capabilities on page 506


Name area on page 260
Data area on page 264
Storage Service area on page 265

Selecting virtual objects to add to an existing dataset


You can select a group of virtual objects and add them directly into an existing dataset.
Before you begin

Have the protection information available that you need to complete this task:
The type of virtual objects that you want to add to an existing dataset.
The names of the virtual objects that you want to add.
The name of the dataset to which you want to add your selected objects.
You must be authorized to perform all the steps of this task; your RBAC administrator can
confirm your authorization in advance.

About this task

Although a dataset might contain more than one object type of the same family, the objects that
you select for this operation must all be of the same virtual object type.

Datasets | 221

The supported object types include VMware datacenter, virtual machine, or datastore objects or
Hyper-V virtual machine objects.
Any VMware datacenter objects that you include in a dataset cannot be empty.
They must contain datastore or virtual machine objects for successful backup.

Steps

1. Click the View menu and click the Servers option.


2. In the Server tab select the type of the virtual objects that you to add to an existing dataset.
The Server tab displays virtual objects of the selected type.
3. Select the objects that you want to add to the new dataset.
To select more than one object, hold down Ctrl while you make your selections.
4. After you complete your selections, click Add to existing dataset.
The Add to existing dataset dialog box appears listing the selected objects and the existing
datasets that can accept objects of their type.
5. Select the dataset to which you want to add your selected objects and click Add.
6. When the OnCommand console displays a confirmation box that informs you that your existing
dataset is updated, complete one of the following actions:

Click Close to close the confirmation box and view the Server tab.
Click the linked dataset name to view the listing of the updated dataset in the Datasets tab.

Related references

Administrator roles and capabilities on page 506

Configuring local backups for multiple datasets of virtual Hyper-V objects


If you want to protect multiple datasets of virtual machines that are associated with one Hyper-V
parent host, you must configure separate local policies with non-overlapping schedules for each
dataset.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

A Hyper-V parent host does not allow simultaneous or overlapping local backups on multiple virtual
machines that are associated with it; therefore each associated dataset of Hyper-V objects that you
want to provide with local protection requires a separate local policy with a schedule that does not
overlap the schedule of any other local policy.

222 | OnCommand Console Help


Steps

1. In the Policies tab, select Hyper-V Local Policy Template and click Copy to create an
alternative local policy for each dataset that is associated with the Hyper-V parent host.
2. Still in the Policies tab, edit the Schedule and Retention area of each alternative local policy that
you just created so that none of those policies has a schedule that overlaps with the schedule of
any other.
3. In the Datasets tab, edit the Local Policy area of each separate dataset that is associated with the
Hyper-V parent host to assign it a separate one of the local policies that you have just edited.
Related references

Administrator roles and capabilities on page 506

Managing datasets
Performing an on-demand backup of a dataset
You can perform a local on-demand dataset backup to protect your virtual objects.
Before you begin

You must have reviewed the Guidelines for performing an on-demand backup on page 277
You must have reviewed the Requirements and restrictions when performing an on-demand
backup on page 279
You must have added the virtual objects to an existing dataset or have created a dataset and added
the virtual objects that you want to back up.
You must be authorized to perform all the steps of this task; your RBAC administrator can
confirm your authorization in advance.
You must have the following information available:

Dataset name
Retention duration
Backup settings
Backup script location
Backup description

About this task

If you perform a backup of a dataset containing Hyper-V virtual machines and you are currently
restoring those virtual machines, the backup might fail.

Datasets | 223
Steps

1. Click the View menu, then click the Datasets option.


2. In the Datasets tab, choose the dataset that you want to back up.
3. Click Back Up Now.
4. In the Back Up Now dialog box, specify the local protection settings, backup script path, and
backup description for the on-demand backup.
If you have already established local policies for the dataset, that information automatically
appears for the local protection settings for the on-demand backup. If you change the local
protection settings, the new settings override any existing application policies for the dataset.
5. If you want a remote backup to begin after the local backup has finished, select the Start remote
backup after local backup box.
6. Click Back Up Now.
After you finish

You can monitor the status of your backup from the Jobs tab.
Guidelines for performing an on-demand backup
Before performing an on-demand backup of a dataset, you must decide how you want to assign
resources and assign protection settings.
General properties information
When performing an on-demand backup, you need to provide information about what objects you
want to back up, to assign protection and retention settings, and to specify script information that
runs before or after the backup operation.
Dataset
name

You must select the dataset that you want to back up.

Local
protection
settings

You can define the retention duration and the backup settings for your on-demand
backup, as needed.
Retention

You can choose to keep a backup until you manually delete it, or
you can assign a retention duration. By specifying a length of time
to keep the on-demand local backup, you can override the retention
duration in the local policy you assigned to the dataset for this
backup. The retention duration of a local backup defaults to a
retention type for the remote backup.
A combination of both the remote backup retention type and storage
service is used to determine the remote backup retention duration.

224 | OnCommand Console Help

For example, if you specify a local backup retention duration of two


days, the retention type of the remote backup is Daily. The dataset
storage service then verifies how long daily remote backups are kept
and applies this to the backup. This is the retention duration of the
remote backup.
The following table lists the local backup retention durations and the
equivalent remote backup retention type:

Backup
settings

Backup
script path

Local retention duration

Remote retention
type

Less than 24 hours

Hourly

1 day up to, but not including, 7 days

Daily

1 week up to, but not including, 31 days

Weekly

More than 31 days

Monthly

You can choose your on-demand backup settings based on the type
of virtual objects you want to back up.
Allow saved state
backup (Hyper-V
only)

You can choose to skip the backup if it


causes one or more of the virtual machines
to go offline. If you do not choose this
option, and your Hyper-V virtual
machines are offline, backup operations
fail.

Create VMware
snapshot
(VMware only)

You can choose to create a VMware


formatted in addition to the storage system
Snapshot copies created during local
backup operations.

Include
independent disks
(VMware only)

You can include independent disks.


VMDKs belong to VMware virtual
machines in the current dataset, but reside
on datastores that are not part of the
current dataset.

You can specify a script that is invoked before and after the local backup. The script
is invoked on the host service and the path is local to the host service. If you use a
PowerShell script, you should use the drive letter convention. For other types of
scripts, you can use either the drive letter convention or the Universal Naming
Convention.

Datasets | 225
Backup
description

You can provide a description for the on-demand backup so you can easily find it
when you need it.

Clustered virtual machine considerations (Hyper-V only)


Dataset backups of clustered virtual machines take longer to complete when the virtual machines run
on different nodes of the cluster. When virtual machines run on different nodes, separate backup
operations are required for each node in the cluster. If all virtual machines run on the same node,
only one backup operation is required, resulting in a faster backup.
Requirements and restrictions when performing an on-demand backup
You must be aware of the requirements and restrictions when performing an on-demand backup.
Some requirements and restrictions apply to all types of objects and some are specific to Hyper-V or
VMware virtual objects.
Requirements

Virtual machines or datastores must first belong to a dataset before backing up.
You can add virtual objects to an existing dataset or create a new dataset and
add virtual objects to it.

Hyper-V specific Each virtual machine contained in the dataset that you want to back up must
requirements
contain at least 300 MB of free disk space. Each Windows volume in the
virtual machine (guest OS) must have at least 300 MB free disk space. This
includes the Windows volumes corresponding to VHDs, iSCSI LUNs, and
pass-through disks attached to the virtual machine.
Hyper-V virtual machine configuration files, snapshot copy files, and VHDs
must reside on Data ONTAP LUNs, otherwise backup operations fail.
VMware specific Backup operations of datasets containing empty VMware datacenters or
datastores will fail. All datacenters must contain datastores or virtual machines
requirements
to successfully perform a backup.
Virtual disks must be contained within folders in the datastore. If virtual disks
exist outside of folders on the datastore, and that data is backed up, restoring
the backup could fail.
NFS backups might take more time than VMFS backups. This is because it
takes more time for VMware to commit snapshots in a NFS environment.
Hyper-V specific Partial backups are not supported. If the Hyper-V VSS writer fails to back up
one of the virtual machines in the backup and the failure occurs at the Hyper-V
restrictions
parent host, the backup fails for all of the virtual machines in the backup.

226 | OnCommand Console Help

Deleting a dataset of virtual objects


You can delete a dataset if you want to stop protection for all of its members and stop conformance
checking against its assigned protection policies.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

When you delete a dataset, the physical resources that compose the dataset are not deleted.
Steps

1. Click the View menu and click the Datasets option.


2. In the Datasets tab, select the dataset that you want to delete and click Delete.
3. Click Yes to confirm the delete request or No to cancel the request and close the dialog box.
After you finish

The deleted dataset is no longer listed in the Datasets tab.


Related references

Administrator roles and capabilities on page 506

Suspending dataset protection and conformance checking


You can temporarily stop data protection and provisioning conformance checks on a dataset and its
members.
Before you begin
About this task

Before performing maintenance on volumes used as destinations for backups or mirror copies,
you might want to stop protection and conformance checking of the dataset to which the volume
belongs to ensure that the protection application does not initiate a new backup or mirror
relationship for the primary data.
Note: If you suspend protection for a dataset and the lag time exceeds the threshold defined for

the dataset, no lag threshold event is generated until protection is resumed. After you resume
protection for the dataset, the protection application generates the backlog of lag threshold

Datasets | 227
events that would have been generated had protection been in effect and triggers any applicable
alarms.
Note: When you suspend services on application datasets, the external application continues to

perform local backups as scheduled.

This task suspends all policies that are assigned to the dataset. You cannot choose to suspend only
protection when both protection and provisioning policies are assigned to a dataset.

Steps

1. Click the View menu and click the Datasets option.


2. In the Datasets tab, select the dataset whose protection you want to suspend, click More then
select Suspend.
3. In the confirmation dialog box, click Yes.
The Details area of the Datasets tab shows a Protection Suspended status if the dataset has a
protection policy assigned, and a Nonconformant status if the dataset has a provisioning policy
assigned.
4. Click Yes to confirm the suspend request or No to cancel the request and close the dialog box.
Result

All scheduled backups and provisioning are cancelled until service is resumed.
After you finish

After you bring the storage system volume online again, you must wait for the DataFabric Manager
server to recognize that the volume is back online. You can check the backup volume status using
Operations Manager.
You can resume data protection from the Datasets tab.

Resuming protection and conformance checking on a suspended dataset


You can resume data protection and conformance checking after those services were suspended on a
dataset.
About this task

If you suspended protection and conformance operations on a dataset in order to perform


maintenance on its volumes used as destinations for backups or mirror copies, you can resume
operations after maintenance is complete.
Note: If you suspend protection for a dataset and the lag time exceeds the threshold defined for the

dataset, no lag threshold event is generated until protection is resumed. After you resume
protection for the dataset, the protection application generates the backlog of lag threshold events
that would have been generated had protection been in effect and triggers any applicable alarms.

228 | OnCommand Console Help


Steps

1. Click the View menu and click the Datasets option.


2. In the Datasets tab, select the dataset on which you want to resume protection and provisioning
activities, click More then select Resume.
3. In the confirmation dialog box, click Yes.
4. Click Yes to confirm the resume request or No to cancel the request and close the dialog box.
Result

All scheduled backups and provisioning operations are resumed.

Changing a storage service on datasets of storage objects


You can use the Change Storage Service wizard to change a storage service on datasets of storage
objects that already have another storage service attached.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
The following conditions must exist:

The new storage service that you want to attach is available in the group in which you want to
locate that dataset.
All datasets to which you want to attach the new storage service are currently attached to the
same current storage service.

About this task

The Change Storage Service wizard allows you to select an alternative storage service, presents you
with possible node remapping alternatives along with rebaselining requirements for each alternative,
carries out a dry run of your request, and then implements your request upon your approval.
During this task, the OnCommand console launches NetApp Management Console. Depending on
your browser configuration, you can return to the OnCommand console by using the Alt-Tab key
combination or clicking the OnCommand console browser tab. After the completion of this task, you
can leave the NetApp Management Console open, or you can close it to conserve bandwidth.
Steps

1. Click the View menu and click the Datasets option.


2. In the Datasets tab, select the dataset whose storage service you want to change, click More then
select Storage Service and Change to start the Change Storage Service wizard.
3. After you complete each property sheet in the wizard, click Next.

Datasets | 229
4. Confirm the details of the storage service and click Finish.
5. Press Alt-Tab or click the OnCommand console browser tab to return to the OnCommand
console.
6. Refresh your browser to update the OnCommand console with the changes you made.
Result

The selected datasets are listed in the datasets table with their newly attached storage service named
in the storage service column.
Related references

Administrator roles and capabilities on page 506

Attaching a storage service to existing datasets of storage objects


You can use the Attach Storage Service wizard to attach a storage service to existing datasets.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Confirm that all datasets to which you want to attach the storage service currently use the same
protection policy or no protection policy.
About this task

The Attach Storage Service wizard allows you to select from a list of possible storage services,
presents you with possible node remappings and associated rebaselining requirements for the storage
service that you select, carries out a dry run of your request, and then implements your request upon
your approval.
After you attach a storage service to an existing dataset, you cannot directly edit that dataset to
change its individual protection policy selection, provisioning policy selections, or resource pool
selections as long as that storage service is attached. You can only edit the attached storage service to
change the protection policy, provisioning policy, or resource pool selections for all datasets attached
to that storage service.
During this task, the OnCommand console launches NetApp Management Console. Depending on
your browser configuration, you can return to the OnCommand console by using the Alt-Tab key
combination or clicking the OnCommand console browser tab. After the completion of this task, you
can leave the NetApp Management Console open, or you can close it to conserve bandwidth.
Steps

1. Click the View menu and click the Datasets option.

230 | OnCommand Console Help


2. In the Datasets tab, select the dataset to which you want to attach a storage service, click More
then select Storage Service and Attach to start the Attach Storage Service wizard.
3. After you complete each property sheet in the wizard, click Next.
4. Confirm the details of the storage service and click Finish.
5. Press Alt-Tab or click the OnCommand console browser tab to return to the OnCommand
console.
6. Refresh your browser to update the OnCommand console with the changes you made.
Result

The selected datasets are listed in the datasets table with their newly attached storage service named
in their storage service column.
Related references

Administrator roles and capabilities on page 506

Restoring data backed up from a dataset of physical storage objects


In addition to restoring backed up virtual objects, you can also restore data backed up from physical
storage objects (aggregates, volumes, and qtrees).
Before you begin

Before restoring backed-up data, have the following information available:

The backup version that you want to restore.


The volumes, qtrees, directories, files or Open Systems SnapVault directories you want to restore.

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

You can restore files contained in volumes, qtrees, and vFiler units that were backed up as members
of a dataset.
During this task, the OnCommand console launches NetApp Management Console. Depending on
your browser configuration, you can return to the OnCommand console by using the Alt-Tab key
combination or clicking the OnCommand console browser tab. After the completion of this task, you
can leave the NetApp Management Console open, or you can close it to conserve bandwidth.
Steps

1. Click the View menu and select the Datasets option.

Datasets | 231
2. In the Datasets tab, select the dataset whose data you want to restore, click More then select
Restore to start the Restore wizard.
The wizard displays the Backup Files window.
3. In the Backup Files window, select the backup copy containing the data that you want to restore.
4. Select the volumes, qtrees, directories and files contained in the backup copies that you want to
restore and continue running the Restore wizard.
5. Click Finish to end the wizard and begin the restore operation.
6. Press Alt-Tab or click the OnCommand console browser tab to return to the OnCommand
console.
After you finish

You can use the Jobs window to track the progress of the restore job and monitor the job for possible
errors.
Related references

Administrator roles and capabilities on page 506

Repairing datasets that contain deleted virtual objects


You can reenable full OnCommand console backup of a dataset after that backup has been partially
disabled by the deletion of virtual object members from inventory by third-party management tools
before those members were removed from the dataset.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

If a virtual object that is a member of a dataset is deleted by use of third-party management tools
from inventory before it is removed as a member from its dataset, any subsequent backup jobs
attempted on that dataset are only partially successful until you complete the following actions to
remove their references from the dataset.
Steps

1. Click the View menu and click the Datasets option.


2. In the Datasets tab, select the dataset whose local or remote protection jobs have been achieving
only partial success and click Edit.
3. In the Edit Dataset dialog box, select the Data option.

232 | OnCommand Console Help


The associated content area displays all the current members of the selected dataset in the list box
that is labeled "Members in the dataset." Any virtual object members that have been deleted from
inventory and thus no longer exist are labeled as "Deleted."
4. Select the virtual objects that have been deleted from inventory and are marked "Deleted."
5. Click the < button.
The selected virtual objects are removed from the dataset.
6. After you finish removing the objects that were deleted from inventory from your dataset, click
OK.
Result

After the update of the dataset is complete, the partially successful backup failures caused by the
dataset containing deleted virtual objects stop, and fully successful backup jobs resume.
Related references

Administrator roles and capabilities on page 506

Evaluating and resolving issues displayed in the Conformance Details


dialog box
You can use the Conformance Details dialog box to evaluate and resolve a dataset's conformance
issues. You can evaluate the error and warning messages that the dialog displays and attempt to
resolve the conformance issues they describe.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

This procedure assumes you are viewing the Conformance Details dialog box that you displayed by
selecting a nonconformant dataset in the Datasets tab and clicking
display.

after its nonconformant status

The dialog box displays Information, Error, Action, Reason, and Suggestion text about the
nonconformance condition.
Steps

1. Evaluate the displayed Information, Error, Action, Reason or Suggestion text.


Information Indicates configuration operations that the OnCommand console successfully
completed without conformance issues.

Datasets | 233
Error

Indicates the configuration operations that the OnCommand console can not
perform on this dataset due to conformance issues.

Action

Indicates what the OnCommand console conformance engine did to discover the
conformance issue.

Reason

Indicates the probable cause of the conformance issue.

Suggestion

Indicates a possible way of resolving the conformance issue. Sometimes the


suggested resolution involves a baseline transfer of data although it is usually
desirable to avoid reinitiating a baseline transfer of data .

2. Based on the dialog box text, decide the best way to resolve the conformance issue.

If the dialog box text indicates that the OnCommand console conformance monitor cannot
automatically resolve the conformance issue, resolve this issue manually.
If Suggestion text indicates that automatically resolving the conformance issues requires a
baseline transfer of data, first attempt to resolve this issue manually. If unsuccessful, consider
resolving the issue automatically even if doing so requires a baseline transfer of data.
If the Suggestion text indicates that the OnCommand console conformance monitor can
resolve the conformance issue automatically without reinitiating a baseline transfer of data,
first, consider resolving the issue manually. If unsuccessful resolve the issue automatically.

After you finish

Proceed to the appropriate task to resolve the dataset conformance issues.


Related references

Administrator roles and capabilities on page 506


Resolving conformance issues manually without a baseline transfer of data
If the text displayed in the Conformance Details dialog box identifies conformance issues that you
cannot resolve automatically, you can resolve the conformance issue manually.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

Because of the probable time and bandwidth required for a baseline transfer completion, a
resolution that avoids a baseline transfer of data is preferable to a resolution that triggers one.
This procedure assumes you are viewing the Conformance Details dialog box that you displayed
by selecting a nonconformant dataset Datasets tab and clicking
display.

after its nonconformant status

234 | OnCommand Console Help


Steps

1. In the Conformance Details dialog box, confirm that the messages indicate that the conformance
issues cannot be resolved automatically.
2. Using the conformance messages, determine what is causing the nonconformance problem and
attempt to correct the condition manually.
You might need to log in to another GUI or CLI console to resolve the issues.
3. After you have attempted to correct the condition, wait at least one hour for the conformance
monitor to update the dataset's conformance status.
4. Return to the Conformance Details dialog box and click Test Conformance to determine if the
conformance issue is resolved.
If the conformance issue is resolved, the Conformance Details dialog box does not display the
"Conform" button.
5. If the conformance issue is resolved, click Cancel.
6. If the conformance issue is not resolved, repeat Steps 2, 3, and 4.
After you finish

After you achieve dataset conformant status, continue with the operation that required the dataset to
be conformant.
Related references

Administrator roles and capabilities on page 506


Resolving conformance issues automatically without a baseline transfer of data
If the text displayed in the Conformance Details dialog box identifies conformance issues that you
can automatically resolve without reinitializing a baseline transfer of data, you can use the dialog box
controls to do so.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

Because of the probable time and bandwidth required for a baseline transfer completion, a
resolution that avoids a baseline transfer of data is preferable to a resolution that triggers one.
This procedure assumes you are viewing the Conformance Details dialog box that you displayed
by selecting a nonconformant dataset Datasets tab and clicking
display.

after its nonconformant status

Datasets | 235
Steps

1. In the Conformance Details dialog box, read the text to determine the ability of the conformance
engine to automatically resolve the nonconformant condition without reinitializing a baseline
transfer of data.
2. If the text suggests that a simple automatic resolution is possible, click Conform .
The OnCommand console conformance engine closes the Conformance Details dialog box and
attempts to reconfigure storage resources to resolve storage service protection and provisioning
policy conformance issues automatically.
3. Monitor the conformance status on the Datasets tab for one of the following values:

Conformant: The conformance issue is resolved.


Nonconformant: The conformance issue is not resolved. Consider manual resolution of the
issue.

After you finish

After you achieve dataset conformant status, continue with the operation that required the dataset to
be conformant.
Related references

Administrator roles and capabilities on page 506


Resolving conformance issues manually when a baseline transfer of data might be
necessary
If the text displayed in the Conformance Details dialog box identifies conformance issues whose
automated resolution might require a baseline transfer of data, attempt a manual resolution before
using automated resolution.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

Because of the probable time and bandwidth required for baseline transfer completion, a
resolution that avoids a baseline transfer of data is preferable to a resolution that triggers one.
This procedure assumes you are viewing the Conformance Details dialog box that you displayed
by selecting a nonconformant dataset in the Datasets tab and clicking
status display.

after its nonconformant

236 | OnCommand Console Help


Steps

1. In the Conformance Details dialog box, confirm that warning text is displayed that indicates that
a reinitialized baseline transfer of data might be required.
You should try to resolve the conformance issues manually before initializing a time-consuming
baseline transfer of your data.
2. Using the conformance messages, determine what is causing the conformance problem and
attempt to correct the condition manually.
You might need to log in to another GUI or CLI console to resolve the issues.
3. After you have attempted to correct the condition, wait at least one hour for the conformance
monitor to update the dataset's conformance status.
4. Return to the Conformance Details dialog box and click Test Conformance to determine if the
conformance issue is resolved.
If the conformance issue is resolved, the Conformance Details dialog box does not display the
"Conform" button.
5. If the conformance issue is not resolved, click Conform to attempt automated resolution and
initiate a rebaseline of your data.
After you finish

After you achieve dataset conformant status, continue with the operation that required the dataset to
be conformant.
Related references

Administrator roles and capabilities on page 506

Monitoring datasets
Overview of dataset status types
The OnCommand console reports on each dataset's protection status, conformance status, and
resource status.
Although protection policies are not assigned directly to datasets of virtual objects, the OnCommand
console still displays the statuses related to protection policies that are assigned indirectly to datasets
of virtual objects, as components of assigned storage services.

Datasets | 237

Descriptions of dataset protection status


A dataset is remotely protected only if the secondary storage system specified in the protection
relationship is successfully backing up data, and if the copies of the data can be restored. You can
monitor dataset status using the Datasets tab.
You should regularly monitor a dataset's protection, because the OnCommand console cannot
sufficiently protect the dataset under the following conditions:

If a secondary storage system runs out of storage space necessary to meet the retention duration
required by the protection policy
If the lag thresholds specified by the policy are exceeded

The following list describes protection status values and their descriptions:
Baseline Failed

No initial baseline data transfers have registered a backup version.

Initializing

The dataset is conforming to the protection policy and the initial baseline data
transfer is in process.

Job Failure

The most recent protection job did not succeed.


This status might result for any of the following reasons:

Lag Error

A backup from a SnapVault or Qtree SnapMirror relationship failed or could


not be registered.
A mirror copy from a SnapMirror relationship failed or could not be
registered.
Local backups (Snapshot copies) failed on the primary node.

The dataset has reached or exceeded the lag error threshold specified in the
assigned protection policy. This value indicates that there has been no successful
backup or mirror copy of a node's data within a specified period of time.
This status might result for any of the following reasons:

Lag Warning

The most recent local backup (Snapshot copy) on the primary node is older
than the threshold setting permits.
The most recent backup (SnapVault or Qtree SnapMirror) is older than the
lag threshold setting or no backup jobs have completed since the dataset was
created.
The most recent mirror (SnapMirror) copy is older than the lag threshold
setting or no mirror jobs have completed since the dataset was created.

The dataset has reached or exceeded the lag warning threshold specified in the
assigned protection policy. This value indicates that there has been no successful
backup or mirror copy of a node's data within a specified period of time.
This status might result for any of the following reasons:

238 | OnCommand Console Help

The most recent local backup (Snapshot copy) on the primary node is older
than the threshold setting permits.
The most recent backup (SnapVault or Qtree SnapMirror) is older than the
lag threshold setting or no backup jobs have completed since the dataset was
created.
The most recent mirror (SnapMirror) copy is older than the lag threshold
setting or no mirror jobs have completed since the dataset was created.

No Protection
Policy

The dataset is managed by the OnCommand console but no protection policy has
been assigned to the dataset.

Protected

The dataset has an assigned policy and it has conformed to that policy at least
once.

Protection
Suspended

An administrator has requested that scheduled backups be temporarily halted


until the administrator requests that they be resumed.

Uninitialized

This status might result for any of the following reasons:

The dataset has a protection policy that does not have any protection
operations scheduled.
The dataset does not contain any data to be protected.
The dataset does not contain storage for one or more destination nodes.
The single node dataset does not have any backup versions.
An application dataset requires at least one backup version associated with it.
The dataset does not contain any backup or mirror relationships.

No Local Policy A dataset of virtual objects has no local policy assigned.


Attention: Importing an external relationship into a dataset temporarily changes the dataset's

protection status to Unitialized. When the next scheduled backup or mirror backup job runs or
when you run an on-demand backup the protection status changes to reflect the results of the
protection job.
Descriptions of dataset conformance status
The dataset conformance status indicates whether a dataset is configured according to its local policy
or storage service's protection policy. To be in conformance, all secondary and tertiary storage that is
part of the backup relationship must be successfully provisioned and the provisioned objects must
match the requirements of the primary data. You can monitor dataset status using the Datasets tab.
The OnCommand console regularly checks a dataset for conformance. If it detects changes in the
dataset's membership or policy definition, the console does one of three things:

Automatically performs corrective steps to bring a dataset back into conformance


Presents you with a list of actions for your approval prior to correction
Lists conditions that it cannot resolve

Datasets | 239
You can view these actions and approve them in the Conformance Details dialog box.
A dataset might be nonconformant because there are no available resources from which to provision
the storage or because the NetApp Management Console data protection capability does not have the
necessary credentials to provision the storage resources.
The following list describes dataset conformance values:
Conformant

The dataset is conformant with all associated policies.

Conforming

The dataset is not in conformance with all associated policies. The OnCommand
console is performing actions to bring the dataset into conformance.

Nonconformant The OnCommand console cannot bring the dataset into conformance with all
associated policies and might require your approval or intervention to complete
this task.
Descriptions of dataset resource status
The dataset resource status indicates the event status for all resource objects that are assigned to the
dataset. The resources include those that are members of the secondary and tertiary storage systems.
If, for example, a tertiary member's status is critical, the dataset 's resource status also is displayed as
critical.
You can monitor dataset status using the Datasets tab . You can troubleshoot the resource objects
using .
The following list describes resource status values:
Normal

A previous abnormal condition for the resource returned to a normal state and the
resource is operating within the desired thresholds. No action is required.

Information A normal resource event occurred. No action is required.


Warning

The resource experienced an occurrence that you should be aware of. This event
severity does not cause service disruption, and corrective action might not be
required.

Error

The resource is still performing, but corrective action is required to avoid service
disruption.

Critical

A problem occurred that might lead to service disruption if you do not take
immediate corrective action.

Emergency The resource unexpectedly stopped working and experienced unrecoverable data
loss. You must take corrective action immediately to avoid extended downtime.

How to evaluate dataset conformance to policy


The OnCommand console periodically checks that the dataset conforms to its data protection policy.
If either the dataset membership or policy changes, the OnCommand console either tries to bring the

240 | OnCommand Console Help


dataset back into conformance or notifies the administrator that the dataset conformance status
changed to nonconformant.
You can view the Conformance Results from the Datasets tab the Datasets window the by clicking
next to Conformance in the Status area.
Why datasets fail to conform to policy
A dataset must meet several conditions to be conformant to its assigned policy.
A dataset is conformant with its policy when it meets the following conditions:

Its member storage systems are properly configured.


Its assigned secondary storage system is provisioned and has enough backup space.
Its protection policy includes all necessary relationships to enforce data backups or mirror copies.

Following are some of the common reasons datasets fail to conform to their protection policy:

Dataset storage service and protection policy definitions changed.


Dataset membership changed.
Volumes or qtrees were created or deleted at the storage system (external to the OnCommand
console).
The configuration of a dataset of virtual objects has been updated, but not yet communicated from
the DataFabric Manager server to the host service.
This nonconformance condition normally lasts only a few minutes until the DataFabric Manager
server updates the host service with the latest configuration.

How the OnCommand console monitors dataset conformance


The OnCommand console conformance monitor regularly checks the DataFabric Manager server
database for configuration information to determine if a dataset is in conformance with its assigned
policy.
Conformance status is determined based on data gathered from SNMP queries by system monitors.
The monitors update the DataFabric Manager server database at scheduled intervals. The
conformance monitor queries the DataFabric Manager server database for the information, which is
then displayed in the OnCommand console. As a result, the information displayed by the
conformance monitor is not real-time data. This can result in the conformance monitor results being
temporarily out of date with actual changes to a storage system or configuration.
Should a dataset be nonconformant, you can view the conformance details from the Datasets tab by
selecting an item in the dataset list then clicking next to Conformance in the Status area. This opens
the Conformance Results window, that shows the results of the last conformance run on the selected
dataset. The results provide a description of any problems found during the last conformance run and
suggestions for resolving the problems. Depending on the results shown, you can then either make
changes manually to your system configuration or click Conform to allow in the OnCommand
console. to automatically make changes in an attempt to bring the dataset into conformance.
When you click Conform, you give the OnCommand console full control to do whatever it can to
bring everything into conformance for the selected dataset. This could include initiating a rebaseline

Datasets | 241
of your data, which might require significant time and bandwidth. As a result, if you would not want
a rebaseline to occur, you should try manual corrections to your system to resolve conformance
issues before you choose to use the Conform option.
After making manual corrections to your system, you can return to the Conformance Results window
and click the Test Conformance button to see if any changes made to the system have brought the
dataset into conformance with the policy assigned to it. Test Conformance initiates a new check on
the dataset but does not execute a conformance run. The results of the check reflect the latest system
updates that have been identified by the monitors and captured in the DataFabric Manager server
database. Therefore, the information displayed in the Conformance Results window might not reflect
recent changes made to a storage system or configuration and could be outdated by a few minutes or
a few hours, depending on the changes made and the scanning interval for each monitor.
You can view a list of monitor intervals by using the command dfm option list | grep
Interval. Following are some common monitoring actions, the default update intervals for
standard DataFabric Manager server configurations, and the associated monitors:
Discover new hosts

Default interval: 15 minutes


Monitor: discover

Update sizes for aggregates, volumes, and free or used space

Default interval: 30 minutes


Monitor: dfmon

Find new disks, aggregates, volumes, qtrees

Default interval: 15 minutes


Monitors: fsmon, diskmon

Find vFiler units

Default interval: 1 hour


Monitor: vfiler

SnapMirror, SnapVault, and OSSV directory discovery

Default interval: 30 minutes


Monitor: relationships

Update license capabilities on storage systems

Default interval: 4 hours


Monitor: license

Dataset conformance conditions


The OnCommand console displays a dataset's conformance status in the Datasets tab . If the dataset
is nonconformant, there are several ways in which the dataset can be brought back into conformance.
You can view the Conformance Results from the Datasets tab by clicking
in the Status area.

next to Conformance

When the conformance monitor detects a change in the dataset's membership or policy definition, the
conformance monitor does one of three things:

Automatically performs corrective steps to bring a dataset back into conformance


Presents you with a list of actions for your approval prior to correction

242 | OnCommand Console Help

Lists conditions that it cannot resolve

Conditions the monitor can resolve


The following list describes some of the conditions that the conformance monitor can detect, the
actions it can take to bring the dataset back into conformance with its policy, and whether those
actions are automatic or they require your approval for completion.

The OnCommand console provisions a destination volume but the aggregate in which the volume
is contained is no longer a member of the assigned resource pool.
Corrective action: The OnCommand console creates a new volume and moves the
relationship to it.
Does the action require your approval? Yes
Corrective action: The OnCommand console moves the relationship to an existing volume.
Does the action require your approval? Yes

The OnCommand console provisions a destination volume but the aggregate in which the volume
is contained is no longer a member of the assigned resource pool.
Corrective action: The OnCommand console creates a new volume and moves the
relationship to it.
Does the action require your approval? Yes
Corrective action: The OnCommand console moves the relationship to an existing volume.
Does the action require your approval? Yes

The destination volume does not have enough backup space or it is over its "nearly full"
threshold.
Corrective
action:

The OnCommand console expands the volume in an aggregate that is a


member of the resource pool.
Does the action require your approval? No

Corrective
action:

The OnCommand console provisions the volume and migrates the physical
relationship to a new destination volume.
Does the action require your approval? Yes

The destination volume contains expired backup versions.


Corrective
action:

The OnCommand console deletes the backup versions. The console also
deletes the copies of the data if those copies do not contain other backup
versions.
Does the action require your approval? No

Policy calls for the source data to be mirrored but the source volume is not protected in a mirror
relationship.

Datasets | 243
Corrective
action:

The OnCommand console creates a new relationship and performs a baseline


transfer of the data. A baseline transfer is defined as an initial backup (also
known as a level-0 backup) of a primary volume to a secondary volume in
which the entire contents of the primary volume are transferred.
Does the action require your approval? No

Policy calls for the source data to be backed up but the source qtree is not protected in a backup
relationship.
Corrective
action:

The primary volume has extra mirror relationships.


Corrective action:

The OnCommand console creates a new relationship and performs a


baseline transfer of the data.
Does the action require your approval? No
The OnCommand console deletes the extra relationships.
Does the action require your approval? No

The primary qtree has extra backup relationships.


Corrective action:

The OnCommand console deletes the extra relationships.


Does the action require your approval? No

Conditions the monitor cannot resolve


The following list describes some of the conditions that the conformance monitor can detect but
cannot resolve. These conditions require manual intervention from an administrator to bring the
dataset back into conformance with its policy.

An imported relationship has been detected in which the secondary volume exceeds the
volFullThreshold.
Corrective action: You must manually increase the secondary volume size. The conformance
monitor cannot resolve this condition.

A dataset's assigned secondary resources do not offer appropriate backup space.


Corrective action: You must reconfigure the resource pool membership so that The
OnCommand console can successfully continue data protection.

The application does not have the appropriate credentials to access the assigned resources.
Corrective action: You must provide the credentials for access to the hosts or storage systems.

What is a test conformance check?


When you create or edit a dataset, you can run a test conformance check of your dataset
configuration or reconfiguration before you commit the OnCommand console to implementing those
configuration changes.
Because initializing a dataset protection configuration requires a baseline transfer of data from
primary storage to secondary storage and tertiary storage and can take a lot of time, common practice

244 | OnCommand Console Help


is to run a quick test on your dataset configuration settings to ensure their conformance to the
assigned protection and provisioning policies before you commit to that initializing process.
A test conformance button that is located in the Edit Dataset dialog box and, when appropriate, in the
Conformance Details dialog box enables you to test conformance when necessary.
What is the difference between Test Conformance and Conform Now?
You can use both of the command buttons in the Conformance Details dialog box to help resolve
conformance issues. However, these two buttons perform very different functions and you should
become familiar with their functions before using these buttons.
Test
Conformance

Allows you to test whether manual changes that you have made to your dataset
configuration have brought it into conformance with its protection and
provisioning policies before you execute a conformance run. The results of the
test reflect the latest system updates that have been identified by the monitors and
changes you have specified but not yet executed in the current OnCommand
console session.

Conform Now

Initiates the automated attempt by the OnCommand console to reconfigure the


current dataset to be in conformance with its protection and provisioning policies.
Clicking this button allows the conformance monitor to do whatever it can to
bring the dataset into conformance.
Possible automated actions include re-initiating a baseline transfer of your data,
which usually requires significant time and bandwidth. If the Conformance
Details dialog box message text indicates that a reinitialized baseline transfer
might be triggered, and you want to avoid the associated time and bandwidth
cost, you should try manual corrections to your system to resolve conformance
issues before you choose to use the Conform Now option.

Monitoring dataset status


You can monitor the status of your datasets for possible errors.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu and click the Datasets option.


2. In the Datasets tab, select the dataset whose status you want to monitor.
The Overview area of the Datasets tab displays the protection, conformance, resources, and space
statuses of the selected dataset.

Datasets | 245
3.

To view status details, click the

button, if displayed.

button is displayed for resources, clicking it displays a dialog box that lists events
If the
related to warning-level or critical-level resource issues. You can use the Acknowledge button to
mark an event as acknowledged. If you take actions outside of the dialog box that resolves an
event issue, you can use the Resolve button to mark that event as resolved.
4. To view details about secondary or tertiary nodes, click the corresponding tabs for these nodes.
Related references

Administrator roles and capabilities on page 506


Window layout customization on page 16

Monitoring backup and mirror relationships


You can monitor backup or mirror relationships that are governed by a protection policy that is
assigned to a dataset. For example, you can monitor the lag status of each relationship or monitor that
backups are being created as scheduled.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

Although protection policies are not assigned directly to datasets of virtual objects, the OnCommand
console still displays the backup and mirror relationships of a protection policy that is assigned
indirectly to datasets of virtual objects, as a component of an assigned storage service.
Steps

1. Click the Storage menu and click the View option.


2. In the Datasets tab, select the dataset whose backup or mirror relationships you want to monitor.
3. (Optional) You can customize the Datasets tab to your needs.
Related references

Administrator roles and capabilities on page 506


Window layout customization on page 16

246 | OnCommand Console Help

Listing nonconformant datasets and viewing details


You can list existing datasets that have become nonconformant with their protection and provisioning
policies. Then, you can view and attempt to resolve their nonconformance. Conformance issues can
prevent completion of dataset protection and provisioning operations.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu and click the Datasets option to display the Datasets tab.
2. In the Datasets tab click the column header labeled Conformance Status and select
Nonconformant.
3. If the Datasets tab lists a dataset with nonconformant status, select that dataset to display its
Details area.
4.

In the selected dataset's Details area, click the

button.

The Conformance Details dialog box displays the results of the most recent conformance check
and suggestions for resolving the issues encountered.
After you finish

After you display the Conformance Details dialog box, you must address and resolve the issues that
are indicated by its warning and error messages.
Related references

Administrator roles and capabilities on page 506

Evaluating and resolving issues displayed in the Conformance Details


dialog box
You can use the Conformance Details dialog box to evaluate and resolve a dataset's conformance
issues. You can evaluate the error and warning messages that the dialog displays and attempt to
resolve the conformance issues they describe.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.

Datasets | 247
About this task

This procedure assumes you are viewing the Conformance Details dialog box that you displayed by
selecting a nonconformant dataset in the Datasets tab and clicking
display.

after its nonconformant status

The dialog box displays Information, Error, Action, Reason, and Suggestion text about the
nonconformance condition.
Steps

1. Evaluate the displayed Information, Error, Action, Reason or Suggestion text.


Information Indicates configuration operations that the OnCommand console successfully
completed without conformance issues.
Error

Indicates the configuration operations that the OnCommand console can not
perform on this dataset due to conformance issues.

Action

Indicates what the OnCommand console conformance engine did to discover the
conformance issue.

Reason

Indicates the probable cause of the conformance issue.

Suggestion

Indicates a possible way of resolving the conformance issue. Sometimes the


suggested resolution involves a baseline transfer of data although it is usually
desirable to avoid reinitiating a baseline transfer of data .

2. Based on the dialog box text, decide the best way to resolve the conformance issue.

If the dialog box text indicates that the OnCommand console conformance monitor cannot
automatically resolve the conformance issue, resolve this issue manually.
If Suggestion text indicates that automatically resolving the conformance issues requires a
baseline transfer of data, first attempt to resolve this issue manually. If unsuccessful, consider
resolving the issue automatically even if doing so requires a baseline transfer of data.
If the Suggestion text indicates that the OnCommand console conformance monitor can
resolve the conformance issue automatically without reinitiating a baseline transfer of data,
first, consider resolving the issue manually. If unsuccessful resolve the issue automatically.

After you finish

Proceed to the appropriate task to resolve the dataset conformance issues.


Related references

Administrator roles and capabilities on page 506

248 | OnCommand Console Help

Resolving conformance issues manually without a baseline transfer of data


If the text displayed in the Conformance Details dialog box identifies conformance issues that you
cannot resolve automatically, you can resolve the conformance issue manually.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

Because of the probable time and bandwidth required for a baseline transfer completion, a
resolution that avoids a baseline transfer of data is preferable to a resolution that triggers one.
This procedure assumes you are viewing the Conformance Details dialog box that you displayed
by selecting a nonconformant dataset Datasets tab and clicking
display.

after its nonconformant status

Steps

1. In the Conformance Details dialog box, confirm that the messages indicate that the conformance
issues cannot be resolved automatically.
2. Using the conformance messages, determine what is causing the nonconformance problem and
attempt to correct the condition manually.
You might need to log in to another GUI or CLI console to resolve the issues.
3. After you have attempted to correct the condition, wait at least one hour for the conformance
monitor to update the dataset's conformance status.
4. Return to the Conformance Details dialog box and click Test Conformance to determine if the
conformance issue is resolved.
If the conformance issue is resolved, the Conformance Details dialog box does not display the
"Conform" button.
5. If the conformance issue is resolved, click Cancel.
6. If the conformance issue is not resolved, repeat Steps 2, 3, and 4.
After you finish

After you achieve dataset conformant status, continue with the operation that required the dataset to
be conformant.
Related references

Administrator roles and capabilities on page 506

Datasets | 249

Resolving conformance issues automatically without a baseline transfer of data


If the text displayed in the Conformance Details dialog box identifies conformance issues that you
can automatically resolve without reinitializing a baseline transfer of data, you can use the dialog box
controls to do so.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

Because of the probable time and bandwidth required for a baseline transfer completion, a
resolution that avoids a baseline transfer of data is preferable to a resolution that triggers one.
This procedure assumes you are viewing the Conformance Details dialog box that you displayed
by selecting a nonconformant dataset Datasets tab and clicking
display.

after its nonconformant status

Steps

1. In the Conformance Details dialog box, read the text to determine the ability of the conformance
engine to automatically resolve the nonconformant condition without reinitializing a baseline
transfer of data.
2. If the text suggests that a simple automatic resolution is possible, click Conform .
The OnCommand console conformance engine closes the Conformance Details dialog box and
attempts to reconfigure storage resources to resolve storage service protection and provisioning
policy conformance issues automatically.
3. Monitor the conformance status on the Datasets tab for one of the following values:

Conformant: The conformance issue is resolved.


Nonconformant: The conformance issue is not resolved. Consider manual resolution of the
issue.

After you finish

After you achieve dataset conformant status, continue with the operation that required the dataset to
be conformant.
Related references

Administrator roles and capabilities on page 506

250 | OnCommand Console Help

Resolving conformance issues manually when a baseline transfer of data might be


necessary
If the text displayed in the Conformance Details dialog box identifies conformance issues whose
automated resolution might require a baseline transfer of data, attempt a manual resolution before
using automated resolution.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

Because of the probable time and bandwidth required for baseline transfer completion, a
resolution that avoids a baseline transfer of data is preferable to a resolution that triggers one.
This procedure assumes you are viewing the Conformance Details dialog box that you displayed
by selecting a nonconformant dataset in the Datasets tab and clicking
status display.

after its nonconformant

Steps

1. In the Conformance Details dialog box, confirm that warning text is displayed that indicates that
a reinitialized baseline transfer of data might be required.
You should try to resolve the conformance issues manually before initializing a time-consuming
baseline transfer of your data.
2. Using the conformance messages, determine what is causing the conformance problem and
attempt to correct the condition manually.
You might need to log in to another GUI or CLI console to resolve the issues.
3. After you have attempted to correct the condition, wait at least one hour for the conformance
monitor to update the dataset's conformance status.
4. Return to the Conformance Details dialog box and click Test Conformance to determine if the
conformance issue is resolved.
If the conformance issue is resolved, the Conformance Details dialog box does not display the
"Conform" button.
5. If the conformance issue is not resolved, click Conform to attempt automated resolution and
initiate a rebaseline of your data.
After you finish

After you achieve dataset conformant status, continue with the operation that required the dataset to
be conformant.

Datasets | 251
Related references

Administrator roles and capabilities on page 506

Page descriptions
Datasets tab
The Datasets tab enables you to create, edit, survey, and manage protection of your datasets.
From the Datasets tab you can launch the configuration of datasets of both virtual objects and
physical objects, monitor their status, backup dataset content on demand, suspend and resume
protection, and initiate restore operations.

Command buttons on page 251


Datasets list on page 253
Members list on page 255
Related objects pane on page 255
Graph area on page 255
Overview tab on page 255
Primary Node tab on page 256
Primary Node to Backup tab on page 257
Backup tab on page 257

Command buttons
Create

Enables you to create datasets to manage physical storage objects or virtual objects. The
Create command gives you the following sub-options:

Edit

If you select the Dataset with Hyper-V objects sub-option, starts the Create
Dataset dialog box for adding a Hyper-V virtual object dataset.
If you select the Dataset with VMware objects sub-option, starts the Create
Dataset dialog box for adding a VMware virtual object dataset.
If you select the Dataset with Storage objects sub-option, starts the NetApp
Management Console Add Dataset wizard for adding a storage dataset.

Enables you edit the configuration of a selected dataset.

If you select a dataset of VMware virtual objects or Hyper-V virtual objects, starts
the OnCommand console Edit Dataset dialog box for editing datasets that hold
virtual objects.
If you select a dataset of storage objects, starts the NetApp Management Console
Edit Dataset window for editing a storage dataset.

252 | OnCommand Console Help

If you select a dataset that still contains no objects, enables you to specify the type
of dataset that you want it to be (Dataset with Hyper-V objects, Dataset with
VMware objects, or Dataset with Storage objects).

Delete

Deletes the selected dataset or datasets and thereby removes the protection relationships
among its member objects.

More

Displays a menu of additional commands.


Suspend

Suspends dataset protection for a dataset.


During the time that the OnCommand console suspends protection, it
displays a Protection Suspended status for the dataset.

Resume

Resumes suspended dataset backup protection for a dataset.

Restore

(Applies to datasets of physical storage objects only).


For datasets of physical storage objects, opens the NetApp
Management Console Restore wizard, from which you can select
backed-up copies of a selected resource to restore as the primary
data.
Restore of virtual objects is supported on the Backups tab and the
Server tab of the OnCommand console.

Back Up
Now

Storage
Service
Attach

(Enabled for datasets of physical storage objects)

Storage
Service
Detach

(Enabled for datasets of physical storage objects)

Storage
Service
Change

Opens the NetApp Management Console Edit Dataset window to


enable you to change the storage service that is assigned to the
selected dataset.

Detaches a currently attached storage service from the selected


dataset of physical storage objects.

Enables you to perform on-demand backup or mirror protection operations on datasets.

Refresh

Attaches a storage service to the selected dataset of physical storage


objects.

For datasets of VMware or Hyper-V objects, opens the Back Up Now dialog box in
the OnCommand console.
For datasets of physical storage objects, opens the Protect Now dialog box in
NetApp Management Console.

Updates the datasets list.

Datasets | 253
Datasets list
A list that provides information about existing datasets. Click a row in the list to view information in
the Details area about the selected dataset.
Name

Displays the name of the dataset.

Data Type

Displays the type of object that the selected dataset contains.


Valid data types include:

Overall Status

Physical
VMware
Hyper-V
Undefined (the dataset is empty of objects)

Displays the status derived from the combined status conditions for disaster
recovery, protection, conformance, space, and resources.
Overall status is computed based upon the following other status values:
Overall
Status:
Error

Overall
Status:
Warning

DR status condition: Error


Protection status condition: Lag error or Baseline failed
Conformance status condition: Nonconformant
Space status condition: Error
Resource status condition: Emergency, Critical, or Error
DR status condition: Warning
Protection status condition: Job failure, Lag warning,
Uninitialized, or No protection policy for a non-empty
dataset
Conformance status condition: NA
Space status condition: Warning
Resource status condition: Warning

Storage Service

Displays what storage service, if any, is attached to the dataset. A dataset that is
attached to a storage service uses the protection policy, provisioning policies,
resource pools, and vFiler unit configurations specified by that storage service.

Local Policy

Displays the name of the local protection policy, if any, that is attached to the
selected dataset. Datasets of virtual objects might have a local policy attached to
them.

Protection
Policy

Displays the name of the protection policy that is either assigned directly to a
dataset of physical storage objects or to a storage service that is then assigned to
a dataset. This information is hidden by default.

254 | OnCommand Console Help


Primary
Provisioning
Policy

Displays the name of the provisioning policy currently assigned to the primary
node of the dataset. If a provisioning policy is assigned to a secondary node in
the dataset, that name is displayed in the details area when you select the
secondary node in the graph area. This information is hidden by default.

Space Status

Displays the status of the available space for the selected dataset node (OK,
Warning, Error, or Unknown).

Conformance
Status

Indicates whether the dataset is Conformant, Nonconformant, or Conforming.

Protection
Status

Displays protection status.


Valid status values, in alphabetical order, are as follows:

Baseline Failure
Initializing
Job Failure
Lag Error
Lag Warning
No Local Policy
No Protection Policy Attached
Not Protected
Protected
Protection Suspended
Uninitialized

Resource Status Displays the most severe of all current events on all direct and indirect members
of the dataset nodes. Values can be Emergency, Critical, Error, Warning, or
Normal.
Failed Over

Indicates whether a disaster recovery-capable dataset has failed over. This


information is hidden by default. Valid values are as follows:

Description

Yes
Failover on the dataset was invoked and completed successfully, completed
with warnings, or completed with errors.
No
Failover on the dataset has not been invoked.
In Progress
Failover on the dataset is currently in progress.
Not Applicable
The dataset is not assigned a disaster recovery protection policy and,
therefore, is not capable of failover.

Describes the dataset.

Datasets | 255
Application

Displays the name of the application that created an application dataset, such as
SnapManager for Oracle. This item is not included in the dataset list by default.

Application
Version

Displays the version of the application that created the application dataset. This
item is not included in the dataset list by default.

Application
Server

Displays the name of the server that runs the application that created the
application dataset. This item is not included in the dataset list by default.

Members list
A folder list of the physical storage object types or the virtual object types that are currently included
as members of the selected dataset.

Clicking the folder for an object type displays the names of the dataset members of that object
type.
The names of virtual objects are linked to a dataset inventory page for their object type.
The names of physical storage objects are only listed.
If a dataset of physical objects contains more than three members of an object type, clicking
More displays all members of that type in a popup dialog box.
If the dataset of virtual objects contains more than three members of an object type, clicking
More displays all members of that type in the inventory page for that object type.

Related objects pane


A folder list of all the object types in primary storage and the backup copies that are created in
relation to the protection activities executed by the OnCommand console on this dataset.

Clicking the folder for an object type displays the names of the dataset's related objects of that
object type.
If the dataset contains more than three objects of that type, clicking More displays all objects of
that type, either in a popup dialog box or in the inventory page for that object type.

Graph area
The graphical representation of the nodes for the selected dataset are displayed in the lower section
of the page.
Overview tab
This tab displays the following status and general property details of the selected dataset:
Protection

Displays protection status. Click


next to the status value to view status details
and a list of any jobs that are associated with the status.

Baseline Failure
Initializing

256 | OnCommand Console Help

Job Failure
Lag Error
Lag Warning
No Protection Policy Attached
Not Protected
Protected
Protection Suspended
Uninitialized

Conformance Indicates whether the dataset is conformant. If a dataset is nonconformant, click


to evaluate errors and warnings and to run the conformance checker.
Resources

Represents the most severe of all current events on all direct and indirect members
of the dataset nodes. Values can be Emergency, Critical, Error, Warning, or
Normal. For Emergency, Critical, Error, or Warning conditions, click
evaluate the events and sources causing those conditions.

to

Space

Displays the status of the available space for the selected dataset node (OK,
Warning, Error, or Unknown). If any volume, qtree, or LUN of a dataset has space
allocation error or warning conditions, the dataset's space status indicates that
condition. You can select the dataset to scan its volumes, LUNs, or qtrees to
determine which member is the cause of the warning or error condition.

Failed over

Indicates whether a disaster recovery-capable dataset has failed over. This is


displayed for disaster recovery-capable dataset only. Valid values are the
following:

Description

Displays a description of the dataset if one is entered.

Owner

Displays the owner of the current dataset if one is specified.

Contact

Displays the e-mail contact address for this dataset if one is specified.

Time Zone

Displays the time zone in which the primary node of the selected dataset is located.
This detail applies only to empty datasets or datasets of physical objects.

Custom Label Displays values that the user might have defined for this dataset.
Primary Node tab
This subtab displays the schedule of local Snapshot copy backups to be executed on the primary
dataset node. If you change the name of the primary node the title of this subtab matches your
change.
Local Backup
Schedule

Displays the names of the local backup schedules that are assigned to the
primary data node of this dataset.

Datasets | 257
Primary Node to Backup tab
If the selected dataset is configured with secondary backup or mirror protection, this subtab is
displayed. If you change the default names of your primary or secondary nodes, the title of this
subtab matches your changes.
This subtab lists the following information about the connection between the primary node and
secondary node:
Relationships Displays the number of existing backup or mirror relationships for the connection
between the volume and qtree objects on the primary node and volume and qtree
objects on the secondary node.
Schedule

Displays the name of the schedule that is assigned to the backup or mirror
connection.

Throttle

Displays the name of the throttle schedule, if any, that is assigned to the backup or
mirror connection.

Lag Status

Displays the worst current lag status for the backup or mirror connection.

Note: If the selected dataset contains multiple connections between the primary node and multiple
secondary or tertiary nodes, then a subtab similar to this one is displayed for every such
connection.

Backup tab
If the selected dataset is configured with secondary backup or mirror protection, this subtab is
displayed. If you change the name of the secondary node, the title of this subtab matches your
change.
This lists the following information about the secondary node:
Provisioning Policy Lists the provisioning policy, if any, that are assigned to the secondary node.
Physical Resources Lists the physical resources that are assigned to the secondary node.
Resource Pools

Lists the resource pools, if any, that are assigned to the secondary node.

Failed over

Indicates whether a disaster recovery-capable dataset has failed over. This is


displayed for disaster recovery-capable dataset only. Valid values are the
following:

Note: If the selected dataset contains multiple secondary or tertiary nodes, then a subtab similar to
this one is displayed for every such node.
Related references

Window layout customization on page 16


Descriptions of dataset protection status on page 197

258 | OnCommand Console Help

Create Dataset dialog box or Edit Dataset dialog box


Both the Create Dataset dialog box and the Edit Dataset dialog box enables you to view and
configure virtual object membership and protection of a dataset.

Options on page 258


Dataset name and topology diagram on page 258
Dataset nodes table on page 258
Command buttons on page 259

Options
Name

Enables you to name, rename, view, and edit the dataset properties and naming
formats of the current dataset.

Data

Enables you to view and edit the virtual object membership of this dataset.

Local Policy Enables you to assign a local policy to this dataset (to execute local protection of this
dataset's virtual object members).

Storage
service

You can create and configure a new local policy.


You can assign, and, if necessary, edit an existing local policy.

Enables you select a storage service for this dataset (to execute remote protection of
this dataset's virtual object members) or review an existing storage service
assignment.
A storage service assignment cannot be changed.

Dataset name and topology diagram


This area displays the dataset name and the protection topology of the current dataset after you
configure the dataset Name area and assign a storage service.
The topology diagram reflects both the dataset's local policy and the type of protection specified by
the protection policy that is associated with any assigned storage service.
Dataset nodes table
Below the dataset name and topology, a table provides the node and job schedule details of the
current dataset.
Primary node Lists the following information related to primary storage:
name table
The names of virtual object members of the current dataset.
column
These are the objects that are located in the primary storage node.
The local backup jobs that are scheduled in primary storage by the current
dataset's assigned local policy.

Datasets | 259

The job times are based on the time settings of the systems running associated
host services.
Secondary
node name
table column

If the current dataset's topology includes secondary storage, lists the following
information related to secondary storage:

Storage systems and resource pools that provision the secondary storage node
of the current dataset.
The schedule for remote backup and mirror protection jobs between primary
and secondary storage nodes and retention times for the backed up data.
The listed jobs and retention times are those that are specified by the protection
policy that is associated with the storage service that is assigned to the current
dataset. The times are based on the time settings of the DataFabric Manager
server
If the retention duration and retention count for a secondary node are set to 0,
the table displays the text "No Transfer (No retention)" to indicate no actual
secondary back up of primary data has occurred.

Tertiary node If the current dataset's topology includes tertiary storage, lists the following
information related to tertiary storage:
name table
column
Storage systems and resource pools that provision the tertiary storage node of
the current dataset.
The schedule for remote backup and mirror protection jobs between secondary
and tertiary storage nodes and retention times for the backed up data.
The listed jobs and retention times are those that are specified by the protection
policy that is associated with the storage service that is assigned to the current
dataset.
If the retention duration and retention count for a tertiary node are set to 0, the
table displays the text "No Transfer (No retention)" to indicate no actual
tertiary back up of data has occurred.
Command buttons
Test
Conformance

Allows you to pretest the conformance of the latest modifications that you have
made to the dataset configuration in this dialog box before you save and apply
those modifications.

OK

Saves the latest changes that you have made to the data in the Create Dataset
dialog box or Edit Dataset dialog box as the latest configuration for this dataset.

Cancel

Cancels any changes you have made to the settings in the Create Dataset dialog
box or Edit Dataset dialog box since the last time you opened it.

260 | OnCommand Console Help

Name area
The Name area displays administrative and naming property information about the current dataset of
virtual objects.
General
Properties tab

Displays the name of the dataset; enables you to enter dataset description, owner,
and contact information about the dataset; and enables you to assign a resource
group to the dataset.

Naming
Properties tab

Enables you to accept global naming formats for the related objects of this dataset
or enables you to configure dataset-level naming formats to be applied to related
objects of this dataset.
Related objects are Snapshot copy, primary volume, secondary volume, or
secondary qtree objects that are generated by local policy or storage service
protection jobs on this dataset.

General Properties tab


The General Properties tab displays the name and administrative information about the current
dataset.
Name

Enables you to change the name of the current dataset.


Unless you change it, the OnCommand console displays a default dataset name that it
has assigned.

Description Enables you to enter or modify a description of the current dataset.


Owner

Enables you to enter the name for the owner of this dataset.

Contact

Enables you to enter an e-mail contact address.

Naming Properties tab


The Naming Properties tab enables you to specify dataset-level naming properties of Snapshot
copies, primary volumes, secondary volumes, and secondary qtrees that are generated when the
OnCommand console runs protection jobs on this dataset.

Custom label on page 261


Snapshot copy on page 261
Secondary volume on page 262
Secondary qtree on page 263

The Naming Properties tab of the Create Dataset dialog box or Edit Dataset dialog box enables you
to select or specify values for the following dataset-level Naming Settings.

Datasets | 261
Custom label
Specifies a dataset-specific identification string that can be included in the name of all objects that a
protection job generates for this dataset.
Use dataset name Includes the dataset name in the name of all objects that a protection job
generates for this dataset.
Use custom label Enables you to enter a custom character string to be included in the name of all
objects that a protection job generates for this dataset.
Snapshot copy
Specifies the name format used for Snapshot copies that are generated by protection jobs run on this
dataset. The display of some of these options depends on your previous configuration choices:
Use global
naming
format

If you previously chose this format in the Global Naming Settings Snapshot Copy
area in the Setup Options dialog box, selecting this option applies the global naming
format for Snapshot copies to all Snapshot copies that a protection job generates for
this dataset.

Use custom Enables you to specify a dataset-level naming format to apply to all Snapshot copies
that a protection job generates for this dataset. You can enter the following attributes
format
(separated by the underscore character) in this field in any order:

%T (timestamp attribute)

The year, month, day, and time of the Snapshot copy.


%R (retention type attribute)
The Snapshot copy's retention class (Hourly, Daily, Weekly, or Monthly)
%L (custom label attribute)
The custom label, if any, that is specified for the Snapshot copy's containing
dataset. If no custom label is specified, then the dataset name is included in the
Snapshot copy name.
The custom label enables you to specify a custom alphanumeric character, .
(period), _ (underscore), or - (hyphen) to include in the names of the related
objects that are generated by protection jobs that are run on this dataset. If the
naming format for a related object type includes the Custom label attribute,
then the value that you specify is included in the related object names. If you do
not specify a value, then the dataset name is used as the custom label. If you
include a blank space in the custom label string, the blank space is converted to
letter x in any Snapshot copy, volume, or qtree object name that includes the
custom label as part of its syntax.
%H (storage system attribute)
The name of the storage system that contains the volume from which a Snapshot
copy is made
%N (volume name attribute)

262 | OnCommand Console Help

Name
preview

The name of the volume from which a Snapshot copy is made


%A (application field attribute)
Data inserted by outside applications into the name of the Snapshot copy. For the
NetApp Management Console data protection capability, it is a list of qtrees
present in the volume from which a Snapshot copy is made.
%1, %2, %3 (digit suffix)
A one-digit, two-digit, or three-digit suffix, if required, to distinguish Snapshot
copies with otherwise matching names

Displays a sample Snapshot copy name that uses the default or custom naming
format that you selected or specified.

Secondary volume
Specifies the name format used for secondary volumes that are generated by protection jobs run on
this dataset. The display of some of these options depends on your previous configuration choices:
Use global
naming
format

If you previously chose this format in the Global Naming Settings Secondary Volume
area in the Setup Options dialog box, selecting this option applies the global naming
script for secondary volumes to all secondary volumes that a protection job generates
for this dataset.

Use custom Selecting this option enables you to specify a dataset-level naming format to apply to
all secondary volumes that a protection job generates for this dataset. You can enter
format
the following attributes (separated by the underscore character) in this field in any
order:

%L (custom label attribute)

The custom label, if any, that is specified for the secondary volume's containing
dataset. If no custom label is specified, then the dataset name is included in the
secondary volume name.
It enables you to specify a custom alphanumeric character, . (period), _
(underscore), or - (hyphen) to include in the names of the related objects that are
generated by protection jobs that are run on this dataset. If the naming format for a
related object type includes the Custom label attribute, then the value that you
specify is included in the related object names. If you do not specify a value, then
the dataset name is used as the custom label. If you include a blank space in the
custom label string, the blank space is converted to letter x in any Snapshot copy,
volume, or qtree object name that includes the custom label as part of its syntax.
%S (primary storage system name)
The name of the primary storage system
%V (primary volume name)
The name of the primary volume
%C (type)

Datasets | 263

Name
preview

The connection type (backup or mirror)


%1, %2, %3 (digit suffix)
A one-digit, two-digit, or three-digit suffix, if required, to distinguish secondary
volumes with otherwise matching names

Displays a sample secondary volume name that uses the default or custom naming
format that you selected or specified.

Secondary qtree
Specifies how secondary qtrees for this dataset that are generated by local policies or storage service
protection jobs are named. Options and information fields include the following:
Use global
naming
format

Selecting this option applies the global naming format for secondary qtrees to all
secondary qtrees that a protection job generates for this dataset.

Use custom Selecting this option enables you to specify a dataset-level naming format to apply to
all secondary qtrees that a protection job generates for this dataset. You can enter the
format
following attributes (separated by the underscore character) in this field in any order:

%L (custom label attribute)

The custom label, if any, that is specified for the secondary volume's containing
dataset. If no custom label is specified, then the dataset name is included in the
secondary volume name.
It enables you to specify a custom alphanumeric character, . (period), _
(underscore), or - (hyphen) to include in the names of the related objects that are
generated by protection jobs that are run on this dataset. If the naming format for
a related object type includes the Custom label attribute, then the value that
you specify is included in the related object names. If you do not specify a value,
then the dataset name is used as the custom label. If you include a blank space in
the custom label string, the blank space is converted to letter x in any Snapshot
copy, volume, or qtree object name that includes the custom label as part of its
syntax.
%S (primary storage system name)
The name of the primary storage system
%V (primary volume name)
The name of the primary volume
%C (type)
The connection type (backup or mirror)
%1, %2, %3 (digit suffix)
A one-digit, two-digit, or three-digit suffix if required to distinguish secondary
volumes with otherwise matching names

264 | OnCommand Console Help


Name
Preview

Displays a sample secondary qtree name that uses the default or custom naming
format that you selected or specified.

Data area
The Data area of the Create Dataset dialog box or Edit Dataset dialog box provides tabs and options
that enable you to add various kinds of virtual object types to the current dataset.
Data tab
The Data tab enables you to filter and select the virtual objects to include in the dataset.
Group

Specifies the OnCommand console resource group from which you want to select
virtual objects to include in the dataset.

Resource
Type

Specifies the virtual object types that you want to include in the dataset.

VMware associated object types include datacenter, datastores, and virtual


machines (VM).
Hyper-V associated object types include virtual machines (VM).

You cannot include both VMware object types and Hyper-V object types in one
dataset.
Available
Resources

Lists the virtual objects that you can select for inclusion in the dataset.
Only virtual objects in the selected resource group that match the selected resource
type are displayed.
Any VMware datacenter objects that you include in a dataset cannot be empty.
They must contain datastore or virtual machine objects for successful backup.

Selected
Resources

Lists the virtual objects that you have selected for inclusion in the dataset.

Spanned Entities tab


The Spanned Entities tab appears only for datasets that contain VMware objects. If one of your
selected virtual machine objects spans two or more datastores, then the Spanned Entities tab displays
all the datastores that include that virtual machine and allows you to include or exclude any of those
datastores from this dataset's local or remote protection jobs by selecting or deselecting their
selection boxes.
Any listed datastore that is not deselected will have the dataset's local and remote protection jobs
applied to it.

Datasets | 265

Storage Service area


The Storage Service area of the Create Dataset dialog box or Edit Dataset dialog box displays basic
information about the storage service that you selected to execute remote protection jobs on the
virtual object members of this dataset.
Storage service

Selects the storage service that you want to assign to this dataset.
Selecting a storage service using this option displays the basic information for
that storage service in the content area.
Storage services enabled for disaster recovery support are not displayed.
After a storage service is selected and assigned to a dataset, the assignment
cannot be changed.

Name

Displays the name of the selected storage service.

Description

Displays a short description of the selected storage service.

Owner

Displays the name of the owner of the selected description service.

Contact

Displays the contact e-mail of the person in charge of the selected storage
service.

Backup Script
Path

Specifies a path to an optional backup script (located on the DataFabric


Manager server) that can specify additional operations to be executed in
association with protection policy specified backups.

Topology

Displays the name and the graphical topology of the protection policy that the
selected storage service uses to execute remote protection of the selected
dataset.

Dataset nodes table


In the Storage Service area, a table provides the node and job schedule details of the protection
policy associated with the storage service that is assigned to the current dataset.
Secondary
node name
table column

If the current dataset's topology includes secondary storage, lists the following
information related to secondary storage:

Storage systems and resource pools that provision the secondary storage node
of the current dataset.
The schedule for remote backup and mirror protection jobs between primary
and secondary storage nodes and retention times for the backed up data.
The listed jobs and retention times are those that are specified by the protection
policy that is associated with the storage service that is assigned to the current
dataset.

266 | OnCommand Console Help


Tertiary node If the current dataset's topology includes tertiary storage, lists the following
information related to tertiary storage:
name table
column
Storage systems and resource pools that provision the tertiary storage node of
the current dataset.
The schedule for remote backup and mirror protection jobs between secondary
and tertiary storage nodes and retention times for the backed up data.
The listed jobs and retention times are those that are specified by the protection
policy that is associated with the storage service that is assigned to the current
dataset.
Local Policy area
The Local Policy area enables you to configure local protection for this virtual object dataset by
selecting an existing local policy or creating a new local policy and assigning it to the dataset.
Local Policy

Either specifies the local policy to be assigned to execute local backup


protection of the current dataset, or enables you to start creation of a new local
policy to execute local backup protection of the current dataset.

You can select an appropriate the OnCommand console-supplied local


policy.
You can create a new local policy.

Policy Name

Specifies the name of the local policy.

Description

Displays an editable description of the selected local policy.

Add

Adds a schedule to be applied to local backups of virtual objects on this dataset.

Delete

Deletes an existing schedule.

Schedule list

Displays details of the local backup schedules that are in effect for the displayed
local policy. For each local backup schedule, the following details are displayed:
Schedule
Type

The type of schedule (Hourly, Daily, Weekly, or Monthly)


whose details you are viewing

Start Time

The time of day that the local backups start for the associated
schedule

End Time

The time of day at which new local backups stop occurring


for the associated schedule (applies to Hourly type schedules
only)

Recurrence

The frequency with which local backups occur for the


associated schedule

Datasets | 267

Retention

The period of time that the backup copies associated with this
schedule remain on the storage systems before becoming
subject to automatic purging

Backup
Options

Additional options that you can enable for the selected local
policy.
Create
VMware
Snapshot

Whether to creates VMware quiesced


snapshots before taking the storage
system Snapshot copies during local
backup operations (displayed for
datasets of VMware objects).

Include
independent
disks

Whether to include independent


VMDKs in local backups associated
with this schedule for this dataset
(displayed for datasets of VMware
objects only).

Allow saved
state backups

Whether to allow the backup of a


dataset of virtual machines to proceed
even if some of the virtual machines in
that dataset are in saved state or shut
down. Virtual machines in saved state
or that are shut down, receive saved
state or offline backup. Performing a
saved state or offline backup can cause
downtime (displayed for datasets of
Hyper-V object.)
If this option is not selected,
encountering a virtual machine that is
in saved state or that is shutdown fails
the dataset backup.

Start a remote
backup after
local backup

If remote backup of data to secondary


storage is specified by a storage service
assigned to this dataset, whether to start
that remote backup immediately after
the local backup is finished.
If this option is not selected, any
remote backup to secondary storage
starts according the schedules
configured for the protection policy
specified for the storage service.

268 | OnCommand Console Help


Issue a warning
if there are no
backups for:

Specifies a period of time after which the OnCommand console issues a warning
event if no local backup has successfully finished.

Issue an error if Specifies a period of time after which the OnCommand console issues an error
event if no local backup has successfully finished.
there are no
backups for:
Backup Script
Path

Specifies a path to an optional backup script (located on the system upon which
the host service is installed) that can specify additional operations to be executed
in association with local backups.

Dataset
Dependencies

If clicked, displays the other datasets that use this local policy and would thus be
affected by changes made to the settings in this content area.
This button appears only if the local policy is also assigned to other datasets.

Save

Saves any settings or changes made to settings in this content area for a new or
an existing local policy.
The OnCommand console enables this button if you make or modify any setting
in this content area.

Cancel

Cancels any changes to the dataset's local policy settings that have not yet been
saved during the current session.

269

Backups
Understanding backups
Types of backups
You can perform scheduled or on-demand local backups, or remote backups of datasets. Depending
on what you need, the different types of backups offer variation in how you protect your data.
You can create scheduled local backups by adding or editing datasets and their
Scheduled
local backups application policies. The host service runs local backups so even when the
OnCommand console is down, your backups continue to run.
You can create on-demand local backups as you need them. On-demand backups
On-demand
local backups apply to datasets. You can add specific virtual machines or datastores to existing or
new datasets for backup. You can also select specific settings for on-demand
backups that might differ from the local policy that might be attached to a dataset,
including starting a remote backup after the local backup operation.
Remote
backups

You can create remote backups by assigning a storage service to the selected
dataset. The storage service you assign to the dataset determines when and how the
remote backup occurs. To create a remote backup, you must first create or edit a
local dataset backup, or perform an on-demand dataset backup. During dataset
creation, you can add the storage service to the dataset you want to back up.

Backup version management


Through backup version management, you can more easily locate and track information, and you can
optimize your system by keeping only current or necessary information.
When you transfer a local backup to a secondary node, you can track which virtual objects are
contained and which of those objects are restorable. These can be different from the virtual objects
contained in the primary backup.
You can use retention settings in a dataset protection policy, local policy, or local on-demand backup
setting to retain backups for set periods of time. When backups reach their expiration, they are
deleted from the storage system.
You can locate any backup, across datasets, using a part of its description. You can also locate
backups that are no longer in a dataset or that have been renamed. You can locate renamed backups
using the current or former name of a virtual object in the backup.
Mounted backups are not deleted.

270 | OnCommand Console Help

Backup scripting information


You can specify a script that is invoked before and after the local backup. The script is invoked on
the host service and the path is local to the host service.
If you would like to configure a script to run before or after your backup operation, you must first run
the script independently from the DOS command line to determine if the script generates any pop-up
messages. Pop-up messages cause the backup operation to not complete properly. Clear out the
condition that generates the user pop-up message to ensure that the script runs automatically without
requiring any user interaction.
Backup script conventions
If you use a PowerShell script, you should use the drive-letter convention. For other types of scripts,
you can use either the drive-letter convention or the Universal Naming Convention.
Pre-backup script arguments
Note: If the pre-backup script fails, the backup might fail.

The following arguments apply to scripts that run before the backup occurs:
prebackup

Indicates that the script is invoked prior to the backup.

resourceids

Specifies the colon-separated list of resource IDs that are backed up.

datasetid

Specifies the dataset identifier.

Post-backup script arguments


The following arguments apply to scripts that run after the backup occurs:
postbackup Indicates that the script is invoked after the backup.
resourceids Specifies the colon-separated list of resource IDs that are backed up.
datasetid

Specifies the dataset identifier.

backupid

Specifies the host service backup identifier.

snapshots

Specifies the comma-separated list of Snapshot copies that constitute the backup. You
should use one of the following formats:

storage system:/vol/volx:snapshot
storage system:/vol/volx/lun:snapshot
storage system:/vol/volx/lun/qtree:snapshot

Backups | 271

Example
The following script (.bat file) is invoked after the backup:
echo "********************* > c:\post.txt
IF %1 == postbackup echo %2 >> C:\post.txt
IF %1 == postbackup
echo %3 >> C:\post.txt
IF %1 == postbackup
echo %4 >> C:\post.txt
IF %1 == postbackup echo %5 >> C:\post.txt
IF %1 == postbackup echo %6 >> C:\post.txt
IF %1 == postbackup
echo %7 >> C:\post.txt
IF %1 == postbackup
echo %8 >> C:\post.txt
IF %1 == postbackup echo %9 >> C:\post.txt

The following output displays:


datasetid:17798
resourceids:netfs://172.17.167.151//vol/vol_nfs_rlabf1_new/
backupid:2ad744f8-97c7-48f6-8e1b-65aec531bb5d
snapshots:172.17.167.151:/vol/vol_nfs_rlabf1_new:
2011-06-10_1605-0700_weekly_datastore_1_rlabf1_vol_nfs_rlabf1_new_novm
snap
172.17.164.186:/vol/
vol_nfs_86:2011-06-10_1605-0700_weekly_datastore_1_sdwatf3_vol_nfs_86_
novmsnap
ECHO is on.
ECHO is on.
ECHO is on.

Retention of job progress information


When you install or upgrade the OnCommand Core Package, the DataFabric Manager server retains
the job progress information for your backup jobs. The default retention period is 90 days. If you
need to retain this information longer than 90 days, for example because of legal or corporate
requirements, you can specify a longer period of time to retain this information using the
purgeJobsOlderThan option with the dfbm option set command.

Guidelines for mounting or unmounting backups in a VMware environment


After you create a full or partial backup of your virtual machine or datastore, you can mount the
backup onto an ESX server to verify what the virtual machine or datastore contains. After you verify
the newly created backup to compare its content to the original and to look for mismatches, you can
unmount the mounted backup.
When you mount or unmount backups, follow these guidelines:

You cannot mount or unmount Hyper-V backups just like VMware backups by clicking a button.

272 | OnCommand Console Help

You can mount Hyper-V backups using a Snapshot copy, virtual hard disk (VHD), LUN and
storage system information available in the OnCommand console GUI.
You cannot mount a backup on the same or a different ESX server if that backup is already
mounted.
You must unmount this backup from the first ESX server prior to mounting a backup to a
different ESX server.
You can mount a local backup and a remote backup on any ESX host that is managed by the
same host service that was used when the backup was created.
If you include the same datastore in multiple backups and those backups are mounted, that
datastore is mounted multiple times.
These datastores can be differentiated because the name includes a mounted timestamp and
contains the dataset name.
Backup and restore of mounted objects is not supported.
If there is some data written in the mounted datastore, that data is lost when you unmount the
backup.
If a backup is mounted, you cannot delete it, even if it has expired, until you unmount the backup.
While mounting a remote mirror backup, if the corresponding primary mirror backup has already
been deleted, the mount request fails with a backup not found error.
After you mount a backup, the time it takes to copy data from the datastore depends on your
network bandwidth and whether this datastore is on a secondary storage system.

How the Hyper-V plug-in uses VSS


The Hyper-V plug-in provides integration with Microsoft Hyper-V Volume Shadow Copy Service
(VSS) writer to quiesce a virtual machine before making an application-consistent Snapshot copy of
the virtual machine. The Hyper-V plug-in is a VSS requestor and coordinates the backup operation to
create a consistent Snapshot copy, using VSS Hardware Provider for Data ONTAP.
The Hyper-V plug-in enables you to make application-consistent backups of a virtual machine, if you
have Microsoft Exchange, Microsoft SQL, or any other VSS aware application running on virtual
hard disks (VHDs) in the virtual machine. The Hyper-V plug-in coordinates with the application
writers inside the virtual machine to ensure that application data is consistent when the backup
occurs.
You can also restore a virtual machine from an application-consistent backup. The applications that
exist in the virtual machine restore to the same state as at the time of the backup. The Hyper-V plugin restores the virtual machine to its original location.
Overview of VSS
VSS (Volume Shadow Copy Service) is a feature of Microsoft Windows Server that coordinates
among data servers, backup applications, and storage management software to support the creation
and management of consistent backups. These backups are called shadow copies, or Snapshot copies.
VSS coordinates Snapshot copy-based backup and restore and includes these additional components:

VSS requestor

Backups | 273

The VSS requestor is a backup application, such as Hyper-V plug-in or NTBackup. It initiates
VSS backup and restore operations. The requestor also specifies Snapshot copy attributes for
backups it initiates.
VSS writer
The VSS writer owns and manages the data to be captured in the Snapshot copy. Hyper-V plug-in
is an example of a VSS writer.
VSS provider
The VSS provider is responsible for the creation and management of the Snapshot copy. A
provider can be either a hardware provider or a software provider:

A hardware provider integrates storage array-specific Snapshot copy and cloning functionality
into the VSS framework. The Data ONTAP VSS Hardware Provider integrates the SnapDrive
service and storage systems running Data ONTAP into the VSS framework.
Note: The Data ONTAP VSS Hardware Provider is installed automatically as part of the
SnapDrive software installation.

A software provider implements Snapshot copy or cloning functionality in software that is


running on the Windows system.
Note: To ensure the Data ONTAP VSS Hardware Provider works properly, do not use the
VSS software provider on Data ONTAP LUNs. If you use the VSS software provider to
create Snapshot copies on a Data ONTAP LUN, you will be unable to delete that LUN
using the VSS hardware provider.

Viewing installed VSS providers


To view the VSS providers installed on your host, complete these steps.
Steps

1. Select Start > Run and enter the following command to open a Windows command prompt:
cmd

2. At the prompt, enter the following command:


vssadmin list providers

The output should be similar to the following:


Provider
Provider
Provider
Version:

name: Data ONTAP VSS Hardware Provider


type: Hardware
Id: {ddd3d232-a96f-4ac5-8f7b-250fd91fd102}
6.4.0.xxxx

274 | OnCommand Console Help

Verifying that the VSS Hardware Provider was used successfully


To verify that the Data ONTAP VSS Hardware Provider was used successfully after a Snapshot copy
was taken, complete this step.
Step

1. Navigate to System Tools > Event Viewer > Application in MMC and look for an event with
the following values.
Source

Event ID

Description

Navsspr

4089

The VSS provider has


successfully completed
CommitSnapshots for
SnashotSetId id in n
milliseconds.

Note: VSS requires that the provider initiate a Snapshot copy within 10 seconds. If this time
limit is exceeded, the Data ONTAP VSS Hardware Provider logs Event ID 4364. This limit
could be exceeded due to a transient problem. If this event is logged for a failed backup, retry
the backup.

How the Hyper-V plug-in handles saved-state backups


Although the default behavior of the Hyper-V plug-in is to cause backups containing virtual
machines that are in the saved state or shut down to fail, you can perform a saved-state backup by
moving the virtual machines to a dataset that has a policy that allows saved-state backups.
You can also create or edit your dataset policy to allow a saved-state virtual machine backup. If you
choose this option, the Hyper-V plug-in does not cause the backup to fail when the Hyper-V VSS
writer backs up the virtual machine using the saved state or performs an offline backup of the virtual
machine. However, performing a saved-state or offline backup can cause downtime.
For more information about online and offline virtual machine backups, see the Hyper-V Planning
for Backup information in the Microsoft TechNet library.
Related information

Hyper-V Planning for Backup - http://technet.microsoft.com/en-us/library/

Overlapping policies and Hyper-V hosts


Because a Hyper-V parent host does not allow simultaneous or overlapping local backups on
multiple virtual machines that are associated with it, each associated dataset of Hyper-V objects that

Backups | 275
you want to provide with local protection requires a separate local policy with a schedule that does
not overlap the schedule of any other local policy in effect.

Co-existence of SnapManager for Hyper-V with the Hyper-V plug-in


To protect your Hyper-V virtual objects, you can install both SnapManager for Hyper-V and the
Hyper-V plug-in on the same Hyper-V parent hosts. Some actions, like using both SnapManager for
Hyper-V and the Hyper-V plug-in to back up the same virtual machines, are unsupported.
When the OnCommand console and the Hyper-V plug-in are installed on hosts that also use
SnapManager for Hyper-V, the version of SnapDrive for Windows that SnapManager for Hyper-V
uses is upgraded as well.
If you uninstall OnCommand console using the host installer, so that you can use SnapManager for
Hyper-V as a stand-alone interface, it also uninstalls SnapDrive for Windows. After uninstalling
OnCommand console, you must reinstall the appropriate version of SnapDrive for Windows. The
following actions are unsupported when using both SnapManager for Hyper-V and the Hyper-V
plug-in:
Backing up the same After you create a dataset with protection policies using OnCommand
console, you must delete the corresponding protection policy in
virtual machines
SnapManager for Hyper-V. If you use the same protection schedule for the
same group of Hyper-V virtual machines, and attempt to back them up
using both SnapManager for Hyper-V and OnCommand console, the
backups fail.
Automatically
deleting backups

SnapManager for Hyper-V does not automatically delete backups after you
remove the protection policy. When you no longer need the backup, you
must manually delete it in SnapManager for Hyper-V.

Reinstalling
SnapManager for
Hyper-V

After you have transitioned all of your SnapManager for Hyper-V dataset
information to the OnCommand console, and uninstalled SnapManager for
Hyper-V, you should not reinstall SnapManager for Hyper-V.

How to manually transition SnapManager for Hyper-V dataset information


You can use the stand-alone SnapManager for Hyper-V software to view your dataset protection
schedules so that you can use them as a reference when adding and editing Hyper-V datasets in the
OnCommand console.
You can still use the dataset and backup policies that you created using SnapManager for Hyper-V to
protect your data until you have fully transitioned to the OnCommand console. As you create new
datasets and protection schedules in the OnCommand console, you can begin deleting your
protection policies in SnapManager for Hyper-V.
After you create a dataset and protection schedule in the OnCommand console, you should not use
SnapManager for Hyper-V to back up the same virtual machines. If you use overlapping schedules
for the same virtual machines, the backups might fail, because only one VSS backup can run at a
time on a host. SnapManager for Hyper-V does not automatically delete old backups after you
disable the protection policy, so you must manually delete them when they are no longer necessary.

276 | OnCommand Console Help


You can also use SnapManager for Hyper-V backups to restore virtual machines until the retention
policies have timed out. You should not restore the same virtual machines in SnapManager for
Hyper-V and the OnCommand console at the same time. If you do so, the restore operation might
fail.
After you transition all of your datasets and protection schedules to the OnCommand console, you
can safely uninstall SnapManager for Hyper-V.

Managing backups
Performing an on-demand backup of virtual objects
You can protect your virtual machines or datastores by adding them to an existing or new dataset and
performing an on-demand backup.
Before you begin

You must have reviewed the Guidelines for performing an on-demand backup on page 277
You must have reviewed the Requirements and restrictions when performing an on-demand
backup on page 279
You must have added the virtual objects to an existing dataset or have created a dataset and added
the virtual objects that you want to back up.
You must be authorized to perform all the steps of this task; your RBAC administrator can
confirm your authorization in advance.
You must have the following information available:

Dataset name
Retention duration
Backup settings
Backup script location
Backup description

About this task

If you perform a backup of a dataset containing Hyper-V virtual machines and you are currently
restoring those virtual machines, the backup might fail.
Steps

1. Click the View menu, then click the Server option.


2. In the Server tab, choose the virtual machine or datastore that you want to back up.

Backups | 277
If you want to back up...

Then...

A virtual machine or datastore that does not


belong to an existing dataset

You must first add it to an existing dataset.

A virtual machine or datastore, but no datasets


currently exist

You must first create a dataset and then add to it


the required virtual machines or datastores.

3. Click Backup and select the Back Up Now option.


4. In the Back Up Now dialog box, select the dataset that you want to back up.
If the virtual machine or datastore belongs to multiple datasets, you must select one dataset to
back up.
5. Specify the local protection settings, backup script path, and backup description for the ondemand backup.
If you have already established local policies for the dataset, that information automatically
appears for the local protection settings for the on-demand backup. If you change the local
protection settings, the new settings override only the existing application policies for this ondemand backup.
6. If you want a remote backup to begin after the local backup has finished, select the Start remote
backup after local backup check box.
7. Click Back Up Now.
After you finish

You can monitor the status of your backup from the Jobs tab.
Related references

Jobs tab on page 46


Administrator roles and capabilities on page 506
Guidelines for performing an on-demand backup
Before performing an on-demand backup of a dataset, you must decide how you want to assign
resources and assign protection settings.
General properties information
When performing an on-demand backup, you need to provide information about what objects you
want to back up, to assign protection and retention settings, and to specify script information that
runs before or after the backup operation.
Dataset
name

You must select the dataset that you want to back up.

278 | OnCommand Console Help


Local
protection
settings

You can define the retention duration and the backup settings for your on-demand
backup, as needed.
Retention

You can choose to keep a backup until you manually delete it, or
you can assign a retention duration. By specifying a length of time
to keep the on-demand local backup, you can override the retention
duration in the local policy you assigned to the dataset for this
backup. The retention duration of a local backup defaults to a
retention type for the remote backup.
A combination of both the remote backup retention type and storage
service is used to determine the remote backup retention duration.
For example, if you specify a local backup retention duration of two
days, the retention type of the remote backup is Daily. The dataset
storage service then verifies how long daily remote backups are kept
and applies this to the backup. This is the retention duration of the
remote backup.
The following table lists the local backup retention durations and the
equivalent remote backup retention type:

Backup
settings

Local retention duration

Remote retention
type

Less than 24 hours

Hourly

1 day up to, but not including, 7 days

Daily

1 week up to, but not including, 31 days

Weekly

More than 31 days

Monthly

You can choose your on-demand backup settings based on the type
of virtual objects you want to back up.
Allow saved state
backup (Hyper-V
only)

You can choose to skip the backup if it


causes one or more of the virtual machines
to go offline. If you do not choose this
option, and your Hyper-V virtual
machines are offline, backup operations
fail.

Create VMware
snapshot
(VMware only)

You can choose to create a VMware


formatted in addition to the storage system
Snapshot copies created during local
backup operations.

Backups | 279

Include
independent disks
(VMware only)

You can include independent disks.


VMDKs belong to VMware virtual
machines in the current dataset, but reside
on datastores that are not part of the
current dataset.

Backup
script path

You can specify a script that is invoked before and after the local backup. The script
is invoked on the host service and the path is local to the host service. If you use a
PowerShell script, you should use the drive letter convention. For other types of
scripts, you can use either the drive letter convention or the Universal Naming
Convention.

Backup
description

You can provide a description for the on-demand backup so you can easily find it
when you need it.

Clustered virtual machine considerations (Hyper-V only)


Dataset backups of clustered virtual machines take longer to complete when the virtual machines run
on different nodes of the cluster. When virtual machines run on different nodes, separate backup
operations are required for each node in the cluster. If all virtual machines run on the same node,
only one backup operation is required, resulting in a faster backup.
Requirements and restrictions when performing an on-demand backup
You must be aware of the requirements and restrictions when performing an on-demand backup.
Some requirements and restrictions apply to all types of objects and some are specific to Hyper-V or
VMware virtual objects.
Requirements

Virtual machines or datastores must first belong to a dataset before backing up.
You can add virtual objects to an existing dataset or create a new dataset and
add virtual objects to it.

Hyper-V specific Each virtual machine contained in the dataset that you want to back up must
requirements
contain at least 300 MB of free disk space. Each Windows volume in the
virtual machine (guest OS) must have at least 300 MB free disk space. This
includes the Windows volumes corresponding to VHDs, iSCSI LUNs, and
pass-through disks attached to the virtual machine.
Hyper-V virtual machine configuration files, snapshot copy files, and VHDs
must reside on Data ONTAP LUNs, otherwise backup operations fail.
VMware specific Backup operations of datasets containing empty VMware datacenters or
datastores will fail. All datacenters must contain datastores or virtual machines
requirements
to successfully perform a backup.

280 | OnCommand Console Help

Virtual disks must be contained within folders in the datastore. If virtual disks
exist outside of folders on the datastore, and that data is backed up, restoring
the backup could fail.
NFS backups might take more time than VMFS backups. This is because it
takes more time for VMware to commit snapshots in a NFS environment.
Hyper-V specific Partial backups are not supported. If the Hyper-V VSS writer fails to back up
one of the virtual machines in the backup and the failure occurs at the Hyper-V
restrictions
parent host, the backup fails for all of the virtual machines in the backup.

Performing an on-demand backup of a dataset


You can perform a local on-demand dataset backup to protect your virtual objects.
Before you begin

You must have reviewed the Guidelines for performing an on-demand backup on page 277
You must have reviewed the Requirements and restrictions when performing an on-demand
backup on page 279
You must have added the virtual objects to an existing dataset or have created a dataset and added
the virtual objects that you want to back up.
You must be authorized to perform all the steps of this task; your RBAC administrator can
confirm your authorization in advance.
You must have the following information available:

Dataset name
Retention duration
Backup settings
Backup script location
Backup description

About this task

If you perform a backup of a dataset containing Hyper-V virtual machines and you are currently
restoring those virtual machines, the backup might fail.
Steps

1. Click the View menu, then click the Datasets option.


2. In the Datasets tab, choose the dataset that you want to back up.
3. Click Back Up Now.
4. In the Back Up Now dialog box, specify the local protection settings, backup script path, and
backup description for the on-demand backup.

Backups | 281
If you have already established local policies for the dataset, that information automatically
appears for the local protection settings for the on-demand backup. If you change the local
protection settings, the new settings override any existing application policies for the dataset.
5. If you want a remote backup to begin after the local backup has finished, select the Start remote
backup after local backup box.
6. Click Back Up Now.
After you finish

You can monitor the status of your backup from the Jobs tab.
Related references

Jobs tab on page 46


Administrator roles and capabilities on page 506
Guidelines for performing an on-demand backup
Before performing an on-demand backup of a dataset, you must decide how you want to assign
resources and assign protection settings.
General properties information
When performing an on-demand backup, you need to provide information about what objects you
want to back up, to assign protection and retention settings, and to specify script information that
runs before or after the backup operation.
Dataset
name

You must select the dataset that you want to back up.

Local
protection
settings

You can define the retention duration and the backup settings for your on-demand
backup, as needed.
Retention

You can choose to keep a backup until you manually delete it, or
you can assign a retention duration. By specifying a length of time
to keep the on-demand local backup, you can override the retention
duration in the local policy you assigned to the dataset for this
backup. The retention duration of a local backup defaults to a
retention type for the remote backup.
A combination of both the remote backup retention type and storage
service is used to determine the remote backup retention duration.
For example, if you specify a local backup retention duration of two
days, the retention type of the remote backup is Daily. The dataset
storage service then verifies how long daily remote backups are kept

282 | OnCommand Console Help

and applies this to the backup. This is the retention duration of the
remote backup.
The following table lists the local backup retention durations and the
equivalent remote backup retention type:

Backup
settings

Local retention duration

Remote retention
type

Less than 24 hours

Hourly

1 day up to, but not including, 7 days

Daily

1 week up to, but not including, 31 days

Weekly

More than 31 days

Monthly

You can choose your on-demand backup settings based on the type
of virtual objects you want to back up.
Allow saved state
backup (Hyper-V
only)

You can choose to skip the backup if it


causes one or more of the virtual machines
to go offline. If you do not choose this
option, and your Hyper-V virtual
machines are offline, backup operations
fail.

Create VMware
snapshot
(VMware only)

You can choose to create a VMware


formatted in addition to the storage system
Snapshot copies created during local
backup operations.

Include
independent disks
(VMware only)

You can include independent disks.


VMDKs belong to VMware virtual
machines in the current dataset, but reside
on datastores that are not part of the
current dataset.

Backup
script path

You can specify a script that is invoked before and after the local backup. The script
is invoked on the host service and the path is local to the host service. If you use a
PowerShell script, you should use the drive letter convention. For other types of
scripts, you can use either the drive letter convention or the Universal Naming
Convention.

Backup
description

You can provide a description for the on-demand backup so you can easily find it
when you need it.

Backups | 283
Clustered virtual machine considerations (Hyper-V only)
Dataset backups of clustered virtual machines take longer to complete when the virtual machines run
on different nodes of the cluster. When virtual machines run on different nodes, separate backup
operations are required for each node in the cluster. If all virtual machines run on the same node,
only one backup operation is required, resulting in a faster backup.
Requirements and restrictions when performing an on-demand backup
You must be aware of the requirements and restrictions when performing an on-demand backup.
Some requirements and restrictions apply to all types of objects and some are specific to Hyper-V or
VMware virtual objects.
Requirements

Virtual machines or datastores must first belong to a dataset before backing up.
You can add virtual objects to an existing dataset or create a new dataset and
add virtual objects to it.

Hyper-V specific Each virtual machine contained in the dataset that you want to back up must
requirements
contain at least 300 MB of free disk space. Each Windows volume in the
virtual machine (guest OS) must have at least 300 MB free disk space. This
includes the Windows volumes corresponding to VHDs, iSCSI LUNs, and
pass-through disks attached to the virtual machine.
Hyper-V virtual machine configuration files, snapshot copy files, and VHDs
must reside on Data ONTAP LUNs, otherwise backup operations fail.
VMware specific Backup operations of datasets containing empty VMware datacenters or
datastores will fail. All datacenters must contain datastores or virtual machines
requirements
to successfully perform a backup.
Virtual disks must be contained within folders in the datastore. If virtual disks
exist outside of folders on the datastore, and that data is backed up, restoring
the backup could fail.
NFS backups might take more time than VMFS backups. This is because it
takes more time for VMware to commit snapshots in a NFS environment.
Hyper-V specific Partial backups are not supported. If the Hyper-V VSS writer fails to back up
one of the virtual machines in the backup and the failure occurs at the Hyper-V
restrictions
parent host, the backup fails for all of the virtual machines in the backup.

On-demand backups using the command-line interface


If you use the graphic user interface, the OnCommand console performs all backups at the dataset
level. However, you can also use the DataFabric Manager server command-line interface to choose a
subset of a dataset and perform an on-demand backup.
A subset of a dataset includes one or more virtual machines or datastores.

284 | OnCommand Console Help


You can also use Windows PowerShell cmdlets to perform on-demand backups. For more
information on using Windows PowerShell, see the OnCommand Windows PowerShell Cmdlets
Guide.
For more information on using the CLI, see the OnCommand Operations Manager Administration
Guide.
Related information

Operations Manager Administration Guide - http://support.netapp.com


NetApp OnCommand Windows PowerShell Cmdlets Guide for Use with Core Package 5.0
and Host Package 1.0 - http://support.netapp.com

Mounting or unmounting backups in a VMware environment


Mounting backups in a VMware environment from the Backups tab
You can mount existing backups onto an ESX server for backup verification prior to completing a
restore operation or to restore a virtual machine to an alternate location. All the datastores and the
virtual machines within the backup are mounted to the ESX server that you specify. Both the Mount
and Unmount buttons are disabled for Hyper-V backups.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

In the OnCommand console Backups tab, deleting a mirror source backup copy prevents you from
mounting its partner mirror destination backup copy. For a Mirror-generated destination backup copy
to be mountable, its associated mirror source backup copy must still exist on the source node.
Steps

1. Click the View menu, then click the Backups option.


2. In the Backups tab, select an unmounted backup that you want to mount.
3. Click Mount.
You cannot mount Hyper-V backups using this button.
4. In the Mount Backup dialog box, select from the drop-down list the name of the ESX server to
which you want to mount the backup.
You can only mount one backup each time and you cannot mount a mounted backup.
5. Click Mount.

Backups | 285
A dialog box appears with a link to the mount job and when you click the link, the Jobs tab
appears.
After you finish

You can monitor the status of your mount and unmount jobs in the Jobs tab.
Related references

Administrator roles and capabilities on page 506


Guidelines for mounting or unmounting backups in a VMware environment on page 271
Unmounting backups in a VMware environment from the Backups tab
After you are done using a mounted backup for verification or to restore a virtual machine to an
alternate location, you can unmount the mounted backup from the ESX server that it was mounted to.
When you unmount a backup, all the datastores in that backup are unmounted and can no longer be
seen from the ESX server that you specify. Both the Mount and Unmount buttons are disabled for
Hyper-V backups.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

If there are virtual objects in use from the previously mounted datastores of a backup, the unmount
operation fails. You must manually clean up the backup prior to mounting the backup again because
its state reverts to not mounted.
If all the datastores of the backup are in use, the unmount operation fails but this backup's state
changes to mounted. You can unmount the backup after determining the datastores are not in use.
Steps

1. Click the View menu, then click the Backups option.


2. In the Backups tab, select a mounted backup to unmount.
3. Click Unmount.
4. At the confirmation prompt, click Yes.
A dialog box opens with a link to the unmount job displays and when you click the link, the Jobs
tab appears.

286 | OnCommand Console Help


After you finish

If the ESX server becomes inactive or restarts during an unmount operation, the job is terminated and
the mount state remains mounted and the backup stays mounted on the ESX server.
You can monitor the status of your mount and unmount jobs in the Jobs tab.
Related references

Administrator roles and capabilities on page 506


Guidelines for mounting or unmounting backups in a VMware environment on page 271
Mounting backups in a VMware environment from the Server tab
You can mount existing backups onto an ESX server for backup verification prior to completing a
restore operation or to restore a virtual machine to an alternate location. All the datastores and the
virtual machines within the backup are mounted to the ESX server that you specify. Both the Mount
and Unmount buttons are disabled for Hyper-V backups.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

In the OnCommand console Backups tab, deleting a mirror source backup copy prevents you from
mounting its partner mirror destination backup copy. For a Mirror-generated destination backup copy
to be mountable, its associated mirror source backup copy must still exist on the source node.
Steps

1. Click the View menu, then click the Server option.


2. In the Server tab, click the VMware option, then click VMware VMs or Datastores.
3. Select a virtual machine or datastore and click Mount.
You cannot mount Hyper-V backups using this button.
4. In the Mount Backup dialog box, select an unmounted backup that you want to mount.
You can only mount one backup each time and you cannot mount a mounted backup.
5. Select from the drop-down list the name of the ESX server to which you want to mount the
backup.
6. Click Mount.
A dialog box appears with a link to the mount job and when you click the link, the Jobs tab
appears.

Backups | 287
After you finish

You can monitor the status of your mount and unmount jobs in the Jobs tab.
Related references

Administrator roles and capabilities on page 506


Guidelines for mounting or unmounting backups in a VMware environment on page 271
Unmounting backups in a VMware environment from the Server tab
After you are done using a mounted backup for verification or to restore a virtual machine to an
alternate location, you can unmount the mounted backup from the ESX server that it was mounted to.
When you unmount a backup, all the datastores in that backup are unmounted and can no longer be
seen from the ESX server that you specify. Both the Mount and Unmount buttons are disabled for
Hyper-V backups.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

If there are virtual objects in use from the previously mounted datastores of a backup, the unmount
operation fails. You must manually clean up the backup prior to mounting the backup again because
its state reverts to not mounted.
If all the datastores of the backup are in use, the unmount operation fails but this backup's state
changes to mounted. You can unmount the backup after determining the datastores are not in use.
Steps

1. Click the View menu, then click the Server option.


2. In the Server tab, click the VMware option, then click VMware VMs or Datastores.
3. Select a virtual machine or datastore and click Unmount.
4. In the Unmount Backup dialog box, select a mounted backup to unmount.
5. Click Unmount.
6. At the confirmation prompt, click Yes.
A dialog box opens with a link to the unmount job displays and when you click the link, the Jobs
tab appears.

288 | OnCommand Console Help


After you finish

If the ESX server becomes inactive or reboots during an unmount operation, the job is terminated and
the mount state remains mounted and the backup stays mounted on the ESX server.
You can monitor the status of your mount and unmount jobs in the Jobs tab.
Related references

Administrator roles and capabilities on page 506


Guidelines for mounting or unmounting backups in a VMware environment on page 271

Manually mounting or unmounting backups in a Hyper-V environment


using SnapDrive for Windows
You can use SnapDrive for Windows to assist you in manually mounting or unmounting a Hyper-V
Snapshot copy so that you can verify the contents of the backup prior to restoring it.
You must have SnapDrive for Windows installed on your Hyper-V parent host to use the manual
mount process.
For more information on connecting and disconnecting LUNs, see the SnapDrive for Windows
documentation.
Locating specific backups
You can locate a specific backup copy by searching for different criteria such as names and
descriptions. After you locate a backup, you can then restore it, view the status of it, or delete it.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

You can locate a specific backup copy by searching one of the following criteria:

Whole or partial backup description


Partial current name of a primary object in the backup
Partial name of a virtual object when the backup was taken
UUID of a virtual object within the backup

Steps

1. Click the View menu, then click the Backups option.


2. From the Backups tab, locate the Search field, and type in all or part of the backup description or
version.

Backups | 289
You can locate multiple backup versions by inserting a comma between search terms and you can
clear the search field to view all backups.
3. Click Find.
Related references

Administrator roles and capabilities on page 506


How to find Snapshot copies within a backup
You can locate a specific Snapshot copy within a backup so that you can use its name to connect to a
LUN within the copy using SnapDrive for Windows.
After you have located a specific backup, you can view the virtual machines that belong to it. If you
select a virtual machine, you can view the list of associated Snapshot copies and VHDs. By selecting
a Snapshot copy and VHD, you can use the following information to connect the Snapshot copy to a
LUN using SnapDrive for Windows:

Snapshot copy name


Primary storage system name
Volume name
Mount point
VHD LUN path

Two Snapshot copy names are displayed for a Hyper-V backup. You must choose the Snapshot copy
with the suffix _backup to mount the backup. This ensures that you select the copy containing
application-consistent data.
Connecting to a LUN in a Snapshot copy
You can connect to a LUN in a Snapshot copy using either a FlexClone volume or a read/write
connection to a LUN in a Snapshot copy depending on what version of Data ONTAP you have
installed on your storage system.
Before you begin

You must have the FlexClone license enabled to connect to a LUN that resides on a volume with a
SnapMirror or SnapVault destination.
Steps

1. Under SnapDrive in the left MMC pane, expand the instance of SnapDrive you want to manage,
then expand Disks and select the disk you want to manage.
2. Expand the LUN whose Snapshot copy you want to connect, then click on Snapshot Copies to
display the list of Snapshot copies. Select the Snapshot copy you want to connect.

290 | OnCommand Console Help


Note: If you cannot see the Snapshot copy list, make sure that cifs.show_snapshot is set to
on and vol options nosnapdir is set to off on your storage system.

3. From the menu choices at the top of MMC, navigate to Action > Connect Disk to launch the
Connect Disk wizard.
4. In the Connect Disk Wizard, click Next.
5. In the Provide a Storage System Name, LUN Path and Name panel, the information for the
LUN and Snapshot copy you selected is automatically filled in. Click Next.
6. In the Select a LUN Type panel, Dedicated is automatically selected because a Snapshot copy
can be connected only as a dedicated LUN. Click Next.
7. In the Select LUN Properties panel, either select a drive letter from the list of available drive
letters or type a volume mount point for the LUN you are connecting, then click Next.
When you create a volume mount point, type the drive path that the mounted drive will use: for
example, G:\mount_drive1\.
8. In the Select Initiators panel, select the FC or iSCSI initiator for the LUN you are connecting
and click Next.
9. In the Select Initiator Group management panel, specify whether you will use automatic or
manual igroup management.
If you specify...

Then...

Automatic igroup
management

Click Next.

Manual igroup
management

Click Next, and then perform the following actions:

SnapDrive uses existing igroups, one igroup per initiator, or, when necessary, creates
new igroups for the initiators you specified in the Select Initiators panel.

a. In the Select Initiator Groups panel, select from the list the igroups to which you
want the new LUN to belong.
Note: A LUN can be mapped to an initiator only once.
OR
Click Manage Igroups and for each new igroup you want to create, type a name
in the Igroup Name text box, select initiators from the initiator list, click Create,
and then click Finish to return to the Select Initiator Groups panel.
b. Click Next.

10. In the Completing the Connect Disk Wizard panel, perform the following actions.
a. Verify all the settings
b. If you need to change any settings, click Back to go back to the previous Wizard panels.
c. Click Finish.

Backups | 291
Result

The newly connected LUN appears under Disks in the left MMC pane.
Viewing the contents of a LUN
You can view the contents in a LUN, including VHDs and other files. Viewing the contents of the
LUN enables you to confirm that you have the correct data before performing a restore operation on
the whole Hyper-V backup.
Before you begin

You must have installed Windows 2008 R2 and the Windows Disk Management Snap-In.
You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. To view the contents of a specific VHD, use Windows Explorer to locate the VHD in the LUN
you mounted using SnapDrive for Windows.
2. Using the Windows Disk Management Snap-In, right-click the VHD and select Attach VHD.
3. Specify the VHD location and click OK.
The VHD is mounted on the Hyper-V parent host.
4. Verify the contents of the VHD.
5. Using the Windows Disk Management Snap-In, right-click the VHD and select Detach VHD.
After you finish

You can now disconnect the LUN using SnapDrive for Windows to unmount the disk mounted from
the Snapshot copy.
Disconnecting a LUN
You can use the SnapDrive for Windows MMC snap-in to disconnect a dedicated or shared LUN, or
a LUN in a Snapshot copy or in a FlexClone volume.
Before you begin

Make sure that neither Windows Explorer nor any other Windows application is using or
displaying any file on the LUN you intend to disconnect. If any files on the LUN are in use, you
will not be able to disconnect the LUN except by forcing the disconnect.
If you are disconnecting a disk that contains volume mount points, change, move, or delete the
volume mount points on the disk first before disconnecting the disk containing the mount points;
otherwise, you will not be able to disconnect the root disk. For example, disconnect G:
\mount_disk1\, then disconnect G:\.

292 | OnCommand Console Help

Before you decide to force a disconnect of a SnapDrive LUN, be aware of the following
consequences:

Any cached data intended for the LUN at the time of forced disconnection is not committed to
disk.
Any mount points associated with the LUN are also removed.
A pop-up message announcing that the disk has undergone "surprise removal" appears in the
console session.

About this task

Under ordinary circumstances, you cannot disconnect a LUN that contains a file being used by an
application such as Windows Explorer or the Windows operating system. However, you can force a
disconnect to override this protection. When you force a disk to disconnect, it results in the disk
being unexpectedly disconnected from the Windows host.
Steps

1. Under SnapDrive in the left MMC pane, expand the instance of SnapDrive you want to manage,
then expand Disks and select the disk you want to manage.
2. From the menu choices at the top of MMC, navigate to either Action > Disconnect Disk to
disconnect normally, or Action > Force Disconnect Disk to force a disconnect.
3. When prompted, click Yes to proceed with the operation.
Note: This procedure will not delete the folder that was created at the time the volume mount

point was added. After you remove a mount point, an empty folder will remain with the same
name as the mount point you removed.
The icons representing the disconnected LUN disappear from both the left and right MMC
panels.

Locating specific backups


You can locate a specific backup copy by searching for different criteria such as names and
descriptions. After you locate a backup, you can then restore it, view the status of it, or delete it.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

You can locate a specific backup copy by searching one of the following criteria:

Whole or partial backup description


Partial current name of a primary object in the backup

Backups | 293

Partial name of a virtual object when the backup was taken


UUID of a virtual object within the backup

Steps

1. Click the View menu, then click the Backups option.


2. From the Backups tab, locate the Search field, and type in all or part of the backup description or
version.
You can locate multiple backup versions by inserting a comma between search terms and you can
clear the search field to view all backups.
3. Click Find.
Related references

Administrator roles and capabilities on page 506

Deleting a dataset backup


You can delete one or more dataset backups if you do not want to wait for the retention period
associated with the dataset to expire. You can not delete a mounted backup. You must first unmount
the backup so that you can delete it.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

In the OnCommand console Backups tab, deleting a mirror source backup copy prevents you from
mounting its partner mirror destination backup copy. For a Mirror-generated destination backup copy
to be mountable, its associated mirror source backup copy must still exist on the source node.
Steps

1. Click the View menu, then click the Backups option.


2. In the Backups tab, choose the backup version that you want to delete.
3. Click Delete.
Related references

Administrator roles and capabilities on page 506

294 | OnCommand Console Help

Monitoring backups
Monitoring local backup progress
You can monitor the progress of your backup job to see whether it is running, succeeded, or failed.
Steps

1. Click the View menu, then click the Jobs option.


2. Select the backup job that you want to monitor.
You can view information about the backup in the Details pane.
Related references

Administrator roles and capabilities on page 506

Page descriptions
Backups tab
The Backups tab displays scheduled and on-demand backup versions that contain virtual objects, but
does not display storage backup versions. You can view detailed information about each virtual
object backup version and restore virtual objects.

Command buttons on page 294


Backups list on page 295
Restorable Entities on page 295
Resource Properties on page 295

Command buttons
Edit

Enables you to modify the backup description.

Delete

Deletes the dataset backup. The Delete button is disabled for mounted backups.

Mount

Enables you to mount a selected VMware backup to an ESX server if you want to
verify its content before restoring it.

Unmount

Enables you to unmount a VMware backup after you mount it on an ESX server and
verify its contents.

Refresh

Updates the backups list.

Backups | 295
Search field Enables you to search for a backup by description, resource name, or vendor object
ID.
Find

Executes the search to find the specified backup.

Clear

Clears the search and returns to default backup list.

Backups list
Backup ID

The unique identifier of the backup job.

Dataset

The name of the dataset to which the backup belongs.

Version

The time at which the backup was taken.

Description

Description of the backup.

Retention Duration The length of time the backup is retained.


Node (Name)

Specifies whether the backup is local or remote, as well as backup and mirror
information. Values are Local, Local (Primary data) for Hyper-V, Remote
(Backup), and Remote (Mirror).

VMware Snapshot Displays as Yes if the backup contains a VMware snapshot, No if does not,
and Not Applicable if it contains Hyper-V virtual objects.
Mount State

Specifies the mount state of the VMware backup. Values are Not Mounted,
Mounted, Mounting, and Unmounting.

Type

Specifies whether the backup contains VMware or Hyper-V virtual objects.

Restorable Entities
Restore button

Restores the specified resource.

Resource

Displays the virtual machines or datastores that belong to the specified backup
in the Backups list.

Vendor Object ID The universal ID assigned to a virtual object by VMware or Hyper-V.


Resource Properties
Partial backup

Specifies whether the resource selected is completely or partially backed up.


The Entire virtual machine button and the Start virtual machine after restore
check box are disabled.

Datastore type

Displays the type of datastore. Values are VMFS or NFS. If the virtual
machine resides on an NFS datastore, and belongs to a local backup, the ESX
Host Name field is disabled.

Is Data ONTAP

Specifies whether the resource is stored on Data ONTAP storage or not.

296 | OnCommand Console Help


Snapshot taken for Specifies whether a Snapshot copy of the resource was created.
VM
Clustered

Specifies whether the resource is part of a cluster.

Is template

Specifies whether the resource is a template. The Start virtual machine after
restore check box is disabled.

State

Displays the state of the resource.

The Groups drop-down list is not applicable to the Manage Backups window.
Related references

Window layout customization on page 16

297

Restore
Understanding restore
Restoring data from backups
The OnCommand console allows you to restore your virtual machines and datastores from legacy
backups and from backups taken of newly created datasets that contain your virtual machines and
datastores. The OnCommand console supports restore from local and remote backups and from
backups that contain VMware-based snapshots.
Backup selection
The OnCommand console allows you to browse for backups to restore from on the Backups tab
panel or the Server tab when determining which datastore, virtual machine or virtual disk files in the
virtual machine to restore. When you select a datastore, all virtual machines in the datastore will be
restored.
The table of backups provides centralized management of VMware and Hyper-V entities. You can
filter the table to show only backups with a backup ID, dataset, description, node name (full or
partial), VMware snapshot, mount state, or resource name of a datastore or virtual machine.
If you search for a comma-separated list, all backups that have either a matching backup ID or
backup name will display. You can specify multiple backup names and IDs by entering them in a
comma-separated list. The result lists all of the backups that have either one of the matching criteria
by entering them in a comma-separated list. The result lists all of the backups that have either one of
the given names or IDs.
Related tasks

Locating specific backups on page 288


Where to restore a backup
OnCommand console allows you to select a restore destination from the Restore Wizard.

Original location
The backup of an entire datastore, a single virtual machine, or a virtual machine's disk files are
restored to the original location. You set this location by choosing The entire virtual machine
option.
Different location
The backup of the virtual machine disk files are restored to a different location. You select the
destination location by setting the Particular virtual disks option.

298 | OnCommand Console Help


Note: If you are recovering data from a remote backup, data access from the ESX server to a
Hyper-V controller using FC, iSCSI, or NFS protocols must exist to ensure that an NFS mount can
be done from a secondary node to the ESX server.

Restore scripting information


You can specify a script that is invoked before and after a restore operation. The script is invoked on
the host service and the path is local to the host service.
If you would like to configure a script to run before or after your restore operation, you must first run
the script independently from the DOS command line to determine if the script generates any pop-up
messages. Pop-up messages cause the restore operation to not complete properly. Clear out the
condition that generates the user pop-up message to ensure that the script runs automatically without
requiring any user interaction.
Restore script conventions
If you use a PowerShell script, you should use the drive-letter convention. For other types of scripts,
you can use either the drive-letter convention or the UNC. When using UNC (Universal Naming
Convention) for the script path, the script executes on the system running the host service.
Pre-restore scripting arguments
Note: If the pre-backup script fails, the backup might fail.

The following arguments apply to scripts that run before the restore operation occurs:
prerestore

Indicates that the script is invoked prior to the restore operation.

resourceids

Species the colon-separated list of resource IDs that are backed up.

backupid

Specifies the backup identifier.

snapshots

Specifies the comma-separated list of Snapshot copies that constitute the


backup. You should use one of the following formats:

storage system:/vol/volx:snapshot
storage system:/vol/volx/lun:snapshot
storage system:/vol/volx/lun/qtree:snapshot

secondarysnapshots Specifies the comma-separated list of secondary snapshot copies that


constitute the backup. This argument applies only for a restore from a
secondary backup.
Post-restore scripting arguments
The following arguments apply to scripts that run after the restore operation occurs:
postbackup

Indicates that the script is invoked after the restore operation.

Restore | 299
resourceids

Specifies the colon-separated list of resource IDs that are being backed up.

backupid

Specifies the backup identifier.

Managing restore
Restoring data from backups created by the OnCommand console
You can restore a datastore, virtual machine, or its disk files to its original location or an alternate
location. From the Backup Management panel, you can sort the backup listings by vendor type to
help you find your backups.
From the Backup Management panel, you can do the following:

Restore a datastore, virtual machine, or its disk files from a local or remote backup to an original
location
Restore virtual machine disk files from a local or remote backup to a different location.
Restore from a backup that has a VMware snapshot.

Restoring a datastore
You can use the OnCommand console to restore a datastore. By doing so, you overwrite the existing
content with the backup you select.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

If you use a PowerShell script, you should use the drive letter convention. For other types of scripts,
you can use either the drive letter convention or the Universal Naming Convention.
Steps

1. Click the View menu, then click the Backups option.


2. In the Backups tab, select a backup of the datastore.
3. Click Restore.
Restoring a datastore powers off all virtual machines.
4. In the Restore dialog box, click Restore.

300 | OnCommand Console Help

Restoring a VMware virtual machine


You can use the OnCommand console to recover a VMware virtual machine from a local or remote
backup. By doing so, you overwrite the existing content with the backup you select.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

The process for restoring a VMware virtual machine differs from restoring a Hyper-V virtual
machine in that you can restore an entire virtual machine or its disk files. Once you start the
restoration, you cannot stop the process, and you cannot restore from a backup of a virtual machine
after you delete the dataset the virtual machine belonged to.
If you use a PowerShell script, you should use the drive letter convention. For other types of scripts,
you can use either the drive letter convention or the Universal Naming Convention.
Steps

1. Click the View menu, then click the Backups option.


2. In the Backups tab, click Type and Node (Name) to sort the backup table by VMware virtual
machines and by local or remote backups.
3. Select a backup, and then select the virtual machine from the list of restorable entities.
4. Click Restore.
The Restore wizard opens and lists all backups which include the virtual machine.
5. Select one of the following recovery options:
Option

Description

Entire virtual
machine

Restores the contents of your virtual machine from a Snapshot copy to its original
location. The Start virtual machine after restore checkbox is enabled if you select
this option and the virtual machine is registered.

Particular virtual Restores the contents of the virtual disks on a virtual machine to a different location.
This option is enabled if you uncheck the entire virtual machine option. You can set a
disks
Destination datastore for each virtual disk.

6. In the ESX host name field, select the name of the ESX host. The ESX host is used to mount the
virtual machine components.
This option is available if you want to restore virtual disk files or the virtual machine is on a
VMFS datastore.

Restore | 301
7. In the Pre/Post Restore Script Path field, type the name of the script that you want to run before
or after the restore operation.
8. From this wizard, click Restore to begin the restoration.
Related tasks

Adding a virtual machine to inventory on page 59


Where to restore a backup on page 297
Restoring a Hyper-V virtual machine
You can use the OnCommand console to recover a Hyper-V virtual machine from a local or remote
backup. By doing so, you overwrite the existing content with the backup you select.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Before attempting a restore operation on a Hyper-V virtual machine, you must ensure that
connectivity to the storage system exists. If there is no connectivity, the restore operation fails.
About this task

If you use a PowerShell script, you should use the drive letter convention. For other types of scripts,
you can use either the drive letter convention or the Universal Naming Convention.
If storage system volumes are renamed, you cannot restore a virtual machine created prior to
renaming the volumes.
Steps

1. Click the View menu, then click the Backups option.


2. In the Backups tab, click Type and Node (Name) to sort the backup table by VMware or HyperV virtual machines, or local and remote, backups.
3. Select a dataset that has the backup you want, and then select the virtual machine from the list of
backed-up entities.
4. Click Restore.
The Restore wizard opens.
5. Select one of the following recovery options:
Option

Description

Start VM after restore

Restores the contents of your virtual machine from a Snapshot copy and
restarts the virtual machine after the operation completes.

302 | OnCommand Console Help


Option

Description

Pre/Post Restore Script: Runs a script that is stored on the host service before or after the restore
operation.

The Restore wizard displays the location of the virtual hard disk (.vhd) file.
6. From this wizard, click Restore to begin the restoration.

Monitoring restore
Viewing restore job details
After you start the restore job, you can track the progress of the restore job from the Jobs tab and
monitor the job for possible errors.
Steps

1. Click the View menu, then click the Jobs option.


2. From the Jobs tab, select Restore from the Jobs type field.

303

Reports
Understanding reports
Reports management
You can print, export, and share data in the reports that are generated by the OnCommand console.
You can also schedule a report and send the report schedule to one or more users.
If you want to do this...
Print a report

Export a report

Export data in a report

Specify parameter values

Perform this action...


Select the report, click the toolbar icon (
and select the Print option.

),

Select the report, click the toolbar icon (


and select the Export Content option.

),

Select the report, click the toolbar icon (


and select the Export Data option.

),

),
Select the report, click the toolbar icon (
and select the Parameters option.
This option is valid only for the Events reports.

Save a report with the existing name

Make changes to a report, click Save, and select


the Save option.
This option is enabled only for custom reports.

Save a report with a new name

Make changes to a report, click Save, and select


the Save As option.

Share a report

Select the report, and then click Share.

Schedule a report

Select the report, and then click Schedule.

Delete a report

Select the report, and then click Delete.


This option is enabled only for custom reports.

304 | OnCommand Console Help

Reports tab customization


The OnCommand console enables you to customize the data in a report and the report layout.
If you want to do this...

Perform this action...

Edit column labels

Select the column, click the arrow on the right, and click
the Header option.

Organize data into groups

Select the column, click the arrow on the right, and click
Group > Add Group.

Remove groups

Select the column, click the arrow on the right, and click
Group > Delete Inner Group.
This option is displayed only when data is organized into
groups.

Hide group details

Select the column, click the arrow on the right, and click
Group > Hide Detail.
This option is displayed only when data is organized into
groups.

Show group details

Select the grouped column, click the arrow on the right,


and click Group > Show Detail.
This option is displayed only when the group details are
hidden.

Start each group on a new page

Select the column, click the arrow on the right, and click
Group > Page Break.

Hide columns

Select the column, click the arrow on the right, and click
Column > Hide Column.

Show hidden columns

Select the column, click the arrow on the right, and click
Column > Show Column.

Delete a column

Select the column, click the arrow on the right, and click
Column > Delete Column.

Compute a column

Select the column, click the arrow on the right, and click
Column > New Computed Column.

Reorder columns

Select the column, click the arrow on the right, and click
Column > Reorder Columns.

Hide duplicate values in a column

Select the column, click the arrow on the right, and click
Column > Do Not Repeat Values.

Reports | 305
If you want to do this...

Perform this action...

Redisplay duplicate values in a


column

Select the column, click the arrow on the right, and click
Column > Repeat Values.

Aggregate data

Select the column, click the arrow on the right, and click
Aggregation.

Filter data

Select the column, click the arrow on the right, and click
Filter > Filter.

Sort data

Select the column, click the arrow on the right, and click
Sort.

Format a column

Select the column, click the arrow on the right, and click
Format > Font.

Format data in a report based on


conditions

Select the column, click the arrow on the right, and click
Format > Conditional Formatting.

Types of object status


The physical or logical objects can display one of six status types: Normal, Warning, Error, Critical,
Emergency, or Unknown. An object's status alerts you to when you must take action to prevent
downtime or failure.
Normal

The object is operating within the desired thresholds.

Warning

The object experienced an occurrence that you should be aware of.

Error

The object is still performing without service disruption, but its performance might be
affected.

Critical

The object is still performing but service disruption might occur if corrective action is
not taken immediately.

Emergency The object unexpectedly stopped performing and experienced unrecoverable data loss.
You must take corrective action immediately to avoid extended downtime.
Unknown

The object is in an unknown transitory state. This status is displayed only for a brief
period.

What report scheduling is


The scheduler service in the DataFabric Manager server enables you to schedule a report to be
generated at a specific date and time. The report is sent by e-mail message to one or more users at the
specified date and time.
Note: If you change the system time when the DataFabric Manager server scheduler service is
running, schedules configured for a report might not work as expected. To ensure that the service
functions correctly, you must stop and start the service using the dfm service command.

306 | OnCommand Console Help

Managing reports
Scheduling reports
You can use the Reports tab to schedule reports to be generated and sent by e-mail message to one or
more users on a recurring basis at a specified date and time. For example, you can schedule a report
to be sent as e-mail, in the HTML format, every Monday.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu, then click the Reports option.


2. In the navigation pane, select the report for which you want to create a schedule, and then click
Schedule.
3. In the Schedule Report dialog box, specify the e-mail address or the alias of the administrator or
the user to whom you want to send the report schedule.
The E-mail field is mandatory. You can specify one or more entries, separated by commas.
4. Specify other report schedule parameters, such as the format, frequency, and scope.
The scope specifies the group, storage system, resource pool, or quota user for which the report is
generated.
Note: For groups, you can navigate only to the direct members of a group.

5. Click Apply, then click Ok.


Result

The new schedule for a report is saved in the DataFabric Manager server, and is displayed in the
Saved Settings.
Related references

Administrator roles and capabilities on page 506

Reports | 307

Viewing the scheduled reports log


You can view the scheduled reports log from the Scheduled Log tab. The log provides details about a
scheduled report's status, run time, and the reason for any failed schedule.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu, then click the Reports option.


2. In the Reports tab, click the Scheduled Log tab.
The list of scheduled reports are displayed in a tabular format.
3. Select a scheduled report to view the details.
Related references

Administrator roles and capabilities on page 506

Sharing reports
You can share a report with one or more users. The report is sent by e-mail message to the specified
users instantly.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

If you customize a report, you should save the changes before you share it. If not, when you share the
report, the changes are not displayed. You can save changes made to a custom report without
changing the report's current name. For detailed reports, you can save the changes with a new report
name.
Steps

1. Click the View menu, then click the Reports option.


2. In the navigation pane, select the report you want to share, and then click Share.
3. In the Share Report dialog box, specify the e-mail address or the alias of the administrator or the
user with whom you want to share the report.

308 | OnCommand Console Help


The E-mail field is mandatory. You can specify one or more entries, separated by commas.
4. Specify other report sharing parameters, such as subject, format, and scope.
By default, the name of the report is displayed as the subject. The scope specifies the group,
storage system, resource pool, or quota user for which the report is generated.
Note: For groups, you can navigate only to the direct members of a group.

5. Click Ok.
Related references

Administrator roles and capabilities on page 506

Deleting a report
You can delete one or more custom reports when they are no longer necessary.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu, then click the Reports option.


2. In the navigation pane, select Custom Reports.
3. Select one or more reports that you want to delete, and then click Delete.
Related references

Administrator roles and capabilities on page 506

Page descriptions
Reports tab
The Reports tab enables you to view detailed information about the reports that you generate. You
can search for a specific report, save a detailed report as a custom report, and delete a custom report.
You can also share and schedule a report.
Reports tab details
You can view the following information in the Viewer tab:

Reports | 309
Navigation tree Displays detailed and custom reports along with the available subcategories (for
example, Events, Inventory, and Storage capacity). You can click the name of a
report in the navigation tree to view the report contents in the reporting area.
Search filter

Enables you to search for a specific report by entering the report name in the
search filter.

Toolbar

Enables you to navigate to specific pages of the report, export the report, undo or
redo your actions, and so on.

Reporting area Displays, in tabular format or as a combination of charts and tabular data, the
contents of a selected report.
Command buttons
The command buttons enable you to perform the following tasks for a selected report:
Save

Save
Saves the changes made to a custom report without changing the report's current
name.
Note: This option is disabled for detailed reports.

Delete

Save As
Displays the Save As dialog box, which enables you to save changes made to a
report and to give the report a new name.

Enables you to delete one or more custom reports when they are no longer necessary.
Note: This option is disabled for detailed reports.

Schedule Enables you to schedule the report to be generated on a recurring basis at a specified
date and time. The report is sent by e-mail message to one or more users at the specified
date and time.
Share

Enables you to share a report with one or more users instantly. The report is sent by email message to the specified users instantly.

Refresh

Refreshes the contents of the current report.

Note: You can also save, delete, schedule, or share a report by right-clicking the selected report in
the navigation tree.

Scheduled Log tab


You can view the list of logs for scheduled reports by clicking the Scheduled Log tab. You can view
details such as the status and run time of a schedule, and the reason for any schedule failure. The
details are displayed in the details pane.

310 | OnCommand Console Help


Operations Manager Reports link
In addition to the reports displayed in the OnCommand console, you can view other reports by
clicking the Operations Manager Reports link. During this task, the OnCommand console launches
the Operations Manager console. Depending on your browser configuration, you can return to the
OnCommand console by using the Alt-Tab key combination or clicking the OnCommand console
browser tab. After the completion of this task, you can leave the Operations Manager console open,
or you can close it to conserve bandwidth.

Schedule Report dialog box


The Schedule Report dialog box enables you to schedule the report to be generated on a recurring
basis at a specified date and time. The report is sent by e-mail message to one or more users at the
specified date and time.

Saved Settings on page 310


Properties on page 310
Command buttons on page 310

Saved Settings
Saved Settings displays all of the schedules for a report. You can create a new schedule by selecting
the New option. You can rename or delete a schedule by right-clicking the schedule and choosing the
appropriate option. You can also modify a schedule by selecting the schedule and making the
required changes.
Properties
E-mail

Specifies either the e-mail address or the alias of the administrator or the user to whom
you want to send the report schedule. You can specify one or more entries, separated
by commas. This is a mandatory field.

Format

Specifies the format in which you want to schedule the report. The HTML option is
selected by default.

Frequency Specifies the frequency at which you want to schedule the report. The Hourly at
Minute option is selected by default.
Scope

Specifies the group, storage system, resource pool, or quota user for which the report is
generated.
Note: For groups, you can navigate only to the direct members of a group.

Command buttons
The command buttons enable you to perform the following tasks:
Browse Enables you to browse through the available resources. The resource you select then
defines the scope of the generated report.

Reports | 311
Apply

Updates the properties that you specify for a report schedule.

Ok

Updates the properties that you specify for a report schedule, and closes the Schedule
Report dialog box.

Cancel Enables you to undo the schedule report configuration and closes the Schedule Report
dialog box.

Share Report dialog box


The Share Report dialog box enables you to e-mail a report to one or more users instantly. If you
customize a report, you should save the changes before you share it. If not, when you share the
report, the changes are not displayed.

Properties on page 311


Command buttons on page 311

Properties
E-mail Specifies either the e-mail address or the alias of the administrator or the user with whom
you want to share the report. You can specify one or more entries, separated by commas.
This is a mandatory field.
Subject Specifies the subject of the e-mail. By default, the name of the report is displayed.
Format Specifies the format in which you want to share the report. The HTML option is selected
by default.
Scope

Specifies the group, storage system, resource pool, or quota user for which the report is
generated.
Note: For groups, you can navigate only to the direct members of a group.

Command buttons
The command buttons enable you to perform the following tasks:
Browse Enables you to browse through the available resources. The resource you select then
defines the scope of the generated report.
Ok

Updates the share report properties that you specify.

Cancel Enables you to undo the share report configuration and closes the Share Report dialog
box.

Events reports

312 | OnCommand Console Help

Understanding events reports


What event reports are
Event reports provide information about the name, severity, status, and cause of an event. By default,
the report displays events with severity level Warning, Error, Critical, and Emergency. You can filter
the reports based on severity, status, and so on, and save the result as a new report.
The event reports display the following information:

Severity of the event


Time that the event was triggered
Status of the event
The name of the administrator who acknowledged the event
The date and time when the event was acknowledged

Related references

Description of event severity types on page 31

Page descriptions
Events Current report
The Events Current report displays information about events that are current and yet to be resolved.
The information includes the name, severity, and cause of the event.
Note: To display the charts and icons in a report, you must ensure that a DNS mapping is
established between the client machine from which you are starting the Web connection and the
host name of the system on which the DataFabric Manager server is installed.

Chart
You can customize the data displayed in the chart, change the subtype and format of the chart, and
export the data from the chart.
Report details
Severity

Displays the severity of the event.


By default, the report displays events with severity level Warning, Error,
Critical, and Emergency. To view events with severity level Normal and
Information, you can select the Parameters option from the toolbar, clear the
prefilter, and run the report.

Event

Displays the event name.

Reports | 313
Triggered On

Displays the date and time when the event was generated.

Acknowledged By Displays the name of the administrator who acknowledged the event.
Acknowledged

Displays the date and time when the event was acknowledged.

Source

Displays the name of the object with which the event is associated.

Related references

Description of event severity types on page 31


Reports tab customization on page 304
Reports management on page 303
Events All report
The Events All report displays, in tabular format, information about all the events, including resolved
and unresolved events. The information includes the name, severity, and cause of the event, and
details about the resolved event.
Note: To display the icons in a report, you must ensure that a DNS mapping is established between

the client machine from which you are starting the Web connection and the host name of the
system on which the DataFabric Manager server is installed.
Severity

Displays the severity of the event.


By default, the report displays events with severity level Warning, Error,
Critical, and Emergency. To view events with severity level Normal or
Information, you can select the Parameters option from the toolbar, clear the
prefilter, and run the report.

Event

Displays the event name.

Triggered On

Displays the date and time when the event was generated.

Acknowledged By Displays the name of the administrator who acknowledged the event.
Acknowledged

Displays the date and time when the event was acknowledged.

Source

Displays the name of the object with which the event is associated.

Resolved On

Displays the date and time when the event was resolved.

Resolved By

Displays the name of the administrator who resolved the event.

Related references

Description of event severity types on page 31


Reports tab customization on page 304
Reports management on page 303

314 | OnCommand Console Help

Inventory reports
Understanding inventory reports
What inventory reports are
Inventory reports provide information about objects such as aggregates, volumes, qtrees, and LUNs.
Inventory report types are as follows:

Aggregates report
File Systems report
LUNs report
Qtrees report
Storage Systems report
vFiler Units report
Volumes report
Storage Services report
Storage Service Policies report
Storage Service Datasets report

Page descriptions
Aggregates report
The Aggregates report displays information such as the type, state, and status of the aggregate, and
the SnapLock feature in the aggregate.
Note: To display the icons in a report, you must ensure that a DNS mapping is established between
the client machine from which you are starting the Web connection and the host name of the
system on which the DataFabric Manager server is installed.

Aggregate

Displays the name of the aggregate.

Storage
System

Displays the name of the storage system that contains the aggregate.

Type

Displays the type of the aggregate:


Traditional

An aggregate that contains only one traditional volume. A


traditional aggregate is created automatically in response to a
traditional volume create request.

Reports | 315

Aggregate

An aggregate that contains one or more flexible volumes, RAID


groups, and disks.

Striped
Aggregate

An aggregate that is striped across multiple nodes, allowing its


volumes to also be striped across multiple nodes. Striped
aggregates are made up of multiple member aggregates.
Member aggregates are the aggregates of actual cluster nodes,
which form a striped aggregate.

Block Type

Displays the block format of the aggregate as 32_bit or 64_bit. By default, this
column is hidden.

RAID

Displays the RAID protection scheme. The RAID protection scheme can be one of
the following:
raid0

All the raid groups in the aggregate are of type raid0.

raid4

All the raid groups in the aggregate are of type raid4.

raid_dp

All the raid groups in the aggregate are of type raid_dp.

mixed_raid_type The aggregate contains RAID groups of different RAID types


(raid0, raid4, and raid_dp).
Note: Data ONTAP uses raid4 or raid_dp protection to ensure data integrity
within a group of disks even if one or two of those disks fail.

State

Displays the current state of the aggregate. An aggregate can be in one of the
following three states:
Offline

Read or write access is not allowed.

Restricted Some operations such as parity reconstruction are allowed, but data
access is not allowed.
Online

Read and write access to volumes hosted on this aggregate are


allowed.

Status

Displays the current status of the aggregate based on the events generated for the
aggregate. The status can be Normal, Warning, Error, Critical, Emergency, or
Unknown.

Mirrored

Displays whether the aggregate is mirrored.


Note: For an aggregate to be enabled for mirroring, the storage system must have

a SyncMirror license for syncmirror_local or cluster_remote installed and


enabled, and the storage system's disk configuration must support RAID-level
mirroring.

316 | OnCommand Console Help


SnapLock
Type

Displays the type of the SnapLock feature (if it is enabled) used in the aggregate:
SnapLock
Compliance

Provides WORM protection of files and restricts the storage


administrator's ability to perform any operations that might
modify or erase retained WORM records.

SnapLock
Enterprise

Provides WORM protection of files, but uses a trusted


administrator model of operation that allows the storage
administrator to manage the system with few restrictions.

No

Indicates the SnapLock feature is disabled.

Related references

Types of object status on page 305


Reports tab customization on page 304
Reports management on page 303
File Systems report
The File Systems report displays information about the volumes and qtrees in a storage server. You
can monitor and manage only qtrees created by the user. Therefore, the default qtree, qtree 0, is not
monitored or managed.
Note: To display the icons in a report, you must ensure that a DNS mapping is established between

the client machine from which you are starting the Web connection and the host name of the
system on which the DataFabric Manager server is installed.
Type

Displays whether the volume or qtree is clustered or nonclustered.

File System

Displays the name and path of the storage object.

Storage
Server

Displays the name of the storage server. The storage server can be a storage
controller, Vserver, or a vFiler unit that contains the volume.

Status

Displays the current status of the volume based on the events generated for a
specific volume, qtree, or LUN. The status can be Normal, Warning, Error,
Critical, Emergency, or Unknown.

Related references

Types of object status on page 305


Reports tab customization on page 304
Reports management on page 303

Reports | 317

LUNs report
The LUNs report displays information about the LUNs and LUN initiator groups in your storage
server.
Note: To display the icons in a report, you must ensure that a DNS mapping is established between
the client machine from which you are starting the Web connection and the host name of the
system on which the DataFabric Manager server is installed.

LUN Path

Displays the path name of the LUN (volume or qtree that contains the LUN).

Initiator Group
(LUN ID)

Displays the name of the initiator group to which the LUN is mapped.

Description

Displays the description (comment) that you specified when creating the LUN
on your storage server. By default, this column is hidden.

Size (GB)

Displays the size of the LUN.

Storage Server

Displays the name of the storage server that contains the LUN. The storage
server can be a storage controller or a vFiler unit.

Status

Displays the current status of the LUN. The status can be Normal, Warning,
Error, Critical, Emergency, or Unknown. The status of a LUN is determined
by the DataFabric Manager server based on the information that it obtains
from the storage controller or the vFiler unit in which the LUN exists. For
example, if the storage controller or the vFiler unit reports that a LUN is
offline, the DataFabric Manager server displays the status of the LUN as
Warning.

Read/Sec (Bytes)

Displays the rate of bytes (number of bytes per second) read from the LUN.
By default, this column is hidden.

Write/Sec (Bytes)

Displays the rate of bytes (number of bytes per second) written to the LUN.
By default, this column is hidden.

Operations/Sec

Displays the rate of the total operations performed on the LUN. By default,
this column is hidden.

Related references

Types of object status on page 305


Reports tab customization on page 304
Reports management on page 303

318 | OnCommand Console Help

Qtrees report
The Qtrees report displays information about all the qtrees in a volume. You can monitor the capacity
and status of the qtree, and the used and available space in the qtree.
You can monitor and manage only qtrees created by the user. Therefore, the default qtree, qtree 0, is
not monitored or managed.
Note: To display the icons in a report, you must ensure that a DNS mapping is established between
the client machine from which you are starting the Web connection and the host name of the
system on which the DataFabric Manager server is installed.

Qtree

Displays the name of the qtree. The icon indicates if the qtree is clustered (
or nonclustered (

).

Storage Server

Displays the name of the storage server that contains the qtree. The storage
server can be a storage controller or a vFiler unit that contains the qtree.

Volume

Displays the name of the volume that contains the qtree.

Status

Displays the current status of the qtree based on the events generated for the
qtree. The status can be Normal, Warning, Error, Critical, Emergency, or
Unknown.

Used Capacity
(GB)

Displays the amount of space used in the qtree.

Disk Space Limit Displays the hard limit on disk space as specified in the /etc/quotas file of
(GB)
the storage system. By default, this column is hidden.
Possible Addition Displays the amount of additional storage that can be installed on the storage
server to increase the available space for this qtree.
(GB)
Possible
Available (GB)

Displays the total amount of storage (currently available and possible addition)
that is available for increasing the capacity of the qtree.

Related references

Types of object status on page 305


Reports tab customization on page 304
Reports management on page 303

Reports | 319

Storage Systems report


The Storage Systems report displays information such as the type, model, serial number, and the
operating system version of the storage system.
Note: To display the charts and icons in a report, you must ensure that a DNS mapping is

established between the client machine from which you are starting the Web connection and the
host name of the system on which the DataFabric Manager server is installed.
Chart
You can customize the data displayed in the chart, change the subtype and format of the chart, and
export the data from the chart.
Report details
Type

Displays whether the storage system is a cluster or a controller. The controller


can be a stand-alone system, a clustered system, or part of an HA pair.

Status

Displays the current status of the storage system based on the events generated
for the storage system. The status can be Normal, Warning, Error, Critical,
Emergency, or Unknown.

Storage System

Displays the name of the storage system.

Model

Displays the model of the storage system.

Serial Number

Displays the serial number of the storage system. The system serial number is
usually provided on the chassis and is used to identify a system.

OS Version

Displays the version of the operating system running on the storage system.

Firmware
Version

Displays the version of the firmware running on the storage system.

System ID

Displays the product ID of the storage system. It is a unique number to identify


a system, and is usually specified for the storage system head. By default, this
column is hidden.

Related references

Reports tab customization on page 304


Types of object status on page 305
Reports management on page 303

320 | OnCommand Console Help

vFiler Units report


The vFiler Units report displays information about the vFiler units that are discovered by the
DataFabric Manager server, and the network addresses that are assigned to them.
Note: To display the icons in a report, you must ensure that a DNS mapping is established between
the client machine from which you are starting the Web connection and the host name of the
system on which the DataFabric Manager server is installed.

Status

Displays the current status of the vFiler unit based on the events generated for
the vFiler unit. The status can be Normal, Warning, Error, Critical, Emergency,
or Unknown.

vFiler Units

Displays the name of the vFiler unit.

Hosting Storage Displays the name of the hosting storage system.


System
Network
Address

Displays the IP address of the vFiler unit.

IP Space

Defines a distinct IP address space in which vFiler units can participate. By


default, this column is hidden.

System ID

Displays the universal unique identifier (UUID) of the vFiler unit. By default,
this column is hidden.

Ping Status

Displays the status of the ping request sent to the vFiler unit. A vFiler unit might
be up or down. The ping status is displayed as "Down" if the vFiler unit has
stopped. The ping status is displayed as "Down (inconsistent)" if the vFiler unit
state is inconsistent.

Ping Timestamp

Displays the date and time when the vFiler unit was last queried.

Down Timestamp

Displays the date and time when the vFiler unit went offline.

Related references

Types of object status on page 305


Reports tab customization on page 304
Reports management on page 303

Reports | 321

Volumes report
The Volumes report displays information about the type, state, and status of volumes, and the RAID
protection scheme.
Note: To display the icons in a report, you must ensure that a DNS mapping is established between
the client machine from which you are starting the Web connection and the host name of the
system on which the DataFabric Manager server is installed.

Volume

Displays the name of the volume. The icon indicates if the volume is clustered
(

) or nonclustered (

).

Aggregate

Displays the name of the aggregate that contains the volume.

Storage
Server

Displays the name of the storage server that contains the volume. The storage
server can be a storage controller, Vserver, or a vFiler Unit.

Type

Specifies the type of volume that is selected: Traditional, Flexible or Striped. By


default, this column is hidden.

Block Type

Displays the block format of the volume as 32_bit or 64_bit. By default, this
column is hidden.

RAID

Displays the RAID protection scheme. The RAID protection scheme can be one of
the following:
raid0

All the raid groups in the volume are of type raid0.

raid4

All the raid groups in the volume are of type raid4.

raid_dp

All the raid groups in the volume are of type raid_dp.

mixed_raid_type The volume contains RAID groups of different RAID types


(raid0, raid4, and raid_dp).
Note: Data ONTAP uses raid4 or raid_dp protection to ensure data integrity
within a group of disks even if one or two of those disks fail.

State

Displays the state of the volume. A volume can be in one of the following three
states (also called mount states):
Online

Read and write access is allowed.

Offline

Read and write access is not allowed.

Restricted Some operations such as copying volumes and parity reconstruction


are allowed, but data access is not allowed.
Status

Displays the current status of the volume based on the events generated for the
volume. The status can be Normal, Warning, Error, Critical, Emergency, or
Unknown.

322 | OnCommand Console Help


Clones

Displays the clone of the specified volume, if any.

Parent
Clones

Displays the parent volume from which the clone is derived. By default, this
column is hidden.

Related references

Types of object status on page 305


Reports tab customization on page 304
Reports management on page 303
Storage Services report
The Storage Services report displays information about the available storage services.
Storage Service

Displays the name of the storage service.

Storage Service
Description

Displays the description about the corresponding storage service.

Protection Policy

Displays the protection policy that is assigned to the storage service.

Primary Provisioning
Policy

Displays the provisioning policy that is assigned to the primary node


of the storage service.

Dataset Count

Displays the number of datasets that are associated with the storage
service.

Related references

Reports tab customization on page 304


Reports management on page 303
Storage Service Policies report
The Storage Service Policies report displays information about the policies and resource pools that
are associated with each policy node of a storage service.
Storage Service

Displays the name of the storage service.

Protection Policy

Displays the protection policy that is assigned to the storage service.

Policy Node

Displays the name of the data protection policy node, depending on the type
of protection policy that is assigned to the storage service.

Provisioning Policy Displays the provisioning policy that is assigned to the policy node of the
storage service.
vFiler Template

Displays the vFiler template that specifies the configuration settings that are
required to create a new vFiler unit.

Reports | 323

The vFiler template is assigned to the policy node of the storage service.
Resource Pools

Displays the resource pools that are associated with the policy node of the
storage service.

Related references

Reports tab customization on page 304


Reports management on page 303
Storage Service Datasets report
The Storage Service Datasets report displays information about the datasets associated with a storage
service.
Storage Service

Displays the name of the storage service.

Dataset

Displays the name of the dataset that is associated with a storage


service.

Protection Policy

Displays the protection policy that is assigned to the storage service.

Primary Provisioning
Policy

Displays the provisioning policy that is assigned to the primary policy


node of the storage service.

Related references

Reports tab customization on page 304


Reports management on page 303

Storage capacity reports


Understanding storage capacity reports
Overview of storage capacity reports
Storage capacity reports provide information about a storage object's capacity, committed space,
growth, space savings and reservation, and usage metrics.
Storage capacity reports are as follows:

Capacity reports
Committed capacity reports
Capacity growth reports
Space reservation reports
Space efficiency reports

324 | OnCommand Console Help

Usage metrics reports

Capacity reports
Capacity reports provide information about the total capacity, used space, and free space available in
the storage object.
You can view the following capacity reports:

Storage Systems Capacity report


Aggregates Capacity report
Volumes Capacity report
Qtrees Capacity report
User Quotas Capacity report

Committed capacity reports


Committed capacity reports provide information about the capacity, space usage, and committed
space statistics. You can determine the amount of space committed in the qtrees or flexible volumes
from the committed capacity reports.
You can view the following committed capacity reports:

Aggregates Committed Capacity report


Volumes Committed Capacity report

Capacity growth reports


Capacity growth reports provide information about the space usage and growth of the storage objects.
These reports enable you to determine the daily growth rate of aggregates, volumes, or qtrees, and
the time remaining before these storage objects run out of storage space.
You can view the following capacity growth reports:

Aggregates Capacity Growth report


Qtrees Capacity Growth report
Volumes Capacity Growth report

Space reservation reports


Space reservation reports provide information about the space reservation statistics in a volume. You
can determine the size and availability of the space reserve, and the thresholds at which the space
reserve is likely to be depleted using the space reservation reports.
You can view the following space reservation report:

Volumes Space Reservation report

Reports | 325

Space efficiency reports


Space efficiency reports provide information about the space usage and the space savings achieved
for the volumes through deduplication.
You can determine the space savings achieved through deduplication from these reports.
You can view the following space savings reports:

Aggregates Space Savings report


Volumes Space Savings report

Usage metrics reports


Usage metric reports provide information about the space used by the dataset nodes.
You can determine the total space used by a dataset, space used by the dataset nodes in primary and
secondary physical storage, allocated primary storage for a dataset node, the node's read and write
statistics, and so on.
You can view the following usage metrics reports:

Datasets Average Space Usage Metrics report


Datasets Maximum Space Usage Metrics report
Datasets IO Usage Metrics report
Datasets Average Space Usage Metric Samples report
Datasets Maximum Space Usage Metric Sample report
Datasets IO Usage Metric Sample report

Understanding usage metrics reports


What usage metric reports are
Usage metric reports provide information about the space used by the dataset nodes, such as total
space used by a dataset, space used by the dataset nodes in primary and secondary physical storage,
allocated primary storage for a dataset node, and the node's read and write statistics.
Usage Metric reports are generated for each node of the dataset at specified intervals. You can set the
interval at which you want a report to be generated. To set the interval, you must use the command
dfm option.
There are six Usage Metric reports. The reports are broadly classified asspace utilization and input/
output measurement.
The space utilization reports provide information about effective used data space, physical used data
space, total data space, used Snapshot space, Snapshot reserve, total space, and guaranteed space.
The space utilization values are presented in the following reports:

Datasets Average Space Usage Metric report

326 | OnCommand Console Help

Datasets Maximum Space Usage Metric report


Datasets Average Space Usage Metric Sample report
Datasets Maximum Space Usage Metric Sample report

The input/output measurement reports provide information about the data read from or written to
each dataset node. The input/output measurement values are presented in the following reports:

Datasets IO Usage Metric report


Datasets IO Usage Metric Sample report

The Usage Metric reports do not include information earlier than the most recent 12 months.
Guidelines for solving usage metric report issues
You must follow certain guidelines to avoid being unable to view or generate a usage metric report or
receiving excess data.
The following guidelines can help you to avoid issues related to report generation:

Reports cannot be created for the destination node if a mirror relationship is not created for the
primary node.
If a mirror relationship is not created in the primary node, then the destination volumes are
deleted. Therefore, metrics are not calculated as there are no volumes in the dataset of the
destination node.
Reports cannot be generated for the second node of a node pair if it is not accessible by the
DataFabric Manager server.
Reports cannot be generated for a dataset if the space utilization monitors or the input/output
monitors are turned off.
The Notes field of a usage metric report might display overcharge if a dataset has one or more
qtrees in the primary node.
Having one or more qtrees in the primary node results in the datasets volume information being
included in the metric computation of the qtrees.

How space utilization values are calculated


The space utilization values are calculated using different formulas. You can verify the values
presented in the Usage Metric report by using either the volume-list-info API or the dfTables
file.
You must use the volume-list-info API to view the sample data collected for the space saved by
data sharing, the volume size, total data space, and the Snapshot reserve. To view the sample data
collected for used data space, the Snapshot overflow, and the used Snapshot space, you must use the
snmpwalk command on the dfTable file.
Effective Used Data
Space formula

Effective Used Data Space = used data space - (Snapshot overflow +


available overwrite reserve + hole reserve) + dedupe returns

Reports | 327
Physical Used Data
Space formula

Physical Used Data Space = used data space - (Snapshot overflow +


available overwrite reserve + hole reserve)

Total Data Space


formula

Total Data Space = Total space of all the volumes in the dataset node

Used Snapshot Space


formula

Used Snapshot Space = Sum of the physical space used by the volume
Snapshot copies

Snapshot Reserve
formula

Snapshot Reserve = Sum of the Snapshot Reserve of all the volumes of


the dataset node

Total Space formula

Total Space = Total Data Space + Snapshot Reserve

Guaranteed Space
formula

Guaranteed Space for volume = Total Data Space + Snapshot Reserve


Guaranteed Space for file = (used data space + Used Snapshot Space) Snapshot overflow
Guaranteed Space for none = (Physical Used Data Space + Used
Snapshot Space) - Snapshot overflow
Note: Guaranteed Space is calculated based on the guarantee level set
on a volume.

How maximum and average space utilization values are calculated


You can calculate the average space utilization and the maximum space utilization for the volumes in
a dataset node at specified intervals.
Space utilization computation using sample data collected from different volumes
Assume the following about a dataset:

volx, voly, volz are the volumes of the dataset's primary node.

Each volume has two samples. The samples are sx1, sx2 for volx.
All the samples are collected at the same time.

Maximum space utilization = MAX [(sx1 + sy1 + sz1), (sx2 + sy2 + sz2)]
Average space utilization = AVG [(sx1 + sy1 + sz1), (sx2 + sy2 + sz2)]
Example: Space utilization computation when sample data is not present for
some volumes
Assume the following about a dataset:

volx, voly, and volz are the volumes of the dataset's primary node.

Sample sx1 is collected at time t1 for volume volx.


Samples sy1 and sy2 are collected at time t1 and t2, respectively, for volume voly.

328 | OnCommand Console Help

Samples sz1, sz2, sz3, and sz4 are collected at time t1, t2, t3, and t4, respectively, for
volume volz.

Maximum space utilization = MAX [(sx1 + sy1 + sz1), (sy2 + sz2), sz3, sz4]
Average space utilization = AVG [(sx1 + sy1 + sz1), (sy2 + sz2), sz3, sz4]
How input/output measurement values are calculated
The total input/output measurement value is the sum of the total data read to and written from all the
volumes in each dataset node. You can either use the Data ONTAP APIs or the dfTables file to view
that sample data.
How total input/output measurement values are calculated
You can calculate the total data read and written for each dataset node. The total input/output
measurement value is the sum of data read and written for all volumes in a dataset node, at specified
intervals for a specified period.
Example: Input/output measurement for data collected from different volumes
Assume the following about a dataset:

volx, voly, volz are the volumes of the dataset's primary node.

Each volume has three samples for a input/output metric. Let the samples be sx1, sx2, and
sx3 for volx, collected at time t1,t2, and t3 respectively.

Total Data Read for volx between interval t1 and t3 = (sx1-sx0) + (sx2-sx1) + (sx3-sx2)
Total Data Read from a dataset node between time t1 and t3 = { [(sx1-sx0) + (sx2-sx1) +
(sx3-sx2)] + [(sy1-sy0) + (sy2-sy1) + (sy3-sy2)] + [(sz1-sz0) + (sz2-sz1) + (sz3-sz2)] }
The total data read between time t01 to t3, where t01 is the timestamp between t0 and t1, is
calculated by normalizing the t0 and t1 samples.
Therefore, Total Data Read for volx for the interval between t01 to t3 = [(sx1-sx0) * (t1t01)/(t1-t0)] + (sx2-sx1) + (sx3-sx2)

Monitoring storage capacity reports


Aggregate capacity thresholds and their events
You can configure capacity thresholds for aggregates and events for these thresholds from
DataFabric Manager server. You can set alarms to monitor the capacity and committed space of an
aggregate. You can also take corrective actions based on the event generated.
You can configure alarms to send notification whenever an event related to the capacity of an
aggregate occurs. For the Aggregate Full threshold, you can also configure an alarm to send
notification only when the condition persists over a specified time.

Reports | 329
By default, if you have configured an alarm to alert you to an event, the DataFabric Manager server
issues the alarm only once per event. You can configure the alarm to repeat until you receive an
acknowledgment.
Note: If you want to set an alarm for a specific aggregate, you must create a group with that

aggregate as the only member.


You can set the following aggregate capacity thresholds:
Aggregate Full
(%)

Description: Specifies the percentage at which an aggregate is full.


Note: To reduce the number of Aggregate Full Threshold events generated,
you can set an Aggregate Full Threshold Interval. This causes DataFabric
Manager server to generate an Aggregate Full event only if the condition
persists for the specified time.

Default value: 90 percent


Event generated: Aggregate Full
Event severity: Error
Corrective Action
Perform one or more of the following actions:

To free disk space, ask your users to delete files that are no longer needed
from volumes contained in the aggregate that generated the event.
You must add one or more disks to the aggregate that generated the event.
Note: After you add a disk to an aggregate, you cannot remove it
without first destroying all flexible volumes present in the aggregate to
which the disk belongs. You must destroy the aggregate after all the
flexible volumes are removed from the aggregate.

Aggregate Nearly
Full (%)

You must temporarily reduce the Snapshot reserve.


By default, the reserve is 20 percent of disk space. If the reserve is not in
use, reducing the reserve can free disk space, giving you more time to add
a disk.
There is no way to prevent Snapshot copies from consuming disk space
greater than the amount reserved for them. It is, therefore, important to
maintain a large enough reserve for Snapshot copies so that the active file
system always has space available to create new files or modify existing
ones. For more information about the Snapshot reserve, see the Data
ONTAP Data Protection Online Backup and Recovery Guide.

Description: Specifies the percentage at which an aggregate is nearly full.


Default value: 80 percent

330 | OnCommand Console Help

The value for this threshold must be lower than the value for Aggregate Full
Threshold for DataFabric Manager server to generate meaningful events.
Event generated: Aggregate Almost Full
Event severity: Warning
Corrective action
Perform one or more of the actions mentioned in Aggregate Full.
Aggregate
Overcommitted
(%)

Description: Specifies the percentage at which an aggregate is


overcommitted.
Default value: 100 percent
Event generated: Aggregate Overcommitted
Event severity: Error
Corrective action
You should perform one or more of the following actions:

You must create new free blocks in the aggregate by adding one or more
disks to the aggregate that generated the event.
Note: You must add disks with caution. After you add a disk to an
aggregate, you cannot remove it without first destroying all flexible
volumes present in the aggregate to which the disk belongs. You must
destroy the aggregate after all the flexible volumes are destroyed.

You must temporarily free some already occupied blocks in the aggregate
by taking unused flexible volumes offline.
Note: When you take a flexible volume offline, it returns any space it
uses to the aggregate. However, when you bring the flexible volume
online again, it requires the space again.

Aggregate Nearly
Overcommitted
(%)

Permanently free some already occupied blocks in the aggregate by


deleting unnecessary files.

Description: Specifies the percentage at which an aggregate is nearly


overcommitted.
Default value: 95 percent
The value for this threshold must be lower than the value for Aggregate Full
Threshold for DataFabric Manager server to generate meaningful events.
Event generated: Aggregate Almost Overcommitted
Event severity: Warning
Corrective action

Reports | 331

Perform one or more of the actions provided in Aggregate Overcommitted.


Aggregate
Snapshot Reserve
Nearly Full
Threshold (%)

Description: Specifies the percentage of the Snapshot copy reserve on an


aggregate that you can use before the system generates the Aggregate
Snapshots Nearly Full event.
Default value: 80 percent
Event generated: Aggregate Snapshot Reserve Almost Full
Event severity: Warning
Corrective action: None
There is no way to prevent Snapshot copies from consuming disk space
greater than the amount reserved for them. If you disable the aggregate
Snapshot autodelete option, it is important to maintain a large enough
reserve.
See the Operations Manager Help for instructions on how to identify
Snapshot copies you can delete. For more information about the Snapshot
reserve, see the Data ONTAP Data Protection Online Backup and Recovery
Guide.

Aggregate
Snapshot Reserve
Full Threshold
(%)

Description: Specifies the percentage of the Snapshot copy reserve on an


aggregate that you can use before the system generates the Aggregate
Snapshots Full event.
Default value: 90 percent
Event generated: Aggregate Snapshot Reserve Full
Event severity: Warning
Corrective action: None
There is no way to prevent Snapshot copies from consuming disk space
greater than the amount reserved for them.

Note: A newly created traditional volume tightly couples with its containing aggregate so that the
capacity of the aggregate determines the capacity of the new traditional volume. Therefore, you
should synchronize the capacity thresholds of traditional volumes with the thresholds of their
containing aggregates.
Related information

Data ONTAP Data Protection Online Backup and Recovery Guide - now.netapp.com/NOW/
knowledge/docs/ontap/ontap_index.shtml

332 | OnCommand Console Help

Qtree capacity thresholds and events


The OnCommand console enables you to monitor qtree capacity and set alarms. You can also take
corrective actions based on the event generated.
DataFabric Manager server features thresholds to help you monitor the capacity of qtrees. Quotas
must be enabled on the storage systems. You can configure alarms to send notification whenever an
event related to the capacity of a qtree occurs.
By default, if you have configured an alarm to alert you to an event, the DataFabric Manager server
issues the alarm only once per event. You can configure the alarm to continue to alert you with
events until it is acknowledged. For the Qtree Full threshold, you can also configure an alarm to send
notification only when the condition persists over a specified period.
Note: If you want to set an alarm for a specific qtree, you must create a group with that qtree as the
only member.

You can set the following qtree capacity thresholds:


Qtree Full
(%)

Description: Specifies the percentage at which a qtree is considered full.


Note: To reduce the number of Qtree Full Threshold events generated, you can

set a Qtree Full Threshold Interval to a non-zero value. By default, the Qtree Full
threshold Interval is set to zero. The Qtree Full Threshold Interval specifies the
time during which the condition must persist before the event is generated. If the
condition persists for the specified amount of time, DataFabric Manager server
generates a Qtree Full event.

If the threshold interval is 0 seconds or a value less than the volume


monitoring interval, DataFabric Manager server generates Qtree Full events.
If the threshold interval is greater than the volume monitoring interval,
DataFabric Manager server waits for the specified threshold interval, which
includes two or more monitoring intervals, and generates a Qtree Full event
only if the condition persisted throughout the threshold interval.

For example, if the monitoring cycle time is 60 seconds and the threshold
interval is 90 seconds, the threshold event is generated only if the condition
persists for two monitoring intervals.
Default value: 90 percent
Event generated: Qtree Full
Event severity: Error
Corrective action
Perform one or more of the following actions:

Ask users to delete files that are no longer needed, to free disk space.

Reports | 333

Increase the hard disk space quota for the qtree.

Qtree Nearly Description: Specifies the percentage at which a qtree is considered nearly full.
Full
Default value: 80 percent
Threshold
Event severity: Warning
(%)
Corrective action
Perform one or more of the following actions:

Ask users to delete files that are no longer needed, to free disk space.
Increase the hard disk space quota for the qtree.

Related information

Data ONTAP Data Protection Online Backup and Recovery Guide - now.netapp.com/NOW/
knowledge/docs/ontap/ontap_index.shtml
User quota thresholds
You can set a user quota threshold to all the user quotas present in a volume or a qtree.
When you configure a user quota threshold for a volume or qtree, the settings apply to all user quotas
on that volume or qtree.
DataFabric Manager server uses the user quota thresholds to monitor the hard and soft quota limits
configured in the /etc/quotas file of each storage system.
Volume capacity thresholds and events
DataFabric Manager server features thresholds to help you monitor the capacity of flexible and
traditional volumes. You can configure alarms to send notification whenever an event related to the
capacity of a volume occurs. You can also take corrective actions based on the event generated. For
the Volume Full threshold, you can configure an alarm to send notification only when the condition
persists over a specified period.
By default, if you have configured an alarm to alert you to an event, the DataFabric Manager server
issues the alarm only once per event. You can configure the alarm to repeat until it is acknowledged.
Note: If you want to set an alarm for a specific volume, you must create a group with that volume
as the only member.

You can set the following volume capacity thresholds:


Volume Full
Threshold (%)

Description: Specifies the percentage at which a volume is considered full.


Note: To reduce the number of Volume Full Threshold events
generated, you can set the Volume Full Threshold Interval to a non-zero
value. By default, the Volume Full threshold Interval is set to zero.

334 | OnCommand Console Help

Volume Full Threshold Interval specifies the time during which the
condition must persist before the event is triggered. Therefore, if the
condition persists for the specified time, DataFabric Manager server
generates a Volume Full event.

If the threshold interval is 0 seconds or a value less than the volume


monitoring interval, DataFabric Manager server generates the
Volume Full events.
If the threshold interval is greater than the volume monitoring
interval, DataFabric Manager server waits for the specified
threshold interval, which includes two or more monitoring intervals,
and generates a Volume Full event only if the condition persisted
throughout the threshold interval.

For example, if the monitoring cycle time is 60 seconds and the


threshold interval is 90 seconds, the threshold event is generated only if
the condition persists for two monitoring intervals.
Default value: 90
Event generated: Volume Full
Event severity: Error
Corrective action
Perform one or more of the following actions:

Ask your users to delete files that are no longer needed, to free disk
space.
For flexible volumes containing enough aggregate space, you can
increase the volume size.
For traditional volumes containing aggregates with limited space, you
can increase the size of the volume by adding one or more disks to the
aggregate.
Note: Add disks with caution. After you add a disk to an aggregate,
you cannot remove it without destroying the volume and its
aggregate.

For traditional volumes, temporarily reduce the Snapshot copy reserve.


By default, the reserve is 20 percent of the disk space. If the reserve is
not in use, reduce the reserve free disk space, giving you more time to
add a disk. There is no way to prevent Snapshot copies from
consuming disk space greater than the amount reserved for them.
Therefore, it is important to maintain a large enough reserve for
Snapshot copies. By maintaining the reserve for Snapshot copies, the
active file system always has space available to create new files or

Reports | 335

modify existing ones. For more information about the Snapshot copy
reserve, see the Data ONTAP Data Protection Online Backup and
Recovery Guide.
Volume Nearly Full Description: Specifies the percentage at which a volume is considered
nearly full.
Threshold (%)
Default value: 80. The value for this threshold must be lower than the
value for the Volume Full Threshold in order for DataFabric Manager
server to generate meaningful events.
Event generated: Volume Almost Full
Event severity: Warning
Corrective action
Perform one or more of the actions mentioned in Volume Full.
Volume Space
Reserve Nearly
Depleted Threshold
(%)

Description: Specifies the percentage at which a volume is considered to


have consumed most of its reserved blocks. This option applies to
volumes with LUNs, Snapshot copies, no free blocks, and a fractional
overwrite reserve of less than 100%. A volume that crosses this threshold
is getting close to having write failures.
Default value: 80
Event generated: Volume Space Reservation Nearly Depleted
Event severity: Warning

Volume Space
Reserve Depleted
Threshold (%)

Description: Specifies the percentage at which a volume is considered to


have consumed all its reserved blocks. This option applies to volumes
with LUNs, Snapshot copies, no free blocks, and a fractional overwrite
reserve of less than 100%. A volume that has crossed this threshold is
getting dangerously close to having write failures.
Default value: 90
Event generated: Volume Space Reservation Depleted
Event severity: Error
When the status of a volume returns to normal after one of the preceding
events, events with severity 'Normal' are generated. Normal events do not
generate alarms or appear in default event lists, which display events of
warning or worse severity.

Volume Quota
Overcommitted
Threshold (%)

Description: Specifies the percentage at which a volume is considered to


have consumed the whole of the overcommitted space for that volume.
Default value: 100

336 | OnCommand Console Help

Event generated: Volume Quota Overcommitted


Event severity: Error
Corrective action
Perform one or more of the following actions:

Volume Quota
Nearly
Overcommitted
Threshold (%)

Create new free blocks by increasing the size of the volume that
generated the event.
Permanently free some of the occupied blocks in the volume by
deleting unnecessary files.

Description: Specifies the percentage at which a volume is considered to


have consumed most of the overcommitted space for that volume.
Default Value: 95
Event generated: Volume Quota Almost Overcommitted
Event Severity: Warning
Corrective action:
Perform one or more of the actions mentioned in Volume Quota
Overcommitted.

Volume Growth
Event Minimum
Change (%)

Description: Specifies the minimum change in volume size (as a


percentage of total volume size) that is acceptable. If the change in volume
size is more than the specified value, and the growth is abnormal in
relation to the volume-growth history, DataFabric Manager server
generates a Volume Growth Abnormal event.
Default value: 1
Event generated: Volume Growth Abnormal

Volume Snap
Reserve Full
Threshold (%)

Description: Specifies the value (percentage) at which the space that is


reserved for taking volume Snapshot copies is considered full.
Default value: 90
Event generated: Volume Snap Reserve Full
Event severity: Error
Corrective action: None
There is no way to prevent Snapshot copies from consuming disk space
greater than the amount reserved for them. If you disable the volume
Snapshot autodelete option, it is important to maintain a large
enough reserve. Disabling would ensure Snapshot copies that there is
always space available to create new files or modify present ones. For

Reports | 337

instructions on how to identify Snapshot copies you can delete, see the
Operations Manager Help.
User Quota Full
Threshold (%)

Description: Specifies the value (percentage) at which a user is considered


to have consumed all the allocated space (disk space or files used) as
specified by the user quota. The user quota includes hard limit in
the /etc/quotas file. If this limit is exceeded, DataFabric Manager
server generates a User Disk Space Quota Full event or a User Files Quota
Full event.
Default value: 90
Event generated: User Quota Full

User Quota Nearly Description: Specifies the value (percentage) at which a user is considered
Full Threshold (%) to have consumed most of the allocated space (disk space or files used) as
specified by the user quota. The user quota includes hard limit in
the /etc/quotas file. If this limit is exceeded, DataFabric Manager
server generates a User Disk Space Quota Almost Full event or a User
Files Quota Almost Full event.
Default value: 80
Event generated: User Quota Almost Full
Volume No First
Snapshot
Threshold (%)

Description: Specifies the value (percentage) at which a volume is


considered to have consumed all the free space for its space reservation.
This is the space that the volume needs when the first Snapshot copy is
created.
This option applies to volumes that contain space-reserved files, no
Snapshot copies, a fraction of Snapshot copies overwrite reserve set to
greater than 0, and where the sum of the space reservations for all LUNs
in the volume is greater than the free space available to the volume.
Default value: 90
Event generated: Volume No First Snapshot

Volume Nearly No
First Snapshot
Threshold (%)

Description: Specifies the value (percentage) at which a volume is


considered to have consumed most of the free space for its space
reservation. This is the space that the volume needs when the first
Snapshot copy is created.
This option applies to volumes that contain space-reserved files, no
Snapshot copies, a fractional overwrite reserve set to greater than 0, and
where the sum of the space reservations for all LUNs in the volume is
greater than the free space available to the volume.
Default value: 80

338 | OnCommand Console Help

Event generated: Volume Almost No first Snapshot


Note: When a traditional volume is created, it is tightly coupled with its containing aggregate so

that its capacity is determined by the capacity of the aggregate. For this reason, you should
synchronize the capacity thresholds of traditional volumes with the thresholds of their containing
aggregates.
Related information

Data ONTAP Data Protection Online Backup and Recovery Guide - now.netapp.com/NOW/
knowledge/docs/ontap/ontap_index.shtml

Page descriptions
Aggregates Capacity report
The Aggregates Capacity report displays information about the used and available space in an
aggregate and its capacity.
Note: To display the charts and icons in a report, you must ensure that a DNS mapping is
established between the client machine from which you are starting the Web connection and the
host name of the system on which the DataFabric Manager server is installed.

Chart
You can customize the data displayed in the chart, change the subtype and format of the chart, and
export the data from the chart.
Report details
Aggregate

Displays the name of the aggregate.

Storage System

Displays the name of the storage system that contains the aggregate.

Available Capacity (GB)

Displays the amount of space available for data in the aggregate. By


default, this column is hidden.

Total Capacity (GB)

Displays the total space in the aggregate.

Used Capacity (%)

Displays the percentage of space used for data in the aggregate.

Snap Reserve Total (GB) Displays the size of the Snapshot reserve for this aggregate.
Snap Reserve Used (%)

Displays the percentage of the Snapshot reserve currently in use.

Aggregate Used Capacity Displays the amount of space used for data in the aggregate. By
default, this column is hidden.
(GB)

Reports | 339
Aggregate Available
Capacity (%)

Displays the percentage of free space for data in the aggregate. By


default, this column is hidden.

Snapshot Autodelete

Specifies whether a Snapshot copy is deleted to free storage space


when a write to the volume fails due to lack of space in the aggregate.
By default, this column is hidden.

Snap Reserve Used (GB)

Displays the amount of the Snapshot reserve currently in use. By


default, this column is hidden.

Snapshots Disabled

Specifies whether Snapshot copies are disabled for this aggregate. If


the status is displayed as Off, then Snapshot copies are not disabled.
By default, this column is hidden.

Status

Displays the current status of the aggregate based on the events


generated for the aggregate. The status can be Normal, Warning, Error,
Critical, Emergency, or Unknown.

Full Threshold (%)

Displays the percentage of physical space that is used in the aggregate


before the system generates the Aggregate Full event.

Nearly Full Threshold


(%)

Displays the percentage of physical space that is used in the aggregate


before the system generates the Aggregate Almost Full event.

Overcommitted
Threshold (%)

Displays the percentage of physical space that is used in the aggregate


before the system generates the Aggregate Overcommitted event. By
default, this column is hidden.

Nearly Overcommitted
Threshold (%)

Displays the percentage of physical space that is used in the aggregate


before the system generates the Aggregate Nearly Overcommitted
event. By default, this column is hidden.

Bytes Used (%)

Displays the percentage of space in the aggregate that is used by


flexible volumes. By default, this column is hidden.

Bytes Committed (%)

Displays the percentage of space in the aggregate that is committed to


flexible volumes. By default, this column is hidden.

Related references

Types of object status on page 305


Reports tab customization on page 304
Reports management on page 303

340 | OnCommand Console Help

Qtrees Capacity report


The Qtrees Capacity report displays information about the status of a qtree, used and available space
in a qtree, and its capacity. You can monitor and manage only qtrees created by the user. Therefore,
the default qtree, qtree 0, is not monitored or managed.
Note: To display the icons in a report, you must ensure that a DNS mapping is established between
the client machine from which you are starting the Web connection and the host name of the
system on which the DataFabric Manager server is installed.

Qtree

Displays the name of the qtree. The icon indicates if the qtree is clustered
(

) or nonclustered (

).

Storage Server

Displays the name of the storage server that contains the qtree. The storage
server can be a storage controller or a vFiler unit that contains the qtree.

Volume

Displays the name of the volume that contains the qtree.

Status

Displays the current status of the qtree based on the events generated for the
qtree. The status can be Normal, Warning, Error, Critical, Emergency, or
Unknown.

Used (%)

Displays the percentage of space used in the qtree.

Soft Limit (GB)

Displays the soft limit on disk space as specified in the /etc/quotas file of
the storage system. By default, this column is hidden.

Disk Space Limit


(GB)

Displays the hard limit on disk space as specified in the /etc/quotas file
of the storage system. By default, this column is hidden.

Available Capacity Displays the percentage of available space in the qtree that is not committed.
By default, this column is hidden.
(%)
Full Threshold (%) Displays the limit, as a percentage, at which a qtree is considered full.
Nearly Full
Threshold (%)

Displays the limit, as a percentage, at which a qtree is considered nearly full.

Bytes Used (%)

Displays the percentage of storage space used by the qtree. By default, this
column is hidden.

Files Used (% )

Displays the percentage of space used by files in the qtree. By default, this
column is hidden.

Related references

Types of object status on page 305


Reports tab customization on page 304
Reports management on page 303

Reports | 341

Storage Systems Capacity report


The Storage Systems Capacity report displays information about the capacity and the space used in a
storage system.
Note: To display the icons in a report, you must ensure that a DNS mapping is established between
the client machine from which you are starting the Web connection and the host name of the
system on which the DataFabric Manager server is installed.

Type

Displays whether the storage system is a cluster or a controller. The


controller can be a stand-alone system, a clustered system, or part of an
HA pair. By default, this column is hidden.

Status

Displays the current status of the storage system based on the events
generated for the storage system. The status can be Normal, Warning,
Error, Critical, Emergency, or Unknown.

Storage System

Displays the name of the storage system.

Volume Used
Capacity (GB)

Displays the amount of space used for data in all the volumes. By default,
this column is hidden.

Volume Total
Capacity (GB)

Displays the total space available for data in all the volumes.

Volume Used
Capacity (%)

Displays the percentage of space used for data in all the volumes.

Aggregate Used
Capacity (GB)

Displays the amount of space used for data in an aggregate. By default,


this column is hidden.

Aggregate Total
Capacity (GB)

Displays the total space available for data in the aggregate.

Aggregate Used
Capacity (%)

Displays the percentage of space used for data in the aggregate.

Related references

Types of object status on page 305


Reports tab customization on page 304
Reports management on page 303

342 | OnCommand Console Help

User Quotas Capacity report


The User Quotas Capacity report displays information about the capacity of the quota users and the
quota user groups that exist on the storage systems.
Note: To display the icons in a report, you must ensure that a DNS mapping is established between

the client machine from which you are starting the Web connection and the host name of the
system on which the DataFabric Manager server is installed.
User Name

Displays the name of the user or user group.


The user name contains a comma-separated list of users when a storage system is
configured to track quotas of multiple users as a single entity. For example, the /
etc/ quotas file of storage system F1 contains the following entry:
joe,finance\joe user@/vol/vol0/fin 50M - -

In this case, the DataFabric Manager server displays joe, finance\joe in the User
Name column.
When the user name of a storage system cannot be reported, the DataFabric
Manager server reports one of the following:

User identifier (UID)


Security identifier (SID)
Group identifier (GID)

File System

Displays the name, path, and quota information of the volumes or qtrees on
which the user quota or group quota is enabled.

Status

Displays the status of a user's quotas. The status can be Normal, Warning, Error,
Critical, Emergency, or Unknown.
If the status for a user is not Normal, an event related to the user's quotas has
occurred. For details about the events, you must go to the Events tab.

Disk Space
Used (MB)

Displays the total amount of disk space used. By default, this column is hidden.

Disk Space
Threshold
(MB)

Displays the disk space threshold as specified in the /etc/quotas file of the
storage system.
Note: This threshold is different from the user quota thresholds that you can
configure in the DataFabric Manager server.

Disk Space Soft Displays the soft limit on disk space as specified in the /etc/quotas file of the
Limit (MB)
storage system. By default, this column is hidden.

Reports | 343
Disk Space
Hard Limit
(MB)

Displays the hard limit on disk space as specified in the /etc/quotas file of the
storage system. By default, this column is hidden.

Disk Space
Used (%)

Displays the total percentage of disk space used.

Files Used

Displays the total number of files used. By default, this column is hidden.

Files Soft Limit Displays the soft limit on files as specified in the /etc/quotas file of the
storage system. By default, this column is hidden.
Files Hard
Limit (Million)

Displays the hard limit on files as specified in the /etc/quotas file of the
storage system. By default, this column is hidden.

Files Used (%)

Displays the percentage of files used.


The percentage of files used is calculated using the following formula:
Percentage = (Files Used / Files Hard Limit) x 100

SID

Displays the identifier of the user or user group.


The Security Identifier (SID) contains a comma-separated list of identifiers of
users when a storage system is configured to track quotas of multiple users as a
single entity. For Windows users or user groups, this identifier specifies the SID
of the user or user group. For other platforms, the identifier specifies the Relative
Identifier (RID) on the storage system.

Nearly Full
Threshold (%)

Displays the percentage value at which a user is likely to consume most of the
allocated space (disk space or files used) as specified by the user's quota (hard
limit in the /etc/quotas file).
If this threshold is crossed, the DataFabric Manager server generates a User Disk
Space Quota Almost Full event when the disk space is consumed or a User Files
Quota Almost Full event when the file space is consumed.

Full Threshold
(%)

Displays the percentage value at which a user is likely to consume the entire
allocated space (disk space or files used) as specified by the user's quota (hard
limit in the /etc/quotas file).
If this threshold is crossed, the DataFabric Manager server generates a User Disk
Space Quota Full event when the disk space is consumed or a User Files Quota
Full event when the file space is consumed.

Related references

Types of object status on page 305


Reports tab customization on page 304
Reports management on page 303

344 | OnCommand Console Help

Volumes Capacity report


The Volumes Capacity report displays information about the capacity of the volumes and the used
and available space in these volumes.
Note: To display the icons in a report, you must ensure that a DNS mapping is established between
the client machine from which you are starting the Web connection and the host name of the
system on which the DataFabric Manager server is installed.

Volume

Displays the name of the volume. The icon indicates if the volume is
clustered (

) or nonclustered (

).

Aggregate

Displays the name of the aggregate that contains the volume.

Storage Server

Displays the name of the storage server that contains the volume. The
storage server can be a storage controller, Vserver, or a vFiler unit.

Available Capacity
(GB)

Displays the amount of space available for data in the volume. By default,
this column is hidden.

Used Capacity (GB) Displays the amount of space that is used for data, in GB, in the volume. By
default, this column is hidden.
Total Capacity (GB) Displays the total space available for data in the volume.
Used Capacity (%)

Displays the amount of space that is used for data, in percentage, in the
volume.

Used Snapshot
Space (GB)

Displays the amount of space used to store Snapshot copies in the volume.
This value can be larger than the specified size of the Snapshot reserve.

Used Snapshot
Space (%)

Displays the percentage of space used to store Snapshot copies in the


volume. This value can be larger than the specified size of the Snapshot
reserve.

Available Capacity
(%)

Displays the percentage of space available for data in the volume. By


default, this column is hidden.

Status

Displays the current status of the volume based on the events generated for
the volume. The status can be Normal, Warning, Error, Critical, Emergency,
or Unknown.

Full Threshold (%)

Displays the limit, as a percentage, at which a volume is considered full.

Nearly Full
Threshold (%)

Displays the limit, as a percentage, at which a volume is considered nearly


full.

Files Used (%)

Displays the percentage of files used by the volume. By default, this column
is hidden.

Reports | 345
Related references

Types of object status on page 305


Reports tab customization on page 304
Reports management on page 303
Aggregates Committed Capacity report
The Aggregates Committed Capacity report displays information about the type of the aggregate, and
the space committed to flexible volumes in the aggregate.
Note: To display the icons in a report, you must ensure that a DNS mapping is established between
the client machine from which you are starting the Web connection and the host name of the
system on which the DataFabric Manager server is installed.

Aggregate

Displays the name of the aggregate.

Storage System

Displays the name of the storage system that contains the aggregate.

Type

Displays the type of the aggregate:

Bytes
Committed
(GB)

Traditional

An aggregate that contains only one traditional volume. A


traditional aggregate is created automatically in response to
a volume create request on the storage system.

Aggregate

An aggregate that contains one or more flexible volumes,


RAID groups, and disks.

Striped
Aggregate

An aggregate that is striped across multiple nodes, allowing


its volumes to also be striped across multiple nodes. Striped
aggregates are made up of multiple member aggregates.

Displays the amount of space committed to flexible volumes, in GB, in the


aggregate.

Displays the amount of space committed to flexible volumes, in percentage, in


Bytes
Committed (%) the aggregate.
Status

Displays the current status of the aggregate based on the events generated for the
aggregate. The status can be Normal, Warning, Error, Critical, Emergency, or
Unknown.

Related references

Types of object status on page 305


Reports tab customization on page 304
Reports management on page 303

346 | OnCommand Console Help

Volumes Committed Capacity report


The Volumes Committed Capacity report displays information about the capacity and space used in
the volume, and the space committed in the qtrees.
Note: To display the icons in a report, you must ensure that a DNS mapping is established between

the client machine from which you are starting the Web connection and the host name of the
system on which the DataFabric Manager server is installed.
Volume

Displays the name of the volume. The icon indicates if the volume is
clustered (

) or nonclustered (

).

Aggregate

Displays the name of the aggregate that contains the volume.

Storage Server

Displays the name of the storage server that contains the volume. The
storage server can be a storage controller, Vserver, or a vFiler Unit.

Quota OverCommitted Displays the amount of physical space in the qtrees that can be used
before the system generates the Volume Quota OverCommitted event.
Space (GB)
Used Capacity (GB)

Displays the amount of space used for data in the volume.

Committed (GB)

Displays the amount of space, in GB, committed in the qtrees. By


default, this column is hidden.

Committed (%)

Display the amount of space, in percentage, committed in the qtrees.

Total/Max Size (GB)

Displays the total amount of storage allocated for this volume, if the
volume autosize option is disabled on this volume.
Displays the maximum size to which the volume can grow, if the
volume autosize option is enabled on this volume.

Related references

Reports tab customization on page 304


Reports management on page 303
Aggregates Capacity Growth report
The Aggregates Capacity Growth report displays information about the space used in and the growth
in capacity of the aggregate.
Note: To display the charts and icons in a report, you must ensure that a DNS mapping is

established between the client machine from which you are starting the Web connection and the
host name of the system on which the DataFabric Manager server is installed.

Reports | 347
Chart
You can customize the data displayed in the chart, change the subtype and format of the chart, and
export the data from the chart.
Report details
Aggregate

Displays the name of the aggregate.

Storage System

Displays the name of the storage system that contains the aggregate.

Data Days To Full

Displays the number of days required for the aggregate to reach the actual
Aggregate Full (in terms of capacity), and is based on the daily growth rate
(GB) value.

Daily Growth Rate


(GB)

Displays, in GB, the amount of disk space used in the aggregate if the
amount of change between the last two samples continues for 24 hours. The
default sample collection interval is four hours.
For example, if an aggregate uses 10 GB of disk space at 2 pm and 12 GB at
6 pm, the daily growth rate (GB) for this aggregate is 2 GB.

Daily Growth Rate


(%)

Displays the percentage of the rate of growth in capacity. This is determined


by dividing the daily growth rate by the total amount of space in the
aggregate.

Used Capacity (%)

Displays the percentage of the total space currently in use in the aggregate.

Committed %

Displays the amount of space, in percentage, committed to flexible volumes


in the aggregate.

Committed (GB)

Displays the amount of space, in GB, committed to flexible volumes in the


aggregate. By default, this column is hidden.

Total Capacity (GB) Displays the total amount of space in the aggregate. By default, this column
is hidden.
Used Capacity (GB) Displays the amount of used space in the aggregate. By default, this column
is hidden.
Related references

Reports tab customization on page 304


Reports management on page 303

348 | OnCommand Console Help

Qtrees Capacity Growth report


The Qtrees Capacity Growth report displays information about the space usage and the growth in
capacity of the qtree. You can monitor and manage only qtrees created by the user. Therefore, the
default qtree, qtree 0, is not monitored or managed.
Note: To display the charts and icons in a report, you must ensure that a DNS mapping is
established between the client machine from which you are starting the Web connection and the
host name of the system on which the DataFabric Manager server is installed.

Chart
You can customize the data displayed in the chart, change the subtype and format of the chart, and
export the data from the chart.
Report details
Qtree

Displays the name of the qtree. The icon indicates if the qtree is clustered (
or nonclustered (

).

Storage Server

Displays the name of the storage server that contains the qtree. The storage
server can be a storage controller or a vFiler unit that contains the qtree.

Volume

Displays the name of the volume that contains the qtree.

Data Days to
Full

Displays the estimated amount of time left before this qtree runs out of storage
space.
If the time is less than one day, the current storage status of the qtree is
displayed.

Daily Growth
Rate (GB)

Displays, in GB, the amount of disk space used in the qtree if the amount of
change between the last two samples continues for 24 hours.

Daily Growth
Rate (%)

Displays the percentage of change in the disk space used in the qtree if the
amount of change between the last two samples continues for 24 hours.

Related references

Reports tab customization on page 304


Reports management on page 303

Reports | 349

Volumes Capacity Growth report


TheVolumes Capacity Growth report displays information about the space usage and the growth in
capacity of the volume.
Note: To display the charts and icons in a report, you must ensure that a DNS mapping is
established between the client machine from which you are starting the Web connection and the
host name of the system on which the DataFabric Manager server is installed.

Chart
You can customize the data displayed in the chart, change the subtype and format of the chart, and
export the data from the chart.
Report details
Volume

Displays the name of the volume. The icon indicates if the volume is
clustered (

) or nonclustered (

).

Aggregate

Displays the name of the aggregate that contains the volume.

Storage Server

Displays the name of the storage server that contains the volume. The storage
server can be a storage controller, Vserver, or a vFiler unit.

Daily Growth Rate Displays, in GB, the amount of disk space used in the volume, if the amount
of change between the last two samples continues for 24 hours.
(GB)
Data Days To Full Displays the estimated time left before this volume runs out of storage space.
If the estimated time is less than one day, the current storage status of the
(GB)
volume is displayed.
Daily Growth Rate Displays the percentage of change in the used space in the volume reserve, if
the change between the last two samples continues for 24 hours.
(%)
Related references

Reports tab customization on page 304


Reports management on page 303
Volumes Space Reservation report
The Volumes Space Reservation report displays information about the space reservation and the
space-reserved files in the volume.
Note: To display the icons in a report, you must ensure that a DNS mapping is established between
the client machine from which you are starting the Web connection and the host name of the
system on which the DataFabric Manager server is installed.

The following details are displayed for volumes that have space-reserved files.

350 | OnCommand Console Help


Volume

Displays the name of the volume. The icon indicates if the volume is
clustered (

) or nonclustered (

).

Aggregate

Displays the name of the aggregate that contains the volume.

Storage Server

Displays the name of the storage server that contains the volume. The
storage server can be a storage controller, Vserver, or a vFiler unit.

Reserved Files Total


Size (GB)

Displays the total size of the space-reserved files in this volume.

Fractional Reserve
(%)

Controls the size of the overwrite reserve. If the fractional reserve is less
than 100 percent, the reserved space for all the space-reserved files in that
volume is reduced to the fractional reserve percentage. By default, this
column is hidden.

Reservation Used
(GB)

Displays the total space reservation used for overwrites in this volume. By
default, this column is hidden.

Reservation Available Displays the amount of free space remaining in the space reservation. By
default, this column is hidden.
(GB)
Space Reserve Total
(GB)

Displays the total size of the space reserve for this volume.

Space Reservation
Used (GB)

Displays the total space reserve currently in use.

Status

Displays the current status of the volume based on the events generated
for the volume. The status can be Normal, Warning, Error, Critical,
Emergency, or Unknown.

Space Reserve
Depleted Threshold
(%)

Displays the percentage of the threshold at which a volume is considered


to have consumed all of its reserved space. A volume that has crossed this
threshold is getting dangerously close to having write failures.

Space Reserve Nearly


Depleted Threshold
(%)

Displays the percentage of the threshold at which the space reserve is


likely to be nearly depleted. By default, this column is hidden.

Space Reserve Used


(%)

Displays the percentage of the total space reserve currently in use.

Related references

Types of object status on page 305


Reports tab customization on page 304
Reports management on page 303

Reports | 351

Aggregates Space Savings report


The Aggregates Space Savings report displays information about the space savings achieved in the
aggregate through deduplication.
Note: To display the charts and icons in a report, you must ensure that a DNS mapping is

established between the client machine from which you are starting the Web connection and the
host name of the system on which the DataFabric Manager server is installed.
Chart
You can customize the data displayed in the chart, change the subtype and format of the chart, and
export the data from the chart.
Report details
The following details are displayed when the space savings on the aggregate is enabled.
Aggregate

Displays the name of the aggregate.

Storage System

Displays the name of the storage system that contains the aggregate.

Physical Used (GB)

Displays the active file system data of all the deduplication-enabled


volumes in the aggregate.

Dedupe Space Savings


(GB)

Displays the savings achieved in an aggregate through deduplication.

Effective Used (GB)

Displays the active file system data of all the deduplicated volumes in
the aggregate without deduplication space savings.

Dedupe Space Savings


(%)

Displays the percentage of savings achieved in the aggregate through


deduplication.

Volume Enabled

Displays the number of deduplication-enabled volumes in the


aggregate.

Available Capacity (GB) Displays the amount of space available for data in the aggregate.
Total Capacity (GB)

Displays the total amount of storage allocated for this aggregate.

Related references

Reports tab customization on page 304


Reports management on page 303

352 | OnCommand Console Help

Volumes Space Savings report


The Volumes Space Savings report displays information about the space savings achieved in the
volume through deduplication.
Note: To display the charts and icons in a report, you must ensure that a DNS mapping is
established between the client machine from which you are starting the Web connection and the
host name of the system on which the DataFabric Manager server is installed.

Chart
You can customize the data displayed in the chart, change the subtype and format of the chart, and
export the data from the chart.
Report details
The following details are displayed when the space savings on the volume is enabled.
Volume

Displays the name of the volume. The icon indicates if the volume is
clustered (

) or nonclustered (

).

Storage Server

Displays the name of the storage server that contains the volume. The
storage server can be a storage controller, Vserver, or a vFiler Unit.

Dedupe Status

Specifies whether deduplication is enabled or disabled on the volume.

Last Dedupe Run

Displays the timestamp of the last run deduplication operation on the


volume.

Physical Used (GB)

Displays the active file system data in the volume with deduplication space
savings.

Dedupe Space
Savings (GB)

Displays the savings achieved in a volume through deduplication.

Effective Used (GB)

Displays the active file system data in the volume without deduplication
space savings (that is, if deduplication has not been enabled on the
volume).

Space Savings (%)

Displays the percentage of savings achieved in a volume through


deduplication.

Available Capacity
(GB)

Displays the amount of space available for data in the volume.

Total Capacity (GB)

Displays the total amount of storage allocated for this volume.

Related references

Reports tab customization on page 304

Reports | 353

Reports management on page 303


Datasets Average Space Usage Metric report
The Datasets Average Space Usage Metric report provides the average space utilization for each
node computed at specified intervalsfor example, one hour. The computed metrics is consolidated
and reported for a specific periodfor example, one month.
Dataset

Displays the name of the dataset.

Storage Service

Displays the name of the storage service associated with the dataset at the
time of computing the metrics. If the storage service changes for a dataset, the
new dataset details are displayed in a new row. The field is blank if a storage
service is not associated with the dataset.

Protection Policy

Displays the name of the protection policy associated with the dataset at the
time of computing the metrics. If the protection policy changes for a dataset,
the new dataset details are displayed in a new row. The field is blank if a
protection policy is not associated with the dataset.

Dataset Node

Displays the name of the dataset node.

Provisioning
Policy

Displays the name of the provisioning policy associated with the dataset at
the time of computing the metrics. If the provisioning policy changes for a
dataset, the new dataset details are displayed in a new row. The field is blank
if a provisioning policy is not associated with the dataset.

Effective Used
Data Space

Displays the space used by user data in the dataset node, without accounting
for data sharing.

Physical Used
Data Space

Displays the actual space used by user data in the dataset node, accounting for
the space saved by data sharing.

Used Snapshot
Space

Displays the physical space used by the volume Snapshot copies in the
dataset node.

Total Data Space

Displays the space allocated to the dataset's primary node. The field is blank
for non-primary nodes.

Snapshot Reserve

Displays the space allocated for Snapshot copies in the dataset's primary
node. The field is blank for non-primary nodes.

Total Space

Displays the space allocated for data and Snapshot copies in the dataset's
primary node. The field is blank for non-primary nodes.

Guaranteed Space Displays the physical space allocated to the dataset node.
Metric Period

Displays the period (in days) for which metrics is calculated for the dataset.

Deleted

Displays the time at which the dataset is deleted.

354 | OnCommand Console Help


Comments

Displays the custom comments specified by the user. If custom comments


change during the reporting period, both the new and the old comments are
displayed.

Notes

Displays any discrepancy in metrics computation. The field value is


overcharge if the dataset has one or more qtrees in the primary node. The
field value is partial if data of some volume does not exist for a period.
The field is blank if there are no issues.

Related references

Reports tab customization on page 304


Reports management on page 303
Datasets Maximum Space Usage Metric report
The Datasets Maximum Space Usage Metric report provides the maximum space utilization for each
node computed at specified intervalsfor example, one hour. The computed metrics is consolidated
and reported for a specific periodfor example, one month.
Dataset

Displays the name of the dataset.

Storage Service

Displays the name of the storage service associated with the dataset at the
time of computing the metrics. If the storage service changes for a dataset, the
new dataset details are displayed in a new row. The field is blank if a storage
service is not associated with the dataset.

Protection Policy

Displays the name of the protection policy associated with the dataset at the
time of computing the metrics. If the protection policy changes for a dataset,
the new dataset details are displayed in a new row. The field is blank if a
protection policy is not associated with the dataset.

Dataset Node

Displays the name of the dataset node.

Provisioning
Policy

Displays the name of the provisioning policy associated with the dataset at
the time of computing the metrics. If the provisioning policy changes for a
dataset, the new dataset details are displayed in a new row. The field is blank
if a provisioning policy is not associated with the dataset.

Effective Used
Data Space

Displays the space used by user data in the dataset node, without accounting
for data sharing.

Physical Used
Data Space

Displays the actual space used by user data in the dataset node, accounting for
the space saved by data sharing.

Used Snapshot
Space

Displays the physical space used by the volume Snapshot copies in the
dataset node.

Total Data Space

Displays the space allocated in the dataset's primary node. The field is blank
for non primary nodes.

Reports | 355
Snapshot Reserve

Displays the space allocated for Snapshot copies in the dataset's primary
node. The field is blank for non primary nodes.

Total Space

Displays the space allocated for data and Snapshot copies in the dataset's
primary node. The field is blank for non primary nodes.

Guaranteed Space Displays the physical space allocated to the dataset node.
Metric Period

Displays the period (in days) for which metrics is calculated for the dataset.

Deleted

Displays the time at which the dataset is deleted.

Comments

Displays the custom comments specified by the user. If custom comments


change during the reporting period, both the new and the old comments are
displayed.

Notes

Displays any discrepancy in metrics computation. The field value is


overcharge if the dataset has one or more qtrees in the primary node. The
field value is partial if data of some volume does not exist for a period.
The field is blank if there are no issues.

Related references

Reports tab customization on page 304


Reports management on page 303
Datasets IO Usage Metric report
The Datasets IO Usage Metric report provides user data read and written for each dataset node for a
specific periodfor example, one week. The consolidated sample is reported for a specific period
for example, one month.
Dataset

Displays the name of the dataset.

Storage Service

Displays the name of the storage service associated with the dataset at the time
of computing the metrics. If the storage service changes for a dataset, the new
dataset details are displayed in a new row. The field is blank if a storage service
is not associated with the dataset.

Protection Policy Displays the name of the protection policy associated with the dataset at the
time of computing the metrics. If the protection policy changes for a dataset,
the new dataset details are displayed in a new row. The field is blank if a
protection policy is not associated with the dataset.
Dataset Node

Displays the name of the dataset node.

Provisioning
Policy

Displays the name of the provisioning policy associated with the dataset at the
time of computing the metrics. If the provisioning policy changes for a dataset,
the new dataset details are displayed in a new row. The field is blank if a
provisioning policy is not associated with the dataset.

356 | OnCommand Console Help


Data Read

Displays the total data read by the user from all volumes of the dataset node.

Data Written

Displays the total data written by the user to all volumes of the dataset node.

Metric Period

Displays the period (in days) for which metrics is calculated for the dataset.

Deleted

Displays the time at which the dataset is deleted.

Notes

Displays any discrepancy in metrics computation. The field value is


overcharge if the dataset has one or more qtrees in the primary node. The
field value is partial if data of some volume does not exist for a period. The
field is blank if there are no issues.

Comments

Displays the custom comments specified by the user. If custom comments


change during the reporting period, both the new and the old comments are
displayed.

Related references

Reports tab customization on page 304


Reports management on page 303
Datasets Average Space Usage Metric Samples report
The Datasets Average Space Usage Metric Samples report provides the average space utilization for
each node computed at specified intervals. The computed metrics is consolidated and reported at the
intervals at which the sample is collected. For example, if the interval is set to one hour, the report
can collect and display the average space utilization for a dataset at each hour.
Dataset

Displays the name of the dataset.

Storage Service

Displays the name of the storage service associated with the dataset at the
time of computing the metrics. If the storage service changes for a dataset, the
new dataset details are displayed in a new row. The field is blank if a storage
service is not associated with the dataset.

Protection Policy

Displays the name of the protection policy associated with the dataset at the
time of computing the metrics. If the protection policy changes for a dataset,
the new dataset details are displayed in a new row. The field is blank if a
protection policy is not associated with the dataset.

Dataset Node

Displays the name of the dataset node.

Provisioning
Policy

Displays the name of the provisioning policy associated with the dataset at
the time of computing the metrics. If the provisioning policy changes for a
dataset, the new dataset details are displayed in a new row. The field is blank
if a provisioning policy is not associated with the dataset.

Timestamp

Displays the time at which sample is collected.

Reports | 357
Effective Used
Data Space

Displays the space used by user data in the dataset node, without accounting
for data sharing.

Physical Used
Data Space

Displays the actual space used by user data in the dataset node, accounting for
the space saved by data sharing.

Used Snapshot
Space

Displays the physical space used by the volume Snapshot copies in the
dataset node.

Total Data Space

Displays the space allocated in the dataset's primary node. The field is blank
for non primary nodes.

Snapshot Reserve

Displays the space allocated for Snapshot copies in the dataset's primary
node. The field is blank for non primary nodes.

Total Space

Displays the space allocated for data and Snapshot copies in the dataset's
primary node. The field is blank for non primary nodes.

Guaranteed Space Displays the physical space allocated to the dataset node.
Deleted

Displays the time at which the dataset is deleted.

Notes

Displays any discrepancy in metrics computation. The field value is


overcharge if the dataset has one or more qtrees in the primary node. The
field value is partial if data of some volume does not exist for a period.
The field is blank if there are no issues.

Comments

Displays the custom comments specified by the user. If custom comments


change during the reporting period, both the new and the old comments are
displayed.

Related references

Reports tab customization on page 304


Reports management on page 303
Datasets Maximum Space Usage Metric Samples report
The Datasets Maximum Space Usage Metric Samples report provides the maximum space utilization
for each node computed at specified intervals. The computed metrics is consolidated and reported at
the intervals at which the sample is collected. For example, if the interval is set to one hour, the
report can collect and display the maximum space utilization for a dataset at each hour.
Dataset

Displays the name of the dataset.

Storage Service

Displays the name of the storage service associated with the dataset at the
time of computing the metrics. If the storage service changes for a dataset,
the new dataset details are displayed in a new row. The field is blank if a
storage service is associated with the dataset.

358 | OnCommand Console Help


Protection Policy

Displays the name of the protection policy associated with the dataset at the
time of computing the metrics. If the protection policy changes for a dataset,
the new dataset details are displayed in a new row. The field is blank if a
protection policy is not associated with the dataset.

Dataset Node

Displays the name of the dataset node.

Provisioning Policy Displays the name of the provisioning policy associated with the dataset at
the time of computing the metrics. If the provisioning policy changes for a
dataset, the new dataset details are displayed in a new row. The field is blank
if a provisioning policy is not associated with the dataset.
Timestamp

Displays the time at which sample is collected.

Effective Used Data Displays the space used by user data in the dataset node, without accounting
for data sharing.
Space
Physical Used Data Displays the actual space used by user data in the dataset node, accounting
for the space saved by data sharing.
Space
Used Snapshot
Space

Displays the physical space used by the volume Snapshot copies in the
dataset node.

Total Data Space

Displays the space allocated in the dataset's primary node.

Snapshot Reserve

Displays the space allocated for Snapshot copies in the dataset's primary
node.

Total Space

Displays the space allocated for data and Snapshot copies in the dataset's
primary node.

Guaranteed Space

Displays the physical space allocated to the dataset node.

Deleted

Displays the time at which the dataset is deleted.

Notes

Displays any discrepancy in metrics computation. The field value is


overcharge if the dataset has one or more qtrees in the primary node. The
field value is partial if data of some volume does not exist for a period.
The field is blank if there are no issues.

Related references

Reports tab customization on page 304


Reports management on page 303
Datasets IO Usage Metric Samples report
The Datasets IO Usage Metric Sample report provides user data read and written for each dataset
node for a specific period. Consolidated sample is reported at the same interval at which the sample

Reports | 359
is collected. For example, if the interval is set to one hour, the report can collect and display the data
read for a dataset at each hour.
Dataset

Displays the name of the dataset.

Storage Service

Displays the name of the storage service associated with the dataset at the time
of computing the metrics. If the storage service changes for a dataset, the new
dataset details are displayed in a new row. The field is blank if a storage service
is not associated with the dataset.

Protection Policy Displays the name of the protection policy associated with the dataset at the
time of computing the metrics. If the protection policy changes for a dataset,
the new dataset details are displayed in a new row. The field is blank if a
protection policy is not associated with the dataset.
Dataset Node

Displays the name of the dataset node.

Provisioning
Policy

Displays the name of the provisioning policy associated with the dataset at the
time of computing the metrics. If the provisioning policy changes for a dataset,
the new dataset details are displayed in a new row. The field is blank if a
provisioning policy is not associated with the dataset.

Timestamp

Displays the time at which sample is collected.

Data Read

Displays the total data read by the user from all volumes of the dataset node.

Data Written

Displays the total data written by the user to all volumes of the dataset node.

Deleted

Displays the time at which the dataset is deleted.

Notes

Displays any discrepancy in metrics computation. The field value is


overcharge if the dataset has one or more qtrees in the primary node. The
field value is partial if data of some volume does not exist for a period. The
field is blank if there are no issues.

Related references

Reports tab customization on page 304


Reports management on page 303

Database schema
How to access DataFabric Manager server data
By using third-party tools, you can create customized reports from the data you export from the
DataFabric Manager server. By default, you cannot access the DataFabric Manager server views. To

360 | OnCommand Console Help


gain access to the views that are defined within the embedded database of the DataFabric Manager
server, you need to first create a database user and then enable database access to this user.
You can access the DataFabric Manager server data through views, which are dynamic virtual tables
collated from data in the database. These views are defined and exposed within the embedded
database of the DataFabric Manager server.
Note: A database user is user-created and authenticated by the database server. Database users are
not related to the DataFabric Manager server users.

Before you can create and give access to a database user, you must have the CoreControl capability.
The CoreControl capability allows you to perform the following operations:

Creating a database user


Deleting a database user
Enabling database access to a database user
Disabling database access to a database user
Disabling the database access denies the read permission on the DataFabric Manager server views
for the user account.
Changing the password for the database user

All of these operations can be performed only through the CLI. For more information about the CLI
commands, see the DataFabric Manager server manual (man) pages.
You can use a third-party reporting tool to connect to the DataFabric Manager server database for
accessing views. Following are the connection parameters:

Database name: monitordb


User name: <database user name>
Password: <database user password>
Port: 2638
dobroad: none
Links: tcpip
Note: The .jar files required for iAnywhere and jConnect JDBC drivers are copied as part of the
DataFabric Manager server installation. The new jar files are saved in the following directory
path: .../install/misc/dbconn.

Supported database views


By using third-party tools, you can create customized reports from the data you export from the
DataFabric Manager server. You can access the DataFabric Manager server database through views.
Following is the list of supported views:

alarmView
cpuView
designerReportView

Reports | 361

datasetIOMetricView
datasetSpaceMetricView
datasetUsageMetricCommentView
hbaInitiatorView
hbaView
initiatorView
reportOutputView
sanhostLunView
usersView
volumeDedupeDetailsView

alarmView
Column name

Type

Length Description

alarmId

Unsigned
integer

Primary key in identifying alarms

alarmScript

Varchar

254

The script that is run when an alarm is


triggered

alarmScriptRunAs

Varchar

64

The user account specified to run the


script

alarmTrapHosts

Varchar

254

The SNMP traphost system that receives


the alarm notification in the form of
SNMP traps

alarmGroupId

Unsigned
integer

Group ID of the group with which the


alarm is associated

alarmGroupName

Varchar

1024

Name of the group with which the alarm


is associated

alarmEventClass

Varchar

254

The event class configured to trigger the


alarm

alarmEventType

Varchar

128

The event name that triggers the alarm

alarmEventSeverity

Varchar

16

The event severity level

alarmEventTimeFrom

Date time

The time at which the alarm becomes


active

alarmEventTimeTo

Date time

The time at which the alarm becomes


inactive

alarmsRepeatNotify

Varchar

The alarm repeat notification

362 | OnCommand Console Help


Column name

Type

Length Description

alarmRepeatInterval

Unsigned
small integer

The time period (in minutes) before the


DataFabric Manager server continues to
send a repeated notification until the event
is acknowledged or resolved

alarmDisabled

Varchar

The alarm is disabled on enabled

alarmEmailAddrs

Varchar

254

The e-mail address to which the alarm


notification is sent

alarmEventName

Varchar

256

The event name associated with the alarm

alarmEventPageLoginName

Text

32767

The login name used by DataFabric


Manager server for sending pager
notification

alarmAdminEmailLoginName Text

32767

The login name of the administrator who


receives the alarm notification

alarmAdminPageAddress

Text

32767

The pager address of the administrator, to


which the alarm notification is sent

alarmAdminEmailAddress

Text

32767

The e-mail address of the administrator, to


which the alarm notification is sent

alarmPageAddrs

Varchar

254

The pager address of the


nonadministrator, to which the alarm
notification is sent

cpuView
Column name

Type

Length Description

cpuId

Unsigned
integer

Primary key to identify CPU

Reports | 363
Column name

Type

Length Description

cpuBusyPercentInterval Float

The value (as time interval) controls the


generation of "cpu too busy" event
The value, cpubusy (in percentage) is compared
with cpuTooBusyThreshold value. If this value is
greater than or equal to cpuTooBusyThreshold,
and persists throughout the
cpuBusyThresholdInterval period, then "cpu too
busy" event is generated.
If cpuBusyThresholdInterval is set to "15
minutes" (default interval), these values are
checked every cpu monitoring cycle.

cpuStatTimestamp

Timestamp

The time-stamp when the statistics of CPU is


taken.

designerReportView
Column name Type

Length Description

drId

Unsigned integer 4

The designer report ID

drCliName

Varchar

64

The CLI name of the designer report

drGuiName

Varchar

64

The GUI name of the designer report

drIsCustom

Unsigned integer 4

The flag to differentiate if the designer report is a


custom report or a canned report
0 indicates Canned and 1 indicates Custom

drDescription

Varchar

1024

Brief description of the designer Report

Database view datasetIOMetricView


Column Name

Data type

Length

Description

dsIOMetricDatasetId

unsigned int 4

Dataset ID

dsIOMetricDatasetName

varchar

255

Dataset name

dsIOMetricProtectionPolicyName

varchar

255

Protection policy name that is


associated with the dataset

dsIOMetricProtectionPolicyId

unsigned int 4

Protection policy ID that is associated


with the dataset

364 | OnCommand Console Help


Column Name

Data type

Length

Description

dsIOMetricStorageServiceName

varchar

255

Storage service name that is


associated with the dataset

dsIOMetricStorageServiceId

unsigned int 4

Storage service ID that is associated


with the dataset

dsIOMetricNodeName

varchar

255

Dataset node name

dsIOMetricProvisioningPolicyNam varchar
e

255

Provisioning policy name that is


associated with the dataset node

dsIOMetricProvisioningPolicyId

unsigned int 4

Provisioning policy ID that is


associated with the dataset node

dsIOMetricMetricTimestamp

timestamp

Time at which metric is calculated


for the dataset

dsIOMetricTotalDataRead

unsigned
big int

Total data read by the user from all


volumes of the dataset node

dsIOMetricTotalDataWritten

unsigned
big int

Total data written by the user to all


volumes of the dataset node

dsIOMetricDeletedTimestamp

timestamp

Time at which the dataset is deleted

dsIOMetricOvercharge

bit

Flag true if the computed metric is


overcharged

dsIOMetricPartialData

bit

Flag true if some samples required


for metric computation are missing

dsIOMetricCommentId

unsigned int 4

ID of custom comment name and


custom value pair set for the dataset

Database view datasetSpaceMetricView


Column Name

Data type

Length

Description

dsSpaceMetricDatasetId

unsigned
int

Dataset ID

dsSpaceMetricDatasetName

varchar

255

Dataset name

dsSpaceMetricProtectionPolicyNa
me

varchar

255

Protection policy name that is


associated with the dataset

dsSpaceMetricProtectionPolicyId

unsigned
int

Protection policy ID that is


associated with the dataset

dsSpaceMetricNodeName

varchar

255

Dataset node name

Reports | 365
Column Name

Data type

Length

Description

dsSpaceMetricStorageServiceNam
e

varchar

255

Storage service name that is


associated with the dataset

dsSpaceMetricStorageServiceId

unsigned
int

Storage service ID that is associated


with the dataset

dsSpaceMetricProvisioningPolicyN varchar
ame

255

Provisioning policy name that is


associated with the dataset node

dsSpaceMetricProvisioningPolicyI
d

unsigned
int

Provisioning policy ID that is


associated with the dataset node

dsSpaceMetricTimestamp

timestamp

Time at which metric is calculated


for the dataset

dsSpaceMetricAvgEffectiveUsedD unsigned
ataSpace
big int

The average space used by user data


in the dataset node, without
accounting for data sharing

dsSpaceMetricAvgPhysicalUsedDa unsigned
taSpace
big int

The average actual space used by


user data in the dataset node,
accounting for the space saved by
data sharing

dsSpaceMetricAvgUsedSnapshotS
pace

unsigned
big int

The average physical space used by


the volume Snapshot copies in the
dataset node

dsSpaceMetricAvgTotalDataSpace unsigned
big int

The average space allocated in the


dataset's primary node. The field is
blank for non primary nodes

dsSpaceMetricAvgSnapshotReserv unsigned
e
big int

The average space allocated for


Snapshot copies in the dataset's
primary node. The field is blank for
non primary nodes

dsSpaceMetricAvgTotalSpace

unsigned
big int

The average space allocated for data


and Snapshot copies in the dataset's
primary node. The field is blank for
non primary nodes

dsSpaceMetricAvgGuaranteedSpac unsigned
e
big int

The average physical space


allocated to the dataset node

dsSpaceMetricMaxEffectiveUsedD unsigned
ataSpace
big int

The maximum space used by user


data in the dataset node, without
accounting for data sharing

366 | OnCommand Console Help


Column Name

Data type

Length

Description

dsSpaceMetricMaxPhysicalUsedD
ataSpace

unsigned
big int

The maximum actual space used by


user data in the dataset node,
accounting for the space saved by
data sharing

dsSpaceMetricMaxUsedSnapshotS unsigned
pace
big int

The maximum physical space used


by the volume Snapshot copies in
the dataset node

dsSpaceMetricMaxTotalDataSpace unsigned
big int

The maximum space allocated in


the dataset's primary node. The field
is blank for non primary nodes

dsSpaceMetricMaxSnapshotReserv unsigned
e
big int

The maximum space allocated for


Snapshot copies in the dataset's
primary node. The field is blank for
non primary nodes

dsSpaceMetricMaxTotalSpace

unsigned
big int

The maximum space allocated for


data and Snapshot copies in the
dataset's primary node. The field is
blank for non primary nodes

dsSpaceMetricMaxGuaranteedSpac unsigned
e
big int

The maximum physical space


allocated to the dataset node

dsSpaceMetricDeletedTimestamp

timestamp

Time at which the dataset is deleted

dsSpaceMetricOvercharge

bit

Flag true if the computed metric is


overcharged

dsSpaceMetricPartialData

bit

Flag true if some samples required


for metric computation are missing

dsSpaceMetricCommentId

unsigned
int

ID of custom comment name and


custom value pair set for the dataset

Database view datasetUsageMetricCommentView


Column Name

Data type

Length Description

datasetId

unsigned int

Dataset ID

dsUsageMetricCommentId unsigned int

ID of custom comment name and comment


value set for the dataset at a specified time

dsUsageMetricCommentN varchar
ame

255

Custom comment name

Reports | 367
Column Name

Data type

Length Description

dsUsageMetricCommentV varchar
alue

255

Custom comment value

hbaInitiatorView
Column name

Type

Length

Description

initiatorId

Unsigned integer

Primary key for initiator

hbaId

Unsigned integer

WWPN ID of the HBA port

hbaView
Column name

Type

Length

Description

hbaId

Unsigned integer

Primary key for WWPN of the HBA port

hbaName

Varchar

64

WWPN of the HBA port

initiatorView
Column name

Type

Length Description

initiatorId

Unsigned integer

Primary key for initiator

iGroupId

Unsigned integer

Initiator group ID to which the LUN is mapped

initiatorName

Varchar

255

Name of the initiator to which the LUN is mapped

reportOutputView
Column name

Type

Length Description

reportOutputId

Unsigned
integer

Report output ID

reportScheduleId

Unsigned
integer

ID of the report schedule that generates the


report output

reportId

Unsigned
integer

ID of the report for which the output is


generated
If it is null, it means the report is a canned
legacy report.

368 | OnCommand Console Help


Column name

Type

Length Description

reportName

Varchar

64

Name of the report for which the output is


generated

reportBaseCatalog

Varchar

64

The name of the base catalog that


corresponds to the custom report
The DataFabric Manager server provides
report catalogs that you use to customize
reports. You can set basic report properties
from the CLI. If reportBaseCatalog is null,
it means the report is either a canned legacy
report or a canned OnCommand report.

reportOutputTargetObjId

Unsigned big
integer

ID of the object for which the report is


scheduled

reportOutputTimestamp

Timestamp

The time which the report output is


generated

reportOutputRunStatus

Unsigned
small integer

The status of the report output

reportOutputRunBy

Varchar

255

The user (with privileges) who generated


the report

reportOutputFailureReason Varchar

255

The reason for failure of a report generation

reportOutputFileName

128

The file in which the report output is saved

Varchar

sanhostlunview
Column name Type

Length Description

shlunId

Unsigned integer 4

LUN ID

hostId

Unsigned integer 4

SAN host ID

shInitiatorId

Unsigned integer 4

Initiator ID to which the LUN is mapped

shlunpathId

Unsigned integer 4

Path ID of the storage system on which the LUN is


located

usersView
Column name

Type

Length Description

userId

Unsigned
integer

The User Quota ID

Reports | 369
Column name

Type

userNearlyFullThreshold Unsigned
small
integer

Length Description
2

The value (in percentage) at which most of the


allocated space (disk space or files used), as
specified by the users quota (hard limit in
the /etc/quotas file), is consumed
If this threshold is crossed, the DataFabric
Manager server generates a User Disk Space
Quota Almost Full event or a User Files Quota
Almost Full event.

userFullThreshold

Unsigned
small
integer

The value (in percentage) at which most of the


allocated space (disk space or files used), as
specified by the users quota (hard limit in
the /etc/quotas file), is consumed
If this threshold is crossed, the DataFabric
Manager server generates a User Disk Space
Quota Full event or a User Files Quota Full
event.

volumeDedupeDetailsView
Column name

Type

Length Description

volumeId

Unsigned 4
integer

Volume ID

volumeOverDedupeThreshold

Unsigned 2
small
integer

The threshold (as an integer


percentage), used to generate "Volume
Over Deduplicated" event on a dedupe
enabled volume
This event is generated if the sum of
volume used space and saved space (as
a result of deduplication), expressed as
a percentage of the volume total size,
exceeds the value of the
volOverDeduplicatedThreshold set on
the volume. If the threshold is set on the
volume, the value overrides the one at
the global level. If
volumeOverDedupeThreshold is null, it
means that the threshold is not set at the
volume level.

370 | OnCommand Console Help


Column name

Type

Length Description

volumeNearlyOverDedupeThreshold Unsigned 2
small
integer

The threshold (as an integer


percentage), used to generate "Volume
Nearly Over Deduplicated" event on a
dedupe enabled volume
This event is generated if the sum of
volume used space and saved space (as
a result of deduplication), expressed as
a percentage of the volume total size,
exceeds the value of the
volNearlyOverDeduplicatedThreshold
set on the volume. If the threshold is set
on the volume, the value overrides the
one at the global level. If
volumeNearlyOverDedupeThreshold is
null, it means that the threshold is not
set at the volume level.

371

Administration
Users and roles
Understanding users and roles
What RBAC is
RBAC (role-based access control) provides the ability to control who has access to various features
and resources in DataFabric Manager server.
How RBAC is used
Applications use RBAC to authorize user capabilities. Administrators use RBAC to manage groups
of users by defining roles and capabilities.
For example, if you need to control user access to resources, such as groups, datasets, and resource
pools, you must set up administrator accounts for them. Additionally, if you want to restrict the
information these administrators can view and the operations they can perform, you must apply roles
to the administrator accounts you create.
Note: RBAC permission checks occur in the DataFabric Manager server. RBAC must be
configured using the Operations Manager console or command line interface.

How roles relate to administrators


Role management allows the administrator who logs in with super-user access to restrict the use of
certain DataFabric Manager server functions to other administrators.
The super-user can assign roles to administrators on an individual basis, by group, or globally (and
for all objects in DataFabric Manager server).
You can list the description of an operation by using the dfm role operation list [ -x ]
[ <operation-name> ] command.
The ability to configure administrative users and roles is supported in the Operations Manager
console, which can be accessed from the OnCommand console Administration menu.
Example of how to use RBAC to control access
This example describes how a storage architect can use RBAC to control the operations that can be
performed by a virtual server administrator.
Suppose you are a storage architect and you want to use RBAC to enable the virtual server
administrator (abbreviated to "administrator") to do the following operations: see the VMs associated

372 | OnCommand Console Help


with the servers managed by the administrator, create datasets to include them, and attach storage
services and application policies to these datasets to back up the data.
Assume for this example that the host service registration, validation, and so on have been
successfully completed and that DataFabric Manager server has discovered the virtual server and its
VMs, datastores, and so on. However, at this point, only you and other administrators with the global
read permission in DataFabric Manager server can see these VMs. To enable the virtual server
administrator to perform the desired operations, you need to perform the following steps:
1. Add the administrator as a user.
You add the administrator as an authorized DataFabric Manager server user by using the
Operations Manager console. If DataFabric Manager server is running on Linux, then you must
add the administrator's UNIX identity or LDAP identity. If DataFabric Manager server is running
on Microsoft Windows, you can add an active directory user group that the administrator belongs
to, which allows all administrators in that user group to log onto DataFabric Manager server.
2. Create a resource group.
You next create a resource group by using the Operations Manager console. For this example, we
call the group "virtual admin resource group." Then you add the virtual server and its objects to
the resource group. Any new datasets or policies that the administrator creates will be placed in
this resource group.
3. Assign a role to the administrator.
Assign a role, by using the Operations Manager console, that gives an appropriate level of control
to the administrator. For example, if there is only one administrator, you can assign the default
GlobalApplicationProtection role or you can create a custom role by choosing custom
capabilities. If there are multiple administrators, then a few of the capabilities assigned to any
given administrator should be at that administrator's group level and not on the global level. That
prevents an administrator from reading or modifying objects owned by other administrators.
Related tasks

Accessing the Users and Roles capability (RBAC) on page 375


Administrator roles and capabilities
The RBAC administrator roles determine the tasks you can perform in the OnCommand console.
One or more capabilities must be specified for every role, and you can assign multiple capabilities if
you want the administrator to have more control than a specific role provides. For example, if you
want an administrator to perform both the backup and restore operations, you can create and assign to
the administrator a single role that has both of these capabilities.
You can use the Operations Manager console to create new roles and to customize the default global
roles provided by the DataFabric Manager server and the client applications. For more information
about configuring RBAC, see the OnCommand Operations Manager Administration Guide .
Note: If you want a role with global host service management capability, create a role with the
following properties:

Administration | 373

The role inherits from the GlobalHostService role.


The role includes the DFM.Database.Read operation on a global level.

Note: A user who is part of the local administrators group is treated as a super-user and

automatically granted full control.


Default global roles
GlobalApplicationProtection Enables you to create and manage application policies, create
datasets with application policies for local backups, use storage
services for remote backups, perform scheduled and on-demand
backups, perform restore operations, and generate reports.
GlobalBackup

Enables you to initiate a backup to any secondary volume and


ignore discovered hosts.

GlobalDataProtection

Enables you to initiate a backup to any secondary volume; view


backup configurations, events and alerts, and replication or failover
policies; and import relationships into datasets.

GlobalDataset

Enables you to create, modify, and delete datasets.

GlobalDelete

Enables you to delete information in the DataFabric Manager


server database, including groups and members of a group,
monitored objects, custom views, primary and secondary storage
systems, and backup relationships, schedules, and retention
policies.

GlobalHostService

Enables you to authorize, configure, and unregister a host service.

GlobalEvent

Enables you to view, acknowledge, and delete events and alerts.

GlobalFullControl

Enables you to view and perform any operation on any object in the
DataFabric Manager server database and configure administrator
accounts. You cannot apply this role to accounts with group access
control.

GlobalMirror

Enables you to create, destroy, and update replication or failover


policies.

GlobalRead

Enables you to view the DataFabric Manager server database,


backup and provisioning configurations, events and alerts,
performance data, and policies.

GlobalRestore

Enables you to restore the primary data to a point in time or to a


new location.

GlobalWrite

Enables you to view or write both primary and secondary data to


the DataFabric Manager server database.

374 | OnCommand Console Help


GlobalResourceControl

Enables you to add members to dataset nodes that are configured


with provisioning policies.

GlobalProvisioning

Enables you to provision primary dataset nodes and can attach


resource pools to secondary or tertiary dataset nodes. The
GlobalProvisioning role also includes all the capabilities of the
GlobalResourceControl, GlobalRead, and GlobalDataset roles for
dataset nodes that are configured with provisioning and protection
policies.

GlobalPerfManagement

Enables you to manage views, event thresholds, and alarms apart


from viewing performance information in Performance Advisor.

Related information

Operations Manager Administration Guide - http://now.netapp.com/NOW/knowledge/docs/


DFM_win/dfm_index.shtml
Access permissions for the Virtual Infrastructure Administrator role
When you create a virtual infrastructure administrator, you must assign specific permissions to
ensure that the administrator can view, back up, and recover the appropriate virtual objects.
A virtual infrastructure administrator role must have the following permissions for the resources:
Groups

Policies

The VI administrator will need the following operation permissions for the group
created for the VI administrator role.
DFM.Database

All

DFM.BackManager

All

DFM.ApplicationPolicy

All

DFM.Dataset

All

DFM.Resource

Control

The VI administrator will need the following operation permissions for each policy
template, located under Local Policies, that you want the virtual administrator to be
able to copy.
DFM.ApplicationPolicy

Storage
services

Read

The VI administrator will need the following operation permissions for each of the
storage services that you want to allow the VI administrator to use.
DFM.StorageService

Attach, read, detach, and clear

Administration | 375
Protection
Policies

These are the policies contained within the storage services that you selected above.
DFM.Policy

All

Configuring users and roles


Accessing the Users and Roles capability (RBAC)
You can configure users and roles from the Administrators page in the Operations Manager console,
which you can access from the OnCommand console Administration menu.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

During this task, the OnCommand console launches the Operations Manager console. Depending on
your browser configuration, you can return to the OnCommand console by using the Alt-Tab key
combination or clicking the OnCommand console browser tab. After the completion of this task, you
can leave the Operations Manager console open, or you can close it to conserve bandwidth.
Steps

1. Click the Administration menu, then click the Users and Roles option.
A separate browser window opens to the Administrators page in the Operations Manager console.
2. Configure users and roles.
For more information, see the Operations Manager Help.
3. When finished, press Alt-Tab or click the OnCommand console browser tab to return to the
OnCommand console.
Related references

Administrator roles and capabilities on page 506

Groups
Understanding groups

376 | OnCommand Console Help

What a global group is


By default, a group called Global exists in the DataFabric Manager server database. All objects
monitored by DataFabric Manager server belong to the global group. All subgroups created also
belong to the global group.
You cannot delete or rename the global group. When you delete an object, the DataFabric Manager
server stops monitoring and reporting data for that object. Data collection and reporting is not
resumed until the object is added back (recovered) to the database. You can recover the object
from the Deleted Objects view.
Note: You can perform group management tasks, such as copying and moving for groups that you

create in the OnCommand console. However, you cannot perform management tasks for the global
group.
Related references

Deleted Objects view on page 117


What groups and objects are
A group is a collection of objects that are discovered and monitored by the OnCommand console.
You can group objects based on characteristics such as the operating system version, location of the
storage systems, or projects to which the all the file systems belong.
By default, a group called Global exists in the DataFabric Manager server database. If you create
additional groups, you can select them from the Group menu.
Objects that are monitored by the OnCommand console, such as storage systems, aggregates, file
systems (volumes and qtrees), and logical unit numbers (LUNs), datasets, are referred to as objects.
Objects directly added to the groups are called direct members. You can also add child objects of the
direct members to the groups and categorize them as indirect members.
You can add the following list of objects that can be added to a group.
Following are the server objects:

Datacenters
Datastores
ESX Servers
Host Agents
Host services
Hyper-V Servers
Hyper-V VMs
Virtual Centers
VMware VMs

Following are the service automation objects:

Administration | 377

Datasets
Local policies
Resource pools
Storage services

Following are the storage objects:

Aggregates
Clusters
LUNs
Qtrees
SRM paths
Storage controllers
vFiler Units
Vservers
Volumes

Configuring groups
Creating groups
You can create groups to contain multiple objects so that you easily manage these objects. You can
create a group directly under the global group or create a subgroup to a parent group you already
created.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.

You must review the Guidelines for creating groups on page 378.

The following information must be available:

Name of the group


Members of the group

When you create a group, you can add an object to the group membership only if you have
permission to view that object.

Steps

1. Click the Administration menu, then click the Groups option.


2. From the Groups tab, select the default global group.
3. Click Create.

378 | OnCommand Console Help


4. Enter the name of the group in the name field.
5. Enter the owner's name, e-mail address, annual rate (per GB), and resource tag.
6. Select the appropriate member type from the drop-down menu.
7. Select the required member type and use appropriate arrow keys to move it to the list on the right.
The selected member types are added to the group and the membership list is updated.
8. Click Create.
Result

The new group appears in the Groups list. You can select any group in the Groups list from the
Groups menu.
Related references

Guidelines for creating groups on page 378


Administrator roles and capabilities on page 506
Guidelines for creating groups
You should follow a set of guidelines when you create groups.
Use the following guidelines when you create groups:

You can group similar or mix-types objects in a group.


An object can be a member of any number of groups.
You can group a subset of group members to create a new group.
You can create any number of groups.
You can copy to a group or move to a group in a group hierarchy.
However, parent group cannot be moved to its child group.

Deleting groups
You can delete groups that you no longer find useful.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

Deleting a group removes only the group container from the OnCommand database. The objects
contained in the deleted group are not removed from the database. When you delete a group, you also
delete all its subgroups, if any. If you want to preserve the subgroups, you must move them to a
different parent group before deleting the current parent group.

Administration | 379
Steps

1. Click the Administration menu, then click the Groups option.


2. From the Groups tab, select the group you want to delete from the list.
3. Click Delete.
4. Click Yes.
Related references

Administrator roles and capabilities on page 506

Managing groups
Editing groups
You can edit a group name, add or delete members of a group, and modify the contact information of
a group from the Edit Group dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click the Groups option.


2. From the Groups tab, select the group you want to modify.
3. Click Edit.
4. In the Edit Group dialog box, specify the group settings, as required.
5. Click Ok.
Related references

Administrator roles and capabilities on page 506

380 | OnCommand Console Help

Copying a group
You can copy a group and assign it to a different parent group from the Copy To dialog box. When
you copy a group, you create a copy of the selected group, and assign the copy to a different parent
group.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
To copy a group to a new parent group, you must be logged in as an administrator with Database
Write capability on the new parent group.
Steps

1. Click the Administration menu, then click the Groups option.


2. From the Groups tab, select the group you want to copy from the list.
3. Click Copy.
4. Click Ok.
Result

The selected group is copied to the target group.


Related references

Administrator roles and capabilities on page 506


Moving a group
When you move a group, you assign the selected group to a new parent group. If the group you
moved has any subgroups, those subgroups are also moved to the new parent, maintaining the same
hierarchical structure and membership.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
To move a group to a new parent group, you must have following capabilities:

Database Delete capability on the group that you want to move.


Database Write capability on the new parent group.

The group you want to move must not already exist in the target group. If the group already exists, an
appropriate message is displayed, and you cannot perform the move operation.

Administration | 381
Steps

1. Click the Administration menu, then click the Groups option.


2. From the Groups tab, select the group you want to move from the list.
3. Click Move.
4. Click Ok.
Related references

Administrator roles and capabilities on page 506

Page descriptions
Groups tab
The Groups tab enables you to view existing groups and perform tasks such as creating, deleting,
modifying, moving, and copying groups.

Command buttons on page 381


Groups list on page 382
Members tab on page 382
Graph tab on page 382

Command buttons
The command buttons enable you to perform the following management tasks for a selected group:
Create

Launches the Create Group dialog box, which enables you to create groups with different
member types.

Edit

Launches the Edit Group dialog box. You can edit the group name, add or delete group
members, or modify group contact information.

Delete

Deletes the selected group.


Note: When you delete a group, you also delete all of its subgroups, if any. If you want
to preserve the subgroups, you must move them to a different parent group before
deleting the current parent group.

Copy

Launches the Copy To dialog box. When you copy a group, you create a copy of the
selected group and assign it to a different parent group.

Move

Launches the Move To dialog box. When you move a group, you assign the selected
group to a new parent group. If the group you moved has any subgroups, those subgroups
are also moved to the new parent, maintaining the same hierarchical structure and
membership.

Refresh Refreshes the list of groups.

382 | OnCommand Console Help


Groups list
The groups list displays a list of all of the groups that are created. You can select the name of a group
to see the details for that group.
Group Name

Specifies the name of the group.

Owner

Displays the name of the owner of the group.

Email

Displays the e-mail address of the owner of the group.

Resource Tag

Specifies the resource tag of the group. This is a system-generated custom


comment field.

Annual Rate
(currency unit/GB)

Specifies the amount to charge for storage space usage per GB per year.

Total Capacity (GB)

Displays the total space allocated for the group.

Used Capacity (GB)

Displays the amount of storage used by the group.

Used (%)

Displays the percentage of storage used by the group.

Status

Displays the current status of each group. The status can be Normal,
Information, Warning, Error, Critical, Emergency, or Unknown.

ID

Displays the group ID. By default, this column is hidden.

Members tab
The Members tab displays detailed information about the selected group.
The Members tab displays the current status of each group as mentioned in the groups list, and
includes the following additional information:
Member Name Specifies the name of the group member.
Member Type Specifies the object type of the group member.
Status

Displays the current status of the group member. The status can be Normal,
Information, Warning, Error, Critical, Emergency, or Unknown.

Member Of

Specifies the name of the parent group to which the group member belongs.

Graph tab
The Graph tab displays information about the performance of the selected group. You can select the
graph you want to view from the drop-down menu in the area.
You can display information for a specified time period, such as one day, one week, one year, one
month, and three months. By clicking the export icon, you can export the graphical data in CSV
format.

Administration | 383
Related references

Window layout customization on page 16


Create Group dialog box
The Create Group dialog box enables you to create groups so that you can easily manage multiple
storage objects.

Properties on page 383


Command buttons on page 383

Properties
You can create groups by specifying properties such as group name, owner name, e-mail address of
the owner, and annual rate.
Name

Specifies the name of the group.

Owner

Specifies the user who owns the group.

Email

Specifies the e-mail address of the user who owns the group.

Resource Tag

Specifies the resource tag of the group. This is a system-generated custom


comment field.

Annual Rate (Per Specifies the amount to charge for storage space usage per GB per year. You
must enter a value in the x.y notation, where x is the integer part of the number
GB)
and y is the fractional part. For example, to specify an annual charge rate of
$150.55, you must enter 150.55.
Member Type

Displays the object types of the group in the drop-down menu.

Available
Members

Displays the list of members based on the object type selected. You can use the
filter to search for the objects. You can use appropriate arrow keys to move the
objects to the list on the right.

Selected
Members

Displays the selected object types.

Command buttons
You can use command buttons to perform the following management tasks:
Create

Creates a group based on the properties that you specify.

Cancel

Does not save the group configuration and closes the Create Group dialog box.

384 | OnCommand Console Help

Edit Group dialog box


The Edit Group dialog box enables you to edit the properties of the groups you created. You can edit
group properties such as name, owner, owner e-mail address, members, and chargeback details.

General tab on page 384


Group Member tab on page 384
Chargeback tab on page 384
Command buttons on page 385

General tab
You can edit properties groups such as group name, owner name, and e-mail address of the owner.
Name

Specifies the name of the group.

Owner

Specifies the user who owns the group.

Email

Specifies the e-mail address of the user who owns the group.

Resource Tag Specifies the resource tag of the group. This is a system-generated custom comment
field.
Group Member tab
You can edit properties of group members such as member type, available members, and selected
members.
Member Type

Displays the object types of the group in the drop-down menu.

Available
Members

Displays the list of members based on the object type selected. You can use
appropriate arrow keys to move the member types to the list on the right.

Selected Members

Displays the selected member types.

Chargeback tab
You can edit chargeback properties of groups such as annual rate and format of the annual rate.
Annual Rate (Per
GB)

Specifies the amount to charge for storage space usage per GB per year. You
must enter a value in the x.y notation, where x is the integer part of the
number and y is the fractional part. For example, to specify an annual charge
rate of $150.55, you must enter 150.55.

Formatted Annual Displays the format of the annual rate.


Rate

Administration | 385
Command buttons
You can use command buttons to perform the following management tasks:
OK

Updates the group properties that you specify.

Cancel Does not save the modification of the group configuration, and closes the Edit Group
dialog box.

Alarms
Understanding alarms
Alarm configuration
DataFabric Manager server uses alarms to notify you when events occur. DataFabric Manager server
sends the alarm notification to one or more specified recipients in different formats, such as e-mail
notification, pager alert, an SNMP traphost, or a script you wrote (you should attach the script to the
alarm).
You should determine the events that cause alarms, whether the alarm repeats until it is
acknowledged, and how many recipients an alarm has. Not all events are severe enough to require
alarms, and not all alarms are important enough to require acknowledgment. Nevertheless, to avoid
multiple responses to the same event, you should configure DataFabric Manager server to repeat
notification until an event is acknowledged.
Note: DataFabric Manager server does not automatically send alarms for the events.

Configuring alarms
Creating alarms for events
The OnCommand console enables you to configure alarms for immediate notification of events. You
can also configure alarms even before a particular event occurs. You can add an alarm based on the
event, event severity type, or event class from the Create Alarm dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
You must have your mail server configured so that the DataFabric Manager server can send e-mails
to specified recipients when an event occurs.
You must have the following information available to add an alarm:

The group with which you want the alarm associated.

386 | OnCommand Console Help

The event name, event class, or event severity type that triggers the alarm.
The recipients and the modes of event notifications.
The period during which the alarm is active.

You must have the following capabilities to perform this task:

DFM.Event.Write
DFM.Alarm.Write

About this task

Alarms you configure based on the event severity type are triggered when that event severity level
occurs.
Steps

1. Click the Administration menu, then click the Alarms option.


2. From the Alarms tab, click Create.
3. In the Create Alarm dialog box, specify the condition for which you want the alarm to be
triggered.
Note: An alarm is configured based on event type, event severity, or event class.

4. Specify either one or more means of alarm notification properties.


5. Click Create, then click Close.
Related concepts

Guidelines for creating alarms on page 30


Related tasks

Configuring the mail server for alarm notifications on page 36


Related references

Administrator roles and capabilities on page 506

Administration | 387

Creating alarms for a specific event


The OnCommand console enables you to configure an alarm when you want immediate notification
for a specified event name or event class, or if events (of a specified severity level) occur. You can
add an alarm from the Create Alarm dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
You must have your mail server configured so that the DataFabric Manager server can send e-mails
to specified recipients when an event occurs.
You must have the following information available to add an alarm:

The group with which you want the alarm associated


The event name, event class, or severity type that triggers the alarm
The recipients and the modes of event notifications
The period during which the alarm is active

You must have the following capabilities to perform this task:

DFM.Event.Write
DFM.Alarm.Write

About this task

Alarms you configure for a specific event are triggered when that event occurs.
Steps

1. Click the View menu, then click the Events option.


2. From the events list in the Events tab, analyze the events, and then determine the event for which
you want to create an alarm notification.
3. Select the event for which you want to create an alarm.
4. Click Create Alarm.
The Create Alarm dialog box opens, and the event is selected by default.
5. Specify either one or more means of alarm notification properties.
6. Click Create, then click Close.
Related concepts

Guidelines for creating alarms on page 30

388 | OnCommand Console Help


Related tasks

Configuring the mail server for alarm notifications on page 36


Related references

Administrator roles and capabilities on page 506

Managing events and alarms


Resolving events
After you have taken corrective action of a particular event, you should mark the event as resolved to
avoid multiple event notifications. You can mark the events as resolved from the Events tab.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the View menu, then click the Events option.


2. From the events list in the Events tab, select the event that you want to acknowledge.
3. Click Acknowledge.
When you do not acknowledge and mark an event resolved, you will receive multiple event
notifications for the same event.
4. Find the cause of the event and take corrective action.
5. Click Resolve to mark the event as resolved.
Related references

Administrator roles and capabilities on page 506


Editing alarm properties
You can edit the configuration of an existing alarm from the Edit Alarm dialog box. For example, if
you have created a script that is executed when there is an event notification, you can provide the
complete script path in Edit Alarm dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
You must have the following capabilities to perform this task:

Administration | 389

DFM.Event.Write
DFM.Alarm.Write

Steps

1. Click the Administration menu, then click the Alarms option.


2. From the alarms list, select the alarm whose properties you want to modify.
3. From the Alarms tab, click Edit.
4. In the Edit Alarm dialog box, edit the properties of the alarm as required.
5. Click Edit.
Result

The new configuration is immediately activated and displayed in the alarms list.
Related references

Administrator roles and capabilities on page 506


Configuring the mail server for alarm notifications
You must configure the mail server so that when an event occurs the DataFabric Manager server can
send e-mails to specified recipients.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

During this task, the OnCommand console launches the Operations Manager console. Depending on
your browser configuration, you can return to the OnCommand console by using the Alt-Tab key
combination or clicking the OnCommand console browser tab. After the completion of this task, you
can leave the Operations Manager console open, or you can close it to conserve bandwidth.
Steps

1. From the File menu, click Operations Manager.


2. In the Operations Manager console, click the Control Center tab.
3. Click Setup menu, and then click Options.
4. In Edit options, click the Events and Alerts option.
5. In the Events and Alerts Options page, specify the name of your mail server.

390 | OnCommand Console Help


Related references

Administrator roles and capabilities on page 506

Page descriptions
Alarms tab
The Alarms tab provides a single location from which you can view a list of alarms configured based
on event, event severity type, and event class. You can also perform various actions from this
window, such as edit, delete, test, and enable or disable alarms.

Command buttons on page 390


Alarms list on page 390
Details area on page 391

Command buttons
The command buttons enable you to perform the following management tasks for a selected event:
Create

Launches the Create Alarm dialog box in which you can create an alarm based on event,
event severity type, and event class.

Edit

Launches the Edit Alarm dialog box in which you can modify alarm properties.

Delete

Deletes the selected alarm.

Test

Tests the selected alarm to check its configuration, after creating or editing the alarm.

Enable Enables an alarm to send notifications.


Disable Disables the selected alarm when you want to temporarily stop its functioning.
Refresh Refreshes the list of alarms.
Alarms list
The Alarms list displays a list of all the configured alarms. You can select an alarm to see the details
for that alarm.
Alarm ID

Displays the ID of the alarm.

Event

Displays the event name for which the alarm is created.

Event Severity

Displays the severity type of the event.

Group

Displays the group name with which the alarm is associated.

Enabled

Displays Yes if the selected alarm is enabled or No if the selected alarm is


disabled.

Administration | 391
Start

Displays the time at which the selected alarm becomes active. By default, this
column is hidden.

End

Displays the time at which the selected alarm becomes inactive. By default, this
column is hidden.

Repeat Interval Displays the time period (in minutes) before the DataFabric Manager server
continues to send a repeated notification until the event is acknowledged or
(Minutes)
resolved. By default, this column is hidden.
Repeat Notify

Displays Yes if the selected alarm is enabled for repeated notification, or


displays No if the selected alarm is disabled for repeated notification. By
default, this column is hidden.

Event Class

Displays the class of event that is configured to trigger an alarm. By default, this
column is hidden.
You can configure a single alarm for multiple events using the event class. The
event class is a regular expression that contains rules, or pattern descriptions, that
typically use the word "matches" in the expression. For example, the
userquota.*|qtree.* expression matches all user quota or qtree events.

Details area
Apart from the alarm details displayed in the alarms list, you can view other additional properties of
the alarms in the area below the alarms list.
Effective Time Range

The time during which an alarm is active.

Administrators (Email
Address)

The e-mail address of the administrator, to which the alarm


notification is sent.

Administrators (Pager
Number)

The pager number of the administrator, to which the alarm


notification is sent.

Email Addresses (Others)

The e-mail addresses of nonadministrator users, to which the alarm


notification is sent.

Pager Numbers (Others)

The pager numbers of nonadministrator users, to which the alarm


notification is sent.

SNMP Trap Host

The SNMP traphost system that receives the alarm notification in the
form of SNMP traps.

Script Path

The name, along with the path of the script that is run when an alarm
is triggered.

Related references

Window layout customization on page 16

392 | OnCommand Console Help

Create Alarm dialog box


The Create Alarm dialog box enables you to create alarms based on the event type, event severity, or
event class. You can create alarms for specific event or many events.

Event Options on page 392


Notification Options on page 392
Command buttons on page 393

Event Options
You can create an alarm based on event name, event severity type, or event class:
Group

Displays the group that receives an alert when an event or event type triggers an
alarm.

Event

Displays the names of the events that triggers an alarm.

Event
Severity

Displays the severity types of the event that triggers an alarm. The event severity
types are Normal, Information, Warning, Error, Critical, and Emergency.

Event Class

Specifies the events classes that triggers an alarm.


The event class is a regular expression that contains rules, or pattern descriptions,
that typically use the word "matches" in the expression. For example, the
expression userquota.* matches all user quota events.

Notification Options
You can specify alarm notification properties by selecting one of the following check boxes:
SNMP Trap Host

Specifies the SNMP traphost that receives the notification.

E-mail Administrator
(Admin Name)

Specifies the name of the administrator who receives the e-mail


notification. You can specify multiple administrator names, separated
by commas.

Page Administrator
(Admin Name)

Specifies the administrator who receives the pager notification. You


can specify multiple administrator names, separated by commas.

E-mail Addresses
(Others)

Specifies the e-mail addresses of nonadministrator users who receive


the notification. You can specify multiple e-mail addresses, separated
by commas.

Pager Numbers (Others)

Specifies the pager numbers of other nonadministrator users who


receive the notification. You can specify multiple pager numbers,
separated by commas.

Script Path

Specifies the name of the script that is run when the alarm is triggered.

Administration | 393
Repeat Interval
(Minutes)

Specifies whether an alarm notification is repeated until the event is


acknowledged, and, if so, how often the notification is repeated.

Effective Time Range

Specifies the time during which the alarm is active.

Command buttons
You can use command buttons to perform the following management tasks for a selected event:
Create

Creates an alarm based on the properties that you specify.

Cancel

Does not save the alarm configuration and closes the Create Alarm dialog box.

Host services
Understanding host services
What a host service is
The host service is software that runs on a physical machine, a Hyper-V parent, or in a virtual
machine. The host service software includes plug-ins that enable the DataFabric Manager server to
discover, back up, and restore virtual objects, such as virtual machines and datastores. The host
service also enables you to view virtual objects in the OnCommand console.
Guidelines for managing host services
Resource discovery by a host service can be initiated manually by an administrator and by default,
automatic notification is available in response to changes in resources. When you make changes to
the virtual infrastructure, the results are available immediately because of the automatic notification
from the host service to the DataFabric Manager server. You can manually start a rediscovery job to
see your changes. You might need to refresh the host service information to see the updates in the
OnCommand console.
During the process of installing the NetApp OnCommand management software, you must register at
least one host service with the DataFabric Manager server and with the virtual infrastructure
(VMware or Hyper-V). You can register additional host services after installation, from the Host
Services tab accessible from the Administration menu in the OnCommand console. After
registration, you can monitor and manage host services from the Host Services tab.
Note: If the Hyper-V parent is part of a cluster, you must install the OnCommand Host Package on
each node of the cluster and all the cluster nodes must have the same TCP/IP port number to
enable communication between host services on different nodes. You must register and authorize
each node with the same DataFabric Manager server.
Note: When you register a host service with the DataFabric Manager server, you can type the fully
qualified domain name or IP address in the IPv4 format.

394 | OnCommand Console Help


Note: If you change the certificate on a host service or uninstall a host service then reinstall it, you
must unregister the host service from DataFabric Manager server using the -f flag. See the
unregister man page for more information.

The host service is included as part of the installation of the OnCommand Host Package. You can
install multiple host services on multiple vCenter Servers, virtual machines, or Hyper-V parents.
Note: In a Hyper-V cluster only, if you manually shut down a host service on the node that is

designated as the owner of a cluster and the node is active, the host service on both the cluster and
the node become inactive.
Note: The OnCommand Host Package upgrade does not force host services to reregister with

DataFabric Manager server. Therefore, if you unregister a host service from DataFabric Manager
server prior to an OnCommand Host Package upgrade, you must manually register the host service
to DataFabric Manager server after the upgrade is finished.
Messages from host services are stored persistently in the DataFabric Manager server database
hsNotifications table so that even if DataFabric Manager server goes down, information is not lost
and incomplete operations are automatically restarted or resumed after the server comes back up
again. This table continues to grow over time, and can quickly become huge in a large environment.
You can use the following global options to manage the size of this table:

hsNotificationsMaxCount
hsNotificationsPurgingInterval

Configuring host services


Adding and registering a host service
Before you can use a VMware or Hyper-V host, you must add the host service and register it with the
DataFabric Manager server.
Before you begin

The host service firewall must be disabled for the administration and management ports.
You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

DataFabric Manager server does not support a host service created as a generic service from failover
cluster manager in Microsoft Windows.
If you unregister a cluster level host service, DataFabric Manager server will not automatically
register the host service when you re-register the node. You must re-register or add the host service
using the cluster IP address.

Administration | 395
Attention: If you change the name of the machine after installing the OnCommand Host Package,

you must uninstall the OnCommand Host Package and perform a fresh installation.
Steps

1. Click the Administration menu, then click the Host Services option.
2. In the Host Services tab, click Add.
3. In the Add Host Service dialog box, type the IP address or the DNS name of the host on which
the host service is installed.
4. The default number of the administrative port is automatically entered.
This is the port that is used by plug-ins to discover information about the host service. If the port
number has been changed in the host service, type in the changed port number.
5. Click Add.
Result

The host service is added and registered with the DataFabric Manager server.
Tip: If you see an error stating that the requested operation did not complete in 60 seconds, wait

several minutes and then click Refresh to see if the host service was actually added.
Attention: Host services can be registered with only one DataFabric Manager server at a time.

Before you register a host service with a new DataFabric Manager server, you must first manually
unregister the host service from the old DataFabric Manager server. To unregister a host service
you must use the DataFabric Manager server hsid command.
After you finish

To make the host service fully operational, you might need to authorize the host service. In a
VMware environment, you must edit the host service to add the vCenter Server credentials.
Related references

Administrator roles and capabilities on page 506


Configuring a new host service
After you add a host service to OnCommand console, you must configure the host service to
communicate with DataFabric Manager server, virtual centers, and storage systems that are required

396 | OnCommand Console Help


for backup and recovery of your data. You also must reconfigure a host service when the credentials
on a storage server have been changed.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1.
2.
3.
4.
5.

Verifying that a host service is registered with the DataFabric Manager server on page 396
Authorizing a host service to access storage system credentials on page 397
Associating a host service with the vCenter Server on page 397
Associating storage systems with a host service on page 399
Editing storage system login and NDMP credentials from the Host Services tab on page 400

Related references

Administrator roles and capabilities on page 506


Verifying that a host service is registered with the DataFabric Manager server
You must be properly registered a host service with a DataFabric Manager server before the server
can discover objects and before you can perform a backup in a virtual environment.
About this task

A host service can be registered with the DataFabric Manager server during installation, or later from
the OnCommand console. However, you might want to verify that the registration is still valid when
troubleshooting problems or prior to performing an action involving a host service, such as adding a
storage system to a host service.
Steps

1. From the Administration menu, select Host Services.


The Host Services tab opens.
2. In the Host Services list, verify that the name of the host service is listed.
When a host service is registered with the DataFabric Manager server, it displays in the host
services list.
After you finish

If the host service is not displayed in the list, you must add and configure the new host service.

Administration | 397

Authorizing a host service to access storage system credentials


If the host service is not authorized, you must authorize the host service to access the storage system
credentials before you can create backup jobs.
Before you begin

The host service must be registered with the DataFabric Manager server prior to performing this task.
About this task

DataFabric Manager server does not support a host service created as a generic service from failover
cluster manager in Microsoft Windows.
Steps

1. From the Administration menu, select the Host Services option.


2. In the Host Services list, select the host service that you want and click Edit.
If the host service is not displayed in the list, you must add the host service and verify that it is
registered with the DataFabric Manager server.
3. In the Edit Host Service dialog box, click Authorize, review the certificate, and then click OK.
If the Authorize area is unavailable, the host service is already authorized.
When authorization is complete, the Authorize area becomes disabled.
4. Click OK.
After you finish

If you do not have storage systems associated with the host service, you must associate at least one
storage system to be able to perform backups.
After you finish editing the host service properties, you can view job progress from the Jobs subtab
on the Manage Host Services window and you can view details about each job from the Jobs tab.
Associating a host service with the vCenter Server
In a VMware environment, you must authorize each host service and associate it with a vCenter
Server. This provides part of the communication needed for discovery, monitoring, backup, and
recovery of virtual server objects such as virtual machines and datastores.
Before you begin

The host service must be registered with the DataFabric Manager server prior to performing this task.
Have the following information available:

398 | OnCommand Console Help

Name or IP address of the vCenter Server


You must specify the hostname or the fully qualified domain name for host service registration.
Do not use localhost.
User name and password for access to the vCenter Server

About this task

Authorization is required to create backup jobs as it allows the host service to access the storage
system credentials.
DataFabric Manager server does not support a host service created as a generic service from failover
cluster manager in Microsoft Windows.
Steps

1. From the Administration menu, select the Host Services option.


2. In the Host Services list, select the host service that you want and then click Edit.
If the host service is not displayed in the list, you must add the host service and verify that it is
registered with the DataFabric Manager server.
The Edit Host Services dialog box opens.
3. Click Authorize, review the certificate, and then click OK.
If the Authorize area is disabled, the host service is already authorized.
4. Enter the vCenter Server properties.
If the properties fields are populated, then a server is already associated with the host service that
you selected.
If the vCenter Properties section is not displayed, you might have selected a host service that is
installed in an environment other than VMware.
5. Click OK.
After you finish

If you do not have storage systems associated with the host service, you must associate at least one
storage system to be able to perform backups.
After you finish editing the host service properties, you can view job progress from the Jobs subtab
on the Manage Host Services window and you can view details about each job from the Jobs tab.

Administration | 399

Associating storage systems with a host service


For each host service instance, you must associate one or more storage systems that host virtual
machines for the host service. This enables communication between the service and storage to ensure
that storage objects, such as virtual disks, are discovered and that host service features work properly.
Before you begin

If you add a new storage system to associate with the host service, you must have the following
storage system information available:

IP address or name
Login and NDMP credentials
Access protocol (HTTP or HTTPS)

Steps

1. From the Administration menu, select Host Services.


2. From the Host Services list, select the host service with which you want to associate storage.
In the Storage Systems subtab is a list of storage systems currently associated with the host
service you selected.
3. Click Edit to associate a new storage system.
The Edit Host Service dialog box opens.
4. In the Storage Systems area, click Associate.
5. Associate the storage systems with the host service, as follows:

To associate storage systems shown in the Available Storage Systems list, select the system
names and click OK.
To associate a storage system not listed in Available Storage Systems, click Add, enter the
required information, and click OK.

The newly associated storage system displays in the Storage Systems area.
6. In the list of storage systems, verify that the status is Good for the login and NDMP credentials
for each storage system.
After you finish

If the login or NDMP status is other than Good for any storage system, you must edit the storage
system properties to provide the correct credentials before you can use that storage system.
After you finish editing the host service properties, you can view job progress from the Jobs subtab
on the Manage Host Services window and you can view details about each job from the Jobs tab.

400 | OnCommand Console Help

Editing storage system login and NDMP credentials from the Host Services tab
You must have valid login and NDMP credentials for storage systems so they can be accessed by the
DataFabric Manager server. If the server cannot access the storage, your backups might fail.
Before you begin

Have the following storage system information available:

IP address or name
Login and NDMP credentials
Access protocol (HTTP or HTTPS)

Steps

1. From the Administration menu, select Host Services.


2. In the Host Services list, select a host service.
The storage systems associated with the selected host service display in the Host Services tab.
3. In the Host Services tab, select a storage system with login or NDMP status of Bad or Unknown.
4. Click Edit.
5. In the Edit Host Service dialog box, click Edit.
6. Enter the appropriate login and NDMP credentials and click OK.
7. In the Host Services tab, verify that the Login Status and NDMP Status are Good.
8. Click OK.
The storage system status columns in the Host Services tab update with the new status.
After you finish

After you finish editing the storage system properties, you can view job progress from the Jobs
subtab on the Manage Host Services window and you can view details about each job from the Jobs
tab.

Managing host services


Editing a host service
After you add a host service to OnCommand console, you can edit the properties of the host service,
the storage systems associated with the host service, and the system login and NDMP credentials. If

Administration | 401
you did not authorize the host service when you added it, you can also authorize it from the Edit Host
Service dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

If you are adding a new host service, see Configuring a new host service.
Attention: If you change the name of the machine after installing the OnCommand Host Package,

you must uninstall the OnCommand Host Package and perform a fresh installation.
Steps

1.
2.
3.
4.

Verifying that a host service is registered with the DataFabric Manager server on page 401
Associating a host service with the vCenter Server on page 402
Associating storage systems with a host service on page 403
Editing storage system login and NDMP credentials from the Host Services tab on page 404

Related references

Configuring a new host service on page 395


Administrator roles and capabilities on page 506
Verifying that a host service is registered with the DataFabric Manager server
You must be properly registered a host service with a DataFabric Manager server before the server
can discover objects and before you can perform a backup in a virtual environment.
About this task

A host service can be registered with the DataFabric Manager server during installation, or later from
the OnCommand console. However, you might want to verify that the registration is still valid when
troubleshooting problems or prior to performing an action involving a host service, such as adding a
storage system to a host service.
Steps

1. From the Administration menu, select Host Services.


The Host Services tab opens.
2. In the Host Services list, verify that the name of the host service is listed.
When a host service is registered with the DataFabric Manager server, it displays in the host
services list.

402 | OnCommand Console Help


After you finish

If the host service is not displayed in the list, you must add and configure the new host service.
Associating a host service with the vCenter Server
In a VMware environment, you must authorize each host service and associate it with a vCenter
Server. This provides part of the communication needed for discovery, monitoring, backup, and
recovery of virtual server objects such as virtual machines and datastores.
Before you begin

The host service must be registered with the DataFabric Manager server prior to performing this task.
Have the following information available:

Name or IP address of the vCenter Server


You must specify the hostname or the fully qualified domain name for host service registration.
Do not use localhost.
User name and password for access to the vCenter Server

About this task

Authorization is required to create backup jobs as it allows the host service to access the storage
system credentials.
DataFabric Manager server does not support a host service created as a generic service from failover
cluster manager in Microsoft Windows.
Steps

1. From the Administration menu, select the Host Services option.


2. In the Host Services list, select the host service that you want and then click Edit.
If the host service is not displayed in the list, you must add the host service and verify that it is
registered with the DataFabric Manager server.
The Edit Host Services dialog box opens.
3. Click Authorize, review the certificate, and then click OK.
If the Authorize area is disabled, the host service is already authorized.
4. Enter the vCenter Server properties.
If the properties fields are populated, then a server is already associated with the host service that
you selected.
If the vCenter Properties section is not displayed, you might have selected a host service that is
installed in an environment other than VMware.
5. Click OK.

Administration | 403
After you finish

If you do not have storage systems associated with the host service, you must associate at least one
storage system to be able to perform backups.
After you finish editing the host service properties, you can view job progress from the Jobs subtab
on the Manage Host Services window and you can view details about each job from the Jobs tab.
Associating storage systems with a host service
For each host service instance, you must associate one or more storage systems that host virtual
machines for the host service. This enables communication between the service and storage to ensure
that storage objects, such as virtual disks, are discovered and that host service features work properly.
Before you begin

If you add a new storage system to associate with the host service, you must have the following
storage system information available:

IP address or name
Login and NDMP credentials
Access protocol (HTTP or HTTPS)

Steps

1. From the Administration menu, select Host Services.


2. From the Host Services list, select the host service with which you want to associate storage.
In the Storage Systems subtab is a list of storage systems currently associated with the host
service you selected.
3. Click Edit to associate a new storage system.
The Edit Host Service dialog box opens.
4. In the Storage Systems area, click Associate.
5. Associate the storage systems with the host service, as follows:

To associate storage systems shown in the Available Storage Systems list, select the system
names and click OK.
To associate a storage system not listed in Available Storage Systems, click Add, enter the
required information, and click OK.

The newly associated storage system displays in the Storage Systems area.
6. In the list of storage systems, verify that the status is Good for the login and NDMP credentials
for each storage system.

404 | OnCommand Console Help


After you finish

If the login or NDMP status is other than Good for any storage system, you must edit the storage
system properties to provide the correct credentials before you can use that storage system.
After you finish editing the host service properties, you can view job progress from the Jobs subtab
on the Manage Host Services window and you can view details about each job from the Jobs tab.
Editing storage system login and NDMP credentials from the Host Services tab
You must have valid login and NDMP credentials for storage systems so they can be accessed by the
DataFabric Manager server. If the server cannot access the storage, your backups might fail.
Before you begin

Have the following storage system information available:

IP address or name
Login and NDMP credentials
Access protocol (HTTP or HTTPS)

Steps

1. From the Administration menu, select Host Services.


2. In the Host Services list, select a host service.
The storage systems associated with the selected host service display in the Host Services tab.
3. In the Host Services tab, select a storage system with login or NDMP status of Bad or Unknown.
4. Click Edit.
5. In the Edit Host Service dialog box, click Edit.
6. Enter the appropriate login and NDMP credentials and click OK.
7. In the Host Services tab, verify that the Login Status and NDMP Status are Good.
8. Click OK.
The storage system status columns in the Host Services tab update with the new status.
After you finish

After you finish editing the storage system properties, you can view job progress from the Jobs
subtab on the Manage Host Services window and you can view details about each job from the Jobs
tab.

Administration | 405

Deleting a host service


You can delete a VMware or Hyper-V host service from DataFabric Manager server.
Before you begin

The host service must not be in use.


You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click the Host Services option.
2. In the Host Services tab, click Delete.
3. In the Delete Host Service Confirmation dialog box, click Yes to delete the host service or click
No to terminate the deletion request.
Result

The host service is deleted from DataFabric Manager server and the associated virtual objects are
removed from the inventory lists.
If you restart the host service on the server side, the host service attempts to register with the
DataFabric Manager server. You can prevent this by manually changing the dfm_server attribute in
HSServiceHost.exe.config to a different value before restarting the host service plug-in.
Related references

Administrator roles and capabilities on page 506


Re-registering a host service if the hostname changes
If the hostname of the host service machine changes, but the IP address remains the same, you must
unregister the host service and re-register it using the IP address.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Unregister the host service by completing the following steps:


a. Click the Administration menu, then click the Host Services option.
b. In the Host Services tab, click Delete.

406 | OnCommand Console Help


c. In the Delete Host Service Confirmation dialog box, click Yes.
2. Re-register the host service with DataFabric Manager server using the IP address in the Add Host
Service dialog box.
Related tasks

Adding and registering a host service on page 394


Related references

Administrator roles and capabilities on page 506


Bringing up a host service after restoring it to a different DataFabric Manager server
After you restore a host service to a different DataFabric Manager server, you must manually bring
up the host service before it can receive notifications.
Steps

1. Copy the DataFabric Manager server keys to the new DataFabric Manager server.
2. Enter the following command to reload the SSL service: dfm ssl service reload
3. Enter the following command for each host service: dfm hs configure -i <New
DataFabric Manager server IP> <host service name or ID>.
Moving a host service to a different DataFabric Manager server
If you uninstall DataFabric Manager server and start with a fresh database or make your host service
point to a new DataFabric Manager server with brand new databases without first unregistering the
host service, you might need to clean up the host service repository either by reinstalling the host
service or by cleaning up the old, leftover data from the host service repository.
Before you begin

If any of the resources from the host service you want to move are in a dataset, they must be removed
from the dataset prior to unregistering the host service.
You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

This procedure describes how to manually delete the old dataset information rather than reinstalling
the host service.
Steps

1. Unregister the host service by completing the following steps:

Administration | 407
a. Click the Administration menu, then click the Host Services option.
b. In the Host Services tab, click Delete.
c. In the Delete Host Service Confirmation dialog box, click Yes.
2. Stop the host service by using the Service Control Manager on the host service machine.
3. Clean up the data by performing the following steps:
a. Remove policyenforcementdata.xml and eventrepository.xml from the data stores
folder in the host service installation directory.
b. Delete any leftover messages in the messages queues.
Message queue folders end with "queue" and are located in the installation directory.
c. Clean up the scheduled jobs.
This step is done from the Microsoft Windows Task Scheduler on the host service machine.
4. Restart the host service by using the Service Control Manager on the host service machine.
5. Re-register the host service with DataFabric Manager server.
Related tasks

Adding and registering a host service on page 394


Related references

Administrator roles and capabilities on page 506

Monitoring host services


Viewing configured hosts
You can view information about the configured VMware or Hyper-V hosts and the status of jobs
associated with those hosts from the Host Services tab.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Step

1. Click the Administration menu, then click the Host Services option.
Related references

Administrator roles and capabilities on page 506

408 | OnCommand Console Help

Rediscovering virtual object inventory


If you have made changes to the virtual infrastructure managed by a host service and don't want to
wait for the host service to automatically notify DataFabric Manager server of the changes, you can
update the list of configured VMware or Hyper-V hosts and virtual objects by manually starting a
rediscovery job in the Host Services tab.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task
Note: If the host service FQDN changes, DataFabric Manager server will be unable to discover
any new virtual objects for the host service until you delete it and then add it again.
Steps

1. Click the Administration menu, then click the Host Services option.
2. In the Host Services tab, select a host service, then click Rediscover.
Result

A discovery job is started for the selected host service. When the discovery job finishes, DataFabric
Manager server reflects the current list of configured VMware or Hyper-V hosts and virtual objects
managed by the host service. You might need to refresh the host service information to see the
updated list.
Related tasks

Refreshing host service information on page 408


Related references

Administrator roles and capabilities on page 506


Refreshing host service information
Since a host service can be added or deleted while you are viewing the Host Services tab, you may
want to refresh the list of host services.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.

Administration | 409
Steps

1. Click the Administration menu, then click the Host Services option.
2. In the Host Services tab, click Refresh.
Result

The current list of host services is retrieved from the DataFabric Manager server.
Related references

Administrator roles and capabilities on page 506

Page descriptions
Host Services tab
You can view information about registered virtual host services; add, configure, edit, and delete host
services; rediscover virtual objects; and refresh the virtual object inventory display from the Host
Services tab. You can access this window by clicking Administration > Host Services.

Command buttons on page 409


Host services list on page 410
Storage Systems tab on page 410
More Details tab on page 411
Jobs tab on page 411

Command buttons
Add

Opens the Add Host Service dialog box, which allows you to add a virtual host service
to OnCommand console.
You can add a host service by using the Add button and then configure it later by
using the Edit button.

Edit

Configures the credentials for a selected host service, which enables the host to be
used by the OnCommand console. Typically, you first add a host service and then
configure it. Thereafter, you use the Edit button to edit the configuration for a host, if
needed.

Delete

Deletes a host service from DataFabric Manager server.

Refresh

Updates the virtual object inventory information that is displayed.

Rediscover Starts a discovery job for the selected host service.

410 | OnCommand Console Help


Host services list
Displays information about the host services that are registered on DataFabric Manager server.
ID

The ID of the host on which the host service is installed.

Name

The name of the host on which the host service is installed. This might be a
fully qualified domain name if the host is on a domain.

IP Address

The IP address of the host on which the host service is installed.

Admin Port

The host service port that is used for administrative operations.

Management Port The host service port that is used for management operations.
Version

The version of the host service software.

Discovery Status

Indicates whether the discovery of the host service was successful. "Error"
indicates that the discovery was not completely successful. The Jobs tab at the
bottom of the list displays the reason that the discovery failed.

Status

Indicates whether the host service is running (up) or not running (down).

Storage Systems tab


This section displays information about the storage system associated with the selected host service.
Storage System
Name

Indicates the name of the storage system associated with the host service.

IP Address

Displays the IP address of the storage system.

System Status

Indicates the current status of the storage system.

Login Status (Host


Service)

Indicates whether or not the host service has valid credentials for the storage
system.

NDMP Status

Indicates whether or not DataFabric Manager server has the valid network
data management protocol credentials for the storage system.

Login Status
(Server)

Indicates whether or not DataFabric Manager server has the valid login
credentials for the storage system.

Transport Protocol

Specifies the transport protocol used by the host service to communicate


with the storage system.

You can click on the storage system name to display the respective entry in
the storage inventory.

Valid values are http, https, and rpc.

Administration | 411
More Details tab
This section displays detailed information about the components of the selected host service. The
components that are listed vary depending upon the virtual infrastructure type the host service is
managing.
Type

Specifies the type of host service plugin.

Version

Specifies the version of the host service plugin.

Jobs tab
This section displays information about the most recent jobs that ran on the host service. Host service
jobs are typically discovery and host service software upgrade jobs.
Job ID

The identification number of the job.

Job Type

The type of job, which is determined by the policy assigned to the dataset or by the
direct request initiated by a user.

Description A description of the job taken from the policy configuration or the job description
entered when the job was manually started.
Started By The ID of the user who started the job.
Start

The date and time the job started.

Status

The running status of the job.

End

The date and time the job ended.

Related references

Window layout customization on page 16

Storage systems users


Understanding storage system users
Who local users are
Local users are the users created on storage systems and vFiler units.

Configuring storage system users

412 | OnCommand Console Help

Adding a storage system user


To manage your storage systems efficiently, you can configure a local user to a host from the Add
Local User to Host page in the Operations Manager console. You can access this page from the
OnCommand console Administration menu.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

During this task, the OnCommand console launches the Operations Manager console. Depending on
your browser configuration, you can return to the OnCommand console by using the Alt-Tab key
combination or clicking the OnCommand console browser tab. After the completion of this task, you
can leave the Operations Manager console open, or you can close it to conserve bandwidth.
Steps

1. Click the Administration menu, then click the Storage Systems user option.
A separate browser window opens to the Host Users page in the Operations Manager console.
2. Click the Local Users tab.
3. Configure the local users.
For more information, see the Operations Manager Help.
4. When finished, press Alt-Tab or click the OnCommand console browser tab to return to the
OnCommand console.
Related references

Administrator roles and capabilities on page 506

Storage system configuration


Understanding storage system configuration
Configuration of storage systems
You can remotely configure multiple storage systems using Operations Manager.
By creating configuration resource groups and applying configuration settings to them,
administrators can remotely configure multiple storage systems from the server on which the

Administration | 413
DataFabric Manager server is installed. Administrators can also manage CIFS data through
configuration management.
List of configuration management tasks for storage systems
You can perform a variety of configuration management tasks by using the storage system
configuration management feature.
Following are some of the tasks you can perform:

Pull a configuration file from a storage system


View the contents of each configuration file.
Edit the configuration file settings (registry options and /etc files).
Copy or rename configuration files.
Edit a configuration file to create a partial configuration file.
Compare configuration files against a standard template.
View the list of existing configuration files.
Upgrade or revert file versions.
Delete a configuration file.
Import and export configuration files.
Remove an existing configuration file from a groups configuration list.
Change the order of files in the configuration list.
Specify configuration overrides for a storage system assigned to a group.
Exclude configuration settings from being pushed to a storage system.
View Groups configuration summary for a version of Data ONTAP.
Push configuration files to a storage system or a group of storage systems.
Delete push configuration jobs.
View the status of push configuration jobs.

Configuring storage systems


Configuring storage systems
You can configure storage systems from the Storage System Configurations page in the Operations
Manager console, which you can access from the OnCommand console Administration menu.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

During this task, the OnCommand console launches the Operations Manager console. Depending on
your browser configuration, you can return to the OnCommand console by using the Alt-Tab key

414 | OnCommand Console Help


combination or clicking the OnCommand console browser tab. After the completion of this task, you
can leave the Operations Manager console open, or you can close it to conserve bandwidth.
Steps

1. Click the Administration menu, then click the Storage Systems Configuration option.
A separate browser window opens to the Storage System Configurations page in the Operations
Manager console.
2. Configure the storage system.
For more information, see the Operations Manager Help.
3. When finished, press Alt-Tab or click the OnCommand console browser tab to return to the
OnCommand console.
Related references

Administrator roles and capabilities on page 506

vFiler configuration
Understanding vFiler unit configuration
List of configuration management tasks for vFiler units
You can perform a variety of configuration management tasks by using the vFiler units configuration
management feature.
Following are some of the tasks you can perform:

Pull a configuration file from a vFiler unit


View the contents of each configuration file.
Edit the configuration file settings (registry options and /etc files).
Copy or rename configuration files.
Edit a configuration file to create a partial configuration file.
Compare configuration files against a standard template.
View the list of existing configuration files.
Upgrade or revert file versions.
Delete a configuration file.
Import and export configuration files.
Remove an existing configuration file from a groups configuration list.
Change the order of files in the configuration list.
Specify configuration overrides for a vFiler unit assigned to a group.

Administration | 415

Exclude configuration settings from being pushed to a vFiler unit.


View Groups configuration summary for a version of Data ONTAP.
Push configuration files to a group of vFiler units.
Delete push configuration jobs.
View the status of push configuration jobs.

Configuring vFiler units


Configuring vFiler units
You can configure vFiler units from the vFiler Configurations page in the Operations Manager
console, which you can access from the OnCommand console Administration menu.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

During this task, the OnCommand console launches the Operations Manager console. Depending on
your browser configuration, you can return to the OnCommand console by using the Alt-Tab key
combination or clicking the OnCommand console browser tab. After the completion of this task, you
can leave the Operations Manager console open, or you can close it to conserve bandwidth.
Steps

1. Click the Administration menu, then click the vFiler Configuration option.
A separate browser window opens to the vFiler Configurations page in the Operations Manager
console.
2. Configure the vFiler units.
For more information, see the Operations Manager Help.
3. When finished, press Alt-Tab or click the OnCommand console browser tab to return to the
OnCommand console.
Related references

Administrator roles and capabilities on page 506

Options
Page descriptions

416 | OnCommand Console Help

Setup Options dialog box


The Setup Options dialog box enables you to configure various options from the DataFabric Manager
server such as chargeback settings, database backup, default threshold values, SNMP settings,
monitoring intervals, and naming settings.

Backup on page 416


Global Naming Settings on page 416
Costing on page 417
Database Backup on page 417
Default Thresholds on page 417
Discovery on page 417
File SRM on page 418
LDAP on page 418
Monitoring on page 418
Management on page 418
Systems on page 419
Command buttons on page 419

Setup Options
You can configure the following options from the Setup Options dialog box:
Backup
The Backup option enables you to configure the lag threshold values for all volumes or specific
secondary volumes.

Default Thresholds
Specifies the Backup Manager monitoring intervals.

Global Naming Settings


The Global Naming Settings options specifies how to name a dataset's related objects, such as
Snapshot copies, volumes, and qtrees that result from protection jobs run on that dataset.
Global naming settings apply to all related objects except those that belong to datasets configured
with dataset-level naming settings.

Snapshot copy
Specifies the global-level names for Snapshot copies that are generated by dataset protection jobs.
Primary volume
Specifies the global-level names of primary volumes that are generated by protection jobs.
Secondary volume
Specifies the global-level names of secondary volumes that are generated by protection jobs.

Administration | 417

Secondary qtree
Specifies the global-level names of secondary qtrees that are generated by protection jobs.

Costing
The Costing option enables you to configure the chargeback settings to obtain billing reports for the
space used by a specific storage object or a group of objects.

Chargeback
Specifies the parameters you can configure to generate billing reports for the amount of space
used by a specific object or a group of objects.

Database Backup
The Database Backup option enables you to configure the backup destination directory and retention
count for the DataFabric Manager server database backup, and also manages existing backups.

Schedule
Specifies the parameters that you can configure to schedule a database backup.
Completed
Displays the ongoing database backups and the associated database backup events.

Default Thresholds
The Default Thresholds option enables you to configure the global default threshold values for
objects such as aggregates, volumes, qtrees, user quotas, resource pools, HBA ports and Hosts.

Aggregates
Specifies the global default threshold values for monitored aggregates.
Volumes
Specifies the global default threshold values for monitored volumes.
Other
Specifies the global default threshold values for host agents, HBA ports, qtrees, user quotas, and
resource pools.

Discovery
The Discovery option enables you to set host discovery options, discovery methods, timeout and
interval values. You can also configure networks, and settings for the discovery of storage objects
such as networks, storage systems (including clusters), host agents, and Open Systems SnapVault
agents.

Options
Specifies the discovery methods and the monitoring interval for discovery.
Addresses
Specifies the network addresses which are discovered for new hosts.

418 | OnCommand Console Help

Credentials
Specifies the credentials for network addresses that are used for network and host discovery.

File SRM
The File SRM option enables you to configure the File SRM settings, such as the number of largest
files, recently modified files, least accessed files, and least modified files.

Options
Specifies the file parameters that you can configure.

LDAP
The LDAP option enables you to configure the LDAP settings to successfully retrieve data from the
LDAP server.

Authentication
Specifies the authentication settings that help the DataFabric Manager server to communicate
with the LDAP servers.
Server Types
Specifies settings that are configured to establish compatibility with the LDAP server.
Servers
Specifies the LDAP server properties and the last authentication status.

Monitoring
The Monitoring option enables you to configure the monitoring intervals for various storage objects
monitored by the DataFabric Manager server.

Storage
Specifies the monitoring parameters for storage objects.
Protection
Specifies the monitoring parameters for the protection of storage objects.
Networking
Specifies the monitoring parameters for networking objects.
Inventory
Specifies the monitoring parameters for inventory objects.
System
Specifies the monitoring parameters for system objects.

Management
The Management option enables you to configure the connection protocols settings for management
purposes.

Client

Administration | 419

Specifies the HTTP and HTTPS settings that you can configure to establish a connection between
the client and the DataFabric Manager server.
Managed Host
Specifies settings that you can configure to establish a connection between the managed host and
the DataFabric Manager server.
Host Agent
Specifies settings that you can configure to establish a connection between the host agent and the
DataFabric Manager server.

Systems
The Systems option enables you to configure the system settings such as event notifications (e-mail
notification, pager alerts, and SNMP traps), create custom comment fields, and set audit log options.

Alarms
Specifies settings that you can configure to send event notifications in different formats.
Annotations
Enables you to create annotations for the DataFabric Manager server that can be assigned to any
resource objects.
Miscellaneous
Specifies miscellaneous settings that you can configure such as audit log options, credential TTL
cache, and options to preserve your local configuration settings.

Command buttons
The command buttons enable you to save or cancel the setup options.
Save and Close

Saves the configuration settings for the selected option.

Cancel

Does not save the recent changes and closes the Setup Options dialog box.

Save

Saves the configuration settings for the selected option.

Backup setup options


Understanding backup options
Backup relationships and lag thresholds
The DataFabric Manager server uses the SnapVault technology of Data ONTAP to manage backup
and restore operations. The DataFabric Manager server discovers and imports existing SnapVault
relationships by using NDMP.
To use SnapVault, you must have separate SnapVault licenses for both the primary and secondary
storage systems.
Lag thresholds are limits set on the time elapsed since the last successful backup. When those limits
are exceeded, the DataFabric Manager server generates events that indicate the severity of the event.

420 | OnCommand Console Help


After you add a secondary volume, the default values for lag thresholds are applied. However, you
can change these lag thresholds, either for all volumes or for specific secondary volumes, from the
Setup Options dialog box.
Managing backup options
Editing backup default threshold options
You can edit the lag threshold values for all volumes or specific secondary volumes from the Setup
Options dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click Setup Options.


2. In the Setup Options dialog box, click the Backup option.
3. In the Default Thresholds area, specify the new values, as required.
4. Click Save and Close.
Related references

Administrator roles and capabilities on page 506


Page descriptions
Backup Default Thresholds area
You can use the Backup Default Threshold options to configure the lag threshold values for all
volumes or specific secondary volumes. When a secondary volume is added to Backup Manager, the
default values for the lag thresholds are applied automatically, but you can change them for all
volumes or specific secondary volumes.

Options on page 420


Command buttons on page 421

Options
You can configure the lag threshold values using the following options:
SnapVault
Replica

Displays options to configure thresholds for secondary volumes.

Administration | 421

Out-of-Date
Threshold

Specifies the limit at which the backups on a secondary


volume are considered obsolete. If this limit is exceeded,
DataFabric Manager server generates a SnapVault Replica
Out of Date event.
The default is 2 days.

Nearly Out-ofDate Threshold

Specifies the limit at which the backups on a secondary


volume are considered nearly obsolete. If this limit is
exceeded, DataFabric Manager server generates the
SnapVault Replica Nearly Out of Date event.
The default is 1.5 days.

Purge Backup
Jobs

Displays options to list purge backup job files.


Older
Than

Specifies whether backup job files that are older than the
designated period of time are purged.
The default is 12.86 weeks.

Command buttons
The command buttons enable you to save or cancel the setup options.
Save and Close

Saves the configuration settings for the selected option.

Cancel

Does not save the recent changes and closes the Setup Options dialog box.

Save

Saves the configuration settings for the selected option.

Global naming settings setup options


Understanding naming settings options
Differences between global and dataset-level naming settings
Global naming settings apply, by default, to all objects that are generated in datasets that are not
configured with dataset-level naming settings. Dataset-level naming settings are configured for
individual datasets and apply only to the objects that are generated in those datasets.
Global naming settings
Protection-related objects are, by default, automatically named using a global naming format for each
object type.
The OnCommand console enables you to customize the global naming settings for one or more
related object type. After you customize the global naming format for a particular object type, that

422 | OnCommand Console Help


customization then applies by default to all newly generated objects of that type that do not belong to
a dataset with conflicting dataset-level naming settings.
However, if the newly generated object is part of a dataset that already has customized dataset-level
Snapshot copy, primary volume, secondary volume, or secondary qtree naming formats, then the
customized naming formats apply to that object.
Dataset-level naming settings
Both the OnCommand console and NetApp Management Console enable you to configure datasetlevel naming settings for one or more object types in a particular dataset. Those dataset-level naming
settings then apply to the naming of their object types within that dataset.
Within that dataset, dataset-level naming settings have priority over any conflicting global naming
settings.
When to customize naming of protection-related objects
To support backup tasks, storage administration tasks, and application administration tasks, you can
customize naming of the OnCommand console protection-related objects.
Custom naming and backup tasks
If you are a backup administrator, configuration of custom naming presents you with the following
advantages:

Easy location of secondary and tertiary storage backup items for restoration
Custom naming applied to Snapshot copies, secondary volumes, or secondary qtrees enables you
to easily identify the backup objects in which to locate files for restoration.
Easy identification of backup data by an assortment of criteria
Custom naming, configured with specific naming conventions and applied to the backup objects,
enables you to identify those objects by priority, business unit, administrator, backup time,
physical container, or logical container.
Customization of formats to maintain naming used by imported protection relationships
Custom naming can be used by the OnCommand console to assign naming formats for related
objects that matches the related object naming conventions originally used by imported protection
relationships.
Easy identification of backup data related to protection, SnapManager, or SnapDrive operations
A consistent naming convention enables you to easily find the right data to restore even if that
backup data is related to such disparate activities as the OnCommand console protection
operations, SnapManager activity, or SnapDrive activity.
Easy identification of source and secondary volumes in case of application shutdown
If the database application that is generating the backed up data shuts down, the naming formats
of the associated Snapshot copy, secondary volume, or secondary qtree objects, if set properly,
enable you to identify primary volumes from the names of the secondary volumes and Snapshot
copies that are associated with that application.

Administration | 423
Custom naming and storage management tasks
If you are a storage administrator, configuration of custom naming enables you to specify a
company-wide naming convention for related objects at the global level or at the dataset level.
Custom naming and application management tasks
If you are an application administrator and have specified a particular dataset in which to store data
generated by a particular application, custom naming enables you to specify distinctive naming
conventions for that dataset's related object types. The distinctive naming enables you to track the
objects that are generated by that application more easily.
Naming settings by format strings
The OnCommand console naming settings enable you to enter custom format strings that contain
letter attributes for identifying information to include in the names of protection-related objects.
The attributes that you can include in the format strings cause such identifying characteristics as
timestamps, storage system names, custom labels, dataset names, and retention type to be included in
the name of a protection-related object type.
Naming settings by naming script
Naming scripts are user-authored scripts for naming some protection-related object types (Snapshot
copies, primary volumes, or secondary volumes) that are generated by protection jobs being executed
on a dataset.
For each supported protection-related object type (secondary qtrees are the only related object type
for which naming scripts are not supported), you can write a script that makes use of DataFabric
Manager server supported environmental variables to generate a name for objects of that type that are
generated when a protection job is executed on a dataset. When you configure global naming settings
in the Setup Options dialog box, you can specify this naming script and path as an alternative to
accepting the default naming settings or to entering a custom naming format string in the Setup
Options dialog box itself.
You can specify naming scripts for global naming settings only. Naming scripts are not applied to a
dataset's protection-related objects if that dataset's dataset-level naming settings specify a custom
naming format string instead.
Naming script restrictions and precautions
As you author and apply naming scripts for your dataset's protection-related object types, keep in
mind the following points:

You can specify naming scripts for global naming settings only.
Naming scripts are not applied to a dataset's protection-related objects if that dataset's datasetlevel naming settings specify a custom naming format string instead.
The output of your naming scripts must be tested.

424 | OnCommand Console Help


The output of the naming script determines what name is generated and given to the dataset's
protection-related objects for whom the script is written. If the script generates an error message,
part of that error message might be included in the names generated for the objects in question.
Naming script limitations
Special limitations apply to protection-related object names that are specified by naming scripts.
These limitations are in addition to the naming limitations that apply to protection-related object
names that are specified by custom format.
Special limitations that apply to the use of naming scripts include the following:

The naming script must be in a location that is readable from the DataFabric Manager server.
You can assign a naming script only to the global naming settings for Snapshot copy, primary
volume, and secondary volume objects.
You cannot assign a naming script to secondary qtree objects.
A naming script does not apply to an object type in a dataset if that dataset has a dataset-level
custom naming format enabled for that object type.
If you specify an incomplete script path in the global naming settings, an error event and job
failure results when a protection job is run.
If you specify a script that does not exist, an error event and job failure results when a protection
job is run.
If the script generates the name of a volume that already exists, an error event and job failure
results when a protection job is run.

Naming script environment variables for Snapshot copies


The OnCommand console supports a set of environment variables that you can reference or use when
authoring a global naming script to apply to Snapshot copies that are generated when a protection job
is run on a dataset.
You can author your Snapshot copy naming script to use the following variables when specifying the
information to include in Snapshot copy names.
ENV_DATASET_ID

The system-assigned dataset ID

ENV_DATASET_NAME

The user-specified dataset name

ENV_DATASET_LABEL

An optional user-specified custom label

ENV_NODE_ID

The system-assigned ID of the node on which the


operation is being performed

ENV_NODE_NAME

The user-specified name of the node on which the


operation is being performed

ENV_STORAGE_SYSTEM_ID

The system-assigned ID of the storage system on which the


operation is being performed

ENV_STORAGE_SYSTEM_NAME The user-specified name of the storage system on which


the operation is being performed

Administration | 425
ENV_VOLUME_ID

The system-assigned ID of the volume on which the


operation is being performed

ENV_VOLUME_NAME

The user-specified name of the volume on which the


operation is being performed

ENV_TIMESTAMP

The timestamp indicating the date and time of Snapshot


copy generation

ENV_RETENTION_TYPE

The retention type (Hourly, Daily, Weekly, or Monthly) of


the Snapshot copy being generated

Naming script environment variables for primary volumes


The OnCommand console supports a set of environment variables that you can reference or use when
authoring a global naming script to apply to primary volumes that are generated when provisioning
or protection jobs are run on a dataset.
You can author your primary volume naming script to use the following variables when specifying
the information to include in primary volume names.
ENV_DATASET_ID

The system-assigned dataset ID

ENV_DATASET_NAME The user-specified dataset name


ENV_DATASET_LABEL The user-specified custom label
ENV_NODE_ID

The system-assigned ID of the node on which the operation is being


performed (not supported if no protection policy is assigned)

ENV_NODE_NAME

The user-specified name of the node on which the operation is being


performed (not supported if no protection policy is assigned)

Naming script environment variables for secondary volumes


The OnCommand console supports a set of environment variables that you can reference or use when
authoring a global naming script to apply to secondary volumes that are generated when protection
jobs are run on datasets.
You can author your secondary volume naming script to use the following variables when specifying
the information to include in secondary volume names.
ENV_DATASET_ID

The system-assigned dataset ID

ENV_DATASET_NAME

The user-specified dataset name

ENV_DATASET_LABEL

The user-specified custom label

ENV_NODE_ID

The system-assigned ID of the node on which the


operation is being performed

ENV_NODE_NAME

The user-specified name of the node on which the


operation is being performed

426 | OnCommand Console Help


ENV_PRI_STORAGE_SYSTEM_ID

The system-assigned ID of the primary storage


system (not supported for Open Systems SnapVault
configurations)

ENV_PRI_STORAGE_SYSTEM_NAME The user-specified name of the primary storage


system (not supported for Open Systems SnapVault
configurations)
ENV_PRI_VOLUME_ID

The system-assigned ID of the primary root volume


(not supported for Open Systems SnapVault
configurations)

ENV_PRI_VOLUME_NAME

The user-specified name of the primary root volume


(not supported for Open Systems SnapVault
configurations)

ENV_CONNECTION_TYPE

The type of relationship ("backup or "mirror") this


secondary volume has with its primary volume

ENV_DP_POLICY_ID

The system-assigned ID of the protection policy


attached

ENV_DP_POLICY_NAME

The user-specified name of the protection policy


attached

Snapshot copy naming settings


Snapshot copy naming settings determine how Snapshot copies that are generated by OnCommand
console protection jobs are named.
Default Snapshot copy naming format
The default global naming attribute format for Snapshot copies generated by OnCommand console
protection operations is %T_%R_%L_%H_%N_%A (Timestamp_Retention type_Custom
label_Storage system name_Volume name_Application field ).
This format applies to the naming of all Snapshot copies generated by OnCommand console
protection operations.
Snapshot copy naming attributes
Snapshot copy names can be configured to contain the following attributes:
%T (Timestamp)

Indicates the year, month, date, and time of the Snapshot copy. Timestamp
is in the format yyyy-mm-dd_hhmm (along with UTC offset).

%R (Retention type) Indicates whether the Snapshot copy's retention class is hourly, daily,
weekly, monthly, or unlimited.
%L (Custom label)

Enables you to specify a custom alphanumeric character, . (period), _


(underscore), or - (hyphen), to include in the names of the related objects

Administration | 427
that are generated by protection jobs that are run on this dataset. If the
naming format for a related object type includes the Custom label
attribute, then the value that you specify is included in the related object
names. If you do not specify a value, then the dataset name is used as the
custom label. If you include a blank space in the custom label string, the
blank space is converted to letter x in any Snapshot copy, volume, or qtree
object name that includes the custom label as part of its syntax.
%H (Storage system Indicates the name of the storage system that contains the volume from
which a Snapshot copy is made.
name)
%N (Volume name)

Indicates the name of the volume from which a Snapshot copy is made.

%A (Application
fields)

Indicates application-inserted fields. For the NetApp Management Console


data protection capability, it is a list of qtrees present in the volume from
which a Snapshot copy is made.

%1 (One-digit
suffix), %2 (Twodigit suffix), or %3
(Three-digit suffix)

Specifies a one-digit, two-digit, or three-digit suffix if required to


distinguish Snapshot copies.

Example: Snapshot copy custom naming


If a dataset has a custom label of "my_data," includes a volume named "myVol" on a storage
system named "mgt-u35," and is configured for Hourly backup, then the following Snapshot
copy format strings result in the following names for the resulting Snapshot copies:
Format string

Resulting name

%L_%R_%T_%H

my_data_hourly_2010-03-04_0330+0430_mgt-u35

%T_myunit_%Lmysection-%R

2010-03-04_0330_myunit_my_data-mysection-hourly

myunit-mydept-%R_%H_
%T

myunit-mydept-hourly_mgt-u35_2010-03-04_0403-0800

%R_%T_%N_%A

hourly_2010-03-04_0330_myVol_qtree1_qtree2_qtree3

%T

2010-03-04_0330 (if no UTC offset)


2010-03-04_0330+0530 (for IST)

%L_%R_%H_%2

my_data_hourly_mgt-u35_01
my_data_hourly_mgt-u35_02

428 | OnCommand Console Help


Snapshot copy naming exceptions
The following are Snapshot copy naming exceptions:

When a SMHV plug-in creates a Snapshot copy in a host system, the plug-in creates two
Snapshot copies for every backup.
The second Snapshot copy has the string "_backup" appended to the end of the Snapshot copy
name irrespective of the order of attribute selection.
When a SMHV plug-in creates a Snapshot copy, and the Snapshot copy name exceeds the SMHV
Snapshot copy character limit, the SMHV plug-in does not truncate the name by removing the
characters of the Application fields attribute.
Instead, it truncates the name by removing characters before the Application fields
attribute, from right to left.
For Snapshot copies created by a SMHV plug-in, if the Application fields attribute is not
specified, it is added automatically at the end of the naming format.
For Snapshot copies created by a SMVI plug-in, if the Application fields attribute is not
specified, it is not added to the naming format.
For Snapshot copies created by the NetApp Management Console data protection capability, if
the Application fields attribute is not mentioned, it is not added implicitly in the naming
format.
When SMVI plug-in creates a Snapshot copy, and the Snapshot copy name exceeds the SMVI
Snapshot copy character limit, SMVI plug-in does not truncate the name by removing the
characters of the Application fields attribute.
Instead, it truncates the name by removing characters before the Application fields
attribute, from right to left.
If you want to use scripts to generate the Snapshot copy name, and if the Snapshot copy is
generated by the SnapManager plug-in in the host system, the plug-in does not use the user script.
Instead, the plug-in uses the global naming format to create the Snapshot copy name. The user
script is used only if the Snapshot copy is created by the NetApp Management Console data
protection capability.
Snapshot copies created by the host system are in the local time zone of the host system.

Snapshot copy naming restrictions


The following are Snapshot copy naming restrictions:

The timestamp attribute is mandatory.


The Application fields attribute is mandatory in the custom Snapshot copy naming for all
application datasets controlled by the SMHV plug-in Host services.
The Snapshot copy name can only display ASCII alphanumeric characters, _ (underscore), (hyphen), + (plus sign), and . (dot).
Any other characters cause errors.
If there is no custom label for the dataset, the Snapshot copy name defaults to the dataset name.
UTF-8 encoded characters are not supported.

Administration | 429

Including four digits reserved for suffixes, a Snapshot copy name cannot exceed 128 characters.
The Snapshot copy name, excluding the suffixes, can be no more than 124 characters. If the
generated Snapshot copy name exceeds 124 characters, then the name is truncated by removing
characters from right to left.
To avoid possible truncation of timestamp information from the Snapshot copy name, best
practice is to place the timestamp %T attribute at the left end of the format string.

Primary volume naming settings


Primary volume naming settings determine how primary storage volumes generated by OnCommand
console provisioning operations are named.
Default primary volume naming format
The default global naming format for primary volumes generated by OnCommand console
provisioning operations is %L (Custom label).
The global format applies to the naming of all NAS primary volumes generated by OnCommand
console provisioning operations. The default global naming format does not apply to SAN volumes,
because they are named directly at the time of provisioning.
Primary volume naming attributes
Primary volume names can be configured to contain the following attributes:
%L (Custom
label)

Enables you to specify a custom alphanumeric character, . (period), _


(underscore), or - (hyphen) to include in the names of the related objects that
are generated by protection jobs that are run on this dataset. If the naming
format for a related object type includes the Custom label attribute, then the
value that you specify is included in the related object names. If you do not
specify a value, then the dataset name is used as the custom label. If you
include a blank space in the custom label string, the blank space is converted to
letter x in any Snapshot copy, volume, or qtree object name that includes the
custom label as part of its syntax.

%D (Dataset
name)

Indicates the actual name of the dataset in which a volume is created.

%1 (One-digit
suffix), %2
(Two-digit
suffix), or %3
(Three-digit
suffix)

Displays a one-digit, two-digit, or three-digit suffix if required to distinguish


primary volumes.

430 | OnCommand Console Help

Example: Primary volume custom naming


If a dataset named "mydataset" has a custom label of "mydata", and a new volume is
provisioned through the OnCommand console, then the following primary volume format
strings result in the following names for the resulting primary volumes:
Format string

Resulting name

%L_%D

mydata_mydataset

%L

mydata

pri_%L

pri_mydata

myunit-privol

myunit-privol

%L_%D_%3

mydata_mydataset_001
mydata_mydataset_002

Primary volume naming exceptions


The following are primary volume naming exceptions:

If the primary volume's naming format in one or more datasets is customized, then the primary
volumes generated in those datasets are named according to the dataset-level format.
For a primary volume to be provisioned, if you specify the primary volume name from the
OnCommand console user interface, then the name that you specify in the user interface takes
precedence over the options for primary volume naming settings.

Primary volume naming restrictions


The following are the naming restrictions of primary volume:

All naming attributes are optional.


At least one attribute must be enabled at any point of time, or there should be some free-form
text.
In case of name conflict, numerical suffixes are appended to the names.
If the Custom label attribute is included in the naming format, but no custom name exists for a
dataset, the resulting names use the actual dataset name instead.
Including four digits reserved for suffixes, a primary volume name cannot exceed 64 characters.
The primary volume name, excluding the suffixes, can be no more than 60 characters. If the
generated primary volume name exceeds 60 characters, then the name gets truncated by removing
characters from left to right.
The primary volume name can only display ASCII alphanumeric characters and _ (underscore).
Any other characters cause errors. The primary volume name cannot start with a number.

Administration | 431

Secondary volume naming settings


Secondary volume naming settings determine how secondary storage volumes that are generated by
OnCommand console protection operations are named.
Default secondary volume naming format
The default global naming format for secondary volumes generated by OnCommand console
protection operations is %V (Primary volume name).
This format applies to the naming of all secondary volumes generated by OnCommand console
protection operations.
Secondary volume naming attributes
Secondary volume names can be configured to contain the following attributes:
%L (Custom label)

Enables you to specify a custom alphanumeric character, . (period), _


(underscore), or - (hyphen) to include in the names of the related objects that
are generated by protection jobs that are run on this dataset. If the naming
format for a related object type includes the Custom label attribute, then
the value that you specify is included in the related object names. If you do
not specify a value, then the dataset name is used as the custom label. If you
include a blank space in the custom label string, the blank space is converted
to letter x in any Snapshot copy, volume, or qtree object name that includes
the custom label as part of its syntax.

%C (Type)

Indicates the connection type (backup or mirror).

%S (Primary
storage system
name)

Indicates the name of the primary storage system.

%V (Primary
volume name)

Indicates the name of the primary volume.

%1 (One-digit
suffix), %2 (Twodigit suffix), or %3
(Three-digit suffix)

Displays a one-digit, two-digit, or three-digit suffix if required to distinguish


secondary volumes.

Example: Secondary volume custom naming


If a dataset has a custom label of "mydata," includes a primary volume named "myVol1" on a
primary storage system named "myhost1," and is configured for Hourly backup, then the
following secondary volume format strings result in the following names for the resulting
secondary volumes:

432 | OnCommand Console Help


Format string

Resulting name

%L_C_%S_%V

mydata_backup_myhost1_myVol1

%C-%S-%L-destVol

backup_muhost1_mydata_destVol

%C_%L

backup_mydata

%V

myhost1

%C_%L_%1

backup_mydata_1
backup_mydata_2

Secondary volume naming exceptions


The following is secondary volume naming exception:

If secondary volume's naming format in one or more datasets is customized, then the secondary
volumes generated in those datasets are named accordingly.

Secondary volume naming restrictions


The following are secondary volume naming restrictions:

All naming attributes are optional.


At least one attribute must be enabled at any point of time, or there must be some free-form text.
In case of name conflict, numerical suffixes are appended to the names.
If the Fan-in feature is enabled for a backup destination, and two or more qtrees from different
primary volumes are backed up into the same secondary volume, then the Primary storage
system name and Primary volume name attributes of the source volumes are randomly
selected to form the names of the secondary volumes.
For example, if host1:/vol1/qtr1, host2:/vol2/qtr2, and host3:/vol3/qt3 are backed up to one
secondary volume, then all the names for the secondary qtrees in that volume include one
common <Host name> and <Volume name> attribute combination character string. That
common string is either: "host1_vol1", "host2_vol2," or "host3_vol3."
If the Custom label attribute is included in the naming format, but no custom name exists for a
dataset, the resulting names use the actual dataset name instead.
Including four digits reserved for suffixes, a secondary volume name cannot exceed 64
characters.
The secondary volume name, excluding the suffixes, can be no more than 60 characters. If the
generated secondary volume name exceeds 60 characters, then the name gets truncated by
removing characters from left to right.
The secondary volume name can only display ASCII alphanumeric characters and _ (underscore).
Any other characters cause errors. The secondary volume name cannot start with a number.

Administration | 433

When taking a backup of Open Systems SnapVault (OSSV) directories, if you include %V
(Primary volume name) attribute in the naming format, %V will be replaced with %S
(Primary storage system name).
This is because OSSV does not have the concept of volume.

Secondary qtree naming settings


Secondary qtree naming settings determine how secondary storage qtrees that are generated by
OnCommand console protection operations are named.
Default secondary qtree naming format
The default global naming format for secondary qtrees generated by OnCommand console protection
operations is %Q (Primary qtree name).
This format applies to the naming of all secondary qtrees generated by OnCommand console
protection operations.
Secondary qtree attributes
Secondary qtree names can be configured to contain the following attributes:
%Q (Primary qtree The name of the primary qtree.
name)
%L (Custom label)

Enables you to specify a custom alphanumeric character, . (period), _


(underscore), or - (hyphen) to include in the names of the related objects
that are generated by protection jobs that are run on this dataset. If the
naming format for a related object type includes the Custom label
attribute, then the value that you specify is included in the related object
names. If you do not specify a value, then the dataset name is used as the
custom label. If you include a blank space in the custom label string, the
blank space is converted to letter x in any Snapshot copy, volume, or qtree
object name that includes the custom label as part of its syntax.

%S (Primary
storage system)

Indicates the name of the primary storage system.

%V (Primary
volume name)

Indicates the name of the primary storage volume.

%1 (One-digit
suffix), %2 (Twodigit suffix), or %3
(Three-digit suffix)

Displays a one-digit, two-digit, or three-digit suffix if required to distinguish


secondary qtrees.

Example: Secondary qtree custom naming


If a dataset has a custom label of "mydata," includes a primary qtree named "qtree1", and
primary volume named "myVol1" on a primary storage system named "myhost1," and is

434 | OnCommand Console Help


configured for Hourly backup, then the following secondary qtree format strings result in the
following names for the resulting secondary qtrees:
Format string

Resulting name

%L_%S_%V_%Q

mydata_myhost1_myVol1_qtree1

%L_%S_%Q

mydata_myhost1_qtree1

%V_%Q

myVol1_qtree1

%L_%S_%V_%Q_%3

mydata_myhost1_myVol1_qtree1_001
mydata_myhost1_myVol1_qtree1_002

Secondary qtree naming exceptions


The following is secondary qtree naming exception:

If secondary qtree naming format in one or more datasets is customized, the secondary qtrees
generated in those datasets are named accordingly.

Secondary qtree naming restrictions


The following are secondary qtree naming restrictions:

All naming attributes are optional.


At least one attribute must be enabled at any point of time, or there must be some free-form text.
In case of name conflict, numerical suffixes are appended to the names.
If the Custom label attribute is included in the naming format, but no custom name exists for a
dataset, the resulting names use the actual dataset name instead.
Including four digits reserved for suffixes, a secondary qtree name cannot exceed 64 characters.
The secondary qtree name, excluding the suffixes, can be no more than 60 characters. If the
generated secondary qtree name exceeds 60 characters, then the name gets truncated by removing
characters from left to right.
The secondary qtree name can only display ASCII alphanumeric characters, _ (underscore), (hyphen), and . (dot).
Any other characters cause errors.
When taking a backup of Open Systems SnapVault directories, if you include the Primary
volume name attribute in the naming format, it is replaced by the Primary storage system
attribute.
When taking a backup of Open Systems SnapVault directories, if you include the Primary
qtree name attribute in the naming format, it is replaced by the path of the root directory.
The characters of the root directory path that are not supported are converted to the letter x.
When taking a backup of Open Systems SnapVault directories, if you include the Primary
qtree name attribute in the naming format, and if the directory path contains non-ASCII

Administration | 435
characters, or if the directory path is / (slash), then the Primary qtree name attribute is
replaced with the directory ID.
Managing naming settings options
Globally customizing naming of protection-related objects
To improve recognition and usability, you can globally customize the naming formats of all
protection-related objects (Snapshot copies, primary volumes, secondary volumes, or secondary
qtrees that are generated by the OnCommand console protection or provisioning operations).
Before you begin

You must have reviewed the Guidelines for globally customizing naming of protection-related
objects on page 436
You must have reviewed the Requirements and restrictions when customizing naming of
protection-related objects on page 438
Have the following custom naming information available:

The protection-related object types whose naming you want to customize


If you want to customize by selecting and ordering OnCommand console-supported attributes
in the Name format field, the naming attributes that you want to include in the naming format
If you want to customize naming by using a pre-authored naming script, the name and
location of that script
You must be authorized to perform all the steps of this task; your RBAC administrator can
confirm your authorization in advance.

About this task

Configuring global custom naming for a related object type applies to all objects of that type that are
generated by OnCommand console protection or provisioning jobs unless those objects belong to a
dataset that is already configured with dataset-level custom naming for object type in question.
Steps

1. In the OnCommand console Setup Options dialog box, select Naming Settings and select the
object type (Snapshot copy, primary volume, secondary volume, or secondary qtree) whose
global naming format you want to customize.
2. Customize the naming settings for the selected object type.

Select Use naming format if you want to customize naming by selecting and ordering
naming attributes.
Then use the name format field to complete your selection and ordering of naming attributes.
Select Use naming script if you want to customize naming by a pre-authored script.

436 | OnCommand Console Help


Then complete the Script path and Run as fields to specify the path to find the script and the
authorized user under whose identity to execute the script.
3. Click OK.
Result

The OnCommand console applies the custom global naming format to all protection job or
provisioning job generated objects of the type that you customized. Exception is made only to those
objects that belong to a dataset that has dataset-level naming customized for it.
Related references

Administrator roles and capabilities on page 506


Guidelines for globally customizing naming of protection-related objects
Before you use the Settings options to customize global naming for protection-related objects (the
Snapshot copies, primary volumes, secondary volumes, or secondary qtrees that are generated by
OnCommand console protection jobs or provisioning jobs) you need to decide how you want to
customize the global naming formats.
Custom naming by script
You can use a script to customize global naming settings of your Snapshot copy, primary volume, or
secondary volume objects if custom naming by OnCommand console-supported attribute selection
and ordering is inadequate for your global Snapshot copy, primary volume, and secondary volume
naming needs.
Note: The OnCommand console does not support global customization of secondary qtree naming

by script.
Note: You do not have to assign a policy to create a new dataset. You can assign a policy to the
dataset later by running the Dataset Policy Change wizard.

If you intend to customize global naming by script, you should specify the path location to a preauthored script and the name of an application to read and execute that script.
Note: The output of the naming script determines what name is generated and given to the
dataset's related objects for whom the script is written. If the script generates an error message,
part of that error message might be included in the names generated for the objects in question.

Custom naming by attribute selection and ordering


The most common method of customizing global naming of your Snapshot copy, primary volume,
secondary volume, or secondary qtree objects is by specifying the selection and ordering of their
OnCommand console-supported naming attributes. Each object type for which custom naming is
supported supports a different set of attributes.

Administration | 437
Snapshot
copy naming
attributes

Primary
volume
naming
attributes

Secondary
volume
naming
attributes

Attributes that can be specified and ordered to customize global Snapshot copy
naming include the following:

%T (Timestamp) (required)
%R (Retention type) (retention class of the Snapshot copy)
%L (Custom label) (custom label of the containing dataset, if one exists)
%H (Storage system name) (storage system of the containing volume)
%N (Volume name) (the containing volume)
%A (Application fields)
%1 (One-digit suffix), %2 (Two-digit suffix), or %3 (Threedigit suffix) (applied as necessary if naming differentiation is required)

Attributes that can be specified and ordered to customize global primary volume
naming include the following:

%L (Custom label) (custom label of the containing dataset, if one exists)


%D (Dataset name) (actual name of the containing dataset)
%1 (One-digit suffix), %2 (Two-digit suffix), or %3 (Threedigit suffix) (applied as necessary if naming differentiation is required)

Attributes that can be specified and ordered to customize global secondary volume
naming include the following:

%L (Custom label) (custom label of the containing dataset, if one exists)


%C (Type) (the connection type: backup or mirror).
%S (Primary storage system name) (storage system of the volume being

backed up or mirrored)

%V (Primary volume name) (volume being backed up or mirrored)


%1 (One-digit suffix), %2 (Two-digit suffix), or %3 (Threedigit suffix) (applied as necessary if naming differentiation is required)

Attributes that can be specified and ordered to customize global secondary volume
Secondary
qtree naming naming include the following:
attributes
%Q (Primary qtree name) (qtree being backed up)
%L (Custom label) (custom label of the containing dataset, if one exists)
%S (Primary storage system name) (storage system of the volume being
backed up or mirrored)
%V (Primary volume name) (volume being backed up or mirrored)
%1 (One-digit suffix), %2 (Two-digit suffix), or %3 (Threedigit suffix) (applied as necessary if naming differentiation is required)

438 | OnCommand Console Help

Requirements and restrictions when customizing naming of protection-related


objects
You must follow the requirements and restrictions when customizing naming of protection-related
objects (Snapshot copies, volumes, or qtrees that are generated by protection jobs run on datasets).
Snapshot copies
The following are the naming restrictions of Snapshot copies:

The timestamp attribute is mandatory in the name.


The Application fields attribute is mandatory in the custom Snapshot copy naming for all
application datasets controlled by Host services.
Snapshot copy name can only display ASCII alphanumeric characters, _ (underscore), - (hyphen),
+ (plus sign), and . (dot).
Any other characters cause errors.
If there is no custom label for the dataset, the Snapshot copy name defaults to the dataset name.
UTF-8 encoded characters are not supported.
Snapshot copy name can be 124 characters long.
If the generated Snapshot copy name exceeds 124 characters, then the name gets truncated by
removing characters from right to left. Additional 4 digits are dedicated for suffixes. Therefore,
the maximum length of Snapshot copy name including the suffix can be 128 characters.

Primary volume
The following are the naming restrictions of primary volume:

All naming attributes are optional.


At least one attribute must be enabled at any point of time, or there should be some free-form
text.
In case of name conflict, numerical suffixes are appended to the names.
If the Custom label attribute is included in the naming format, but no custom name exists for a
dataset, the resulting names use the actual dataset name instead.
Primary volume name can be 60 characters long.
If the generated primary volume name exceeds 60 characters, then the name gets truncated by
removing characters from left to right. Additional 4 digits are dedicated for suffixes. Therefore,
the maximum length of primary volume name including the suffix can be 64 characters long.
Primary volume name can only display ASCII alphanumeric characters, and _ (underscore).
Any other characters cause errors. Primary volume name cannot start with a number.

Secondary volume
The following are the naming restrictions of secondary volume:

All naming attributes are optional.

Administration | 439

At least one attribute must be enabled at any point of time, or there must be some free-form text.
In case of name conflict, numerical suffixes are appended to the names.
If the Fan-in feature is enabled for a backup destination, and two or more qtrees from different
primary volumes are backed up into the same secondary volume, then the Primary storage
system name and Primary volume name attributes of the source volumes are randomly
selected to form the names of the secondary volumes.
For example, if host1:/vol1/qtr1, host2:/vol2/qtr2, and host3:/vol3/qt3 are backed up to one
secondary volume, then all the names for the secondary qtrees in that volume include one
common <Host name> and <Volume name> attribute combination character string. That
common string is either: "host1_vol1", "host2_vol2," or "host3_vol3."
If the Custom label attribute is included in the naming format, but no custom name exists for a
dataset, the resulting names use the actual dataset name instead.
Secondary volume name can be 60 characters long.
If the generated secondary volume name exceeds 60 characters, then the name gets truncated by
removing characters from left to right. Additional 4 digits are dedicated for suffixes. Therefore,
the maximum length of secondary volume name including the suffix can be 64 characters long.
Secondary volume name can only display ASCII alphanumeric characters, and _ (underscore).
Any other characters cause errors. Secondary volume name cannot start with a number.

Secondary qtree
The following are the naming restrictions of secondary qtree naming:

All naming attributes are optional.


At least one attribute must be enabled at any point of time, or there must be some free-form text.
In case of name conflict, numerical suffixes are appended to the names.
If the Custom label attribute is included in the naming format, but no custom name exists for a
dataset, the resulting names use the actual dataset name instead.
Secondary qtree name can be 60 characters long.
If the generated secondary qtree name exceeds 60 characters, then the name gets truncated by
removing characters from left to right. Additional 4 digits are dedicated for suffixes. Therefore,
the maximum length of secondary qtree name including the suffix can be 64 characters long.
Secondary qtree name can only display ASCII alphanumeric characters, _ (underscore), (hyphen), and . (dot).
Any other characters cause errors.
For non qtree data (Qtree0) naming on the secondary, if the secondary qtree name end up being (hyphen), or "etc", then the naming format for the non qtree data secondary qtree is (Custom
label)_(Primary storage name)_(Primary volume name).
When taking a backup of Open Systems SnapVault directories, if you include the Primary
volume name attribute in the naming format, it is replaced by the Primary storage system
attribute.
When taking a backup of Open Systems SnapVault directories, if you include the Primary
qtree name attribute in the naming format, it is replaced by the path of the root directory.

440 | OnCommand Console Help

The characters of the root directory path that are not supported are converted to letter "x".
When taking a backup of Open Systems SnapVault directories, if you include the Primary
qtree name attribute in the naming format, and if the directory path contains non-ASCII
characters, or if the directory path is / (slash), then the Primary qtree name attribute is
replaced with the directory ID.

Page descriptions
Global Naming Settings Snapshot Copy area
You can customize the Snapshot copy naming settings in the Setup Options dialog box to determine
global-level names for Snapshot copies generated by dataset protection jobs.
Options
You can specify a Snapshot copy name at the global level by specifying either a naming format or
the path to a naming script.
Note: The global-level Snapshot copy naming settings apply to any Snapshot copy that is
generated by a OnCommand console protection job that is executed on a dataset that does not have
dataset-level Snapshot copy naming settings specified. Dataset-level Snapshot copy naming
settings take precedence over global-level naming settings.

Selecting this option specifies that global-level Snapshot copy naming is determined by
Use
Naming the format that is specified in the Name Format field.
Format
Enables you to specify global-level naming attributes of any Snapshot copy
Name
Format that is generated by OnCommand console protection jobs. You can type
the following attributes (separated by the underscore character) in this field
in any order:

%T (timestamp attribute)

The year, month, day, and time of the Snapshot copy.


%R (retention type attribute)
The Snapshot copy's retention class (Hourly, Daily, Weekly, Monthly,
or Unlimited)
%L (custom label attribute)
The custom label, if any, that is specified for the Snapshot copy's
containing dataset. If no custom label is specified, then the dataset
name is included in the Snapshot copy name.
It enables you to specify a custom alphanumeric character, . (period), _
(underscore), or - (hyphen) to include in the names of the related
objects that are generated by protection jobs that are run on this dataset.
If the naming format for a related object type includes the Custom
label attribute, then the value that you specify is included in the
related object names. If you do not specify a value, then the dataset

Administration | 441

name is used as the custom label. If you include a blank space in the
custom label string, the blank space is converted to letter x in any
Snapshot copy, volume, or qtree object name that includes the custom
label as part of its syntax.
%H (storage system name attribute)
The name of the storage system that contains the volume from which a
Snapshot copy is made
%N (volume name attribute)
The name of the volume from which a Snapshot copy is made
%A (application fields attribute)
Data inserted by outside applications into the name of the Snapshot
copy. In the case of regular datasets, %A contains a list of qtrees on the
volume for which the Snapshot copy is made.
%1, %2, %3 (digit suffix)
A one-digit, two-digit, or three-digit suffix, if required, to distinguish
Snapshot copies with otherwise matching names

The default Snapshot copy naming format is %T_%R_%L_%H_%N_%A.


Name
Preview

Displays a sample Snapshot copy name based on the attributes that you
entered in the Name Format field.
For example, if Snapshot copy naming format is customized as %T_%R_
%L_%H_%N_%A , and if a dataset with custom label "mydata" has some
data backed up "hourly" from primary to the backup destination mgt-u35:/
myVol, then the name of the Snapshot copy on myVol is
"2010-03-04_03.30.45+0430_hourly_mydata_mgtu35_myVol_(Application fields)".

Selecting this option specifies at a global level the path and name of a user-authored
Use
Naming naming script. The script specifies how a Snapshot copy that is generated by
OnCommand console protection jobs is named. Naming scripts for Snapshot copies
Script
apply to datasets of physical storage objects only. They do not apply to datasets of
virtual objects. Host services will use the Name Format instead of the script.
Script Path The path and file name of a user-supplied naming script. The naming
script must be in a location that is readable by DataFabric Manager
server.
Run As:

The name of the authorized user under whose identity the operations
that are specified in the naming script are executed.

442 | OnCommand Console Help

Global Naming Settings Primary Volume area


You can customize the primary volume naming settings in the Setup Options dialog box to determine
global-level names of primary volumes that are generated by OnCommand console protection jobs.
Options
You can specify a primary volume naming at the global level by specifying either a naming format or
the path to a naming script.
Note: The global-level primary volume naming settings apply to any primary volume that is

generated by an OnCommand console protection job that is executed on a dataset that does not
have dataset-level primary volume naming settings specified. Dataset-level primary volume
naming settings take precedence over global-level naming settings.
Use
Naming
Format

Selecting this option specifies that global-level primary volume naming is determined
by the format that is specified in the Name Format field.
Name
Format

Enables you to specify at a global-level the naming attributes that are to


be included in the name of any primary volume that is generated by
OnCommand console protection jobs. You can type the following
attributes (separated by the underscore character) in this field in any
order:

%L (custom label attribute)

The custom label, if any, that is specified for the primary volume's
containing dataset. If no custom label is specified, then the dataset
name is included in the primary volume name.
%D (dataset name)
The actual name of the dataset in which a volume was created
%1, %2, %3 (digit suffix)
A one-digit, two-digit, or three-digit suffix, if required, to distinguish
primary volumes with otherwise matching names

The default primary volume naming format is %L.


Name
Preview

Displays a sample primary volume name based on the attributes that you
entered in the Name Format field.
For example, if primary volume naming format is customized as %L_
%D , and if a dataset named "mydataset" with custom label "mydata"
and a new volume is provisioned through the OnCommand console, then
the name of the primary volume is "mydata_mydataset."

Use
Naming
Script

Selecting this option specifies at a global level the path location and name of a userauthored naming script. The script specifies how a primary volume that is generated by

Administration | 443
OnCommand console protection jobs is named. Host services will use the Name
Format instead of the script.
Script Path The path and file name of a user-supplied naming script. The naming
script must be in a location that is readable by DataFabric Manager
server.
Run As:

The name of the authorized user under whose identity the operations
that are specified in the naming script are executed.

Global Naming Settings Secondary Volume area


You can customize the secondary volume naming settings in the Setup Options dialog box to
determine the global-level names of secondary volumes that are generated by OnCommand console
protection jobs.
Options
You can specify a secondary volume naming at the global level by specifying either a naming format
or the path to a naming script.
Note: The global-level secondary volume naming settings apply to any secondary volume that is
generated by an OnCommand console protection job that is executed on a dataset that does not
have dataset-level secondary volume naming settings specified. Dataset-level secondary volume
naming settings take precedence over global-level naming settings.

Use
Naming
Format

Selecting this option specifies that global-level secondary volume naming is determined
by the format that is specified in the Name Format field of this option.
Name
Format

Enables you to specify at a global level the naming attributes that are to be
included in the name of any secondary volume that is generated by
OnCommand console protection jobs. You can type the following
attributes (separated by the underscore character) in this field in any order:

%L (custom label attribute)

The custom label, if any, that is specified for the secondary volume's
containing dataset. If no custom label is specified, then the dataset
name is included in the secondary volume name.
It enables you to specify a custom alphanumeric character, . (period), _
(underscore), or - (hyphen) to include in the names of the related
objects that are generated by protection jobs that are run on this
dataset. If the naming format for a related object type includes the
Custom label attribute, then the value that you specify is included in
the related object names. If you do not specify a value, then the dataset
name is used as the custom label. If you include a blank space in the
custom label string, the blank space is converted to letter x in any

444 | OnCommand Console Help

Snapshot copy, volume, or qtree object name that includes the custom
label as part of its syntax.
%S (primary storage system name)
The name of the primary storage system
%V (primary volume name)
The name of the primary volume
%C (type)
The connection type (backup or mirror)
%1, %2, %3 (digit suffix)
A one-digit, two-digit, or three-digit suffix, if required, to distinguish
secondary volumes with otherwise matching names

The default secondary volume naming format is %V .


Name
Preview

Displays a sample secondary volume name based on the attributes that


you entered in the Name Format field.
For example, if secondary volume naming format is customized as %L_
%C_%S_%V, and if a dataset with custom label "mydata" has some data
backed up "hourly" from primary myhost1:/myvol1 to the backup
destination "myhost2:/myvol2", then the name of the secondary volume is
"mydata_backup_myhost1_myvol1".

Use
Naming
Script

Selecting this option specifies at a global level the path location and name of a userauthored naming script. The script specifies how a secondary volume that is generated
by OnCommand console protection jobs is named. Host services will use the Name
Format instead of the script.
Script Path The path and file name of a user-supplied naming script. The naming
script must be in a location that is readable by DataFabric Manager
server.
Run As:

The name of the authorized user under whose identity the operations
that are specified in the naming script are executed.

Global Naming Settings Secondary Qtree area


You can customize the secondary qtree naming settings in the Setup Options dialog box to determine
the global-level names of secondary qtrees that are generated by OnCommand console protection
jobs.
Options
You can specify a secondary qtree naming at the global level by specifying a naming format.
Note: The global-level secondary qtree naming settings apply to any secondary qtree that is
generated by an OnCommand console protection job that is executed on a dataset that does not

Administration | 445
have dataset-level secondary qtree naming settings specified. Dataset-level secondary qtree
naming settings take precedence over global-level naming settings.
Name
Format

Enables you to specify at a global level the naming attributes that are to be included in
the name of any secondary qtree that is generated by OnCommand console protection
jobs. You can type the following attributes (separated by the underscore character) in
this field in any order:

%L (custom label attribute)

The custom label, if any, that is specified for the secondary qtree's containing
dataset. If no custom label is specified, then the dataset name is included in the
secondary qtree name.
It enables you to specify a custom alphanumeric character, . (period), _ (underscore),
or - (hyphen) to include in the names of the related objects that are generated by
protection jobs that are run on this dataset. If the naming format for a related object
type includes the Custom label attribute, then the value that you specify is
included in the related object names. If you do not specify a value, then the dataset
name is used as the custom label. If you include a blank space in the custom label
string, the blank space is converted to letter x in any Snapshot copy, volume, or
qtree object name that includes the custom label as part of its syntax.
%S (primary storage system name)
The name of the primary storage system
%V (primary volume name)
The name of the primary volume
%Q (primary qtree name)
The name of the primary qtree
%1, %2, %3 (digit suffix)
A one-digit, two-digit, or three-digit suffix, if required, to distinguish secondary
qtrees with otherwise matching names.

The default secondary qtree naming format is %Q .


Name
Preview

Displays a sample secondary qtree name based on the attributes that you entered in the
Name Format field.
For example, if secondary qtree naming format is customized as %L_%S_%V_%Q, and
if a dataset with custom label "mydata" has some data backed up "hourly" from primary
myhost1:/myvol1/qtree1 to the backup destination myhost2:/myvol2, then the name of
the secondary qtree is "qtree1_mydata_host1_vol1".

Costing setup options


Understanding costing options

446 | OnCommand Console Help

What chargeback is
You can configure the DataFabric Manager server to collect data related to space usage by
individual, or groups of, appliances and file systems. The statistical information you collect can be
used for chargeback, and planning space utilization.
The chargeback reports provide an easy way to track space usage and to generate bills based on your
specifications. If your organization bills other organizations or groups in your company for the
storage services they use, you can use the chargeback reports.
Managing costing options
Editing chargeback options
You can edit the chargeback options for objects to customize the billing information.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click the Setup Options option.
2. In the Setup Options dialog box, click the Costing option.
3. Click the Chargeback option.
The Costing Chargeback area appears.
4. Specify the chargeback increment, currency display format, the amount to charge for disk space
usage, and the day of the month when the billing cycle begins and ends.
5. Click Save and Close.
Related references

Administrator roles and capabilities on page 506


Costing Chargeback area on page 446
Page descriptions
Costing Chargeback area
The chargeback feature in DataFabric Manager server enables you to obtain billing reports for the
amount of space used by a specific object or a group of objects.

Options on page 447

Administration | 447

Command buttons on page 447

Options
You can configure the following chargeback options:
Chargeback
Increment

Displays how the charge rate is calculated. You can specify this setting only at
the global level.
You can specify the following values for the Chargeback increment option:

Daily: Charges are variable and they are adjusted based on the number of
days in the billing period.
Formula used by DataFabric Manager server to calculate the charges: Annual
Rate / 365 x number of days in the billing period
Monthly: Charges are fixed and there is a flat rate for each billing period
regardless of the number of days in the period.
Formula used by DataFabric Manager server to calculate the charges: Annual
Rate / 12

The default is Daily.


Currency
Format

Displays the format of the currency to be specified.


The default is $ #,###.##.

Annual Charge Displays the amount to charge for storage space usage, per GB, per year.
Rate (Per GB)
You must specify the value in the x.y notation, where x is the integer of the
number and y is the fraction. For example, to specify an annual charge rate of $
150.55, you must enter 150.55.
Day of the
Month for
Billing

Displays the day of the month when the billing cycle begins.
You can specify the following values for the day of the month for the billing
option:

1 through 28: These values specify the day of the month.


For example, if you specify 15, the billing cycle begins on the 15th day of the
month.
-27 through 0: These values specify the number of days before the last day of
the month. Therefore, 0 specifies the last day of the month.
For example, if you want the bill on the fifth day before the month ends
every month, specify -4.

The default is one day.


Command buttons
The command buttons enable you to save or cancel the setup options.

448 | OnCommand Console Help


Save and Close

Saves the configuration settings for the selected option.

Cancel

Does not save the recent changes and closes the Setup Options dialog box.

Save

Saves the configuration settings for the selected option.

Database backup setup options


Understanding database backup options
Overview of the DataFabric Manager server database backup process
You can back up the DataFabric Manager server database, script plug-ins, and performance data
without stopping any of the DataFabric Manager server services. However, data collection and view
modifications of Performance Advisor are suspended during the backup process.
There are two types of backups:

Archive
This backup process backs up your critical data in compressed form as a .zip file. The DataFabric
Manager server data is automatically converted to an archive format and the DataFabric Manager
server stores the backup in a local or remote directory. You can easily move an archive-based
backup to a different system and restore it. This backup process is time-consuming.
Snapshot
This backup process uses the Snapshot technology to back up the database. You can quicken the
backup process through this approach. But you cannot transfer a Snapshot backup to a different
system and restore it.

Configuring database backup options


Deleting a database backup
You can delete a database backup that you do not need from OnCommand to save space.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click the Setup Options option.
2. In the Setup Options dialog box, click the Database Backup option.
3. Click the Completed option.
4. In the Database Backup Completed area, select the backup file you want to delete.

Administration | 449
5. Click Delete.
Related references

Database Backup Completed area on page 453


Administrator roles and capabilities on page 506
Managing database backup options
Scheduling a database backup
You can schedule a database backup to occur at a specific time on a recurring basis, to preserve your
data.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can
confirm your authorization in advance.
You must log in to the DataFabric Manager server as an administrator with the GlobalFullControl
role.

Steps

1. Click the Administration menu, then click the Setup Options option.
2. In the Setup Options dialog box, click the Database Backup option.
3. Click the Schedule option.
4. In the Database Backup Schedule area, specify the database backup properties, such as backup
type, backup path, retention count, and schedule.
You can select between the Archive and Snapshot backup types.
5. Select Schedule.
You can configure the time of your database backup schedule in minutes, hours, days, and weeks.
6. Click Save and Close.
Related references

Administrator roles and capabilities on page 506


Database Backup Schedule area on page 451

450 | OnCommand Console Help

Changing the directory path for database backups


You can use the Setup Options dialog box to change the directory path for database backups if you
want to back up to a different location.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click the Setup Options option.
2. In the Setup Options dialog box, click the Database Backup option.
3. Click the Schedule option.
4. In the Database Backup Schedule area, select a backup type:

Archive
Snapshot

5. Change the Backup path.


6. Click Save and Close.
Related references

Administrator roles and capabilities on page 506


Database Backup Schedule area on page 451
Starting a database backup
You must back up your data before any upgrade operation, and before any maintenance on the
system host on the DataFabric Manager server. You can start a database backup from the
OnCommand console in one of the two ways; schedule based or on-demand. Follow the steps to
perform an on-demand backup.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click the Setup Options option.
2. In the Setup Options dialog box, click the Database Backup option.

Administration | 451
3. Click the Schedule option.
4. In the Database Backup Schedule area, select a backup type:

Archive
Snapshot

5. Specify the location where you want to save the backup.


6. Specify a Retention Count.
7. Click Back Up Now.
A message is displayed confirming that the backup has started successfully.
Related references

Administrator roles and capabilities on page 506


Database Backup Schedule area on page 451
Page descriptions
Database Backup Schedule area
You can schedule a database backup, manage available database information, and view the list of
completed backups and related events using the Database Backup setup option.

Options on page 451


Command buttons on page 452

Options
The following options are available in the Database Backup Schedule area:
Status

Displays the status of the scheduled backup and current backup.

Backup
Type

Pending: Indicates that a backup is scheduled, but yet to start.


Schedule Active: Indicates that a backup is scheduled and enabled.
Schedule Inactive: Indicates that a backup is scheduled and disabled.
Running: Indicates that a backup is currently running.
Not Scheduled: Indicates that a backup is not scheduled.
Enables you to select a backup type. You can select one of the following backup types:

Archive: Performs only critical data backup in a compressed form using the ZIP
format. You can transfer the backup to a different system and restore it with ease;
however, this process is time-consuming. By default, Archive is enabled.

452 | OnCommand Console Help

Snapshot: Uses the Snapshot technology to perform the database backup. Although
this is a faster option, you cannot transfer the backup to a different system and
restore it.
You can export a Snapshot backup to Archive backup by using the dfm backup
export snapshot-name command.
Note: Snapshot backup is enabled only when both of the following conditions

exist:

Backup Path

Either SnapDrive for Linux 2.2.1 (and later) or SnapDrive for Windows 4.2
(and later) is installed.
DataFabric Manager server data resides on a dedicated LUN managed by
SnapDrive.

Displays the location at which the backed-up database is stored.

Retention
Count

Specifies the maximum number of backups that the OnCommand console can store
simultaneously. If you exceed this limit, the old backups are automatically deleted
to provide space for new backups. The default retention count is 0.

Schedule Enables you to plan the database backup.


The following options are available to schedule your backup:

Hourly at Minute: Displays the time (in minutes) at which the hourly backup must be
performed.
Every (Hours): Displays the time (in hours) at which the backup must be performed.
Starting Every Day At: Displays the time during the day when the backup must start.
This backup is performed based on the time interval set to the Every (Hours) field.
Daily At: Displays the time during the day when the backup must start. This backup
is performed once every 24 hours.
Weekly On: Displays the day for the weekly backup schedule.
At: Displays the time for the weekly backup schedule.
Note: Hourly backups are not possible with Archive backup. By default, Snapshot
backup is selected when the DataFabric Manager server data resides on a LUN.

Command buttons
The command buttons enable you to save or cancel the setup options, and back up data.
Save and Close

Saves the configuration settings for the selected option.

Cancel

Does not save the recent changes and closes the Setup Options dialog box.

Save

Saves the configuration settings for the selected option.

Administration | 453
Back Up Now

Enables you to start a manual backup.

Database Backup Completed area


You can view information about the database backups that have been completed and the related
events.

Options on page 453


Command buttons on page 453

Options
You can view the following information about the completed database backup:
File Name

Displays the name of the database backup file.

File Size

Displays the size of the database backup file.

Creation Time Displays the time when the database backup file was created.
Backup Events Display the events that are triggered during the database backup. You can also
view the time when the event was triggered.
Command buttons
The command buttons enable you to save or cancel the setup options, and delete the database backup.
Save and Close

Saves the configuration settings for the selected option.

Cancel

Does not save the recent changes and closes the Setup Options dialog box.

Save

Saves the configuration settings for the selected option.

Delete

Enables you to delete a database backup.

Default thresholds setup options


Understanding default thresholds options
Default thresholds
Default thresholds are values or limits set on a storage object attribute, that allow the DataFabric
Manager server to trigger event(s) when the value or limit is reached. You can modify these values to
set specific thresholds for individual storage objects or groups of storage objects.
You can access the global default threshold values from the Setup Options dialog box. From the
Default Thresholds options area, you can change the global default threshold values for the following
objects:

Aggregates

454 | OnCommand Console Help

Volumes
Qtrees
Hosts
Resource pools
User quotas
HBA ports

Managing default thresholds options


Editing threshold conditions for an aggregate
You can change the default threshold settings for an aggregate. When you edit thresholds, you edit
values that are associated with the individual aggregates, not the aggregate groups.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click the Setup Options option.
2. In the Setup Options dialog box, click the Default Thresholds option.
3. Click the Aggregates option.
4. In the Default Thresholds Aggregates area, specify the new values, as required.
5. Click Save and Close.
Related references

Administrator roles and capabilities on page 506


Default Thresholds Aggregates area on page 456
Editing threshold conditions for a volume
You can change the default threshold settings for a volume or all volumes in a group. If you modify
the settings of a volume that belongs to more than one group, the settings are applied to the volume
across all groups it belongs to.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.

Administration | 455
Steps

1. Click the Administration menu, then click the Setup Options option.
2. In the Setup Options dialog box, click the Default Thresholds option.
3. Click the Volumes option.
4. In the Default Thresholds Volumes area, specify the new values, as required.
5. Click Save and Close.
Related references

Administrator roles and capabilities on page 506


Default Thresholds Aggregates area on page 456
Editing threshold conditions for other objects
You can change the default threshold settings for qtrees, hosts, user quotas, resource pools, and HBA
ports.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
About this task

User quota thresholds that you configure for a qtree do not apply to quotas on a qtree.
Steps

1. Click the Administration menu, then click the Setup Options option.
2. In the Setup Options dialog box, click the Default Thresholds option.
3. Click the Other option.
4. In the Default Thresholds Other area, specify the new values, as required.
5. Click Save and Close.
Related references

Administrator roles and capabilities on page 506


Default Thresholds Aggregates area on page 456
Page descriptions

456 | OnCommand Console Help

Default Thresholds Aggregates area


You can change the global default threshold values for monitored aggregates using the Default
Threshold setup option.

Options on page 456


Command buttons on page 457

Options
The following thresholds apply to all monitored aggregates. You can override the default values of
any aggregate from the Aggregates View page.
Full Threshold (%)

Displays the percentage of physical space on an aggregate that can be used


before the system generates an Aggregate Full event. All requests for space
exceeding this threshold are dropped.
The default is 90%.

Nearly Full
Threshold (%)

Displays the percentage of physical space on an aggregate that can be used


before the system generates an Aggregate Almost Full event.
You should to specify a limit that is less than the value specified for
Aggregate Full Threshold.
The default is 80%.

Full Threshold
Interval

Displays the time that a condition can persist before the event is generated.
If the condition persists for the specified amount of time, the DataFabric
Manager server generates an Aggregate Full event. Threshold intervals
apply only to error and informational events.
If the threshold interval is 0 seconds, or a value less than the aggregate
monitoring interval, the DataFabric Manager server continuously generates
Aggregate Full events until an event is resolved. If the threshold interval is
greater than the aggregate monitoring interval, the DataFabric Manager
server waits for the specified threshold interval (which includes two or
more monitoring intervals), and generates an Aggregate Full event only if
the condition persisted throughout the threshold interval.
For example, if the monitoring cycle time is 60 seconds and the threshold
interval is 90 seconds, the threshold event is generated only if the condition
persists for two monitoring cycles.
The default is 0 seconds.

Overcommitted
Threshold (%)

Displays the percentage of physical space on an aggregate that can be used


before the system generates an Aggregate Overcommitted event.
The default is 100%.

Administration | 457
Nearly
Overcommitted
Threshold (%):

Displays the percentage of physical space on an aggregate that can be used


before the system generates an Aggregate Nearly Overcommitted event.
You should specify a limit that is less than the value specified for
Aggregate Overcommitted Threshold.
The default is 95%.

Snapshot Reserve
Full Threshold (%)

Displays the percentage of Snapshot reserve on an aggregate that can be


used before the system generates an Aggregate Snapshot Full event.
The default is 90%.

Snapshot Reserve
Nearly Full
Threshold (%)

Displays the percentage of Snapshot reserve on an aggregate that can be


used before the system generates an Aggregate Snapshot Nearly Full event.
You should specify a limit that is less than the value specified for
Aggregate Snapshot Reserve Full Threshold.
The default is 80%.

Over-Deduplicated
Threshold (%)

Displays the percentage of user data that can be deduplicated and stored on
a volume before the system generates a Volume Over Deduplicated event.
The default is 150%.

Nearly overDeduplicated
Threshold (%)

Displays the percentage of user data that can be deduplicated and stored on
a volume before the system generates a Volume Nearly Over Deduplicated
event.
The default is 140%.

Command buttons
The command buttons enable you to save or cancel the setup options.
Save and Close

Saves the configuration settings for the selected option.

Cancel

Does not save the recent changes and closes the Setup Options dialog box.

Save

Saves the configuration settings for the selected option.

Default Thresholds Volumes area


You can change the global default threshold values for monitored volumes using the Default
Threshold setup option.

Options on page 457


Command buttons on page 460

Options
The following thresholds apply to all monitored volumes.

458 | OnCommand Console Help


Full Threshold (%)

Displays the percentage at which a volume is considered full. If this limit is


exceeded, the DataFabric Manager server generates a Volume Full event.
The default is 90%.

Nearly Full
Threshold (%)

Displays the percentage at which a volume is considered nearly full. If this


limit is exceeded, the DataFabric Manager server generates a Volume
Nearly Full event.
You should specify a limit that is less than the value specified for the
Volume Full Threshold option.
The default is 80%.

Full Threshold
Interval

Displays the maximum duration of a threshold breach before a full


threshold event is generated.
If the condition persists for the specified amount of time, the DataFabric
Manager server generates a Volume Full event. Threshold intervals apply
only to error and informational events.
If the threshold interval is 0 seconds or a value less than the volume
monitoring interval, the DataFabric Manager server generates Volume Full
events as they occur. If the threshold interval is greater than the volume
monitoring interval, the DataFabric Manager server waits for the specified
threshold interval, which includes two or more monitoring intervals, and
generates a Volume Full event only if the condition persisted throughout
the threshold interval. For instance, if the monitoring cycle time is 60
seconds and the threshold interval is 90 seconds, the threshold event is
generated only if the condition persists for two monitoring cycles.
The default is 0 seconds.

Quota
Overcommitted
Threshold (%)

Displays the percentage of allocated space (disk space or files used) on a


quota, as specified by the users quota file that can be used before the
system generates an Quota Overcommitted event.
The default is 50%.

Quota Nearly
Overcommitted
Threshold (%):

Displays the percentage of allocated space (disk space or files used) on a


quota, as specified by the users quota file that can be used before the
system generates an Quota Nearly Overcommitted event.
The default is 40%.

Growth Event
Minimum Change
(%)

Displays the minimum change in volume size (as a percentage of total


volume size) that is considered normal. If the change in volume size is more
than the specified value, the DataFabric Manager server generates a
Volume Growth Abnormal event.
The default is 1%.

Administration | 459
Snap Reserve Full
Threshold (%)

Displays the value at which the space reserved for making volume Snapshot
copies is considered full.
The default is 90%.

No First Snapshot
Threshold (%)

Displays the value at which a volume is considered to have consumed all


the free space that it needs when the first Snapshot copy is created.
This option applies to volumes that contain space-reserved files, no
Snapshot copies, and a fractional overwrite reserve set to greater than 0, and
for which the sum of the space reservations for all LUNs in the volume is
greater than the free space available to the volume.
The default is 90%.

Nearly No First
Displays the value at which a volume is considered to have consumed most
Snapshot Threshold of the free space that it needs when the first Snapshot copy is created.
(%)
This option applies to volumes that contain space-reserved files, no
Snapshot copies, and a fractional overwrite reserve set to greater than 0, and
for which the sum of the space reservations for all LUNs in the volume is
greater than the free space available to the volume.
You should specify a limit that is less than the value specified for the
Volume No First Snapshot Threshold option.
The default is 80%.
Space Reserve
Depleted Threshold
(%)

Displays the value at which a volume is considered to have consumed all its
reserved space.
This option applies to volumes with LUNs, Snapshot copies, no free space,
and a fractional overwrite reserve of less than 100%. A volume that has
crossed this threshold is getting dangerously close to having write failures.
The default is 90%.

Space Reserve
Nearly Depleted
Threshold (%)

Displays the value at which a volume is considered to have consumed most


of its reserved space.
This option applies to volumes with LUNs, Snapshot copies, no free space,
and a fractional overwrite reserve of less than 100%. A volume that crosses
this threshold is getting close to having write failures.
You should specify a limit that is less than the value specified for the
Volume Space Reserve Depleted Threshold option.
The default is 80%.

Snapshot Count
Threshold

Displays the limit to the number of Snapshot copies allowed on the volume.
A volume is allowed up to 255 Snapshot copies.

460 | OnCommand Console Help

The default is 250 Snapshot copies.


Too Old Snapshot
Threshold

Specifies the limit to the age of a Snapshot copy allowed for the volume.
The Snapshot copy age can be specified in seconds, minutes, hours, days, or
weeks.
The default is 52 weeks.

Nearly overDeduplicated
threshold (%)

Displays the percentage of user data that can be deduplicated and stored on
a volume before the system generates a Nearly Over-Deduplicated event.

Over Deduplicated
threshold (%)

Displays the percentage of user data that can be deduplicated and stored on
a volume before the system generates an Over Deduplicated event.

The default is 140%.

The default is 150%.


Command buttons
The command buttons enable you to save or cancel the setup options.
Save and Close Saves all the configuration settings and then closes the Setup Options dialog box.
Cancel

Enables you to undo all the configuration settings and then closes the Setup
Options dialog box.

Save

Saves all your configuration settings.

Default Thresholds Other area


You can change the global default threshold values for host agents, HBA ports, qtrees, user quotas,
and resource pools using Default threshold options.

Options on page 460


Command buttons on page 462

Options
The following information is available under Other thresholds:
HBA Port Too Displays the percentage of maximum traffic the HBA port can handle without
Busy Threshold adversely affecting performance. If this threshold is crossed, the DataFabric
Manager server generates an HBA Port Traffic High event.
(%)
The default is 90%.
Host CPU Too Displays the percentage of maximum traffic the host CPU can handle without
Busy Threshold adversely affecting performance.
(%)
The default is 95%.

Administration | 461
Host CPU Busy Specifies the maximum duration of a threshold breach that can persist before the
CPU busy event is generated.
Threshold
Interval
If the condition persists for the specified amount of time, the DataFabric
Manager server generates a CPU-too-busy event. Threshold intervals apply only
to error and informational events.

If the threshold interval is 0 seconds or a value less than the CPU monitoring
interval, the DataFabric Manager server generates CPU-too-busy events as
they occur.
If the threshold interval is greater than the CPU monitoring interval, the
DataFabric Manager server waits for the specified threshold interval, which
includes two or more monitoring intervals, and generates a CPU-too-busy
event only if the condition persisted throughout the threshold interval.

For example, if the monitoring cycle time is 60 seconds and the threshold
interval is 90 seconds, the event is generated only if the condition persists for
two monitoring cycles.
The default is 15 minutes.
Qtree Full
Threshold (%)

Displays the percentage at which a qtree is considered full. If this limit is


exceeded, the DataFabric Manager server generates a Qtree Full event.
The default is 90%.

Qtree Nearly
Full Threshold
(%)

Displays the percentage at which a qtree is considered nearly full. If this limit is
exceeded, the DataFabric Manager server generates a Qtree Nearly Full event.
You should to specify a limit that is less than the value specified for the Qtree
Full Threshold option
The default is 80%.

Qtree Full
Threshold
Interval

Displays the maximum duration of a threshold breach before the Qtree full
threshold event is generated.
If the condition persists for a specified amount of time, the DataFabric Manager
server generates a Qtree Full event. Threshold intervals apply only to error and
informational events.
If the threshold interval is 0 seconds or a value less than the qtree monitoring
interval, the DataFabric Manager server continuously generates Qtree Full
events until an event is resolved.. If the threshold interval is greater than the
qtree monitoring interval, the DataFabric Manager server waits for the specified
threshold interval, which includes two or more monitoring intervals, and
generates a Qtree Full event only if the condition persisted throughout the
threshold interval. For instance, if the monitoring cycle time is 60 seconds and
the threshold interval is 90 seconds, the threshold event is generated only if the
condition persists for two monitoring cycles.

462 | OnCommand Console Help

The default is 0 seconds.


Qtree Growth
Event
Minimum
Change (%)

Displays the minimum change in qtree size (as a percentage of total volume
size). If the change in qtree size is more than the specified value, and the growth
is abnormal with respect to the qtree-growth history, the DataFabric Manager
server generates a Qtree Growth Abnormal event.
The default is 1%.

User Quota Full Displays the value at which a user is considered to have consumed all the
Threshold (%) allocated space (disk space or files used) as specified by the user's quota (hard
limit in the /etc/quotas file).
If this limit is exceeded, the DataFabric Manager server generates a User Disk
Space Quota Full or User Files Quota Full event.
The default is 90%.
User Quota
Nearly Full
Threshold (%)

Displays the value at which a user is considered to have consumed most of the
allocated space (disk space or files used) as specified by the user's quota (hard
limit in the /etc/quotas file).
If this limit is exceeded, the DataFabric Manager server generates a User Disk
Space Quota Almost Full or User Files Quota Almost Full event.
You should to specify a limit that is less than the value specified for the User
Quota Full Threshold option.
The default is 80%.

Resource Pool
Full Threshold
(%)

Displays the percentage of space used in a resource pool before DataFabric


Manager server generates a Resource Pool Full event.

Resource Pool
Nearly Full
Threshold (%)

Displays the percentage of space used in a resource pool before DataFabric


Manager server generates a Resource Pool Nearly Full event.

The default is 90%.

The default is 80%.

Command buttons
The command buttons enable you to save or cancel the setup options.
Save and Close

Saves the configuration settings for the selected option.

Cancel

Does not save the recent changes and closes the Setup Options dialog box.

Save

Saves the configuration settings for the selected option.

Discovery setup options

Administration | 463

Understanding discovery options


What the discovery process is
The DataFabric Manager server discovers all the storage systems in your organization's network by
default. You can add other networks to the discovery process or enable discovery on all the networks.
Depending on your network setup, you can disable discovery entirely. You can disable autodiscovery if you do not want SNMP network walking.
SNMP version setup
You must know of the settings that are used for the SNMP version preferred at the storage system
level or network level.
If the SNMP version is...

Then...

Specified at the storage system level

The version preferred takes precedence over the


network and global settings.

Not specified at the storage system level

Network setting is used.

Not specified at the network level

Global setting is used.

When the DataFabric Manager server is installed for the first time or updated, by default, the global
and network setting uses SNMPv1 as the preferred version. However, you can configure the global
and network setting to use SNMPv3 as the default version.
Guidelines for editing discovery options
You must follow a set of guidelines for changing the default values of the discovery options.
Interval

This option specifies the period after which the DataFabric Manager server scans
for new storage systems and networks.
You can change the default value if you want to increase the minimum time
interval between system discovery attempts. This option affects the discovery
interval only at the time of installation. After storage systems are discovered, the
user should determine the interval based on the number of networks and their size.
If you choose a longer interval, there might be a delay in discovering new storage
systems, but the discovery process is less likely to affect the network load.
The default is 15 minutes.

Timeout

This option specifies the time interval after which the DataFabric Manager server
considers a discovery query to have failed.
You can change the default value if you want to lengthen the time before
considering a discovery to have failed (to avoid discovery queries on a local area
network failing due to the long response times of a storage system).

464 | OnCommand Console Help

The default is 5 seconds.


Host
discovery

This option enables the discovery of storage systems, host agents and vFiler Units
through SNMP.
You can change the default value if any of the following situations exist:

All storage systems that you expected the DataFabric Manager server to
discover have been discovered and you do not want the DataFabric Manager
server to continue scanning for new storage systems.
You want to manually add storage systems to the DataFabric Manager server
database.
Manually adding storage systems is faster than discovering storage systems in
the following cases:

You want the DataFabric Manager server to manage a small number of


storage systems.
You want to add a single new storage system to the DataFabric Manager
server database.

Host agent
discovery

This option allows you to enable or disable discovery of host agents.

Network
discovery

This option enables the discovery of networks, including SAN and cluster
networks.

You can change the default value if you want to disable the discovery of LUNs or
storage area network (SAN) hosts and host agents.

You can change the default value if you want the DataFabric Manager server to
automatically discover storage systems on your entire network.
Note: When the Network Discovery option is enabled, the list of networks on
the Networks to Discover page can expand considerably as the DataFabric
Manager server discovers additional networks attached to previously discovered
networks.

Network
Discovery
Limit (in
hops)

This option sets the boundary of network discovery as a maximum number of hops
(networks) from the DataFabric Manager server.
You can change the default value if you want to increase this limit if the storage
systems that you want the DataFabric Manager server to discover are connected to
networks that are more than 15 hops (networks) away from the network to which
the DataFabric Manager server is attached. The other method for discovering these
storage systems is to add them manually.
You can decrease the discovery limit if a smaller number of hops includes all the
networks with storage systems you want to discover. For example, reduce the limit
to six hops if there are no storage systems that must be discovered on networks
beyond six hops. Reducing the limit prevents the DataFabric Manager server from

Administration | 465

using cycles to probe networks that contain no storage systems that you want to
discover.
The default is 15 hops.
Networks to
discover

This option enables you to manually add or delete networks that the DataFabric
Manager server scans for new storage systems.
You can change the default value if you want to add a network to the DataFabric
Manager server that it cannot discover automatically, or you want to delete a
network for which you no longer want storage systems to be discovered.

Network
Credentials

This option enables you to specify, change, or delete an SNMP community that the
DataFabric Manager server uses for a specific network or host.
You can change the default value if storage systems and routers that you want to
include in the DataFabric Manager server do not use the default SNMP
community.

Configuring discovery options


Adding addresses for discovery
The DataFabric Manager server uses the automatic discovery process to find storage systems on your
organizations network. You can add the addresses of your network or storage systems for discovery
from the Setup Options dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click Setup Options.


2. In the Setup Options dialog box, click the Discovery option, then click the Addresses option.
3. In the Discovery Addresses area, click Add.
4. In the Add Network Address dialog box, specify your network address and the prefix length.
5. Click OK.
6. Click Save and Close.
Related references

Administrator roles and capabilities on page 506

466 | OnCommand Console Help

Managing discovery options


Editing SNMP communities and network credentials
You can change the default network credentials and SNMP settings if storage systems and routers
that you want to include in the DataFabric Manager server , do not use the default SNMP
community.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click the Setup Options option.
2. In the Setup Options dialog box, click the Discovery option.
3. Click the Credentials option.
4. In the Discovery Credentials area, select the network address you want to edit.
5. Click Edit.
6. In the Edit Network Credentials dialog box, under Preferred SNMP version, you can select one
of the following options:

SNMP v1; if you select SNMP v1, you can configure the SNMP communities.
SNMP v3; if you select SNMP v3, you can configure the following:

Auth protocol: You can choose either of the two protocol options: MD5 and SHA. The
default option is MD5.
Login: Type your login information.
Password: Type your password.
Privacy password: Type your privacy password.

7. Click OK.
8. Click Save and Close.
Related references

Administrator roles and capabilities on page 506


Guidelines for editing discovery options on page 463
Discovery Credentials area on page 469
Page descriptions

Administration | 467

Discovery Options area


You can use Discovery options to discover networks, storage systems (including clusters), Open
Systems SnapVault agents, and host agents.

Options on page 467


Command buttons on page 468

Options
Displays the settings you can configure to discover storage objects and the time taken for discovery.
Host discovery

Enables (default) or disables the discovery of hosts.

Host agent
discovery

Enables (default) or disables the discovery of hosts running the NetApp Host
Agent software.

vFiler Unit
discovery

Enables (default) or disables discovery of vFiler units. When disabled, the


DataFabric Manager server does not discover new vFiler units but continues
to monitor existing vFiler units.

Host-initiated
discovery

Enables (default) or disables host-initiated discovery. When enabled, the


DataFabric Manager server accepts communication requests initiated by host
agents.
Note: Currently, host-initiated discovery is supported only in NetApp Host

Agent.
Other Discovery
Methods

Displays the methods you can use to discover storage systems and
networks.

SAN discovery

Enables(default) or disables the discovery of systems in a SAN


environment.

Cluster discovery

Enables (default) or disables the discovery of cluster systems.

Network discovery

Enables or disables (default) the discovery of new networks.

Default Discovery
Options (except
vFiler Units)

Displays default settings that are used by the DataFabric Manager server to
start or end the discovery process.

Interval

Displays the minimum time interval during which the DataFabric Manager
server scans for new storage systems and networks.
The default is 15 minutes.

Timeout

Displays the time interval after which the DataFabric Manager server
considers a discovery query to have failed.
The default is 5 seconds.

468 | OnCommand Console Help


Network Discovery
Limit (in hops)

Sets the boundary of network discovery as a maximum number of hops


(networks) from the DataFabric Manager server.

Command buttons
The command buttons enable you to save or cancel the setup options.
Save and Close

Saves the configuration settings for the selected option.

Cancel

Does not save the recent changes and closes the Setup Options dialog box.

Save

Saves the configuration settings for the selected option.

Discovery Addresses area


The Discovery setup options enable the DataFabric Manager server to discover storage systems and
host agents from the network.

Options on page 468


Command buttons on page 468

Options
Displays the properties of the discovered network address in a tabular format. You can sort data
based on the filters applied to the columns. Multiple filters and single column sorting can be applied
at the same time.
Address

Displays the network address of the object which will be discovered.

Prefix Length

Displays the prefix length of the network address.

Hop Count

Displays the hop count of the network address.

Last Searched

Displays the date and timestamp of the last searched network address.

Command buttons
The command buttons enable you to save or cancel the setup options, and add, edit, or delete network
addresses.
Add

Enables you to add a network address and its prefix length.

Edit

Enables you to edit the specified network address.

Delete

Enables you to delete the specified network address.


You can choose to delete Network Only or Network and Hosts from the dropdown list.

Save and Close Saves all the configuration settings and then closes the Setup Options dialog box.

Administration | 469
Cancel

Enables you to undo all the configuration settings and then closes the Setup
Options dialog box.

Save

Saves all your configuration settings.

Discovery Credentials area


You can discover storage objects from their network credentials by using the Discovery credentials
option.

Options on page 469


Command buttons on page 469

Options
Displays, in tabular format, the properties of the discovered network credential. You can sort data
based on the filters applied to the columns. Multiple filters and single column sorting can be applied
at the same time.
Address

Displays the network address.

Prefix Length

Displays the prefix length of the network address.

Preferred SNMP Version

Displays the preferred SNMP version.

Command buttons
The command buttons enable you to save or cancel the setup options, and add, delete, or edit network
credentials.
Add

Enables you to add a network and configure the SNMP settings.

Edit

Enables you to edit the selected network and configure the SNMP settings.

Delete

Enables you to delete the selected network.

Save and Close

Saves the configuration settings for the selected option.

Cancel

Does not save the recent changes and closes the Setup Options dialog box.

Save

Saves the configuration settings for the selected option.

File SRM setup options


Understanding File SRM options

470 | OnCommand Console Help

What File Storage Resource Management does


File Storage Resource Manager (FSRM) enables you to gather file-system metadata and generate
reports on different characteristics of that metadata.
The DataFabric Manager server interacts with the NetApp Host Agent residing on remote Windows,
Solaris, or Linux workstations or servers (called hosts) to recursively examine the directory
structures (paths) you have specified.
For example, if you suspect that certain file types are consuming excessive storage space on your
storage systems, you can perform the following tasks:
1. Deploy one or more host agents.
2. Configure FSRM to walk a path.
The host agents might have a NetApp LUN, volume, or qtree mounted. You can configure FSRM to
generate reports periodically. These reports contain the following details:

Files that are consuming the most space


Files that are outdated or have been accessed recently
Types of files (.doc, .gif, .mp3, and so on) on the file system

You can then decide how to most efficiently use your existing space.
How FSRM monitoring works
The DataFabric Manager server monitors directory paths that are visible to the host agent. Therefore,
if you want to enable FSRM monitoring of NetApp storage systems, the remote host must mount a
NetApp share using NFS or CIFS, or the host must use a LUN on the storage system.
Note: The DataFabric Manager server cannot obtain FSRM data for files that are located in

NetApp volumes, which are not exported by CIFS or NFS. Host agents can also gather FSRM data
about other file system paths that are not on a NetApp storage system: for example, local disk or
third-party storage systems.
Configuring File SRM options
Adding new file types for monitoring file-level statistics
DataFabric Manager server collects file-system metadata using File SRM. You can add new file
types from the Setup Options dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.

Administration | 471
Steps

1. Click the Administration menu, then click Setup Options.


2. In the Setup Options dialog box, click the File SRM option, then click Options.
3. In the File SRM Options area, click Add to add a SRM file type.
4. Specify the file type in the Add SRM File Type dialog box, then click OK.
5. Click Save and Close.
Related references

Administrator roles and capabilities on page 506


Deleting File SRM file types
When you do not want DataFabric Manager server to collect file-system metadata for a particular file
type, you can delete those file types from the Setup Options dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click Setup Options.


2. In the Setup Options dialog box, click the File SRM option, then click Options.
3. In the File SRM Options area, select one or more file types you want to delete from SRM File
Types.
4. Click Delete.
5. Click Save and Close.
Related references

Administrator roles and capabilities on page 506


Managing File SRM options

472 | OnCommand Console Help

Editing File SRM options


You can edit the values for the largest files, recently modified files, and recently accessed file-reports
from the Setup Options dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click Setup Options.


2. In the Setup Options dialog box, click the File SRM option, then click Options.
3. In the File SRM Options area, specify the new values, as required.
4. Click Save and Close.
Related references

Administrator roles and capabilities on page 506


Page descriptions
File SRM area
You can use the File SRM setup option to configure the number of largest files, recently modified
files, accessed files, and file types.

Options on page 472


Command buttons on page 473

Options
You can configure the following File SRM setting options:
Largest Files (Max)

Displays the maximum number of largest files for each File SRM
path.

Least Recently Modified


Files (Max)

Displays the maximum number of files that were least modified


recently, for each File SRM path.

Least Recently Accessed


Files (Max)

Displays the maximum number of files that were least accessed


recently, for each File SRM path.

Recently Accessed Files


(Max)

Displays the maximum number of files that were recently accessed


for each File SRM path.

Administration | 473
SRM File Types

Displays the SRM file types to add or delete.

Note: The default value for all the options is 20.

Command buttons
The command buttons enable you to save or cancel the setup options, and add or delete SRM file
types.
Add

Adds the specified SRM file type.

Delete

Deletes the specified SRM file type.

Save and Close

Saves the configuration settings for the selected option.

Cancel

Does not save the recent changes and closes the Setup Options dialog box.

Save

Saves the configuration settings for the selected option.

LDAP setup options


Understanding authentication
Authentication methods on the DataFabric Manager server
The DataFabric Manager server uses the information available in the native operating system for
authentication. The server does not maintain its own database of the administrator names and the
passwords.
You can also configure the DataFabric Manager server to use Lightweight Directory Access Protocol
(LDAP). If you configure LDAP, then the server uses it as the preferred method of authentication.
Authentication with LDAP
You can enable LDAP authentication on the DataFabric Manager server and configure it to
communicate with your LDAP servers in order to retrieve relevant data.
The DataFabric Manager server provides predefined templates for the most common LDAP server
types. These templates provide predefined LDAP settings that make the DataFabric Manager server
compatible with your LDAP server.
Configuring LDAP options

474 | OnCommand Console Help

Adding LDAP servers for authentication


You can enable LDAP authentication on the DataFabric Manager server and configure it to work
with your LDAP servers. You can add the LDAP servers from the Setup Options dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click Setup Options.


2. In the Setup Options dialog box, click the LDAP option, then click Servers.
3. In the LDAP Servers area, click Add.
4. In the Add LDAP Server dialog box, specify the server address or host name, and port details.
5. Click OK.
6. Click Save and Close.
Related references

Administrator roles and capabilities on page 506


Deleting LDAP servers
If you want to disable authentication, you can delete your LDAP servers from the Setup Options
dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click Setup Options.


2. In the Setup Options dialog box, click the LDAP option, then click Servers.
3. In the LDAP Servers area, select one or more LDAP server you want to delete, and then click
Delete.
4. Click Save and Close.

Administration | 475
Related references

Administrator roles and capabilities on page 506


Managing LDAP options
Editing server type options for authentication
When configuring the DataFabric Manager server for LDAP authentication, you can use a template
to automatically select the settings compatible with your LDAP server. You can edit the default
template settings from the Setup Options dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click Setup Options.


2. In the Setup Options dialog box, click the LDAP option, then click Server Types.
3. In the LDAP Server Types area, modify the template settings, as required.
4. Click Save and Close.
Related references

Administrator roles and capabilities on page 506


Page descriptions
LDAP Authentication area
You can use the LDAP Authentication setup option to configure the DataFabric Manager server to
communicate with your LDAP servers.

Options on page 475


Command buttons on page 476

Options
The LDAP authentication setting options are as follows:
LDAP Is Enabled

Enables you to enable LDAP.

LDAP Bind DN

Specifies the bind distinguished name (DN) that DataFabric Manager server
uses to identify itself to the LDAP server.

476 | OnCommand Console Help


LDAP Bind
Password

Specifies the password that DataFabric Manager server uses to gain access
to the bind distinguished name.

LDAP Base DN

Specifies the directory on the LDAP server that DataFabric Manager server
uses to search for the LDAP server.
For example, dc=domain, dc=com specifies whether LDAP is enabled or
disabled. By default, LDAP is disabled.

Command buttons
The command buttons enable you to save or cancel the setup options.
Save and Close

Saves the configuration settings for the selected option.

Cancel

Does not save the recent changes and closes the Setup Options dialog box.

Save

Saves the configuration settings for the selected option.

LDAP Server Types area


You can use the LDAP Server types setup option to configure the DataFabric Manager server for
LDAP authentication and use a template to automatically select the settings compatible with your
LDAP server.

Options on page 476


Command buttons on page 477

Options
The LDAP server type setting options are as follows:
Templates

Specifies the templates you can select to facilitate LDAP configuration.


You can select one of three templates; each file provides different default
information. Templates provide predefined LDAP settings to configure the
DataFabric Manager server with your LDAP server.

Product Line Specifies the template that provides the predefined LDAP settings designed to
make DataFabric Manager server compatible with your LDAP server. You can
select Netscape/iPlanet (default), UMich/OpenLDAP, or Lotus Domino from the
drop-down menu.
Netscape/iPlanet

Selects the access control settings to use with your


Netscape/iPlanet servers.

UMich/OpenLDAP Selects the access control settings to use with your


Umich/OpenLDAP servers.

Administration | 477

Lotus Domino

Selects the access control settings to use with your Lotus


Domino servers.

Custom

Displays LDAP server attributes that you can modify.

Protocol
Version

Specifies the LDAP protocol version. You can either select 2 or 3 (default) from
the drop-down menu.
The default is 3.

UID
Attribute

Specifies the name of the attribute in the LDAP directory that contains user login
names to be authenticated by the DataFabric Manager server.
The default is UID.

GID
Attribute

Specifies a value that assigns the DataFabric Manager server group membership to
LDAP users based on an attribute and value specified in their LDAP user objects.

UGID
Attribute

If the LDAP users are included as members of a GroupOfUniqueNames object in


the LDAP directory, this option enables you to assign the DataFabric Manager
server group membership to them based on a specified attribute in that
GroupOfUniqueNames object.
The default is CN.

Member
Attribute

Specifies the attribute name that your LDAP server uses to store information about
the individual members of a group.
The default is uniqueMember.

Command buttons
The command buttons enable you to save or cancel the setup options.
Save and Close

Saves the configuration settings for the selected option.

Cancel

Does not save the recent changes and closes the Setup Options dialog box.

Save

Saves the configuration settings for the selected option.

LDAP Servers area


You can use the LDAP Servers setup option to identify the LDAP server that DataFabric Manager
server queries for authentication information.

Options on page 477


Command buttons on page 478

Options
The LDAP server setting options are as follows:

478 | OnCommand Console Help


Address or
Hostname

Displays the IP address or host name of the LDAP server that is used to
authenticate the user on the DataFabric Manager server.

Port

Displays the port number of the LDAP server.

Last Used

Displays the date and timestamp of the most recent authentication success.

Last Failed

Displays the date and timestamp of the most recent permission failure.

Command buttons
The command buttons enable you to add, delete, save or cancel the setup options.
Add

Adds a new LDAP server.

Delete

Deletes the selected LDAP server.

Save and Close

Saves the configuration settings for the selected option.

Cancel

Does not save the recent changes and closes the Setup Options dialog box.

Save

Saves the configuration settings for the selected option.

Monitoring setup options


Understanding monitoring options
The DataFabric Manager server monitoring process
The DataFabric Manager server discovers the storage systems supported on your network. The
DataFabric Manager server periodically monitors data that it collects from the discovered storage
systems, such as CPU usage, interface statistics, free disk space, qtree usage, and chassis
environmental. The DataFabric Manager server generates events when it discovers a storage system,
when the status is abnormal, or when a predefined threshold is breached. If configured to do so, the
DataFabric Manager server sends a notification to a recipient when an event triggers an alarm.
The following flow chart illustrates the DataFabric Manager server monitoring process.

Administration | 479

Start

DataFabric Manager server


discovers a storage system

DataFabric Manager server


periodically collects data from
the storage system

No

Abnormal
status received
from the storage
system?

Ye

No

Generate an event

Alarm
configured for
the event?

Ye
Generate the alarm
No

No

Repeat
notification for
the alarm
configured?

Ye

Alarm
acknowledged?

Ye

480 | OnCommand Console Help

SNMP queries
The DataFabric Manager server uses periodic SNMP queries to collect data from the storage systems
it discovers. The data is reported by the DataFabric Manager server in the form of tabular and
graphical reports and event generation.
The time interval at which an SNMP query is sent depends on the data being collected. For example,
although the DataFabric Manager server pings each storage system every minute to ensure that the
storage system is reachable, the amount of free space on the disks of a storage system is collected
every 30 minutes.
Guidelines for changing monitoring intervals
Although you should generally keep the default values, you might need to change some of the
options to suit your environment. All the monitoring option values apply to all storage systems in all
groups.
If you decrease the monitoring intervals, you receive more real-time data. However, the DataFabric
Manager server queries the storage systems more frequently, thereby increasing the network traffic
and the load on the DataFabric Manager server and the storage systems responding to the queries.
If you increase the monitoring interval, the network traffic and the storage system load are reduced.
However, the reported data might not reflect the current status or condition of a storage system.
Managing monitoring options
Editing storage options for monitoring
Although the DataFabric Manager server is configured with defaults that enable you to manage the
global default threshold values immediately, you might need to change some of the storage options to
suit your environment. You can change the storage options from the Setup Options dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click Setup Options.


2. In the Setup Options dialog box, click the Monitoring option, then click the Storage option.
3. In the Monitoring Storage area, specify the new values, as required.
4. Click Save and Close.

Administration | 481
Related concepts

Guidelines for changing monitoring intervals on page 480


Related references

Administrator roles and capabilities on page 506


Editing protection options for monitoring
Although the DataFabric Manager server is configured with defaults that enable you to manage the
global default threshold values immediately, you might need to change some of the protection
options to suit your environment. You can change the protection options from the Setup Options
dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click Setup Options.


2. In the Setup Options dialog box, click the Monitoring option, then click the Protection option.
3. In the Monitoring Protection area, specify the new values, as required.
4. Click Save and Close.
Related concepts

Guidelines for changing monitoring intervals on page 480


Related references

Administrator roles and capabilities on page 506


Editing network options for monitoring
Although the DataFabric Manager server is configured with defaults that enable you to manage the
global default threshold values immediately, you might need to change some of the network options
to suit your environment. You can change the network options from the Setup Options dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.

482 | OnCommand Console Help


Steps

1. Click the Administration menu, then click Setup Options.


2. In the Setup Options dialog box, click the Monitoring option, then click the Networking option.
3. In the Monitoring Networking area, specify the new values, as required.
4. Click Save and Close.
Related concepts

Guidelines for changing monitoring intervals on page 480


Related references

Administrator roles and capabilities on page 506


Editing inventory options for monitoring
Although the DataFabric Manager server is configured with defaults that enable you to manage the
global default threshold values immediately, you might need to change some of the inventory options
to suit your environment. You can change the inventory options from the Setup Options dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click Setup Options.


2. In the Setup Options dialog box, click the Monitoring option, then click the Inventory option.
3. In the Monitoring Inventory area, specify the new values, as required.
4. Click Save and Close.
Related concepts

Guidelines for changing monitoring intervals on page 480


Related references

Administrator roles and capabilities on page 506

Administration | 483

Editing system options for monitoring


Although the DataFabric Manager server is configured with defaults that enable you to manage the
global default threshold values immediately, you might need to change some of the system options to
suit your environment. You can change the system options from the Setup Options dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click Setup Options.


2. In the Setup Options dialog box, click the Monitoring option, then click the System option.
3. In the Monitoring System area, specify the new values, as required.
4. Click Save and Close.
Related concepts

Guidelines for changing monitoring intervals on page 480


Related references

Administrator roles and capabilities on page 506


Editing ping intervals
You can perform a ping operation to ensure that a storage object is available.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click the Setup Options option.
2. In the Setup Options dialog box, click the Monitoring option.
3. Click the Networking option.
4. In the Monitoring Networking area, specify the following settings:

The ping monitoring interval


Ping method that the DataFabric Manager server uses to contact a storage object

484 | OnCommand Console Help

Ping timeout interval


Ping retry delay

The ping is declared successful only if the reply is received before the ping timeout interval.
5. Click Save and Close.
Related references

Administrator roles and capabilities on page 506


Monitoring Networking area on page 486
Page descriptions
Monitoring Storage area
You can use the Setup Options dialog box to configure and customize global options for storage
objects.

Options on page 484


Command buttons on page 485

Options
You can use the monitoring options to configure monitoring intervals for various storage objects that
are monitored by DataFabric Manager server.
Cluster
Interval

Specifies the time at which the DataFabric Manager server gathers status
information from each cluster.
The default is 15 minutes.

Cluster
Failover
Interval

Specifies the time at which the DataFabric Manager server gathers highavailability configuration status information from each controller.

vFiler Unit
Interval

Specifies the time at which the DataFabric Manager server gathers information
about vFiler units that are configured or destroyed on hosting storage systems.

The default is 5 minutes.

The default is 1 hour.


Vserver
Interval

Specifies the time at which the DataFabric Manager server gathers information
about virtual servers on hosting cluster systems.
The default is 1 hour.

User Quota
Interval

Specifies the time at which the DataFabric Manager server collects the user quota
information from the monitored storage systems.
The default is 1 day.

Administration | 485

Note: The process of collecting the user quota information from storage systems

is resource-intensive for storage systems. Decreasing the User Quota


Monitoring Interval option to a low value to increase the frequency of collecting
the information might affect the performance of the monitored storage systems.
Qtree Interval Specifies the time at which the DataFabric Manager server gathers statistics about
monitored qtrees.
The default is 8 hours.
LUN Interval Specifies the time at which the DataFabric Manager server gathers information
about LUNs on each storage system.
The default is 30 minutes.
Command buttons
The command buttons enable you to save or cancel the setup options.
Save and Close Saves all the configuration settings and then closes the Setup Options dialog box.
Cancel

Enables you to undo all the configuration settings and then closes the Setup
Options dialog box.

Save

Saves all your configuration settings.

Monitoring Protection area


You can use the Setup Options dialog box to configure and customize global options for the
protection policies of your storage objects. You can configure the dataset details, resource pool
space, and the time at which the DataFabric Manager server gathers SnapMirror, Snapshot, and
SnapVault information from the storage system.

Options on page 485


Command buttons on page 486

Options
You can use the monitoring options to configure monitoring intervals for various protection objects
that are monitored by DataFabric Manager server.
Dataset
Conformance

Specifies the time at which the DataFabric Manager server checks whether
each dataset conforms to its protection policy.
The default is 1 hour.

Dataset Disaster
Recovery Status

Specifies the time at which the DataFabric Manager server checks the
disaster recovery status of each dataset.
The default is 15 minutes.

486 | OnCommand Console Help


Dataset Protection
Status

Specifies the time at which the DataFabric Manager server checks the
protection status of each dataset.
The default is 15 minutes.

Resource Pool Space Specifies the time at which the DataFabric Manager server checks the space
usage in each resource pool.
The default is 1 hour.
SnapMirror

Specifies the time at which the DataFabric Manager server gathers


SnapMirror information from each storage system.
The default is 30 minutes.

SnapShot

Specifies the time at which the DataFabric Manager server gathers Snapshot
information from each storage system.
The default is 30 minutes.

SnapVault

Specifies the time at which the DataFabric Manager server gathers


SnapVault information from each storage system.
The default is 30 minutes.

Command buttons
The command buttons enable you to save or cancel the setup options.
Save and Close Saves all the configuration settings and then closes the Setup Options dialog box.
Cancel

Enables you to undo all the configuration settings and then closes the Setup
Options dialog box.

Save

Saves all your configuration settings.

Monitoring Networking area


You can use the Setup Options dialog box to configure and customize global options for networking.
You can configure details about the ping methods, SNMP retries and timeout, SAN host, and Fibre
Channel switch.

Options on page 486


Command buttons on page 488

Options
You can use the monitoring options to configure monitoring intervals for various networking objects
that are monitored by DataFabric Manager server.
Ping

Specifies the time at which the DataFabric Manager server pings a storage system.

Administration | 487

A short ping interval is recommended when you want to quickly detect if a storage
system is available. The minimum ping monitoring interval is 1 second.
The default is 1 minute.
Note: The actual ping interval depends on variables such as networking conditions
and the number of monitored hosts. Therefore, the ping interval can be longer than
the specified value.

Ping
Method

Specifies the ping method that the DataFabric Manager server uses to check that a
storage system is accessible. By default, the ICMP echo and SNMP ping method is
selected.
You can select one of the following options from the drop-down menu:

ICMP echo and SNMP


Enables the DataFabric Manager server to use ICMP ping first, and then SNMP,
to determine if the storage system is running. ICMP echo and SNMP is valid for
all storage systems.
ICMP echo
Enables the DataFabric Manager server to ICMP ping a storage system. ICMP
echo is valid for all storage systems.
HTTP (port 80)
Enables HTTP connect on port 80 of the storage system. HTTP is valid for all
storage systems.
Note: Do not use the HTTP ping method on networks where transparent
caching is enabled, to avoid a transparent redirect when a storage system is
down. In this scenario, the DataFabric Manager server mistakes the storage
system to be running.

SNMP (port 161)


Enables SNMP service to listen to the storage system on port 161. SNMP is valid
for all storage systems.
NDMP (port 10000)
Enables NDMP service to listen to the storage system on port 10000.
All methods

The default is ICMP echo and SNMP.


Ping
timeout

Specifies the time after which a storage system is considered to be not responsive if
the DataFabric Manager server does not receive a reply from the storage system to a
ping request.
The default is 3 seconds.

Ping Retry Specifies the time period that the ping utility remains inactive before retrying an
unresponsive host.
Delay

488 | OnCommand Console Help

The default is 3 seconds.


You might want to select a different value if you are experiencing network difficulties
that cause false Host Down events.
SNMP
Retries

When an SNMP timeout occurs, this option specifies the time between SNMP
connection attempts.
The default is 4 seconds.
When an SNMP timeout occurs, the SNMP monitor attempts to reconnect to the
device for the number of times specified in the SNMP retries option. If the number of
retries is exceeded, the DataFabric Manager server generates a Host SNMP Not
Responding event.

SNMP
Timeout

Specifies the time that can elapse before an SNMP timeout occurs.
The default is 5 seconds.
When an SNMP timeout occurs, the SNMP monitor attempts to reconnect to the
device for the number of times specified in the SNMP retries option. If the number of
retries is exceeded, the DataFabric Manager server generates a Host SNMP Not
Responding event.

Fibre
Channel

Specifies the time at which the DataFabric Manager server gathers status information
from each Fibre Channel switch.
The default is 5 minutes.

SAN Host

Specifies the time at which the DataFabric Manager server gathers information from
each SAN host.
The default is 5 minutes.

Command buttons
The command buttons enable you to save or cancel the setup options.
Save and Close Saves all the configuration settings and then closes the Setup Options dialog box.
Cancel

Enables you to undo all the configuration settings and then closes the Setup
Options dialog box.

Save

Saves all your configuration settings.

Monitoring Inventory area


You can use the Setup Options dialog box to configure and customize global options for storage
inventory, such as the time at which the DataFabric Manager server gathers information about CPU

Administration | 489
usage, disk space and status, environmental status, file system, and global status from the storage
system.

Options on page 489


Command buttons on page 489

Options
You can use the monitoring options to configure monitoring intervals for various inventory objects
that are monitored by DataFabric Manager server.
CPU

Specifies the time at which the DataFabric Manager server gathers CPU usage
information from each storage object.
The default is 5 minutes.

Disk Free Space Specifies the time at which the DataFabric Manager server gathers available disk
space information from each storage object.
In addition to monitoring free disk space on storage systems, the DataFabric
Manager server also monitors disk space on the workstation. If the workstation is
running low on disk space, monitoring is turned off.
The default is 30 minutes.
Disk

Specifies the time at which the DataFabric Manager server gathers disk status
information, such as disks that are not functioning or spare disk count.
The default is 4 hours.

Environmental

Specifies the time at which the DataFabric Manager server gathers


environmental status information from each storage object.
The default is 5 minutes.

File System

Specifies the time at which the DataFabric Manager server gathers file system
information from each storage object.
The default is 15 minutes.

Global Status

Specifies the time at which the DataFabric Manager server gathers global status
information from each storage object.
The default is 10 minutes.

Interface

Specifies the time at which the DataFabric Manager server gathers network
interface information from each storage object.
The default is 15 minutes.

Command buttons
The command buttons enable you to save or cancel the setup options.

490 | OnCommand Console Help


Save and Close Saves all the configuration settings and then closes the Setup Options dialog box.
Cancel

Enables you to undo all the configuration settings and then closes the Setup
Options dialog box.

Save

Saves all your configuration settings.

Monitoring System area


You can use the Setup Options dialog box to configure and customize global options for systems,
such as the details for host agent discovery, configuration conformance of storage systems, RBAC
monitoring, gathering license status, gathering information about the number of operations, and
gathering storage system information.

Options on page 490


Command buttons on page 491

Options
You can use the monitoring options to configure monitoring intervals for various system objects that
are monitored by DataFabric Manager server.
Agent

Specifies the time at which the DataFabric Manager server gathers status
information from each host agent.
The default is 2 minutes.

Config
Conformance

Specifies the time at which the DataFabric Manager server verifies that the
configuration on the storage system conforms with the configuration
provided by the DataFabric Manager server.
The default is 4 hours.

Host RBAC

Specifies the time interval at which the host RBAC Monitor should run.
The default is 1 day.

License

Specifies the time at which the DataFabric Manager server gathers license
status information from each appliance.
The default is 4 hours.

Operation Count

Specifies the time at which the DataFabric Manager server gathers


information about the number of operations that have taken place.
The default is 10 minutes.

SRM Host

Specifies the time at which the DataFabric Manager server gathers


information from each SRM host.
The default is 10 minutes.

Administration | 491
System
Information

Specifies the time at which the DataFabric Manager server gathers system
information from each system.
The default is 1 hour.

Command buttons
The command buttons enable you to save or cancel the setup options.
Save and Close Saves all the configuration settings and then closes the Setup Options dialog box.
Cancel

Enables you to undo all the configuration settings and then closes the Setup
Options dialog box.

Save

Saves all your configuration settings.

Management setup options


Understanding management options
Managed host options
You can configure the managed host options to ensure secure communication between the
DataFabric Manager server and storage systems.
You can select conventional (HTTP) or secure (HTTPS) administration transport for API
communication and conventional (RSH) or secure (SSH) login protocol for login connection. You
can use RSH/SSH connection for executing commands on a storage system. You can use SSH
connection for executing commands on a remote LAN module (RLM) card.
You can set managed host options globally (for all storage systems) or individually (for specific
storage systems). If you set storage system-specific options, the DataFabric Manager server retains
information about the security settings for each managed storage system. It references this
information when choosing one of the following options to connect to the storage system:

HTTP or HTTPS
RSH or SSH
Administration Port
hosts.equiv authentication

Guidelines for changing managed host options


You can change managed host options, such as the login protocol, transport protocol, port, and
hosts.equiv option.
Login Protocol

This option enables you to set login protocols (RSH or SSH) that the
DataFabric Manager server uses when connecting to the managed hosts.

492 | OnCommand Console Help

Login connections.
Active/active configuration operations
The dfm run command for running commands on the storage system

Change the default value if you want a secure connection for active/active
configuration operations, running commands on the storage system.
Administration
Transport

This options enables you to select a conventional (HTTP) or secure (HTTPS)


connection to monitor and manage storage systems through APIs (XML).
Change the default value if you want a secure connection for monitoring and
management.

Administration
Port

This options enables you to configure the administration port which, along
with administration transport, monitors and manages storage systems.
If you do not configure the port option at the storage system-level, the default
value for the corresponding protocol is used.

hosts.equiv option This option enables users to authenticate storage systems when the user name
and password are not provided.
You must change the default value if you have selected the global default
option and if you do not want to set authentication for a specific storage
system.
Note: If you do not set the transport and port options for a storage system, then the DataFabric
Manager server uses SNMP to get storage system-specific transport and port options for
communication. If SNMP fails, then the DataFabric Manager server uses the options set at the
global level.

Managing management options


Editing managed host options
You can modify the managed host options to control the connection between DataFabric Manager
server and storage systems from the Setup Options dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click Setup Options.


2. In the Setup Options dialog box, click the Management option, then click the Managed Host
option.

Administration | 493
3. In the Management Managed Host area, specify the new values, as required.
4. Click Save and Close.
Related concepts

Guidelines for changing managed host options on page 491


Related references

Administrator roles and capabilities on page 506


Editing Host Agent options
The DataFabric Manager server stores authentication credentials globally for all host agents, based
on the Host Agent Options settings. You can change a login name or password for an individual host
agent from the Setup Options dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click Setup Options.


2. In the Setup Options dialog box, click the Management option, then click Host Agent option.
3. In the Management Host Agent area, specify the new values, as required.
4. Click Save and Close.
Related references

Administrator roles and capabilities on page 506


Page descriptions
Management Client area
You can use the Setup Options dialog box for clients to configure a secure connection between the
DataFabric Manager server and your storage systems and your browsers. You can enable HTTP or
HTTPS for establishing the secure connection.

Options on page 494


Command buttons on page 494

494 | OnCommand Console Help


Options
You can configure the following options:
Enable
HTTP

Enables you to connect the clients to the DataFabric Manager server. By default,
HTTP is enabled with the default port 8080.

Enable
HTTPS

Enables you to connect the clients to the DataFabric Manager server. By default,
HTTPS is enabled with the default port 8443.
Note: To enable HTTPS on the DataFabric Manager server, you must configure
SSL using the dfm ssl server setup command.

Command buttons
The command buttons enable you to save or cancel the setup options.
Save and Close

Saves the configuration settings for the selected option.

Cancel

Does not save the recent changes and closes the Setup Options dialog box.

Save

Saves the configuration settings for the selected option.

Management Managed Host area


You can use the Managed Host option in the Setup Options dialog box to configure connections
between the DataFabric Manager server and the storage systems. You can select conventional
(HTTP) or secure (HTTPS) administration transport protocol, and conventional (RSH) or secure
(SSH) login protocol.

Options on page 494


Command buttons on page 495

Options
You can configure the following options:
Login Protocol

Specifies the login protocol that the DataFabric Manager server must use
when connecting to managed hosts.
The default is Remote Shell (RSH).

Administration
Transport

Specifies the transport protocol that the DataFabric Manager server must
use when connecting to storage systems.
The default is HTTP.

Administration Port Specifies the port that the DataFabric Manager server must use when
connecting to storage systems.
The default is port 80.

Administration | 495
Enable hosts.equiv

Specifies the use of the hosts.equiv option to authenticate the storage


system. By default, the hosts.equiv option is disabled. For more
information about configuring the hosts.equiv option, see the Operations
Manager Administration Guide.

Command buttons
The command buttons enable you to save or cancel the setup options.
Save and Close

Saves the configuration settings for the selected option.

Cancel

Does not save the recent changes and closes the Setup Options dialog box.

Save

Saves the configuration settings for the selected option.

Management Host Agent area


You can use the Setup Options dialog box for host agents to configure login credentials for the
NetApp Host Agent software. The DataFabric Manager server interacts with the NetApp Host Agent
residing on remote Windows, Solaris, or Linux workstations or servers (called hosts) to recursively
examine the directory structures (paths) you have specified.

Options on page 495


Command buttons on page 496

Options
You can configure the following options:
Login

Specifies the administrator access level that the DataFabric Manager server
uses to connect with the host agent.
The default login option is guest.
You can specify the following login options for NetApp Host Agent:

guest
Enables the user to log in to the host agent to monitor its LUNs.
admin
Enables the user to log in to the host agent to monitor and manage its
LUNs and successfully execute file walk on directory structures (paths).

Monitoring
Password

Specifies the password that the NetApp Host Agent uses to authenticate a
"guest" user that has monitoring access privilege on the host agent. This
value is the NetApp Host Agent Software option, Monitoring API Password.

Management
Password

Specifies the password that the NetApp Host Agent uses to authenticate an
"admin" user account that has monitoring and management access privileges

496 | OnCommand Console Help


on the host agent. This value is the NetApp Host Agent Software option,
Management API Password.
Administration
Transport

Specifies the transport protocol that runs on the host agent. You can select
either HTTP or HTTPS from the drop-down menu.
The default is http.

Administration
Port

Specifies the port used for communication between the DataFabric Manager
server and NetApp Host Agent.
The default is port 4092.

CIFS Account

Specifies the CIFS account name. This information is required for host agents
running Windows during file walk on CIFS shares.

CIFS Password

Specifies the Host Agent CIFS password. This information is required for
Windows during file walk on CIFS shares.

Command buttons
The command buttons enable you to save or cancel the setup options.
Save and Close

Saves the configuration settings for the selected option.

Cancel

Does not save the recent changes and closes the Setup Options dialog box.

Save

Saves the configuration settings for the selected option.

Systems setup options


Understanding system options
Overview of system options
You can configure alarm settings, system properties such as owner e-mail, audit log settings,
credential cache settings, storage system configurations, script plug-ins, and table settings by using
the system setup options.
You can configure the following system options:
Alarms

You can configure the following alarm properties:

Annotations

Mail server for sending the e-mail notification


Time interval for deleting events
SNMP trap (if you want to receive SNMP traps from storage systems.)
User quota alert details

You can add, edit, or delete user-defined properties such as the e-mail address
of the owner, owner name, and resource tag information.

Administration | 497
Audit Log

You can use the audit log option to set the global auditLogForever option
to keep the audit log files forever in the default log directory of the DataFabric
Manager server. You can view the specific operation in the audit.log file
and determine who performed certain actions from the CLI.

Credential Cache You can use the credential cache option to specify the Time-To-Live (TTL) for
web responses cached by the DataFabric Manager server.
Storage System
Configuration

You can use the storage system configuration option to manage local
configuration file changes on all storage systems that are discovered and
managed by the DataFabric Manager server.

Script Plugins

You can configure the script plug-ins to specify the search path that the
DataFabric Manager server uses to find script interpreters.

Paged Tables

You can configure the number of rows for display in a table. This is applicable
only to the Operations Manager console
Note: This option is not applicable to OnCommand console list tables. It is
applicable only to the Operations Manager console tables.

Configuring system options


Adding custom annotations for alarm recipients
You can add customized annotations for the DataFabric Manager server that can be associated with
storage systems, SAN hosts, FC switches, volumes, qtrees, LUNs, groups, and quota users. For
example, you can include properties such as the asset number, department code, location name, or
support contact for any storage object.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click Setup Options.


2. In the Setup Options dialog box, click the Systems option, then click the Annotations option.
3. In the Systems Annotations area, click Add.
4. Type the name of your annotation in the Add System Annotation dialog box.
5. Click OK.
6. Click Save and Close.

498 | OnCommand Console Help


Related references

Administrator roles and capabilities on page 506


Deleting custom annotations for alarm recipients
You can delete the customized annotations that you created. You cannot delete system-defined
annotations such as the owner name, owner e-mail, and resource tag.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click Setup Options.


2. In the Setup Options dialog box, click the Systems option, then click the Annotations option.
3. In the Systems Annotations area, select the annotation you want to delete.
4. Click Delete.
5. Click Yes.
6. Click Save and Close.
Related references

Administrator roles and capabilities on page 506


Managing system options
Editing alarm settings
You can modify the alarm settings such as mail server details, event purge intervals, SNMP trap
details, and user quota alerts from the Setup Options dialog box.
Before you begin

You must be authorized to perform all the steps of this task; your RBAC administrator can confirm
your authorization in advance.
Steps

1. Click the Administration menu, then click Setup Options.


2. In the Setup Options dialog box, click the Systems option, then click Alarms option.
3. In the Systems Alarms area, specify the new values, as required.

Administration | 499
4. Click Save and Close.
Related references

Administrator roles and capabilities on page 506


Page descriptions
Systems Alarms area
You can use the Setup Options dialog box to configure the DataFabric Manager server to send event
notifications in different formats. You can also configure the mail server, event purge interval,
SNMP trap details, and user quota alert details.

Options on page 499


Command buttons on page 500

Options
You can configure the following options:
E-mail

Specifies the following e-mail settings for the notification of alarms:

Mail Server
Specifies the name of your mail server.
From Field
Specifies the e-mail address of the owner who is marked on all the e-mails.

Events

Specifies the event evaluation details.

Purge Interval
Specifies the period of time after which events are removed from the DataFabric
Manager server database. The DataFabric Manager server evaluates events for
deletion on a daily basis.

The default is 25.71 weeks.


SNMP
Traps

Specifies the SNMP trap settings to be received from storage systems. By default, the
SNMP trap is enabled.

Listener Port
Specifies the UDP port on which the DataFabric Manager server Trap Listener
receives traps. To use this feature, you must also configure the DataFabric Manager
server as a trap destination in the systems you are monitoring. The DataFabric
Manager server Trap Listener communicates through port 162, by default.
Window Size

500 | OnCommand Console Help

Displays the SNMP Maximum Traps Received per window option to determine the
number of SNMP traps that can be received by the trap listener within a specified
period of time. By default, 5 minutes is specified.
Max Traps/Window
Specifies the maximum number of SNMP traps that the workstation receives within
the time specified in the SNMP Trap Window Size option. The trap listener attempts
to limit the incoming rate of traps to this value.
The default is 250 traps per window.

User
Quota
Alerts

Enables you to configure or change the values of options that specify e-mail domains
and enable or disable alerts based on quota events. By default, the user quota alerts
option is enabled.

Default E-mail Domain


Specifies the domain that the DataFabric Manager server appends to the user name
when sending a user quota alert.

Command buttons
The command buttons enable you to save or cancel the setup options.
Save and Close

Saves the configuration settings for the selected option.

Cancel

Does not save the recent changes and closes the Setup Options dialog box.

Save

Saves the configuration settings for the selected option.

Systems Annotations area


You can use the Setup Options dialog box to view system-defined annotations such as owner e-mail
address, owner name, and resource tag and modify them.

Options on page 500


Command buttons on page 501

Options
You can configure the following options:
Name

Displays the list of comment field names that include both system-defined and userdefined annotations:

ownerEmail
Specifies the system-defined annotation used for owner's e-mail address.
ownerName
Specifies the system-defined annotation used for the name of the owner.
resourceTag

Administration | 501

Specifies the system-defined annotation used for resource tag.


System
Defined

Displays
Yes

if the corresponding annotation is system generated, and displays


No

if the corresponding annotation is not system generated.


Command buttons
The command buttons enable you to save or cancel the setup options.
Add

Adds a new system annotation name.

Edit

Edits the selected system annotations.

Delete

Deletes the selected system annotations.

Save and Close

Saves the configuration settings for the selected option.

Cancel

Does not save the recent changes and closes the Setup Options dialog box.

Save

Saves the configuration settings for the selected option.

Systems Miscellaneous area


You can use the Setup Options dialog box to configure miscellaneous settings such as audit log
options, credential TTL cache, storage system configuration, script plug-ins, and options to preserve
your local configuration settings.

Options on page 501


Command buttons on page 502

Options
You can configure the following options:
Audit Log

Enables you to set the global option auditLogForever to keep the audit log
files forever in the default log directory of the DataFabric Manager server.
You can view the specific operation in the audit.log file and determine who
performed certain actions in the OnCommand console, the command line
interface, and the APIs. By default, the audit log option is enabled.

Keep Log Files


Enables you to keep the audit log files forever in the default log directory of
the DataFabric Manager server.

502 | OnCommand Console Help


Credential
Cache

Enables you to specify the Time-To-Live (TTL) for web responses cached by the
DataFabric Manager server.
When a user authenticates to the DataFabric Manager server, the DataFabric
Manager server caches the web response and reuses the information to satisfy
subsequent authentication queries for the amount of time specified in this option.

TTL
Displays the Time-To-Live for LDAP server responses cached by the
DataFabric Manager server. By default, LDAP and Windows authentication
information is cached for 20 minutes.

Storage System Enables you to manage local configuration file changes on all storage systems
that are recognized and managed by the DataFabric Manager server.
Configuration

Preserve Local Configuration Changes


Enables you to keep the local configuration file changes on the storage
systems. By default, this option is selected.
Note: The "Preserve Local Configuration Changes" is a global option and
affects all storage systems that are managed by the DataFabric Manager
server.

Script Plugins

Enables you to specify the search path that the DataFabric Manager server uses
to find script interpreters. The value you enter for this option must be a string
that contains multiple paths, delimited by colons or semicolons, depending on
your DataFabric Manager server's platform.

Paged Tables

Search Path
Displays the path for locating the script interpreters. The DataFabric
Manager server uses this path information before using the system path when
searching for script interpreters.

Enables you to specify the number of rows in a report for display.

Rows Per Page


Displays the number of rows displayed per page. By default, the number of
rows selected for display is 20.
Note: This option is applicable only to the Operations Manager console

reports.

Command buttons
The command buttons enable you to save or cancel the setup options.
Save and Close

Saves the configuration settings for the selected option.

Administration | 503
Cancel

Does not save the recent changes and closes the Setup Options dialog box.

Save

Saves the configuration settings for the selected option.

504 | OnCommand Console Help

505

Security and access


Understanding RBAC
What RBAC is
RBAC (role-based access control) provides the ability to control who has access to various features
and resources in DataFabric Manager server.

How RBAC is used


Applications use RBAC to authorize user capabilities. Administrators use RBAC to manage groups
of users by defining roles and capabilities.
For example, if you need to control user access to resources, such as groups, datasets, and resource
pools, you must set up administrator accounts for them. Additionally, if you want to restrict the
information these administrators can view and the operations they can perform, you must apply roles
to the administrator accounts you create.
Note: RBAC permission checks occur in the DataFabric Manager server. RBAC must be
configured using the Operations Manager console or command line interface.

How roles relate to administrators


Role management allows the administrator who logs in with super-user access to restrict the use of
certain DataFabric Manager server functions to other administrators.
The super-user can assign roles to administrators on an individual basis, by group, or globally (and
for all objects in DataFabric Manager server).
You can list the description of an operation by using the dfm role operation list [ -x ]
[ <operation-name> ] command.
The ability to configure administrative users and roles is supported in the Operations Manager
console, which can be accessed from the OnCommand console Administration menu.

Example of how to use RBAC to control access


This example describes how a storage architect can use RBAC to control the operations that can be
performed by a virtual server administrator.
Suppose you are a storage architect and you want to use RBAC to enable the virtual server
administrator (abbreviated to "administrator") to do the following operations: see the VMs associated
with the servers managed by the administrator, create datasets to include them, and attach storage
services and application policies to these datasets to back up the data.

506 | OnCommand Console Help


Assume for this example that the host service registration, validation, and so on have been
successfully completed and that DataFabric Manager server has discovered the virtual server and its
VMs, datastores, and so on. However, at this point, only you and other administrators with the global
read permission in DataFabric Manager server can see these VMs. To enable the virtual server
administrator to perform the desired operations, you need to perform the following steps:
1. Add the administrator as a user.
You add the administrator as an authorized DataFabric Manager server user by using the
Operations Manager console. If DataFabric Manager server is running on Linux, then you must
add the administrator's UNIX identity or LDAP identity. If DataFabric Manager server is running
on Microsoft Windows, you can add an active directory user group that the administrator belongs
to, which allows all administrators in that user group to log onto DataFabric Manager server.
2. Create a resource group.
You next create a resource group by using the Operations Manager console. For this example, we
call the group "virtual admin resource group." Then you add the virtual server and its objects to
the resource group. Any new datasets or policies that the administrator creates will be placed in
this resource group.
3. Assign a role to the administrator.
Assign a role, by using the Operations Manager console, that gives an appropriate level of control
to the administrator. For example, if there is only one administrator, you can assign the default
GlobalApplicationProtection role or you can create a custom role by choosing custom
capabilities. If there are multiple administrators, then a few of the capabilities assigned to any
given administrator should be at that administrator's group level and not on the global level. That
prevents an administrator from reading or modifying objects owned by other administrators.
Related tasks

Accessing the Users and Roles capability (RBAC) on page 375

Administrator roles and capabilities


The RBAC administrator roles determine the tasks you can perform in the OnCommand console.
One or more capabilities must be specified for every role, and you can assign multiple capabilities if
you want the administrator to have more control than a specific role provides. For example, if you
want an administrator to perform both the backup and restore operations, you can create and assign to
the administrator a single role that has both of these capabilities.
You can use the Operations Manager console to create new roles and to customize the default global
roles provided by the DataFabric Manager server and the client applications. For more information
about configuring RBAC, see the OnCommand Operations Manager Administration Guide .
Note: If you want a role with global host service management capability, create a role with the
following properties:

The role inherits from the GlobalHostService role.

Security and access | 507

The role includes the DFM.Database.Read operation on a global level.

Note: A user who is part of the local administrators group is treated as a super-user and

automatically granted full control.


Default global roles
GlobalApplicationProtection Enables you to create and manage application policies, create
datasets with application policies for local backups, use storage
services for remote backups, perform scheduled and on-demand
backups, perform restore operations, and generate reports.
GlobalBackup

Enables you to initiate a backup to any secondary volume and


ignore discovered hosts.

GlobalDataProtection

Enables you to initiate a backup to any secondary volume; view


backup configurations, events and alerts, and replication or failover
policies; and import relationships into datasets.

GlobalDataset

Enables you to create, modify, and delete datasets.

GlobalDelete

Enables you to delete information in the DataFabric Manager


server database, including groups and members of a group,
monitored objects, custom views, primary and secondary storage
systems, and backup relationships, schedules, and retention
policies.

GlobalHostService

Enables you to authorize, configure, and unregister a host service.

GlobalEvent

Enables you to view, acknowledge, and delete events and alerts.

GlobalFullControl

Enables you to view and perform any operation on any object in the
DataFabric Manager server database and configure administrator
accounts. You cannot apply this role to accounts with group access
control.

GlobalMirror

Enables you to create, destroy, and update replication or failover


policies.

GlobalRead

Enables you to view the DataFabric Manager server database,


backup and provisioning configurations, events and alerts,
performance data, and policies.

GlobalRestore

Enables you to restore the primary data to a point in time or to a


new location.

GlobalWrite

Enables you to view or write both primary and secondary data to


the DataFabric Manager server database.

GlobalResourceControl

Enables you to add members to dataset nodes that are configured


with provisioning policies.

508 | OnCommand Console Help


GlobalProvisioning

Enables you to provision primary dataset nodes and can attach


resource pools to secondary or tertiary dataset nodes. The
GlobalProvisioning role also includes all the capabilities of the
GlobalResourceControl, GlobalRead, and GlobalDataset roles for
dataset nodes that are configured with provisioning and protection
policies.

GlobalPerfManagement

Enables you to manage views, event thresholds, and alarms apart


from viewing performance information in Performance Advisor.

Related information

Operations Manager Administration Guide - http://now.netapp.com/NOW/knowledge/docs/


DFM_win/dfm_index.shtml

Access permissions for the Virtual Infrastructure Administrator role


When you create a virtual infrastructure administrator, you must assign specific permissions to
ensure that the administrator can view, back up, and recover the appropriate virtual objects.
A virtual infrastructure administrator role must have the following permissions for the resources:
Groups

Policies

The VI administrator will need the following operation permissions for the group
created for the VI administrator role.
DFM.Database

All

DFM.BackManager

All

DFM.ApplicationPolicy

All

DFM.Dataset

All

DFM.Resource

Control

The VI administrator will need the following operation permissions for each policy
template, located under Local Policies, that you want the virtual administrator to be
able to copy.
DFM.ApplicationPolicy

Storage
services

The VI administrator will need the following operation permissions for each of the
storage services that you want to allow the VI administrator to use.
DFM.StorageService

Protection
Policies

Read

Attach, read, detach, and clear

These are the policies contained within the storage services that you selected above.
DFM.Policy

All

Security and access | 509

Understanding authentication
Authentication methods on the DataFabric Manager server
The DataFabric Manager server uses the information available in the native operating system for
authentication. The server does not maintain its own database of the administrator names and the
passwords.
You can also configure the DataFabric Manager server to use Lightweight Directory Access Protocol
(LDAP). If you configure LDAP, then the server uses it as the preferred method of authentication.

Authentication with LDAP


You can enable LDAP authentication on the DataFabric Manager server and configure it to
communicate with your LDAP servers in order to retrieve relevant data.
The DataFabric Manager server provides predefined templates for the most common LDAP server
types. These templates provide predefined LDAP settings that make the DataFabric Manager server
compatible with your LDAP server.

510 | OnCommand Console Help

511

Plug-ins
Hyper-V troubleshooting
Error: Vss Requestor - Backup Components failed with partial writer error.
Description

This message occurs when backing up a dataset using the Hyper-V plug-in. This
error causes the backup to fail for some of the virtual machines in the dataset.
The following message appears:
Error: Vss Requestor - Backup Components failed with partial
writer error.Writer Microsoft Hyper-V VSS Writer involved in
backup or restore operation reported partial failure. Writer
returned failure code 0x80042336. Writer state is 5.
Application specific error information:
Application error code: 0x1
Application error message: Failed component information:
Failed component: VM GUID XXX
Writer error code: 0x800423f3
Application error code: 0x8004230f
Application error message: Failed to revert to VSS
snapshot on the virtual hard disk 'volume_guid' of the
virtual machine 'vm_name'. (Virtual machine ID XXX)

The following errors appear in the Windows Application event log on the Hyper-V
host:
Volume Shadow Copy Service error: Unexpected error calling
routine GetOverlappedResult. hr = 0x80070057, The parameter
is incorrect.
Operation:
Revert a Shadow Copy
Context:
Execution Context: System Provider
Volume Shadow Copy Service error: Error calling a routine on
a Shadow Copy Provider {b5946137-7b9f-4925af80-51abd60b20d5}. Routine details RevertToSnashot [hr =
0x80042302, A Volume Shadow Copy Service component
encountered an unexpected error.
Check the Application event log for more information.].
Operation:
Revert a Shadow Copy

512 | OnCommand Console Help

Context:
Execution Context: Coordinator

Corrective
action

Retry the dataset backup.

Error: Failed to start VM. Job returned error 32768


Description

After a successfully restore operation, you might get an error message stating that
your Hyper-V virtual machine did not restart. The Hyper-V plug-in gives this error
because the virtual machine is not yet ready to start.

Corrective
action

Currently the Hyper-V plug-in waits two seconds before restarting the virtual
machine. You can configure a longer delay by adding the following attribute in the
Windows registry:
KEY: System\CurrentControlSet\Services\OnCommandHyperV
\Parameters"; attribute (DWORD) : 'vm_restart_sleep'

Error: Failed to start VM. You might need to start the VM using Hyper-V
Manager
Description

After a successfully restore operation, you might get an error message stating that
your Hyper-V virtual machine did not restart. The Hyper-V plug-in gives this error
because the virtual machine is not yet ready to start.

Corrective
action

Currently the Hyper-V plug-in waits two seconds before restarting the virtual
machine. You can configure a longer delay by adding the following attribute in the
Windows registry:
KEY: System\CurrentControlSet\Services\OnCommandHyperV
\Parameters"; attribute (DWORD) : 'vm_restart_sleep'

Error: Vss Requestor - Backup Components failed. An expected disk did


not arrive in the system
Description

This message occurs when you back up a dataset using the Hyper-V plug-in and the
following error appears in the Windows Application event log on the Hyper-V host.
A Shadow Copy LUN was not detected in the system and did not
arrive.
LUN ID

guid

Version
0x0000000000000001
Device Type
0x0000000000000000
Device TypeModifier 0x0000000000000000

Plug-ins | 513

Command Queueing 0x0000000000000001


Bus Type
0x0000000000000006
Vendor Id
vendor
Product Id
LUN
Product Revision
number
Serial Number
serial_number
Storage Identifiers
Version
0
Identifier Count 0

Operation:
Exposing Disks
Locating shadow-copy LUNs
PostSnapshot Event
Executing Asynchronous Operation
Context:
Execution Context: Provider
Provider Name: Data ONTAP VSS Hardware Provider
Provider Version: 6. 1. 0. 4289
Provider ID: {ddd3d232-a96f-4ac5-8f7b-250fd91fd102}
Current State: DoSnapshotSet

Corrective
action

Retry the dataset backup.

Error: Vss Requestor - Backup Components failed. Writer Microsoft HyperV VSS Writer involved in backup or restore encountered a retryable error
Description

If you receive a VSS retry error that causes your backup to fail, the Hyper-V plug-in
retries the backup three times with a wait of one minute between each attempt.
The following error message is displayed in the Hyper-V plug-in report and the
Windows Event log:
Error: Vss Requestor - Backup Components failed.Writer
Microsoft Hyper-V VSS Writer involved in backup or restore
encountered a retryable error. Writer returned failure code
0x800423f3. Writer state is XXX. For more information, see the
Hyper-V-VMMS event log in the Windows Event Viewer.

Corrective
action

You can configure the number of retries (retry count) and the duration of wait time
between the retries (retry interval) using the following registry keys:
Key: HKLM\System\CurrentControlSet\Services\OnCommandHyperV
\Parameters DWORD value in seconds: vss_retry_sleep (The time
duration to wait between retries) DWORD value: vss_retry
(Number of retries)

514 | OnCommand Console Help

These settings are at the Hyper-V host level and the keys and values should be set on
the Hyper-V host for each virtual machine. If the virtual machine is clustered, the
keys should be set on each node in the cluster.

Hyper-V virtual objects taking too long to appear in OnCommand console


Issue

After either first configuring the Hyper-V plug-in or after a failover, Hyper-V
virtual objects take a long time to appear in the OnCommand console.

Cause

The Hyper-V plug-in uses SnapDrive for Windows to enumerate virtual machines.
With large numbers of virtual machines in a clustered setup, it can take SnapDrive
for Windows a significant amount of time to enumerate all of the virtual machines,
so the discovery of Hyper-V objects takes time.

Corrective
action

Depending on the size of your setup, discovery might take longer than you expect.

Increasing SnapDrive operations timeout value in the Windows registry


SnapDrive for Windows has a default operations timeout of 60 seconds; however, you can increase
the timeout value by creating a new registry key in the Windows registry. You can change the
timeout value in instances when you need SnapDrive for Windows to wait longer for operations,
such as backups, to complete.
To increase the SnapDrive for Windows operations timeout value, in the Windows registry, add a
DWORD value named OperationsTimeout in HKEY_LOCAL_MACHINE\SYSTEM
\CurrentControlSet\Services\SWSvc\Parameters, and set a timeout to a value greater than
the default of 60 seconds.

MBR unsupported in the Hyper-V plug-in


Issue

The Hyper-V plug-in does not support MBR LUNs for virtual machines running
on shared volumes or cluster shared volumes.

Cause

A Microsoft API issue returns different volume GUIDs when the cluster shared
volume disk ownership changes from active to passive, for example Node A to
Node B. The volume GUID is not the same as the GUID in the cluster disk
resource property. This issue also applies to virtual machines made highly
available using Microsoft Failover clustering.

Corrective
action

See Knowledge Base article 2006163 on the Microsoft support site.

Related information

Knowledge Base article 2006163 - http://support.microsoft.com/

Plug-ins | 515

Some types of backup failures do not result in partial backup failure


If one virtual machine in a dataset has an error, the Hyper-V plug-in does not successfully complete
the dataset backup, and in some scenarios, does not generate a partial failure. In these situations, the
entire dataset backup fails.
The following examples illustrate some of the scenarios that do not generate partial failure, even
though the problem is associated to a subset of virtual machines in the dataset :

One virtual machine has data on a non-Data ONTAP LUN


One storage system volume exceeds the 255 Snapshot copy limit
One virtual machine is in a Critical state

To successfully complete the backup operation, you need to fix the virtual machine that has the issue.
If that is not possible, you can temporarily move the virtual machine out of the dataset, or create a
dataset that only contains virtual machines known not to have a problem.

Space consumption when taking two snapshot copies for each backup
Issue

For every backup containing Hyper-V objects two snapshots are created, which can
lead to concerns over space consumption.

Cause

Microsoft Hyper-V VSS Writer creates both VM and application-consistent backups


within the VMs, with the applications residing on VHDs. In order to create both
software and VM consistent backups, VSS employs the native auto-recovery process,
which sets the VM to a state consistent with the software snapshot. Hyper-V VSS
writer contacts each VM in the backup, and creates a software consistent snapshot.
Once the snapshots are created, the parent partition creates a VSS snapshot of the
entire disk (LUN) that houses these VMs. Once the parent partition snapshot is
created, VSS requires mounting of the previously created parent partition to roll each
of the VMs back to the software consistent state, and remove any changes that were
made to the VMs after the software snapshot was created. These modifications to the
VHDs must be made persistent. Since these snapshots are read-only by default, a new
snapshot must be made to retain the updated copies of the VHDs. For this reason, a
second snapshot of the volume is created. This snapshot is labeled with a the suffix
_backup and is backup used in restore operations.

Corrective
action

The two snapshots are considered a pair. When the retention period ends for the
backup, both the snapshots are deleted. You should not manually delete the first
snapshot since it is necessary for restore operations.
Microsoft VSS only supports backing up VMs on the host that owns the Cluster
Shared Volume (CSV), so CSV ownership moves between the nodes to create
backups of the VMs on each host in the cluster.
When backing up a CSV, the Hyper-V plug-in creates two snapshots per host in the
cluster that runs a VM from that CSV. This means that if you backup up 15 VMs on a

516 | OnCommand Console Help

single CSV and those VMs are evenly split across three Hyper-V Servers that there
will be a total of six snapshots per backup.

Virtual machine snapshot file location change can cause the Hyper-V plugin backup to fail
If you change a virtual machine snapshot file location to a different Data ONTAP LUN after creating
the virtual machine, you should create at least one virtual machine snapshot using Hyper-V manager
before making a backup using the Hyper-V plug-in. If you change the snapshot file location to a
different LUN and do not make a virtual machine snapshot before making a backup, the backup
operation could fail.

Virtual machine backups taking too long to complete


Issue

If a virtual machine contains several direct-attached iSCSI LUNs or pass-through


LUNs, and SnapDrive for Windows is installed on the virtual machine, the virtual
machine backup can take a long time.

Cause

The Hyper-V writer takes a hardware snapshot of all the LUNs in the virtual
machine using the SnapDrive for Windows VSS hardware provider.

Corrective
action

You can use a Microsoft hotfix that uses the default system provider (software
provider) in the virtual machine to make the snapshot. As a result, the Data
ONTAP VSS hardware provider is not used for snapshot creation inside the child
OS and the backup speed increases.

The Hyper-V writer takes a hardware snapshot of all the LUNs in the virtual machine using the
SnapDrive for Windows VSS hardware provider. There is a Microsoft hotfix that uses the default
system provider (software provider) in the virtual machine to make the snapshot. As a result, the
Data ONTAP VSS hardware provider is not used for snapshot creation inside the child OS and the
backup speed increases. See Knowledge Base article 975354 on the Microsoft support site.
Related information

Knowledge Base article 975354 - http://support.microsoft.com/

Virtual machine backups made while a restore operation is in progress


might be invalid
Issue

A backup created while a restore operation is in progress might be invalid, because


the virtual machine configuration information is missing in the backup copy. The
backup operation is successful, but the backup copy is invalid because the virtual
machine configuration information is not included. Restoring a virtual machine
from this incomplete backup results in data loss and the virtual machine is deleted.

Plug-ins | 517
Cause

The Hyper-V plug-in restore operations delete the virtual machine configuration
information from the Hyper-V host before performing a restore operation. This
behavior is by design from the Microsoft Hyper-V Writer.

Corrective
action

Ensure that the backup schedule does not coincide with the restore operation, or that
the on-demand backup you want to perform does not overlap with a restore
operation of the same data.

Manually locating a lost virtual machine


If your virtual machine is deleted after a failed restore operation, you can manually copy the virtual
machine data from backup snapshot copies so you can restore it.
Before you begin

You must have manually copied all of the virtual machine files, including the virtual machine
configuration, VHD's, and virtual machine snapshot files, from the backup snapshot copy to the
virtual machine's original path.
Steps

1. Enter the following registry key: HKLM\System\CurrentControlSet\Services


\OnCommandHyperV\Parameters DWORD value name: RestoreVMFromExistingData
value data: 1

When you set the value to 1 and perform a restore operation, the Hyper-V plug-in does not copy
the data from the Data ONTAP snapshot copies to the original virtual machine location.
2. Restore the virtual machine from any backup using the Hyper-V plug-in or the Restore-Backup
PowerShell cmdlet.
The Hyper-V plug-in notifies the Hyper-V VSS writer to restore the virtual machine from
existing data.
3. When you are finished, delete the registry value created during Step 1.

Volume Shadow Copy Service error: An internal inconsistency was


detected in trying to contact shadow copy service writers.
Description

When you perform a backup of a virtual machine that uses Windows Server 2003,
it repeatedly fails due to a retry error.

Corrective
action

Check the Windows Application event log inside the virtual machine for any VSS
errors. You can also see Knowledge Base article 940184 on the Microsoft support
site if you see the following error:
Volume Shadow Copy Service error: An internal inconsistency
was detected in trying to contact shadow copy service

518 | OnCommand Console Help

writers. Please check to see that the Event Service and


Volume Shadow Copy Service are operating properly.
Related information

Knowledge Base article 940184 - http://support.microsoft.com/

Hyper-V VHDs do not appear in the OnCommand console


Issue

Hyper-V VHDs do not appear in the OnCommand console after being properly
added to a virtual machine.

Cause

After a VHD is added to a virtual machine, a Windows WMI event for virtual
machine changes is not generated and pass-through LUNS are not properly listed
in the OnCommand console.

Corrective
action

Resend the storage system credentials for one of the storage systems managed by
the host service.

519

Copyright information
Copyright 19942011 NetApp, Inc. All rights reserved. Printed in the U.S.A.
No part of this document covered by copyright may be reproduced in any form or by any means
graphic, electronic, or mechanical, including photocopying, recording, taping, or storage in an
electronic retrieval systemwithout prior written permission of the copyright owner.
Software derived from copyrighted NetApp material is subject to the following license and
disclaimer:
THIS SOFTWARE IS PROVIDED BY NETAPP "AS IS" AND WITHOUT ANY EXPRESS OR
IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED
WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE,
WHICH ARE HEREBY DISCLAIMED. IN NO EVENT SHALL NETAPP BE LIABLE FOR ANY
DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE
GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER
IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR
OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF
ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
NetApp reserves the right to change any products described herein at any time, and without notice.
NetApp assumes no responsibility or liability arising from the use of products described herein,
except as expressly agreed to in writing by NetApp. The use or purchase of this product does not
convey a license under any patent rights, trademark rights, or any other intellectual property rights of
NetApp.
The product described in this manual may be protected by one or more U.S.A. patents, foreign
patents, or pending applications.
RESTRICTED RIGHTS LEGEND: Use, duplication, or disclosure by the government is subject to
restrictions as set forth in subparagraph (c)(1)(ii) of the Rights in Technical Data and Computer
Software clause at DFARS 252.277-7103 (October 1988) and FAR 52-227-19 (June 1987).

520 | OnCommand Console Help

521

Trademark information
NetApp, the NetApp logo, Network Appliance, the Network Appliance logo, Akorri,
ApplianceWatch, ASUP, AutoSupport, BalancePoint, BalancePoint Predictor, Bycast, Campaign
Express, ComplianceClock, Cryptainer, CryptoShred, Data ONTAP, DataFabric, DataFort, Decru,
Decru DataFort, DenseStak, Engenio, Engenio logo, E-Stack, FAServer, FastStak, FilerView,
FlexCache, FlexClone, FlexPod, FlexScale, FlexShare, FlexSuite, FlexVol, FPolicy, GetSuccessful,
gFiler, Go further, faster, Imagine Virtually Anything, Lifetime Key Management, LockVault,
Manage ONTAP, MetroCluster, MultiStore, NearStore, NetCache, NOW (NetApp on the Web),
Onaro, OnCommand, ONTAPI, OpenKey, PerformanceStak, RAID-DP, ReplicatorX, SANscreen,
SANshare, SANtricity, SecureAdmin, SecureShare, Select, Service Builder, Shadow Tape,
Simplicity, Simulate ONTAP, SnapCopy, SnapDirector, SnapDrive, SnapFilter, SnapLock,
SnapManager, SnapMigrator, SnapMirror, SnapMover, SnapProtect, SnapRestore, Snapshot,
SnapSuite, SnapValidator, SnapVault, StorageGRID, StoreVault, the StoreVault logo, SyncMirror,
Tech OnTap, The evolution of storage, Topio, vFiler, VFM, Virtual File Manager, VPolicy, WAFL,
Web Filer, and XBB are trademarks or registered trademarks of NetApp, Inc. in the United States,
other countries, or both.
IBM, the IBM logo, and ibm.com are trademarks or registered trademarks of International Business
Machines Corporation in the United States, other countries, or both. A complete and current list of
other IBM trademarks is available on the Web at www.ibm.com/legal/copytrade.shtml.
Apple is a registered trademark and QuickTime is a trademark of Apple, Inc. in the U.S.A. and/or
other countries. Microsoft is a registered trademark and Windows Media is a trademark of Microsoft
Corporation in the U.S.A. and/or other countries. RealAudio, RealNetworks, RealPlayer,
RealSystem, RealText, and RealVideo are registered trademarks and RealMedia, RealProxy, and
SureStream are trademarks of RealNetworks, Inc. in the U.S.A. and/or other countries.
All other brands or products are trademarks or registered trademarks of their respective holders and
should be treated as such.
NetApp, Inc. is a licensee of the CompactFlash and CF Logo trademarks.
NetApp, Inc. NetCache is certified RealSystem compatible.

522 | OnCommand Console Help

523

How to send your comments


You can help us to improve the quality of our documentation by sending us your feedback.
Your feedback is important in helping us to provide the most accurate and high-quality information.
If you have suggestions for improving this document, send us your comments by e-mail to
doccomments@netapp.com. To help us direct your comments to the correct division, include in the
subject line the product name, version, and operating system.
You can also contact us in the following ways:

NetApp, Inc., 495 East Java Drive, Sunnyvale, CA 94089


Telephone: +1 (408) 822-6000
Fax: +1 (408) 822-4501
Support Telephone: +1 (888) 4-NETAPP

524 | OnCommand Console Help

Index | 525

Index
A
access roles (RBAC)
See RBAC
Add Dataset wizard
decisions to make for protection 200
administration
HTTP transport 494, 495
port 494496
transport 495, 496
Administration Port option 491
Administration Transport option 491
administrator roles
list and descriptions 372, 373, 506, 507
See also RBAC
aggregate full threshold 94, 328
aggregate nearly overcommitted threshold 94, 328
aggregate overcommitted threshold 94, 328
aggregates
editing threshold conditions 454
monitoring inventory 98
relation to traditional volume 141, 333
viewing inventory 98
Aggregates Capacity Growth report 346, 347
Aggregates Capacity report 338
Aggregates Committed Capacity report 345
Aggregates report 314
Aggregates Space Savings report 351
Aggregates view 109111, 114, 115
alarm conditions
event class 32, 33, 385, 387
event severity 32, 33, 385, 387
alarm formats
e-mail format 32, 33, 385, 387
pager alert 32, 33, 385, 387
script 32, 33, 385, 387
SNMP trap 32, 33, 385, 387
alarms
adding 30, 32, 3741, 385, 390, 391
adding for a specific event 33, 387
alarm begin 32, 33, 385, 387
alarm end 32, 33, 385, 387
configurations 32, 385
configuring 41, 42, 392, 393
creating 30, 41, 42, 392, 393
deleting 40, 41, 390, 391

details 37
disabling 40, 41, 390, 391
editing 35, 40, 41, 43, 44, 388, 390, 391
enabling 40, 41, 390, 391
guidelines for creating 30
modifying 35, 43, 44, 388
notification 32, 385
repeat notification 32, 33, 385, 387
settings 496
testing 40, 41, 390, 391
viewing details 37
Alarms tab 29, 40, 41, 390, 391
alarmView 361
archive backup 448
associating storage systems 399, 403
audit log
file 501, 502
settings 496
authentication
LDAP 473, 509
methods 473, 509
AutoSupport
information provided for 19
Availability dashboard panel 22

B
backing up
virtual objects on-demand 61, 276
datasets on-demand 222, 280
Backup Management panel 299
backup options
editing 420
backup relationships 419
Backup Settings area
for configuring local policies 179
backups
before performing on-demand backup 62, 64, 223,
225, 277, 279, 281, 283
deleting 293
guidelines for mounting or unmounting backups in
a VMware environment 271
guidelines for performing on-demand backup 62,
64, 223, 225, 277, 279, 281, 283
Hyper-V saved-state backups 274
locating 288, 292

526 | OnCommand Console Help


monitoring 294
mounting using Backups tab 69, 284
mounting using Server tab 67, 286
on-demand 269
performing on-demand 222, 280
properties 294, 295
remote 269
requirements and restrictions when performing an
on-demand backup 64, 225, 279, 283
restoring data from 297
retaining job progress information 271
scheduled local 269
scripting 270
searching for 297
selecting 297
selecting the restore destination 297
unmounting using Backups tab 70, 285
unmounting using Server tab 68, 287
using CLI 283
version management 269
Backups tab 294, 295
bookmarks to favorite topics 15

C
calculations
effective used data space 326
guaranteed space 326
physical used data space 326
Snapshot reserve 326
total data space 326
total space 326
used Snapshot space 326
chargeback rates for groups
configuring 446
CIFS
account 495, 496
password 495, 496
cluster-related objects 86
clusters
adding 100102
deleting 100102
editing settings 100102
grouping 100102
monitoring inventory 97
viewing inventory 97
Clusters view 100102
configured threshold
reached 24
configuring

mail server 36, 389


non-overlapping schedules for Hyper-V objects
221
storage systems 413
users and roles 375
vFiler units 415
conformance
conditions for datasets 241243
dataset status 238
datasets, failure to conform 240
evaluating error text 232, 246
evaluating for datasets 239
how dataset conformance is monitored 240
monitor intervals 240
monitoring and correction 184
resolving dataset issues manually with a baseline
transfer of data 233, 248
resolving issues automatically without a baseline
transfer of data 234, 249
resolving manually in datasets when a baseline
transfer of data might be necessary 235,
250
test button 243
troubleshooting 246
costing options
chargeback 446, 447
cpuView 362
Create Alarm dialog box 4042, 390393
Create Group dialog box 383
Create Local Policy dialog box 176, 177
credential cache
settings 496
credentials
storage systems 400, 404
custom annotations
adding 497
deleting 498
for alarm recipients 497, 498
custom comment
field names 500, 501
custom label 200
customization
window layout 16

D
dashboard panels
Availability 22
Dataset Overall Status 25
descriptions 21

Index | 527
Events 23
External Relationship Lags 26
Fastest Growing Storage 24
Full Soon Storage 24
monitoring objects 22
Resource Pools 25
Unprotected Data 26
Data ONTAP
licenses, described 195
database backup
archive backup 448
deleting 448
process 448
scheduling to reoccur 449
Snapshot backup 448
starting 450
types 448
database backup option
completed 453
scheduling 451, 452
datacenters
viewing VMware inventory of 54
VMware Datacenters view 72, 73
DataFabric Manager server
verifying host service registration 396, 401
Dataset Overall Status dashboard panel 25
dataset-level naming settings
configuring while adding datasets of virtual objects
215
editing a dataset of virtual objects for 217
datasets
adding to manage physical storage objects 199
adding virtual objects to 209
adding, decisions to make for protection 200
attaching storage services to 229
best practices when configuring datasets of virtual
objects 187, 208, 209
changing storage services on 228
conformance conditions 241243
conformance status values 238
conformance to policy, evaluating 239
creating for protection of virtual objects 203
decisions to make before adding to manage
physical storage objects 200
deleting backups 293
editing to specify naming settings 218
evaluating conformance issues 232, 246
general concepts 181
guidelines for adding a dataset of virtual objects
204207

how conformance is monitored 240


Hyper-V objects that a dataset can include 186
listing nonconformant datasets 246
monitoring backup and mirror relationships 245
monitoring conformance to policy 239
monitoring status 244
names, acceptable characters 200
object types that can be members of datasets of
physical objects 184
of virtual objects 185
properties of, for protection 200
protection of physical storage 184
protection status values 197, 237
provisioning policies
role in dataset management 183
reasons for failure to conform to policy 240
removing a virtual object 213
repairing datasets that contain deleted virtual
objects 231
resolving conformance issues automatically
without a baseline transfer of data 234,
249
resolving conformance issues manually when a
baseline transfer of data might be
necessary 235, 250
resolving conformance issues manually without a
baseline transfer of data 233, 248
resource status values 239
restoring data from backed up storage objects 230
role of protection policies 183
role of provisioning policy 183
status types 236
testing conformance 243
troubleshooting 246
types of conformance status, defined 238
types of protection status, defined 197, 237
types of resource status, defined 239
virtual objects, best practices when configuring
datasets of 187, 208, 209
VMware objects that a dataset can include 186
what a dataset is 181
Datasets Average Space Usage Metric report 353
Datasets Average Space Usage Metric Samples report
356
Datasets IO Usage Metric report 355
Datasets IO Usage Metric Samples report 358
Datasets Maximum Space Usage Metric report 354
Datasets Maximum Space Usage Metric Samples report
357
datastores

528 | OnCommand Console Help


restoring 65, 299
selecting a backup 297
viewing VMware inventory of 56
VMware Datastores view 77, 78, 80
days to full 24
deduplication
license, described 195
default role 371, 505
default tabs 15
default threshold options
aggregates 456, 457
volumes 457, 460
default thresholds options
other 460, 462
deleted objects
repairing datasets that contain deleted virtual
objects 231
retrieving 99
viewing 117, 118
viewing inventory 99
what deleted objects are 87
what happens when storage objects are deleted 87
Deleted Objects view 117, 118
designerReportView 363
directory path for archive backups
changing 450
disconnecting
a LUN 291
forced (of LUN) 291
discovery option
addresses 468
credentials 469
disks
monitoring inventory 99
viewing inventory 99
Disks view 116, 117
display
minimum setting 15

E
Edit Alarm dialog box 35, 43, 44, 388
Edit Group dialog box 384, 385
Edit Local Policy dialog box 176, 177
ESX host name field
selecting the host server for a virtual machine
restore 65, 300
ESX servers
viewing VMware inventory of 55
VMware ESX Servers view 73, 74
event purge interval 499, 500

event reports 312


event severity types
critical 31
emergency 31
error 31
information 31
normal 31
warning 31
events
acknowledging 34, 3739, 388
aggregate almost full 94, 328
aggregate almost overcommitted 94, 328
aggregate full 94, 328
aggregate overcommitted 94, 328
definition of 29
details 3639
how to know when events occur 31
qtree full 146, 332
qtree nearly full 146, 332
resolving 34, 3739, 388
severity types 23, 31
triggered 3739
viewing details 36
volume almost full 141, 333
Events All report 313
Events Current report 312
Events dashboard panel 23
Events tab 3739
External Relationship Lags dashboard panel 26

F
Fastest Growing Storage dashboard panel 24
favorite topics, adding to list 15
File SRM
editing options 472
See also FSRM
File SRM area 472, 473
File Storage Resource Management (FSRM)
See FSRM
File Systems report 316
file types
adding 470
File SRM 470
file-level metadata
monitoring 470
file-level statistics 470
FlexClone
license, described 195
forced disconnect (of LUN) 291

Index | 529
FSRM
monitoring requirements 470
what FSRM does 470
Full Soon Storage dashboard panel 24

G
General Properties tab 260
global access control
precedence over group access control 371, 505
global groups 376
global naming settings
customizing 435
guidelines for customizing 436
requirements for customizing 438, 439
use of a naming script 423
Global Naming Settings Primary Volume area 442
Global Naming Settings Secondary Qtree area 444
Global Naming Settings Secondary Volume area 443
Global Naming Settings Snapshot Copy area 440
group access control
precedence over global access control 371, 505
groups
adding 381, 382
chargeback 384, 385
copying 380382
creating 377, 383
deleting 378, 381, 382
editing 379, 384, 385
global 376
managing 378
member types 383385
moving 380382
what groups are 376
Groups tab 381, 382
growth rate
of storage space utilization 24

H
hbaInitiatorView 367
hbaView 367
health
of managed objects 23
Host Agent
editing options 493
login 495, 496
host services
adding 394
associating with vCenter Server 397, 402

authorizing access to storage 397


configuring 395
deleting 405
overview 393
rediscovering virtual object inventory 408
refreshing the list of 408
registering 394
registering with vCenter Server 397, 402
verifying server registration 396, 401
viewing configured hosts 407
Host Services tab 409411
hosts
Data ONTAP licenses, described 195
hosts.equiv 494, 495
hosts.equiv option 491
how to use Help 15
HTTP
enabling 493, 494
HTTPS
enabling 493, 494
Hyper-V
best practices when configuring datasets of HyperV objects 187, 208, 209
how virtual objects are discovered 53
local protection of virtual objects 167, 192, 193
objects that a dataset can include 186
overlapping policies 274
parent hosts, local policies 274
remote protection of virtual objects 188
viewing server inventory 57
viewing VM inventory 58
Hyper-V plug-in
application-consistent backups 272
co-existence with SnapManager for Hyper-V 275
saved-state backups 274
VSS 272
Hyper-V Servers view 80, 81
Hyper-V VMs view 8183

I
initiatorView 367
interface groups 86
inventory options
editing 482
monitoring 482
inventory reports
overview 314

530 | OnCommand Console Help

J
jobs
canceling 45
defined 45
monitoring 46
understanding 45
viewing details of restore 302
Jobs tab 46, 47, 51, 52
junctions 86

L
lag thresholds 419
LDAP
adding servers 474
authentication 473, 475, 476, 509
deleting servers 474
disabling authentication 474
editing authentication settings 475
enabling 475, 476
enabling authentication 474
server types 476, 477
servers 477, 478
template settings 475
LDAP Authentication area 475, 476
LDAP Server Types area 476, 477
LDAP Servers area 477, 478
licenses
Data ONTAP 195
LIFs
cluster management LIF 86
data LIF 86
node management LIF 86
local backups
scheduling for virtual objects 172, 211
local policies
adding 170
and local backup of virtual objects 167, 192, 193
Backup Settings area 179
copying 173
deleting 174
editing 171
effect of time zones 194
guidelines for adding or editing 169, 170
Name area 177
Policies tab 175, 176
Schedule and Retention area 177
scheduling local backup of virtual objects 172, 211
local protection

of virtual objects 167, 192, 193


local users 411
locating
backup copies 288, 292
Snapshot copies in a backup 289
logical interfaces (LIFs)
See LIFs
logical storage
configuring
LUN path settings 138
qtree quota 137
quota settings 139
volume quota 136
grouping
LUNs 141
qtrees 140
volumes 139
LUNs 135
LUNs view 157159
monitoring
qtree capacity threshold and events 146, 332
volume capacity threshold and events 141,
333
overview 135
qtrees 135
Qtrees view 160162
Quota Settings view 162, 163
Snapshot Copies view 164, 165
volumes 135
Volumes view 150153, 156, 157
login
admin 495, 496
guest 495, 496
login credentials 400, 404
login protocol
RSH 494, 495
SSH 494, 495
Login Protocol option 491
LUNs
definition of 135
disconnecting 291
forced disconnect 291
grouping 141
monitoring inventory 148
path settings 138
viewing inventory 148
LUNs report 317
LUNs view 157159

Index | 531

M
mail server
configuring 499, 500
configuring for alarm notifications 36, 389
managed host options
editing 492
guidelines for changing 491
overview 491
Management Console
how the OnCommand console works with 17
installing 18
management options
for clients 493, 494
for Host Agent 495, 496
for managed hosts 494, 495
miscellaneous options 501, 502
monitoring
dataset conformance to policy 239
flow chart of process 478
local backup progress 294
process 478
query intervals 480
system options 490, 491
monitoring options
for inventory 488, 489
for networking 486, 488
for protection 485, 486
for storage 484, 485
guidelines for changing 480
location to change 480
mounting backups
manually in a Hyper-V environment 288
MultiStore Option
license, described 195

N
Name area
for configuring local policies 177
namespace 120
namespaces 86
naming properties
custom label 200
Naming Properties tab 260263
naming scripts
environment variables for naming primary volumes
425
environment variables for naming secondary
volumes 425

environment variables for naming Snapshot copies


424
limitations 424
understanding 423
naming settings
adding datasets of virtual objects with dataset-level
custom naming 215
configuring naming settings using the Add Dataset
wizard 214
customizing global settings 435
definition of 181
descriptions for primary volumes 429, 430
descriptions for secondary qtrees 433, 434
descriptions for secondary volumes 431, 432
descriptions for Snapshot copies 426, 428
editing a dataset of virtual objects for dataset-level
custom naming 217
editing a dataset to specify dataset level custom
naming 218
global and dataset-level differences 421, 422
guidelines for customizing global settings 436
requirements for customizing global settings 438,
439
understanding naming scripts 423
use of format strings 423
when to customize for related objects 422, 423
navigation 15
NDMP credentials 400, 404
NearStore Option
license, described 195
network address
adding 465
for discovery 465
network credentials
modifying 466
network options
editing 481
monitoring 481
notification
of events 29

O
objects
status types 305
what objects are 376
on-demand backups
before performing 62, 64, 223, 225, 277, 279, 281,
283

532 | OnCommand Console Help


guidelines for performing 62, 64, 223, 225, 277,
279, 281, 283
performing 222, 280
requirements and restrictions 64, 225, 279, 283
Operations Manager console
how the OnCommand console works with 17
Operations Manager Console
launching 17

P
paged tables 496
password
management 495, 496
monitoring 495, 496
percentage
of space availability 22
physical storage
adding
clusters 88
storage controllers 88
aggregates 85
Aggregates view 109111, 114, 115
clusters 85
Clusters view 100102
configuring
aggregate settings 91
cluster settings 90
storage controller settings 89
Deleted Objects view 117, 118
disks 85
Disks view 116, 117
grouping
aggregates 93
clusters 92
storage systems 92
monitoring
aggregate capacity thresholds and events 94,
328
discovery of storage systems 94
overview 85
Storage Controllers view 102104, 107109
storage systems 85
ping intervals
configuring 483
policies
evaluating dataset conformance to 239
monitoring dataset conformance to 239
Policies tab

for listing, configuring, and managing local policies


175, 176
Pre/Post Restore Script field
defining which script to run 65, 300
primary volumes
naming settings descriptions 429, 430
properties
of backups 294, 295
of datasets 200
of events 3739
of jobs 46, 47, 51, 52
protected data
dataset conformance conditions 241243
dataset conformance status, described 238
dataset protection status, described 197, 237
dataset resource status, described 239
evaluating conformance to policy 239
how dataset conformance is monitored 240
reasons for failure to conform to policy 240
protection
decisions before adding a dataset 200
protection options
editing 481
monitoring 481
protection policies
overview 181
role in dataset management 183
provisioning policies
overview 181

Q
qtree threshold conditions
editing 455
qtrees
configuring quota settings 137
definition of 135
grouping 140
monitoring inventory 149
viewing inventory 149
Qtrees Capacity Growth report 348
Qtrees Capacity report 340
Qtrees report 318
Qtrees view 160162
quota settings
monitoring inventory 149
viewing inventory 149
Quota Settings view 162, 163
quotas
configuring settings 139

Index | 533
monitoring quota settings inventory 149
process 135
viewing quota settings inventory 149
why you use 135

R
RBAC
capabilities 372, 373, 506, 507
default roles 372, 373, 506, 507
definition 371, 505
example of how to use 371, 505
how RBAC is used 371, 505
how roles relate to administrators 371, 505
related objects
configuring naming settings using the Add Dataset
wizard 214
datasets
configuring naming settings using the Add
Dataset wizard 214
definition of 181
global and dataset-level naming settings 421, 422
primary volume naming settings 429, 430
secondary qtrees naming settings 433, 434
secondary volume naming settings 431, 432
Snapshot copy naming settings 426, 428
specifying naming by editing a dataset 218
when to customize naming for 422, 423
remote backups 269
remote configuration 412
remote protection
assigning to virtual objects 210
of virtual objects 188
reportOutputView 367
reports
Aggregates 314
Aggregates Capacity 338
Aggregates Capacity Growth 346, 347
Aggregates Committed Capacity 345
Aggregates Space Savings 351
aggregating data 304
computing new columns 304
Datasets Average Space Usage Metric 353
Datasets Average Space Usage Metric Samples 356
Datasets IO Usage Metric 355
Datasets IO Usage Metric Samples 358
Datasets Maximum Space Usage Metric 354
Datasets Maximum Space Usage Metric Samples
357
deleting 303, 308

deleting columns 304


displaying group details 304
displaying hidden columns 304
Events All 313
Events Current 312
exporting content 303
exporting data 303
File Systems 316
filtering data 304
formatting a column 304
formatting data based on conditions 304
hiding columns 304
hiding duplicate values in a column 304
hiding group details 304
LUNs 317
management 303
parameters 303
printing 303
Qtrees 318
Qtrees Capacity 340
Qtrees Capacity Growth 348
reordering columns 304
scheduling 303, 306
scheduling, defined 305
sharing 303, 307
sorting data 304
starting each group on a new page 304
Storage Service Datasets 323
Storage Service Policies 322
Storage Services 322
Storage Systems 319
Storage Systems Capacity 341
tab 308310
usage metric reports
overview 325
User Quotas Capacity 342
vFiler Units 320
Volumes 321
Volumes Capacity 344
Volumes Capacity Growth 349
Volumes Committed Capacity 346
Volumes Space Reservation 349
Volumes Space Savings 352
Reports tab 304, 308310
resource groups
global 376
Resource Pools dashboard panel 25
resource status, dataset 239
restore
scripting 298

534 | OnCommand Console Help


viewing job details 302
where to restore a backup 297
restoring
a datastore 65, 299
a Hyper-V virtual machine 66, 301
a VMware virtual machine 65, 300
data from backups 299
role-based access control
See RBAC
roles
global 371, 505

S
sanhostlunview 368
saved-state backups
how Hyper-V plug-in handles 274
Schedule and Retention area
for configuring local policies 177
Schedule Report dialog box 310
scheduled backups 269
scheduled reports log
viewing 307
script plugins
configuring 496
scripting
arguments 270, 298
backups 270
restore 298
searching for backup copies 288, 292
secondary qtrees
naming settings descriptions 433, 434
secondary volumes
naming settings descriptions 431, 432
secure connections 493, 494
selection of backups 297
server types
editing 475
servers
grouping virtual objects 59
Hyper-V Servers view 80, 81
Hyper-V VMs view 8183
viewing Hyper-V server inventory 57
viewing Hyper-V VM inventory 58
viewing VMware datacenter inventory 54, 55
viewing VMware datastore inventory 56
viewing VMware virtual center inventory 53
viewing VMware virtual machine inventory 56
VMware Datacenters view 72, 73
VMware Datastores view 77, 78, 80

VMware ESX Servers view 73, 74


VMware Virtual Centers view 71, 72
VMware VMs view 7477
Setup Options dialog box 495, 496
Share Report dialog box 311
SnapDrive for Windows
Hyper-V manual mount and unmount process 288
SnapManager for Hyper-V
co-existence with Hyper-V plug-in 275
co-existence with OnCommand console 275
manually transitioning dataset information 275
SnapMirror
license, described 195
SnapMirror Sync
license, described 195
Snapshot backup 448
Snapshot copies
locating in a backup 289
monitoring inventory 150
naming settings descriptions 426, 428
overview 136
viewing inventory 150
Snapshot Copies view 164, 165
SnapVault 419
SnapVault Data
primary license, described 195
secondary license, described 195
SnapVault Linux
license, described 195
SnapVault UNIX
license, described 195
SnapVault Windows Open File Manager
license, described 195
SNMP communities
adding 466
editing 466
SNMP traps 499, 500
space utilization 24
SRM, File
See FSRM
status definitions
dataset conformance 238
dataset protection 197, 237
dataset resource 239
status of an object
critical 305
emergency 305
error 305
normal 305
unknown 305

Index | 535
warning 305
storage capacity reports
overview 323
storage chargeback
configuring 446
storage controllers
adding 102104, 107109
deleting 102104, 107109
editing settings 102104, 107109
grouping 102104, 107109
monitoring inventory 98
viewing inventory 98
Storage Controllers view 102104, 107109
storage objects
adding again 117, 118
deleted by 117, 118
deleted date 117, 118
deleted objects 87
deleting 117, 118
restoring data from 230
undeleting 117, 118
viewing deleted objects 117, 118
storage options
editing 480
monitoring 480
Storage Service Datasets report 323
Storage Service Policies report 322
storage services
assigning to a dataset of virtual objects 210
attaching to existing datasets 229
changing on a dataset 228
for executing remote protection of virtual objects
188
overview 181
supplied with the product 189
Storage Services report 322
storage systems
adding users 412
associating with a host service 399, 403
authorizing host service access 397
configuration 501, 502
configuring 413, 496
Data ONTAP licenses, described 195
discovering 467, 468
login credentials 400, 404
managing configuration files 413
NDMP credentials 400, 404
remote configuration 412
Storage Systems Capacity report 341
Storage Systems report 319

surprise removal (of LUN) 291


system annotations
settings 496
system options
configuring 496
editing 483
monitoring 483
Systems Alarms area 499, 500
Systems Annotations area 500, 501
Systems Miscellaneous area 501, 502

T
table settings 496, 501, 502
test conformance check 243
thresholds
aggregate full 94, 328
aggregate full interval 94, 328
aggregate nearly full 94, 328
aggregate nearly overcommitted 94, 328
aggregate overcommitted 94, 328
time zone
effect on protection job schedules in datasets of
virtual objects 194
traditional volumes
See volumes
transitioning legacy Hyper-V dataset information
manually 275
trends
of storage space utilization 24
troubleshooting
dataset conformance conditions 241243
dataset conformance issues 246
dataset failure to conform 240
evaluating dataset conformance 239
listing nonconformant datasets 246

U
unmounting backups
manually in a Hyper-V environment 288
Unprotected Data dashboard panel 26
usage metric reports
guidelines for solving issues 326
user quotas
alerts 499, 500
editing 123
User Quotas Capacity report 342
Users and Roles capability
See RBAC
usersView 368

536 | OnCommand Console Help

V
vCenter Server
associating a host service 397, 402
registering a host service 397, 402
version management
backups 269
vFiler units
configuration tasks to manage 414
configuring 415
defined 119
deleting 125127, 129, 130
discovery 119
editing settings 120, 125127, 129, 130
grouping 122, 125127, 129, 130
monitoring inventory 124
viewing inventory 124
vFiler Units report 320
vFiler Units view 125127, 129, 130
vFilers
discovery 119
editing settings 120
grouping 122
threshold settings 120
virtual centers
viewing VMware inventory of 53
VMware Virtual Centers view 71, 72
virtual disk files
selecting a backup 297
virtual inventory
adding virtual machines to 59
deleting virtual objects from 60
virtual machines
adding to inventory 59
restoring Hyper-V virtual machines 66, 301
restoring VMware virtual machines 65, 300
selecting a backup 297
viewing Hyper-V inventory 58
viewing VMware inventory 56
VMware VMs view 7477
virtual object inventory
rediscovery 408
virtual objects
adding to a dataset 209
best practices when configuring datasets of virtual
objects 187, 208, 209
configuring remote protection for 210
definition of 181
deleting from virtual inventory 60

discovery 53
grouping 59
guidelines for adding datasets containing virtual
objects 204207
local protection of 167, 192, 193
performing on-demand backups 61, 276
remote protection of 188
removing from a dataset 213
scheduling local backup for 172, 211
unprotecting 213
virtual servers
definition 120
deleting 130134
editing settings 121, 130134
grouping 123, 130134
monitoring inventory 125
viewing inventory 125
virtual storage
grouping vFiler units 122
grouping Vservers 123
vFiler Units view 125127, 129, 130
Vservers view 130134
VMs (virtual machines)
See virtual machines
VMware
best practices when configuring datasets of
VMware objects 187, 208, 209
guidelines for mounting or unmounting backups in
a VMware environment 271
how virtual objects are discovered 53
local protection of virtual objects 167, 192, 193
objects that a dataset can include 186
remote protection of virtual objects 188
viewing datacenter inventory 54, 55
viewing datastore inventory 56
viewing virtual center inventory 53
viewing virtual machine inventory 56
VMware Datacenters view 72, 73
VMware Datastores view 77, 78, 80
VMware ESX Servers view 73, 74
VMware Virtual Centers view 71, 72
VMware VMs view 7477
volume full threshold 141, 333
volume nearly full threshold 141, 333
volume threshold conditions
editing 454
volumeDedupeDetailsView 369
volumes
configuring quota settings 136
definition of 135

Index | 537
grouping 139
monitoring inventory 148
viewing inventory 148
Volumes Capacity Growth report 349
Volumes Capacity report 344
Volumes Committed Capacity report 346
Volumes report 321
Volumes Space Reservation report 349
Volumes Space Savings report 352
Volumes view 150153, 156, 157
Vservers
definition 120
deleting 130134
editing settings 121, 130134
grouping 123, 130134
monitoring inventory 125

viewing inventory 125


Vservers view 130134
VSS
about 272
Hyper-V plug-in 272
verifying provider used 274
viewing installed providers 273

W
welcome 15
window layout
customization 16
navigation 15

You might also like