Storage Profiles
This chapter includes the following sections:
Storage Profiles
To allow flexibility in defining the number of storage disks, roles and usage of these disks, and other storage parameters, you can create and use storage profiles. A storage profile encapsulates the storage requirements for one or more service profiles. LUNs configured in a storage profile can be used as boot LUNs or data LUNs, and can be dedicated to a specific server. You can also specify a local LUN as a boot device.
Storage profiles allow you to do the following:
-
Configure multiple virtual drives and select the physical drives that are used by a virtual drive.
-
Configure the storage capacity of a virtual drive.
-
Configure the number, type and role of disks in a disk group.
-
Associate a storage profile with a service profile.
Note | LUN resizing is not supported. |
- Virtual Drives
- Virtual Drive Naming
- RAID Levels
- Supported LUN Modifications
- Unsupported LUN Modifications
- LUN Dereferencing
- Creating or Editing a Storage Profile
Virtual Drives
A disk group can be partitioned into virtual drives. Each virtual drive appears as an individual physical device to the Operating System.
All virtual drives in a disk group must be managed by using a single disk group policy.
Configuration States
-
Applying—Creation of the virtual drive is in progress.
-
Applied—Creation of the virtual drive is complete, or virtual disk policy changes are configured and applied successfully.
-
Failed to apply—Creation, deletion, or renaming of a virtual drive has failed due to errors in the underlying storage subsystem.
-
Orphaned—The service profile that contained this virtual drive is deleted or the service profile is no longer associated with a storage profile.
-
Not in use—The service profile that contained this virtual drive is in the disassociated state.
Deployment States
Operability States
-
Optimal—The virtual drive operating condition is good. All configured drives are online.
-
Degraded—The virtual drive operating condition is not optimal. One of the configured drives has failed or is offline.
-
Cache-degraded—The virtual drive has been created with a write cache policy of Write Back Good BBU mode, but the BBU has failed, or there is no BBU.
Note
This state does not occur if you select Always Write Back mode.
-
Partially degraded—The operating condition in a RAID 6 virtual drive is not optimal. One of the configured drives has failed or is offline. RAID 6 can tolerate up to two drive failures.
-
Offline—The virtual drive is not available to the RAID controller. This is essentially a failed state.
-
Unknown—The state of the virtual drive is not known.
Presence States
Virtual Drive Naming
When you use Cisco UCS Central to create a virtual drive, Cisco UCS Central assigns a unique ID that can be used to reliably identify the virtual drive for further operations. Cisco UCS Central also provides the flexibility to provide a name to the virtual drive at the time of service profile association. Any virtual drive without a service profile or a server reference is marked as an orphan virtual drive.
In addition to a unique ID, a name is assigned to the drive. Names can be assigned in two ways:
-
When configuring a virtual drive, you can explicitly assign a name that can be referenced in storage profiles.
-
If you have not preprovisioned a name for the virtual drive, Cisco UCS Central generates a unique name for the virtual drive.
You can rename virtual drives that are not referenced by any service profile or server.
RAID Levels
The RAID level of a disk group describes how the data is organized on the disk group for the purpose of ensuring availability, redundancy of data, and I/O performance.
-
Striping—Segmenting data across multiple physical devices. This improves performance by increasing throughput due to simultaneous device access.
-
Mirroring—Writing the same data to multiple devices to accomplish data redundancy.
-
Parity—Storing of redundant data on an additional device for the purpose of error correction in the event of device failure. Parity does not provide full redundancy, but it allows for error recovery in some scenarios.
-
Spanning—Allows multiple drives to function like a larger one. For example, four 20 GB drives can be combined to appear as a single 80 GB drive.
-
RAID 0 Striped—Data is striped across all disks in the array, providing fast throughput. There is no data redundancy, and all data is lost if any disk fails.
-
RAID 1 Mirrored—Data is written to two disks, providing complete data redundancy if one disk fails. The maximum array size is equal to the available space on the smaller of the two drives.
-
RAID 5 Striped Parity—Data is striped across all disks in the array. Part of the capacity of each disk stores parity information that can be used to reconstruct data if a disk fails. RAID 5 provides good data throughput for applications with high read request rates.
RAID 5 distributes parity data blocks among the disks that are part of a RAID-5 group and requires a minimum of three disks.
-
RAID 6 Striped Dual Parity—Data is striped across all disks in the array and two sets of parity data are used to provide protection against failure of up to two physical disks. In each row of data blocks, two sets of parity data are stored.
Other than addition of a second parity block, RAID 6 is identical to RAID 5 . A minimum of four disks are required for RAID 6.
-
RAID 10 Mirrored and Striped—RAID 10 uses mirrored pairs of disks to provide complete data redundancy and high throughput rates through block-level striping. RAID 10 is mirroring without parity and block-level striping. A minimum of four disks are required for RAID 10.
-
RAID 50 Striped Parity and Striped—Data is striped across multiple striped parity disk sets to provide high throughput and multiple disk failure tolerance.
-
RAID 60 Striped Dual Parity and Striped—Data is striped across multiple striped dual parity disk sets to provide high throughput and greater disk failure tolerance.
Supported LUN Modifications
Some modifications that are made to the LUN configuration when LUNs are already deployed on an associated server are supported.
The following are the types of modifications that can be performed:
-
Creation of a new virtual drive.
-
Deletion of an existing virtual drive, which is in the orphaned state.
-
Non-disruptive changes to an existing virtual drive. These changes can be made on an existing virtual drive without loss of data, and without performance degradation:
The removal of a LUN will cause a warning to be displayed. Ensure that you take action to avoid loss of data.
Unsupported LUN Modifications
Some modifications to existing LUNs are not possible without destroying the original virtual drive and creating a new one. All data is lost in these types of modification, and these modifications are not supported.
-
RAID-level changes that do not support reconstruction. For example, RAID5 to RAID1.
-
Shrinking the size of a virtual drive.
-
RAID-level changes that support reconstruction, but where there are other virtual drives present on the same drive group.
-
Disk removal when there is not enough space left on the disk group to accommodate the virtual drive.
-
Explicit change in the set of disks used by the virtual drive.
LUN Dereferencing
A LUN is dereferenced when it is no longer used by any service profile. This can occur as part of the following scenarios:
-
The LUN is no longer referenced from the storage profile
-
The storage profile is no longer referenced from the service profile
-
The server is disassociated from the service profile
-
The server is decommissioned
When the LUN is no longer referenced, but the server is still associated, re-association occurs. When the service profile that contained the LUN is disassociated, the LUN state is changed to Not in Use. When the service profile that contained the LUN is deleted, the LUN state is changed to Orphaned. When decommissioning the server, the state of all the LUNs associated with the server is changed to Not in use or Orphaned. However, no action is taken to delete the actual LUNs.
Creating or Editing a Storage Profile
Disk Groups and Disk Group Configuration Policies
Servers in a chassis can use storage that is centralized in that chassis. You can select and configure the disks to be used for storage. A logical collection of these physical disks is called a disk group. Disk groups allow you to organize local disks. The storage controller controls the creation and configuration of disk groups.
A disk group configuration policy defines how a disk group is created and configured. The policy specifies the RAID level to be used for the disk group. It also specifies either a manual or an automatic selection of disks for the disk group, and roles for disks. You can use a disk group policy to manage multiple disk groups. However, a single disk group can be managed only by one disk group policy.
Creating or Editing a Disk Group Configuration Policy
Step 1 | In the Task
bar, type
Create
Disk Group Configuration Policy and press Enter.
This launches the Create Disk Group Configuration Policy dialog box. |
Step 2 | In Basic, click Organization and select the location in which you want to create the disk group configuration policy. |
Step 3 | Enter the
Name and optional
Description.
The name is case sensitive. |
Step 4 | Select the
Raid
Level.
This can be one of the following: |
Step 5 | In Disk Group, select the Drive Type, type values for the drive information, and choose whether to use the remaining disks. |
Step 6 | In Virtual Drive icon, complete the fields as necessary. |
Step 7 | Click Create. |