- Preface
- Overview of Cisco Unified Computing System
- Overview of Cisco UCS Manager
- Overview of Cisco UCS Manager CLI
- Configuring the Fabric Interconnects
- Configuring Ports and Port Channels
- Configuring Communication Services
- Configuring Authentication
- Configuring Organizations
- Configuring Role-Based Access Control
- Configuring DNS Servers
- Configuring System-Related Policies
- Managing Licenses
- Managing Virtual Interfaces
- Registering Cisco UCS Domains with Cisco UCS Central
- VLANs
- Configuring LAN Pin Groups
- Configuring MAC Pools
- Configuring Quality of Service
- Configuring Network-Related Policies
- Configuring Upstream Disjoint Layer-2 Networks
- Configuring Named VSANs
- Configuring SAN Pin Groups
- Configuring WWN Pools
- Configuring Storage-Related Policies
- Configuring Fibre Channel Zoning
- Configuring Server-Related Pools
- Setting the Management IP Address
- Configuring Server-Related Policies
- Configuring Server Boot
- Deferring Deployment of Service Profile Updates
- Service Profiles
- Configuring Storage Profiles
- Managing Power in Cisco UCS
- Managing Time Zones
- Managing the Chassis
- Managing Blade Servers
- Managing Rack-Mount Servers
- CIMC Session Management
- Managing the I/O Modules
- Backing Up and Restoring the Configuration
- Recovering a Lost Password
- Storage Profiles
- Disk Groups and Disk Group Configuration Policies
- RAID Levels
- Automatic Disk Selection
- Supported LUN Modifications
- Unsupported LUN Modifications
- Disk Insertion Handling
- Virtual Drive Naming
- LUN Dereferencing
- Controller Constraints and Limitations
- Configuring Storage Profiles
- Configuring a Disk Group Policy
- Creating a Storage Profile
- Deleting a Storage Profile
- Creating a Storage Profile PCH Controller Definition
- Deleting a Storage Profile PCH Controller Definition
- Creating Local LUNs
- Associating a Storage Profile with a Service Profile
- Displaying Details of All Local LUNs Inherited By a Service Profile
- Importing Foreign Configurations for a RAID Controller on a Blade Server
- Importing Foreign Configurations for a RAID Controller on a Rack Server
- Configuring Local Disk Operations on a Blade Server
- Configuring Local Disk Operations on a Rack Server
- Configuring Virtual Drive Operations
- Boot Policy for Local Storage
- Local LUN Operations in a Service Profile
- Viewing the Local Disk Locator LED State
Configuring Storage
Profiles
This part contains the following chapters:
- Storage Profiles
- Disk Groups and Disk Group Configuration Policies
- RAID Levels
- Automatic Disk Selection
- Supported LUN Modifications
- Unsupported LUN Modifications
- Disk Insertion Handling
- Virtual Drive Naming
- LUN Dereferencing
- Controller Constraints and Limitations
- Configuring Storage Profiles
Storage Profiles
You can create a storage profile both at an org level and at a service-profile level. A service profile can have a dedicated storage profile as well as a storage profile at an org level.
Disk Groups and Disk Group Configuration Policies
You can select and configure the disks to be used for storage. A logical collection of these physical disks is called a disk group. Disk groups allow you to organize local disks. The storage controller controls the creation and configuration of disk groups.
A disk group configuration policy defines how a disk group is created and configured. The policy specifies the RAID level to be used for the disk group. It also specifies either a manual or an automatic selection of disks for the disk group, and roles for disks. You can use a disk group policy to manage multiple disk groups. However, a single disk group can be managed only by one disk group policy.
A hot spare is an unused extra disk that can be used by a disk group in the case of failure of a disk in the disk group. Hot spares can be used only in disk groups that support a fault-tolerant RAID level. In addition, a disk can be allocated as a global hot spare, which means that it can be used by any disk group.
Virtual Drives
A disk group can be partitioned into virtual drives. Each virtual drive appears as an individual physical device to the Operating System.
All virtual drives in a disk group must be managed by using a single disk group policy.
Configuration States
-
Applying—Creation of the virtual drive is in progress.
-
Applied—Creation of the virtual drive is complete, or virtual disk policy changes are configured and applied successfully.
-
Failed to apply—Creation, deletion, or renaming of a virtual drive has failed due to errors in the underlying storage subsystem.
-
Orphaned—The service profile that contained this virtual drive is deleted or the service profile is no longer associated with a storage profile.
Deployment States
Operability States
-
Optimal—The virtual drive operating condition is good. All configured drives are online.
-
Degraded—The virtual drive operating condition is not optimal. One of the configured drives has failed or is offline.
-
Cache-degraded—The virtual drive has been created with a write policy of write back mode, but the BBU has failed, or there is no BBU.
Note
This state does not occur if you select the always write back mode.
-
Partially degraded—The operating condition in a RAID 6 virtual drive is not optimal. One of the configured drives has failed or is offline. RAID 6 can tolerate up to two drive failures.
-
Offline—The virtual drive is not available to the RAID controller. This is essentially a failed state.
-
Unknown—The state of the virtual drive is not known.
Presence States
RAID Levels
The RAID level of a disk group describes how the data is organized on the disk group for the purpose of ensuring availability, redundancy of data, and I/O performance.
-
Striping—Segmenting data across multiple physical devices. This improves performance by increasing throughput due to simultaneous device access.
-
Mirroring—Writing the same data to multiple devices to accomplish data redundancy.
-
Parity—Storing of redundant data on an additional device for the purpose of error correction in the event of device failure. Parity does not provide full redundancy, but it allows for error recovery in some scenarios.
-
Spanning—Allows multiple drives to function like a larger one. For example, four 20 GB drives can be combined to appear as a single 80 GB drive.
-
RAID 0 Striped—Data is striped across all disks in the array, providing fast throughput. There is no data redundancy, and all data is lost if any disk fails.
-
RAID 1 Mirrored—Data is written to two disks, providing complete data redundancy if one disk fails. The maximum array size is equal to the available space on the smaller of the two drives.
-
RAID 5 Striped Parity—Data is striped across all disks in the array. Part of the capacity of each disk stores parity information that can be used to reconstruct data if a disk fails. RAID 5 provides good data throughput for applications with high read request rates.
RAID 5 distributes parity data blocks among the disks that are part of a RAID-5 group and requires a minimum of three disks.
-
RAID 6 Striped Dual Parity—Data is striped across all disks in the array and two sets of parity data are used to provide protection against failure of up to two physical disks. In each row of data blocks, two sets of parity data are stored.
Other than addition of a second parity block, RAID 6 is identical to RAID 5 . A minimum of four disks are required for RAID 6.
-
RAID 10 Mirrored and Striped—RAID 10 uses mirrored pairs of disks to provide complete data redundancy and high throughput rates through block-level striping. RAID 10 is mirroring without parity and block-level striping. A minimum of four disks are required for RAID 10.
-
RAID 50 Striped Parity and Striped—Data is striped across multiple striped parity disk sets to provide high throughput and multiple disk failure tolerance.
-
RAID 60 Striped Dual Parity and Striped—Data is striped across multiple striped dual parity disk sets to provide high throughput and greater disk failure tolerance.
Automatic Disk Selection
When you specify a disk group configuration, and do not specify the local disks in it, Cisco UCS Manager determines the disks to be used based on the criteria specified in the disk group configuration policy. Cisco UCS Manager can make this selection of disks in multiple ways.
When all qualifiers match for a set of disks, then disks are selected sequentially according to their slot number. Regular disks and dedicated hot spares are selected by using the lowest numbered slot.
The following is the disk selection process:
-
Iterate over all local LUNs that require the creation of a new virtual drive. Iteration is based on the following criteria, in order:
-
Select regular disks depending on the minimum number of disks and minimum disk size. Disks are selected sequentially starting from the lowest numbered disk slot that satisfies the search criteria.
Note
If you specify Any as the type of drive, the first available drive is selected. After this drive is selected, subsequent drives will be of a compatible type. For example, if the first drive was SATA, all subsequent drives would be SATA.
-
Select dedicated hot spares by using the same method as normal disks. Disks are only selected if they are in an Unconfigured Good state.
-
If a provisioned LUN has the same disk group policy as a deployed virtual drive, then try to deploy the new virtual drive in the same disk group. Otherwise, try to find new disks for deployment.
Supported LUN Modifications
Some modifications that are made to the LUN configuration when LUNs are already deployed on an associated server are supported.
The following are the types of modifications that can be performed:
-
Creation of a new virtual drive.
-
Deletion of an existing virtual drive, which is in the orphaned state.
-
Non-disruptive changes to an existing virtual drive. These changes can be made on an existing virtual drive without loss of data, and without performance degradation:
The removal of a LUN will cause a warning to be displayed. Ensure that you take action to avoid loss of data.
Unsupported LUN Modifications
Some modifications to existing LUNs are not possible without destroying the original virtual drive and creating a new one. All data is lost in these types of modification, and these modifications are not supported.
-
RAID-level changes that do not support reconstruction. For example, RAID5 to RAID1.
-
Shrinking the size of a virtual drive.
-
RAID-level changes that support reconstruction, but where there are other virtual drives present on the same drive group.
-
Disk removal when there is not enough space left on the disk group to accommodate the virtual drive.
-
Explicit change in the set of disks used by the virtual drive.
Disk Insertion Handling
When the following sequence of events takes place:
-
The LUN is created in one of the following ways: -
The LUN is successfully deployed, which means that a virtual drive is created, which uses the slot.
-
You remove a disk from the slot, possibly because the disk failed.
-
You insert a new working disk into the same slot.
- Non-Redundant Virtual Drives
- Redundant Virtual Drives with No Hot Spare Drives
- Redundant Virtual Drives with Hot Spare Drives
- Replacing Hot Spare Drives
- Inserting Physical Drives into Unused Slots
Non-Redundant Virtual Drives
For non-redundant virtual drives (RAID 0), when a physical drive is removed, the state of the virtual drive is Inoperable. When a new working drive is inserted, the new physical drive goes to an Unconfigured Good state.
For non-redundant virtual drives, there is no way to recover the virtual drive. You must delete the virtual drive and re-create it.
Redundant Virtual Drives with No Hot Spare Drives
For redundant virtual drives (RAID 1, RAID 5, RAID 6, RAID 10, RAID 50, RAID 60) with no hot spare drives assigned, virtual drive mismatch, virtual drive member missing, and local disk missing faults appear until you insert a working physical drive into the same slot from which the old physical drive was removed.
If the physical drive size is greater than or equal to that of the old drive, the storage controller automatically uses the new drive for the virtual drive. The new drive goes into the Rebuilding state. After rebuild is complete, the virtual drive goes back into the Online state.
Redundant Virtual Drives with Hot Spare Drives
For redundant virtual drives (RAID 1, RAID 5, RAID 6, RAID 10, RAID 50, RAID 60) with hot spare drives assigned, when a drive fails, or when you remove a drive, the dedicated hot spare drive, if available, goes into the Rebuilding state with the virtual drive in the Degraded state. After rebuilding is complete, that drive goes to the Online state.
Cisco UCSM raises a disk missing and virtual drive mismatch fault because although the virtual drive is operational, it does not match the physical configuration that Cisco UCSM expects.
if you insert a new disk in the slot with the disk missing, automatic copy back starts from the earlier hot spare disk to the newly inserted disk. After copy back, the hot spare disk is restored. In this state all faults are cleared.
If automatic copy back does not start, and the newly inserted disk remains in the Unconfigured Good, JBOD, or Foreign Configuration state, remove the new disk from the slot, reinsert the earlier hot spare disk into the slot, and import foreign configuration. This initiates the rebuilding process and the drive state becomes Online. Now, insert the new disk in the hot spare slot and mark it as hot spare to match it exactly with the information available in Cisco UCSM.
Replacing Hot Spare Drives
If a hot spare drive is replaced, the new hot spare drive will go to the Unconfigured Good, Unconfigured Bad, JBOD, or Foreign Configuration state.
Cisco UCSM will raise a virtual drive mismatch or virtual drive member mismatch fault because the hot spare drive is in a state different from the state configured in Cisco UCSM.
You must manually clear the fault. To do this, you must perform the following actions:
Inserting Physical Drives into Unused Slots
If you insert new physical drives into unused slots, neither the storage controller nor Cisco UCSM will make use of the new drive even if the drive is in the Unconfigured Good state and there are virtual drives that are missing good physical drives.
The drive will simply go into the Unconfigured Good state. To make use of the new drive, you will need to modify or create LUNs to reference the newly inserted drive.
Virtual Drive Naming
When you use UCSM to create a virtual drive, UCSM assigns a unique ID that can be used to reliably identify the virtual drive for further operations. UCSM also provides the flexibility to provide a name to the virtual drive at the time of service profile association. Any virtual drive without a service profile or a server reference is marked as an orphan virtual drive.
In addition to a unique ID, a name is assigned to the drive. Names can be assigned in two ways:
-
When configuring a virtual drive, you can explicitly assign a name that can be referenced in storage profiles.
-
If you have not preprovisioned a name for the virtual drive, UCSM generates a unique name for the virtual drive.
You can rename virtual drives that are not referenced by any service profile or server.
LUN Dereferencing
A LUN is dereferenced when it is no longer used by any service profile. This can occur as part of the following scenarios:
-
The LUN is no longer referenced from the storage profile
-
The storage profile is no longer referenced from the service profile
-
The server is disassociated from the service profile
-
The server is decommissioned
When the LUN is no longer referenced, but the server is still associated, re-association occurs.
When the service profile that contained the LUN is deleted, the LUN state is changed to Orphaned.
Controller Constraints and Limitations
-
For Cisco UCS C240, C220, C24, and C22 servers, the storage controller allows 24 virtual drives per server. For all other servers, the storage controller allows 16 virtual drives per server.
-
In Cisco UCS Manager Release 2.2(4), blade servers do not support drives with a block size of 4K, but rack-mount servers support such drives. If a drive with a block size of 4K is inserted into a blade server, discovery fails and the following error message appears:Unable to get Scsi Device Information from the system.
Configuring Storage Profiles
Configuring a Disk Group Policy
You can choose to configure a disk group policy through automatic or manual disk selection. Configuring a disk group involves the following:
- Setting the RAID Level
- Automatically Configuring Disks in a Disk Group
- Manually Configuring Disks in a Disk Group
- Configuring Virtual Drive Properties
Setting the RAID Level
This example shows how to set the RAID level for a disk group configuration policy.
UCS-A# scope org UCS-A /org # create disk-group-config-policy raid5policy UCS-A /org/disk-group-config-policy* # set raid-level raid-5-striped-parity UCS-A /org/disk-group-config-policy* # commit-buffer
Automatically or manually configure disks as part of the disk group configuration policy.
Automatically Configuring Disks in a Disk Group
You can allow UCSM to automatically select and configure disks in a disk group.
This example shows how to automatically configure disks for a disk group configuration policy.
UCS-A# scope org UCS-A /org # enter disk-group-config-policy raid5policy UCS-A /org/disk-group-config-policy* # enter disk-group-qual UCS-A /org/disk-group-config-policy/disk-group-qual* # set drive-type hdd UCS-A /org/disk-group-config-policy/disk-group-qual* # set min-drive-size 1000 UCS-A /org/disk-group-config-policy/disk-group-qual* # set num-ded-hot-spares 2 UCS-A /org/disk-group-config-policy/disk-group-qual* # set num-drives 7 UCS-A /org/disk-group-config-policy/disk-group-qual* # set num-glob-hot-spares 2 UCS-A /org/disk-group-config-policy/disk-group-qual* # set use-remaining-disks no UCS-A /org/disk-group-config-policy/disk-group-qual* # commit-buffer
Configure Virtual Drives.
Manually Configuring Disks in a Disk Group
You can manually configure disks for a disk group.
Command or Action | Purpose | |
---|---|---|
Step 1 | UCS-A# scope org org-name |
Enters the organization mode for the specified organization. To enter the root organization mode, enter / as the org-name. |
Step 2 | UCS-A /org# enter disk-group-config-policy disk-group-name |
Enters disk group configuration policy mode for the specified disk group name. |
Step 3 | UCS-A /org/disk-group-config-policy* # create local-disk-config-ref slot-num |
Creates a local disk configuration reference for the specified slot and enters local disk configuration reference mode. |
Step 4 | UCS-A /org/disk-group-config-policy/local-disk-config-ref *# set role role |
|
Step 5 | UCS-A /org/disk-group-config-policy/local-disk-config-ref *# set span-id span-id |
Specifies the ID of the span group to which the disk belongs. Disks belonging to a single span group can be treated as a single disk with a larger size. The values range from 0 to 8. You can also set the Span ID as Unspecified when spanning information is not required. |
Step 6 | UCS-A /org/disk-group-config-policy/local-disk-config-ref *# commit-buffer |
Commits the transaction to the system configuration. |
This example shows how to manually configure disks for a disk group configuration policy.
UCS-A# scope org UCS-A /org # enter disk-group-config-policy raid5policy UCS-A /org/disk-group-config-policy* # create local-disk-config-ref 1 UCS-A /org/disk-group-config-policy/local-disk-config-ref *# set role ded-hot-spare UCS-A /org/disk-group-config-policy/local-disk-config-ref* # set span-id 1 UCS-A /org/disk-group-config-policy/local-disk-config-ref *# commit-buffer
Configure Virtual Drive Properties.
Configuring Virtual Drive Properties
All virtual drives in a disk group must be managed by using a single disk group policy.
If you try to associate to a server that does not support these properties, a configuration error will be generated.
For the LSI MegaRAID SAS 2208 ROMB controller, these properties are supported only in the B420-M3 blade server. For the other controllers, these properties are supported in multiple rack servers.
Command or Action | Purpose | |
---|---|---|
Step 1 | UCS-A# scope org org-name |
Enters the organization mode for the specified organization. To enter the root organization mode, enter / as the org-name. |
Step 2 | UCS-A /org# scope disk-group-config-policy disk-group-name |
Enters disk group configuration policy mode for the specified disk group name. |
Step 3 | UCS-A /org/disk-group-config-policy* # create virtual-drive-def |
Creates a virtual drive definition and enters the virtual drive definition mode. |
Step 4 | UCS-A /org/disk-group-config-policy/virtual-drive-def* # set access-policy policy-type |
|
Step 5 | UCS-A /org/disk-group-config-policy/virtual-drive-def* # set drive-cache state |
|
Step 6 | UCS-A /org/disk-group-config-policy/virtual-drive-def* # set io-policy policy-type |
|
Step 7 | UCS-A /org/disk-group-config-policy/virtual-drive-def* # set read-policy policy-type |
|
Step 8 | UCS-A /org/disk-group-config-policy/virtual-drive-def* # set strip-size strip-size |
Specifies the strip size. This can be one of the following: |
Step 9 | UCS-A /org/disk-group-config-policy/virtual-drive-def* # set write-cache-policy policy-type |
|
Step 10 | UCS-A /org/disk-group-config-policy/virtual-drive-def* # commit-buffer |
Commits the transaction to the system configuration. |
Step 11 | UCS-A /org/disk-group-config-policy/virtual-drive-def* # show |
Displays the configured virtual drive properties. |
This example shows how to configure virtual disk properties:
UCS-A# scope org UCS-A /org # scope disk-group-config-policy raid0policy UCS-A /org/disk-group-config-policy # create virtual-drive-def UCS-A /org/disk-group-config-policy/virtual-drive-def* # set access-policy read-write UCS-A /org/disk-group-config-policy/virtual-drive-def* # set drive-cache enable UCS-A /org/disk-group-config-policy/virtual-drive-def* # set io-policy cached UCS-A /org/disk-group-config-policy/virtual-drive-def* # set read-policy normal UCS-A /org/disk-group-config-policy/virtual-drive-def* # set strip-size 1024 UCS-A /org/disk-group-config-policy/virtual-drive-def* # set write-cache-policy write-through UCS-A /org/disk-group-config-policy/virtual-drive-def* # commit-buffer UCS-A /org/disk-group-config-policy/virtual-drive-def # show Virtual Drive Def: Strip Size (KB): 1024KB Access Policy: Read Write Read Policy: Normal Configured Write Cache Policy: Write Through IO Policy: Cached Drive Cache: Enable UCS-A /org/disk-group-config-policy/virtual-drive-def #
Create a Storage Profile
Creating a Storage Profile
You can create a storage profile at the org level and at the service-profile level.
Command or Action | Purpose | |
---|---|---|
Step 1 | UCS-A# scope org org-name |
Enters the organization mode for the specified organization. To enter the root organization mode, enter / as the org-name. |
Step 2 | UCS-A /org # create storage-profile storage-profile-name |
Creates a storage profile with the specified name at the org level and enters storage-profile configuration mode. |
Step 3 | UCS-A /org/storage-profile* # commit-buffer |
Commits the transaction to the system configuration. |
Step 4 | UCS-A /org* # enter service-profile service-profile-name | (Optional)
Enters the specified service profile. |
Step 5 | UCS-A /org/service-profile* # create storage-profile-def | (Optional)
Creates a storage profile at the service-profile level. |
Step 6 | UCS-A /org/service-profile/storage-profile-def* # commit-buffer |
Commits the transaction to the system configuration. |
This example shows how to create a storage profile at the org level.
UCS-A# scope org UCS-A /org # create storage-profile stp2 UCS-A /org/storage-profile* # commit-buffer
This example shows how to create a storage profile at the service-profile level.
UCS-A# scope org UCS-A /org* # enter service-profile sp1 UCS-A /org/service-profile* # create storage-profile-def UCS-A /org/service-profile/storage-profile-def* # commit-buffer
Create Local LUNs
Deleting a Storage Profile
You can delete a storage profile that was created at the org level or at the service-profile level.
Command or Action | Purpose | |
---|---|---|
Step 1 | UCS-A# scope org org-name |
Enters the organization mode for the specified organization. To enter the root organization mode, enter / as the org-name. |
Step 2 | UCS-A /org # delete storage-profile storage-profile-name |
Deletes the storage profile with the specified name at the org level. |
Step 3 | UCS-A /org # scope service-profile service-profile-name | (Optional)
Enters the specified service profile. |
Step 4 | UCS-A /org/service-profile # delete storage-profile-def | (Optional)
Deletes the dedicated storage profile at the service-profile level. |
This example shows how to delete a storage profile at the org level.
UCS-A # scope org UCS-A /org # delete storage-profile stor1
This example shows how to delete a storage profile at the service-profile level.
UCS-A # scope org UCS-A /org # scope service-profile sp1 UCS-A /org/service-profile # delete storage-profile-def
Creating a Storage Profile PCH Controller Definition
You can create a PCH controller definition under a storage profile at the org level or at the service profile level.
Command or Action | Purpose | |||
---|---|---|---|---|
Step 1 | UCS-A# scope org org-name |
Enters the organization mode for the specified organization. To enter the root organization mode, enter / as the org-name.
| ||
Step 2 | UCS-A /org # scope storage-profile storage-profile-name |
Enters storage-profile configuration mode for the selected storage profile. | ||
Step 3 | UCS-A /org/storage-profile # create controller-def controller-definition-name |
Creates a PCH controller definition with the specified name and enters controller-definition configuration mode. | ||
Step 4 | UCS-A /org/storage-profile/controller-def* # create controller-mode-config |
Creates a PCH controller configuration and enters controller-mode configuration mode. | ||
Step 5 | UCS-A /org/storage-profile/controller-def/controller-mode-config* # set protect-config {yes|no} |
Specifies whether the server retains the configuration in the PCH controller even if the server is disassociated from the service profile. | ||
Step 6 | UCS-A /org/storage-profile/controller-def/controller-mode-config* # set raid-mode {any-configuration | no-local-storage | no-raid | raid-0-striped | raid-1-mirrored | raid-5-striped-parity | raid-50--striped-parity-and-striped | raid-6-striped-dual-parity | raid-60-striped-dual-parity-and-striped | raid-10-mirrored-and-striped} |
Specifies the raid mode for the PCH controller. | ||
Step 7 | UCS-A /org/storage-profile/controller-def/controller-mode-config* # commit-buffer |
Commits the transaction to the system configuration. |
This example shows how to add a PCH controller definition called "raid1-controller" with raid mode set to RAID 1 Mirrored to the org-level storage profile named "storage-profile-A".
UCS-A# scope org / UCS-A /org # scope storage-profile storage-profile-A UCS-A /org/storage-profile # create controller-def raid1-controller UCS-A /org/storage-profile/controller-def* # create controller-mode-config UCS-A /org/storage-profile/controller-def/controller-mode-config* # set protect-config yes UCS-A /org/storage-profile/controller-def/controller-mode-config* # set raid-mode raid-1-mirrored UCS-A /org/storage-profile/controller-def/controller-mode-config* # commit buffer
This example shows how to scope to the service profile called "Service-Profile1", create a storage profile, then create a PCH controller definition called "Raid60Ctrlr" within that storage profile. The controller definition has protection mode off and uses RAID 60 Striped Dual Parity and Striped.
UCS-A /org/service-profile # scope org / UCS-A /org # scope service-profile Service-Profile1 UCS-A /org/service-profile # create storage-profile-def UCS-A /org/service-profile/storage-profile-def* # create controller-def Raid60Ctrlr UCS-A /org/service-profile/storage-profile-def/controller-def* # create controller-mode-config UCS-A /org/service-profile/storage-profile-def/controller-def/controller-mode-config* # set protect-config no UCS-A /org/service-profile/storage-profile-def/controller-def/controller-mode-config* # set raid-mode raid-60-striped-dual-parity-and-striped UCS-A /org/service-profile/storage-profile-def/controller-def/controller-mode-config* # commit-buffer
Deleting a Storage Profile PCH Controller Definition
Command or Action | Purpose | |||
---|---|---|---|---|
Step 1 | UCS-A# scope org org-name |
Enters the organization mode for the specified organization. To enter the root organization mode, enter / as the org-name.
| ||
Step 2 | UCS-A /org # scope storage-profile storage-profile-name |
Enters storage-profile configuration mode for the selected storage profile. | ||
Step 3 | UCS-A /org/storage-profile # delete controller-def controller-definition-name |
Deletes a PCH controller definition with the specified name. | ||
Step 4 | UCS-A /org/storage-profile* # commit-buffer |
Commits the transaction to the system configuration. |
This example shows how to delete a PCH controller definition called "raid1-controller" from the org-level storage profile named "storage-profile-A".
UCS-A# scope org UCS-A /org # scope storage-profile storage-profile-A UCS-A /org/storage-profile # delete controller-def raid1-controller UCS-A /org/storage-profile* # commit-buffer
Creating Local LUNs
You can create local LUNs within a storage profile at the org level and within a dedicated storage profile at the service-profile level.
Command or Action | Purpose | |||
---|---|---|---|---|
Step 1 | UCS-A# scope org org-name |
Enters the organization mode for the specified organization. To enter the root organization mode, enter / as the org-name. | ||
Step 2 | UCS-A /org # enter storage-profile storage-profile-name |
Enters storage-profile mode for the specified storage profile. | ||
Step 3 | UCS-A /org/storage-profile* # create local-lun lun-name |
Creates a local LUN with the specified name. | ||
Step 4 | UCS-A /org/storage-profile/local-lun* # set auto-deploy {auto-deploy | no-auto-deploy} |
Specifies whether the LUN should be auto-deployed or not. | ||
Step 5 | UCS-A /org/storage-profile/local-lun* # set disk-policy-name disk-policy-name |
Specifies the name of the disk policy name for this LUN. | ||
Step 6 | UCS-A /org/storage-profile/local-lun* # set expand-to-avail {no | yes} |
Specifies whether the LUN should be expanded to the entire available disk group. For each service profile, only one LUN can be configured to use this option. | ||
Step 7 | UCS-A /org/storage-profile/local-lun* # set size size |
Specifies the size of this LUN in GB. The size can range from 1 GB to 10240 GB.
| ||
Step 8 | UCS-A /org/storage-profile/local-lun* # commit-buffer |
Commits the transaction to the system configuration. |
This example shows how to configure a local LUN within a storage profile at the org level.
UCS-A# scope org UCS-A /org # enter storage-profile stp2 UCS-A /org/storage-profile* # create local-lun lun2 UCS-A /org/storage-profile/local-lun* # set auto-deploy no-auto-deploy UCS-A /org/storage-profile/local-lun* # set disk-policy-name dpn2 UCS-A /org/storage-profile/local-lun* # set expand-to-avail yes UCS-A /org/storage-profile/local-lun* # set size 1000 UCS-A /org/storage-profile/local-lun* # commit-buffer
This example shows how to configure a local LUN within a dedicated storage profile at the service-profile level.
UCS-A# scope org UCS-A /org # enter service-profile sp1 UCS-A /org/service-profile* # enter storage-profile-def UCS-A /org/service-profile/storage-profile-def # create local-lun lun1 UCS-A /org/service-profile/storage-profile-def/local-lun* # set auto-deploy no-auto-deploy UCS-A /org/service-profile/storage-profile-def/local-lun* # set disk-policy-name dpn1 UCS-A /org/service-profile/storage-profile-def/local-lun* # set expand-to-avail yes UCS-A /org/service-profile/storage-profile-def/local-lun* # set size 1000 UCS-A /org/service-profile/storage-profile-def/local-lun* # commit-buffer
Associate a Storage Profile with a Service Profile
Deleting Local LUNs In a Storage Profile
When a LUN is deleted, the corresponding virtual drive is marked as orphan after the virtual drive reference is removed from the server.
Command or Action | Purpose | |
---|---|---|
Step 1 | UCS-A# scope org org-name |
Enters the organization mode for the specified organization. To enter the root organization mode, enter / as the org-name. |
Step 2 | UCS-A /org # enter storage-profile storage-profile-name |
Enters storage-profile mode for the specified storage profile. |
Step 3 | UCS-A /org/storage-profile* # show local-lun | (Optional)
Displays the local LUNs in the specified storage profile. |
Step 4 | UCS-A /org/storage-profile* # delete local-lun lun-name |
Deletes the specified LUN. |
Step 5 | UCS-A /org/storage-profile* # commit-buffer |
Commits the transaction to the system configuration. |
This example shows how to delete a LUN in a storage profile.
UCS-A # scope org UCS-A /org # enter storage-profile stp2 UCS-A /org/storage-profile # show local-lun Local SCSI LUN: LUN Name Size (GB) Order Disk Policy Name Auto Deploy ---------- ----------- ---------------- ---------------- ----------- luna 1 2 raid0 Auto Deploy lunb 1 1 raid0 Auto Deploy UCS-A /org/storage-profile # delete local-lun luna UCS-A /org/storage-profile* # commit-buffer UCS-A /org/storage-profile* # show local-lun Local SCSI LUN: LUN Name Size (GB) Order Disk Policy Name Auto Deploy ---------- ----------- ---------------- ---------------- ----------- lunb 1 1 raid0 Auto Deploy
Associating a Storage Profile with a Service Profile
A storage profile created under org can be referred by multiple service profiles, and a name reference in service profile is needed to associate the storage profile with a service profile.
Storage profiles can be defined under org and under service profile (dedicated). Hence, a service profile inherits local LUNs from both possible storage profiles. A service profile can have a maximum of two such local LUNs.
Command or Action | Purpose | |||
---|---|---|---|---|
Step 1 | UCS-A# scope org org-name |
Enters the organization mode for the specified organization. To enter the root organization mode, enter / as the org-name. | ||
Step 2 | UCS-A /org # scope service-profile service-profile-name |
Enters the specified service profile mode. | ||
Step 3 | UCS-A /org/service-profile # set storage-profile-name storage-profile-name |
Associates the specified storage profile with the service profile.
| ||
Step 4 | UCS-A /org/service-profile* # commit-buffer |
Commits the transaction to the system configuration. |
This example shows how to associate a storage profile with a service profile.
UCS-A# scope org UCS-A /org # scope service-profile sp1 UCS-A /org/service-profile # set storage-profile-name stp2
This example shows how to dissociate a service profile from a storage profile.
UCS-A# scope org UCS-A /org # scope service-profile sp1 UCS-A /org/service-profile # set storage-profile-name ""
Displaying Details of All Local LUNs Inherited By a Service Profile
Storage profiles can be defined under org and as a dedicated storage profile under service profile. Thus, a service profile inherits local LUNs from both possible storage profiles. It can have a maximum of 2 such local LUNs. You can display the details of all local LUNs inherited by a service profile by using the following command:
Command or Action | Purpose | |
---|---|---|
Step 1 | UCS-A /org/service-profile # show local-lun-ref |
|
UCS-A /org/service-profile # show local-lun-ref Local LUN Ref: Profile LUN Name Admin State RAID Level Provisioned Size (GB) Assigned Size (MB) Config State Referenced Lun Deploy Name ID Drive State ---------------- ----------- ---------------------- ---------------------- -------------------- ------------ -------------- ----------- ----------- ----------- luna Online RAID 0 Striped 1 1024 Applied luna-1 luna-1 1003 Optimal lunb Online RAID 0 Striped 1 1024 Applied lunb-1 lunb-1 1004 Optimal UCS-A /org/service-profile # Local LUN Ref: Name Admin State RAID Level Provisioned Size (GB) Assigned Size (MB) Config State Referenced Lun Deploy Name ID Drive State ---------------- ----------- ---------------------- ---------------------- -------------------- ------------ -------------- ----------- ----------- ----------- lun111 Online RAID 0 Striped 30 30720 Applied lun111-1 lun111-1 1001 Optimal lun201 Online Unspecified 1 0 Not Applied
Importing Foreign Configurations for a RAID Controller on a Blade Server
Command or Action | Purpose | |
---|---|---|
Step 1 | UCS-A# scope server [chassis-num/server-num | dynamic-uuid] |
Enters server mode for the specified server. |
Step 2 | UCS-A /chassis/server # scope raid-controller raid-contr-id {sas | sata} |
Enters RAID controller mode. |
Step 3 | UCS-A /chassis/server/raid-controller # set admin-state import-foreign-configuration |
Allows import of configurations from local disks that are in the Foreign Configuration state. |
This example shows how to import foreign configurations from local disks that are in the Foreign Configuration state:
UCS-A# scope server 1/3 UCS-A /chassis/server # scope raid-controller 1 sas UCS-A /chassis/server/raid-controller # set admin-state import-foreign-configuration UCS-A /chassis/server/raid-controller* #
Importing Foreign Configurations for a RAID Controller on a Rack Server
Command or Action | Purpose | |
---|---|---|
Step 1 | UCS-A # scope server server-id |
Enters server mode for the specified server. |
Step 2 | UCS-A /server # scope raid-controller raid-contr-id {sas | sata} |
Enters RAID controller mode. |
Step 3 | UCS-A /server/raid-controller # set admin-state import-foreign-configuration |
Allows import of configurations from local disks that are in the Foreign Configuration state. |
This example shows how to import foreign configurations from local disks that are in the Foreign Configuration state:
UCS-A# scope server 1 UCS-A /server # scope raid-controller 1 sas UCS-A /server/raid-controller # set admin-state import-foreign-configuration UCS-A /server/raid-controller* #
Configuring Local Disk Operations on a Blade Server
Command or Action | Purpose | |
---|---|---|
Step 1 | UCS-A# scope server [chassis-num/server-num | dynamic-uuid] |
Enters server mode for the specified server. |
Step 2 | UCS-A /chassis/server # scope raid-controller raid-contr-id {sas | sata} |
Enters RAID controller mode. |
Step 3 | UCS-A /chassis/server/raid-controller # scope local-disk local-disk-id |
Enters local disk configuration mode. |
Step 4 | UCS-A /chassis/server/raid-controller/local-disk # set admin-state {clear-foreign-configuration | dedicated-hot-spare [admin-vd-id] | prepare-for-removal | remove-hot-spare | unconfigured-good | undo-prepare-for-removal} |
|
This example shows how to clear any foreign configuration from a local disk:
UCS-A /chassis/server/raid-controller/local-disk # set admin-state clear-foreign-configuration
This example shows how to specify a local disk as a dedicated hot spare:
UCS-A /chassis/server/raid-controller/local-disk* # set admin-state dedicated-hot-spare 1001
This example shows how to specify that a local disk is marked for removal from the chassis:
UCS-A /chassis/server/raid-controller/local-disk* # set admin-state prepare-for-removal
This example shows how to specify that a local disk is marked for removal as a hot spare:
UCS-A /chassis/server/raid-controller/local-disk* # set admin-state remove-hot-spare
This example shows how to specify that a local disk is working, but is unconfigured for use:
UCS-A /chassis/server/raid-controller/local-disk* # set admin-state unconfigured-good
This example shows how to specify that a local disk is no longer marked for removal from the chassis:
UCS-A /chassis/server/raid-controller/local-disk* # set admin-state undo-prepare-for-removal
Configuring Local Disk Operations on a Rack Server
Command or Action | Purpose | |
---|---|---|
Step 1 | UCS-A # scope server server-id |
Enters server mode for the specified server. |
Step 2 | UCS-A /server # scope raid-controller raid-contr-id {sas | sata} |
Enters RAID controller mode. |
Step 3 | UCS-A /server/raid-controller # scope local-disk local-disk-id |
Enters local disk configuration mode. |
Step 4 | UCS-A /server/raid-controller/local-disk # set admin-state {clear-foreign-configuration | dedicated-hot-spare [admin-vd-id] | prepare-for-removal | remove-hot-spare | unconfigured-good | undo-prepare-for-removal} |
|
This example shows how to clear any foreign configuration from a local disk:
UCS-A /server/raid-controller/local-disk # set admin-state clear-foreign-configuration
This example shows how to specify a local disk as a dedicated hot spare:
UCS-A /server/raid-controller/local-disk* # set admin-state dedicated-hot-spare 1001
This example shows how to specify that a local disk is marked for removal:
UCS-A /server/raid-controller/local-disk* # set admin-state prepare-for-removal
This example shows how to specify that a local disk is marked for removal as a hot spare:
UCS-A /server/raid-controller/local-disk* # set admin-state remove-hot-spare
This example shows how to specify that a local disk is working, but is unconfigured for use:
UCS-A /server/raid-controller/local-disk* # set admin-state unconfigured-good
This example shows how to specify that a local disk is no longer marked for removal:
UCS-A /server/raid-controller/local-disk* # set admin-state undo-prepare-for-removal
Configuring Virtual Drive Operations
The following operations can be performed only on orphaned virtual drives:
- Deleting an Orphaned Virtual Drive on a Blade Server
- Deleting an Orphaned Virtual Drive on a Rack Server
- Renaming an Orphaned Virtual Drive on a Blade Server
- Renaming an Orphaned Virtual Drive on a Rack Server
Deleting an Orphaned Virtual Drive on a Blade Server
Command or Action | Purpose | |
---|---|---|
Step 1 | UCS-A# scope server [chassis-num/server-num | dynamic-uuid] |
Enters server mode for the specified server. |
Step 2 | UCS-A /chassis/server # scope raid-controller raid-contr-id {sas | sata} |
Enters RAID controller chassis mode. |
Step 3 | UCS-A /chassis/server/raid-controller # delete virtual-drive id virtual-drive-id | (Optional)
Deletes the orphaned virtual drive with the specified virtual drive ID. |
Step 4 | UCS-A /chassis/server/raid-controller # delete virtual-drive name virtual-drive-name | (Optional)
Deletes the orphaned virtual drive with the specified virtual drive name. |
Step 5 | UCS-A /chassis/server/raid-controller # scope virtual-drive virtual-drive-id | (Optional)
Enters virtual drive mode for the specified orphaned virtual drive. |
Step 6 | UCS-A /chassis/server/raid-controller/virtual-drive # set admin-state delete |
Deletes the orphaned virtual drive. |
Step 7 | UCS-A /chassis/server/raid-controller/virtual-drive # commit-buffer |
Commits the transaction to the system configuration. |
This example shows how to delete an orphan virtual drive by specifying the virtual drive ID.
UCS-A# scope server 1/3 UCS-A /chassis/server # scope raid-controller 1 sas UCS-A /chassis/server/raid-controller # show virtual-drive Virtual Drive: ID: 1001 Name: lun111-1 Block Size: 512 Blocks: 62914560 Size (MB): 30720 Operability: Operable Presence: Equipped Oper Device ID: 0 Change Qualifier: No Change Config State: Applied Deploy Action: No Action ID: 1002 Name: luna-1 Block Size: 512 Blocks: 2097152 Size (MB): 1024 Operability: Operable Presence: Equipped Oper Device ID: 1 Change Qualifier: No Change Config State: Orphaned Deploy Action: No Action ID: 1003 Name: lunb-1 Block Size: 512 Blocks: 2097152 Size (MB): 1024 Operability: Operable Presence: Equipped Oper Device ID: 2 Change Qualifier: No Change Config State: Orphaned Deploy Action: No Action ID: 1004 Name: lunb-2 Block Size: 512 Blocks: 2097152 Size (MB): 1024 Operability: Operable Presence: Equipped Oper Device ID: 3 Change Qualifier: No Change Config State: Orphaned Deploy Action: No Action ID: 1005 Name: luna-2 Block Size: 512 Blocks: 2097152 Size (MB): 1024 Operability: Operable Presence: Equipped Oper Device ID: 4 Change Qualifier: No Change Config State: Orphaned Deploy Action: No Action ... UCS-A /chassis/server/raid-controller # delete virtual-drive id 1002 Warning: When committed, the virtual drive will be deleted, which may result in data loss. UCS-A /chassis/server/raid-controller # commit-buffer
This example shows how to delete an orphan virtual drive by specifying the virtual drive name.
UCS-A# scope server 1/3 UCS-A /chassis/server # scope raid-controller 1 sas UCS-A /chassis/server/raid-controller # show virtual-drive Virtual Drive: ID: 1001 Name: lun111-1 Block Size: 512 Blocks: 62914560 Size (MB): 30720 Operability: Operable Presence: Equipped Oper Device ID: 0 Change Qualifier: No Change Config State: Applied Deploy Action: No Action ID: 1003 Name: lunb-1 Block Size: 512 Blocks: 2097152 Size (MB): 1024 Operability: Operable Presence: Equipped Oper Device ID: 2 Change Qualifier: No Change Config State: Orphaned Deploy Action: No Action ID: 1004 Name: lunb-2 Block Size: 512 Blocks: 2097152 Size (MB): 1024 Operability: Operable Presence: Equipped Oper Device ID: 3 Change Qualifier: No Change Config State: Orphaned Deploy Action: No Action ID: 1005 Name: luna-2 Block Size: 512 Blocks: 2097152 Size (MB): 1024 Operability: Operable Presence: Equipped Oper Device ID: 4 Change Qualifier: No Change Config State: Orphaned Deploy Action: No Action ... UCS-A /chassis/server/raid-controller # delete virtual-drive name lunb-1 Warning: When committed, the virtual drive will be deleted, which may result in data loss. UCS-A /chassis/server/raid-controller # commit-buffer
This example shows how to delete an orphan virtual drive by setting the admin-state.
UCS-A# scope server 1/3 UCS-A /chassis/server # scope raid-controller 1 sas UCS-A /chassis/server/raid-controller # scope virtual-drive 1004 UCS-A /chassis/server/raid-controller/virtual-drive # set admin-state delete Warning: When committed, the virtual drive will be deleted, which may result in data loss. UCS-A /chassis/server/raid-controller/virtual-drive # commit-buffer
Deleting an Orphaned Virtual Drive on a Rack Server
Command or Action | Purpose | |
---|---|---|
Step 1 | UCS-A # scope server server-id |
Enters server mode for the specified server. |
Step 2 | UCS-A /server # scope raid-controller raid-contr-id {sas | sata} |
Enters RAID controller mode. |
Step 3 | UCS-A /server/raid-controller # delete virtual-drive id virtual-drive-id | (Optional)
Deletes the orphaned virtual drive with the specified virtual drive ID. |
Step 4 | UCS-A /server/raid-controller # delete virtual-drive name virtual-drive-name | (Optional)
Deletes the orphaned virtual drive with the specified virtual drive name. |
Step 5 | UCS-A /server/raid-controller # scope virtual-drive virtual-drive-id | (Optional)
Enters virtual drive mode for the specified orphaned virtual drive. |
Step 6 | UCS-A /server/raid-controller/virtual-drive # set admin-state delete |
Deletes the orphaned virtual drive. |
Step 7 | UCS-A /server/raid-controller/virtual-drive # commit-buffer |
Commits the transaction to the system configuration. |
This example shows how to delete an orphan virtual drive by specifying the virtual drive ID.
UCS-A# scope server 1 UCS-A /server # scope raid-controller 1 sas UCS-A /server/raid-controller # show virtual-drive Virtual Drive: ID: 1001 Name: lun111-1 Block Size: 512 Blocks: 62914560 Size (MB): 30720 Operability: Operable Presence: Equipped Oper Device ID: 0 Change Qualifier: No Change Config State: Applied Deploy Action: No Action ID: 1002 Name: luna-1 Block Size: 512 Blocks: 2097152 Size (MB): 1024 Operability: Operable Presence: Equipped Oper Device ID: 1 Change Qualifier: No Change Config State: Orphaned Deploy Action: No Action ID: 1003 Name: lunb-1 Block Size: 512 Blocks: 2097152 Size (MB): 1024 Operability: Operable Presence: Equipped Oper Device ID: 2 Change Qualifier: No Change Config State: Orphaned Deploy Action: No Action ID: 1004 Name: lunb-2 Block Size: 512 Blocks: 2097152 Size (MB): 1024 Operability: Operable Presence: Equipped Oper Device ID: 3 Change Qualifier: No Change Config State: Orphaned Deploy Action: No Action ID: 1005 Name: luna-2 Block Size: 512 Blocks: 2097152 Size (MB): 1024 Operability: Operable Presence: Equipped Oper Device ID: 4 Change Qualifier: No Change Config State: Orphaned Deploy Action: No Action ... UCS-A /server/raid-controller # delete virtual-drive id 1002 Warning: When committed, the virtual drive will be deleted, which may result in data loss. UCS-A /server/raid-controller # commit-buffer
This example shows how to delete an orphan virtual drive by specifying the virtual drive name.
UCS-A# scope server 1 UCS-A /server # scope raid-controller 1 sas UCS-A /server/raid-controller # show virtual-drive Virtual Drive: ID: 1001 Name: lun111-1 Block Size: 512 Blocks: 62914560 Size (MB): 30720 Operability: Operable Presence: Equipped Oper Device ID: 0 Change Qualifier: No Change Config State: Applied Deploy Action: No Action ID: 1003 Name: lunb-1 Block Size: 512 Blocks: 2097152 Size (MB): 1024 Operability: Operable Presence: Equipped Oper Device ID: 2 Change Qualifier: No Change Config State: Orphaned Deploy Action: No Action ID: 1004 Name: lunb-2 Block Size: 512 Blocks: 2097152 Size (MB): 1024 Operability: Operable Presence: Equipped Oper Device ID: 3 Change Qualifier: No Change Config State: Orphaned Deploy Action: No Action ID: 1005 Name: luna-2 Block Size: 512 Blocks: 2097152 Size (MB): 1024 Operability: Operable Presence: Equipped Oper Device ID: 4 Change Qualifier: No Change Config State: Orphaned Deploy Action: No Action ... UCS-A /server/raid-controller # delete virtual-drive name lunb-1 Warning: When committed, the virtual drive will be deleted, which may result in data loss. UCS-A /server/raid-controller # commit-buffer
This example shows how to delete an orphan virtual drive by setting the admin-state.
UCS-A# scope server 1 UCS-A /server # scope raid-controller 1 sas UCS-A /server/raid-controller # scope virtual-drive 1004 UCS-A /server/raid-controller/virtual-drive # set admin-state delete Warning: When committed, the virtual drive will be deleted, which may result in data loss. UCS-A /server/raid-controller/virtual-drive # commit-buffer
Renaming an Orphaned Virtual Drive on a Blade Server
Command or Action | Purpose | |
---|---|---|
Step 1 | UCS-A# scope server [chassis-num/server-num | dynamic-uuid] |
Enters server mode for the specified server. |
Step 2 | UCS-A /chassis/server # scope raid-controller raid-contr-id {sas | sata} |
Enters RAID controller chassis mode. |
Step 3 | UCS-A /chassis/server/raid-controller # scope virtual-drive virtual-drive-id |
Enters virtual drive mode for the specified virtual drive. |
Step 4 | UCS-A /chassis/server/raid-controller/virtual-drive # set name virtual-drive-name |
Specifies a name for the orphan virtual drive. |
Step 5 | UCS-A /chassis/server/raid-controller/virtual-drive # commit-buffer |
Commits the transaction to the system configuration. |
This example shows how to specify a name for an orphan virtual drive.
UCS-A /chassis/server # scope raid-controller 1 sas UCS-A /chassis/server/raid-controller # scope virtual-drive 1060 UCS-A /chassis/server/raid-controller/virtual-drive # set name vd1 UCS-A /chassis/server/raid-controller/virtual-drive # commit-buffer
Renaming an Orphaned Virtual Drive on a Rack Server
Command or Action | Purpose | |
---|---|---|
Step 1 | UCS-A # scope server server-id |
Enters server mode for the specified server. |
Step 2 | UCS-A /server # scope raid-controller raid-contr-id {sas | sata} |
Enters RAID controller mode. |
Step 3 | UCS-A /server/raid-controller # scope virtual-drive virtual-drive-id |
Enters virtual drive mode for the specified virtual drive. |
Step 4 | UCS-A /server/raid-controller/virtual-drive # set name virtual-drive-name |
Specifies a name for the orphan virtual drive. |
Step 5 | UCS-A /server/raid-controller/virtual-drive # commit-buffer |
Commits the transaction to the system configuration. |
This example shows how to specify a name for an orphan virtual drive.
UCS-A /server # scope raid-controller 1 sas UCS-A /server/raid-controller # scope virtual-drive 1060 UCS-A /server/raid-controller/virtual-drive # set name vd1 UCS-A /server/raid-controller/virtual-drive # commit-buffer
Boot Policy for Local Storage
You can specify the primary boot device for a storage controller as a local LUN or a JBOD disk. Each storage controller can have one primary boot device. However, in a storage profile, you can set only one device as the primary boot LUN.
Configuring the Boot Policy for a Local LUN
Command or Action | Purpose | |
---|---|---|
Step 1 | UCS-A# scope org org-name |
Enters organization mode for the specified organization. To enter the root organization mode, type / as the org-name . |
Step 2 | UCS-A /org # scope boot-policy policy-name |
Enters organization boot policy mode for the specified boot policy. |
Step 3 | UCS-A /org/boot-policy # create storage |
Creates a storage boot for the boot policy and enters organization boot policy storage mode. |
Step 4 | UCS-A /org/boot-policy/storage # create local |
Creates a local storage location and enters the boot policy local storage mode. |
Step 5 | UCS-A /org/boot-policy/storage/local/ # create local-lun |
Specifies a local hard disk drive as the local storage. |
Step 6 | UCS-A /org/boot-policy/storage/local/local-lun # create local-lun-image-path {primary | secondary} |
Specifies the boot order for the LUN that you specify. Cisco UCS Manager Release 2.2(4) does not support secondary boot order. |
Step 7 | UCS-A /org/boot-policy/storage/local/local-lun/local-lun-image-path # set lunname lun_name |
Specifies the name of the LUN that you want to boot from. |
Step 8 | UCS-A /org/boot-policy/storage/local/local-storage-device # commit-buffer |
Commits the transaction to the system configuration. |
The following example shows how to create a boot policy named lab1-boot-policy, create a local hard disk drive boot for the policy, specify a boot order and a LUN to boot from, and commit the transaction:
UCS-A# scope org / UCS-A /org* # scope boot-policy lab1-boot-policy UCS-A /org/boot-policy* # create storage UCS-A /org/boot-policy/storage* # create local UCS-A /org/boot-policy/storage/local* # create local-lun UCS-A /org/boot-policy/storage/local/local-lun # create local-lun-image-path primary UCS-A /org/boot-policy/storage/local/local-lun/local-lun-image-path # set lunname luna UCS-A /org/boot-policy/storage/local/local-lun/local-lun-image-path # commit-buffer UCS-A /org/boot-policy/storage/local/local-lun/local-lun-image-path #
Include the boot policy in a service profile and template.
Configuring the Boot Policy for a Local JBOD Disk
Command or Action | Purpose | |
---|---|---|
Step 1 | UCS-A# scope org org-name |
Enters organization mode for the specified organization. To enter the root organization mode, type / as the org-name . |
Step 2 | UCS-A /org # scope boot-policy policy-name |
Enters organization boot policy mode for the specified boot policy. |
Step 3 | UCS-A /org/boot-policy # create storage |
Creates a storage boot for the boot policy and enters organization boot policy storage mode. |
Step 4 | UCS-A /org/boot-policy/storage # create local |
Creates a local storage location and enters the boot policy local storage mode. |
Step 5 | UCS-A /org/boot-policy/storage/local/ # create local-jbod |
Specifies a local JBOD disk as the local storage. |
Step 6 | UCS-A /org/boot-policy/storage/local/local-jbod # create local-disk-image-path {primary | secondary} |
Specifies the boot order for the local JBOD disk. Cisco UCS Manager Release 2.2(4) does not support secondary boot order. |
Step 7 | UCS-A /org/boot-policy/storage/local/local-jbod/local-disk-image-path # set slotnumber slot_number |
Specifies the slot number of the JBOD disk that you want to boot from. |
Step 8 | UCS-A /org/boot-policy/storage/local/local-jbod/local-disk-image-path # commit-buffer |
Commits the transaction to the system configuration. |
The following example shows how to create a boot policy named lab1-boot-policy, create a local hard disk drive boot for the policy, specify a boot order and a JBOD disk to boot from, and commit the transaction:
UCS-A# scope org / UCS-A /org* # scope boot-policy lab1-boot-policy UCS-A /org/boot-policy* # create storage UCS-A /org/boot-policy/storage* # create local UCS-A /org/boot-policy/storage/local* # create local-jbod UCS-A /org/boot-policy/storage/local/local-jbod # create local-disk-image-path primary UCS-A /org/boot-policy/storage/local/local-jbod/local-disk-image-path* # set slotnumber 5 UCS-A /org/boot-policy/storage/local/local-jbod/local-disk-image-path # commit-buffer UCS-A /org/boot-policy/storage/local/local-jbod/local-disk-image-path #
Include the boot policy in a service profile and template.
Local LUN Operations in a Service Profile
Note | Preprovisioning a LUN name, claiming an orphan LUN, and deploying or undeploying a LUN result in server reboot. |
- Preprovisioning a LUN Name or Claiming an Orphan LUN
- Deploying and Undeploying a LUN
- Renaming a Service Profile Referenced LUN
Preprovisioning a LUN Name or Claiming an Orphan LUN
You can preprovision a LUN name or claim an orphan LUN by using the set ref-name command. Preprovisioning a LUN name or claiming an orphan LUN can be done only when the admin state of the LUN is Undeployed. You can also manually change the admin state of the LUN to Undeployed and claim an orphan LUN.
If the LUN name is empty, set a LUN name before claiming it.
Command or Action | Purpose | |
---|---|---|
Step 1 | UCS-A# scope org org-name |
Enters the organization mode for the specified organization. To enter the root organization mode, enter / as the org-name. |
Step 2 | UCS-A /org# scope service-profile service-profile-name |
Enters the specified service profile mode. |
Step 3 | UCS-A /org/service-profile# enter local-lun-ref lun-name |
Enters the specified LUN. |
Step 4 | UCS-A /org/service-profile/local-lun-ref# set ref-name ref-lun-name |
Sets the referenced LUN name. If this LUN name exists and the LUN is orphaned, its is claimed by the service profile. If this LUN does not exist, a new LUN is created with the specified name. |
-
If the LUN exists and is not orphaned, a configuration failure occurs.
-
If a LUN is already referred to and the ref-name is changed, it will release the old LUN and will claim or create a LUN with the ref-name. The old LUN is marked as an orphan after the LUN reference is removed from the server.
This examples shows how to preprovision a LUN name.
UCS-A# scope org UCS-A /org # scope service-profile sp1 UCS-A /org/service-profile* # enter local-lun-ref lun1 UCS-A /org/service-profile/local-lun-ref* # set ref-name lun2
Deploying and Undeploying a LUN
You can deploy or undeploy a LUN by using the admin-state command. If the admin state of a local LUN is Undeployed, the reference of that LUN is removed and the LUN is not deployed.
Command or Action | Purpose | |
---|---|---|
Step 1 | UCS-A# scope org org-name |
Enters the organization mode for the specified organization. To enter the root organization mode, enter / as the org-name. |
Step 2 | UCS-A /org# scope service-profile service-profile-name |
Enters the specified service profile mode. |
Step 3 | UCS-A /org/service-profile# enter local-lun-ref lun-name |
Enters the specified LUN. |
Step 4 | UCS-A /org/service-profile/local-lun-ref# set admin-state {online | undeployed} |
Sets the admin state of the specified LUN to online or undeployed. If a LUN is already referred to and the admin state is set to undeployed, it will release the old LUN. The old LUN is marked as orphan after the LUN reference is removed from the server. |
This examples shows how to deploy a LUN.
UCS-A# scope org UCS-A /org # scope service-profile sp1 UCS-A /org/service-profile* # enter local-lun-ref lun1 UCS-A /org/service-profile/local-lun-ref* # set admin-state online
This examples shows how to undeploy a LUN.
UCS-A# scope org UCS-A /org # scope service-profile sp1 UCS-A /org/service-profile* # enter local-lun-ref lun1 UCS-A /org/service-profile/local-lun-ref* # set admin-state undeployed
Renaming a Service Profile Referenced LUN
Command or Action | Purpose | |
---|---|---|
Step 1 | UCS-A# scope org org-name |
Enters the organization mode for the specified organization. To enter the root organization mode, enter / as the org-name. |
Step 2 | UCS-A /org# scope service-profile service-profile-name |
Enters the specified service profile mode. |
Step 3 | UCS-A /org/service-profile# enter local-lun-ref lun-name |
Enters the specified LUN. |
Step 4 | UCS-A /org/service-profile/local-lun-ref# set name |
Renames the referenced LUN. |
This examples shows how to rename a LUN referenced by a service profile.
UCS-A# scope org UCS-A /org # scope service-profile sp1 UCS-A /org/service-profile* # enter local-lun-ref lun1 UCS-A /org/service-profile/local-lun-ref* # set name lun11
Viewing the Local Disk Locator LED State
The following example shows that the state of the local disk Locator LED is on:
USA-A# scope server 1 USA-A /server # scope local-disk 2 USA-A /serverlocal-disk # show locator-led Locator LED: Equipment Operational State ---------------- ----------------- 1/SAS-1/2 On
Turning On the Local Disk Locator LED
Step 1 | UCS-A#
scope server
id
Enters server mode for the specified server. |
Step 2 | UCS-A/server #
scope
local-disk
id
Enters the RAID controller for the specified local disk. |
Step 3 | UCS-A
/server/local-disk #
enable locator-led
Turns on the disk locator LED. |
Step 4 |
UCS-A/server/local-disk* #
commit-buffer
Commits the command to the system configuration. |
The following example displays how to turn on the local disk Locator LED:
UCS-A# scope server 1 UCS-A /server/raid-controller # scope local-disk 2 USA-A /server/raid-controller/local-disk # enable locator-led USA-A /server/raid-controller/local-disk* # commit-buffer
Turning Off the Local Disk Locator LED
Step 1 | UCS-A#
scope server
id
Enters server mode for the specified server. |
Step 2 | UCS-A/server #
scope
local-disk
id
Enters the RAID controller for the specified local disk. |
Step 3 |
UCS-A/server/local-disk #
disable locator-led
Turns off the disk locator LED. |
Step 4 |
UCS-A/server/raid-controller/local-disk* #
commit-buffer
Commits the command to the system configuration. |
The following example displays how to disable the local disk Locator LED:
UCS-A# server 1 UCS-A /server # scope local-disk 2 USA-A /server/local-disk # disable locator-led USA-A /server/local-disk* # commit-buffer