Design and Deployment of Cisco UCS Servers and MapR Converged Data Platform with SUSE Linux Enterprise Server
Last Updated: February 21, 2018
About the Cisco Validated Design Program
The Cisco Validated Design (CVD) program consists of systems and solutions designed, tested, and documented to facilitate faster, more reliable, and more predictable customer deployments. For more information, visit:
http://www.cisco.com/go/designzone.
ALL DESIGNS, SPECIFICATIONS, STATEMENTS, INFORMATION, AND RECOMMENDATIONS (COLLECTIVELY, "DESIGNS") IN THIS MANUAL ARE PRESENTED "AS IS," WITH ALL FAULTS. CISCO AND ITS SUPPLIERS DISCLAIM ALL WARRANTIES, INCLUDING, WITHOUT LIMITATION, THE WARRANTY OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT OR ARISING FROM A COURSE OF DEALING, USAGE, OR TRADE PRACTICE. IN NO EVENT SHALL CISCO OR ITS SUPPLIERS BE LIABLE FOR ANY INDIRECT, SPECIAL, CONSEQUENTIAL, OR INCIDENTAL DAMAGES, INCLUDING, WITHOUT LIMITATION, LOST PROFITS OR LOSS OR DAMAGE TO DATA ARISING OUT OF THE USE OR INABILITY TO USE THE DESIGNS, EVEN IF CISCO OR ITS SUPPLIERS HAVE BEEN ADVISED OF THE POSSIBILITY OF SUCH DAMAGES.
THE DESIGNS ARE SUBJECT TO CHANGE WITHOUT NOTICE. USERS ARE SOLELY RESPONSIBLE FOR THEIR APPLICATION OF THE DESIGNS. THE DESIGNS DO NOT CONSTITUTE THE TECHNICAL OR OTHER PROFESSIONAL ADVICE OF CISCO, ITS SUPPLIERS OR PARTNERS. USERS SHOULD CONSULT THEIR OWN TECHNICAL ADVISORS BEFORE IMPLEMENTING THE DESIGNS. RESULTS MAY VARY DEPENDING ON FACTORS NOT TESTED BY CISCO.
CCDE, CCENT, Cisco Eos, Cisco Lumin, Cisco Nexus, Cisco StadiumVision, Cisco TelePresence, Cisco WebEx, the Cisco logo, DCE, and Welcome to the Human Network are trademarks; Changing the Way We Work, Live, Play, and Learn and Cisco Store are service marks; and Access Registrar, Aironet, AsyncOS, Bringing the Meeting To You, Catalyst, CCDA, CCDP, CCIE, CCIP, CCNA, CCNP, CCSP, CCVP, Cisco, the Cisco Certified Internetwork Expert logo, Cisco IOS, Cisco Press, Cisco Systems, Cisco Systems Capital, the Cisco Systems logo, Cisco Unified Computing System (Cisco UCS), Cisco UCS B-Series Blade Servers, Cisco UCS C-Series Rack Servers, Cisco UCS S-Series Storage Servers, Cisco UCS Manager, Cisco UCS Management Software, Cisco Unified Fabric, Cisco Application Centric Infrastructure, Cisco Nexus 9000 Series, Cisco Nexus 7000 Series. Cisco Prime Data Center Network Manager, Cisco NX-OS Software, Cisco MDS Series, Cisco Unity, Collaboration Without Limitation, EtherFast, EtherSwitch, Event Center, Fast Step, Follow Me Browsing, FormShare, GigaDrive, HomeLink, Internet Quotient, IOS, iPhone, iQuick Study, LightStream, Linksys, MediaTone, MeetingPlace, MeetingPlace Chime Sound, MGX, Networkers, Networking Academy, Network Registrar, PCNow, PIX, PowerPanels, ProConnect, ScriptShare, SenderBase, SMARTnet, Spectrum Expert, StackWise, The Fastest Way to Increase Your Internet Quotient, TransPath, WebEx, and the WebEx logo are registered trademarks of Cisco Systems, Inc. and/or its affiliates in the United States and certain other countries.
All other trademarks mentioned in this document or website are the property of their respective owners. The use of the word partner does not imply a partnership relationship between Cisco and any other company. (0809R)
© 2018 Cisco Systems, Inc. All rights reserved.
Table of Contents
Cisco Unified Computing System
Cisco UCS 6332UP Fabric Interconnect
Cisco UCS 2304XP Fabric Extender
Cisco UCS B460 M4 Blade Server
Cisco UCS C460 M4 Rack Servers
Cisco UCS C240 M4 Rack Servers
Cisco I/O Adapters for Blade and Rack-Mount Servers
Cisco VIC 1380 Virtual Interface Card
Cisco VIC 1385 Virtual Interface Card
Cisco Unified Computing System Performance Manager
Hardware Requirements for the SAP HANA Database
Deployment Hardware and Software
Cisco Nexus 3500 Series Switch Network Configuration
Cisco Nexus 3524 Initial Configuration
Enable Appropriate Cisco Nexus 3524 Switch Features and Settings
Create Global Policy to Enable Jumbo Frame and Apply the Policy to System Wide
Create VLANs for SAP HANA Traffic
Configure Network Interfaces Connecting to Cisco UCS Fabric Interconnect
Cisco Nexus 9000 Series Switch Network Configuration
Cisco Nexus 9000 A Initial Configuration
Cisco Nexus 9000 B Initial Configuration
Enable Appropriate Cisco Nexus 9000 Series Switches─Features and Settings
Create VLANs for SAP HANA Traffic
Configure Virtual Port-Channel Domain
Configure Network Interfaces for the VPC Peer Links
Configure Network Interfaces with Cisco UCS Fabric Interconnect
(Optional) Configure Network Interfaces for SAP HANA Backup/Data Source/Replication
(Optional) Management Plane Access for Cisco UCS Servers
Uplink into Existing Network Infrastructure
Initial Setup of Cisco UCS 6332 Fabric Interconnect
Cisco UCS 6332 Fabric Interconnect A
Cisco UCS 6332 Fabric Interconnect B
Upgrade Cisco UCS Manager Software to Version 3.1(2b)
Add Block of IP Addresses for KVM Access
Cisco UCS Blade Chassis Connection Options
Enable Server and Uplink Ports
Acknowledge Cisco UCS Chassis and Rack-Mount Servers
Create Local Disk Configuration Policy
Update Default Maintenance Policy
Set Jumbo Frames in Cisco UCS Fabric
Create Network Control Policy for Internal Network
Create a vNIC Template for Storage Network
Create a vNIC Template for AppServer Network
Create a vNIC Template for Backup Network
Create a vNIC Template for Client Network
Create a vNIC Template for DataSource Network
Create a vNIC Template for Replication Network
Create a vNIC Template for Management Network
Create a vNIC Template for IPMI Inband Access Network
Create a vNIC Template for MapR-01 Internal Network
Create a vNIC template for MapR-02 Internal Network
Create a vNIC Template for MapR-03 Internal Network
Create IP Pool for Inband Management
Create Service Profile Templates for SAP HANA Scale-Out Servers
Create Service Profile from the Template
Create Service Profile Templates for MapR Servers
MapR Server RAID Configuration
MapR Server Operating System Installation
Operating System Network Configuration
SUSE Linux Enterprise Server 12 System Update and OS Customization for MapR Servers
SUSE Operating System Installation
Operating System Network Configuration
High Availability (HA) Configuration for Scale-Out
High Availability Configuration
Enable the SAP HANA Storage Connector API
Organizations in every industry are generating and using more data than ever before; from customer transactions and supplier delivery considerations to real-time user-consumption statistics. Without scalable infrastructure that can store, process, and analyze big data sets in real time, companies are unable to use this information to their advantage. The Cisco UCS Integrated Infrastructure for SAP HANA Scale-Out with the Cisco Unified Computing System™ (Cisco UCS) helps companies more easily harness information and make better business decisions that let them stay ahead of the competition. Our solutions help improve access to all of your data, accelerate business decision making with policy-based, simplified management; lower deployment risk; and reduce total cost of ownership (TCO). Our innovations give you the key to unlock the intelligence in your data and interpret it with a new dimension of context and insight to help you create a sustainable, competitive business advantage.
Cisco Validated Designs include systems and solutions that are designed, tested, and documented to facilitate and improve customer deployments. These designs incorporate a wide range of technologies and products into a portfolio of solutions have been developed to address the business needs of customers. The Cisco UCS Integrated Infrastructure for SAP HANA with MapR Converged Data Platform provides an end-to-end architecture that demonstrate support for multiple SAP HANA workloads with high availability and server redundancy. The solution consists of Cisco UCS B-series B460-M4 blade servers and or Cisco UCS C-series C460-M4 rack mount servers as compute nodes for SAP HANA and Cisco UCS C-Series C240-M4 rack mount servers as storage nodes. The next generation Cisco UCS Fabric Interconnect 6332 with 40 Gb Ethernet for Server Management and Storage Connectivity. Cisco UCS service profiles enable rapid and consistent server configuration, and automation simplifies ongoing system maintenance activities such as deploying firmware updates across the entire cluster as a single operation. Advanced monitoring capabilities raise alarms and send notifications about the health of the entire cluster so that you can proactively address concerns before they affect data analysis. The Storage Nodes are composed of Cisco UCS Servers with MapR Converged Data Platform, which is a modern NFS-mountable distributed file-system. MapR-FS is a complete POSIX file system that handles raw disk I/O for big data workload with direct access to storage hardware which dramatically improving performance and scale to thousands of nodes, and trillions of files, with extremely high throughput. MapR-FS includes enterprise-grade features such as block-level mirroring for mission-critical disaster recovery as well as load balancing, and consistent snapshots for easy data recovery.
Cisco UCS Integrated Infrastructure provides a pre-validated, ready-to-deploy infrastructure, which reduces the time and complexity involved in configuring and validating a traditional data center deployment. Cisco UCS Platforms is flexible, reliable and cost effective to facilitate various deployment options of the applications while being easily scalable and manageable. The reference architecture detailed in this document highlights the resiliency, cost benefit, and ease of deployment of a SAP HANA solution. This document describing the infrastructure installation and configuration to run SAP HANA on a dedicated or shared infrastructure.
SAP HANA is SAP SE’s implementation of in-memory database technology. The SAP HANA database takes advantage of the low cost main memory (RAM), data-processing capabilities of multicore processors, and faster data access to provide better performance for analytical and transactional applications. SAP HANA offers a multi-engine, query-processing environment that supports relational data (with both row- and column-oriented physical representations in a hybrid engine) as well as a graph and text processing for semi-structured and unstructured data management within the same system. As an appliance, SAP HANA combines software components from SAP optimized for certified hardware. However, this solution has a preconfigured hardware set-up and preinstalled software package that is dedicated for SAP HANA. In 2013, SAP introduced SAP HANA Tailored Datacenter Integration (TDI) option; TDI solution offers a more open and flexible way for integrating SAP HANA into the data center by reusing existing enterprise storage hardware, thereby reducing hardware costs. With the introduction of SAP HANA TDI for shared infrastructure, the Cisco UCS Integrated Infrastructure solution provides the advantage of having the compute, storage, and network stack integrated with the programmability of the Cisco Unified Computing System (Cisco UCS). SAP HANA TDI option enables organizations to run multiple SAP HANA production systems on a shared infrastructure. It also enables customers to run the SAP applications servers and SAP HANA database hosted on the same infrastructure.
For more information about SAP HANA, see the SAP Help Portal: http://help.sap.com/hana/.
The intended audience for this document includes, but is not limited to, sales engineers, field consultants, professional services, IT managers, partner engineering, and customers deploying the Cisco Integrated Infrastructure for SAP HANA with MapR Converged Data Platform. External references are provided wherever applicable, but readers are expected to be familiar with the technology, infrastructure, and database security policies of the customer installation.
This document describes the steps required to deploy and configure a Cisco Datacenter Solution for SAP HANA. Cisco’s validation provides further confirmation with regard to component compatibility, connectivity and correct operation of the entire integrated stack. This document showcases one of the variants of Cisco Integrated Infrastructure for SAP HANA. While readers of this document are expected to have sufficient knowledge to install and configure the products used, configuration details that are important to the deployment of this solution are provided in this CVD.
Cisco UCS Integrated Infrastructure for SAP HANA solution designed with next generation Fabric Interconnect which 40GbE ports. The solution is designed with 40GbE end-to-end network including Storage network. The persistent storage is configured on C240 C-series servers with MapR Converged Data Platform. MapR-FS provides distributed, reliable, high performance, scalable, and full read/write data storage for SAP HANA.
The Cisco UCS Integrated Infrastructure for SAP HANA with MapR Converged Data Platform provides an end-to-end architecture with Cisco Hardware that demonstrate support for multiple SAP HANA workloads with high availability and server redundancy. The solution supports up to 16 x Cisco UCS B460-M4 B-series blade servers or 16 x Cisco UCS C460-M4 C-series rack mount servers for SAP HANA and up to 8 x Cisco UCS C240-M4 C-Series rack mount servers for storage. The next Gen Cisco UCS Fabric Interconnect 6332 with 40 GbE network bandwidth for Server Management and Network connectivity. The Cisco UCS C240-M4 servers provides persistent storage with MapR Converged Data Platform, which is a modern NFS-mountable distributed file-system. The Nexus 3000 series Ethernet switch is used for failover purpose. In case of 2304 Fabric Expander failure or link failure between Server and Fabric Interconnect, the data traffic path will use Nexus 3000 series switch for High Availability and redundancy. Figure 1 shows the Cisco UCS Integrated Infrastructure for SAP HANA block diagram with Cisco UCS C460 –M4.
The reference architecture documented in the CVD consists of 8 x Cisco UCS B/C460-M4 servers for SAP HANA and 4 x Cisco UCS C240-M4 C-Series rack mount servers for storage.
Figure 1 Cisco UCS Integrated Infrastructure for SAP HANA with Cisco UCS C460 M4
Figure 2 shows the Cisco UCS Integrated Infrastructure for SAP HANA block diagram with Cisco UCS B460 M4.
Figure 2 Cisco UCS Integrated Infrastructure for SAP HANA with Cisco UCS B460 M4
The solution can be designed with a pair for Cisco Nexus 9000 series switches, alternative to single Nexus 3000 series switch. Two Cisco Nexus 9000 series switches is configured with vPC between network switch and Cisco Fabric Interconnect as show in Figure 3.
Figure 3 Cisco UCS B/C460 M4 Scale-Out for SAP HANA Pair of Nexus Switches
The Cisco Unified Computing System is a state-of-the-art data center platform that unites computing, network, storage access, and virtualization into a single cohesive system.
The main components of the Cisco Unified Computing System are:
· Computing - The system is based on an entirely new class of computing system that incorporates rack mount and blade servers based on Intel Xeon Processor E5 and E7. The Cisco UCS Servers offer the patented Cisco Extended Memory Technology to support applications with large datasets and allow more virtual machines per server.
· Network - The system is integrated onto a low-latency, lossless, 40-Gbps unified network fabric. This network foundation consolidates LANs, SANs, and high-performance computing networks which are separate networks today. The unified fabric lowers costs by reducing the number of network adapters, switches, and cables, and by decreasing the power and cooling requirements.
· Virtualization - The system unleashes the full potential of virtualization by enhancing the scalability, performance, and operational control of virtual environments. Cisco security, policy enforcement, and diagnostic features are now extended into virtualized environments to better support changing business and IT requirements.
· Storage access - The system provides consolidated access to both SAN storage and Network Attached Storage (NAS) over the unified fabric. By unifying the storage access, the Cisco Unified Computing System can access storage over Ethernet (NFS or iSCSI) and Fibre Channel over Ethernet (FCoE). This provides customers with choice for storage access and investment protection. In addition, the server administrators can pre-assign storage-access policies for system connectivity to storage resources, simplifying storage connectivity, and management for increased productivity.
The Cisco Unified Computing System is designed to deliver:
· A reduced Total Cost of Ownership (TCO) and increased business agility.
· Increased IT staff productivity through just-in-time provisioning and mobility support.
· A cohesive, integrated system, which unifies the technology in the data center.
· Industry standards supported by a partner ecosystem of industry leaders.
Cisco Unified Computing System Manager (UCSM) provides unified, embedded, policy-driven management programmatically controls server, network, and storage resources, so the can be efficiently managed at scale through software. It is embedded on a pair of Cisco UCS 6300 or 6200 Series Fabric Interconnects, supporting an end-to-end Ethernet or Fibre Channel over Ethernet (FCoE) solution of up to 40 Gb and up to 16 Gb Fibre Channel. The manager participates in server, fabric, and storage provisioning, device discovery, inventory, configuration, diagnostics, monitoring, fault detection, auditing, and statistics collection. Cisco UCS Manager works with HTML 5, Java, or CLI graphical user interfaces. An open API facilitates integration of Cisco UCS Manager with a wide variety of monitoring, analysis, configuration, deployment, and orchestration tools from other independent software vendors. The API also facilitates customer development through the Cisco UCS PowerTool for Powershell and a Python SDK.
The Cisco UCS 6300 Series Fabric Interconnects are a core part of Cisco UCS, providing both network connectivity and management capabilities for the system. The Cisco UCS 6300 Series offers line-rate, low-latency, lossless 10 and 40 Gigabit Ethernet, Fibre Channel over Ethernet (FCoE), and Fibre Channel functions. The Cisco UCS 6300 Series provides the management and communication backbone for the Cisco UCS B-Series Blade Servers, 5100 Series Blade Server Chassis, and C-Series Rack Servers managed by Cisco UCS. All servers attached to the fabric interconnects become part of a single, highly available management domain. In addition, by supporting unified fabric, the Cisco UCS 6300 Series provides both LAN and SAN connectivity for all servers within its domain.
From a networking perspective, the Cisco UCS 6300 Series uses a cut-through architecture, supporting deterministic, low-latency, line-rate 10 and 40 Gigabit Ethernet ports, switching capacity of 2.56 terabits per second (Tbps), and 320 Gbps of bandwidth per chassis, independent of packet size and enabled services. The product family supports Cisco® low-latency, lossless 10 and 40 Gigabit Ethernet unified network fabric capabilities, which increase the reliability, efficiency, and scalability of Ethernet networks. The fabric interconnect supports multiple traffic classes over a lossless Ethernet fabric from the server through the fabric interconnect. Significant TCO savings can be achieved with an FCoE optimized server design in which network interface cards (NICs), host bus adapters (HBAs), cables, and switches can be consolidated.
The Cisco UCS 6332 Fabric Interconnect is the management and communication backbone for Cisco UCS B-Series Blade Servers, C-Series Rack Servers, and 5100 Series Blade Server Chassis. All servers attached to 6332 Fabric Interconnects become part of one highly available management domain. The Cisco UCS 6332UP 32-Port Fabric Interconnect is a 1-rack-unit 40 Gigabit Ethernet, FCoE and Fibre Channel switch offering up to 2.56 Tbps throughput and up to 32 ports. The switch has 32 fixed 40-Gbps Ethernet and FCoE ports. Cisco UCS 6332UP 32-Port Fabric Interconnect have ports that can be configured for the breakout feature that supports connectivity between 40 Gigabit Ethernet ports and 10 Gigabit Ethernet ports. This feature provides backward compatibility to existing hardware that supports 10 Gigabit Ethernet. A 40 Gigabit Ethernet port can be used as four 10 Gigabit Ethernet ports. Using a 40 Gigabit Ethernet SFP, these ports on a Cisco UCS 6300 Series Fabric Interconnect can connect to another fabric interconnect that has four 10 Gigabit Ethernet SFPs.
Figure 4 Cisco UCS 6332 UP Fabric Interconnect
The Cisco UCS 2304 Fabric Extender has four 40 Gigabit Ethernet, FCoE-capable, Quad Small Form-Factor Pluggable (QSFP+) ports that connect the blade chassis to the fabric interconnect. Each Cisco UCS 2304 has four 40 Gigabit Ethernet ports connected through the midplane to each half-width slot in the chassis. Typically configured in pairs for redundancy, two fabric extenders provide up to 320 Gbps of I/O to the chassis.
Figure 5 Cisco UCS 2304 XP
The Cisco UCS 5100 Series Blade Server Chassis is a crucial building block of the Cisco Unified Computing System, delivering a scalable and flexible blade server.
The Cisco UCS 5108 Blade Server Chassis is six rack units (6RU) high and can mount in an industry standard 19-inch rack. A single chassis can house up to eight half-width Cisco UCS B-Series Blade Servers and can accommodate both half-width and full-width blade form factors. Four hot-swappable power supplies are accessible from the front of the chassis, and single-phase AC, –48V DC, and 200 to 380V DC power supplies and chassis are available. These power supplies are up to 94 percent efficient and meet the requirements for the 80 Plus Platinum rating. The power subsystem can be configured to support nonredundant, N+1 redundant, and grid-redundant configurations. The rear of the chassis contains eight hot-swappable fans, four power connectors (one per power supply), and two I/O bays that can support either Cisco UCS 2000 Series Fabric Extenders or the Cisco UCS 6324 Fabric Interconnect. A passive midplane provides up to 80 Gbps of I/O bandwidth per server slot and up to 160 Gbps of I/O bandwidth for two slots.
The Cisco UCS Blade Server Chassis is shown in Figure 6.
Figure 6 Cisco Blade Server Chassis (front and back view)
The Cisco UCS B460 M4 Blade Server provides outstanding performance and enterprise-critical stability for memory-intensive workloads such as large-scale databases, in-memory analytics, and business intelligence. The Cisco UCS B460 M4 harnesses the power of four Intel Xeon processor E7 v2, E7 v3, or E7 v4 product family and accelerates access to critical data. This blade server supports up to 96 processor cores, 6.0 TB of memory, 6.4 TB of internal storage, and 320 Gbps of overall Ethernet throughput.
The Cisco UCS B460 M4 provides:
· Four Intel Xeon processor E7 v2, E7 v3, or E7 v4 product family
· 96 DDR3 or DDR4 memory DIMM slots
· Four hot-pluggable drive bays for Hard Disk Drives (HDDs) or Solid State Disks (SSDs)
· SAS controller on board with RAID 0 and 1 support
· Two modular LAN on motherboard (mLOM) slots for Cisco UCS Virtual Interface Card (VIC)
· Six PCIe mezzanine slots, with two dedicated for optional Cisco UCS VIC 1340, and four slots for Cisco UCS VIC 1380, VIC port expander, or flash cards
Figure 7 Cisco UCS B460 M4 Blade Server
The Cisco UCS C460 M4 Rack Server offers exceptionally high performance and reliability to power the most compute- and memory-intensive, mission-critical enterprise applications and virtualized workloads. The C460 M4 Rack Server is well suited for the most demanding enterprise and mission-critical workloads, large-scale virtualization, and database applications. Either as a standalone server or in UCS-managed operations, customers gain the benefits of this server's high-capacity memory when very large memory footprints are required. Product highlights include:
· Either 2 or 4 Intel® Xeon® processor E7-4800/8800 v2, v3, or v4 product family CPUs.
· Up to 6 terabytes (TB) of double-data-rate 3 (DDR3) memory or double-data-rate 4 (DDR4) memory in 96 dual inline memory (DIMM) slots.
· 12 front accessible Small Form Factor (SFF) disk drive bays with support for hot-pluggable SAS/SATA/SSD disk drives. Two disk drive bays can be used for PCIe SSDs.
· Abundant I/O capability with 10 PCI Express (PCIe) Generation 3 (Gen 3) slots supporting the Cisco UCS virtual interface cards (VICs). An internal slot is reserved for a hard-disk drive array controller card.
· Two Gigabit Ethernet LAN-on-motherboard (LOM) ports, two 10-Gigabit Ethernet ports, and a dedicated out-of-band (OOB) management port that provides additional networking options.
Figure 8 Cisco UCS C460 M4 Rack Server
The Cisco UCS C240 M4 Rack Server is an enterprise-class server designed to deliver exceptional performance, expandability, and efficiency for storage and I/O-intensive infrastructure workloads. This includes big data analytics, virtualization, and graphics-rich and bare-metal applications.
The Cisco UCS C240 M4 Rack Server delivers outstanding levels of expandability and performance for standalone or Cisco UCS-managed environments in a two rack-unit (2RU) form factor. It provides:
· Dual Intel® Xeon® E5-2600 v3 or v4 processors for improved performance suitable for nearly all two-socket applications
· Next-generation double-data-rate 4 (DDR4) memory, 12-Gbps SAS throughput, and NVMe PCIe SSD support
· Innovative Cisco UCS virtual interface card (VIC) support in PCIe or modular LAN-on-motherboard (mLOM) form factor
· Graphics-rich experiences to more virtual users with support for the latest NVIDIA graphics processing units (GPUs)
The Cisco UCS C240 M4 server also offers maximum reliability, availability, and serviceability (RAS) features, including:
· Tool-free CPU insertion
· Easy-to-use latching lid
· Hot-swappable and hot-pluggable components
· Redundant Cisco Flexible Flash SD cards
The Cisco UCS C240 M4 server can be deployed standalone or as part of the Cisco Unified Computing System. Cisco UCS unifies computing, networking, management, virtualization, and storage access into a single integrated architecture that can enable end-to-end server visibility, management, and control in both bare-metal and virtualized environments. With Cisco UCS-managed deployment, Cisco UCS C240 M4 takes advantage of our standards-based unified computing innovations to significantly reduce customers’ TCO and increase business agility.
Figure 9 Cisco UCS C240 M4 Rack Server
This section discusses the Cisco I/O Adapters used in this solution.
The Cisco UCS blade server has various Converged Network Adapters (CNA) options.
The Cisco UCS Virtual Interface Card (VIC) 1340 is a 2-port 40-Gbps Ethernet or dual 4 x 10-Gbps Ethernet, Fibre Channel over Ethernet (FCoE)-capable modular LAN on motherboard (mLOM) designed exclusively for the Cisco UCS B-Series Blade Servers. When used in combination with an optional port expander, the Cisco UCS VIC 1340 capabilities is enabled for two ports of 40-Gbps Ethernet.
Figure 10 Cisco UCS 1340 VIC Card
The Cisco UCS VIC 1340 enables a policy-based, stateless, agile server infrastructure that can present over 256 PCIe standards-compliant interfaces to the host that can be dynamically configured as either network interface cards (NICs) or host bus adapters (HBAs). In addition, the Cisco UCS VIC 1340 supports Cisco® Data Center Virtual Machine Fabric Extender (VM-FEX) technology, which extends the Cisco UCS fabric interconnect ports to virtual machines, simplifying server virtualization deployment and management.
The Cisco UCS Virtual Interface Card (VIC) 1380 is a dual-port 40-Gbps Ethernet, or dual 4 x 10 Fibre Channel over Ethernet (FCoE)-capable mezzanine card designed exclusively for the M5 generation of Cisco UCS B-Series Blade Servers. The card enables a policy-based, stateless, agile server infrastructure that can present over 256 PCIe standards-compliant interfaces to the host that can be dynamically configured as either network interface cards (NICs) or host bus adapters (HBAs). In addition, the Cisco UCS VIC 1380 supports Cisco® Data Center Virtual Machine Fabric Extender (VM-FEX) technology, which extends the Cisco UCS fabric interconnect ports to virtual machines, simplifying server virtualization deployment and management.
Figure 11 Cisco UCS 1380 VIC Card
The Cisco UCS Virtual Interface Card (VIC) 1385 is a Cisco® innovation. It provides a policy-based, stateless, agile server infrastructure for your data center. This dual-port Enhanced Quad Small Form-Factor Pluggable (QSFP) half-height PCI Express (PCIe) card is designed exclusively for Cisco UCS C-Series Rack Servers. The card supports 40 Gigabit Ethernet and Fibre Channel over Ethernet (FCoE). It incorporates Cisco’s next-generation converged network adapter (CNA) technology and offers a comprehensive feature set, providing investment protection for future feature software releases. The card can present more than 256 PCIe standards-compliant interfaces to the host, and these can be dynamically configured as either network interface cards (NICs) or host bus adapters (HBAs). In addition, the VIC supports Cisco Data Center Virtual Machine Fabric Extender (VM-FEX) technology. This technology extends the Cisco UCS Fabric Interconnect ports to virtual machines, simplifying server virtualization deployment.
Figure 12 Cisco UCS 1385 VIC Card
Cisco UCS Performance Manager is a purpose-built data center operations management solution. It unifies the monitoring of key applications, business services, and integrated infrastructures across dynamic, heterogeneous, physical, and virtual Cisco UCS-powered data centers. Cisco UCS Performance Manager uses Cisco UCS APIs to collect data from Cisco UCS Manager to display comprehensive information about all Cisco UCS infrastructure components. With a customizable view, data center staff can see application services and view performance and component or service availability information for Cisco UCS integrated infrastructures.
Cisco UCS Performance Manager dynamically collects information about Cisco UCS servers, network, storage, and virtual machine hosts using an agentless information gathering approach. The solution provided the following:
· Unifies performance monitoring and management of Cisco UCS integrated infrastructure solutions
· Delivers real-time views of fabric and data center switch bandwidth usage and capacity thresholds
· Discovers and creates a relationship model of each system, giving staff a single, accurate view of all components
· Allows staff to navigate into individual Cisco UCS infrastructure components when troubleshooting and resolving issues
Cisco UCS Performance Manager provides deep visibility of Cisco UCS integrated infrastructure performance for service profiles, chassis, fabric extenders, adapters, virtual interface cards, ports, and uplinks for granular data center monitoring. Customers can use Cisco UCS Performance Manager to maintain service-level agreements (SLAs) by managing optimal resource allocation to prevent under-provisioning and avoid performance degradation. By defining component or application-centric views of critical resources, administrators can monitor SLA health and performance from a single console, eliminating the need for multiple tools.
For detailed information, see the Cisco UCS Performance Manager Install Guide.
Cisco’s Unified Compute System is revolutionizing the way servers are managed in data-center. The following are the unique differentiators of Cisco Unified Computing System and Cisco UCS Manager:
· Embedded management: In Cisco Unified Computing System, the servers are managed by the embedded firmware in the Fabric Interconnects, eliminating need for any external physical or virtual devices to manage the servers. Also, a pair of FIs can manage up to 40 chassis, each containing 8 blade servers. This gives enormous scaling on management plane.
· Unified fabric: In Cisco Unified Computing System, from blade server chassis or rack server fabric extender to FI, there is a single Ethernet cable used for LAN, SAN and management traffic. This converged I/O, results in reduced cables, SFPs and adapters – reducing capital and operational expenses of overall solution.
· Auto discovery: By simply inserting the blade server in the chassis or connecting rack server to the fabric extender, discovery and inventory of compute resource occurs automatically without any management intervention. Combination of unified fabric and auto-discovery enables wire-once architecture of Cisco Unified Computing System, where compute capability of Cisco Unified Computing System can extend easily, while keeping the existing external connectivity to LAN, SAN and management networks.
· Policy based resource classification: When a compute resource is discovered by Cisco UCS Manager, it can be automatically classified to a given resource pool based on policies defined. This capability is useful in multi-tenant cloud computing. This CVD focuses on the policy-based resource classification of Cisco UCS Manager.
· Combined Rack and Blade server management: Cisco UCS Manager can manage Cisco UCS B-Series Blade Servers and Cisco UCS C-Series Rack Servers under the same Cisco UCS domain. This feature, along with stateless computing makes compute resources truly hardware form factor agnostic. This CVD focuses on the combination of B-Series and C-Series Servers to demonstrate stateless and form factor independent computing work load.
· Model-based management architecture: Cisco UCS Manager Architecture and management database is model based and data driven. Open, standard based XML API is provided to operate on the management model. This enables easy and scalable integration of Cisco UCS Manager with other management system, such as VMware vCloud director, Microsoft system center, and Citrix CloudPlatform.
· Policies, Pools, Templates: Management approach in Cisco UCS Manager is based on defining policies, pools and templates, instead of cluttered configuration, which enables simple, loosely coupled, data driven approach in managing compute, network and storage resources.
· Loose referential integrity: In Cisco UCS Manager, a service profile, port profile or policies can refer to other policies or logical resources with loose referential integrity. A referred policy cannot exist at the time of authoring the referring policy or a referred policy can be deleted even though other policies are referring to it. This provides different subject matter experts to work independently from each-other. This provides great flexibilities where different experts from different domains, such as network, storage, security, server and virtualization work together to accomplish a complex task.
· Policy resolution: In Cisco UCS Manager, a tree structure of organizational unit hierarchy can be created that mimics the real life tenants and/or organization relationships. Various policies, pools and templates can be defined at different levels of organization hierarchy. A policy referring to other policy by name is resolved in the org hierarchy with closest policy match. If no policy with specific name is found in the hierarchy till root org, then special policy named “default” is searched. This policy resolution practice enables automation friendly management APIs and provides great flexibilities to owners of different orgs.
· Service profiles and stateless computing: Service profile is a logical representation of a server, carrying its various identities and policies. This logical server can be assigned to any physical compute resource as far as it meets the resource requirements. Stateless computing enables procurement of a server within minutes, which used to take days in legacy server management systems.
· Built-in multi-tenancy support: Combination of policies, pools and templates, loose referential integrity, policy resolution in org hierarchy and service profile based approach to compute resources make Cisco UCS Manager inherently friendly to multi-tenant environment typically observed in private and public clouds.
· Virtualization aware network: VM-FEX technology makes access layer of network aware about host virtualization. This prevents domain pollution of compute and network domains with virtualization when virtual network is managed by port-profiles defined by the network administrators’ team. VM-FEX also offloads hypervisor CPU by performing switching in the hardware, thus allowing hypervisor CPU to do more virtualization related tasks. VM-FEX technology is well integrated with VMware vCenter, Linux KVM and Hyper-V SR-IOV to simplify cloud management.
· Simplified QoS: Even though fibre-channel and Ethernet are converged in Cisco UCS fabric, built-in support for QoS and lossless Ethernet makes it seamless. Network Quality of Service (QoS) is simplified in Cisco UCS Manager by representing all system classes in one GUI panel.
The MapR Converged Data Platform integrates Hadoop and Spark, real-time database capabilities, and global event streaming with big data enterprise storage, for developing and running innovative data applications. The MapR Platform is powered by the industry’s fastest, most reliable, secure, and open data infrastructure that dramatically lowers TCO and enables global real-time data applications.
MapR Platform Services is the set of core capabilities in the MapR Converged Data Platform. Core components include:
· MapR Streams - a global publish-subscribe event streaming system for big data
· MapR-DB - a high performance, in-Hadoop NoSQL database management system
· MapR-FS - the underlying POSIX file system that provides distributed, reliable, high performance, scalable, and full read/write data storage
The Cisco UCS C240 Appliance for SAP HANA takes advantage of the enterprise grade MapR-FS in MapR Platform Services. It offers the appliance a robust storage layer that is highly available, resilient and performant.
MapR converged data platform is backed by the robust MapR-FS, which can be accessed through the NFS gateway. The SAP HANA Servers mounts MapR-FS using NFS client. Data can be persisted to the storage system managed by MapR-FS. MapR-FS is distributed, has a global name space, real-time read/write access, volume based, secure and has many other benefits compared to HDFS.
This section describes the SAP HANA system requirements defined by SAP and Architecture of Cisco UCS Integrated Infrastructure for SAP HANA.
SAP HANA System on a Single Server Scale-Up, is the simplest of the installation types. It is possible to run an SAP HANA system entirely on one host and then scale the system up as needed. All data and processes are located on the same server and can be accessed locally. The network requirements for this option minimum one 1-Gb Ethernet (access) and one 10-Gb Ethernet storage networks are sufficient to run SAP HANA scale-up. SAP HANA Scale-Out option is used if the SAP HANA system does not fit into the main memory of a single server based on the rules defined by SAP. In this method, multiple independent servers are combined to form one system and the load is distributed among multiple servers. In a distributed system, each index server is usually assigned to its own host to achieve maximum performance. It is possible to assign different tables to different hosts (partitioning the database), or a single table can be split across hosts (partitioning of tables). SAP HANA Scale-Out supports failover scenarios and high availability. Individual hosts in a distributed system have different roles master, worker, slave, standby depending on the task.
Some use cases are not supported on SAP HANA Scale-Out configuration and it is recommended to check with SAP whether a use case can be deployed as a Scale-Out solution.
The network requirements for this option are higher than for Scale-Up systems. In addition to the client and application access and storage access network, a node-to-node network is necessary. One 10 Gigabit Ethernet (access) and one 10 Gigabit Ethernet (node-to-node) and one 10 Gigabit Ethernet storage networks are required to run SAP HANA Scale-Out system. Additional network bandwidth is required to support system replication or backup capability.
There are hardware and software requirements defined by SAP to run SAP HANA systems. This Cisco Validated Design uses guidelines provided by SAP.
For additional information, go to: http://saphana.com.
This document does not cover the updated information published by SAP.
SAP HANA supports servers equipped with Intel(R) Xeon(R) Platinum 8176, 8176M, 8180 and 8180M CPU.
SAP HANA Scale-Out solution is supported in the following memory configurations:
· Homogenous symmetric assembly of dual in-line memory modules (DIMMs) for example, DIMM size or speed should not be mixed
· Maximum use of all available memory channels
· Memory of 1.5 TB or 3 TB per 4 Socket Server for SAP NetWeaver Business Warehouse (BW) and DataMart
A SAP HANA data center deployment can range from a database running on a single host to a complex distributed system. Distributed systems can get complex with multiple hosts located at a primary site having one or more secondary sites; supporting a distributed multi-terabyte database with full fault and disaster recovery.
SAP HANA has different types of network communication channels to support the different SAP HANA scenarios and setups:
· Client zone: Channels used for external access to SAP HANA functions by end-user clients, administration clients, and application servers, and for data provisioning through SQL or HTTP
· Internal zone: Channels used for SAP HANA internal communication within the database or, in a distributed scenario, for communication between hosts
· Storage zone: Channels used for storage access (data persistence) and for backup and restore procedures
Table 1 lists all the networks defined by SAP or Cisco or requested by customers.
Table 1 List of Known Networks
Name | Use Case | Solutions | Bandwidth Requirements | Solution Design |
Client Zone Networks | ||||
Application Server Network | SAP Application Server to DB communication | All | 1 or 10 GbE | 10 or 40 GbE |
Client Network | User / Client Application to DB communication | All | 1 or 10 GbE | 10 or 40 GbE |
Data Source Network | Data import and external data integration | Optional for all SAP HANA systems | 1 or 10 GbE | 10 or 40 GbE |
Internal Zone Networks | ||||
Inter-Node Network | Node to node communication within a scale-out configuration | Scale-Out | 10 GbE | 40 GbE |
System Replication Network | SAP HANA System Replication | For SAP HANA Disaster Tolerance | TBD with Customer | TBD with Customer |
Storage Zone Networks | ||||
Backup Network | Data Backup | Optional for all SAP HANA systems | 10 GbE | 10 or 40 GbE |
Storage Network | Node to Storage communication | All | 10 GbE | 20 or 40 GbE |
Infrastructure Related Networks | ||||
Administration Network | Infrastructure and SAP HANA administration | Optional for all SAP HANA systems | 1 GbE | 10 or 40 GbE |
Boot Network | Boot the Operating Systems through PXE/NFS or FCoE | Optional for all SAP HANA systems | 1 GbE | N/A |
For detailed information about the network requirements for SAP HANA see: SAP HANA Network Requirements.
The network need to be properly segmented and must be connected to the same core/ backbone switch as shown in Figure 13 based on customer’s high-availability and redundancy requirements for different SAP HANA network segments.
Figure 13 High-Level SAP HANA Network Overview
Based on the listed network requirements, every server must be equipped with 2x 10 Gigabit Ethernet for scale-up systems to establish the communication with the application or user (Client Zone) and a 10 GbE Interface for Storage access.
For scale-out solutions an additional redundant network for SAP HANA node to node communication with 10 GbE is required (Internal Zone).
For more information on SAP HANA Network security please refer to the SAP HANA Security Guide.
As an in-memory database, SAP HANA uses storage devices to save a copy of the data, for the purpose of startup and fault recovery without data loss. The choice of the specific storage technology is driven by various requirements like size, performance and high availability. To use the storage system in the Tailored Datacenter Integration option, the storage must be certified for the SAP HANA TDI option at: http://scn.sap.com/docs/DOC-48516.
All relevant information about storage requirements is documented in the white paper SAP HANA Storage Requirements.
SAP can only support performance related SAP HANA topics if the installed solution has passed the validation test successfully.
Refer to SAP HANA Administration Guide section 2.8 Hardware Checks for Tailored Datacenter Integration for Hardware check test tool and the related documentation.
Figure 14 shows the file system layout and the required storage sizes to install and operate SAP HANA. For the Linux OS installation (/root) 10 GB of disk size is recommended. Additionally, 50 GB must be provided for the /usr/sap since the volume used for SAP software that supports SAP HANA.
While installing SAP HANA on a host, we specify the mount point for the installation binaries (/hana/shared/<sid>), data files (/hana/data/<sid>) and log files (/hana/log/<sid>), where sid is the instance identifier of the SAP HANA installation.
Figure 14 File System Layout for 2 Node Scale-Out System
The storage sizing for filesystem is based on the amount of memory equipped on the SAP HANA host.
In case of distributed installation of SAP HANA Scale-Out, each server will have the following:
Root-FS: 10 GB
/usr/sap: 50 GB
The installation binaries, trace and configuration files are stored on a shared filesystem, which should be accessible for all hosts in the distributed installation. The size of shared filesystem should be equal to one times memory in each host for every four worker nodes. For example: In a distributed installation with eight hosts with 2 TB of memory each, shared file system should be 4 TB.
For each HANA host there should be a mount point for data and log volume.
Size of the file system for data volume for Appliance option is three times the host memory:
/hana/data/<sid>/mntXXXXX: 3 x Memory
Size of the file system for data volume with TDI option is one times the host memory:
/hana/data/<sid>/mntXXXXX: 1 x Memory
The Log volume must be as follows:
· Half of the server memory for systems ≤ 256 GB memory
· Min 512 GB for systems with ≥ 512 GB memory
The supported operating systems for SAP HANA are as follows:
· SUSE Linux Enterprise Server for SAP Applications
· RedHat Enterprise Linux for SAP HANA
This document provides the installation process for the SUSE Linux Enterprise Server for SAP Applications only. For RedHat Enterprise Linux option, please review the Cisco Validated Design, Cisco UCS II for SAP HANA with RHEL.
The infrastructure for a SAP HANA solution must not have single point of failure. To support high-availability, the hardware and software requirements are:
· Internal storage: A RAID-based configuration is preferred
· External storage: Redundant data paths, dual controllers, and a RAID-based configuration are required
· Ethernet switches: Two or more independent switches should be used
SAP HANA Scale-Out comes with in integrated high-availability function. If a SAP HANA system is configured with a stand-by node, a failed part of SAP HANA will start on the stand-by node automatically. For automatic host failover, storage connector API must be properly configured for the implementation and operation of the SAP HANA.
Please check the latest information from SAP at: http://saphana.com or http://service.sap.com/notes.
The Cisco UCS Integrated Infrastructure for SAP HANA with MapR Converged Data Platform provides an end-to-end architecture with Cisco Hardware that demonstrate support for multiple SAP HANA workloads with high availability and server redundancy. The architecture uses Cisco UCS Manager with combined Cisco UCS B-Series and C-Series Servers with Cisco UCS Fabric Interconnect. The uplink from Cisco UCS Fabric Interconnect is connected to Nexus 3524 switches for High Availability and Failover functionality. The Cisco UCS C-Series Rack Servers are connected directly to Cisco UCS Fabric Interconnect with single-wire management feature, the data traffic between HANA servers and Storage will be contained in the Cisco UCS Fabric Interconnect. This infrastructure is deployed to provide IP based Storage access using NFS protocols with file-level access to shared storage.
Although this is the base design, each of the components can be scaled easily to support specific business requirements. Additional servers or even blade chassis can be deployed to increase compute capacity without additional Network components. Two Cisco UCS 6332UP, 32 port Fabric interconnect can support up to:
· 16 Cisco UCS B-Series B460 M4 Server with 8 Blade Server Chassis
· 8 Cisco UCS C240 M4 Server
Minimum of 3 x UCS C240 M4 Servers are required to install MapR Converged Data Platform with High Availability. 3 x Cisco UCS C240 M4 can support up to 6 Active HANA Servers. The scaling of storage to computer node is linear, for every 2 x Cisco UCS Server for HANA, 1 x Cisco UCS C240 M4 for Storage is required to meet the SAP HANA storage performance defined by SAP SE.
The solution is designed to meet SAP HANA performance requirement defined by SAP SE. The HANA Server and Storage server are connecting to Cisco UCS Fabric Interconnect, all the data traffic between HANA node and Storage node is contained in the Cisco UCS Fabric Interconnect. Each HANA Server and Storage Server are equipped with 2 x 40GbE capable Cisco Virtual Interface Cards, the storage network provides dedicated bandwidth between HANA servers and Storage Servers. The traffic shaping and Quality of Service (QoS) is configured and managed by Cisco UCS Fabric Interconnect. For HANA node-to-node network, 40 Gb dedicated network bandwidth is provided with non-blocking mode. The Cisco UCS Integrated Infrastructure for SAP HANA quarantines the bandwidth and latency for best performance to run SAP HANA.
This document provides details for configuring a fully redundant, highly available configuration for a Cisco UCS B/C460 M4 Scale-Out for SAP HANA.
This document is intended to enable you to fully configure the customer environment. In this process, various steps require you to insert customer-specific naming conventions, IP addresses, and VLAN schemes, as well as to record appropriate MAC addresses. Table 2 lists the configuration variables that are used throughout this document. This table can be completed based on the specific site variables and used in implementing the document configuration steps.
Table 2 Configuration Variables
Variable | Description | Customer Implementation Value |
<<var_nexus_HA_hostname>> | Cisco Nexus 3524 HA Switch host name |
|
<<var_nexus_HA_mgmt0_ip>> | Out-of-band Cisco Nexus 3524 HA Switch management IP address |
|
<<var_nexus_HA_mgmt0_netmask>> | Out-of-band management network netmask |
|
<<var_nexus_HA_mgmt0_gw>> | Out-of-band management network default gateway |
|
<<var_global_ntp_server_ip>> | NTP server IP address |
|
<<var_oob_vlan_id>> | Out-of-band management network VLAN ID |
|
<<var_mgmt_vlan_id>> | Management network VLAN ID |
|
<<var_nexus_vpc_domain_mgmt_id>> | Unique Cisco Nexus switch VPC domain ID for Management Switch |
|
<<var_nexus_vpc_domain_id>> | Unique Cisco Nexus switch VPC domain ID |
|
<<var_nexus_A_hostname>> | Cisco Nexus 9000 A host name |
|
<<var_nexus_A_mgmt0_ip>> | Out-of-band Cisco Nexus 9000 A management IP address |
|
<<var_nexus_A_mgmt0_netmask>> | Out-of-band management network netmask |
|
<<var_nexus_A_mgmt0_gw>> | Out-of-band management network default gateway |
|
<<var_nexus_B_hostname>> | Cisco Nexus 9000 B host name |
|
<<var_nexus_B_mgmt0_ip>> | Out-of-band Cisco Nexus 9000 B management IP address |
|
<<var_nexus_B_mgmt0_netmask>> | Out-of-band management network netmask |
|
<<var_nexus_B_mgmt0_gw>> | Out-of-band management network default gateway |
|
<<var_mapr-01_vlan_id>> | MapR Internal network 01 VLAN ID |
|
<<var_mapr-02_vlan_id>> | MapR Internal network 02 VLAN ID |
|
<<var_mapr-03_vlan_id>> | MapR Internal network 03 VLAN ID |
|
<<var_storage_vlan_id>> | Storage network for HANA Data/log VLAN ID |
|
<<var_internal_vlan_id>> | Node to Node Network for HANA Data/log VLAN ID |
|
<<var_backup_vlan_id>> | Backup Network for HANA Data/log VLAN ID |
|
<<var_client_vlan_id>> | Client Network for HANA Data/log VLAN ID |
|
<<var_appserver_vlan_id>> | Application Server Network for HANA Data/log VLAN ID |
|
<<var_datasource_vlan_id>> | Data source Network for HANA Data/log VLAN ID |
|
<<var_replication_vlan_id>> | Replication Network for HANA Data/log VLAN ID |
|
<<var_inband_vlan_id>> | In-band management network VLAN ID |
|
<<var_ucs_clustername>> | Cisco UCS Manager cluster host name |
|
<<var_ucsa_mgmt_ip>> | Cisco UCS fabric interconnect (FI) A out-of-band management IP address |
|
<<var_ucsa_mgmt_mask>> | Out-of-band management network netmask |
|
<<var_ucsa_mgmt_gateway>> | Out-of-band management network default gateway |
|
<<var_ucs_cluster_ip>> | Cisco UCS Manager cluster IP address |
|
<<var_ucsb_mgmt_ip>> | Cisco UCS FI B out-of-band management IP address |
|
The information in this section is provided as a reference for cabling the network and compute components. For connectivity between 40GbE ports to 10GbE ports on Cisco Switches, Cisco QSFP to Four SFP+ Copper Breakout Cables are used. These breakout cables connect to a 40G QSFP port of a Cisco Fabric Interconnect on one end and to four 10G SFP+ ports of a Cisco switch on the other end
Figure 15 illustrates the cabling topology for Cisco UCS Integrated Infrastructure for SAP HANA configuration using the Cisco Nexus 3000.
Figure 15 Cabling Topology for Cisco UCS Integrated Infrastructure for SAP HANA
Figure 16 Cisco UCS Integrated Infrastructure for SAP HANA configuration using the Cisco Nexus 3000
Figure 17 shows the cabling topology for Cisco UCS Integrated Infrastructure for SAP HANA configuration using the pair of Cisco Nexus 9000 series switches.
Figure 17 Cabling Topology for Cisco UCS Integrated Infrastructure for SAP HANA
The information in this section is provided as a reference for cabling the network and compute components.
To simplify cabling requirements, the tables include both local and remote device and port locations. The following tables show the out-of-band management ports connectivity into preexisting management infrastructure, the Management Ports cabling needs to be adjusted accordingly. These Management interfaces will be used in various configuration steps.
Table 3 through 0provides the details of all the connections.
Table 3 Cisco UCS Fabric Interconnect A - Cabling Information for Cisco UCS B460-M4 Server(s) Option
Local Device | Local Port | Connection | Remote Device | Remote Port |
Cisco UCS fabric interconnect A
| Eth1/1 | 40GbE | Uplink to Customer Data Switch A | Any |
Eth1/2 | 40GbE | Uplink to Customer Data Switch B | Any | |
Eth1/3/1 | 40GbE QSFP to 4 SFP+ break-out cables | Cisco Nexus 3000 HA | Eth 1/3 | |
Eth1/3/2 | Cisco Nexus 3000 HA | Eth 1/4 | ||
Eth1/3/3 | Cisco Nexus 3000 HA | Eth 1/5 | ||
Eth1/3/4 | Cisco Nexus 3000 HA | Eth 1/6 | ||
Eth1/3/1* | 40GbE QSFP to 4 SFP+ break-out cables | Cisco Nexus 9000 A | Eth 1/3 | |
Eth1/3/2* | Cisco Nexus 9000 A | Eth 1/4 | ||
Eth1/3/3* | Cisco Nexus 9000 B | Eth 1/3 | ||
Eth1/3/4* | Cisco Nexus 9000 B | Eth 1/4 | ||
Eth1/4 | 40GbE |
|
| |
Eth1/5 | 40GbE | Cisco UCS Chassis 1 Fabric Extender (FEX) A | IOM 1/1 | |
Eth1/6 | 40GbE | Cisco UCS Chassis 1 Fabric Extender (FEX) A | IOM 1/2 | |
Eth1/7 | 40GbE | Cisco UCS Chassis 1 Fabric Extender (FEX) A | IOM 1/3 | |
Eth1/8 | 40GbE | Cisco UCS Chassis 1 Fabric Extender (FEX) A | IOM 1/4 | |
Eth1/9 | 40GbE | Cisco UCS Chassis 2 Fabric Extender (FEX) A | IOM 1/1 | |
Eth1/10 | 40GbE | Cisco UCS Chassis 2 Fabric Extender (FEX) A | IOM 1/2 | |
Eth1/11 | 40GbE | Cisco UCS Chassis 2 Fabric Extender (FEX) A | IOM 1/3 | |
Eth1/12 | 40GbE | Cisco UCS Chassis 2 Fabric Extender (FEX) A | IOM 1/4 | |
Eth1/13 | 40GbE | Cisco UCS Chassis 3 Fabric Extender (FEX) A | IOM 1/1 | |
Eth1/14 | 40GbE | Cisco UCS Chassis 3 Fabric Extender (FEX) A | IOM 1/2 | |
Eth1/15 | 40GbE | Cisco UCS Chassis 3 Fabric Extender (FEX) A | IOM 1/3 | |
Eth1/16 | 40GbE | Cisco UCS Chassis 3 Fabric Extender (FEX) A | IOM 1/4 | |
Eth1/17 | 40GbE | Cisco UCS Chassis 4 Fabric Extender (FEX) A | IOM 1/1 | |
Eth1/18 | 40GbE | Cisco UCS Chassis 4 Fabric Extender (FEX) A | IOM 1/2 | |
Eth1/19 | 40GbE | Cisco UCS Chassis 4 Fabric Extender (FEX) A | IOM 1/3 | |
Eth1/20 | 40GbE | Cisco UCS Chassis 4 Fabric Extender (FEX) A | IOM 1/4 | |
Eth1/21 | 40GbE | Cisco UCS C240-M5-1 | VIC 1385 Port 1 | |
Eth1/22 | 40GbE | Cisco UCS C240-M5-2 | VIC 1385 Port 1 | |
Eth1/23 | 40GbE | Cisco UCS C240-M5-3 | VIC 1385 Port 1 | |
Eth1/24 | 40GbE | Cisco UCS C240-M5-4 | VIC 1385 Port 1 | |
Eth1/25 | 40GbE | Cisco UCS C240-M5-5 | VIC 1385 Port 1 | |
Eth1/26 | 40GbE | Cisco UCS C240-M5-6 | VIC 1385 Port 1 | |
Eth1/27 | 40GbE | Cisco UCS C240-M5-7 | VIC 1385 Port 1 | |
Eth1/28 | 40GbE | Cisco UCS C240-M5-8 | VIC 1385 Port 1 | |
MGMT0 | GbE | Customer’s Management Switch | Any | |
L1 | GbE | Cisco UCS fabric interconnect B | L1 | |
L2 | GbE | Cisco UCS fabric interconnect B | L2 |
* The ports ETH1/3/1-4 are used with pair Nexus 9000 Switches design option.
Table 4 Cisco UCS Fabric Interconnect B - Cabling Information for Cisco UCS B460 M4 Server(s) Option
Local Device | Local Port | Connection | Remote Device | Remote Port |
Cisco UCS fabric interconnect B
| Eth1/1 | 40GbE | Uplink to Customer Data Switch A | Any |
Eth1/2 | 40GbE | Uplink to Customer Data Switch B | Any | |
Eth1/3/1 | 40GbE QSFP to 4 SFP+ break-out cables | Cisco Nexus 3000 HA | Eth 1/7 | |
Eth1/3/2 | Cisco Nexus 3000 HA | Eth 1/8 | ||
Eth1/3/3 | Cisco Nexus 3000 HA | Eth 1/9 | ||
Eth1/3/4 | Cisco Nexus 3000 HA | Eth 1/10 | ||
Eth1/3/1* | 40GbE QSFP to 4 SFP+ break-out cables | Cisco Nexus 9000 A | Eth 1/5 | |
Eth1/3/2* | Cisco Nexus 9000 A | Eth 1/6 | ||
Eth1/3/3* | Cisco Nexus 9000 B | Eth 1/5 | ||
Eth1/3/4* | Cisco Nexus 9000 B | Eth 1/6 | ||
Eth1/4 | 40GbE |
|
| |
Eth1/5 | 40GbE | Cisco UCS Chassis 1 Fabric Extender (FEX) B | IOM 1/1 | |
Eth1/6 | 40GbE | Cisco UCS Chassis 1 Fabric Extender (FEX) B | IOM 1/2 | |
Eth1/7 | 40GbE | Cisco UCS Chassis 1 Fabric Extender (FEX) B | IOM 1/3 | |
Eth1/8 | 40GbE | Cisco UCS Chassis 1 Fabric Extender (FEX) B | IOM 1/4 | |
Eth1/9 | 40GbE | Cisco UCS Chassis 2 Fabric Extender (FEX) B | IOM 1/1 | |
Eth1/10 | 40GbE | Cisco UCS Chassis 2 Fabric Extender (FEX) B | IOM 1/2 | |
Eth1/11 | 40GbE | Cisco UCS Chassis 2 Fabric Extender (FEX) B | IOM 1/3 | |
Eth1/12 | 40GbE | Cisco UCS Chassis 2 Fabric Extender (FEX) B | IOM 1/4 | |
Eth1/13 | 40GbE | Cisco UCS Chassis 3 Fabric Extender (FEX) B | IOM 1/1 | |
Eth1/14 | 40GbE | Cisco UCS Chassis 3 Fabric Extender (FEX) B | IOM 1/2 | |
Eth1/15 | 40GbE | Cisco UCS Chassis 3 Fabric Extender (FEX) B | IOM 1/3 | |
Eth1/16 | 40GbE | Cisco UCS Chassis 3 Fabric Extender (FEX) B | IOM 1/4 | |
Eth1/17 | 40GbE | Cisco UCS Chassis 4 Fabric Extender (FEX) B | IOM 1/1 | |
Eth1/18 | 40GbE | Cisco UCS Chassis 4 Fabric Extender (FEX) B | IOM 1/2 | |
Eth1/19 | 40GbE | Cisco UCS Chassis 4 Fabric Extender (FEX) B | IOM 1/3 | |
Eth1/20 | 40GbE | Cisco UCS Chassis 4 Fabric Extender (FEX) B | IOM 1/4 | |
Eth1/21 | 40GbE | Cisco UCS C240-M5-1 | VIC 1385 Port 2 | |
Eth1/22 | 40GbE | Cisco UCS C240-M5-2 | VIC 1385 Port 2 | |
Eth1/23 | 40GbE | Cisco UCS C240-M5-3 | VIC 1385 Port 2 | |
Eth1/24 | 40GbE | Cisco UCS C240-M5-4 | VIC 1385 Port 2 | |
Eth1/25 | 40GbE | Cisco UCS C240-M5-5 | VIC 1385 Port 2 | |
Eth1/26 | 40GbE | Cisco UCS C240-M5-6 | VIC 1385 Port 2 | |
Eth1/27 | 40GbE | Cisco UCS C240-M5-7 | VIC 1385 Port 2 | |
Eth1/28 | 40GbE | Cisco UCS C240-M5-8 | VIC 1385 Port 2 | |
MGMT0 | GbE | Customer’s Management Switch | Any | |
L1 | GbE | Cisco UCS fabric interconnect A | L1 | |
L2 | GbE | Cisco UCS fabric interconnect A | L2 |
* The ports ETH1/3/1-4 are used with pair Nexus 9000 Switches design option.
Table 5 Cisco UCS Fabric Interconnect A - Cabling Information for Cisco UCS C460 M4 Server(s) Option
Local Device | Local Port | Connection | Remote Device | Remote Port |
Cisco UCS fabric interconnect A
| Eth1/1 | 40GbE | Uplink to Customer Data Switch A | Any |
Eth1/2 | 40GbE | Uplink to Customer Data Switch B | Any | |
Eth1/3/1 | 40GbE QSFP to 4 SFP+ break-out cables | Cisco Nexus 3000 HA | Eth 1/3 | |
Eth1/3/2 | Cisco Nexus 3000 HA | Eth 1/4 | ||
Eth1/3/3 | Cisco Nexus 3000 HA | Eth 1/5 | ||
Eth1/3/4 | Cisco Nexus 3000 HA | Eth 1/6 | ||
Eth1/3/1* | 40GbE QSFP to 4 SFP+ break-out cables | Cisco Nexus 9000 A | Eth 1/3 | |
Eth1/3/2* | Cisco Nexus 9000 A | Eth 1/4 | ||
Eth1/3/3* | Cisco Nexus 9000 B | Eth 1/3 | ||
Eth1/3/4* | Cisco Nexus 9000 B | Eth 1/4 | ||
Eth1/4 | 40GbE |
|
| |
Eth1/5 | 40GbE | Cisco UCS B460-M4-1 | VIC 1385 Port 1 | |
Eth1/6 | 40GbE | Cisco UCS B460-M4-2 | VIC 1385 Port 1 | |
Eth1/7 | 40GbE | Cisco UCS B460-M4-3 | VIC 1385 Port 1 | |
Eth1/8 | 40GbE | Cisco UCS B460-M4-4 | VIC 1385 Port 1 | |
Eth1/9 | 40GbE | Cisco UCS B460-M4-5 | VIC 1385 Port 1 | |
Eth1/10 | 40GbE | Cisco UCS B460-M4-6 | VIC 1385 Port 1 | |
Eth1/11 | 40GbE | Cisco UCS B460-M4-7 | VIC 1385 Port 1 | |
Eth1/12 | 40GbE | Cisco UCS B460-M4-8 | VIC 1385 Port 1 | |
Eth1/13 | 40GbE | Cisco UCS B460-M4-9 | VIC 1385 Port 1 | |
Eth1/14 | 40GbE | Cisco UCS B460-M4-10 | VIC 1385 Port 1 | |
Eth1/15 | 40GbE | Cisco UCS B460-M4-11 | VIC 1385 Port 1 | |
Eth1/16 | 40GbE | Cisco UCS B460-M4-12 | VIC 1385 Port 1 | |
Eth1/17 | 40GbE | Cisco UCS B460-M4-13 | VIC 1385 Port 1 | |
Eth1/18 | 40GbE | Cisco UCS B460-M4-14 | VIC 1385 Port 1 | |
Eth1/19 | 40GbE | Cisco UCS B460-M4-15 | VIC 1385 Port 1 | |
Eth1/20 | 40GbE | Cisco UCS B460-M4-16 | VIC 1385 Port 1 | |
Eth1/21 | 40GbE | Cisco UCS C240-M5-1 | VIC 1385 Port 1 | |
Eth1/22 | 40GbE | Cisco UCS C240-M5-2 | VIC 1385 Port 1 | |
Eth1/23 | 40GbE | Cisco UCS C240-M5-3 | VIC 1385 Port 1 | |
Eth1/24 | 40GbE | Cisco UCS C240-M5-4 | VIC 1385 Port 1 | |
Eth1/25 | 40GbE | Cisco UCS C240-M5-5 | VIC 1385 Port 1 | |
Eth1/26 | 40GbE | Cisco UCS C240-M5-6 | VIC 1385 Port 1 | |
Eth1/27 | 40GbE | Cisco UCS C240-M5-7 | VIC 1385 Port 1 | |
Eth1/28 | 40GbE | Cisco UCS C240-M5-8 | VIC 1385 Port 1 | |
MGMT0 | GbE | Customer’s Management Switch | Any | |
L1 | GbE | Cisco UCS fabric interconnect B | L1 | |
L2 | GbE | Cisco UCS fabric interconnect B | L2 |
* The ports ETH1/3/1-4 are used with pair Nexus 9000 Switches design option.
Table 6 Cisco UCS Fabric Interconnect B - Cabling Information for Cisco UCS C460 M4 Server(s)
Option
Local Device | Local Port | Connection | Remote Device | Remote Port |
Cisco UCS fabric interconnect B
| Eth1/1 | 40GbE | Uplink to Customer Data Switch A | Any |
Eth1/2 | 40GbE | Uplink to Customer Data Switch B | Any | |
Eth1/3/1 | 40GbE QSFP to 4 SFP+ break-out cables | Cisco Nexus 3000 HA | Eth 1/7 | |
Eth1/3/2 | Cisco Nexus 3000 HA | Eth 1/8 | ||
Eth1/3/3 | Cisco Nexus 3000 HA | Eth 1/9 | ||
Eth1/3/4 | Cisco Nexus 3000 HA | Eth 1/10 | ||
Eth1/3/1* | 40GbE QSFP to 4 SFP+ break-out cables | Cisco Nexus 9000 A | Eth 1/5 | |
Eth1/3/2* | Cisco Nexus 9000 A | Eth 1/6 | ||
Eth1/3/3* | Cisco Nexus 9000 B | Eth 1/5 | ||
Eth1/3/4* | Cisco Nexus 9000 B | Eth 1/6 | ||
Eth1/4 | 40GbE |
|
| |
Eth1/5 | 40GbE | Cisco UCS B460-M4-1 | VIC 1385 Port 2 | |
Eth1/6 | 40GbE | Cisco UCS B460-M4-2 | VIC 1385 Port 2 | |
Eth1/7 | 40GbE | Cisco UCS B460-M4-3 | VIC 1385 Port 2 | |
Eth1/8 | 40GbE | Cisco UCS B460-M4-4 | VIC 1385 Port 2 | |
Eth1/9 | 40GbE | Cisco UCS B460-M4-5 | VIC 1385 Port 2 | |
Eth1/10 | 40GbE | Cisco UCS B460-M4-6 | VIC 1385 Port 2 | |
Eth1/11 | 40GbE | Cisco UCS B460-M4-7 | VIC 1385 Port 2 | |
Eth1/12 | 40GbE | Cisco UCS B460-M4-8 | VIC 1385 Port 2 | |
Eth1/13 | 40GbE | Cisco UCS B460-M4-9 | VIC 1385 Port 2 | |
Eth1/14 | 40GbE | Cisco UCS B460-M4-10 | VIC 1385 Port 2 | |
Eth1/15 | 40GbE | Cisco UCS B460-M4-11 | VIC 1385 Port 2 | |
Eth1/16 | 40GbE | Cisco UCS B460-M4-12 | VIC 1385 Port 2 | |
Eth1/17 | 40GbE | Cisco UCS B460-M4-13 | VIC 1385 Port 2 | |
Eth1/18 | 40GbE | Cisco UCS B460-M4-14 | VIC 1385 Port 2 | |
Eth1/19 | 40GbE | Cisco UCS B460-M4-15 | VIC 1385 Port 2 | |
Eth1/20 | 40GbE | Cisco UCS B460-M4-16 | VIC 1385 Port 2 | |
Eth1/21 | 40GbE | Cisco UCS C240-M5-1 | VIC 1385 Port 2 | |
Eth1/22 | 40GbE | Cisco UCS C240-M5-2 | VIC 1385 Port 2 | |
Eth1/23 | 40GbE | Cisco UCS C240-M5-3 | VIC 1385 Port 2 | |
Eth1/24 | 40GbE | Cisco UCS C240-M5-4 | VIC 1385 Port 2 | |
Eth1/25 | 40GbE | Cisco UCS C240-M5-5 | VIC 1385 Port 2 | |
Eth1/26 | 40GbE | Cisco UCS C240-M5-6 | VIC 1385 Port 2 | |
Eth1/27 | 40GbE | Cisco UCS C240-M5-7 | VIC 1385 Port 2 | |
Eth1/28 | 40GbE | Cisco UCS C240-M5-8 | VIC 1385 Port 2 | |
MGMT0 | GbE | Customer’s Management Switch | Any | |
L1 | GbE | Cisco UCS fabric interconnect A | L1 | |
L2 | GbE | Cisco UCS fabric interconnect A | L2 |
* The ports ETH1/3/1-4 are used with pair Nexus 9000 Switches design option.
Table 7 Cisco Nexus 3000 Cabling Information
Local Device | Local Port | Connection | Remote Device | Remote Port |
Cisco Nexus 3000 HA | Eth1/3 | 10GbE | Cisco UCS fabric interconnect A | Eth1/3/1 |
Eth1/4 | 10GbE | Cisco UCS fabric interconnect A | Eth1/3/2 | |
Eth1/5 | 10GbE | Cisco UCS fabric interconnect A | Eth1/3/3 | |
Eth1/6 | 10GbE | Cisco UCS fabric interconnect A | Eth1/3/4 | |
Eth1/7 | 10GbE | Cisco UCS fabric interconnect B | Eth1/3/1 | |
Eth1/8 | 10GbE | Cisco UCS fabric interconnect B | Eth1/3/2 | |
Eth1/9 | 10GbE | Cisco UCS fabric interconnect B | Eth1/3/3 | |
Eth1/10 | 10GbE | Cisco UCS fabric interconnect B | Eth1/3/4 | |
MGMT0 | GbE | Customer’s Management Switch | Any |
Table 8 Cisco Nexus 9000-A Cabling Information
Local Device | Local Port | Connection | Remote Device | Remote Port |
Cisco Nexus 9000 A
| Eth1/1 | 10GbE | Uplink to Customer Data Switch A | Any |
Eth1/2 | 10GbE | Uplink to Customer Data Switch B | Any | |
Eth1/3 | 10GbE | Cisco UCS fabric interconnect A | Eth1/3/1 | |
Eth1/4 | 10GbE | Cisco UCS fabric interconnect A | Eth1/3/2 | |
Eth1/5 | 10GbE | Cisco UCS fabric interconnect B | Eth1/3/1 | |
Eth1/6 | 10GbE | Cisco UCS fabric interconnect B | Eth1/3/2 | |
Eth1/9* | 10GbE | Cisco Nexus 9000 B | Eth1/9 | |
Eth1/10* | 10GbE | Cisco Nexus 9000 B | Eth1/10 | |
Eth1/11* | 10GbE | Cisco Nexus 9000 B | Eth1/11 | |
Eth1/12* | 10GbE | Cisco Nexus 9000 B | Eth1/12 | |
MGMT0 | GbE | Customer’s Management Switch | Any |
* The ports ETH1/9-12 can be replaced with E2/1 and E2/2 for 40G connectivity.
For devices requiring GbE connectivity, use the GbE Copper SFP+s (GLC–T=).
Table 9 Cisco Nexus 9000-B Cabling Information
Local Device | Local Port | Connection | Remote Device | Remote Port |
Cisco Nexus 9000 B
| Eth1/1 | 10GbE | Uplink to Customer Data Switch A | Any |
Eth1/2 | 10GbE | Uplink to Customer Data Switch B | Any | |
Eth1/3 | 10GbE | Cisco UCS fabric interconnect A | Eth1/3/3 | |
Eth1/4 | 10GbE | Cisco UCS fabric interconnect A | Eth1/3/4 | |
Eth1/5 | 10GbE | Cisco UCS fabric interconnect B | Eth1/3/3 | |
Eth1/6 | 10GbE | Cisco UCS fabric interconnect B | Eth1/3/4 | |
Eth1/9* | 10GbE | Cisco Nexus 9000 A | Eth1/9 | |
Eth1/10* | 10GbE | Cisco Nexus 9000 A | Eth1/10 | |
Eth1/11* | 10GbE | Cisco Nexus 9000 A | Eth1/11 | |
Eth1/12* | 10GbE | Cisco Nexus 9000 A | Eth1/12 | |
MGMT0 | GbE | Customer’s Management Switch | Any |
* The ports ETH1/9-12 can be replaced with E2/1 and E2/2 for 40G connectivity.
For devices requiring GbE connectivity, use the GbE Copper SFP+s (GLC–T=).
Table 10 details the software revisions used for validating various components of the Cisco UCS B/C460 M4 Scale-Out for SAP HANA.
Table 10 Hardware and Software Components of the Cisco UCS B/C460 M4 Scale-Out for SAP HANA
Vendor | Product | Version | Description |
Cisco | Cisco UCSM | 3.1(2b) | Cisco UCS Manager |
Cisco | Cisco UCS 6332 UP FI | 5.0(3)N2(3.12b) | Cisco UCS Fabric Interconnects |
Cisco | Cisco UCS 5108 Blade Chassis | NA | Cisco UCS Blade Server Chassis |
Cisco | Cisco UCS 2304 XP FEX | 3.1(2b) | Cisco UCS Fabric Extenders for Blade Server chassis |
Cisco | Cisco UCS B-Series M5 Servers | 3.1(2b) | Cisco B-Series M4 Blade Servers |
Cisco | Cisco UCS C-Series M5 Servers | 3.1(22a) – CIMC Controller | Cisco C-Series M4 Blade Servers |
Cisco | Cisco UCS C240 M5 Servers | 3.1(22a) – CIMC Controller | Cisco C240 M4 Rack Servers for Management |
Cisco | Cisco UCS VIC 1340/1380 | 4.1(2d) | Cisco UCS VIC 1240/1280 Adapters |
Cisco | Cisco UCS VIC 1385 | 4.1(2d) | Cisco UCS VIC Adapter |
Cisco | Cisco Nexus 9372PX Switches | 7.0(3)I5(1) | Cisco Nexus 9372PX Switches |
Cisco | Cisco Nexus 3524 Switches | 6.0(2)A7(2) | Cisco Nexus N3K-C3524P Switch |
SUSE | SLES 12 / SLES for SAP Applications 12 | 12 SP2 (64 bit) | Operating System to host SAP HANA |
MapR | MapR Converged Enterprise Edition | 5.2.039122.GA | MapR Converged Enterprise Edition |
The following section provides a detailed procedure for configuring the Cisco Nexus 3524 Switch for High Availability in the SAP HANA environment. The switch configuration in this section is based on the cabling plan described in the Device Cabling section. If the systems are connected on different ports, configure the switches accordingly by following the guidelines described below.
These steps provide the details for the initial Cisco Nexus 3524 Series Switch setup.
To set up the initial configuration for the first Cisco Nexus switch, complete the following steps:
On initial boot and connection to the serial or console port of the switch, the NX-OS setup should automatically start and attempt to enter Power on Auto Provisioning.
Abort Power On Auto Provisioning and continue with normal setup ?(yes/no)[n]:yes
---- System Admin Account Setup ----
Do you want to enforce secure password standard (yes/no): yes
Enter the password for "admin":
Confirm the password for "admin":
---- Basic System Configuration Dialog ----
This setup utility will guide you through the basic configuration of
the system. Setup configures only enough connectivity for management
of the system.
Please register Cisco Nexus 3500 Family devices promptly with your
supplier. Failure to register may affect response times for initial
service calls. Nexus devices must be registered to receive entitled
support services.
Press Enter at anytime to skip a dialog. Use ctrl-c at anytime
to skip the remaining dialogs.
Would you like to enter the basic configuration dialog (yes/no): yes
Create another login account (yes/no) [n]:
Configure read-only SNMP community string (yes/no) [n]:
Configure read-write SNMP community string (yes/no) [n]:
Enter the switch name : <<var_nexus_HA_hostname>>
Continue with Out-of-band (mgmt0) management configuration? (yes/no) [y]:
Mgmt0 IPv4 address : <<var_nexus_HA_mgmt0_ip>>
Mgmt0 IPv4 netmask : <<var_nexus_HA_mgmt0_netmask>>
Configure the default gateway? (yes/no) [y]:
IPv4 address of the default gateway : <<var_nexus_HA_mgmt0_gw>>
Enable the telnet service? (yes/no) [n]:
Enable the ssh service? (yes/no) [y]:
Type of ssh key you would like to generate (dsa/rsa) : rsa
Number of key bits <768-2048> : 2048
Configure the ntp server? (yes/no) [n]: y
NTP server IPv4 address : <<var_global_ntp_server_ip>>
Configure default interface layer (L3/L2) [L2]:
Configure default switchport interface state (shut/noshut) [noshut]:
Configure CoPP System Policy Profile ( default / l2 / l3 ) [default]:
The following configuration will be applied:
switchname <<var_nexus_HA_hostname>>
interface mgmt0
ip address <<var_nexus_HA_mgmt0_ip>> <<var_nexus_HA_mgmt0_netmask>>
vrf context management
ip route 0.0.0.0/0 <<var_nexus_A_mgmt0_gw>>
no shutdown
no telnet server enable
ssh key rsa 2048 force
ssh server enable
ntp server <<var_global_ntp_server_ip>>
system default switchport
no system default switchport shutdown
policy-map type control-plane copp-system-policy ( default )
Would you like to edit the configuration? (yes/no) [n]:
Use this configuration and save it? (yes/no) [y]:
[########################################] 100%
Copy complete, now saving to disk (please wait)...
The following commands enable IP switching feature and set default spanning tree behaviors:
1. On each Nexus 3524, enter configuration mode:
config terminal
2. Use the following commands to enable the necessary features:
feature lacp
feature interface-vlan
feature lldp
3. Save the running configuration to start-up:
copy run start
policy-map type network-qos jumbo
class type network-qos class-default
mtu 9216
system qos
service-policy type network-qos jumbo
To create the necessary VLANs, complete the following step on both switches:
1. From the configuration mode, run the following commands:
vlan <<var_storage_vlan_id>>
name HANA-Storage
vlan <<var_internal_vlan_id>>
name HANA-Internal
vlan <<var_inband_vlan_id>>
name NFS-IPMI
vlan <<var_mapr-01_vlan_id>>
name MapR-01
vlan <<var_mapr-02_vlan_id>>
name MapR-02
vlan <<var_mapr-03_vlan_id>>
name MapR-03
1. Define a port description for the interface connecting to <<var_ucs_clustername>>-A.
interface Eth1/3-6
description <<var_ucs_clustername>>-A:1/3
2. Apply it to a port channel and bring up the interface.
interface eth1/3-6
channel-group 3 mode active
no shutdown
3. Define a description for the port channel connecting to <<var_ucs_clustername>>-A.
interface Po3
description <<var_ucs_clustername>>-A
4. Make the port channel a switchport, and configure a trunk to allow internal HANA VLANs.
switchport
switchport mode trunk
switchport trunk allowed vlan <<var_storage_vlan_id>>,<<var_internal_vlan_id>>,<<var_inband_vlan_id>>,
<<var_mapr-01_vlan_id>>,<<var_mapr-02_vlan_id>>,<<var_mapr-03_vlan_id>>
5. Make the port channel and associated interfaces spanning tree edge ports.
spanning-tree port type edge trunk
spanning-tree bpduguard enable
6. Bring up the port-channel.
no shutdown
7. Define a port description for the interface connecting to <<var_ucs_clustername>>-B.
interface Eth1/7-10
description <<var_ucs_clustername>>-B:1/3
8. Apply it to a port channel and bring up the interface.
interface Eth1/7-10
channel-group 4 mode active
no shutdown
9. Define a description for the port-channel connecting to <<var_ucs_clustername>>-B.
interface Po4
description <<var_ucs_clustername>>-B
10. Make the port-channel a switchport, and configure a trunk to allow internal HANA VLANs.
switchport
switchport mode trunk
switchport trunk allowed vlan <<var_storage_vlan_id>>,<<var_internal_vlan_id>>,<<var_inband_vlan_id>>,
<<var_mapr-01_vlan_id>>,<<var_mapr-02_vlan_id>>,<<var_mapr-03_vlan_id>>
11. Make the port channel and associated interfaces spanning tree edge ports.
spanning-tree port type edge trunk
spanning-tree bpduguard enable
12. Bring up the port channel.
no shutdown
13. Save the running configuration to start-up configuration.
copy run start
The following section provides a detailed procedure for configuring the Cisco Nexus 9000 Switches for SAP HANA environment. The switch configuration in this section based on cabling plan described in the Device Cabling section. If the systems connected on different ports, configure the switches accordingly following the guidelines described in this section.
The configuration steps detailed in this section provides guidance for configuring the Cisco Nexus 9000 running r lease 7.0(3)I5(1) within a multi-VDC environment.
These steps provide the details for the initial Cisco Nexus 9000 Series Switch setup.
To set up the initial configuration for the first Cisco Nexus switch, complete the following steps:
On initial boot and connection to the serial or console port of the switch, the NX-OS setup should automatically start and attempt to enter Power on Auto Provisioning.
---- Basic System Configuration Dialog VDC: 1 ----
This setup utility will guide you through the basic configuration of
the system. Setup configures only enough connectivity for management
of the system.
*Note: setup is mainly used for configuring the system initially,
when no configuration is present. So setup always assumes system
defaults and not the current system configuration values.
Press Enter at anytime to skip a dialog. Use ctrl-c at anytime
to skip the remaining dialogs.
Would you like to enter the basic configuration dialog (yes/no): yes
Do you want to enforce secure password standard (yes/no) [y]:
Create another login account (yes/no) [n]:
Configure read-only SNMP community string (yes/no) [n]:
Configure read-write SNMP community string (yes/no) [n]:
Enter the switch name : <<var_nexus_A_hostname>>
Continue with Out-of-band (mgmt0) management configuration? (yes/no) [y]:
Mgmt0 IPv4 address : <<var_nexus_A_mgmt0_ip>>
Mgmt0 IPv4 netmask : <<var_nexus_A_mgmt0_netmask>>
Configure the default gateway? (yes/no) [y]:
IPv4 address of the default gateway : <<var_nexus_A_mgmt0_gw>>
Configure advanced IP options? (yes/no) [n]:
Enable the telnet service? (yes/no) [n]:
Enable the ssh service? (yes/no) [y]:
Type of ssh key you would like to generate (dsa/rsa) [rsa]:
Number of rsa key bits <1024-2048> [2048]:
Configure the ntp server? (yes/no) [n]: y
NTP server IPv4 address : <<var_global_ntp_server_ip>>
Configure CoPP system profile (strict/moderate/lenient/dense/skip) [strict]:
The following configuration will be applied:
password strength-check
switchname <<var_nexus_A_hostname>>
vrf context management
ip route 0.0.0.0/0 <<var_nexus_A_mgmt0_gw>>
exit
no feature telnet
ssh key rsa 2048 force
feature ssh
ntp server <<var_global_ntp_server_ip>>
copp profile strict
interface mgmt0
ip address <<var_nexus_A_mgmt0_ip>> <<var_nexus_A_mgmt0_netmask>>
no shutdown
Would you like to edit the configuration? (yes/no) [n]: Enter
Use this configuration and save it? (yes/no) [y]: Enter
[########################################] 100%
Copy complete.
To set up the initial configuration for the second Cisco Nexus switch complete the following steps:
On initial boot and connection to the serial or console port of the switch, the NX-OS setup should automatically start and attempt to enter Power on Auto Provisioning.
---- Basic System Configuration Dialog VDC: 1 ----
This setup utility will guide you through the basic configuration of
the system. Setup configures only enough connectivity for management
of the system.
*Note: setup is mainly used for configuring the system initially,
when no configuration is present. So setup always assumes system
defaults and not the current system configuration values.
Press Enter at anytime to skip a dialog. Use ctrl-c at anytime
to skip the remaining dialogs.
Would you like to enter the basic configuration dialog (yes/no): yes
Create another login account (yes/no) [n]:
Configure read-only SNMP community string (yes/no) [n]:
Configure read-write SNMP community string (yes/no) [n]:
Enter the switch name : <<var_nexus_B_hostname>>
Continue with Out-of-band (mgmt0) management configuration? (yes/no) [y]:
Mgmt0 IPv4 address : <<var_nexus_B_mgmt0_ip>>
Mgmt0 IPv4 netmask : <<var_nexus_B_mgmt0_netmask>>
Configure the default gateway? (yes/no) [y]:
IPv4 address of the default gateway : <<var_nexus_B_mgmt0_gw>>
Configure advanced IP options? (yes/no) [n]:
Enable the telnet service? (yes/no) [n]:
Enable the ssh service? (yes/no) [y]:
Type of ssh key you would like to generate (dsa/rsa) [rsa]:
Number of rsa key bits <1024-2048> [2048]:
Configure the ntp server? (yes/no) [n]: y
NTP server IPv4 address : <<var_global_ntp_server_ip>>
Configure default interface layer (L3/L2) [L3]: L2
Configure default switchport interface state (shut/noshut) [shut]: Enter
Configure CoPP system profile (strict/moderate/lenient/dense/skip) [strict]:
The following configuration will be applied:
password strength-check
switchname <<var_nexus_B_hostname>>
vrf context management
ip route 0.0.0.0/0 <<var_nexus_B_mgmt0_gw>>
exit
no feature telnet
ssh key rsa 2048 force
feature ssh
ntp server <<var_global_ntp_server_ip>>
copp profile strict
interface mgmt0
ip address <<var_nexus_B_mgmt0_ip>> <<var_nexus_B_mgmt0_netmask>>
no shutdown
Would you like to edit the configuration? (yes/no) [n]: Enter
Use this configuration and save it? (yes/no) [y]: Enter
[########################################] 100%
Copy complete.
The following commands enable the IP switching feature and set the default spanning tree behaviors:
1. On each Nexus 9000, enter configuration mode:
config terminal
2. Use the following commands to enable the necessary features:
feature udld
feature lacp
feature vpc
feature interface-vlan
feature lldp
3. Configure spanning tree defaults:
spanning-tree port type network default
spanning-tree port type edge bpduguard default
spanning-tree port type edge bpdufilter default
4. Save the running configuration to start-up:
copy run start
To create the necessary VLANs, complete the following step on both switches:
1. From the configuration mode, run the following commands:
vlan <<var_storage_vlan_id>>
name HANA-Storage
vlan <<var_mgmt_vlan_id>>
name Management
vlan <<var_internal_vlan_id>>
name HANA-Internal
vlan <<var_backup_vlan_id>>
name HANA-Backup
vlan <<var_client_vlan_id>>
name HANA-Client
vlan <<var_appserver_vlan_id>>
name HANA-AppServer
vlan <<var_datasource_vlan_id>>
name HANA-DataSource
vlan <<var_replication_vlan_id>>
name HANA-Replication
vlan <<var_inband_vlan_id>>
name NFS-IPMI
vlan <<var_mapr-01_vlan_id>>
name MapR-01
vlan <<var_mapr-02_vlan_id>>
name MapR-02
vlan <<var_mapr-03_vlan_id>>
name MapR-03
To configure vPCs for switch A, complete the following steps:
1. From the global configuration mode, create a new vPC domain:
vpc domain <<var_nexus_vpc_domain_id>>
2. Make Nexus 9000A the primary vPC peer by defining a low priority value:
role priority 10
3. Use the management interfaces on the supervisors of the Nexus 9000s to establish a keepalive link:
peer-keepalive destination <<var_nexus_B_mgmt0_ip>> source <<var_nexus_A_mgmt0_ip>>
4. Enable following features for this vPC domain:
peer-switch
delay restore 150
peer-gateway
auto-recovery
To configure vPCs for switch B, complete the following steps:
1. From the global configuration mode, create a new vPC domain:
vpc domain <<var_nexus_vpc_domain_id>>
2. Make Cisco Nexus 9000 B the secondary vPC peer by defining a higher priority value than that of the Nexus 9000 A:
role priority 20
3. Use the management interfaces on the supervisors of the Cisco Nexus 9000s to establish a keepalive link:
peer-keepalive destination <<var_nexus_A_mgmt0_ip>> source <<var_nexus_B_mgmt0_ip>>
4. Enable following features for this vPC domain:
peer-switch
delay restore 150
peer-gateway
auto-recovery
1. Define a port description for the interfaces connecting to VPC Peer <<var_nexus_B_hostname>>.
interface Eth2/1
description VPC Peer <<var_nexus_B_hostname>>:2/1
interface Eth2/2
description VPC Peer <<var_nexus_B_hostname>>:2/2
2. Apply a port channel to both VPC Peer links and bring up the interfaces.
interface Eth2/1-2
channel-group 1 mode active
no shutdown
3. Define a description for the port-channel connecting to <<var_nexus_B_hostname>>.
interface Po1
description vPC peer-link
4. Make the port channel a switchport, and configure a trunk to allow HANA VLANs.
switchport
switchport mode trunk
switchport trunk allowed vlan
<<var_storage_vlan_id>>,<<var_mgmt_vlan_id>>,<<var_boot_vlan_id>>,<<var_internal_vlan_id>>,<<var_inband_vlan_
id>>,<<var_backup_vlan_id>>, <<var_client_vlan_id>>, <<var_appserver_vlan_id>>, <<var_datasource_vlan_id>>,
<<var_replication_vlan_id>>
5. Make this port-channel the VPC peer link and bring it up.
spanning-tree port type network
vpc peer-link
no shutdown
1. Define a port description for the interfaces connecting to VPC peer <<var_nexus_A_hostname>>.
interface Eth2/1
description VPC Peer <<var_nexus_A_hostname>>:2/1
interface Eth2/2
description VPC Peer <<var_nexus_A_hostname>>:2/2
2. Apply a port channel to both VPC peer links and bring up the interfaces.
interface Eth2/1-2
channel-group 1 mode active
no shutdown
3. Define a description for the port-channel connecting to <<var_nexus_A_hostname>>.
interface Po1
description vPC peer-link
4. Make the port-channel a switchport, and configure a trunk to allow HANA VLANs.
switchport
switchport mode trunk
switchport trunk allowed vlan
<<var_storage_vlan_id>>,<<var_mgmt_vlan_id>>,<<var_boot_vlan_id>>,<<var_internal_vlan_id>>,<<var_inband_vlan_
id>>,<<var_backup_vlan_id>>, <<var_client_vlan_id>>, <<var_appserver_vlan_id>>, <<var_datasource_vlan_id>>,
<<var_replication_vlan_id>>
5. Make this port-channel the VPC peer link and bring it up.
spanning-tree port type network
vpc peer-link
no shutdown
1. Define a port description for the interface connecting to <<var_ucs_clustername>>-A.
interface Eth1/3-4
description <<var_ucs_clustername>>-A:1/3
2. Apply it to a port channel and bring up the interface.
interface eth1/3-4
channel-group 11 mode active
no shutdown
3. Define a description for the port-channel connecting to <<var_ucs_clustername>>-A.
interface Po11
description <<var_ucs_clustername>>-A
4. Make the port-channel a switchport, and configure a trunk to allow all HANA VLANs.
switchport
switchport mode trunk
switchport trunk allowed vlan
<<var_storage_vlan_id>>,<<var_mgmt_vlan_id>>,<<var_boot_vlan_id>>,<<var_internal_vlan_id>>,<<var_inband_vlan_
id>>,<<var_backup_vlan_id>>, <<var_client_vlan_id>>, <<var_appserver_vlan_id>>, <<var_datasource_vlan_id>>,
<<var_replication_vlan_id>>
5. Make the port channel and associated interfaces spanning tree edge ports.
spanning-tree port type edge trunk
6. Set the MTU to be 9216 to support jumbo frames.
mtu 9216
7. Make this a VPC port-channel and bring it up.
vpc 11
no shutdown
8. Define a port description for the interface connecting to <<var_ucs_clustername>>-B.
interface Eth1/5-6
description <<var_ucs_clustername>>-B:1/3
9. Apply it to a port channel and bring up the interface.
interface Eth1/5-6
channel-group 12 mode active
no shutdown
10. Define a description for the port-channel connecting to <<var_ucs_clustername>>-B.
interface Po12
description <<var_ucs_clustername>>-B
11. Make the port-channel a switchport, and configure a trunk to allow all HANA VLANs.
switchport
switchport mode trunk
switchport trunk allowed vlan
<<var_storage_vlan_id>>,<<var_mgmt_vlan_id>>,<<var_boot_vlan_id>>,<<var_internal_vlan_id>>,<<var_inband_vlan_
id>>,<<var_backup_vlan_id>>, <<var_client_vlan_id>>, <<var_appserver_vlan_id>>, <<var_datasource_vlan_id>>,
<<var_replication_vlan_id>>
12. Make the port channel and associated interfaces spanning tree edge ports.
spanning-tree port type edge trunk
13. 53. Set the MTU to be 9216 to support jumbo frames.
mtu 9216
14. Make this a VPC port-channel and bring it up.
vpc 12
no shutdown
You can define the port-channel for each type Network to have dedicated bandwidth. Below is an example to create such port-channel for Backup Network, these cables are connected to Storage for Backup. The following example assumes two ports (Ethernet 1/29 and 1/30) are connected to dedicated NFS Storage to backup SAP HANA.
1. Define a port description for the interface connecting to <<var_node01>>.
interface Eth1/29
description <<var_backup_node01>>:<<Port_Number>>
2. Apply it to a port channel and bring up the interface.
interface eth1/29
channel-group 21 mode active
no shutdown
3. Define a description for the port-channel connecting to <<var_backup_node01>>.
interface Po21
description <<var_backup_vlan_id>>
4. Make the port-channel a switchport, and configure a trunk to allow NFS VLAN for DATA.
switchport
switchport mode trunk
switchport trunk allowed vlan <<var_backup_vlan_id>>
5. Make the port channel and associated interfaces spanning tree edge ports.
spanning-tree port type edge trunk
6. Set the MTU to be 9216 to support jumbo frames.
mtu 9216
7. Make this a VPC port-channel and bring it up.
vpc 21
no shutdown
8. Define a port description for the interface connecting to <<var_node02>>.
interface Eth1/30
description <<var_backup_node01>>:<<Port_Number>>
9. Apply it to a port channel and bring up the interface.
channel-group 22 mode active
no shutdown
10. Define a description for the port-channel connecting to <<var_node02>>.
interface Po22
description <<var_backup_node02>>
11. Make the port-channel a switchport, and configure a trunk to allow NFS VLAN for DATA.
switchport
switchport mode trunk
switchport trunk allowed vlan <<var_backup_vlan_id>>
12. Make the port channel and associated interfaces spanning tree edge ports.
spanning-tree port type edge trunk
13. Set the MTU to be 9216 to support jumbo frames.
mtu 9216
14. Make this a VPC port-channel and bring it up.
vpc 22
no shutdown
This is an optional step, which can be used to implement a management plane access for the Cisco UCS servers.
To enable management access across the IP switching environment, complete the following steps:
You may want to create a dedicated Switch Virtual Interface (SVI) on the Nexus data plane to test and troubleshoot the management plane. If an L3 interface is deployed be sure it is deployed on both Cisco Nexus 9000s to ensure Type-2 VPC consistency.
1. Define a port description for the interface connecting to the management plane.
interface Eth1/<<interface_for_in_band_mgmt>>
description IB-Mgmt:<<mgmt_uplink_port>>
2. Apply it to a port channel and bring up the interface.
channel-group 6 mode active
no shutdown
3. Define a description for the port-channel connecting to management switch.
interface Po6
description IB-Mgmt
4. Configure the port as an access VLAN carrying the InBand management VLAN traffic.
switchport
switchport mode access
switchport access vlan <<var_inband_vlan_id>>
5. Make the port channel and associated interfaces normal spanning tree ports.
spanning-tree port type normal
6. Make this a VPC port-channel and bring it up.
vpc 6
no shutdown
7. Save the running configuration to start-up in both Nexus 9000s.
copy run start
Depending on the available network infrastructure, several methods and features can be used to uplink the SAP HANA environment. If an existing Cisco Nexus environment is present, Cisco recommends using vPCs to uplink the Cisco Nexus 9000 switches in the SAP HANA environment to the existing infrastructure. The previously described procedures can be used to create an uplink vPC to the existing environment. Make sure to run copy run start to save the configuration on each switch after configuration is completed.
This section describes the specific configurations on Cisco UCS servers to address SAP HANA requirements.
This section provides detailed procedures for configuring the Cisco Unified Computing System (Cisco UCS) for use in SAP HANA Scale Out Solution environment. These steps are necessary to provision the Cisco UCS C-Series and B-Series servers to meet SAP HANA requirement.
To configure the Cisco UCS Fabric Interconnect A, complete the following steps:
1. Connect to the console port on the first Cisco UCS 6300 Fabric Interconnect.
Enter the configuration method: console
Enter the setup mode; setup newly or restore from backup.(setup/restore)? setup
You have chosen to setup a a new fabric interconnect? Continue? (y/n): y
Enforce strong passwords? (y/n) [y]: y
Enter the password for "admin": <<var_password>>
Enter the same password for "admin": <<var_password>>
Is this fabric interconnect part of a cluster (select 'no' for standalone)?
(yes/no) [n]: y
Which switch fabric (A|B): A
Enter the system name: <<var_ucs_clustername>>
Physical switch Mgmt0 IPv4 address: <<var_ucsa_mgmt_ip>>
Physical switch Mgmt0 IPv4 netmask: <<var_ucsa_mgmt_mask>>
IPv4 address of the default gateway: <<var_ucsa_mgmt_gateway>>
Cluster IPv4 address: <<var_ucs_cluster_ip>>
Configure DNS Server IPv4 address? (yes/no) [no]: y
DNS IPv4 address: <<var_nameserver_ip>>
Configure the default domain name? y
Default domain name: <<var_dns_domain_name>>
Join centralized management environment (UCS Central)? (yes/no) [n]: Enter
2. Review the settings printed to the console. If they are correct, answer yes to apply and save the configuration.
3. Wait for the login prompt to make sure that the configuration has been saved.
To configure the Cisco UCS Fabric Interconnect B, complete the following steps:
1. Connect to the console port on the second Cisco UCS 6332 Fabric Interconnect.
Enter the configuration method: console
Installer has detected the presence of a peer Fabric interconnect. This Fabric
interconnect will be added to the cluster. Do you want to continue {y|n}? y
Enter the admin password for the peer fabric interconnect: <<var_password>>
Physical switch Mgmt0 IPv4 address: <<var_ucsb_mgmt_ip>>
Apply and save the configuration (select ‘no’ if you want to re-enter)? (yes/no):
y
2. Wait for the login prompt to make sure that the configuration has been saved.
To log in to the Cisco Unified Computing System (UCS) environment, complete the following steps:
1. Open a web browser and navigate to the Cisco UCS 6332 Fabric Interconnect cluster address.
2. Click the Launch Cisco UCS Manager.
3. If prompted to accept security certificates, accept as necessary.
4. When prompted, enter admin as the user name and enter the administrative password.
5. Click Login to log in to the Cisco UCS Manager.
This document assumes you are using Cisco UCS Manager Software version 3.1(2b). To upgrade the Cisco UCS Manager software and the Cisco UCS 6332 Fabric Interconnect software to version 3.1(2b), refer to Cisco UCS Manager Install and Upgrade Guides.
To create a block of IP addresses for server Keyboard, Video, Mouse (KVM) access in the Cisco UCS environment, complete the following steps:
This block of IP addresses should be in the same subnet as the management IP addresses for the Cisco UCS Manager.
1. In Cisco UCS Manager, click the LAN tab in the navigation pane.
2. Select Pools > root > IP Pools > IP Pool ext-mgmt.
3. In the Actions pane, select Create Block of IP Addresses.
4. Enter the starting IP address of the block and the number of IP addresses required, and the subnet and gateway information.
5. Click OK to create the IP block.
6. Click OK in the confirmation message.
To synchronize the Cisco UCS environment to the NTP server, complete the following steps:
1. In Cisco UCS Manager, click the Admin tab in the navigation pane.
2. Select All > Timezone Management.
3. In the Properties pane, select the appropriate time zone in the Timezone menu.
4. Click Save Changes, and then click OK.
5. Click Add NTP Server.
6. Enter <<var_global_ntp_server_ip>> and click OK.
7. Click OK.
For the Cisco UCS 6300 Series Fabric Extenders, two configuration options are available: pinning and portchannel. SAP HANA node communicates with every other SAP HANA node using multiple I/O streams and this makes the port-channel option a highly suitable configuration.
Setting the discovery policy simplifies adding the Cisco UCS B-Series chassis and additional fabric extenders for C-Series connectivity.
To modify the chassis discovery policy, complete the following steps:
1. In Cisco UCS Manager, click the Equipment tab in the navigation pane and select Equipment in the list on the left.
2. In the right pane, click the Policies tab.
3. Under Global Policies, set the Chassis/FEX Discovery Policy to match the number of server ports that are connected between the fabric extenders (FEXes) within the chassis and the fabric interconnects.
4. Set the Link Grouping Preference to Port Channel.
5. Click Save Changes.
6. Click OK.
To enable server and uplink ports, complete the following steps:
1. In Cisco UCS Manager, click the Equipment tab in the navigation pane.
2. Select Equipment > Fabric Interconnects > Fabric Interconnect A (primary) > Fixed Module.
3. Expand Ethernet Ports.
4. Select the ports that are connected to the chassis and / or to the Cisco C-Series Server (two per FI),right-click them, and select Configure as Server Port.
5. Click Yes to confirm server ports and click OK.
6. Verify that the ports connected to the chassis and / or to the Cisco C-Series Server are now configured as server ports.
7. Select ports that are connected to the Cisco Nexus switches, right-click them, and select Configure as Uplink Port.
8. Click Yes to confirm uplink ports and click OK.
9. Select Equipment > Fabric Interconnects > Fabric Interconnect B (subordinate) > Fixed Module.
10. Expand Ethernet Ports.
11. Select the ports that are connected to the chassis or to the Cisco C-Series Server (two per FI), right-click them, and select Configure as Server Port.
12. Click Yes to confirm server ports and click OK.
13. Select ports that are connected to the Cisco Nexus switches, right-click them, and select Configure as Uplink Port.
14. Click Yes to confirm the uplink ports and click OK.
The 40 GB Ethernet ports on Cisco UCS 6300 Fabric Interconnects can be configured as four breakout 10 GB ports using a supported breakout cable.
Configuring breakout ports requires rebooting the Fabric Interconnect. Any existing configuration on a port is erased. It is recommended to break out all required ports in a single transaction.
1. In Cisco UCS Manager, click the Equipment tab in the navigation pane.
2. Select Equipment > Fabric Interconnects > Fabric Interconnect A (primary) > Fixed Module.
3. Expand Ethernet Ports.
4. Select ports that are connected to the Cisco Nexus 3524 switches, right-click them, and select Configure Breakout Port.
5. Click Yes to confirm the Breakout ports and click OK.
6. Select Equipment > Fabric Interconnects > Fabric Interconnect B (subordinate) > Fixed Module.
7. Expand Ethernet Ports.
8. Select ports that are connected to the Cisco Nexus 3524 switches, right-click them, and select Configure Breakout Port.
9. Click Yes to confirm the Breakout ports and click OK.
When you configure a breakout port, you can configure each 10 GB sub-port as server, uplink, FCoE uplink, FCoE storage or appliance as required. Unified Ports cannot be configured as Breakout Ports.
To acknowledge all Cisco UCS chassis and Rack Mount Servers, complete the following steps:
1. In Cisco UCS Manager, click the Equipment tab in the navigation pane.
2. Expand Chassis and select each chassis that is listed.
3. Right-click each chassis and select Acknowledge Chassis.
4. Click Yes and then click OK to complete acknowledging the chassis.
5. If C-Series servers are part of the configuration, expand Rack Mounts and FEX.
6. Right-click each Server that is listed and select Acknowledge Server.
7. Click Yes and then click OK to complete acknowledging the Rack Mount Servers.
A separate uplink port channel for each of the network zones are defined as per SAP. For example, create port-channel 11 on fabric interconnect A and port-channel 12 on fabric interconnect B for Client zone network.
To configure the necessary port channels out of the Cisco UCS environment, complete the following steps:
1. In Cisco UCS Manager, click the LAN tab in the navigation pane.
2. In this procedure, two port channels are created: one from fabric A to both Cisco Nexus switches and one from fabric B to both Cisco Nexus switches.
3. Under LAN > LAN Cloud, expand the Fabric A tree.
4. Right-click Port Channels.
5. Select Create Port Channel.
6. Enter 11 as the unique ID of the port channel.
7. Enter vPC-11-Nexus as the name of the port channel.
8. Click Next.
9. Select the following ports to be added to the port channel:
- Slot ID 1 and port 1
- Slot ID 1 and port 2
10. If the breakout cables are used for Uplink connectivity. Select the following ports to be added to the port channel:
- Slot ID 1, Aggregated Port ID 1 and port 1
- Slot ID 1, Aggregated Port ID 1 and port 2
- Slot ID 1, Aggregated Port ID 1 and port 3
- Slot ID 1, Aggregated Port ID 1 and port 4
- Slot ID 1, Aggregated Port ID 2 and port 1
- Slot ID 1, Aggregated Port ID 2 and port 2
- Slot ID 1, Aggregated Port ID 2 and port 3
- Slot ID 1, Aggregated Port ID 2 and port 4
11. Click >> to add the ports to the port channel.
12. Click Finish to create the port channel.
13. Click OK.
14. In the navigation pane, under LAN > LAN Cloud, expand the fabric B tree.
15. Right-click Port Channels.
16. Select Create Port Channel.
17. Enter 12 as the unique ID of the port channel.
18. Enter vPC-12-NEXUS as the name of the port channel.
19. Click Next.
20. Select the following ports to be added to the port channel:
- Slot ID 1 and port 1
- Slot ID 1 and port 2
21. If the breakout cables are used for Uplink connectivity. Select the following ports to be added to the port channel:
- Slot ID 1, Aggregated Port ID 1 and port 1
- Slot ID 1, Aggregated Port ID 1 and port 2
- Slot ID 1, Aggregated Port ID 1 and port 3
- Slot ID 1, Aggregated Port ID 1 and port 4
- Slot ID 1, Aggregated Port ID 2 and port 1
- Slot ID 1, Aggregated Port ID 2 and port 2
- Slot ID 1, Aggregated Port ID 2 and port 3
- Slot ID 1, Aggregated Port ID 2 and port 4
22. Click >> to add the ports to the port channel.
23. Click Finish to create the port channel.
24. Click OK.
Repeat the steps 1-24 to create Additional port-channel for each network zone based on your Data Center requirements.
If you are using single Nexus 3524 for IOM failover, create port-channel 5 on fabric interconnect A and portchannel 6 on fabric interconnect B for Internal network zone.
To configure the necessary port channels out of the Cisco UCS environment, complete the following steps:
1. In Cisco UCS Manager, click the LAN tab in the navigation pane.
2. In this procedure, two port channels are created: one from fabric A to both Cisco Nexus switches and one from fabric B to both Cisco Nexus switches.
3. Under LAN > LAN Cloud, expand the Fabric A tree.
4. Right-click Port Channels.
5. Select Create Port Channel.
6. Enter 5 as the unique ID of the port channel.
7. Enter N3k-Uplink as the name of the port channel.
8. Click Next.
9. Select the following ports to be added to the port channel:
- Slot ID 1, Aggregated Port ID 3 and port 1
- Slot ID 1, Aggregated Port ID 3 and port 2
- Slot ID 1, Aggregated Port ID 3 and port 3
- Slot ID 1, Aggregated Port ID 3 and port 4
10. Click >> to add the ports to the port channel.
11. Click Finish to create the port channel.
12. Click OK.
13. In the navigation pane, under LAN > LAN Cloud, expand the fabric B tree.
14. Right-click Port Channels.
15. Select Create Port Channel.
16. Enter 6 as the unique ID of the port channel.
17. Enter N3k-Uplink as the name of the port channel.
18. Click Next.
19. Select the following ports to be added to the port channel:
- Slot ID 1, Aggregated Port ID 3 and port 1
- Slot ID 1, Aggregated Port ID 3 and port 2
- Slot ID 1, Aggregated Port ID 3 and port 3
- Slot ID 1, Aggregated Port ID 3 and port 4
20. Click >> to add the ports to the port channel.
21. Click Finish to create the port channel.
22. Click OK.
For secure multi-tenancy within the Cisco UCS domain, a logical entity created known as organizations.
To create organization unit, complete the following steps:
1. In Cisco UCS Manager, on the Tool bar click New.
2. From the drop-down list select Create Organization.
3. Enter the Name as HANA.
4. (Optional) Enter the Description as Org for HANA.
5. Click OK to create the Organization.
To configure the necessary MAC address pools for the Cisco UCS environment, complete the following steps:
1. In Cisco UCS Manager, click the LAN tab in the navigation pane.
2. Select Pools > root > Sub-Organization > HANA.
3. In this procedure, two MAC address pools are created, one for each switching fabric.
4. Right-click MAC Pools under the root organization.
5. Select Create MAC Pool to create the MAC address pool.
6. Enter FI-A as the name of the MAC pool.
7. (Optional) Enter a description for the MAC pool.
8. Choose Assignment Order Sequential.
9. Click Next.
10. Click Add.
11. Specify a starting MAC address.
12. The recommendation is to place 0A in the next-to-last octet of the starting MAC address to identify all of the MAC addresses as Fabric Interconnect A addresses.
13. Specify a size for the MAC address pool that is sufficient to support the available blade or server resources.
14. Click OK.
15. Click Finish.
16. In the confirmation message, click OK.
17. Right-click MAC Pools under the HANA organization.
18. Select Create MAC Pool to create the MAC address pool.
19. Enter FI-B as the name of the MAC pool.
20. (Optional) Enter a description for the MAC pool.
21. Click Next.
22. Click Add.
23. Specify a starting MAC address.
The recommendation is to place 0B in the next to last octet of the starting MAC address to identify all the MAC addresses in this pool as fabric B addresses.
24. Specify a size for the MAC address pool that is sufficient to support the available blade or server resources.
25. Click OK.
26. Click Finish.
27. In the confirmation message, click OK.
You can also define separate MAC address Pool for each Network Zone. Follow the above steps 1-16 to create MAC address pool for each Network Zone.
To configure the necessary universally unique identifier (UUID) suffix pool for the Cisco UCS environment, complete the following steps:
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Pools > root.
3. Right-click UUID Suffix Pools.
4. Select Create UUID Suffix Pool.
5. Enter UUID_Pool as the name of the UUID suffix pool.
6. (Optional) Enter a description for the UUID suffix pool.
7. Keep the Prefix as the Derived option.
8. Select Sequential for Assignment Order
9. Click Next.
10. Click Add to add a block of UUIDs.
11. Keep the From field at the default setting.
12. Specify a size for the UUID block that is sufficient to support the available blade or server resources.
13. Click OK.
14. Click Finish.
15. Click OK.
To run Cisco UCS with two independent power distribution units, the redundancy must be configured as Grid. Complete the following steps:
1. In Cisco UCS Manager, click the Equipment tab in the navigation pane and select Equipment in the list on the left.
2. In the right pane, click the Policies tab.
3. Under Global Policies, set the Redundancy field in Power Policy to Grid.
4. Click Save Changes.
5. Click OK.
Power Capping feature in Cisco UCS is designed to save power with a legacy data center use cases. This feature does not contribute much to the high performance behavior of SAP HANA. By choosing the option “No Cap” for power control policy, the SAP HANA server nodes will not have a restricted power supply. It is recommended to have this power control policy to make sure sufficient power supply for high performance and critical applications like SAP HANA.
To create a power control policy for the Cisco UCS environment, complete the following steps:
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Policies > root.
3. Right-click Power Control Policies.
4. Select Create Power Control Policy.
5. Enter HANA as the Power Control Policy name.
6. Change the Power Capping setting to No Cap.
7. Click OK to create the power control policy.
8. Click OK.
Firmware management policies allow the administrator to select the corresponding packages for a given server configuration. These policies often include packages for adapter, BIOS, board controller, FC adapters, host bus adapter (HBA) option ROM, and storage controller properties.
To create a firmware management policy for a given server configuration in the Cisco UCS environment, complete the following steps:
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Policies > root.
3. Right-click Host Firmware Packages.
4. Select Create Host Firmware Package.
5. Enter HANA-FW as the name of the host firmware package.
6. Leave Simple selected.
7. Select the version 3.2(2d) for both the Blade and Rack Packages.
8. Click OK to create the host firmware package.
9. Click OK.
A local disk configuration policy configures SAS local drives that have been installed on a server through the onboard RAID controller of the local drive.
To create a local disk configuration policy for HANA servers for Local OS disks, complete the following steps:
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Policies > root.
3. Right-click Local Disk Config Policies.
4. Select Create Local Disk Configuration Policy.
5. Enter RAID1 as the local disk configuration policy name.
6. Change the mode to RAID 1.
7. Click OK to create the local disk configuration policy.
8. Click OK.
To create a local disk configuration policy for MapR servers, complete the following steps:
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Policies > root.
3. Right-click Local Disk Config Policies.
4. Select Create Local Disk Configuration Policy.
5. Enter MapR as the local disk configuration policy name.
6. Change the mode to Any Configuration.
7. Click OK to create the local disk configuration policy.
To get best performance for HANA it is required to configure the Server BIOS accurately. To create a server BIOS policy for the Cisco UCS environment, complete the following steps:
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Policies > root > Sub-Organization > HANA.
3. Right-click BIOS Policies.
4. Select Create BIOS Policy.
5. Enter HANA-BIOS as the BIOS policy name.
6. Click OK.
7. Under BIOS Policies, Click the newly created HANA-BIOS Policy.
8. In the Main pane, under BIOS Setting choose Disabled for Quiet Boot.
9. Click on Advance tab.
10. The recommendation from SAP for SAP HANA is to disable all Processor C States. This will force the CPU to stay on maximum frequency and allow SAP HANA to run with best performance.
11. Under Processor choose Disabled for all C-States.
12. Set HPC for CPU Performance, Performance for Power Technology, Energy Performance.
13. Click RAS Memory.
14. Select Maximum-Performance for Memory RAS Configuration and Enabled for NUMA optimized.
15. Click Serial Port.
16. Select Enabled for Serial Port A enable.
17. Click Server Management.
18. Choose 115.2k for BAUD Rate, Enabled for Legacy OS redirection, VT100-PLUS for Terminal type. This is used for Serial Console Access over LAN to all SAP HANA servers.
19. Click Save Changes.
To create a server BIOS policy for MapR server, complete the following steps:
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Policies > root > Sub-Organization > HANA.
3. Right-click BIOS Policies.
4. Select Create BIOS Policy.
5. Enter MapR-BIOS as the BIOS policy name.
6. Change the Quiet Boot setting to Disabled.
7. Click Next.
8. Enable Turbo Boost, Enhanced Intel Speedstep, Hyper Threading.
It is recommended to disable all Processor C States. This will force the CPU to stay on maximum frequency and allow MapR Server to run with best performance.
9. Set CPU Performance for high throughput, Power Technology, Energy Performance, DRAM Clock Throttling for maximum performance.
10. Click Next.
11. Keep default values at the Intel Direct IO.
12. Click Next.
13. In the RAS Memory, select maximum-performance and enable NUMA.
14. Click Next.
15. Keep the default values for the rest of the options.
16. Click Next.
17. Click Finish to create the BIOS policy.
The Serial over LAN policy is required to get console access to all the SAP HANA servers through SSH from the management network. This is used in case of the server hang or a Linux kernel crash, where the dump is required. Configure the speed in the Server Management Tab of the BIOS Policy. To create the serial over LAN policy, complete the following steps:
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Policies > root > Sub-Organization > HANA.
3. Right-click the Serial over LAN Policies.
4. Select Create Serial over LAN Policy.
5. Enter SoL-Console as the Policy name.
6. Select Serial over LAN State to enable.
7. Change the Speed to 115200.
8. Click OK.
It is recommended to update the default Maintenance Policy with the Reboot Policy “User Ack” for the SAP HANA server. This policy will wait for the administrator to acknowledge the server reboot for the configuration changes to take effect.
To update the default Maintenance Policy, complete the following steps:
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Policies > root.
3. Select Maintenance Policies > default.
4. Change the Reboot Policy to User Ack.
5. Click Save Changes.
6. Click OK to accept the change.
The Serial over LAN access requires an IPMI access control to the board controller. This is also used for the STONITH function of the SAP HANA mount API to kill a hanging server. The default user is “sapadm” with the password “cisco”.
To create an IPMI Access Profile, complete the following steps:
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Policies > root > Sub-Organization > HANA.
3. Right-click IPMI Access Profiles.
4. Select Create IPMI Access Profile.
5. Enter HANA-IPMI as the Profile name.
6. Click + button.
7. Enter Username in the Name field and password.
8. Select Admin as Role.
9. Click OK to create user.
10. Click OK to Create IPMI Access Profile.
11. Click OK.
The core network requirements for SAP HANA are covered by Cisco UCS defaults. Cisco UCS is based on 40GbE and provides redundancy through the Dual Fabric concept. The Service Profile is configured to distribute the traffic across Fabric Interconnect A and B. During normal operation, the traffic in the inter-node flows on FI A and the Storage traffic is on FI B. The inter-node traffic flows from a Blade Server to the Fabric Interconnect A and back to other Blade Server. The storage traffic flows from a HANA Server to the Fabric Interconnect B and back to MapR Server.
To configure jumbo frames and enable quality of service in the Cisco UCS fabric, complete the following steps:
1. In Cisco UCS Manager, click the LAN tab in the navigation pane.
2. Select LAN > LAN Cloud > QoS System Class.
3. In the right pane, click the General tab.
4. On the MTU Column, enter 9216 in the box.
5. Check Enabled Under Priority for Silver.
6. Click Save Changes in the bottom of the window.
7. Click OK.
QoS policies assign a system class to the network traffic for a vNIC. To create QoS Policy for the Cisco UCS environment, complete the following steps:
1. In Cisco UCS Manager, click the LAN tab in the navigation pane.
2. Select Policies > root > Sub-Organization > HANA.
3. Right-click QoS Policies.
4. Select Create QoS Policy.
5. Enter Storage-Silver as the QoS Policy name.
6. For Priority Select Silver from the drop-down list.
7. Enter 20000000 for Rate(Kbps).
8. Click OK to create QoS Policy.
CDP needs to be enabled to learn the MAC address of the End Point. To update default Network Control Policy, complete the following steps:
1. In Cisco UCS Manager, click the LAN tab in the navigation pane.
2. Select LAN > Policies > root > Network Control Policies > default.
3. In the right pane, click the General tab.
4. For CDP: select Enabled radio button.
5. Click Save Changes in the bottom of the window.
6. Click OK.
In order to keep the vNIC links up in case of Nexus 3524 failure, create the Network Control Policy for Internal Network. To create Network Control Policy for Internal Network, complete the following steps:
1. In Cisco UCS Manager, click the LAN tab in the navigation pane.
2. Select LAN > Policies > root > Network Control Policies > right-click and Select Create Network Control Policy.
3. Enter Internal as the Name of the Policy.
4. For CDP: select Enabled radio button.
5. For Action on Uplink Fail, select Warning radio button.
6. Click OK.
Within Cisco UCS, all the network types for an SAP HANA system are reflected by defined VLANs. Network design from SAP has seven SAP HANA related networks and two infrastructure related networks. The VLAN IDs can be changed if required to match the VLAN IDs in the data center network; for example, ID 221 for backup should match the configured VLAN ID at the data center network switches. Even though nine VLANs are defined, VLANs for all the networks are not necessary if the solution will not use the network. For example, if the Replication Network is not used in the solution, then VLAN ID 225 need not be created.
To configure the necessary VLANs for the Cisco UCS environment, complete the following steps:
1. In Cisco UCS Manager, click the LAN tab in the navigation pane.
In this procedure, nine VLANs are created.
2. Select LAN > LAN Cloud.
3. Right-click VLANs.
4. Select Create VLANs.
5. Enter HANA-Internal as the name of the VLAN to be used for HANA Node to Node network.
6. Keep the Common/Global option selected for the scope of the VLAN.
7. Enter <<var_internal_vlan_id>> as the ID of the HANA Node to Node network.
8. Keep the Sharing Type as None.
9. Click OK.
10. Repeat the Steps 1-9 for each VLAN.
11. Create VLAN for HANA-AppServer.
12. Create VLAN for HANA-Backup.
13. Create VLAN for HANA-Client.
14. Create VLAN for HANA-DataSource.
15. Create VLAN for HANA-Replication.
16. Create VLAN for HANA-Storage.
17. Create VLAN for NFS-IPMI for Inband Access.
18. Create VLAN for Management.
19. Create VLAN for MapR-01.
20. Create VLAN for MapR-02.
21. Create VLAN for MapR-03.
For easier management and bandwidth allocation to a dedicated uplink on the Fabric Interconnect, VLAN Groups are created within the Cisco UCS. SAP HANA uses the following VLAN groups:
· Admin Zone
· Client Zone
· Internal Zone
· Backup Network
· Replication Network
To configure the necessary VLAN Groups for the Cisco UCS environment, complete the following steps:
1. In Cisco UCS Manager, click the LAN tab in the navigation pane.
In this procedure, five VLAN Groups are created. Based on the solution requirement create VLAN groups, it not required to create all five VLAN groups.
2. Select LAN > LAN Cloud.
3. Right-click VLAN Groups.
4. Select Create VLAN Groups.
5. Enter Admin-Zone as the name of the VLAN Group used for Infrastructure network.
6. Select Management.
7. Click Next.
8. Click Next on Add Uplink Ports.
9. Choose the port channels created for the Admin Network.
10. Click Finish.
11. Follow the steps 1-10 for each VLAN Group.
12. Create VLAN Groups for Internal Zone.
13. Click Next.
14. Click Next on Add Uplink Ports.
15. Choose port channels created for Internal Zone.
16. Click Finish.
17. Create VLAN Groups for Client Zone.
18. Click Next.
19. Click Next on Add Uplink Ports.
20. Click Finish.
21. Create VLAN Groups for Backup Network.
22. Click Next.
23. Click Next on Add Uplink Ports.
24. Choose port-channels created for Backup Network.
25. Click Finish.
26. Create VLAN Groups for Replication Network.
27. Click Next.
28. Click Next on Add Uplink Ports.
29. Choose port-channels created for Replication Network.
30. Click Finish.
Each VLAN is mapped to a vNIC template to specify the characteristic of a specific network. The vNIC template configuration settings include MTU size, Failover capabilities and MAC-Address pools.
To create vNIC templates for the Cisco UCS environment, complete the following steps:
1. In Cisco UCS Manager, click the LAN tab in the navigation pane.
2. Select Policies > root > Sub-Organization > HANA.
3. Right-click vNIC Templates.
4. Select Create vNIC Template.
5. Enter HANA-Internal as the vNIC template name.
6. Keep Fabric A selected.
7. Check the Enable Failover checkbox.
8. Under Target, make sure that the VM checkbox is unchecked.
9. Select Updating Template as the Template Type.
10. Under VLANs, check the checkboxes for HANA-Internal.
11. Set HANA-Internal as the native VLAN.
12. For MTU, enter 9000.
13. In the MAC Pool list, select FI-A.
14. For Network Control Policy Select Internal from drop-down list.
15. Click OK to create the vNIC template.
For most SAP HANA use cases the network traffic is well distributed across the two Fabrics (Fabric A and Fabric B) using the default setup. In special cases, it can be required to rebalance this distribution for better overall performance. This can be done in the vNIC template with the Fabric ID setting. Note that the MTU settings must match the configuration in customer data center. MTU setting of 9000 is recommended for best performance.
16. Follow the steps1-15 above to create vNIC template for each Network Interface.
1. Select QoS Policy Storage-Silver from the drop-down list.
2. Select Policies > root > Sub-Organization > HANA.
3. Right-click vNIC Templates.
4. Select Create vNIC Template.
5. Enter HANA-Storage as the vNIC template name.
6. Keep Fabric B selected.
7. Check the Enable Failover checkbox.
8. Under Target, make sure that the VM checkbox is unchecked.
9. Select Updating Template as the Template Type.
10. Under VLANs, check the checkboxes for HANA-Storage.
11. Set HANA-Storage as the native VLAN.
12. For MTU, enter 9000.
13. In the MAC Pool list, select FI-B.
14. For Network Control Policy Select Internal from drop-down list.
15. Click OK to create the vNIC template.
1. In Cisco UCS Manager, click the LAN tab in the navigation pane.
2. Select Policies > root > Sub-Organization > HANA.
3. Right-click vNIC Templates.
4. Select Create vNIC Template.
5. Enter HANA-AppServer as the vNIC template name.
6. Keep Fabric A selected.
7. Check the Enable Failover checkbox.
8. Under Target, make sure that the VM checkbox is unchecked.
9. Select Updating Template as the Template Type.
10. Under VLANs, check the checkboxes for HANA-AppServer.
11. Set HANA-AppServer as the native VLAN.
12. For MTU, enter 9000.
13. In the MAC Pool list, select FI-A.
14. For Network Control Policy Select default from drop-down list.
15. Click OK to create the vNIC template.
1. In Cisco UCS Manager, click the LAN tab in the navigation pane.
2. Select Policies > root > Sub-Organization > HANA.
3. Right-click vNIC Templates
4. Select Create vNIC Template.
5. Enter HANA-Backup as the vNIC template name.
6. Keep Fabric B selected.
7. Check the Enable Failover checkbox.
8. Under Target, make sure that the VM checkbox is unchecked.
9. Select Updating Template as the Template Type.
10. Under VLANs, check the checkboxes for HANA-Backup.
11. Set HANA-Backup as the native VLAN.
12. For MTU, enter 9000.
13. In the MAC Pool list, select FI-B.
14. For Network Control Policy Select default from drop-down list.
15. Click OK to create the vNIC template.
1. In Cisco UCS Manager, click the LAN tab in the navigation pane.
2. Select Policies > root > Sub-Organization > HANA.
3. Right-click vNIC Templates.
4. Select Create vNIC Template.
5. Enter HANA-Client as the vNIC template name.
6. Keep Fabric B selected.
7. Check the Enable Failover checkbox.
8. Under Target, make sure that the VM checkbox is unchecked.
9. Select Updating Template as the Template Type.
10. Under VLANs, check the checkboxes for HANA-Client.
11. Set HANA-Client as the native VLAN.
12. For MTU, enter 9000.
13. In the MAC Pool list, select FI-B.
14. For Network Control Policy Select default from drop-down list.
15. Click OK to create the vNIC template.
1. In Cisco UCS Manager, click the LAN tab in the navigation pane.
2. Select Policies > root > Sub-Organization > HANA.
3. Right-click vNIC Templates.
4. Select Create vNIC Template.
5. Enter HANA-DataSource as the vNIC template name.
6. Keep Fabric A selected.
7. Check the Enable Failover checkbox.
8. Under Target, make sure that the VM checkbox is unchecked.
9. Select Updating Template as the Template Type.
10. Under VLANs, check the checkboxes for HANA-DataSource.
11. Set HANA-DataSource as the native VLAN.
12. For MTU, enter 9000.
13. In the MAC Pool list, select FI-A.
14. For Network Control Policy Select default from drop-down list.
15. Click OK to create the vNIC template.
1. In Cisco UCS Manager, click the LAN tab in the navigation pane.
2. Select Policies > root > Sub-Organization > HANA.
3. Right-click vNIC Templates.
4. Select Create vNIC Template.
5. Enter HANA-Replication as the vNIC template name.
6. Keep Fabric B selected.
7. Check the Enable Failover checkbox.
8. Under Target, make sure that the VM checkbox is unchecked.
9. Select Updating Template as the Template Type.
10. Under VLANs, check the checkboxes for HANA-Replication.
11. Set HANA-Replication as the native VLAN.
12. For MTU, enter 9000.
13. In the MAC Pool list, select FI-B.
14. For Network Control Policy Select default from drop-down list.
15. Click OK to create the vNIC template.
1. In Cisco UCS Manager, click the LAN tab in the navigation pane.
2. Select Policies > root > Sub-Organization > HANA.
3. Right-click vNIC Templates.
4. Select Create vNIC Template.
5. Enter Mgmt as the vNIC template name.
6. Keep Fabric A selected.
7. Check the Enable Failover checkbox.
8. Under Target, make sure that the VM checkbox is unchecked.
9. Select Updating Template as the Template Type.
10. Under VLANs, check the checkboxes for Management.
11. Set Management as the native VLAN.
12. For MTU, enter 9000.
13. In the MAC Pool list, select FI-A.
14. For Network Control Policy Select default from drop-down list.
15. Click OK to create the vNIC template.
1. In Cisco UCS Manager, click the LAN tab in the navigation pane.
2. Select Policies > root > Sub-Organization > HANA.
3. Right-click vNIC Templates.
4. Select Create vNIC Template.
5. Enter NFS-IPMI as the vNIC template name.
6. Keep Fabric A selected.
7. Check the Enable Failover checkbox.
8. Under Target, make sure that the VM checkbox is unchecked.
9. Select Updating Template as the Template Type.
10. Under VLANs, check the checkboxes for NFS-IPMI.
11. Set NFS-IPMI as the native VLAN.
12. For MTU, enter 9000.
13. In the MAC Pool list, select FI-A.
14. For Network Control Policy Select Internal from drop-down list.
15. Click OK to create the vNIC template.
1. In Cisco UCS Manager, click the LAN tab in the navigation pane.
2. Select Policies > root > Sub-Organization > HANA.
3. Right-click vNIC Templates
4. Select Create vNIC Template.
5. Enter MapR-01 as the vNIC template name.
6. Keep Fabric A selected.
7. Check the Enable Failover checkbox.
8. Under Target, make sure that the VM checkbox is unchecked.
9. Select Updating Template as the Template Type.
10. Under VLANs, check the checkboxes for MapR-01.
11. Set MapR-01 as the native VLAN.
12. For MTU, enter 9000.
13. In the MAC Pool list, select FI-A.
14. For Network Control Policy Select Internal from drop-down list.
15. Click OK to create the vNIC template.
1. In Cisco UCS Manager, click the LAN tab in the navigation pane.
2. Select Policies > root > Sub-Organization > HANA.
3. Right-click vNIC Templates.
4. Select Create vNIC Template.
5. Enter MapR-02 as the vNIC template name.
6. Keep Fabric B selected.
7. Check the Enable Failover checkbox.
8. Under Target, make sure that the VM checkbox is unchecked.
9. Select Updating Template as the Template Type.
10. Under VLANs, check the checkboxes for MapR-02.
11. Set MapR-02 as the native VLAN.
12. For MTU, enter 9000.
13. In the MAC Pool list, select FI-B.
14. For Network Control Policy Select Internal from drop-down list.
15. Click OK to create the vNIC template.
1. In Cisco UCS Manager, click the LAN tab in the navigation pane.
2. Select Policies > root > Sub-Organization > HANA.
3. Right-click vNIC Templates.
4. Select Create vNIC Template.
5. Enter MapR-03 as the vNIC template name.
6. Keep Fabric A selected.
7. Check the Enable Failover checkbox.
8. Under Target, make sure that the VM checkbox is unchecked.
9. Select Updating Template as the Template Type.
10. Under VLANs, check the checkboxes for MapR-03.
11. Set MapR-03 as the native VLAN.
12. For MTU, enter 9000.
13. In the MAC Pool list, select FI-A.
14. For Network Control Policy Select Internal from drop-down list.
15. Click OK to create the vNIC template.
To create Local boot policies, complete the following steps:
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Policies > root > Sub-Organization > HANA.
3. Right-click Boot Policies.
4. Select Create Boot Policy.
5. Enter Local-Boot as the name of the boot policy.
6. (Optional) Enter a description for the boot policy.
7. Expand the Local Devices drop-down list and select Add CD/DVD.
8. Expand the Local Devices drop-down list and select Add Local Disk.
9. Click OK to save the boot policy.
For SAP HANA High Availability configuration, we would be using IPMI tool, create a block of IP addresses for SAP HANA servers. In the Cisco UCS environment, complete the following steps:
This block of IP addresses should be in the same subnet as the HANA-Admin IP addresses of the HANA Servers.
1. In Cisco UCS Manager, click the LAN tab in the navigation pane.
2. Select Pools > root > > Sub-Organization > HANA > IP Pools, right-click Create a IP Pool.
3. For the Name enter HANA_IPMI.
4. For Assignment order select Sequential and click Next.
5. Click Add.
6. Enter the starting IP address of the block and the number of IP addresses required, and the subnet and gateway information.
7. Click OK to create the IP block.
8. Click OK in the confirmation message.
To allocate the previously configured IPv4 Address Pool, VLAN, and VLAN Group to the global Inband Profile, complete the following steps:
1. In Cisco UCS Manager, click the LAN tab in the navigation pane.
2. Select LAN Cloud, on the right pane click the Global Policies tab.
3. On the Global Policies page, under the Inband Profile section:
4. For Inband VLAN Group select Internal-Zone from the drop-down list.
5. For Network select NFS-IPMI from the drop-down list.
6. For IP Pool Name select HANA-IPMI from the drop-down list.
7. Click Save Changes.
The LAN configurations and relevant SAP HANA policies must be defined prior to creating a Service Profile Template.
To create the service profile template, complete the following steps:
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Service Profile Templates > root > Sub-Organization > HANA.
3. Right-click HANA.
4. Select Create Service Profile Template to open the Create Service Profile Template wizard.
5. Identify the service profile template:
a. Enter HANA-Server as the name of the service profile template.
b. Select the Updating Template option.
c. Under UUID, select HANA-UUID as the UUID pool.
d. Click Next.
6. Configure the Storage Provisioning:
a. Click the Local Disk Configuration Policy tab.
b. Select RAID1 for Local Storage field from the drop-down list.
c. Click Next.
7. Configure the Networking options:
a. Keep the default settings for Dynamic vNIC Connection Policy.
b. Select the Expert option for How would you like to configure LAN connectivity.
c. Click the Add button to add a vNIC to the template.
d. In the Create vNIC dialog box, enter HANA-Internal as the name of the vNIC.
e. Check the Use vNIC Template checkbox.
f. In the vNIC Template list, select HANA-Internal.
g. In the Adapter Policy list, select Linux.
h. Click OK to add this vNIC to the template.
i. Repeat the above steps c-h for each vNIC.
8. Add vNIC for HANA-Storage.
9. Add vNIC for HANA-Client.
10. Add vNIC for HANA-AppServer.
11. Add vNIC for HANA-DataSource.
12. Add vNIC for HANA-Replication.
13. Add vNIC for HANA-Backup.
14. Add vNIC for Mgmt.
15. Add vNIC for NFS-IPMI.
16. Review the table in the Networking page to make sure that all vNICs were created.
17. Click Next.
18. Configure the SAN Connectivity, Select the No vHBAs option for the “How would you like to configure SAN connectivity?” field.
19. Click Next.
20. Set no Zoning options and click Next.
21. Set the vNIC/vHBA placement options:
a. In the Select Placement list, select the Specify Manually.
b. Select vCon1 and assign the vNICs to the virtual network interfaces policy in the following order:
i. HANA-Storage
ii. HANA-AppServer
iii. HANA-Backup
iv. Mgmt
c. Select vCon2 and assign the vNICs to the virtual network interfaces policy in the following order:
i. HANA-Internal
ii. HANA-Client
iii. HANA-DataSource
iv. HANA-Replication
v. NFS-IPMI
d. Review the table to verify that all vNICs are assigned to the policy in the appropriate order.
e. Click Next.
22. No Change required on the vMedia Policy, click Next.
23. Set the server boot order:
a. Select Local-Boot for Boot Policy.
b. Click Next.
24. For Maintenance policy:
a. Select the default Maintenance Policy.
b. Click Next.
25. Specify the server assignment:
a. Select Up as the power state to be applied when the profile is associated with the server.
b. Expand Firmware Management at the bottom of the page and select HANA-FW from the Host Firmware list.
c. Click Next.
26. For Operational Policies:
a. In the BIOS Policy list, select HANA-BIOS.
b. Expand the External IPMI Management Configuration and select HANA-IPMI in the IPMI Access Profile.
c. Select SoL-Console in the SoL Configuration Profile.
d. Expand Management IP Address,
e. In the Outband IPv4 tab choose ext-mgmt in the Management IP Address Policy.
f. In the Inband IPv4 tab choose NFS-IPMI for Network and choose HANA-IPMI for Management IP Address Policy.
g. Expand Power Control Policy Configuration and select No-Power-Cap in the Power Control Policy list.
27. Click Finish to create the service profile template.
28. Click OK in the confirmation message.
If you are using B460-M4 as HANA Servers, create the service profile template and modify the vNIC placement policy.
To create a clone service profile template created, complete the following steps:
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Service Profile Templates > root > Sub-Organization > HANA.
3. Right-click Service Template HANA-Server.
4. Select Create a Clone.
5. Enter HANA-C480 for Clone name and Choose root as Org.
6. Click OK to Create the clone of HANA-C480.
7. Click OK to confirm.
8. On the cloned Service Profile Template Select HANA-C480 in the Navigation pane and click Network tab.
9. Under Actions Click Modify vNIC/vHBA Placement.
10. Select vCon1 and assign the vNICs to the virtual network interfaces policy in the following order:
a. HANA-Internal
b. HANA-Storage
c. HANA-AppServer
d. HANA-Client
e. Mgmt
f. HANA-Backup
g. HANA-DataSource
h. HANA-Replication
i. NFS-IPMI
11. Click OK to complete the vNIC/vHBA Placement policy.
12. Click OK to confirm.
If you are using C460-M4 as HANA Servers, create the service profile template and modify the vNIC placement policy.
To create a clone service profile template created, complete the following steps:
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Service Profile Templates > root > Sub-Organization > HANA.
3. Right-click Service Template HANA-A.
4. Select Create a Clone.
5. Enter HANA-C460 for Clone name and Choose root as Org.
6. Click OK to Create the clone of HANA-C460.
7. Click OK to confirm.
8. On the cloned Service Profile Template Select HANA-C460 in the Navigation pane and click Network tab.
9. Under Actions Click Modify vNIC/vHBA Placement.
10. Select vCon1 and assign the vNICs to the virtual network interfaces policy in the following order:
a. HANA-Internal
b. HANA-Storage
c. HANA-AppServer
d. HANA-Client
e. Mgmt
f. HANA-Backup
g. HANA-DataSource
h. HANA-Replication
i. NFS-IPMI
11. Click OK to complete the vNIC/vHBA Placement policy.
12. Click OK to confirm.
To create service profiles from the service profile template, complete the following steps:
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Service Profile Templates > root > Sub-Organization > HANA > Service Template HANA-Server.
3. Right-click Service Template HANA-Server and select Create Service Profiles from Template
4. Enter HANA-Server0 as the service profile prefix.
5. Enter 1 as Name Suffix Starting Number.
6. Enter 1 as the Number of Instances.
7. Click OK to create the service profile.
To create service profiles from the service profile template HANA-B, complete the following steps:
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Service Profile Templates > root > Sub-Organization > HANA > Service Template HANA-B.
3. Right-click Service Template HANA-B and select Create Service Profiles from Template.
4. Enter HANA-Server0 as the service profile prefix.
5. Enter 2 as Name Suffix Starting Number.
6. Enter 1 as the Number of Instances.
7. Click OK to create the service profile.
To create service profiles from the service profile template HANA-C460, complete the following steps:
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Service Profile Templates > root > Sub-Organization > HANA > Service Template HANA-C460.
3. Right-click Service Template HANA-C480 and select Create Service Profiles from Template.
4. Enter HANA-Server0 as the service profile prefix.
5. Enter 2 as Name Suffix Starting Number.
6. Enter 1 as the Number of Instances.
7. Click OK to create the service profile.
To associate service profile created for a specific slot, complete the following steps:
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Service Profile > root > Sub-Organization > HANA > HANA-Server01.
3. Right-click HANA-Server01 and select Change Service Profile Association.
4. For Server Assignment Choose Select existing Server for the drop-down list.
5. Click All Servers.
6. Select the Server as recommended.
7. Repeat for above steps 1-6 for each HANA Servers.
To create the service profile template, complete the following steps:
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Service Profile Templates > root > Sub-Organization > HANA.
3. Right-click HANA.
4. Select Create Service Profile Template to open the Create Service Profile Template wizard.
5. Identify the service profile template:
a. Enter MapR as the name of the service profile template.
b. Select the Updating Template option.
c. Under UUID, select HANA-UUID as the UUID pool.
d. Click Next.
6. Configure the Storage Provisioning:
a. Click the Local Disk Configuration Policy tab.
b. Select MapR for Local Storage field from the drop-down list.
c. Click Next.
7. Configure the Networking options:
a. Keep the default settings for Dynamic vNIC Connection Policy.
b. Select the Expert option for How would you like to configure LAN connectivity.
c. Click the Add button to add a vNIC to the template.
d. In the Create vNIC dialog box, enter MapR-01 as the name of the vNIC.
e. Check the Use vNIC Template checkbox.
f. In the vNIC Template list, select MapR-01
g. In the Adapter Policy list, select Linux.
h. Click OK to add this vNIC to the template.
i. Repeat the above steps c-h for each vNIC.
8. Add vNIC for MapR-02.
9. Add vNIC for MapR-03.
10. Add vNIC for HANA-Storage.
11. Add vNIC for Mgmt.
12. Review the table in the Networking page to make sure that all vNICs were created.
13. Click Next.
14. Configure the SAN Connectivity, Select the No vHBAs option for the “How would you like to configure SAN connectivity?” field.
15. Click Next.
16. Set no Zoning options and click Next.
17. Set the vNIC/vHBA placement options, keep the default values.
18. Click Next.
19. Keep the default values on the vMedia Policy, click Next.
20. To set the server boot order, select Local-Boot for Boot Policy.
21. Click Next.
22. For Maintenance policy: Select the default Maintenance Policy.
23. Click Next.
24. Specify the server assignment: Select Up as the power state to be applied when the profile is associated with the server.
25. Expand Firmware Management at the bottom of the page and select HANA-FW from the Host Firmware list.
26. Click Next.
27. For Operational policies, In the BIOS Policy list, select MapR-BIOS.
28. Expand Management IP Address, in the Outband IPv4 tab choose ext-mgmt in the Management IP Address Policy.
29. In the Outbound IPv4 tab, choose ext-mgmt in the Management IP Address Policy.
30. In the Outbound IPv4 tab, choose NFS-IPMI for Network and choose HANA-IPMI for Management IP Address Policy.
31. Expand Power Control Policy Configuration and select No-Power-Cap in the Power Control Policy list.
32. Click Finish to create the service profile template.
33. Click OK in the confirmation message.
To create service profiles from the service profile template, complete the following steps:
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Service Profile Templates > root > Sub-Organization > HANA > Service Template MapR.
3. Right-click MapR and select Create Service Profiles from Template.
4. Enter appropriate name for the service profile prefix. For example, MapR-0.
5. Enter 1 as Name Suffix Starting Number.
6. Enter appropriate number of service profile to be created in the Number of Instances. For example, 4.
7. Click OK to create the service profile.
To associate service profile created for a specific slot, complete the following steps:
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Service Profile > root > Sub-Organization > HANA > MapR-01.
3. Right-click MapR-01 and select Change Service Profile Association.
4. For Server Assignment, choose Select existing Server from the drop-down list.
5. Click All Servers.
6. Select the C240-M5 Rack Mount Server Rack ID 1.
7. Repeat the above steps 1-6 for each MapR Server.
The section describes the procedure for Installing MapR Data Platform on Cisco UCS C240-M5 Servers. Each server has 2 x Internal SSD Boot drives where the Operating System is installed with RAID 1. For MapR Storage 24 x 1.8 TB 10k rpm 4K disks are configured in 4 x RAID 5 group with 6 disks.
To configure RAID on the Cisco UCS C240 Servers used for MapR Data Platform, complete the following steps:
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Service Profile > root > Sub-Organization > HANA > MapR-01.
3. Click KVM Console.
4. When the KVM Console is launched, click Boot Server.
5. When prompted Press <Ctrl><R> to Run MegaRAID Configuration Utility.
6. In the VD Mgmt press F2 for Operations and Select Create Virtual Drive.
7. Choose the following option to create RAID 5 virtual drive with 6 disks:
a. For RAID Level Choose RAID-5.
b. Choose the first 6 disks.
c. Give a Name for virtual drive 1.
8. Choose Advanced option:
a. For Strip Size choose 1MB.
b. For Read Policy Choose Ahead.
c. For Write Policy choose Write Back with BBU.
d. For I/O Policy choose Direct.
e. Keep the Default Disk cache Policy and Emulation.
f. Choose Initialize.
g. Click OK.
9. To create the second RAID 5 Virtual Drive press F2 for Operations and Select Create Virtual Drive.
10. Create a RAID 5 virtual drive with 6 disks:
a. For RAID Level Choose RAID-5.
b. Choose the next 6 disks.
c. Give a Name for virtual drive 2.
d. Choose Advanced option.
e. For Strip Size choose 1MB.
f. For Read Policy Choose Ahead.
g. For Write Policy choose Write Back with BBU.
h. For I/O Policy choose Direct.
i. Keep the Default Disk cache Policy and Emulation.
j. Choose Initialize.
k. Click OK.
11. To create the third RAID 5 Virtual Drive press F2 for Operations and select Create Virtual Drive.
12. Create a RAID 5 virtual drive with 6 disks:
a. For RAID Level Choose RAID-5.
b. Choose the next 6 disks.
c. Give a Name for virtual drive 3.
d. Choose Advanced option.
e. For Strip Size choose 1MB.
f. For Read Policy Choose Ahead.
g. For Write Policy choose Write Back with BBU.
h. For I/O Policy choose Direct.
i. Keep the Default Disk cache Policy and Emulation.
j. Choose Initialize.
k. Click OK.
13. To create the fourth RAID 5 Virtual Drive press F2 for Operations and select Create Virtual Drive.
14. Create a RAID 5 virtual drive with 6 disks:
a. For RAID Level Choose RAID-5.
b. Choose the last 6 disks.
c. Give a Name for virtual drive 4.
d. Choose Advanced option.
e. For Strip Size choose 1MB.
f. For Read Policy Choose Ahead.
g. For Write Policy choose Write Back with BBU.
h. For I/O Policy choose Direct.
i. Keep the Default Disk cache Policy and Emulation.
j. Choose Initialize.
k. Click OK.
15. Press esc to exist MegaRAID Configuration Utility.
16. Press Ctrl+Alt+Del to reboot the server.
17. Repeat the steps 1-16 for each MapR Servers.
The following procedure show the SUSE Linux Enterprise Server 12 SP 1 Operation System installation on Internal Boot Drives.
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Service Profile > root > Sub-Organization > HANA > MapR-01.
3. Click KVM Console.
4. When the KVM Console is launched, click Boot Server.
5. Click Virtual Media > Activate Virtual Devices:
a. Select the option Accept this Session for Unencrypted Virtual Media Session and then click Apply.
b. Click Virtual Media and Choose Map CD/DVD.
c. Click Browse to navigate ISO media location.
d. Click Map Device.
6. On the Initial screen select Installation to begin the installation process.
7. Choose Language, Keyboard layout, Select I Agree the License Terms and click Next.
8. On the Network Settings screen Under Overview click VNIC Ethernet NIC.
9. To configure the network interface on the OS, it is required to identify the mapping of the Ethernet device on the OS to vNIC interface on the Cisco UCS:
a. In Cisco UCS Manager, click the Servers tab in the navigation pane.
b. Expand Servers > Service Profile > root > Sub-Organization > HANA > MapR-01.
c. Click + to Expand.
d. Click vNICs.
e. On the right pane list of the vNICs with MAC Address are listed.
f. Note that the MAC Address of the Mgmt vNIC “00:25:B5:2A:00:22”.
g. With the vNIC order in the Service Profile, vNIC for Management should be eth4.
10. By comparing MAC Address on the OS and Cisco UCS, eth4 on OS will carry the VLAN for Management.
11. Click Edit, under the Address tab
12. Click Statically Assigned IP Address:
a. In the IP Address field enter <<Management IP address>>.
b. In the Subnet Mask field enter <<subnet mask for Management Interface>>.
c. In the Hostname field enter the hostname for Management Interface.
13. Repeat Steps 10-12 for each VNIC device.
14. On the Network Settings screen Select Hostname/DNS:
a. In the Hostname field enter the Hostname.
b. In the Domain Name Field enter the Domain Name.
c. In the Name Server 1 field enter <<DNS server1>> and Name Server 2 field enter <<DNS server2>>.
d. In the Search domains field enter <<domain1.com,domain2.com>>.
e. Click Routing.
f. For the Default IPv4 Gateway enter the <<Default Gateway>>.
g. Click Next.
15. In the Registration screen and click Skip Registration. You will update the system in the following section.
16. In the Add On Product screen and click Next.
17. In the System Role leave Default System and click Next.
18. In the Suggested Partitioning screen click Expert Partitioner.
19. In the Expert Partitioner screen, select the first Internal Drive for Operating System and click Add.
20. In the Add Partition on /dev/sdd, select Primary Partition and click Next.
21. For the New Partition Size, select Custom Size.
22. In the Size field enter 200 MB and click Next.
23. For the Role, select Raw Volume (unformatted) and click Next.
24. For Formatting Options, select Do not format partition.
25. In the Add Partition on /dev/sdd, select Primary Partition and click Next.
26. For the New Partition Size, select Custom Size.
27. In the Size field enter 2 GB and click Next.
28. For the Role, select Raw Volume (unformatted) and click Next.
29. For Formatting Options, select Do not format partition.
30. In the Add Partition on /dev/sdd, select Primary Partition and click Next.
31. For the New Partition Size, select Custom Size.
32. In the Size field enter 100 GB and click Next.
33. For the Role, select Raw Volume (unformatted) and click Next.
34. For Formatting Options, select Do not format partition.
35. In the Expert Partitioner screen, select the Second Internal Drive for Operating System click Add.
36. In the Add Partition on /dev/sdf, select Primary Partition and click Next.
37. For the New Partition Size, select Custom Size
38. In the Size field enter 200 MB and click Next.
39. For the Role select Raw Volume (unformatted) and click Next.
40. For Formatting Options select Do not format partition
41. In the Add Partition on /dev/sdf, select Primary Partition and click Next.
42. For the New Partition Size, select Custom Size.
43. In the Size field enter 2 GB and click Next.
44. For the Role select Raw Volume (unformatted) and click Next.
45. For Formatting Options select Do not format partition.
46. In the Add Partition on /dev/sdf, select Primary Partition and click Next.
47. For the New Partition Size, select Custom Size.
48. In the Size field enter 100 GB and click Next.
49. For the Role select Raw Volume (unformatted) and click Next.
50. For Formatting Options select Do not format partition.
51. There should be six RAID partitions, three on each drive. Continue to create the RAID devices.
52. In the Expert Partitioner Window, select RAID and click Add RAID.
53. In the Add RAID /dev/md0:
a. For RAID Type Select RAID 1 (Mirroring).
b. Select the 200 MB partition on the two devices (/dev/sde1 and /dev/sdf1) and click Add.
c. Click Next.
d. In the RAID Options, for Chuck Size, select 512 KiB.
e. Click Next.
f. For Role, select Operating System.
g. Click Next.
h. Select Format Partition, For File System type select ext3.
i. Select Mount Partition and Enter /boot for Mount Point.
j. Click Finish.
54. In the Add RAID /dev/md0:
a. For RAID Type Select RAID 1 (Mirroring).
b. Select the 2 GB partition on the two devices (/dev/sde2 and /dev/sdf2) and click Add.
c. Click Next.
d. In the RAID Options, For Chuck Size select 512 KiB.
e. Click Next.
f. For Role select Operating System.
g. Click Next.
h. Select Format Partition, For File System type select swap.
i. Select Mount Partition and select swap for Mount Point.
j. Click Finish.
55. In the Add RAID /dev/md0:
a. For RAID Type Select RAID 1 (Mirroring).
b. Select the 100 GB partition on the two devices (/dev/sde3 and /dev/sdf3) and click Add.
c. Click Next.
d. In the RAID Options, For Chuck Size select 512 KiB.
e. Click Next.
f. For Role select Operating System.
g. Click Next.
h. Select Format Partition, For File System type select ext3.
i. Select Mount Partition and select / for Mount Point.
j. Click Finish.
56. In the Expert Partitioner Window, make sure all the RAID devices are created and click Accept.
57. In the Suggested Partitioning screen click Next.
58. Select the appropriate Time Zone and click Next.
59. In the Create New User screen, click Next. We would create user for mapr in the later screen.
60. Enter the Password for root User and Confirm Password.
61. In the Installation Settings screen click Kdump.
62. Select Disable Kdump and click OK.
63. In the Installation Settings screen click on Default Systemd Target.
64. Select Text mode for Available Targets and click OK.
65. In the Installation Settings screen click Install.
66. In the Confirm Installation screen, click Install.
The Install process begins and once complete, it will boot the system.
67. Follow the steps listed above to install SUSE Linux Enterprise Server 12 SP 1 on all MapR servers.
To customize the server in preparation for the MapR Installation, complete the following steps:
The operating system must be configured in such a way that the short name of the server is displayed for the command ‘hostname’ and Full Qualified Host Name is displayed with the command ‘hostname –d’.
1. ssh to the Server using Management IP address assigned to the server during installation.
2. Login as root and password.
3. Edit the Hostname:
vi /etc/HOSTNAME
<<hostname>>.<<Domain Name>>
Each MapR Server is configured with 4 vNIC device. Table 11 lists the IP Address information required to configure the IP address on the Operating System. IP Address and Subnet Mask provided below are example only, please configure IP address as per your environment.
Table 11 List the IP Address for SAP HANA Server
vNIC Name | VLAN ID | IP Address Range | Subnet Mask |
MapR-01 | <<var_mapr-01_vlan_id>> | 10.21.21.101 to 10.21.21.104 | 255.255.255.0 |
MapR-02 | <<var_mapr-02_vlan_id>> | 10.22.22.101 to 10.22.22.104 | 255.255.255.0 |
MapR-03 | <<var_mapr-03_vlan_id>> | 10.23.23.101 to 10.23.23.104 | 255.255.255.0 |
HANA-Storage | <<var_storage_vlan_id>> | 192.168.110.101 to 192.168.110.104 | 255.255.255.0 |
Management | <<var_mgmt_vlan_id>> | 192.168.196.101 to 192.168.196.104 | 255.255.0.0 |
1. To configure the network interface on the OS, it is required to identify the mapping of the ethernet device on the OS to vNIC interface on the Cisco UCS.
2. From the OS, execute the following command to get list of Ethernet device with MAC Address:
ifconfig -a|grep HWaddr
eth0 Link encap:Ethernet HWaddr 00:25:B5:2A:00:2C
eth1 Link encap:Ethernet HWaddr 00:25:B5:2B:00:28
eth2 Link encap:Ethernet HWaddr 00:25:B5:2B:00:29
eth3 Link encap:Ethernet HWaddr 00:25:B5:2A:00:2E
3. In Cisco UCS Manager, click the Servers tab in the navigation pane.
4. Expand Servers > Service Profile > root > Sub-Organization > HANA > MapR-01.
5. Click + to Expand. Click vNICs.
6. In the right pane a list of the vNICs with MAC Address are listed.
7. Note the MAC Address of the MapR-01 vNIC is “00:25:B5:2A:00:2C”.
8. By comparing the MAC Address on the OS and Cisco UCS, eth0 on OS will carry the VLAN for MapR-01.
9. Go to the network configuration directory and create a configuration for eth0:
/etc/sysconfig/network/
vi ifcfg-eth0
##
# Mapr-01 Network
##
BOOTPROTO='static'
BROADCAST=''
ETHTOOL_OPTIONS=''
IPADDR='<<IP address for MapR-01 example:10.21.21.101/24>>'
MTU='9000'
NAME='VIC Ethernet NIC'
NETWORK=''
REMOTE_IPADDR=''
STARTMODE='auto'
10. Repeat the steps 1 to 9 for each vNIC interface.
11. Add default gateway:
vi etc/sysconfig/network/routes
default 192.168.196.1 - -
Domain Name Service configuration must be done based on the local requirements.
Add DNS IP if it is required to access internet:
vi /etc/resolv.conf
nameserver <<IP of DNS Server1>>
nameserver <<IP of DNS Server2>>
search <<Domain_name>>
All nodes should be able to resolve internal network IP address, below is an example of 4 node host file with the entire network defined in the /etc/hosts file.
Example:
127.0.0.1 localhost
# special IPv6 addresses
::1 localhost ipv6-localhost ipv6-loopback
#
## MapR 01
#
10.21.221.131 mapr31.ciscolab.local mapr31
10.21.221.132 mapr32.ciscolab.local mapr32
10.21.221.133 mapr33.ciscolab.local mapr33
10.21.221.134 mapr34.ciscolab.local mapr34
#
## MapR 02
#
10.22.222.131 mapr31b.ciscolab.local mapr31b
10.22.222.132 mapr32b.ciscolab.local mapr32b
10.22.222.133 mapr33b.ciscolab.local mapr33b
10.22.222.134 mapr34b.ciscolab.local mapr34b
#
## HANA Storage Network
#
192.168.110.211 cishanaso11s.ciscolab.local cishanaso11s
192.168.110.212 cishanaso12s.ciscolab.local cishanaso12s
192.168.110.213 cishanaso13s.ciscolab.local cishanaso13s
192.168.110.214 cishanaso14s.ciscolab.local cishanaso14s
192.168.110.215 cishanaso15s.ciscolab.local cishanaso15s
192.168.110.216 cishanaso16s.ciscolab.local cishanaso16s
192.168.110.217 cishanaso17s.ciscolab.local cishanaso17s
192.168.110.218 cishanaso18s.ciscolab.local cishanaso18s
#
## MapR Storage
#
192.168.110.131 mapr31s.ciscolab.local mapr31s
192.168.110.132 mapr32s.ciscolab.local mapr32s
192.168.110.133 mapr33s.ciscolab.local mapr33s
192.168.110.134 mapr34s.ciscolab.local mapr34s
#
## MapR Storage Virtual IP
#
192.168.110.71 maprvip31
192.168.110.72 maprvip32
192.168.110.73 maprvip33
192.168.110.74 maprvip34
192.168.110.75 maprvip35
192.168.110.76 maprvip36
192.168.110.77 maprvip37
192.168.110.78 maprvip38
#
## Management
192.168.196.131 mapr31m.ciscolab.local mapr31m
192.168.196.132 mapr32m.ciscolab.local mapr32m
192.168.196.133 mapr33m.ciscolab.local mapr33m
192.168.196.134 mapr34m.ciscolab.local mapr34m
To updated and customize the SLES 12 SP 1 System for MapR Servers, complete the following steps:
1. Register SUSE Linux Enterprise installations with the SUSE Customer Center:
SUSEConnect -r <<Registration Code>> -e <<email address>>
2. Execute the following command to update the SLES 12 SP 1 to latest patch level:
zypper update
3. Follow the on-screen instruction to complete the update process.
4. Install additional packages:
zypper install tuned rpcbind java-1_8_0-openjdk java-1_8_0-openjdk-headless java-
1_8_0-openjdk-devel ipmitool
5. Disable transparent hugepages. Modify /etc/default/grub search for the line starting with "GRUB_CMDLINE_LINUX_DEFAULT" and append to this line:
transparent_hugepage=never
6. Save your changes and run:
grub2-mkconfig -o /boot/grub2/grub.cfg
After a reboot, THP will be disabled.
7. Add mapr user and group. Make sure that <uid> and <gid> are greater than 1000 and the same on all MapR cluster nodes. The mapr defaults of 2000 for both uid and gid will ensure consistency with other default MapR clusters for inter-cluster mirroring:
groupadd mapr -g <gid>
useradd mapr -u <uid> -g <gid>
8. Activate rpcbind:
service rpcbind start
chkconfig rpcbind on
9. Append following parameters in /etc/sysctl.conf:
sunrpc.tcp_slot_table_entries = 128
sunrpc.tcp_max_slot_table_entries = 128
net.ipv4.tcp_slow_start_after_idle = 0
10. Add the following line into /etc/modprobe.d/sunrpc-local.conf. Create the file, if it does not exist:
options sunrpc tcp_slot_table_entries=128
options sunrpc tcp_max_slot_table_entries=128
11. Activate tuned:
service tuned start
chkconfig tuned on
12. Enable tuned profile:
tuned-adm profile throughput-performance
13. Reboot the OS by issuing reboot command
To download the Cisco UCS Drivers ISO bundle, which contains most of the Cisco UCS Virtual Interface Card drivers, complete the following steps:
1. In a web browser, navigate to https://www.cisco.com.
2. Under Support, click All Downloads.
3. In the product selector, click Products, then click Server - Unified Computing.
4. If prompted, enter your Cisco.com username and password to log in.
You must be signed in to download Cisco Unified Computing System (UCS) drivers.
5. Cisco UCS drivers are available for both Cisco UCS B-Series Blade Server Software and Cisco UCS C-Series Rack-Mount UCS-Managed Server Software.
6. Click UCS B-Series Blade Server Software.
7. Click Cisco Unified Computing System (UCS) Drivers.
The latest release version is selected by default. This document is built on Version 3.2(2a).
8. Click 3.2(2a) Version.
9. Download ISO image of Cisco Unified Computing System (UCS) Drivers.
10. Choose ISO image of UCS-related linux drivers only and click Download and follow the prompts to complete your driver download.
11. After the download is complete browse to \ucs-bxxx-drivers-linux.3.2.2a\Network\Cisco\VIC\SLES\SLES12.2 and copy cisco-enic-kmp-default-2.3.0.30_k3.12.49_11-0.x86_64.rpm to each server.
12. ssh to the Server on the Management IP as root.
13. Update the enic driver with the following command:
rpm -Uvh /tmp/ cisco-enic-usnic-kmp-default-3.0.44.553.545.8_k4.4.21_69-
1.x86_64.rpm
14. Update the enic driver on all the MapR Servers.
The configuration of NTP is important and must be performed on all systems. To configure network time, complete the following step:
1. Configure NTP by adding at least one NTP server to the NTP config file /etc/ntp.conf:
vi /etc/ntp.conf
server <NTP-SERVER IP>
fudge <NTP-SERVER IP> stratum 10
keys /etc/ntp.keys
trustedkey 1
The SSH Keys must be exchanged between all MapR servers for user ‘root’. To exchange the SSH keys, complete the following steps:
1. Generate the rsa public key by executing the following command:
ssh-keygen -b 2048
Generating public/private rsa key pair.
Enter file in which to save the key (/root/.ssh/id_rsa):
Enter passphrase (empty for no passphrase):
Enter same passphrase again:
Your identification has been saved in /root/.ssh/id_rsa.
Your public key has been saved in /root/.ssh/id_rsa.pub.
The key fingerprint is:
62:8a:0b:c4:4c:d7:de:8d:dc:e7:b7:17:ca:c7:02:94 root@mapr21.ciscolab.local
The key's randomart image is:
+--[ RSA 2048]----+
| |
| . |
| . . . . |
|+ . . o + E |
| + .o+So.. |
|. . o . o. . |
|. . . .o.o .|
| . . .+.+ |
| . .+ |
+-----------------+
2. Exchange the rsa public key by executing the below command from First server to rest of the servers. This helps ensure that every host can establish a password-less ssh connection to every other host:
“ssh-copy-id -i /root/.ssh/id_rsa.pub mapr31”
ssh-copy-id -i /root/.ssh/id_rsa.pub mapr32
3. Repeat the steps above for all the servers in the MapR cluster.
The installation of the MapR software needs a repository with the MapR software packages. If the MapR cluster nodes have internet access, it is recommended to prepare an online repository. If the MapR cluster nodes do not have internet access, you can download the packages from a computer which has internet access, unzip and copy them to a server which is reachable by the MapR cluster nodes or to a MapR cluster node itself and prepare an offline repository.
The following information is required to setup the MapR Cluster.
Cluster Name: example: mapr500
For High Availability and network bandwidth distribution, virtual IPs are created on the MapR cluster. These virtual IPs are bind to HANA Storage Ethernet device. HANA servers will use these virtual IPs to mount the NFS volumes. If a MapR node is failed, the virtual IPs are moved to other active nodes and the HANA servers will continue to access NFS volumes without any disruption. Its recommended to create two virtual IP address per MapR server, for example if there are 4 MapR nodes in the cluster – Create 8 Virtual IPs.
IP Address: example: 192.168.110.101-108
Subnet Mask: example: 255.255.255.0 or /24
On every MapR cluster node, to add the MapR repository, complete the following steps:
1. Use the following command to add the repository for MapR packages:
zypper ar http://package.mapr.com/releases/v5.2.0/suse/ maprtech
2. If your connection to the Internet is through a proxy server, you must set the http_proxy environment variable before installation:
http_proxy=http://<host>:<port>
export http_proxy
3. Update the system package index by running the following command:
zypper refresh
4. MapR packages require a compatibility package in order to install and run on SUSE. Execute the following command to install the SUSE compatibility package:
zypper install mapr-compat-suse
Packages are installed based on the role of the node. To install the MapR packages, complete the following steps:
1. Install on MapR NFS and MapR Fileserver on all nodes in the MapR Cluster:
zypper in mapr-nfs mapr-fileserver
2. Install zookeeper service on last three nodes of the MapR Cluster:
zypper in mapr-zookeeper
3. Install cldb service on first two nodes of the MapR Cluster:
zypper in mapr-cldb
4. Install webserver service on all the nodes or at least on two nodes in the MapR Cluster:
zypper in mapr-webserver
After all packages are installed, follow the steps below to configure MapR cluster.
WARNING! Any changes after these steps without re-running the configuration script might result in an unstable cluster state.
1. Increase memory usage of MapR instance:
sed -i
's/service.command.mfs.heapsize.percent=.*/service.command.mfs.heapsize.percent=8
5/g' /opt/mapr/conf/warden.conf
2. Define the MapR subnets to use:
sed -i 's%s#export MAPR_SUBNETS=%export
MAPR_SUBNETS=<<internal_network_01>>/<<subnet_short>>,<<internal_network_02>>/<<s
ubnet_short>> %g' /opt/mapr/conf/env.sh
Example:
sed -i -e 's%#export MAPR_SUBNETS=%export
MAPR_SUBNETS=10.21.221.0/24,10.22.222.0/24 %g' /opt/mapr/conf/env.sh
3. Modify the DrCache size of the NFS server:
sed -i 's/#DrCacheSize =.*/DrCacheSize = 614400/g' /opt/mapr/conf/nfsserver.conf
4. Create the /mapr directory and the /opt/mapr/conf/mapr_fstab file:
mkdir -p /mapr
echo "localhost:/mapr /mapr hard,nolock" > /opt/mapr/conf/mapr_fstab
5. Run the configuration script. -C followed by the cldb nodes, -Z followed by the zookeeper nodes, -N gives the cluster name, and -no-autostart hinders the MapR software to start automatically. If you use hostnames instead of IP addresses for -C and -Z, be sure to use hostnames which reflect to the internal cluster network IP addresses.
/opt/mapr/server/configure.sh -C <<cldbnode1>>,<<cldbnode2>> -Z
<<zookeepernode1>>,<<zookeepernode2>>,<<zookeepernode3>> -N <<clustername>> -noautostart
Example:
/opt/mapr/server/configure.sh -C mapr31, mapr32 -Z mapr32,mapr33,mapr34 -N
mapr600 -no-autostart
6. Create a file with a disk list, containing all Virtual Drives in RAID 5 created:
lsblk | grep 8.2T | awk '{print "/dev/"$1}' > /tmp/disk.list
7. Set up the disks in the disk.list file with a storage pool of 1 disk each on all nodes:
WARNING! The disks will be wiped immediately!
/opt/mapr/server/disksetup -W 1 -F /tmp/disk.list
8. Repeat Step 1 through 7 on each MapR node in the cluster.
To issue the following commands to start the cluster, complete the following steps:
1. On zookeeper nodes:
service mapr-zookeeper start
2. On all nodes, including zookeeper nodes:
service mapr-warden start
3. Wait for about 5 minutes for the cluster to come up.
To add a license to the MapR Cluster, complete the following steps:
1. When MapR Cluster is installed and running, log in to any MapR cluster node and retrieve your unique MapR cluster ID:
maprcli license showid
2. Send the cluster ID and the number of MapR nodes, along with <customer info> to <MapR> to get your MapR License.
3. If your cluster has internet access, install your license directly from the MapR License Server. Select Admin from the drop-down list and select Cluster Setting. In the Admin/ Cluster Setting click License.
4. Select Import License, enter the credentials for mapr.com Account to retrieve the license.
5. Alternatively, if your cluster does not have internet access, you can copy your license file to any MapR cluster node, and install it via the command line interface:
maprcli license add -is_file true -license <license_file>
The MapR license file is a text file starting with the line "-----BEGIN SIGNED MESSAGE-----" and ending with the line "-- ---END MESSAGE HASH-----". You can also use "Add licenses via upload" or "Add licenses via copy/paste" in the "Manage Licenses" dialog in the MapR Control System GUI to install the license from a file.
In order to create virtual IP addresses for the MapR NFS server, MAC address of HANA-Storage vNIC on all the MapR nodes, IP address range for Virtual IP and Network subnet are required.
From each MapR Server execute ifconfig –a|grep HWaddr command to get list of Ethernet device with MAC Address. By comparing MAC Address on the OS and Cisco UCS Service Profile, eth2 on OS will carry the VLAN for HANA-Storage.
ifconfig -a |grep HWaddr
eth0 Link encap:Ethernet HWaddr 00:25:B5:2A:00:2C
eth1 Link encap:Ethernet HWaddr 00:25:B5:2B:00:28
eth2 Link encap:Ethernet HWaddr 00:25:B5:2B:00:29
eth3 Link encap:Ethernet HWaddr 00:25:B5:2A:00:2E
Below is the command to create Virtual IP on the MapR cluster:
maprcli virtualip add -macs <mac_addr_node1> <mac_addr_node2> <mac_addr_nodeN> -
virtualipend <<last_virtual_ip>> -netmask <<subnet_mask> -virtualip
<<first_virtual_ip>>
Example:
maprcli virtualip add -macs 00:25:b5:fb:00:29 00:25:b5:fb:00:22 00:25:b5:fb:00:24
00:25:b5:fb:00:26 -virtualipend 192.168.110.78 -netmask 255.255.255.0 -virtualip
192.168.110.71
The MapR cluster needs an internal directory structure for its volumes. It is recommended to store the volumes in the /apps folder of the MapR cluster directory structure. The MapR file system is mounted on the MapR cluster nodes under /mapr, so the following directory structure should be created on one of the MapR cluster node.
1. Confirm MapR file system is mounted at /mapr:
mount | grep mapr
localhost:/mapr on /mapr type nfs (rw,nolock,addr=127.0.0.1)
2. Create HANA directory:
mkdir -p /mapr/<clustername>/apps/hana/<SID>
Example:
mkdir -p /mapr/mapr600/apps/hana/ANA
When creating a volume, the volume name can differ from the directory name of MapR internal file system.
3. Create volume for /hana/shared:
maprcli volume create -name <volumename> -path /apps/hana/<SID>/<volumename> -
replicationtype high_throughput -replication 2 -minreplication 2
4. Create volumes for /hana/data for every single storage partition:
maprcli volume create -name <volumename> -path
/apps/hana/<SID>/<datamountpoint_storage_partition> -replicationtype
high_throughput -replication 2 -minreplication 2
5. Create volumes for /hana/log for every single storage partition:
maprcli volume create -name <volumename> -path
/apps/hana/<SID>/<logmountpoint_storage_partition> -replicationtype low_latency -
replication 2 -minreplication 2
6. Turn off the compression for log volumes:
hadoop mfs -setcompression off /mapr/<clustername>/apps/hana/<SID>/<volumename>
It is recommended to reflect the SID in the path. For the three different volume types shared, data and log.
7. Example for creating a volume for /hana/shared for SID "ANA":
maprcli volume create -name shared -path /apps/hana/ANA/shared -replicationtype
high_throughput -replication 2 -minreplication 2
8. Example for creating eight volumes for /hana/data for SID "ANA":
maprcli volume create -name data001 -path /apps/hana/ANA/data001 -replicationtype
high_throughput -replication 2 -minreplication 2
maprcli volume create -name data002 -path /apps/hana/ANA/data002 -replicationtype
high_throughput -replication 2 -minreplication 2
maprcli volume create -name data003 -path /apps/hana/ANA/data003 -replicationtype
high_throughput -replication 2 -minreplication 2
maprcli volume create -name data004 -path /apps/hana/ANA/data004 -replicationtype
high_throughput -replication 2 -minreplication 2
maprcli volume create -name data005 -path /apps/hana/ANA/data005 -replicationtype
high_throughput -replication 2 -minreplication 2
maprcli volume create -name data006 -path /apps/hana/ANA/data006 -replicationtype
high_throughput -replication 2 -minreplication 2
maprcli volume create -name data007 -path /apps/hana/ANA/data007 -replicationtype
high_throughput -replication 2 -minreplication 2
maprcli volume create -name data008 -path /apps/hana/ANA/data008 -replicationtype
high_throughput -replication 2 -minreplication 2
9. Example for creating eight volumes for /hana/log for SID "ANA":
maprcli volume create -name log001 -path /apps/hana/ANA/log001 -replicationtype
low_latency -replication 2 -minreplication 2
maprcli volume create -name log002 -path /apps/hana/ANA/log002 -replicationtype
low_latency -replication 2 -minreplication 2
maprcli volume create -name log003 -path /apps/hana/ANA/log003 -replicationtype
low_latency -replication 2 -minreplication 2
maprcli volume create -name log004 -path /apps/hana/ANA/log004 -replicationtype
low_latency -replication 2 -minreplication 2
maprcli volume create -name log005 -path /apps/hana/ANA/log005 -replicationtype
low_latency -replication 2 -minreplication 2
maprcli volume create -name log006 -path /apps/hana/ANA/log006 -replicationtype
low_latency -replication 2 -minreplication 2
maprcli volume create -name log007 -path /apps/hana/ANA/log007 -replicationtype
low_latency -replication 2 -minreplication 2
maprcli volume create -name log008 -path /apps/hana/ANA/log008 -replicationtype
low_latency -replication 2 -minreplication 2
10. Example to turn off compression on eight log volumes for SID "ANA":
hadoop mfs -setcompression off /mapr/mapr600/apps/hana/ANA/log001
hadoop mfs -setcompression off /mapr/mapr600/apps/hana/ANA/log002
hadoop mfs -setcompression off /mapr/mapr600/apps/hana/ANA/log003
hadoop mfs -setcompression off /mapr/mapr600/apps/hana/ANA/log004
hadoop mfs -setcompression off /mapr/mapr600/apps/hana/ANA/log005
hadoop mfs -setcompression off /mapr/mapr600/apps/hana/ANA/log006
hadoop mfs -setcompression off /mapr/mapr600/apps/hana/ANA/log007
hadoop mfs -setcompression off /mapr/mapr600/apps/hana/ANA/log008
This section provides the procedure for SUSE Linux Enterprise Server for SAP Applications 12 SP 2 Operating System and customizing for SAP HANA requirement.
For the latest information on SAP HANA installation and OS customization requirement, see the SAP HANA Installation Guide.
To install the SUSE Linux Enterprise Server for SAP Applications 12 SP on the local drives, complete the following steps:
1. In Cisco UCS Manager, click the Servers tab in the navigation pane.
2. Select Service Profile > root > Sub-Organization > HANA > HANA-Server01.
3. Click KVM Console.
4. When the KVM Console is launched, click Boot Server.
5. Click Virtual Media > Activate Virtual Devices.
a. Select the option Accept this Session for Unencrypted Virtual Media Session and then click Apply.
b. Click Virtual Media and Choose Map CD/DVD.
c. Click Browse to navigate ISO media location.
d. Click Map Device.
6. On the Initial screen select Installation to begin the installation process.
7. Choose Language and Keyboard layout and Select I Agree the License Terms and click Next.
8. On the Network Settings screen Under Overview, click VNIC Ethernet NIC.
9. To configure the network interface on the OS, it is required to identify the mapping of the Ethernet device on the OS to vNIC interface on the Cisco UCS.
a. In Cisco UCS Manager, click the Servers tab in the navigation pane.
b. Expand Servers > Service Profile > root > Sub-Organization > HANA > HANA-Server01.
c. Click + to Expand.
d. Click vNICs.
e. On the right pane list of the vNICs with MAC Address are listed.
f. Note that the MAC Address of the Mgmt vNIC is “00:25:B5:2A:00:02”.
g. With the vNIC order in the Service Profile, vNIC for Management should be eth4.
10. By comparing MAC Address on the OS and Cisco UCS, eth4 on OS will carry the VLAN for Management.
11. Click Edit, under the Address tab.
12. Click Statically Assigned IP Address:
a. In the IP Address field enter <<Management IP address>>.
b. In the Subnet Mask field enter <<subnet mask for Management Interface>>.
c. In the Hostname field enter the hostname for Management Interface.
13. Repeat Steps 9-12 for each VNIC device.
14. On the Network Settings screen Select Hostname/DNS:
a. In the Hostname field enter the Hostname.
b. In the Domain Name Field enter the Domain Name.
c. In the Name Server 1 field enter <<DNS server1>> and Name Server 2 field enter <<DNS server2>>.
d. In the Search domains field enter <<domain1.com,domain2.com>>.
e. Click Routing.
f. For the Default IPv4 Gateway enter the <<Default Gateway for>>.
g. Click Next.
15. In the Registration screen and click Skip Registration. You will update the system in the following section.
16. In the Product Installation Mode, select Proceed with standard SLES for SAP Application installation.
17. In the Add On Product screen and click Next.
18. In the Suggested Partitioning screen click Expert Partitioner.
19. In the Expert Partitioner screen, select the first Internal Drive /dev/sda for Operating System and click Add Partition.
20. In the Add Partition on /dev/sda, select Primary Partition and click Next.
21. For the New Partition Size, select Custom Size.
22. In the Size field enter 200 MB and click Next.
23. For Role, select Operating System and click Next.
24. Select Format Partition; for File System type select ext3.
25. Select Mount Partition and Enter /boot for Mount Point.
26. Click Finish.
27. In the Expert Partitioner screen, select the first internal Drive /dev/sda/ for Operating System and click Add Partition.
28. In the Add Partition on /dev/sda, select Primary Partition and click Next.
29. For Role, select Operating System and click Next.
30. For Format Partition, select Do not format partition.
31. For File system ID, select 0x8E Linux LVM.
32. Click Finish.
33. In the Expert Partitioner screen, select Volume Management. In the right pane click Add and select Volume Group.
34. In the Add Volume Group Screen.
35. Enter the systemvg for Volume Group Name.
36. Select the Available Physical Volume and click Add.
37. In the Expert Partitioner screen, select Volume Management. In the right pane click Add and select Logical Volume.
38. In the Add Logical Volume on /dev/systemvg screen.
39. Enter swapvol as the Name of Logical Volume.
40. For the Type choose Normal.
41. Select Custom Size and Enter 2 GB for Size.
42. Click Next.
43. For Role select Operating System, and click Next.
44. Select Format Partition, For File System type select swap.
45. For Mount Partition and select swap for Mount Point.
46. Click Finish.
47. In the Expert Partitioner screen, select Volume Management. On the right pane click Add and select Logical Volume.
48. In the Add Logical Volume on /dev/systemvg screen.
49. Enter rootvol as the Name of Logical Volume.
50. For the Type choose Normal.
51. Click Next.
52. Select Custom Size and Enter 100 GB for Size.
53. Click Next.
54. For Role select Operating System, and click Next.
55. Select Format Partition, For File System type select ext3.
56. For Mount Partition and select / for Mount Point.
57. Click Finish.
58. In the Expert Partitioner screen, click Accept.
59. In the Suggested Partitioning screen click Next.
60. Select the appropriate Time Zone and click Next.
61. Enter the Password for System Administrator “root” and Confirm Password, then click Next.
62. In the Installation Settings screen click Software.
63. Deselect GNOME Desktop Environment and select SAP HANA Server Base.
64. Click OK.
65. In the Installation Settings screen under Firewall and SSH, click Firewall will be enabled(disabled) to disable Firewall.
66. In the Installation Settings screen click Kdump.
67. Select Disable Kdump and click OK.
68. In the Installation Settings screen click on Default Systemd Target.
69. Select Text mode for Available Targets and click OK.
70. In the Installation Settings screen click Install.
71. In the Confirm Installation screen, click Install.
72. The Install process will begin and the System will reboot after OS Installation.
To customize the server in preparation for the HANA Installation, complete the following steps:
The operating system must be configured in such a way that the short name of the server is displayed for the command ‘hostname’ and Full Qualified Host Name is displayed with the command ‘hostname –d’.
1. ssh to the Server using Management IP address assigned to the server during installation.
2. Login as root and password.
3. Edit the Hostname:
vi /etc/HOSTNAME
<<hostname>>.<<Domain Name>>
Each SAP HANA Server is configured with 9 vNIC device. Table 12 lists the IP Address information required to configure the IP address on the Operating System.
The IP Address and Subnet Mask provided below are examples only, please configure the IP address for your environment.
Table 12 IP Addresses for SAP HANA Server
vNIC Name | VLAN ID | IP Address Range | Subnet Mask |
HANA-AppServer | <<var_appserver_vlan_id>> | 192.168.223.211 to 192.168.223.218 | 255.255.255.0 |
HANA-Backup | <<var_backup_vlan_id>> | 192.168.221.211 to 192.168.211.218 | 255.255.255.0 |
HANA-Client | <<var_client_vlan_id>> | 192.168.222.211 to 192.168.222.218 | 255.255.0.0 |
HANA-DataSource | <<var_datasource_vlan_id>> | 192.168.224.211 to 192.168.224.218 | 255.255.255.0 |
HANA-Internal | <<var_internal_vlan_id>> | 192.168.220.211 to 192.168.220.218 | 255.255.255.0 |
HANA-Replication | <<var_replication_vlan_id>> | 192.168.225.211 to 192.168.225.218 | 255.255.255.0 |
HANA-Storage | <<var_storage_vlan_id>> | 192.168.110.211 to 192.168.110.218 | 255.255.255.0 |
Management | <<var_mgmt_vlan_id>> | 192.168.196.211 to 192.168.196.218 | 255.255.0.0 |
NFS-IPMI | <<var_inband_vlan_id>> | 192.168.197.211 to 192.168.197.218 | 255.255.255.0 |
1. To configure the network interface on the OS, it is required to identify the mapping of the ethernet device on the OS to vNIC interface on the Cisco UCS.
2. From the OS execute the below command to get list of Ethernet device with MAC Address:
ifconfig -a |grep HWaddr
eth0 Link encap:Ethernet HWaddr 00:25:B5:2A:00:00
eth1 Link encap:Ethernet HWaddr 00:25:B5:2B:00:00
eth2 Link encap:Ethernet HWaddr 00:25:B5:2A:00:01
eth3 Link encap:Ethernet HWaddr 00:25:B5:2B:00:01
eth4 Link encap:Ethernet HWaddr 00:25:B5:2A:00:02
eth5 Link encap:Ethernet HWaddr 00:25:B5:2B:00:02
eth6 Link encap:Ethernet HWaddr 00:25:B5:2A:00:03
eth7 Link encap:Ethernet HWaddr 00:25:B5:2B:00:03
eth8 Link encap:Ethernet HWaddr 00:25:B5:2A:00:04
3. In Cisco UCS Manager, click the Servers tab in the navigation pane.
4. Expand Servers > Service Profile > root > Sub-Organization > HANA > HANAServer01.
5. Click + to Expand.
6. Click vNICs.
7. On the right pane list of the vNICs with MAC Address are listed.
8. Note the MAC Address of the HANA-Internal vNIC is “00:25:B5:2A:00:00”.
9. By comparing MAC Address on the OS and Cisco UCS, eth0 on OS will carry the VLAN for HANA-Internal.
10. Go to network configuration directory and create a configuration for eth0:
/etc/sysconfig/network/
vi ifcfg-eth0
BOOTPROTO='static'
BROADCAST=''
ETHTOOL_OPTIONS=''
IPADDR='<<IP subnet for HANA-Internal/subnet mask example:192.168.220.101/24>>
MTU='9000'
NAME='VIC Ethernet NIC'
NETWORK=''
REMOTE_IPADDR=''
STARTMODE='auto'
11. Repeat the steps 1 to 10 for each vNIC interface.
12. Add default gateway.
vi etc/sysconfig/network/routes
default 192.168.196.1 - -
Domain Name Service configuration must be done based on the local requirements.
Add DNS IP if it is required to access internet:
vi /etc/resolv.conf
nameserver <<IP of DNS Server1>>
nameserver <<IP of DNS Server2>>
search <<Domain_name>>
For scale-out system all nodes should be able to resolve internal network IP address, below is an example of 8 node host file with the entire network defined in the /etc/hosts file.
cat /etc/hosts
127.0.0.1 localhost
# special IPv6 addresses
::1 localhost ipv6-localhost ipv6-loopback
#
## Internal Network
#
192.168.220.211 cishanaso11.ciscolab.local cishanaso11
192.168.220.212 cishanaso12.ciscolab.local cishanaso12
192.168.220.213 cishanaso13.ciscolab.local cishanaso13
192.168.220.214 cishanaso14.ciscolab.local cishanaso14
192.168.220.215 cishanaso15.ciscolab.local cishanaso15
192.168.220.216 cishanaso16.ciscolab.local cishanaso16
192.168.220.217 cishanaso17.ciscolab.local cishanaso17
192.168.220.218 cishanaso18.ciscolab.local cishanaso18
#
## Storage Network
#
192.168.110.211 cishanaso11s.ciscolab.local cishanaso11s
192.168.110.212 cishanaso12s.ciscolab.local cishanaso12s
192.168.110.213 cishanaso13s.ciscolab.local cishanaso13s
192.168.110.214 cishanaso14s.ciscolab.local cishanaso14s
192.168.110.215 cishanaso15s.ciscolab.local cishanaso15s
192.168.110.216 cishanaso16s.ciscolab.local cishanaso16s
192.168.110.217 cishanaso17s.ciscolab.local cishanaso17s
192.168.110.218 cishanaso18s.ciscolab.local cishanaso18s
#
## Client Network
#
192.168.222.211 cishanaso11c.ciscolab.local cishanaso11c
192.168.222.212 cishanaso12c.ciscolab.local cishanaso12c
192.168.222.213 cishanaso13c.ciscolab.local cishanaso13c
192.168.222.214 cishanaso14c.ciscolab.local cishanaso14c
192.168.222.215 cishanaso15c.ciscolab.local cishanaso15c
192.168.222.216 cishanaso16c.ciscolab.local cishanaso16c
192.168.222.217 cishanaso17c.ciscolab.local cishanaso17c
192.168.222.218 cishanaso18c.ciscolab.local cishanaso18c
#
## AppServer Network
#
192.168.223.211 cishanaso11a.ciscolab.local cishanaso11a
192.168.223.212 cishanaso12a.ciscolab.local cishanaso12a
192.168.223.213 cishanaso13a.ciscolab.local cishanaso13a
192.168.223.214 cishanaso14a.ciscolab.local cishanaso14a
192.168.223.215 cishanaso15a.ciscolab.local cishanaso15a
192.168.223.216 cishanaso16a.ciscolab.local cishanaso16a
192.168.223.217 cishanaso17a.ciscolab.local cishanaso17a
192.168.223.218 cishanaso18a.ciscolab.local cishanaso18a
#
## Management Network
#
192.168.196.211 cishanaso11m.ciscolab.local cishanaso11m
192.168.196.212 cishanaso12m.ciscolab.local cishanaso12m
192.168.196.213 cishanaso13m.ciscolab.local cishanaso13m
192.168.196.214 cishanaso14m.ciscolab.local cishanaso14m
192.168.196.215 cishanaso15m.ciscolab.local cishanaso15m
192.168.196.216 cishanaso16m.ciscolab.local cishanaso16m
192.168.196.217 cishanaso17m.ciscolab.local cishanaso17m
192.168.196.218 cishanaso18m.ciscolab.local cishanaso18m
#
## Backup Network
#
192.168.221.211 cishanaso11b.ciscolab.local cishanaso11b
192.168.221.212 cishanaso12b.ciscolab.local cishanaso12b
192.168.221.213 cishanaso13b.ciscolab.local cishanaso13b
192.168.221.214 cishanaso14b.ciscolab.local cishanaso14b
192.168.221.215 cishanaso15b.ciscolab.local cishanaso15b
192.168.221.216 cishanaso16b.ciscolab.local cishanaso16b
192.168.221.217 cishanaso17b.ciscolab.local cishanaso17b
192.168.221.218 cishanaso18b.ciscolab.local cishanaso18b
#
## DataSource Network
#
192.168.224.211 cishanaso11d.ciscolab.local cishanaso11d
192.168.224.212 cishanaso12d.ciscolab.local cishanaso12d
192.168.224.213 cishanaso13d.ciscolab.local cishanaso13d
192.168.224.214 cishanaso14d.ciscolab.local cishanaso14d
192.168.224.215 cishanaso15d.ciscolab.local cishanaso15d
192.168.224.216 cishanaso16d.ciscolab.local cishanaso16d
192.168.224.217 cishanaso17d.ciscolab.local cishanaso17d
192.168.224.218 cishanaso18d.ciscolab.local cishanaso18d
## Replication Network
#
192.168.225.211 cishanaso11r.ciscolab.local cishanaso11r
192.168.225.212 cishanaso12r.ciscolab.local cishanaso12r
192.168.225.213 cishanaso13r.ciscolab.local cishanaso13r
192.168.225.214 cishanaso14r.ciscolab.local cishanaso14r
192.168.225.215 cishanaso15r.ciscolab.local cishanaso15r
192.168.225.216 cishanaso16r.ciscolab.local cishanaso16r
192.168.225.217 cishanaso17r.ciscolab.local cishanaso17r
192.168.225.218 cishanaso18r.ciscolab.local cishanaso18r
#
## IPMI External Address
#
192.168.196.181 cishanaso11e-ipmi
192.168.196.182 cishanaso12e-ipmi
192.168.196.183 cishanaso13e-ipmi
192.168.196.184 cishanaso14e-ipmi
192.168.196.185 cishanaso15e-ipmi
192.168.196.186 cishanaso16e-ipmi
192.168.196.187 cishanaso17e-ipmi
192.168.196.188 cishanaso18e-ipmi
#
## NFS-IPMI Inband Address
#
192.168.197.161 cishanaso11-ipmi
192.168.197.162 cishanaso12-ipmi
192.168.197.163 cishanaso13-ipmi
192.168.197.164 cishanaso14-ipmi
192.168.197.165 cishanaso15-ipmi
192.168.197.166 cishanaso16-ipmi
192.168.197.167 cishanaso17-ipmi
192.168.197.168 cishanaso18-ipmi
#
##MapR Storage
#
192.168.110.101 mapr11s.ciscolab.local mapr11s
192.168.110.102 mapr12s.ciscolab.local mapr12s
192.168.110.103 mapr13s.ciscolab.local mapr13s
192.168.110.104 mapr14s.ciscolab.local mapr14s
#
## MapR Storage Virtual IP
#
192.168.110.71 maprvip31
192.168.110.72 maprvip32
192.168.110.73 maprvip33
192.168.110.74 maprvip34
192.168.110.75 maprvip35
192.168.110.76 maprvip36
192.168.110.77 maprvip37
192.168.110.78 maprvip38
To get the IP address assigned to the Cisco UCS service profile for Inband Management, go to Cisco UCS Manager, click the LAN tab in the navigation pane, select Pools > root > Sub-Organization > HANA > IP Pools > IP Pools HANA-Servers. In the right pane click the IP Address tab that shows the IP address and assigned Service Profiles.
To updated and customize the SLES 12 SP 1 System for HANA Servers, complete the following steps:
1. Register SUSE Linux Enterprise installations with the SUSE Customer Center:
SUSEConnect -r <<Registration Code>> -e <<email address>>
2. Execute the below command to update the SLES4SAP 12 SP 1 to latest patch level:
zypper update
3. Follow the on-screen instruction to complete the update process.
4. Disable transparent hugepages, Configure C-States for lower latency in Linux, Auto NUMA settings.
5. Modify /etc/default/grub search for the line starting with "GRUB_CMDLINE_LINUX_DEFAULT" and append to this line:
transparent_hugepage=never intel_idle.max_cstate=1 processor.max_cstate=1
numa_balancing=disabled
6. Save your changes and run:
grub2-mkconfig -o /boot/grub2/grub.cfg
7. Add the following line into /etc/init.d/boot.local, for Energy Performance Bias EPB
cpupower set -b 0
8. Append following parameters in /etc/sysctl.conf:
net.ipv4.tcp_slow_start_after_idle = 0
net.ipv4.conf.all.rp_filter = 0
net.ipv4.ip_local_port_range = 40000 61000
net.ipv4.neigh.default.gc_thresh1 = 256
net.ipv4.neigh.default.gc_thresh2 = 1024
net.ipv4.neigh.default.gc_thresh3 = 4096
net.ipv6.neigh.default.gc_thresh1 = 256
net.ipv6.neigh.default.gc_thresh2 = 1024
net.ipv6.neigh.default.gc_thresh3 = 4096
net.core.rmem_max = 16777216
net.core.wmem_max = 16777216
net.core.rmem_default = 262144
net.core.wmem_default = 262144
net.core.optmem_max = 16777216
net.core.netdev_max_backlog = 300000
net.ipv4.tcp_rmem = 65536 262144 16777216
net.ipv4.tcp_wmem = 65536 262144 16777216
net.ipv4.tcp_no_metrics_save = 1
net.ipv4.tcp_moderate_rcvbuf = 1
net.ipv4.tcp_window_scaling = 1
net.ipv4.tcp_timestamps = 1
net.ipv4.tcp_sack = 1
sunrpc.tcp_slot_table_entries = 128
sunrpc.tcp_max_slot_table_entries = 128
net.ipv4.tcp_slow_start_after_idle = 0
9. Add the following line into /etc/modprobe.d/sunrpc-local.conf, create the file, if it does not exist:
options sunrpc tcp_slot_table_entries=128
options sunrpc tcp_max_slot_table_entries=128
10. Activate tuned:
saptune daemon start
11. Enable tuned profile:
saptune solution apply HANA
12. Reboot the OS by issuing reboot command
To download the Cisco UCS Drivers ISO bundle, which contains most of the Cisco UCS Virtual Interface Card drivers, complete the following steps:
1. In a web browser, navigate to https://www.cisco.com.
2. Under Support, click All Downloads.
3. In the product selector, click Products, then click Server - Unified Computing.
4. If prompted, enter your Cisco.com username and password to log in.
You must be signed in to download Cisco Unified Computing System (UCS) drivers.
5. Cisco UCS drivers are available for both Cisco UCS B-Series Blade Server Software and Cisco UCS C-Series Rack-Mount UCS-Managed Server Software.
6. Click UCS B-Series Blade Server Software.
7. Click Cisco Unified Computing System (UCS) Drivers.
The latest release version is selected by default. This document is built on Version 3.1(2b).
8. Click 3.1(2b) Version.
9. Download ISO image of Cisco Unified Computing System (UCS) Drivers.
10. Choose ISO image of UCS-related linux drivers only and click Download and follow the prompts to complete your driver download.
11. After the download is complete browse to \ucs-bxxx-drivers-linux.3.2.2a\Network\Cisco\VIC\SLES\SLES12.2 and copy cisco-enic-kmp-default-2.3.0.30_k3.12.49_11-0.x86_64.rpm to each server
12. ssh to the Server on the Management IP as root.
13. Update the enic driver with the following command;
rpm -Uvh /tmp/ cisco-enic-usnic-kmp-default-3.0.44.553.545.8_k4.4.21_69-1.x86_64.rpm
14. Update the enic driver on all the HANA Servers.
It is important that the time on all components used for SAP HANA is in sync. The configuration of NTP is important and to be performed on all systems.
1. Configure NTP by adding at least one NTP server to the NTP config file /etc/ntp.conf:
vi /etc/ntp.conf
server <NTP-SERVER IP>
fudge <NTP-SERVER IP> stratum 10
keys /etc/ntp.keys
trustedkey 1
The SSH Keys must be exchanged between all nodes in a SAP HANA Scale-Out system for user ‘root’ and user <SID>adm.
1. Generate the rsa public key by executing the command:
ssh-keygen -b 2048
2. The SSH Keys must be exchanged between all nodes in a SAP HANA Scale-Out system for user ‘root’ and user.
3. Exchange the rsa public key by executing the following command from the first server to rest of the servers in the scale-out system.
ssh-copy-id -i /root/.ssh/id_rsa.pub cishanaso12
4. Repeat the 1- 3 for all the servers in the SAP HANA system.
Mount options vary from the default Linux setting for using NFS for SAP HANA data and log volumes. The following is an example of /etc/fstab entry for an eight node SAP HANA Scale-Out system with SID ANA.
maprvip01:/mapr/mapr500/apps/hana/ANA/shared /hana/shared nfs nolock,hard,timeo=600 0 0
maprvip01:/mapr/mapr500/apps/hana/ANA/data001 /hana/data/ANA/mnt00001 nfs nolock,hard,timeo=600 0 0
maprvip02:/mapr/mapr500/apps/hana/ANA/data002 /hana/data/ANA/mnt00002 nfs nolock,hard,timeo=600 0 0
maprvip03:/mapr/mapr500/apps/hana/ANA/data003 /hana/data/ANA/mnt00003 nfs nolock,hard,timeo=600 0 0
maprvip04:/mapr/mapr500/apps/hana/ANA/data004 /hana/data/ANA/mnt00004 nfs nolock,hard,timeo=600 0 0
maprvip05:/mapr/mapr500/apps/hana/ANA/data005 /hana/data/ANA/mnt00005 nfs nolock,hard,timeo=600 0 0
maprvip06:/mapr/mapr500/apps/hana/ANA/data006 /hana/data/ANA/mnt00006 nfs nolock,hard,timeo=600 0 0
maprvip07:/mapr/mapr500/apps/hana/ANA/data007 /hana/data/ANA/mnt00007 nfs nolock,hard,timeo=600 0 0
maprvip08:/mapr/mapr500/apps/hana/ANA/data008 /hana/data/ANA/mnt00008 nfs nolock,hard,timeo=600 0 0
maprvip08:/mapr/mapr500/apps/hana/ANA/log001 /hana/log/ANA/mnt00001 nfs nolock,hard,timeo=600 0 0
maprvip07:/mapr/mapr500/apps/hana/ANA/log002 /hana/log/ANA/mnt00002 nfs nolock,hard,timeo=600 0 0
maprvip06:/mapr/mapr500/apps/hana/ANA/log003 /hana/log/ANA/mnt00003 nfs nolock,hard,timeo=600 0 0
maprvip05:/mapr/mapr500/apps/hana/ANA/log004 /hana/log/ANA/mnt00004 nfs nolock,hard,timeo=600 0 0
maprvip04:/mapr/mapr500/apps/hana/ANA/log005 /hana/log/ANA/mnt00005 nfs nolock,hard,timeo=600 0 0
maprvip03:/mapr/mapr500/apps/hana/ANA/log006 /hana/log/ANA/mnt00006 nfs nolock,hard,timeo=600 0 0
maprvip02:/mapr/mapr500/apps/hana/ANA/log007 /hana/log/ANA/mnt00007 nfs nolock,hard,timeo=600 0 0
maprvip01:/mapr/mapr500/apps/hana/ANA/log008 /hana/log/ANA/mnt00008 nfs nolock,hard,timeo=600 0 0
Create the required directory to mount /hana/shared /hana/data and /hana/log volumes. Mount all the volumes from /etc/fstab using “mount –a”. Check the status of all mounted volumes using “df –h” command:
maprvip01:/mapr/mapr500/apps/hana/ANA/shared
125T 194G 125T 1% /hana/shared
maprvip01:/mapr/mapr500/apps/hana/ANA/data001
125T 194G 125T 1% /hana/data/ANA/mnt00001
maprvip02:/mapr/mapr500/apps/hana/ANA/data002
125T 194G 125T 1% /hana/data/ANA/mnt00002
maprvip03:/mapr/mapr500/apps/hana/ANA/data003
125T 194G 125T 1% /hana/data/ANA/mnt00003
maprvip04:/mapr/mapr500/apps/hana/ANA/data004
125T 194G 125T 1% /hana/data/ANA/mnt00004
maprvip05:/mapr/mapr500/apps/hana/ANA/data005
125T 194G 125T 1% /hana/data/ANA/mnt00005
maprvip06:/mapr/mapr500/apps/hana/ANA/data006
125T 194G 125T 1% /hana/data/ANA/mnt00006
maprvip07:/mapr/mapr500/apps/hana/ANA/data007
125T 194G 125T 1% /hana/data/ANA/mnt00007
maprvip08:/mapr/mapr500/apps/hana/ANA/data008
125T 194G 125T 1% /hana/data/ANA/mnt00008
maprvip08:/mapr/mapr500/apps/hana/ANA/log001
125T 194G 125T 1% /hana/log/ANA/mnt00001
maprvip07:/mapr/mapr500/apps/hana/ANA/log002
125T 194G 125T 1% /hana/log/ANA/mnt00002
maprvip06:/mapr/mapr500/apps/hana/ANA/log003
125T 194G 125T 1% /hana/log/ANA/mnt00003
maprvip05:/mapr/mapr500/apps/hana/ANA/log004
125T 194G 125T 1% /hana/log/ANA/mnt00004
maprvip04:/mapr/mapr500/apps/hana/ANA/log005
125T 194G 125T 1% /hana/log/ANA/mnt00005
maprvip03:/mapr/mapr500/apps/hana/ANA/log006
125T 194G 125T 1% /hana/log/ANA/mnt00006
maprvip02:/mapr/mapr500/apps/hana/ANA/log007
125T 194G 125T 1% /hana/log/ANA/mnt00007
maprvip01:/mapr/mapr500/apps/hana/ANA/log008
125T 194G 125T 1% /hana/log/ANA/mnt00008
chmod -R 777 /hana/data
chmod -R 777 /hana/log
Change the directory permissions BEFORE installing HANA. Use the chown command after the file systems are mounted on each HANA node:
chmod -R 777 /hana/data
chmod -R 777 /hana/log
Please use the official SAP documentation, which describes the installation process with and without the SAP unified installer: SAP HANA Server Installation Guide
Read the SAP Notes before you start the installation (see, Important SAP Notes). These SAP Notes contain the latest information about the installation, as well as corrections to the installation documentation.
All other SAP installation and administration documentation is available here: http://service.sap.com/instguides
Read the following SAP Notes before you start the installation. These SAP Notes contain the latest information about the installation, as well as corrections to the installation documentation.
The latest SAP Notes can be found at: https://service.sap.com/notes.
SAP Note 1514967 - SAP HANA: Central Note
SAP Note 1523337 - SAP HANA Database: Central Note
SAP Note 2000003 - FAQ: SAP HANA
SAP Note 1730999 - Configuration changes in SAP HANA appliance
SAP Note 1514966 - SAP HANA 1.0: Sizing SAP In-Memory Database
SAP Note 1780950 - Connection problems due to host name resolution
SAP Note 1743225 - SAP HANA: Potential failure of connections with scale out nodes
SAP Note 1755396 - Released DT solutions for SAP HANA with disk replication
SAP Note 1890444 - HANA system slow due to CPU power save mode
SAP Note 1681092 - Support for multiple SAP HANA databases on a single SAP HANA appliance
SAP Note 1514966 - SAP HANA: Sizing SAP HANA Database
SAP Note 1637145 - SAP BW on HANA: Sizing SAP HANA Database
SAP Note 1793345 - Sizing for Suite on HANA
SAP Note 2235581 - SAP HANA: Supported Operating Systems
SAP Note 2009879 - SAP HANA Guidelines for RedHat Enterprise Linux (RHEL)
SAP Note 2247020 - SAP HANA DB: Recommended OS settings for RHEL 6.7
SAP Note 2228351 - SAP HANA Database SPS 11 revision 110 (or higher) on RHEL 6 or SLES 11
SAP Note 1944799 - SAP HANA Guidelines for SLES Operating System
SAP Note 2205917 - SAP HANA DB: Recommended OS settings for SLES 12 / SLES for SAP Applications 12
SAP Note 1731000 - Non-recommended configuration changes
SAP Note 1557506 - Linux paging improvements
SAP Note 1740136 - SAP HANA: wrong mount option may lead to corrupt persistency
SAP Note 1829651 - Time zone settings in SAP HANA scale out landscapes
SAP Note 1658845 - SAP HANA DB hardware check
SAP Note 1637145 - SAP BW on SAP HANA: Sizing SAP In-Memory Database
SAP Note 1661202 - Support for multiple applications on SAP HANA
SAP Note 1681092 - Support for multiple SAP HANA databases one HANA aka Multi SID
SAP Note 1577128 - Supported clients for SAP HANA 1.0
SAP Note 1808450 - Homogenous system landscape for on BW-HANA
SAP Note 1976729 - Application Component Hierarchy for SAP HANA
SAP Note 1927949 - Standard Behavior for SAP Logon Tickets
SAP Note 1577128 - Supported clients for SAP HANA
SAP Note 2186744 - FAQ: SAP HANA Parameters
SAP Note 2267798 - Configuration of the SAP HANA Database during Installation Using hdbparam
SAP Note 2156526 - Parameter constraint validation on section indices does not work correctly with hdbparam
SAP Note 2399079 - Elimination of hdbparam in HANA 2
SAP Note 1730928 - Using external software in a SAP HANA appliance
SAP Note 1730929 - Using external tools in an SAP HANA appliance
SAP Note 1730930 - Using antivirus software in an SAP HANA appliance
SAP Note 1730932 - Using backup tools with Backint for SAP HANA
SAP Note 1788665 - SAP HANA running on VMware vSphere VMs
For HANA Scale-Out, the ha_provider python class supports the STONITH functionality.
STONITH = Shoot the Other Node In The Head. With this python class, we are able to reboot the failing node to prevent a split brain and thus an inconsistency of the database. Since we use NFSv3, we must implement the STONITH functionality to prevent the database from a corruption because of multiple access to mounted file systems. If a HANA node is failed over to another node, the failed node will be rebooted from the master name server. This eliminates the risk of multiple accesses to the same file systems.
The used version of the ucs_ha_class.py must be at least 1.1
vi ucs_ha_class.py
"""
Function Class to call the reset program to kill the failed host and remove NFS locks for the SAP
HANA HA
Class Name ucs_ha_class
Class Phath /usr/sap/<SID>/HDB<ID>/exe/python_support/hdb_ha
Provider Cisco Systems Inc.
Version 1.1 (apiVersion=2 and hdb_ha.client import sudowers)
"""
from hdb_ha.client import StorageConnectorClient
import os
class ucs_ha_class(StorageConnectorClient):
apiVersion = 2
def __init__(self, *args, **kwargs):
super(ucs_ha_class, self).__init__(*args, **kwargs)
def stonith(self, hostname):
os.system ("/bin/logger STONITH HANA Node:" + hostname)
os.system ("/hana/shared/HA/ucs_ipmi_reset.sh " + hostname)
return 0
def about(self):
ver={"provider_company":"Cisco",
"provider_name" :"ucs_ha_class",
"provider_version":"1.0",
"api_version" :2}
self.tracer.debug('about: %s'+str(ver))
print '>> ha about',ver
return ver
@staticmethod
def sudoers():
return """ALL=NOPASSWD: /bin/mount, /bin/umount, /bin/logger"""
def attach(self,storages):
return 0
def detach(self, storages):
return 0
def info(self, paths):
pass
Prepare the script to match the Cisco UCS Manager configured ipmi username and password. Default is ipmi-user sapadm and ipmi-user-password cisco.
vi ucs_ipmi_reset.sh
#!/bin/bash
# Cisco Systems Inc.
# SAP HANA High Availability
# Version NFS_UCS v01
# changelog: 02/18/2016
if [ -z $1 ]
then
echo "please add the hostname to reset to the command line"
exit 1
fi
# Trim the domain name off of the hostname
host=`echo "$1" | awk -F'.' '{print $1}'`
PASSWD=cisco
USER=sapadm
echo $host-ipmi
system_down='Chassis Power is off'
system_up='Chassis Power is on'
power_down='power off'
power_up='power on'
power_status='power status'
#
# Shut down the server via ipmitool power off
#
/bin/logger `whoami`" Resetting the HANA Node $host because of an Nameserver reset command"
#Power Off
rc2=`/usr/bin/ipmitool -I lanplus -H $host-ipmi -U $USER -P $PASSWD $power_down`
sleep 20
#Status
rc3=`/usr/bin/ipmitool -I lanplus -H $host-ipmi -U $USER -P $PASSWD $power_status`
#Chassis Power is still on
if [ "$rc3" = "$system_down" ]
then
/bin/logger `whoami`" HANA Node $host switched from ON to OFF "
else
#Power Off again
/bin/logger `whoami`" HANA Node $host still online second try to shutdown... "
rc2=`/usr/bin/ipmitool -I lanplus -H $host-ipmi -U $USER -P $PASSWD $power_down`
sleep 20
#Status
rc3=`/usr/bin/ipmitool -I lanplus -H $host-ipmi -U $USER -P $PASSWD $power_status`
#Chassis Power is still on
if [ "$rc3" = "$system_down" ]
then
/bin/logger `whoami`" HANA Node $host switched from ON to OFF 2nd try"
else
/bin/logger `whoami`" Resetting the HANA Node $host failed "
exit 1
fi
fi
#Chassis Power is down and the server can be swiched back on
#
#The NFS locks are released
#We will start the server now to bring it back as standby node
#Chassis Power is off
power="off"
/bin/logger `whoami`" HANA Node $host will stay offline for 10 seconds..... "
sleep 10
/bin/logger `whoami`" Switching HANA Node $host back ON "
rc2=`/usr/bin/ipmitool -I lanplus -H $host-ipmi -U $USER -P $PASSWD $power_up`
sleep 20
#Status
rc3=`/usr/bin/ipmitool -I lanplus -H $host-ipmi -U $USER -P $PASSWD $power_status`
#Chassis Power is off
if [ "$rc3" = "$system_up" ]
then
/bin/logger `whoami`" HANA Node $host reset done, system is booting"
power="on"
exit 0
else
/bin/logger `whoami`" Switching HANA Node $host back ON failed first time..."
/bin/logger `whoami`" Switching HANA Node $host back ON second time..."
rc2=`/usr/bin/ipmitool -I lanplus -H $host-ipmi -U $USER -P $PASSWD $power_up`
sleep 20
rc3=`/usr/bin/ipmitool -I lanplus -H $host-ipmi -U $USER -P $PASSWD $power_status`
if [ "$rc3" = "$system_up" ]
then
/bin/logger `whoami`" HANA Node $host reset done, system is booting"
power="on"
exit 0
else
/bin/logger `whoami`" Resetting the HANA Node $host failed "
exit 1
fi
fi
#
# Server is power on and should boot - our work is done
#
Copy the HA scripts to the shared HA directory under /hana/shared/<SID>/HA (HANA nameserver is responsible to reset the failed node)
ssh cishana01
mkdir /hana/shared/HA
chown anaadm:sapsys /hana/shared/HA
scp ucs_ipmi_reset.sh /hana/shared/HA/
scp ucs_ha_class.py /hana/shared/HA/
chown anaadm:sapsys /hana/shared/HA/*
The SAP Storage Connector API provides a way to call a user procedure whenever the SAP HANA Nameserver triggers a node failover. The API requires the files mentioned above.
The procedure is executed on the master nameserver.
To activate the procedure in case of a node failover, the global.ini file in:
<HANA installdirectory>/<SID>/global/hdb/custom/config/
must be edited and the following entry must be added:
[Storage]
ha_provider = ucs_ha_class
ha_provider_path = /hana/shared/HA
cd /hana/shared/<SID>/global/hdb/custom/config
vi global.ini
[communication]
internal_network = 192.168.220/24
listeninterface = .internal
[internal_hostname_resolution]
192.168.220.211 = cishanaso11
192.168.220.212 = cishanaso12
192.168.220.213 = cishanaso13
192.168.220.218 = cishanaso18
192.168.220.214 = cishanaso14
192.168.220.216 = cishanaso16
192.168.220.215 = cishanaso15
192.168.220.217 = cishanaso17
[persistence]
basepath_datavolumes = /hana/data/ANA
basepath_logvolumes = /hana/log/ANA
[storage]
ha_provider = ucs_ha_class
ha_provider_path = /hana/shared/HA
[trace]
ha_ucs_ha_class = info
Modify the /etc/sudoers file and append the below line on all the nodes. By adding the line <sid>adm account can execute commands mentioned without password.
cishana01:/ # vi /etc/sudoers
<sid>adm ALL=NOPASSWD: /bin/mount, /bin/umount, /bin/logger, /sbin/multipath,
/sbin/multipathd, /usr/bin/sg_persist, /etc/init.d/multipathd, /bin/kill,
/usr/bin/lsof, /sbin/vgchange, /sbin/vgscan
To activate the change, please restart the SAP HANA DB.
Test the ipmi connectivity on ALL nodes:
cishana01:~ # ipmitool -I lanplus -H cishana01-ipmi -U sapadm -P cisco power
status
Chassis Power is on
Make sure that all the nodes are responding to the ipmitool command and the IP address for the ipmi network match in the /etc/hosts file of all the servers.
Shailendra Mruthunjaya, Cisco Systems, Inc.
Shailendra is a Technical Marketing Engineer with Cisco UCS Solutions and Performance Group. Shailendra has over six years of experience with SAP HANA on Cisco UCS platform. Shailendra has designed several SAP landscapes in public and private cloud environment. Currently, his focus is on developing and validating infrastructure best practices for SAP applications on Cisco UCS Servers, Cisco Nexus products and Storage technologies.
Matthias Schlarb, Cisco Systems, Inc.
Matthias is part of the Cisco SAP Competence Center in Walldorf where he conducts workshops with customers and partners. As a Technical Marketing Engineer, he develops best practices for SAP Cloud Infrastructure on Cisco Unified Computing System.
For their support and contribution to the design, validation, and creation of this CVD, we would like to thank:
· Ralf Klahr, Cisco Systems, Inc.
· Ulrich Kleidon, Cisco Systems, Inc.
· Erik Lillestolen, Cisco Systems, Inc.
· Arun Kumar Thiruttani Kalathi, Cisco Systems, Inc.
· Andy Lerner, MapR Technologies, Inc.
· James Sun, MapR Technologies, Inc.