Running Oracle OLTP workloads in HP Integrity VM 4.3
5
Integrity VM host configuration
The Integrity BL860c i2 Server Blade is configured as follows:
• VM host
2
CPU: Two sockets, each populated with an Intel
®
Itanium
®
Processor 9340 Series; the Itanium
9340 is a quad-core CPU, and it delivers clock frequency of 1.6 GHz, with a 20 MB cache; it has
hyperthreading enabled
• VM host memory: 96 GB, mostly NUMA memory configuration
• VM host storage: One dual-port 8 Gb/s FC mezzanine card, each port connected to an 8 Gb/s FC
interconnect module in an HP BladeSystem c7000 Enclosure; intctl(1M) confirms interrupts from each port is
serviced by a different CPU core:
– The first module port is connected to controller port A of the HP 4400 Enterprise Virtual Array (EVA)
– The second module port is connected to controller port B of the 4400 EVA
• VM host networking: One 10GbE LAN-on-motherboard (LOM) interface connected to an HP 5400zl switch. All
network virtual switches are created using AVIO. 10GbE interface settings:
– Large receive offload = on
– TCP segmentation offload = on
– Transmit checksum offload = on
– Receive checksum offload = on
– Multi-queue = 4
• HP-UX OS configuration
– HP-UX 11i v3 March 2011 Release
– Integrity VM 4.3 (included in the March 2011 11i v3 release)
– Kernel tunable settings
Ο base_pagesize=64
Storage configuration
The 4400 EVA is configured as follows:
• Three disk shelves, each fully populated with 146 GB 15k RPM disks.
• Dual active controllers, each with 2 GB of battery-backed cache.
Two disk groups created to store Oracle data files and redo logs separately:
• Oracle data disk group
– 28 physical disks
– 72 virtual disks evenly distributed
between controller A and B
– RAID 1
• Oracle redo log disk group
– 8 physical disks
– 16 virtual disks evenly distributed
between controller A and B
– RAID 1
2
Integrity VM hypervisor layer