Hyper-V was an only child in our environment until a few large outages caught the eye of management and end users. A fellow admin and I pitched the perfect solution. Sure, VMware will cost you more, but you are buying rock solid stability. Besides, neither of us had ever observed an ESXi outage before.
We asked for it all, distributed switches and host profiles; we wanted our environment to be consistent. Testing went smooth and we were ready to start accepting production tenants. The first three hosts went smooth and we had VM’s in the double digits. We were confident and more importantly management was confident. So much that it was decided to start moving our XenApp environment off of Hyper-V and on to VMware. Our new platform was going to be in the spotlight in front of 10,000 users.
The migration of our Hyper-V PVS image to VMware was challenging but eventually overcome. Our XenApp users were now on a new hypervisor and had no idea a change of this scale had even occurred. The migration went perfect and things were going great until 11:58 am on May 31st 2014. Sitting in my inbox was an e-mail with the subject “vSphere HA initiated a virtual machine failover action.” Shit.
This was one of our XenApp cluster hosts and affected maybe 40 users. Thankfully it decided to go down on a Saturday. Had it gone down on a weekday it could have impacted 500 users. The PSOD showed Exception 14 in world blah blah blah. A quick Google search came up with VMware KB1020181 which explained that a page was being requested that hadn’t been loaded into memory. Needing further explanation I opened a case with VMware.
If you’ve ever opened a case with VMware you know how complicated it can be just to get things going. What’s the difference between my account number and my customer number? Why do I have to provide the order number for my license? You’ll be doing yourself and your fellow admins a huge favor by compiling all that information beforehand rather than fumbling through the customer portal. We employ Confluence for our in house documentation and all this information is now in there. Aside from that VMware support is amazing.
The support engineer asked for an export of our system logs from vCenter. This was pretty straight forward, but I found the following steps to be more direct and useful for my own use.
- Start the SSH service on your host. You don’t normally keep this service running, do you?
- Open an SSH session and log on to the host
- CD to /var/core
- Execute following to extract the relevant information from the VMkernel dump
# esxcfg-dumppart -L vmkernel-zdumpfilename.1 Created file vmkernel-log.1
- Open WinSCP and download vmkernel-log.1 to your local system.
This log file showed the following:
2014-05-31T15:57:32.579Z cpu14:9607609)@BlueScreen: #PF Exception 14 in world 9607609:vmklinux_9:h IP 0x418018b7e594 addr 0x0 PTEs:0x206ba74023;0x12f727023;0x12f728023;0x0; 2014-05-31T15:57:32.579Z cpu14:9607609)Code start: 0x418018200000 VMK uptime: 29:20:14:10.166 2014-05-31T15:57:32.579Z cpu14:9607609)0x412426e5dda0:[0x418018b7e594]hpsa_update_scsi_devices@#+0x39c stack: 0x410b0c037060 2014-05-31T15:57:32.580Z cpu14:9607609)0x412426e5de20:[0x418018b7f28f]hpsa_scan_start@#+0x187 stack: 0x412426e5de60 2014-05-31T15:57:32.580Z cpu14:9607609)0x412426e5de90:[0x418018b807af]hpsa_kickoff_rescan@#+0x20f stack: 0x0 2014-05-31T15:57:32.580Z cpu14:9607609)0x412426e5df30:[0x41801890175d]email@example.comAPI#9.2+0x185 stack: 0x0 2014-05-31T15:57:32.581Z cpu14:9607609)0x412426e5df80:[0x4180188fee5b]LinuxStartFunc@com.vmware.driverAPI#9.2+0x97 stack: 0x100d 2014-05-31T15:57:32.581Z cpu14:9607609)0x412426e5dfd0:[0x4180182bb14f]vmkWorldFunc@vmkernel#nover+0x83 stack: 0x0 2014-05-31T15:57:32.581Z cpu14:9607609)0x412426e5dff0:[0x418018453532]CpuSched_StartWorld@vmkernel#nover+0xfa stack: 0x0 2014-05-31T15:57:32.584Z cpu14:9607609)base fs=0x0 gs=0x418043800000 Kgs=0x0 2014-05-31T15:57:32.584Z cpu14:9607609)vmkernel 0x0 .data 0x0 .bss 0x0
VMware support verified that our POSD was caused by the HPSA driver as documented in KB2075978 which references HP Advisory c04302261. HPSA version 22.214.171.124 is known to cause PSOD under ESXi 5.5. A quick check verified that we were running the version in question.
# cat /proc/scsi/hpsa/2 HP HPSA Driver (v 126.96.36.199-1OEM)
At this point 188.8.131.52 hadn’t been released yet so the only solution was to roll back to 184.108.40.206.
To install the async driver:
- Browse to the local datastore on the host and upload the offline-bundle.zip. In this case we used hpsa-5.5.0-offline_bundle-1287942.zip.
- Place the host in maintenance mode
- Enable SSH on the host and open an SSH session as root
- Verify the current version of the driver to be replaced as documented above
- CD to the root of the datastore
# cd vmfs/volumes/localdatastore/
- Copy the offline bundle to the /var/log/vmware folder
# cp hpsa-5.5.0-offline_bundle-1287942.zip /var/log/vmware
- Install the drivers
# esxcli software vib install -d /var/log/vmware/hpsa-5.5.0-offline_bundle-1287942.zip Installation Result Message: The update completed successfully, but the system needs to be rebooted for the changes to be effective. Reboot Required: true VIBs Installed: Hewlett-Packard_bootbank_scsi-hpsa_220.127.116.11-1OEM.518.104.22.1688611 VIBs Removed: Hewlett-Packard_bootbank_scsi-hpsa_22.214.171.124-1OEM.5126.96.36.1991820 VIBs Skipped:
- Reboot the host
# reboot -f
- After the reboot I verified that the drivers were properly loaded
# cat /proc/scsi/hpsa/2 HP HPSA Driver (v 188.8.131.52-1OEM)