When connecting to an ESX host for the first time, rpowermt prompts the administrator for the username and password of the ESX host. Each new ESX host managed by rpowermt generates a prompt a username and password. The administrator is prompted for a lockbox password once, rpowermt securely stores the following information in a lockbox(encrypted

-         --ESX host being accessed

-     --Username and password of ESX host being accessed,

Storing the host and passwords in a lockbox enables them to be mainitained across system reboots. If a lockbox is copied from one rpowermt server to anotherm, the user is prompted to enter the lockbox password again.

PowerPath/VE for vSphere rpowermt commands do not require root access to run the executable; however, the ESX root password is required as an entry in the lockbox, Access to rpowermt command executable is based on the native access controls of the server (Windows or RHEL) where rpowermt is installed.

PowerPath/VE  5.4 and supported ESX/ESXi versions

PowerPath/VE release

Supported VMware ESX/ESXi version



4.0 U1/U2

4.0 U1/U2




- vCLI install only

- vCLI install only

- vCLI install only

- vCLI install only




- vCLI install only

- vCLI install only

- Update Manager and vCLI install

- Update Manager and vCLI install

- vCLI install only



Note:- Please check ESM for prequisites for PPVE before installing/upgrading.

& 1) vSphere hosts must be part of a DRS cluster.

2) 2) vSphere hosts must have shared storage. If there is not shared storage, create it.

3) 3) Place the first vSphere host in Maintenance Mode. This will force VMs to fail over to
other cluster members using Vmotion.

4)  4)  Install PowerPath/VE on the first vSphere host using vCLI.

Using Vmware Remote Tool (vCLI) on server, install PowerPath/VE on maintenance mode server, use the following command:

vihostupdate --server “server-IP-Address” --install --bunde=/<path>/EMCPower.VMWARE.5.4.bxxx.zip

Note:- Use vihostupdate.pl for Windows

Once the command completes, verify that package is installed. Use the following command.

vihostupdate --server  “server-IP-Address” –query

5) If necessary, make changes to the first vSphere host’s claim rules.

6)   6) Exit Maintenance Mode.

7)   7)  Reboot the first vSphere host. (Wait host to come online before proceeding next steps.)

8)   8)  Place the second vSphere host in Maintenance Mode. This will force VMs to fail over to
other clusters member using Vmotion.

9)   9)  Install PowerPath/VE on the second vSphere host using vCLI.

10)   If necessary, make changes to the second vSphere host’s claim rules.

11)   Exit Maintenance Mode.

12)   Reboot the second vSphere host.

13)   Perform the same operation for remaining hosts in the cluster.

14)   After PowerPath/VE installation has completed for every node in the cluster, rebalance the VMs across the cluster.

EMC Unified Storage is 20% more efficient that the competition. Even Erik Estrada from CHiPS knows that now.It's simple to be efficient with EMC

EMC unified storage makes it simple for you to contain information growth and satisfy the needs of your data-hungry virtual machines while requiring 20% less raw capacity. This works to your advantage when comparing EMC to other unified storage arrays. If, for some unlikely reason, EMC isn’t 20% more efficient, we'll match the shortfall with additional storage—that's how confident we are.
Find out more :- http://bit.ly/EMCGuaranteeST

EMC Unisphere - Next generation storage management that provides single and simple interface for both current and future clariion and celerra series. Unisphere provides capability to integrate other element manager and provides built in support tool, software download and live chat etc. Note :-  Unisphere features will be available in Q3 2010:

Check out demo about EMC Unisphere, presented by Bob Abraham :


• Task-based navigation and controls offer an intuitive, context-based approach to configuring storage, creating replicas, monitoring the environment, managing host connections, and accessing the Unisphere Support ecosystem.

Self-service Unisphere support ecosystem is accessible with 1-click from Unisphere, providing users with quick access to “real-time” support tools including live chat support, software downloads, product documentation, best practices, FAQs, online customer communities, ordering spares, and submitting service requests.

• Customizable dashboard views and reporting capabilities enable at-a-glance management by automatically presenting end-users with valuable information in context of how they manage storage. For example, customers can develop custom reports 18x faster with EMC Unisphere.

BCV copies can be used for backup, restore, decision support, and application testing., BCV devices contain no data after initial Symmetrix configuration. The full establish operation must be used the first time the standard devices are paired with their BCV devices.

1.  Associate the BCV Device for pairing:

To perform standard/BCV pairing, the standard and BCV mirror devices of your production images must be members of the same device group (Note:- Already discussed in previous post about creating device group and pairing devices). 

To associate BCV001 with device 0ABC,enter:

  symbcv –g DgName –sid SymmID associate dev 0ABC BCV001

Or to associate a range of devices to a device group, enter:

 symbcv –g DgName –sid SymmID associateall dev –RANGE 0ABC:0DEF

Note:- -sid SymmID is optional if you have already defined device group in symcli environment varaiable.

2. Unmount the BCV device:

Prior to using devices for BCV operations, the BCV device should be Windows formatted and assigned a drive letter.

If using basic disks on the Windows platform, you must unmount the BCV devices. If using dynamic disks, you must deport the entire TimeFinder device group.  For basic disks, use the syminq command to determine the SymDevName of the potential BCV device. For dynamic disks,use the TimeFinder  symntctl command to determine the volume and disk group name as follows:

symntctl list –volume [dg DgName]

Note that the term device group and dynamic disk group are the same applied to this command.

Unmount the selected BCV device as follows (with TimeFinder command):

symntctl unmount –drive z

Where z equals the designated drive letter.  If an error occurs, check for an “openhandle” and clear this condition.

For Veritas dynamic disks only, you must deport the disk group and rescan using the following commands:

vxdg deport –g DgName

symntctl rescan

3. Fully Establish BCV and STD:

To obtain a copy of the data on a standard device, the BCV device of the pair must be established.

To initiate a full establish on a specific standard/BCV device pair, target the standard device:

symmir –g DgName –full establish DEV001

Fully Establish all pairs in a group. To initiate a full establish on all BCV pairs in a device group, enter:

symmir –g DgName –full establish

Verify the completed (synchronized) establish operation. To verify when the BCV pairs reach the full copied or Synchronized state, use the verify action as follows:

symmir –g DgName –i 20 verify

With this interval and count, the message is displayed every 20 seconds until the pair is established.

Rescan the drive connections. For Dynamic disks only on a Windows host, you should rescan for drive connections visible to the host:

symntctl rescan

After any standard/BCV pair has been fully established and subsequently split, to save establish (resync) time, you can perform an establish operation omitting the -full option, which updates the BCV copy with only the changed tracks that occurred on the standard device during the elapsed BCV split time. To perform an incremental establish, omit the –full option, targeting the standard device
of the pair:

symmir -g DgName establish DEV001

Optionally, you can also collectively target all devices in a device group, composite group, or defined devices in a device file:

  symmir –g DgName establish [-full]

 symmir –g CgName establish [-full]

  symmir –file FileName establish [-full]

4. Prepare (freeze) Production database for a TimeFinder Split:

To prepare to split the synchronized BCV device from the production standard device, you must suspend I/O at the application layer or unmount the production standard prior to executing the split operation.

symioctl freeze –type DbType [object]

Ensure any residual cache on the Production host is fully flushed to disk. To insure all pending unwritten production file system entries are captured, enter TimeFinder command:

symntctl flush –drive z

Wait 30 to 60 seconds for the flush operation to complete.   

5. Split the BCV devices:

To split all the BCV devices from the standard devices in the production device group, enter:

symmir –g DgName split

To split a specific standard/BCV pair, target the logical device name in the group, enter:

symmir –g DgName split DEV001

6. Verify the split operation completes:

To verify when the BCV device is completely split from the standard, use the verify action as follows:

symmir –g DgName –i 20 verify –split -bg

With this interval and count, the message is displayed every 20 seconds until the pair is split.

7. Rescan for dynamic disks:

For dynamic disks only, you should rescan for drive connections visible to the host:

symntctl rescan

8. Mounting BCV device:

After splitting the BCV device, you can mount the device with captured data on another host and reassign the drive letter.

For basic disks, use the TimeFinder command:

symntctl  mount –drive z –dg DgName

For Dynamic disks, use the TimeFinder command:

 symntctl mount –drive z –vol VolName   –dg DgName | -guid VolGuid

For VERITAS dynamic disk only, you must deport the disk group and rescan, as follows:

vxdg deport –dg DgName

symntctl rescan

For Dynamic disk only (without Veritas), you can use the Microsoft diskpart command to select the disk and import the device using the online and import actions.

Note:- symntctl command available in TimeFinder/IM ( Integration Module).

VMware hosts require few mandatory FA bits setting before SAN storage to be provision. Apart from FA bits a series of procedure require from installing HBAs, HBA Firmware and drivers, zoning, mapping, masking devices, to configure kernel files and devices.

Let’s assume we have already identified Symmetrix FA port for VMware host and completed zoning on switch. It is better to have separate FA pair for VMware host. (You can connect VMware host to 2 pair FA if you have enough FA resources available and going to deploy critical application which require more performance).

You can identify the FA port available on Symmetrix:

symcfg list –connections.

Verify port flag settings-

symcfg list –fa -p -v

( FA-Number and Port where your host connected/zoned)

The following FA bits/flag require being set/Enable:

                     i)    Common Serial Number (C)

                    ii)    VCM State (VCM) --- (ACLX for V-MAX)

                    iii)    SCSI 3 (SC3)

                    iv)    SPC 2

                    v)     Unique World Wide Name (UWWN)

                   vi)     Auto-negotiation (EAN)

                   vii)    Point to Point (P)

Note :- FA bit/flag requirement may vary depending on Symmetrix, but most of times you require to enable above bit for VMware host.

Create a command file for setting FA port flags, call it faflags.cmd with the below entry:

# For C-Bit

set port FA:Port Common_Serial_Number=enable;

set port FA:Port Common_Serial_Number=enable;

# For VCM-Bit

set port FA:Port VCM_State=enable;

set port FA:Port VCM_State=enable;

# For SC3-Bit

set port FA:Port scsi_3=enable;

set port FA:Port scsi_3=enable;

# For SP-2-Bit

set port FA:Port SPC2_Protocol_Version=enable;

set port FA:Port SPC2_Protocol_Version=enable;

# For UWWN-Bit

set port FA:Port Unique_WWN=enable;

set port FA:Port Unique_WWN=enable;

# For EAN-Bit

set port FA:Port Auto_Negotiate=enable;

set port FA:Port Auto_Negotiate=enable;

# For PTOP-Bit

set port FA:Port Init_Point_to_Point=enable;

set port FA:Port Init_Point_to_Point=enable;

Once you prepare command file, you can commit the file:

symconfigure –sid preview –f  faflags.cmd

Verify port flag settings once again, required FA flags should have be enabled by now-

symcfg list –fa -p -v

You are ready to provision SAN storage for VMware host now…

CLARiiON Flare release 29 ( introduce support for several new features as follow:

1) Virtualization-aware Navisphere Manager - Discovery of VMware client always were difficult in earlier release but Flare 29 enables CLARiiON CX4 users and VMware administrators to reduce infrastructure reporting time from hours to minutes. Earlier releases have allowed only a single IP address to be assigned to each iSCSI physical port. With Flare 29, the ability to define multiple virtual iSCSI ports on each physical port has been added along with the ability to tag each virtual port with unique VLAN tag. VLAN tagging has also been added to the single Management Port interface. It should be noted that the IP address and VLAN tag assignments should be carefully coordinated with those supporting the network infrastructure where the storage system will operate

2) Built-in policy-based spin down of idle SATA II drives for CLARiiON CX4 - Lowers power requirements in environments such as test and development in physical and virtual environments. Features include a simple management via a “set it and forget it” policy, complete spin down of inactive drives during times of zero I/O activity, and drives automatically spin back up after a "first I/O" request is received.

3) Virtual Provisioning Phase 2 - Support for MirrorView and SAN Copy replication on thin LUNs has been added.

4) Search feature – Provides users with the ability to search for a wide-variety of objects across their storage systems. Objects can be either logical (e.g., LUN) or physical (e.g., disks).

5) Replication roles - Three additional roles have to be added in Navisphere: “Local Replication Only”, “Replication” and “Replication/ Recovery”.

6) Dedicated VMware software files - VMware software files (i.e. NaviSecCLI, Navisphere Initialization Wizard) are now separate from those of the Linux Operating System.

7) Software filename standardization - all CLARiiON software filenames beginning with FLARE Release 29

8) Changing SP IP addresses - SP IP addresses can now be changed without rebooting the SP. Only the Management Sever will need to be rebooted from the Setup page, which results in no storage system downtime.

9) Linux 64-bit server software – Native 64-bit Linux server software files simplify installation by eliminating the need to gather and load 32-bit DLLs.

10) Solaris x64 Navisphere Host Agent – Release 29 marks the introduction of Solaris 64-bit Navisphere Host Agent software. This Host Agent is backward compatible with older FLARE release.

About Me

My photo
Sr. Solutions Architect; Expertise: - Cloud Design & Architect - Data Center Consolidation - DC/Storage Virtualization - Technology Refresh - Data Migration - SAN Refresh - Data Center Architecture More info:- diwakar@emcstorageinfo.com
Blog Disclaimer: “The opinions expressed here are my personal opinions. Content published here is not read or approved in advance by EMC and does not necessarily reflect the views and opinions of EMC.”