Dynamic Multi-Pathing 7.4.1 Administrator's Guide - AIX
- Understanding DMP
- Setting up DMP to manage native devices
- About setting up DMP to manage native devices
- Displaying the native multi-pathing configuration
- Migrating LVM volume groups to DMP
- Migrating to DMP from EMC PowerPath
- Migrating to DMP from Hitachi Data Link Manager (HDLM)
- Migrating to DMP from IBM Multipath IO (MPIO) or MPIO path control module (PCM)
- Using Dynamic Multi-Pathing (DMP) devices with Oracle Automatic Storage Management (ASM)
- Enabling Dynamic Multi-Pathing (DMP) devices for use with Oracle Automatic Storage Management (ASM)
- Removing Dynamic Multi-Pathing (DMP) devices from the listing of Oracle Automatic Storage Management (ASM) disks
- Migrating Oracle Automatic Storage Management (ASM) disk groups on operating system devices to Dynamic Multi-Pathing (DMP) devices
- Adding DMP devices to an existing LVM volume group or creating a new LVM volume group
- Removing DMP support for native devices
- Dynamic Multi-Pathing for the Virtual I/O Server
- About Dynamic Multi-Pathing in a Virtual I/O server
- About the Volume Manager (VxVM) component in a Virtual I/O server
- Configuring Dynamic Multi-Pathing (DMP) on Virtual I/O server
- Configuring Dynamic Multi-Pathing (DMP) pseudo devices as virtual SCSI devices
- Extended attributes in VIO client for a virtual SCSI disk
- Administering DMP
- About enabling and disabling I/O for controllers and storage processors
- About displaying DMP database information
- Displaying the paths to a disk
- Setting customized names for DMP nodes
- Configuring DMP for SAN booting
- Administering the root volume group (rootvg) under DMP control
- Running the bosboot command when LVM rootvg is enabled for DMP
- Extending an LVM rootvg that is enabled for DMP
- Reducing the native rootvg that is enabled for DMP
- Mirroring the root volume group
- Removing the mirror for the root volume group (rootvg)
- Cloning a LVM rootvg that is enabled for DMP
- Cleaning up the alternate disk volume group when LVM rootvg is enabled for DMP
- Using mksysb when the root volume group is under DMP control
- Upgrading Dynamic Multi-Pathing and AIX on a DMP-enabled rootvg
- Using Storage Foundation in the logical partition (LPAR) with virtual SCSI devices
- Setting up DMP for vSCSI devices in the logical partition (LPAR)
- About disabling DMP for vSCSI devices in the logical partition (LPAR)
- Preparing to install or upgrade Storage Foundation with DMP disabled for vSCSI devices in the logical partition (LPAR)
- Disabling DMP multi-pathing for vSCSI devices in the logical partition (LPAR) after installation or upgrade
- Adding and removing DMP support for vSCSI devices for an array
- How DMP handles I/O for vSCSI devices
- Running alt_disk_install, alt_disk_copy and related commands on the OS device when DMP native support is enabled
- Administering DMP using the vxdmpadm utility
- Retrieving information about a DMP node
- Displaying consolidated information about the DMP nodes
- Displaying the members of a LUN group
- Displaying paths controlled by a DMP node, controller, enclosure, or array port
- Displaying information about controllers
- Displaying information about enclosures
- Displaying information about array ports
- User-friendly CLI outputs for ALUA arrays
- Displaying information about devices controlled by third-party drivers
- Displaying extended device attributes
- Suppressing or including devices from VxVM control
- Gathering and displaying I/O statistics
- Setting the attributes of the paths to an enclosure
- Displaying the redundancy level of a device or enclosure
- Specifying the minimum number of active paths
- Displaying the I/O policy
- Specifying the I/O policy
- Disabling I/O for paths, controllers, array ports, or DMP nodes
- Enabling I/O for paths, controllers, array ports, or DMP nodes
- Renaming an enclosure
- Configuring the response to I/O failures
- Configuring the I/O throttling mechanism
- Configuring Subpaths Failover Groups (SFG)
- Configuring Low Impact Path Probing (LIPP)
- Displaying recovery option values
- Configuring DMP path restoration policies
- Stopping the DMP path restoration thread
- Displaying the status of the DMP path restoration thread
- Configuring Array Policy Modules
- Administering disks
- About disk management
- Discovering and configuring newly added disk devices
- Partial device discovery
- About discovering disks and dynamically adding disk arrays
- About third-party driver coexistence
- How to administer the Device Discovery Layer
- Listing all the devices including iSCSI
- Listing all the Host Bus Adapters including iSCSI
- Listing the ports configured on a Host Bus Adapter
- Listing the targets configured from a Host Bus Adapter or a port
- Listing the devices configured from a Host Bus Adapter and target
- Getting or setting the iSCSI operational parameters
- Listing all supported disk arrays
- Excluding support for a disk array library
- Re-including support for an excluded disk array library
- Listing excluded disk arrays
- Listing disks claimed in the DISKS category
- Displaying details about an Array Support Library
- Adding unsupported disk arrays to the DISKS category
- Removing disks from the DISKS category
- Foreign devices
- Changing the disk device naming scheme
- Discovering the association between enclosure-based disk names and OS-based disk names
- Dynamic Reconfiguration of devices
- About online Dynamic Reconfiguration
- Reconfiguring a LUN online that is under DMP control using the Dynamic Reconfiguration tool
- Manually reconfiguring a LUN online that is under DMP control
- Overview of manually reconfiguring a LUN
- Manually removing LUNs dynamically from an existing target ID
- Manually adding new LUNs dynamically to a new target ID
- About detecting target ID reuse if the operating system device tree is not cleaned up
- Scanning an operating system device tree after adding or removing LUNs
- Manually cleaning up the operating system device tree after removing LUNs
- Manually replacing a host bus adapter online
- Changing the characteristics of a LUN from the array side
- Upgrading the array controller firmware online
- Event monitoring
- Performance monitoring and tuning
- About tuning Dynamic Multi-Pathing (DMP) with templates
- DMP tuning templates
- Example DMP tuning template
- Tuning a DMP host with a configuration attribute template
- Managing the DMP configuration files
- Resetting the DMP tunable parameters and attributes to the default values
- DMP tunable parameters and attributes that are supported for templates
- DMP tunable parameters
- DMP driver tunables
- Appendix A. DMP troubleshooting
- Appendix B. Reference
Adding unsupported disk arrays to the DISKS category
Disk arrays should be added as JBOD devices if no Array Support Library (ASL) is available for the array.
JBODs are assumed to be Active/Active (A/A) unless otherwise specified. If a suitable ASL is not available, an A/A-A, A/P, or A/PF array must be claimed as an Active/Passive (A/P) JBOD to prevent path delays and I/O failures. If a JBOD is ALUA-compliant, it is added as an ALUA array.
Warning:
This procedure ensures that Dynamic Multi-Pathing (DMP) is set up correctly on an array that is not supported by Veritas Volume Manager (VxVM). Otherwise, VxVM treats the independent paths to the disks as separate devices, which can result in data corruption.
To add an unsupported disk array to the DISKS category
- Use the following command to identify the vendor ID and product ID of the disks in the array:
# /etc/vx/diag.d/vxscsiinq device_name
where device_name is the device name of one of the disks in the array. Note the values of the vendor ID (VID) and product ID (PID) in the output from this command. For Fujitsu disks, also note the number of characters in the serial number that is displayed.
The following example output shows that the vendor ID is SEAGATE and the product ID is ST318404LSUN18G.
Vendor id (VID) : SEAGATE Product id (PID) : ST318404LSUN18G Revision : 8507 Serial Number : 0025T0LA3H
- Stop all applications, such as databases, from accessing VxVM volumes that are configured on the array, and unmount all file systems and Storage Checkpoints that are configured on the array.
- If the array is of type A/A-A, A/P, or A/PF, configure it in autotrespass mode.
- Enter the following command to add a new JBOD category:
# vxddladm addjbod vid=vendorid [pid=productid] \ [serialnum=opcode/pagecode/offset/length] \ [cabinetnum=opcode/pagecode/offset/length] policy={aa|ap}]where vendorid and productid are the VID and PID values that you found from the previous step. For example, vendorid might be FUJITSU, IBM, or SEAGATE. For Fujitsu devices, you must also specify the number of characters in the serial number as the length argument (for example, 10). If the array is of type A/A-A, A/P, or A/PF, you must also specify the policy=ap attribute.
Continuing the previous example, the command to define an array of disks of this type as a JBOD would be:
# vxddladm addjbod vid=SEAGATE pid=ST318404LSUN18G
- Use the vxdctl enable command to bring the array under VxVM control.
# vxdctl enable
- To verify that the array is now supported, enter the following command:
# vxddladm listjbod
The following is sample output from this command for the example array:
VID PID SerialNum CabinetNum Policy (Cmd/PageCode/off/len) (Cmd/PageCode/off/len) ============================================================== SEAGATE ALL PIDs 18/-1/36/12 18/-1/10/11 Disk SUN SESS01 18/-1/36/12 18/-1/12/11 Disk - To verify that the array is recognized, use the vxdmpadm listenclosure command as shown in the following sample output for the example array:
# vxdmpadm listenclosure
ENCLR_NAME ENCLR_TYPE ENCLR_SNO STATUS ARRAY_TYPE LUN_COUNT FIRMWARE ======================================================================= Disk Disk DISKS CONNECTED Disk 2 -
The enclosure name and type for the array are both shown as being set to Disk. You can use the vxdisk list command to display the disks in the array:
# vxdisk list DEVICE TYPE DISK GROUP STATUS punr710vm04_disk_1 auto:none - - online invalid punr710vm04_disk_2 auto:none - - online invalid punr710vm04_disk_3 auto:none - - online invalid punr710vm04_disk_4 auto:none - - online invalid sda auto:none - - online invalid xiv0_9148 auto:none - - online invalid thinrclm ...
- To verify that the DMP paths are recognized, use the vxdmpadm getdmpnode command as shown in the following sample output for the example array:
# vxdmpadm getdmpnode enclosure=Disk NAME STATE ENCLR-TYPE PATHS ENBL DSBL ENCLR-NAME ============================================================================== punr710vm04_disk_1 ENABLED Disk 1 1 0 disk punr710vm04_disk_2 ENABLED Disk 1 1 0 disk punr710vm04_disk_3 ENABLED Disk 1 1 0 disk punr710vm04_disk_4 ENABLED Disk 1 1 0 disk sda ENABLED Disk 1 1 0 disk ...
The output in this example shows that there are two paths to the disks in the array.
For more information, enter the command vxddladm help addjbod.
See the vxddladm(1M) manual page.
See the vxdmpadm(1M) manual page.