InfoScale™ 9.0 Solutions Guide - Linux
- Section I. Introducing InfoScale
- Section II. Solutions for InfoScale products
- Solutions for InfoScale products
- Use cases for InfoScale products
- Feature support across InfoScale 9.0 products
- Using SmartMove and Thin Provisioning with Sybase databases
- Running multiple parallel applications within a single cluster using the application isolation feature
- Scaling FSS storage capacity with dedicated storage nodes using application isolation feature
- Finding InfoScale product use cases information
- Solutions for InfoScale products
- Section III. Stack-level migration to IPv6 or dual stack
- Section IV. Improving database performance
- Overview of database accelerators
- Improving database performance with Concurrent I/O
- Improving database performance with atomic write I/O
- About the atomic write I/O
- Requirements for atomic write I/O
- Restrictions on atomic write I/O functionality
- How the atomic write I/O feature of Storage Foundation helps MySQL databases
- VxVM and VxFS exported IOCTLs
- Configuring atomic write I/O support for MySQL on VxVM raw volumes
- Configuring atomic write I/O support for MySQL on VxFS file systems
- Dynamically growing the atomic write capable file system
- Disabling atomic write I/O support
- Section V. Using point-in-time copies
- Understanding point-in-time copy methods
- Backing up and recovering
- Storage Foundation and High Availability solutions backup and recovery methods
- Preserving multiple point-in-time copies
- Online database backups
- Backing up on an off-host cluster file system
- Database recovery using Storage Checkpoints
- Backing up and recovering in a NetBackup environment
- Off-host processing
- Creating and refreshing test environments
- Creating point-in-time copies of files
- Section VI. Maximizing storage utilization
- Optimizing storage tiering with SmartTier
- About SmartTier
- About VxFS multi-volume file systems
- About VxVM volume sets
- About volume tags
- SmartTier use cases for Sybase
- Setting up a filesystem for storage tiering with SmartTier
- Relocating old archive logs to tier two storage using SmartTier
- Relocating inactive tablespaces or segments to tier two storage
- Relocating active indexes to premium storage
- Relocating all indexes to premium storage
- Optimizing storage with Flexible Storage Sharing
- Optimizing storage tiering with SmartTier
- Section VII. Migrating data
- Understanding data migration
- Offline migration from LVM to VxVM
- Offline conversion of native file system to VxFS
- Online migration of a native file system to the VxFS file system
- About online migration of a native file system to the VxFS file system
- Administrative interface for online migration of a native file system to the VxFS file system
- Migrating a native file system to the VxFS file system
- Migrating a source file system to the VxFS file system over NFS v4
- Backing out an online migration of a native file system to the VxFS file system
- VxFS features not available during online migration
- Migrating storage arrays
- Migrating data between platforms
- Overview of the Cross-Platform Data Sharing (CDS) feature
- CDS disk format and disk groups
- Setting up your system to use Cross-platform Data Sharing (CDS)
- Maintaining your system
- Disk tasks
- Disk group tasks
- Changing the alignment of a disk group during disk encapsulation
- Changing the alignment of a non-CDS disk group
- Splitting a CDS disk group
- Moving objects between CDS disk groups and non-CDS disk groups
- Moving objects between CDS disk groups
- Joining disk groups
- Changing the default CDS setting for disk group creation
- Creating non-CDS disk groups
- Upgrading an older version non-CDS disk group
- Replacing a disk in a CDS disk group
- Setting the maximum number of devices for CDS disk groups
- Changing the DRL map and log size
- Creating a volume with a DRL log
- Setting the DRL map length
- Displaying information
- Determining the setting of the CDS attribute on a disk group
- Displaying the maximum number of devices in a CDS disk group
- Displaying map length and map alignment of traditional DRL logs
- Displaying the disk group alignment
- Displaying the log map length and alignment
- Displaying offset and length information in units of 512 bytes
- Default activation mode of shared disk groups
- Additional considerations when importing CDS disk groups
- File system considerations
- Considerations about data in the file system
- File system migration
- Specifying the migration target
- Using the fscdsadm command
- Checking that the metadata limits are not exceeded
- Maintaining the list of target operating systems
- Enforcing the established CDS limits on a file system
- Ignoring the established CDS limits on a file system
- Validating the operating system targets for a file system
- Displaying the CDS status of a file system
- Migrating a file system one time
- Migrating a file system on an ongoing basis
- When to convert a file system
- Converting the byte order of a file system
- Alignment value and block size
- Migrating a snapshot volume
- Migrating from Oracle ASM to VxFS
- Section VIII. InfoScale 4K sector device support solution
- Section IX. REST API support
- Support for configurations and operations using REST APIs
- Support for InfoScale operations using REST APIs
- Supported operations
- Configuring the REST server
- Security considerations for REST API management
- Authorization of users for performing operations using REST APIs
- Reconfiguring the REST server
- Configuring HA for the REST server
- Accessing the InfoScale REST API documentation
- Unconfiguring the REST server
- Troubleshooting information
- Limitations
- Support for configurations and operations using REST APIs
Migrating a source file system to the VxFS file system over NFS v4
InfoScale lets you migrate an Ext4 or XFS source file system to the VxFS file system over NFS v4. When you migrate a file system over NFS v4, the NFS server exports the file system to the server's NFS clients. During the migration, only a single NFS client can mount the exported file system. The application on the host is only available on this NFS client node during migration.
The following procedure describes how to migrate an Ext4 file system to VxFS over NFS v4; you can perform similar steps with the appropriate commands for XFS.
Note:
You can neither unmount the target (VxFS) file system nor the source file system after you start the migration. Only the commit or abort operation can unmount the target file system. Do not force unmount the source file system; use the abort operation to stop the migration and then unmount the source file system.
Do not modify the exported file system from the NFS server during the migration.
To migrate an Ext4 file system to the VxFS file system over NFS v4
- Install Storage Foundation on the physical application host.
Refer to the InfoScale Installation Guide.
- Add new storage to the physical application host on which you will configure Volume Manager (VxVM).
- Create a VxVM volume according to the your desired configuration on the newly added storage. The volume size cannot be less than source file system size.
- Mount the source file system if the file system is not mounted already.
# mount -t ext4 /dev/sda /src_mnt
The source file system to be migrated is exported from NFS server, and the source file system is mounted on a single NFS client. The migration runs on this NFS client.
- Verify whether the exported file system is mounted on the NFS client from where the further tasks are to be performed; if not, run the following command to mount it:
# mount -t nfs -o vers=4 NFSHost:/src_mnt /dest_mnt
- Log on to the NFS client system. All further steps should be performed on the NFS client.
- Run the fsmigadm analyze command and ensure that all checks pass:
# fsmigadm analyze /dev/vx/dsk/dg/vol1 /dest_mnt
- If the application is online, then shut down the application.
- Start the migration by running fsmigadm start:
# fsmigadm start /dev/vx/dsk/dg/vol1 /dest_mnt
- While the migration operation proceeds, you can bring the application online.
- You can get the status of the migration using the fsmigadm status command:
# fsmigadm status /dest_mnt
- Check the log file for any errors during migration. If you find any errors, you must copy the indicated files manually from the source file system after performing the commit operation.
- After the migration operation completes, shut down the application.
- Commit the migration:
# fsmigadm commit /dest_mnt
The fsmigadm command unmounts the source file system, unmounts the target file system, then mounts the target file system.
- Start the application on the Storage Foundation stack.