Storage Foundation 7.4.1 Administrator's Guide - Linux
- Section I. Introducing Storage Foundation
- Overview of Storage Foundation
- How Dynamic Multi-Pathing works
- How Veritas Volume Manager works
- How Veritas Volume Manager works with the operating system
- How Veritas Volume Manager handles storage management
- Volume layouts in Veritas Volume Manager
- Online relayout
- Volume resynchronization
- Dirty region logging
- Volume snapshots
- FastResync
- How VxVM handles hardware clones or snapshots
- Volume encryption
- How Veritas File System works
- Section II. Provisioning storage
- Provisioning new storage
- Advanced allocation methods for configuring storage
- Customizing allocation behavior
- Using rules to make volume allocation more efficient
- Understanding persistent attributes
- Customizing disk classes for allocation
- Specifying allocation constraints for vxassist operations with the use clause and the require clause
- Creating volumes of a specific layout
- Customizing allocation behavior
- Creating and mounting VxFS file systems
- Creating a VxFS file system
- Mounting a VxFS file system
- tmplog mount option
- ioerror mount option
- largefiles and nolargefiles mount options
- Resizing a file system
- Monitoring free space
- Extent attributes
- Section III. Administering multi-pathing with DMP
- Administering Dynamic Multi-Pathing
- Discovering and configuring newly added disk devices
- About discovering disks and dynamically adding disk arrays
- How to administer the Device Discovery Layer
- Administering DMP using the vxdmpadm utility
- Gathering and displaying I/O statistics
- Specifying the I/O policy
- Discovering and configuring newly added disk devices
- Dynamic Reconfiguration of devices
- Reconfiguring a LUN online that is under DMP control using the Dynamic Reconfiguration tool
- Manually reconfiguring a LUN online that is under DMP control
- Managing devices
- Displaying disk information
- Changing the disk device naming scheme
- Adding and removing disks
- Event monitoring
- Administering Dynamic Multi-Pathing
- Section IV. Administering Storage Foundation
- Administering sites and remote mirrors
- About sites and remote mirrors
- Fire drill - testing the configuration
- Changing the site name
- Administering the Remote Mirror configuration
- Failure and recovery scenarios
- Administering sites and remote mirrors
- Section V. Optimizing I/O performance
- Veritas File System I/O
- Veritas Volume Manager I/O
- Managing application I/O workloads using maximum IOPS settings
- Section VI. Using Point-in-time copies
- Understanding point-in-time copy methods
- When to use point-in-time copies
- About Storage Foundation point-in-time copy technologies
- Volume-level snapshots
- Storage Checkpoints
- About FileSnaps
- About snapshot file systems
- Administering volume snapshots
- Traditional third-mirror break-off snapshots
- Full-sized instant snapshots
- Creating instant snapshots
- Adding an instant snap DCO and DCO volume
- Controlling instant snapshot synchronization
- Creating instant snapshots
- Cascaded snapshots
- Adding a version 0 DCO and DCO volume
- Administering Storage Checkpoints
- Storage Checkpoint administration
- Administering FileSnaps
- Administering snapshot file systems
- Understanding point-in-time copy methods
- Section VII. Optimizing storage with Storage Foundation
- Understanding storage optimization solutions in Storage Foundation
- Migrating data from thick storage to thin storage
- Maintaining Thin Storage with Thin Reclamation
- Reclamation of storage on thin reclamation arrays
- Identifying thin and thin reclamation LUNs
- Veritas InfoScale 4k sector device support solution
- Section VIII. Maximizing storage utilization
- Understanding storage tiering with SmartTier
- Creating and administering volume sets
- Multi-volume file systems
- Features implemented using multi-volume file system (MVFS) support
- Adding a volume to and removing a volume from a multi-volume file system
- Volume encapsulation
- Load balancing
- Administering SmartTier
- About SmartTier
- Placement classes
- Administering placement policies
- File placement policy rules
- Multiple criteria in file placement policy rule statements
- Using SmartTier with solid state disks
- Sub-file relocation
- Administering hot-relocation
- How hot-relocation works
- Moving relocated subdisks
- Deduplicating data
- Compressing files
- About compressing files
- Use cases for compressing files
- Section IX. Administering storage
- Managing volumes and disk groups
- Rules for determining the default disk group
- Moving volumes or disks
- Monitoring and controlling tasks
- Performing online relayout
- Adding a mirror to a volume
- Managing disk groups
- Disk group versions
- Displaying disk group information
- Importing a disk group
- Moving disk groups between systems
- Importing a disk group containing hardware cloned disks
- Handling conflicting configuration copies
- Destroying a disk group
- Backing up and restoring disk group configuration data
- Managing plexes and subdisks
- Decommissioning storage
- Rootability
- Encapsulating a disk
- Rootability
- Sample supported root disk layouts for encapsulation
- Encapsulating and mirroring the root disk
- Administering an encapsulated boot disk
- Quotas
- Using Veritas File System quotas
- File Change Log
- Managing volumes and disk groups
- Section X. Reference
- Appendix A. Reverse path name lookup
- Appendix B. Tunable parameters
- Tuning the VxFS file system
- Methods to change Dynamic Multi-Pathing tunable parameters
- Tunable parameters for VxVM
- Methods to change Veritas Volume Manager tunable parameters
- Appendix C. Command reference
Encapsulating a disk
Warning:
Encapsulating a disk requires that the system be rebooted several times. Schedule performance of this procedure for a time when this does not inconvenience users.
This section describes how to encapsulate a disk for use in VxVM. Encapsulation preserves any existing data on the disk when the disk is placed under VxVM control.
A root disk can be encapsulated and brought under VxVM control. However, there are restrictions on the layout and configuration of root disks that can be encapsulated.
Use the format or fdisk commands to obtain a printout of the root disk partition table before you encapsulate a root disk. For more information, see the appropriate manual pages. You may need this information should you subsequently need to recreate the original root disk.
You cannot grow or shrink any volume (rootvol, usrvol, varvol, optvol, swapvol, and so on) that is associated with an encapsulated root disk. This is because these volumes map to physical partitions on the disk, and these partitions must be contiguous.
Disks with msdos disk labels can be encapsulated as auto:sliced disks provided that they have at least one spare primary partition that can be allocated to the public region, and one spare primary or logical partition that can be allocated to the private region.
Disks with sun disk labels can be encapsulated as auto:sliced disks provided that they have at least two spare slices that can be allocated to the public and private regions.
Extensible Firmware Interface (EFI) disks with gpt (GUID Partition Table) labels can be encapsulated as auto:sliced disks provided that they have at least two spare slices that can be allocated to the public and private regions.
The entry in the partition table for the public region does not require any additional space on the disk. Instead it is used to represent (or encapsulate) the disk space that is used by the existing partitions.
Unlike the public region, the partition for the private region requires a small amount of space at the beginning or end of the disk that does not belong to any existing partition or slice. By default, the space required for the private region is 32MB, which is rounded up to the nearest whole number of cylinders. On most modern disks, one cylinder is usually sufficient.
To encapsulate a disk for use in VxVM
- Before encapsulating a root disk, set the device naming scheme used by VxVM to be persistent.
# vxddladm set namingscheme={osn|ebn} persistence=yes
For example, to use persistent naming with enclosure-based naming:
# vxddladm set namingscheme=ebn persistence=yes
- Select Encapsulate one or more disks from the vxdiskadm main menu.
Your system may use device names that differ from the examples shown here.
At the following prompt, enter the disk device name for the disks to be encapsulated:
Select disk devices to encapsulate: [<pattern-list>,all,list,q,?] device name
The pattern-list can be a single disk, or a series of disks. If pattern-list consists of multiple items, those items must be separated by white space.
If you do not know the address (device name) of the disk to be encapsulated, enter l or list at the prompt for a complete listing of available disks.
- To continue the operation, enter y (or press Return) at the following prompt:
Here is the disk selected. Output format: [Device] device name Continue operation? [y,n,q,?] (default: y) y
- Select the disk group to which the disk is to be added at the following prompt:
You can choose to add this disk to an existing disk group or to a new disk group. To create a new disk group, select a disk group name that does not yet exist. Which disk group [<group>,list,q,?]
- At the following prompt, either press Return to accept the default disk name or enter a disk name:
Use a default disk name for the disk? [y,n,q,?] (default: y)
- To continue with the operation, enter y (or press Return) at the following prompt:
The selected disks will be encapsulated and added to the disk group name disk group with default disk names. device name Continue with operation? [y,n,q,?] (default: y) y
- To confirm that encapsulation should proceed, enter y (or press Return) at the following prompt:
The following disk has been selected for encapsulation. Output format: [Device] device name Continue with encapsulation? [y,n,q,?] (default: y) y
A message similar to the following confirms that the disk is being encapsulated for use in VxVM:
The disk device device name will be encapsulated and added to the disk group diskgroup with the disk name diskgroup01.
- For non-root disks, you can now choose whether the disk is to be formatted as a CDS disk that is portable between different operating systems, or as a non-portable sliced disk:
Enter the desired format [cdsdisk,sliced,simple,q,?] (default: cdsdisk)
Enter the format that is appropriate for your needs. In most cases, this is the default format, cdsdisk. Note that only the sliced format is suitable for use with root, boot or swap disks.
- At the following prompt, vxdiskadm asks if you want to use the default private region size of 65536 blocks (32MB). Press Return to confirm that you want to use the default value, or enter a different value. (The maximum value that you can specify is 524288 blocks.)
Enter desired private region length [<privlen>,q,?] (default: 65536)
- If you entered cdsdisk as the format in step 8, you are prompted for the action to be taken if the disk cannot be converted this format:
Do you want to use sliced as the format should cdsdisk fail? [y,n,q,?] (default: y)
If you enter y, and it is not possible to encapsulate the disk as a CDS disk, it is encapsulated as a sliced disk. Otherwise, the encapsulation fails.
- vxdiskadm then proceeds to encapsulate the disks. You should now reboot your system at the earliest possible opportunity, for example by running this command:
# shutdown -r now
The
/etc/fstab
file is updated to include the volume devices that are used to mount any encapsulated file systems. You may need to update any other references in backup scripts, databases, or manually created swap devices. The original/etc/fstab
file is saved as/etc/fstab.b4vxvm
- At the following prompt, indicate whether you want to encapsulate more disks (y) or return to the vxdiskadm main menu (n):
Encapsulate other disks? [y,n,q,?] (default: n) n
The default layout that is used to encapsulate disks can be changed.