Cluster Server 7.3.1 Configuration and Upgrade Guide - Solaris
- Section I. Configuring Cluster Server using the script-based installer
- I/O fencing requirements
- Preparing to configure VCS clusters for data integrity
- About planning to configure I/O fencing
- Setting up the CP server
- Configuring VCS
- Configuring a secure cluster node by node
- Verifying and updating licenses on the system
- Configuring VCS clusters for data integrity
- Setting up disk-based I/O fencing using installer
- Setting up server-based I/O fencing using installer
- Section II. Automated configuration using response files
- Performing an automated VCS configuration
- Performing an automated I/O fencing configuration using response files
- Section III. Manual configuration
- Manually configuring VCS
- Configuring LLT manually
- Configuring VCS manually
- Configuring VCS in single node mode
- Modifying the VCS configuration
- Manually configuring the clusters for data integrity
- Setting up disk-based I/O fencing manually
- Setting up server-based I/O fencing manually
- Configuring server-based fencing on the VCS cluster manually
- Setting up non-SCSI-3 fencing in virtual environments manually
- Setting up majority-based I/O fencing manually
- Manually configuring VCS
- Section IV. Upgrading VCS
- Planning to upgrade VCS
- Performing a VCS upgrade using the installer
- Tasks to perform after upgrading to 2048 bit key and SHA256 signature certificates
- Performing an online upgrade
- Performing a rolling upgrade of VCS
- Performing a phased upgrade of VCS
- About phased upgrade
- Performing a phased upgrade using the product installer
- Performing an automated VCS upgrade using response files
- Upgrading VCS using Live Upgrade and Boot Environment upgrade
- Planning to upgrade VCS
- Section V. Adding and removing cluster nodes
- Adding a node to a single-node cluster
- Adding a node to a single-node cluster
- Adding a node to a multi-node VCS cluster
- Manually adding a node to a cluster
- Setting up the node to run in secure mode
- Configuring I/O fencing on the new node
- Adding a node using response files
- Removing a node from a VCS cluster
- Removing a node from a VCS cluster
- Removing a node from a VCS cluster
- Adding a node to a single-node cluster
- Section VI. Installation reference
- Appendix A. Services and ports
- Appendix B. Configuration files
- Appendix C. Configuring LLT over UDP
- Using the UDP layer for LLT
- Manually configuring LLT over UDP using IPv4
- Manually configuring LLT over UDP using IPv6
- Appendix D. Configuring the secure shell or the remote shell for communications
- Appendix E. Installation script options
- Appendix F. Troubleshooting VCS configuration
- Appendix G. Sample VCS cluster setup diagrams for CP server-based I/O fencing
- Appendix H. Reconciling major/minor numbers for NFS shared disks
- Appendix I. Upgrading the Steward process
Response file variables to configure server-based I/O fencing
You can use a coordination point server-based fencing response file to configure server-based customized I/O fencing.
Table: Coordination point server (CP server) based fencing response file definitions lists the fields in the response file that are relevant for server-based customized I/O fencing.
Table: Coordination point server (CP server) based fencing response file definitions
Response file field | Definition |
---|---|
CFG {fencing_config_cpagent} | Enter '1' or '0' depending upon whether you want to configure the Coordination Point agent using the installer or not. Enter "0" if you do not want to configure the Coordination Point agent using the installer. Enter "1" if you want to use the installer to configure the Coordination Point agent. |
CFG {fencing_cpagentgrp} | Name of the service group which will have the Coordination Point agent resource as part of it. Note: This field is obsolete if the |
CFG {fencing_cps} | Virtual IP address or Virtual hostname of the CP servers. |
CFG {fencing_reusedg} | This response file field indicates whether to reuse an existing DG name for the fencing configuration in customized fencing (CP server and coordinator disks). Enter either a "1" or "0". Entering a "1" indicates reuse, and entering a "0" indicates do not reuse. When reusing an existing DG name for the mixed mode fencing configuration. you need to manually add a line of text , such as "$CFG{fencing_reusedg}=0" or "$CFG{fencing_reusedg}=1" before proceeding with a silent installation. |
CFG {fencing_dgname} | The name of the disk group to be used in the customized fencing, where at least one disk is being used. |
CFG {fencing_disks} | The disks being used as coordination points if any. |
CFG {fencing_ncp} | Total number of coordination points being used, including both CP servers and disks. |
CFG {fencing_ndisks} | The number of disks being used. |
CFG {fencing_cps_vips} | The virtual IP addresses or the fully qualified host names of the CP server. |
CFG {fencing_cps_ports} | The port that the virtual IP address or the fully qualified host name of the CP server listens on. |
CFG{fencing_option} | Specifies the I/O fencing configuration mode.
|
CFG{fencing_auto_refresh_reg} | Enable this variable if registration keys are missing on any of the CP servers. |