Storage Foundation for Sybase ASE CE 7.4.1 Configuration and Upgrade Guide - Linux
- Section I. Configuring SF Sybase ASE CE
- Preparing to configure SF Sybase CE
- Configuring SF Sybase CE
- About configuring SF Sybase CE
- Configuring the SF Sybase CE components using the script-based installer
- Configuring the SF Sybase CE cluster
- Configuring the cluster name
- Configuring private heartbeat links
- Configuring the virtual IP of the cluster
- Configuring SF Sybase CE in secure mode
- Configuring a secure cluster node by node
- Adding VCS users
- Configuring SMTP email notification
- Configuring SNMP trap notification
- Configuring global clusters
- Configuring the SF Sybase CE cluster
- Configuring SF Sybase CE clusters for data integrity
- Setting up disk-based I/O fencing using installer
- Performing an automated SF Sybase CE configuration
- Performing an automated I/O fencing configuration using response files
- Configuring a cluster under VCS control using a response file
- Section II. Post-installation and configuration tasks
- Section III. Upgrade of SF Sybase CE
- Planning to upgrade SF Sybase CE
- Performing a full upgrade of SF Sybase CE using the product installer
- Performing an automated full upgrade of SF Sybase CE using response files
- Performing a phased upgrade of SF Sybase CE
- About phased upgrade
- Performing a phased upgrade of SF Sybase CE from version 6.2.1 and later release
- Step 1: Performing pre-upgrade tasks on the first half of the cluster
- Step 2: Upgrading the first half of the cluster
- Step 3: Performing pre-upgrade tasks on the second half of the cluster
- Step 4: Performing post-upgrade tasks on the first half of the cluster
- Step 5: Upgrading the second half of the cluster
- Step 6: Performing post-upgrade tasks on the second half of the cluster
- Performing a rolling upgrade of SF Sybase CE
- Performing post-upgrade tasks
- Section IV. Installation and upgrade of Sybase ASE CE
- Installing, configuring, and upgrading Sybase ASE CE
- Before installing Sybase ASE CE
- Preparing for local mount point on VxFS for Sybase ASE CE binary installation
- Preparing for shared mount point on CFS for Sybase ASE CE binary installation
- Installing Sybase ASE CE software
- Preparing to create a Sybase ASE CE cluster
- Creating the Sybase ASE CE cluster
- Preparing to configure the Sybase instances under VCS control
- Configuring a Sybase ASE CE cluster under VCS control using the SF Sybase CE installer
- Upgrading Sybase ASE CE
- Installing, configuring, and upgrading Sybase ASE CE
- Section V. Adding and removing nodes
- Adding a node to SF Sybase CE clusters
- About adding a node to a cluster
- Before adding a node to a cluster
- Adding the node to a cluster manually
- Starting Veritas Volume Manager (VxVM) on the new node
- Configuring cluster processes on the new node
- Setting up the node to run in secure mode
- Starting fencing on the new node
- Configuring Cluster Volume Manager (CVM) and Cluster File System (CFS) on the new node
- After adding the new node
- Configuring the ClusterService group for the new node
- Adding a node to a cluster using the Veritas InfoScale installer
- Adding the new instance to the Sybase ASE CE cluster
- Removing a node from SF Sybase CE clusters
- Adding a node to SF Sybase CE clusters
- Section VI. Configuration of disaster recovery environments
- Section VII. Installation reference
- Appendix A. Installation scripts
- Appendix B. Sample installation and configuration values
- Appendix C. Tunable files for installation
- About setting tunable parameters using the installer or a response file
- Setting tunables for an installation, configuration, or upgrade
- Setting tunables with no other installer-related operations
- Setting tunables with an un-integrated response file
- Preparing the tunables file
- Setting parameters for the tunables file
- Tunables value parameter definitions
- Appendix D. Configuration files
- About sample main.cf files
- Sample main.cf files for Sybase ASE CE configurations
- Sample main.cf for a basic Sybase ASE CE cluster configuration under VCS control with shared mount point on CFS for Sybase binary installation
- Sample main.cf for a basic Sybase ASE CE cluster configuration with local mount point on VxFS for Sybase binary installation
- Sample main.cf for a primary CVM VVR site
- Sample main.cf for a secondary CVM VVR site
- Appendix E. Configuring the secure shell or the remote shell for communications
- Appendix F. High availability agent information
Verifying LLT
Use the lltstat command to verify that links are active for LLT. If LLT is configured correctly, this command shows all the nodes in the cluster. The command also returns information about the links for LLT for the node on which you typed the command.
Refer to the lltstat(1M) manual page for more information.
To verify LLT
- Log in as superuser on the node sys1.
- Run the lltstat command on the node sys1 to view the status of LLT.
lltstat -n
The output on sys1 resembles:
LLT node information: Node State Links *0 sys1 OPEN 2 1 sys2 OPEN 2Each node has two links and each node is in the OPEN state. The asterisk (*) denotes the node on which you typed the command.
If LLT does not operate, the command does not return any LLT links information: If only one network is connected, the command returns the following LLT statistics information:
LLT node information: Node State Links * 0 sys1 OPEN 2 1 sys2 OPEN 2 2 sys5 OPEN 1 - Log in as superuser on the node sys2.
- Run the lltstat command on the node sys2 to view the status of LLT.
lltstat -n
The output on sys2 resembles:
LLT node information: Node State Links 0 sys1 OPEN 2 *1 sys2 OPEN 2 - To view additional information about LLT, run the lltstat -nvv command on each node.
For example, run the following command on the node sys1 in a two-node cluster:
lltstat -nvv active
The output on sys1 resembles:
Node State Link Status Address *0 sys1 OPEN eth1 UP 08:00:20:93:0E:34 eth2 UP 08:00:20:93:0E:38 1 sys2 OPEN eth1 UP 08:00:20:8F:D1:F2 eth2 DOWNThe command reports the status on the two active nodes in the cluster, sys1 and sys2.
For each correctly configured node, the information must show the following:
A state of OPEN
A status for each link of UP
An address for each link
However, the output in the example shows different details for the node sys2. The private network connection is possibly broken or the information in the /etc/llttab file may be incorrect.
- To obtain information about the ports open for LLT, type lltstat -p on any node.
For example, type lltstat -p on the node sys1 in a two-node cluster:
lltstat -p
The output resembles:
LLT port information: Port Usage Cookie 0 gab 0x0 opens: 0 2 3 4 5 6 7 8 9 10 11 ... 60 61 62 63 connects: 0 1 7 gab 0x7 opens: 0 2 3 4 5 6 7 8 9 10 11 ... 60 61 62 63 connects: 0 1 31 gab 0x1F opens: 0 2 3 4 5 6 7 8 9 10 11 ... 60 61 62 63 connects: 0 1