oracle 去摸_oracle 11g RAC STEP 2: Install grid

login as: root

Access denied

root@10.101.5.70's password:

Last login: Tue May 21 12:17:46 2013 from 10.101.5.66

[root@node1 ~]# /u01/app/oraInventory/orainstRoot.sh

Changing permissions of /u01/app/oraInventory.

Adding read,write permissions for group.

Removing read,write,execute permissions for world.

Changing groupname of /u01/app/oraInventory to oinstall.

The execution of the script is complete.

[root@node1 ~]# /u01/app/grid/11.2.3/root.sh

Performing root user operation for Oracle 11g

The following environment variables are set as:

ORACLE_OWNER= oracle

ORACLE_HOME=  /u01/app/grid/11.2.3

Enter the full pathname of the local bin directory: [/usr/local/bin]:

Copying dbhome to /usr/local/bin ...

Copying oraenv to /usr/local/bin ...

Copying coraenv to /usr/local/bin ...

Creating /etc/oratab file...

Entries will be added to the /etc/oratab file as needed by

Database Configuration Assistant when a database is created

Finished running generic part of root script.

Now product-specific root actions will be performed.

Using configuration parameter file: /u01/app/grid/11.2.3/crs/install/crsconfig_params

Creating trace directory

Failed to create keys in the OLR, rc = 127, Message:

/u01/app/grid/11.2.3/bin/clscfg.bin: error while loading shared libraries: libcap.so.1: cannot open shared object file: No such file or directory

Failed to create keys in the OLR at /u01/app/grid/11.2.3/crs/install/crsconfig_lib.pm line 7497.

/u01/app/grid/11.2.3/perl/bin/perl -I/u01/app/grid/11.2.3/perl/lib -I/u01/app/grid/11.2.3/crs/install /u01/app/grid/11.2.3/crs/install/rootcrs.pl execution failed

[root@node1 ~]#

node2同样运行一次

login as: root

Access denied

root@10.101.5.71's password:

Last login: Tue May 21 13:04:22 2013 from 10.101.5.66

[root@node2 ~]# /u01/app/oraInventory/orainstRoot.sh

Changing permissions of /u01/app/oraInventory.

Adding read,write permissions for group.

Removing read,write,execute permissions for world.

Changing groupname of /u01/app/oraInventory to oinstall.

The execution of the script is complete.

[root@node2 ~]# /u01/app/grid/11.2.3/root.sh

Performing root user operation for Oracle 11g

The following environment variables are set as:

ORACLE_OWNER= oracle

ORACLE_HOME=  /u01/app/grid/11.2.3

Enter the full pathname of the local bin directory: [/usr/local/bin]:

Copying dbhome to /usr/local/bin ...

Copying oraenv to /usr/local/bin ...

Copying coraenv to /usr/local/bin ...

Creating /etc/oratab file...

Entries will be added to the /etc/oratab file as needed by

Database Configuration Assistant when a database is created

Finished running generic part of root script.

Now product-specific root actions will be performed.

Using configuration parameter file: /u01/app/grid/11.2.3/crs/install/crsconfig_params

Creating trace directory

Failed to create keys in the OLR, rc = 127, Message:

/u01/app/grid/11.2.3/bin/clscfg.bin: error while loading shared libraries: libcap.so.1: cannot open shared object file: No such file or directory

Failed to create keys in the OLR at /u01/app/grid/11.2.3/crs/install/crsconfig_lib.pm line 7497.

/u01/app/grid/11.2.3/perl/bin/perl -I/u01/app/grid/11.2.3/perl/lib -I/u01/app/grid/11.2.3/crs/install /u01/app/grid/11.2.3/crs/install/rootcrs.pl execution failed

[root@node2 ~]#

运行完以后点击ok继续,出现一个错误

[INS-41807] The installer has detected that Oracle Clusterware is not running on the following nodes:  node1 node2.

Are you sure you want to continue ?

Cause - Either there was an error in starting the Oracle Clusterware stack on the specified nodes, or the root scripts on the specified nodes were not run.  Action - Run the root scripts on the nodes listed. If root scripts have already been run on these nodes, then examine the log file /u01/app/grid/11.2.3/cfgtoollogs/crsconfig/rootcrs_.log on each failed node to determine the reason for the Oracle Clusterware stack not starting  Exception Details:

PRCI-1108 : Failed to check CRS running state for CRS home /u01/app/grid/11.2.3 on node node1

PRCT-1003 : Failed to run "crsctl" on node "node1"

PRCI-1108 : Failed to check CRS running state for CRS home /u01/app/grid/11.2.3 on node node2

PRCT-1003 : Failed to run "crsctl" on node "node2"

检查日志

more /u01/app/grid/11.2.3/cfgtoollogs/crsconfig/rootcrs_node1.log

所有日志都有相同的输出:

2013-05-21 13:51:42: Executing cmd: /u01/app/grid/11.2.3/bin/clscfg -localadd

2013-05-21 13:51:42: Command output:

>  /u01/app/grid/11.2.3/bin/clscfg.bin: error while loading shared libraries: libcap.so.1: cannot open shared object file: No such file or directory

>End Command output

2013-05-21 13:51:42: '/u01/app/grid/11.2.3/bin/clscfg -localadd' - successful

2013-05-21 13:51:42: Failed to create keys in the OLR, rc = 127, Message:

/u01/app/grid/11.2.3/bin/clscfg.bin: error while loading shared libraries: libcap.so.1: cannot open shared object file: No such file or directory

2013-05-21 13:51:42: Running as user oracle: /u01/app/grid/11.2.3/bin/cluutil -ckpt -oraclebase /u01/app/oracle -writeckpt -name ROOTCRS_OLR -state FAIL

2013-05-21 13:51:42: s_run_as_user2: Running /bin/su oracle -c ' /u01/app/grid/11.2.3/bin/cluutil -ckpt -oraclebase /u01/app/oracle -writeckpt -name ROOTCRS_OLR -state FAIL '

2013-05-21 13:51:42: Removing file /tmp/filejzyWKa

2013-05-21 13:51:42: Successfully removed file: /tmp/filejzyWKa

2013-05-21 13:51:42: /bin/su successfully executed

2013-05-21 13:51:42: Succeeded in writing the checkpoint:'ROOTCRS_OLR' with status:FAIL

2013-05-21 13:51:42: CkptFile: /u01/app/oracle/Clusterware/ckptGridHA_node1.xml

2013-05-21 13:51:42: Sync the checkpoint file '/u01/app/oracle/Clusterware/ckptGridHA_node1.xml'

2013-05-21 13:51:42: Sync '/u01/app/oracle/Clusterware/ckptGridHA_node1.xml' to the physical disk

2013-05-21 13:51:42: ###### Begin DIE Stack Trace ######

2013-05-21 13:51:42:     Package         File                 Line Calling

2013-05-21 13:51:42:     --------------- -------------------- ---- ----------

2013-05-21 13:51:42:  1: main            rootcrs.pl            375 crsconfig_lib::dietrap

2013-05-21 13:51:42:  2: crsconfig_lib   crsconfig_lib.pm     7497 main::__ANON__

2013-05-21 13:51:42:  3: crsconfig_lib   crsconfig_lib.pm     7397 crsconfig_lib::olr_initial_config

2013-05-21 13:51:42:  4: main            rootcrs.pl            674 crsconfig_lib::perform_olr_config

2013-05-21 13:51:42: ####### End DIE Stack Trace #######

2013-05-21 13:51:42: 'ROOTCRS_OLR' checkpoint has failed

2013-05-21 13:51:42: Running as user oracle: /u01/app/grid/11.2.3/bin/cluutil -ckpt -oraclebase /u01/app/oracle -writeckpt -name ROOTCRS_OLR -state FAIL

2013-05-21 13:51:42: s_run_as_user2: Running /bin/su oracle -c ' /u01/app/grid/11.2.3/bin/cluutil -ckpt -oraclebase /u01/app/oracle -writeckpt -name ROOTCRS_OLR -state FAIL '

2013-05-21 13:51:43: Removing file /tmp/fileqQjNUI

2013-05-21 13:51:43: Successfully removed file: /tmp/fileqQjNUI

2013-05-21 13:51:43: /bin/su successfully executed

2013-05-21 13:51:43: Succeeded in writing the checkpoint:'ROOTCRS_OLR' with status:FAIL

2013-05-21 13:51:43: CkptFile: /u01/app/oracle/Clusterware/ckptGridHA_node1.xml

2013-05-21 13:51:43: Sync the checkpoint file '/u01/app/oracle/Clusterware/ckptGridHA_node1.xml'

2013-05-21 13:51:43: Sync '/u01/app/oracle/Clusterware/ckptGridHA_node1.xml' to the physical disk

[root@node1 ~]#

[root@node1 ~]# rpm -qa|grep libcap

libcap-ng-0.6.4-3.el6_0.1.i686

libcap-2.16-5.5.el6.i686

[root@node1 ~]# rpm -ql libcap|grep libcap.so.1

[root@node1 ~]#

[root@node1 ~]# yum provides libcap.so.1

Loaded plugins: security

compat-libcap1-1.10-1.i686 : Library for getting and setting POSIX.1e capabilities

Repo        : ol6_latest

Matched from:

Other       : libcap.so.1

[root@node1 ~]# rpm -qa|grep compat-libcap

[root@node1 ~]#

缺包,装上再说,所有的节点都装上

[root@node1 ~]# yum install compat-libcap1

Loaded plugins: security

Setting up Install Process

Resolving Dependencies

--> Running transaction check

---> Package compat-libcap1.i686 0:1.10-1 will be installed

--> Finished Dependency Resolution

Dependencies Resolved

=================================================================================================================================================================================

Package                                         Arch                                  Version                                 Repository                                   Size

=================================================================================================================================================================================

Installing:

compat-libcap1                                  i686                                  1.10-1                                  ol6_latest                                   16 k

Transaction Summary

=================================================================================================================================================================================

Install       1 Package(s)

Total download size: 16 k

Installed size: 24 k

Is this ok [y/N]: y

Downloading Packages:

compat-libcap1-1.10-1.i686.rpm                                                                                                                            |  16 kB     00:00

Running rpm_check_debug

Running Transaction Test

Transaction Test Succeeded

Running Transaction

Warning: RPMDB altered outside of yum.

Installing : compat-libcap1-1.10-1.i686                                                                                                                                    1/1

Verifying  : compat-libcap1-1.10-1.i686                                                                                                                                    1/1

Installed:

compat-libcap1.i686 0:1.10-1

Complete!

[root@node1 ~]#

先点No然后点击retry,重新执行root脚本

[root@node1 ~]# /u01/app/oraInventory/orainstRoot.sh

Changing permissions of /u01/app/oraInventory.

Adding read,write permissions for group.

Removing read,write,execute permissions for world.

Changing groupname of /u01/app/oraInventory to oinstall.

The execution of the script is complete.

[root@node1 ~]# /u01/app/grid/11.2.3/root.sh

Performing root user operation for Oracle 11g

The following environment variables are set as:

ORACLE_OWNER= oracle

ORACLE_HOME=  /u01/app/grid/11.2.3

Enter the full pathname of the local bin directory: [/usr/local/bin]:

The contents of "dbhome" have not changed. No need to overwrite.

The contents of "oraenv" have not changed. No need to overwrite.

The contents of "coraenv" have not changed. No need to overwrite.

Entries will be added to the /etc/oratab file as needed by

Database Configuration Assistant when a database is created

Finished running generic part of root script.

Now product-specific root actions will be performed.

Using configuration parameter file: /u01/app/grid/11.2.3/crs/install/crsconfig_params

OLR initialization - successful

root wallet

root wallet cert

root cert export

peer wallet

profile reader wallet

pa wallet

peer wallet keys

pa wallet keys

peer cert request

pa cert request

peer cert

pa cert

peer root cert TP

profile reader root cert TP

pa root cert TP

peer pa cert TP

pa peer cert TP

profile reader pa cert TP

profile reader peer cert TP

peer user cert

pa user cert

Adding Clusterware entries to upstart

CRS-2672: Attempting to start 'ora.mdnsd' on 'node1'

CRS-2676: Start of 'ora.mdnsd' on 'node1' succeeded

CRS-2672: Attempting to start 'ora.gpnpd' on 'node1'

CRS-2676: Start of 'ora.gpnpd' on 'node1' succeeded

CRS-2672: Attempting to start 'ora.cssdmonitor' on 'node1'

CRS-2672: Attempting to start 'ora.gipcd' on 'node1'

CRS-2676: Start of 'ora.cssdmonitor' on 'node1' succeeded

CRS-2676: Start of 'ora.gipcd' on 'node1' succeeded

CRS-2672: Attempting to start 'ora.cssd' on 'node1'

CRS-2672: Attempting to start 'ora.diskmon' on 'node1'

CRS-2676: Start of 'ora.diskmon' on 'node1' succeeded

CRS-2676: Start of 'ora.cssd' on 'node1' succeeded

ASM created and started successfully.

Disk Group OCR created successfully.

clscfg: -install mode specified

Successfully accumulated necessary OCR keys.

Creating OCR keys for user 'root', privgrp 'root'..

Operation successful.

CRS-4256: Updating the profile

Successful addition of voting disk 4433c865f4da4f25bf172a60fcfac49c.

Successful addition of voting disk 88ac6a4f38124fe2bf7f7bfc31b43e5c.

Successful addition of voting disk 14536de7c7a84f7fbf39508ef52e5e6f.

Successfully replaced voting disk group with +OCR.

CRS-4256: Updating the profile

CRS-4266: Voting file(s) successfully replaced

##  STATE    File Universal Id                File Name Disk group

--  -----    -----------------                --------- ---------

1. ONLINE   4433c865f4da4f25bf172a60fcfac49c (/dev/asm-diskf) [OCR]

2. ONLINE   88ac6a4f38124fe2bf7f7bfc31b43e5c (/dev/asm-diski) [OCR]

3. ONLINE   14536de7c7a84f7fbf39508ef52e5e6f (/dev/asm-diskj) [OCR]

Located 3 voting disk(s).

CRS-2672: Attempting to start 'ora.asm' on 'node1'

CRS-2676: Start of 'ora.asm' on 'node1' succeeded

CRS-2672: Attempting to start 'ora.OCR.dg' on 'node1'

CRS-2676: Start of 'ora.OCR.dg' on 'node1' succeeded

Configure Oracle Grid Infrastructure for a Cluster ... succeeded

[root@node1 ~]#

[root@node2 ~]# /u01/app/oraInventory/orainstRoot.sh

Changing permissions of /u01/app/oraInventory.

Adding read,write permissions for group.

Removing read,write,execute permissions for world.

Changing groupname of /u01/app/oraInventory to oinstall.

The execution of the script is complete.

[root@node2 ~]#

[root@node2 ~]# /u01/app/grid/11.2.3/root.sh

Performing root user operation for Oracle 11g

The following environment variables are set as:

ORACLE_OWNER= oracle

ORACLE_HOME=  /u01/app/grid/11.2.3

Enter the full pathname of the local bin directory: [/usr/local/bin]:

The contents of "dbhome" have not changed. No need to overwrite.

The contents of "oraenv" have not changed. No need to overwrite.

The contents of "coraenv" have not changed. No need to overwrite.

Entries will be added to the /etc/oratab file as needed by

Database Configuration Assistant when a database is created

Finished running generic part of root script.

Now product-specific root actions will be performed.

Using configuration parameter file: /u01/app/grid/11.2.3/crs/install/crsconfig_params

OLR initialization - successful

Adding Clusterware entries to upstart

CRS-4402: The CSS daemon was started in exclusive mode but found an active CSS daemon on node node1, number 1, and is terminating

An active cluster was found during exclusive startup, restarting to join the cluster

Configure Oracle Grid Infrastructure for a Cluster ... succeeded

[root@node2 ~]#

到100%左右又出一错误

点击details看看

Cause - The plug-in failed in its perform method

Action - Refer to the logs or contact Oracle Support Services.

Log File Location

/u01/app/oraInventory/logs/installActions2013-05-21_12-56-33PM.log

[root@node1 ~]# more /u01/app/oraInventory/logs/installActions2013-05-21_12-56-33PM.log

最后几行:

INFO: Query of CTSS for time offset passed

INFO: Check CTSS state started...

INFO: CTSS is in Observer state. Switching over to clock synchronization checks using NTP

INFO: Starting Clock synchronization checks using Network Time Protocol(NTP)...

INFO: NTP Configuration file check started...

INFO: NTP Configuration file check passed

INFO: Checking daemon liveness...

INFO: Liveness check failed for "ntpd"

INFO: Check failed on nodes:

INFO:   node2,node1

INFO: PRVF-5494 : The NTP Daemon or Service was not alive on all nodes

INFO: PRVF-5415 : Check to see if NTP daemon or service is running failed

INFO: Clock synchronization check using Network Time Protocol(NTP) failed

INFO: PRVF-9652 : Cluster Time Synchronization Services check failed

INFO: Checking VIP configuration.

INFO: Checking VIP Subnet configuration.

INFO: Check for VIP Subnet configuration passed.

INFO: Checking VIP reachability

INFO: Check for VIP reachability passed.

INFO: Post-check for cluster services setup was unsuccessful on all the nodes.

INFO:

WARNING:

INFO: Completed Plugin named: Oracle Cluster Verification Utility

[root@node1 ~]#

忘了禁用系统ntp服务

[root@node1 ~]# chkconfig --list|grep ntp

ntpd            0:off   1:off   2:off   3:off   4:off   5:off   6:off

ntpdate         0:off   1:off   2:off   3:off   4:off   5:off   6:off

[root@node1 ~]# ll /etc/|grep ntp

drwxr-xr-x.  3 root   root       4096 May 17 16:03 ntp

-rw-r--r--.  1 root   root       1917 Feb 24 06:32 ntp.conf

[root@node1 ~]#

[root@node1 ~]# mv /etc/ntp.conf /etc/ntp.conf.orig

[root@node1 ~]#

又有错误

检查日志

INFO: Checking existence of ONS node application (optional)

INFO: ONS node application check passed

INFO: Checking Single Client Access Name (SCAN)...

INFO: Checking TCP connectivity to SCAN Listeners...

INFO: TCP connectivity to SCAN Listeners exists on all cluster nodes

INFO: Checking name resolution setup for "rac-scan1.momo.org"...

INFO: ERROR:

INFO: PRVF-4657 : Name resolution setup check for "rac-scan1.momo.org" (IP address: 208.87.35.103) failed

INFO: ERROR:

INFO: PRVF-4657 : Name resolution setup check for "rac-scan1.momo.org" (IP address: 10.101.5.77) failed

INFO: ERROR:

INFO: PRVF-4664 : Found inconsistent name resolution entries for SCAN name "rac-scan1.momo.org"

INFO: Verification of SCAN VIP and Listener setup failed

INFO: Checking OLR integrity...

INFO: Checking OLR config file...

INFO: OLR config file check successful

INFO: Checking OLR file attributes...

INFO: OLR file check successful

INFO: WARNING:

INFO: This check does not verify the integrity of the OLR contents. Execute 'ocrcheck -local' as a privileged user to verify the contents of OLR.

INFO: OLR integrity check passed

INFO: User "oracle" is not part of "root" group. Check passed

INFO: Checking if Clusterware is installed on all nodes...

INFO: Check of Clusterware install passed

INFO: Checking if CTSS Resource is running on all nodes...

INFO: CTSS resource check passed

INFO: Querying CTSS for time offset on all nodes...

INFO: Query of CTSS for time offset passed

INFO: Check CTSS state started...

INFO: CTSS is in Active state. Proceeding with check of clock time offsets on all nodes...

INFO: PRVF-9661 :  Time offset is greater than acceptable limit on node "node2" [actual = "-152000.0", acceptable = "1000.0" ]

INFO: PRVF-9652 : Cluster Time Synchronization Services check failed

INFO: Checking VIP configuration.

INFO: Checking VIP Subnet configuration.

INFO: Check for VIP Subnet configuration passed.

INFO: Checking VIP reachability

INFO: Check for VIP reachability passed.

INFO: Post-check for cluster services setup was unsuccessful on all the nodes.

INFO:

WARNING:

INFO: Completed Plugin named: Oracle Cluster Verification Utility

[root@node1 ~]#

临时禁用一下DNS

[root@node1 ~]# nslookup

> rac-scan1.momo.org

Server:         10.101.0.16

Address:        10.101.0.16#53

Non-authoritative answer:

Name:   rac-scan1.momo.org

Address: 208.87.35.103

> ^C[root@node1 ~]#

[root@node1 ~]#

[root@node1 ~]# vi /etc/resolv.conf

[root@node1 ~]#

[root@node1 ~]# cd /etc/

[root@node1 etc]#

[root@node1 etc]# mv resolv.conf resolv.conf.orig

[root@node1 etc]#

[root@node1 etc]#

[root@node1 etc]# nslookup

> rac-scan1.momo.org

^C

[root@node1 etc]#

node2也一样

retry还是报错

INFO: GSD node application is offline on nodes "node2,node1"

INFO: Checking existence of ONS node application (optional)

INFO: ONS node application check passed

INFO: Checking Single Client Access Name (SCAN)...

INFO: Checking TCP connectivity to SCAN Listeners...

INFO: TCP connectivity to SCAN Listeners exists on all cluster nodes

INFO: Checking name resolution setup for "rac-scan1.momo.org"...

INFO: ERROR:

INFO: PRVG-1101 : SCAN name "rac-scan1.momo.org" failed to resolve

INFO: ERROR:

INFO: PRVF-4657 : Name resolution setup check for "rac-scan1.momo.org" (IP address: 10.101.5.77) failed

INFO: ERROR:

INFO: PRVF-4664 : Found inconsistent name resolution entries for SCAN name "rac-scan1.momo.org"

INFO: Verification of SCAN VIP and Listener setup failed

INFO: Checking OLR integrity...

INFO: Checking OLR config file...

INFO: OLR config file check successful

看来想通过必须得在DNS上配置域名解析,skip

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值