Exdata cell 节点配置时遇到的一个问题



问题描述:

[celladmin@vrh4 ~]$ cellcli
CellCLI: Release 11.2.3.2.0 - Production on Sat Jun 14 09:11:08 EDT 2014

Copyright (c) 2007, 2012, Oracle.  All rights reserved.
Cell Efficiency Ratio: 1

CellCLI> create celldisk all

CELL-02559: There is a communication error between MS and CELLSRV.

CellCLI> alter cell restart services all

Stopping the RS, CELLSRV, and MS services...
The SHUTDOWN of services was successful.
Starting the RS, CELLSRV, and MS services...
Getting the state of RS services...  running
Starting CELLSRV services...
The STARTUP of CELLSRV services was not successful.
CELL-01547: CELLSRV startup failed due to unknown reasons.

Starting MS services...
The STARTUP of MS services was successful.

CellCLI>


rs,ms 服务起来了,但cellsrv 服务都起不来


问题处理:

alert.log:


CELLSRV process id=3403
CELLSRV cell host name=vrh4.oracle.com
CELLSRV version=11.2.3.2.0,label=OSS_11.2.3.2.0_LINUX.X64_120713,Fri_Jul_13_12:37:13_PDT_2012
OS Hugepage status:
   Total/free hugepages available=32/32; hugepage size=2048KB
OS Stats: Physical memory: 497 MB. Num cores: 1
CELLSRV configuration parameters:
version=0.0
Cellsrv max memory not set. Total physical mem: 497 MB is less than required minimum: 3891 MB.
celldisk policy config read from /opt/oracle/cell11.2.3.2.0_LINUX.X64_120713/cellsrv/deploy/config/cdpolicy.dat with ver

no. 1 and pol no. 0
Auto Online Feature 1.3
CellServer MD5 Binary Checksum: cf96327cbbec459c6ac80deaec94d5cd
Sat Jun 14 09:12:00 2014
[RS] Started Service MS with pid 3258
OS Hugepage status:
   Total/free hugepages available=39/39; hugepage size=2048KB
WARNING: System has fewer hugepages available than needed.
Cache Allocation: Num 1MB hugepage buffers: 78 Num 1MB non-hugepage buffers: 822
MS_ALERT HUGEPAGE WARNING 78 822
ossmmap_map: mmap failed for Mmap memory len: 1624010752 errno: 12  --------------------mmap 无法映射内存
Physical memory on the system might be low.            ---------------------------这里报错信息很明确,物理内存不够啊
Sat Jun 14 09:12:05 2014
Errors in file /opt/oracle/cell11.2.3.2.0_LINUX.X64_120713/log/diag/asm/cell/vrh4/trace/svtrc_3403_0.trc  (incident=65):
ORA-00600: internal error code, arguments: [Cache: map_failed], [], [], [], [], [], [], [], [], [], [], []
Incident details in:

/opt/oracle/cell11.2.3.2.0_LINUX.X64_120713/log/diag/asm/cell/vrh4/incident/incdir_65/svtrc_3403_0_i65.trc
Sweep [inc][65]: completed
CELLSRV error - ORA-600 internal error
Sat Jun 14 09:12:16 2014
[RS] monitoring process /opt/oracle/cell11.2.3.2.0_LINUX.X64_120713/cellsrv/bin/cellrsomt (pid: 0) returned with error: 126
[RS] Monitoring process for service CELLSRV detected a flood of restarts. Disable monitoring process.
Errors in file /opt/oracle/cell11.2.3.2.0_LINUX.X64_120713/log/diag/asm/cell/vrh4/trace/rstrc_3248_4.trc  (incident=73):
RS-7445 [CELLSRV monitor disabled] [Detected a flood of restarts] [] [] [] [] [] [] [] [] [] []
Incident details in:

/opt/oracle/cell11.2.3.2.0_LINUX.X64_120713/log/diag/asm/cell/vrh4/incident/incdir_73/rstrc_3248_4_i73.trc
Sweep [inc][73]: completed


继续查看其它信息:

[root@vrh4 trace]# more /opt/oracle/cell11.2.3.2.0_LINUX.X64_120713/log/diag/asm/cell/vrh4/trace/svtrc_3403_0.trc
Trace file /opt/oracle/cell11.2.3.2.0_LINUX.X64_120713/log/diag/asm/cell/vrh4/trace/svtrc_3403_0.trc
ORACLE_HOME = /opt/oracle/cell11.2.3.2.0_LINUX.X64_120713
System name:    Linux
Node name:      vrh4.oracle.com
Release:        2.6.18-274.el5
Version:        #1 SMP Mon Jul 25 13:17:49 EDT 2011
Machine:        x86_64
CELL SW Version:        OSS_11.2.3.2.0_LINUX.X64_120713

*** 2014-06-14 09:11:53.184
CellDisk Policy configuration:
1 #version_ossp_cdperf_policy
0 #uniq_pol_num_ossp_cdperf_policy
2 #hang_hd_ossp_cdperf_policy
2 #hang_fd_ossp_cdperf_policy
2 #slow_abs_hd_ossp_cdperf_policy
2 #slow_abs_fd_ossp_cdperf_policy
2 #slow_rltv_hd_ossp_cdperf_policy
2 #slow_rltv_fd_ossp_cdperf_policy
2 #slow_lat_hd_ossp_cdperf_policy
2 #slow_lat_fd_ossp_cdperf_policy
0 #ioerr_hd_ossp_cdperf_policy
2 #ioerr_fd_ossp_cdperf_policy
0 #powercycle_hang_ossp_cdperf_policy
0 #powercycle_hang_wtfc_ossp_cdperf_policy
6 #lat_freq_ossp_cdperf_policy
50 #asm_offline_freq_ossp_cdperf_policy
30 #dmwg_avgrqsize_tolr_ossp_cdperf_policy
30 #dmwg_avgnumreads_tolr_ossp_cdperf_policy
30 #dmwg_avgnumwrites_tolr_ossp_cdperf_policy
100 #dmwg_avgrqsize_min_ossp_cdperf_policy
8 #dmwg_avgrqsizefl_min_ossp_cdperf_policy
10 #dmwg_avgnumreads_min_ossp_cdperf_policy
10 #dmwg_avgnumwrites_min_ossp_cdperf_policy
3 #dmwg_lownumreads_ossp_cdperf_policy
3 #dmwg_lownumwrites_ossp_cdperf_policy
30 #dmwg_lowlatreads_ossp_cdperf_policy
30 #dmwg_lowlatwrites_ossp_cdperf_policy
1 #dmwg_avgqdepreads_min_ossp_cdperf_policy
5 #dmwg_avgqdepreadsfl_min_ossp_cdperf_policy
1 #dmwg_avgqdepwrites_min_ossp_cdperf_policy
5 #dmwg_avgqdepwritesfl_min_ossp_cdperf_policy
100 #dmwg_avgqdepreads_tolr_ossp_cdperf_policy
100 #dmwg_avgqdepwrites_tolr_ossp_cdperf_policy
100 #dmwg_avgqszreads_tolr_ossp_cdperf_policy
100 #dmwg_avgqszwrites_tolr_ossp_cdperf_policy
60 #dmwg_same_pct_ossp_cdperf_policy
3 #conf_hd_max_num_ossp_cdperf_policy
8 #conf_fd_max_num_ossp_cdperf_policy
3 #proa_fail_hd_max_num_ossp_cdperf_policy
8 #proa_fail_fd_max_num_ossp_cdperf_policy
2 #hung_hd_max_num_reboot_ossp_cdperf_policy
9 #hung_fd_max_num_reboot_ossp_cdperf_policy
3 #numtriggers_thld_5hrs_ossp_cdperf_policy
4 #numtriggers_thld_day_ossp_cdperf_policy
5 #numtriggers_thld_week_ossp_cdperf_policy
7 #numtriggers_thld_month_ossp_cdperf_policy
8 #numtriggers_thld_quart_ossp_cdperf_policy
6 #ioerr_numthld_near_ossp_cdperf_policy
10 #ioerr_numnzero_near_ossp_cdperf_policy
20 #ioerr_numthld_far_ossp_cdperf_policy
50 #ioerr_numnzero_far_ossp_cdperf_policy
50 #err_lat_timeout_ossp_cdperf_policy
6 #err_lat_numthld_near_ossp_cdperf_policy
10 #err_lat_numnzero_near_ossp_cdperf_policy
20 #err_lat_numthld_far_ossp_cdperf_policy
50 #err_lat_numnzero_far_ossp_cdperf_policy
90000 95000 100 6 10 20 50 10000 300 200 7 10 30 50 20000 500 200 500 200 14 20 14 20 24 40 24 40

#dmg_params_ossp_cdperf_policy[0]
90000 95000 200 6 10 20 50 30000 300 200 7 10 30 50 60000 500 200 500 200 14 20 14 20 24 40 24 40

#dmg_params_ossp_cdperf_policy[1]
90000 95000 150 6 10 20 50 24000 300 200 7 10 30 50 48000 500 200 500 200 14 20 14 20 24 40 24 40

#dmg_params_ossp_cdperf_policy[2]
90000 95000 100 6 10 20 50 15000 300 200 7 10 30 50 30000 500 200 500 200 14 20 14 10 24 40 24 40

#dmg_params_ossp_cdperf_policy[3]
90000 95000 100 6 10 20 50 6000 300 200 7 10 30 50 12000 500 200 500 200 14 20 14 10 24 40 24 40

#dmg_params_ossp_cdperf_policy[4]
90000 95000 200 6 10 20 50 15000 300 200 25 40 30 50 20000 2000 1500 2000 1500 20 30 20 30 25 40 25 40

#dmg_params_ossp_cdperf_policy[5]
90000 95000 300 6 10 20 50 40000 300 200 25 40 30 50 80000 2000 1500 2000 1500 20 30 20 30 25 40 25 40

#dmg_params_ossp_cdperf_policy[6]
90000 95000 250 6 10 20 50 30000 300 200 25 40 30 50 60000 2000 1500 2000 1500 20 30 20 30 25 40 25 40

#dmg_params_ossp_cdperf_policy[7]
90000 95000 200 6 10 20 50 25000 300 200 25 40 30 50 40000 2000 1500 2000 1500 20 30 20 30 25 40 25 40

#dmg_params_ossp_cdperf_policy[8]
90000 95000 200 6 10 20 50 10000 300 200 25 40 30 50 20000 2000 1500 2000 1500 20 30 20 30 25 40 25 40

#dmg_params_ossp_cdperf_policy[9]
90000 95000 50 6 10 20 50 2000 300 200 20 30 30 50 4000 500 200 500 200 14 20 14 20 24 40 24 40

#dmg_params_ossp_cdperf_policy[10]
90000 95000 25 6 10 20 50 1000 300 200 7 10 30 50 2000 500 200 500 200 14 20 14 20 24 40 24 40

#dmg_params_ossp_cdperf_policy[11]
90000 95000 50 6 10 20 50 2000 300 200 7 10 30 50 4000 500 200 500 200 14 20 14 20 24 40 24 40

#dmg_params_ossp_cdperf_policy[12]
90000 95000 50 6 10 20 50 2000 300 200 7 10 30 50 4000 500 200 500 200 14 20 14 20 24 40 24 40

#dmg_params_ossp_cdperf_policy[13]
400000 410000 3000 6 10 20 50 50000 1000 800 7 10 30 50 100000 2000 2000 2000 2000 20 30 20 30 25 40 25 40

#dmg_params_ossp_cdperf_policy[14]
42346 #checksum_ossp_cdperf_policy
LockPool name:Storage Index Lock Pool type:RWLOCK POOL group:35 numLocks:1024 nextLockIndex:0 totalLockRefs:0

lockArray:0x2accba272660
2014-06-14 09:11:53.898190*: Opened file

/opt/oracle/cell11.2.3.2.0_LINUX.X64_120713/cellsrv/deploy/config/griddisk.owners.dat, version 11.2.2.4.0, descriptor 14
2014-06-14 09:12:01.801656*: CELLSRV needs 463 hugepages, but there are only 32 available. 2014-06-14 09:12:01.838968*:  ----------------------这里的报错已经非常明晰了

CELLSRV trying to reserve 431 more hugepages.
2014-06-14 09:12:02.021569*: Successfully allocated 78MB of hugepages for buffersWriting message type

OSS_PIPE_ERR_FAILED_STARTUP_RESTART to OSS->RS pipe
DDE: Flood control is not active
Incident 65 created, dump file:

/opt/oracle/cell11.2.3.2.0_LINUX.X64_120713/log/diag/asm/cell/vrh4/incident/incdir_65/svtrc_3403_0_i65.trc
ORA-00600: internal error code, arguments: [Cache: map_failed], [], [], [], [], [], [], [], [], [], [], []

2014-06-14 09:12:15.281868*: CELLSRV error - ORA-600 internal error


看来cell 节点要加大内存才能解决问题啊


Oracle Exadata数据库作为一个高性能的数据库存储平台,在数据处理速度和存储可靠性方面都有着优异的表现。但是在实际应用中,Exadata数据库也仍然需要进行一些优化工作,以达到更好的性能。 一、优化存储 Exadata对数据存储的优化主要是通过Exadata存储服务器来实现,这里有几个重要的优化点: 1. 存储单元配置Exadata存储单元具有多个闪存和机械硬盘,选择合适的闪存和机械硬盘使用模式,对于I/O密集型的数据库应用系统能够起到重要的优化作用。 2. 优化RAID设置: 通过RAID0(条带化)设置可以提高性能,同也增加了系统的故障风险。RAID10(1 + 0)和RAID5(分布式条带化和奇偶校验)的选择取决于是否需要更高的灵活性或更高的可靠性。 二、优化性能 Exadata的性能优化可以从以下几个方面进行: 1. SQL优化: Exadata数据库在处理复杂的SQL查询性能表现很好,但是在处理大量的简单SQL查询可能会出现瓶颈,需要对SQL查询进行优化。 2. I/O优化: I/O的性能是整个Exadata数据库的核心,通过对I/O进行优化可以有效地提升数据库的性能。 3. 内存优化: Exadata数据库具有大量内存资源,可以通过适当的内存调整来提高I/O效率。 三、优化管理 Exadata数据库管理可以从以下几个方面进行: 1. 系统监控: 监控系统运行情况以及数据库性能,以便及调整和处理。 2. 数据备份与恢复: 定期备份数据,以防数据丢失,同恢复数据,保证数据库的正常运行。 3. 安全管理:加强系统安全管理和防御,以保障数据不被非法获取或破坏,确保数据库的安全可靠。
评论 2
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值