access和tagware_oracle 11GR2 新特性

New Features for Release 2 (11.2)

The following is a list of new features for installation of Oracle Clusterware and Oracle

ASM 11g release 2 (11.2):

Automatic Storage Management and Oracle Clusterware Installation

With Oracle grid infrastructure 11g release 2 (11.2), Oracle Automatic Storage

Management (Oracle ASM) and Oracle Clusterware are installed into a single home

directory, which is referred to as the Grid Infrastructure home. Configuration

assistants start after the installer interview process that configures Oracle ASM and

Oracle Clusterware.

The installation of the combined products is called Oracle grid infrastructure.

However, Oracle Clusterware and Automatic Storage Management remain separate

products.

Automatic Storage Management and Oracle Clusterware Files

With this release, Oracle Cluster Registry (OCR) and voting disks can be placed on

Oracle Automatic Storage Management (Oracle ASM).

This feature enables Oracle ASM to provide a unified storage solution, storing all the

data for the clusterware and the database, without the need for third-party volume

managers or cluster filesystems.

For new installations, OCR and voting disk files can be placed either on Oracle ASM,

or on a cluster file system or NFS system. Installing Oracle Clusterware files on raw or

block devices is no longer supported, unless an existing system is being upgraded.

Oracle ASM Job Role Separation Option with SYSASM

The SYSASM privilege that was introduced in Oracle ASM 11g release 1 (11.1) is now

fully separated from the SYSDBA privilege. If you choose to use this optional feature,

and designate different operating system groups as the OSASM and the OSDBA

groups, then the SYSASM administrative privilege is available only to members of the

OSASM group. The SYSASM privilege also can be granted using password

authentication on the Oracle ASM instance.

You can designate OPERATOR privileges (a subset of the SYSASM privileges,

including starting and stopping ASM) to members of the OSOPER for ASM group.

Providing system privileges for the storage tier using the SYSASM privilege instead of

the SYSDBA privilege provides a clearer division of responsibility between Oracle

ASM administration and database administration, and helps to prevent different

databases using the same storage from accidentally overwriting each other's files.

Cluster Time Synchronization Service

Cluster node times should be synchronized. With this release, Oracle Clusterware

provides Cluster Time Synchronization Service (CTSS), which ensures that there is a

synchronization service in the cluster. If Network Time Protocol (NTP) is not found

during cluster configuration, then CTSS is configured to ensure time synchronization.

Enterprise Manager Database Control Provisioning

Enterprise Manager Database Control 11g provides the capability to automatically

provision Oracle grid infrastructure and Oracle RAC installations on new nodes, and

then extend the existing Oracle grid infrastructure and Oracle RAC database to these

provisioned nodes. This provisioning procedure requires a successful Oracle RAC

installation before you can use this feature.

Fixup Scripts and Grid Infrastructure Checks

With Oracle Clusterware 11g release 2 (11.2), the installer (OUI) detects when

minimum requirements for installation are not completed, and creates shell script

programs, called fixup scripts, to resolve many incomplete system configuration

requirements. If OUI detects an incomplete task that is marked "fixable", then you can

easily fix the issue by generating the fixup script. by clicking the Fix & Check Again

button.

The fixup script. is generated during installation. You are prompted to run the script. as

root in a separate terminal session. When you run the script, it raises kernel values to

required minimums, if necessary, and completes other operating system configuration

tasks.

You also can have Cluster Verification Utility (CVU) generate fixup scripts before

installation.

Grid Plug and Play

In the past, adding or removing servers in a cluster required extensive manual

preparation. With this release, you can continue to configure server nodes manually, or

use Grid Plug and Play to configure them dynamically as nodes are added or removed

from the cluster.

Grid Plug and Play reduces the costs of installing, configuring, and managing server

nodes by starting a grid naming service within the cluster to allow each node to

perform. the following tasks dynamically:

■ Negotiating appropriate network identities for itself

■ Acquiring additional information it needs to operate from a configuration profile

■ Configuring or reconfiguring itself using profile data, making hostnames and

addresses resolvable on the network

Because servers perform. these tasks dynamically, the number of steps required to add

or delete nodes is minimized.

Oracle Clusterware Out-of-place Upgrade

With this release, you can install a new version of Oracle Clusterware into a separate

home from an existing Oracle Clusterware installation. This feature reduces the

downtime required to upgrade a node in the cluster. When performing an out-of-place

upgrade, the old and new version of the software are present on the nodes at the same

time, each in a different home location, but only one version of the software is active.

Oracle Clusterware Administration with Oracle Enterprise Manager

With this release, you can use Enterprise Manager Cluster Home page to perform. full

administrative and monitoring support for both standalone database and Oracle RAC

environments, using High Availability Application and Oracle Cluster Resource

Management.

When Oracle Enterprise Manager is installed with Oracle Clusterware, it can provide a

set of users that have the Oracle Clusterware Administrator role in Enterprise

Manager, and provide full administrative and monitoring support for High

Availability application and Oracle Clusterware resource management. After you have

completed installation and have Enterprise Manager deployed, you can provision

additional nodes added to the cluster using Enterprise Manager.

SCAN for Simplified Client Access

With this release, the single client access name (SCAN) is the hostname to provide for

all clients connecting to the cluster. The SCAN is a domain name registered to at least

one and up to three IP addresses, either in the domain name service (DNS) or the Grid

Naming Service (GNS). The SCAN eliminates the need to change clients when nodes

are added to or removed from the cluster. Clients using the SCAN can also access the

cluster using EZCONNECT.

SRVCTL Command Enhancements for Patching

With this release, you can use srvctl to shut down all Oracle software running

within an Oracle home, in preparation for patching. Oracle grid infrastructure

patching is automated across all nodes, and patches can be applied in a multi-node,

multi-patch fashion.

Typical Installation Option

To streamline cluster installations, especially for those customers who are new to

clustering, Oracle introduces the Typical Installation path. Typical installation defaults

as many options as possible to those recommended as best practices.

Voting Disk Backup Procedure Change

In prior releases, backing up the voting disks using a dd command was a required

postinstallation task. With Oracle Clusterware release 11.2 and later, backing up and

restoring a voting disk using the dd command is not supported.

Backing up voting disks manually is no longer required, as voting disks are backed up

automatically in the OCR as part of any configuration change and voting disk data is

automatically restored to any added voting disks.

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值