• 习水医院12C RAC 数据库安装文档


     

     

    环境介绍

    OS: Oracle Enterprise Linux 6.4 (For RAC Nodes)

    DB: GI and Database 12.1.0.2

    所需介质

    p17694377_121020_Linux-x86-64_1of8.zip

    p17694377_121020_Linux-x86-64_2of8.zip

    p17694377_121020_Linux-x86-64_3of8.zip

    p17694377_121020_Linux-x86-64_4of8.zip

    上述软件1-2介质是DB软件,3-4是GRID软件。

    – 这里只给出Oracle相关的,操作系统以及其他软件请自身准备。

    操作系统信息

    RAC节点服务器:

    (以node1节点为例)

    [root@xsyy1 ~]# cat /etc/redhat-release

    Red Hat Enterprise Linux Server release 6.4 (Santiago)

    [root@xsyy1 ~]# uname -a

    Linux xsyy1 2.6.39-400.17.1.el6uek.x86_64 #1 SMP Fri Feb 22 18:16:18 PST 2013 x86_64 x86_64 x86_64 GNU/Linux

    [root@xsyy1 ~]# grep MemTotal /proc/meminfo

    MemTotal:       65952032 kB

    [root@xsyy1 ~]# grep SwapTotal /proc/meminfo

    SwapTotal:      131070972 kB

    [root@xsyy1 ~]# df -h

    文件系统           容量  已用  可用 已用%% 挂载点

    /dev/sdd2             1.7T   39G1.6T   3% /

    tmpfs                  32G  624M31G   2% /dev/shm

    /dev/sdd1             9.7G195M  9.0G   3% /boot

    网络配置信息:

    备注: 每个节点服务器至少需要两个网卡,eth0用于PUBLIC,而eth1用于Private

    (节点1)

    [root@xsyy1 ~]# ifconfig

    eth0      Link encap:Ethernet  HWaddr 6C:92:BF:2A:7F:8A 

              inet addr:172.16.28.24Bcast:172.16.28.255Mask:255.255.255.0

              inet6 addr: fe80::6e92:bfff:fe2a:7f8a/64 Scope:Link

              UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1

              RX packets:23832511 errors:0 dropped:93 overruns:0 frame:0

              TX packets:24962108 errors:0 dropped:0 overruns:0 carrier:0

              collisions:0 txqueuelen:1000

              RX bytes:25937636908 (24.1 GiB)  TX bytes:29336937458 (27.3 GiB)

    eth0:1    Link encap:Ethernet  HWaddr 6C:92:BF:2A:7F:8A 

              inet addr:172.16.28.52Bcast:172.16.28.255Mask:255.255.255.0

              UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1

    eth0:2    Link encap:Ethernet  HWaddr 6C:92:BF:2A:7F:8A 

              inet addr:172.16.28.39Bcast:172.16.28.255Mask:255.255.255.0

              UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1

    eth1      Link encap:Ethernet  HWaddr 6C:92:BF:2A:7F:8B 

              inet addr:10.1.20.24  Bcast:10.1.20.255  Mask:255.255.255.0

              inet6 addr: fe80::6e92:bfff:fe2a:7f8b/64 Scope:Link

              UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1

              RX packets:15029998 errors:0 dropped:0 overruns:0 frame:0

              TX packets:12761273 errors:0 dropped:0 overruns:0 carrier:0

              collisions:0 txqueuelen:1000

              RX bytes:12231984673 (11.3 GiB)  TX bytes:8970357010 (8.3 GiB)

    eth1:1    Link encap:Ethernet  HWaddr 6C:92:BF:2A:7F:8B 

              inet addr:169.254.169.99Bcast:169.254.255.255Mask:255.255.0.0

              UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1

    lo        Link encap:Local Loopback 

              inet addr:127.0.0.1  Mask:255.0.0.0

              inet6 addr: ::1/128 Scope:Host

              UP LOOPBACK RUNNING  MTU:16436  Metric:1

              RX packets:6453002 errors:0 dropped:0 overruns:0 frame:0

              TX packets:6453002 errors:0 dropped:0 overruns:0 carrier:0

              collisions:0 txqueuelen:0

              RX bytes:6450405692 (6.0 GiB)  TX bytes:6450405692 (6.0 GiB)

    (节点2)

    [root@xsyy2 ~]# ifconfig

    eth0      Link encap:Ethernet  HWaddr 6C:92:BF:2A:7C:D6 

              inet addr:172.16.28.28  Bcast:172.16.28.255  Mask:255.255.255.0

              inet6 addr: fe80::6e92:bfff:fe2a:7cd6/64 Scope:Link

              UP BROADCAST RUNNING MULTICAST  MTU:1500Metric:1

              RX packets:12895198 errors:0 dropped:93 overruns:0 frame:0

              TX packets:2639578 errors:0 dropped:0 overruns:0 carrier:0

              collisions:0 txqueuelen:1000

              RX bytes:17559461040 (16.3 GiB)  TX bytes:231832259 (221.0 MiB)

    eth0:1    Link encap:Ethernet  HWaddr 6C:92:BF:2A:7C:D6 

              inet addr:172.16.28.53  Bcast:172.16.28.255  Mask:255.255.255.0

              UP BROADCAST RUNNING MULTICAST  MTU:1500Metric:1

    eth1      Link encap:Ethernet  HWaddr 6C:92:BF:2A:7C:D7 

              inet addr:10.1.20.28  Bcast:10.1.20.255  Mask:255.255.255.0

              inet6 addr: fe80::6e92:bfff:fe2a:7cd7/64 Scope:Link

              UP BROADCAST RUNNING MULTICAST  MTU:1500Metric:1

              RX packets:12764698 errors:0 dropped:0 overruns:0 frame:0

              TX packets:15034054 errors:0 dropped:0 overruns:0 carrier:0

              collisions:0 txqueuelen:1000

              RX bytes:8973128258 (8.3 GiB)  TX bytes:12234670411 (11.3 GiB)

    eth1:1    Link encap:Ethernet  HWaddr 6C:92:BF:2A:7C:D7 

              inet addr:169.254.87.215  Bcast:169.254.255.255  Mask:255.255.0.0

              UP BROADCAST RUNNING MULTICAST  MTU:1500Metric:1

    lo        Link encap:Local Loopback 

              inet addr:127.0.0.1  Mask:255.0.0.0

              inet6 addr: ::1/128 Scope:Host

              UP LOOPBACK RUNNING  MTU:16436Metric:1

              RX packets:2213075 errors:0 dropped:0 overruns:0 frame:0

              TX packets:2213075 errors:0 dropped:0 overruns:0 carrier:0

              collisions:0 txqueuelen:0

              RX bytes:1213870175 (1.1 GiB)  TX bytes:1213870175 (1.1 GiB)

    确认防火墙和SELinux是禁用的

    (以Node1为例,两个节点相同)

    [root@xsyy1 ~]# iptables -L

    Chain INPUT (policy ACCEPT)

    target     prot opt source               destination        

    Chain FORWARD (policy ACCEPT)

    target     prot opt source               destination        

    Chain OUTPUT (policy ACCEPT)

    target     prot opt source               destination 

    如果防火墙没禁用,那么通过如下方式修改:

    [root@xsyy1 ~]# service iptables stop

    [root@xsyy1 ~]# chkconfig iptables off

    [root@xsyy1 ~]# getenforce

    Disabled

    如果SELinux没有被禁止,那就通过如下方式修改:

    [root@xsyy1 ~]# cat /etc/selinux/config

    -- 改成SELINUX=disabled

    # This file controls the state of SELinux on the system.

    # SELINUX= can take one of these three values:

    #     enforcing - SELinux security policy is enforced.

    #     permissive - SELinux prints warnings instead of enforcing.

    #     disabled - No SELinux policy is loaded.

    SELINUX=disabled

    # SELINUXTYPE= can take one of these two values:

    #     targeted - Targeted processes are protected,

    #     mls - Multi Level Security protection.

    SELINUXTYPE=targeted

    临时关闭selinux

    [root@xsyy1 ~]# setenforce 0

    正式部署安装

    1、配置/etc/hosts

    修改/etc/hosts文件,前两行不动,添加hostname对应信息。

    (以Node1为例,两个节点相同)

    [root@xsyy1 ~]# cat /etc/hosts

    127.0.0.1   localhost localhost.localdomain localhost4 localhost4.localdomain4

    ::1         localhost localhost.localdomain localhost6 localhost6.localdomain6

    # For Public

    172.16.28.24   xsyy1

    172.16.28.28   xsyy2

    # For VIP

    172.16.28.52    xsyyvip1

    172.16.28.53    xsyyvip2

    # For Private IP

    10.1.20.24  xsyypriv1

    10.1.20.28  xsyypriv2

    # For SCAN IP

    172.16.28.39 scanip

    3、系统配置

    【两个节点相同】

    修改/etc/sysctl.conf,添加如下内容:

    fs.file-max = 6815744

    kernel.sem = 250 32000 100 128

    kernel.shmmni = 4096

    kernel.shmall = 1073741824

    kernel.shmmax = 4398046511104

    net.core.rmem_default = 262144

    net.core.rmem_max = 4194304

    net.core.wmem_default = 262144

    net.core.wmem_max = 1048576

    fs.aio-max-nr = 1048576

    net.ipv4.ip_local_port_range = 9000 65500

     生效:

    [root@12c ~]# sysctl -p

     修改/etc/security/limits.conf,添加如下内容:

    grid   soft   nofile    1024

    grid   hard   nofile    65536

    grid   soft   nproc    2047

    grid   hard   nproc    16384

    grid   soft   stack    10240

    grid   hard   stack    32768

    oracle   soft   nofile    1024

    oracle   hard   nofile    65536

    oracle   soft   nproc    2047

    oracle   hard   nproc    16384

    oracle   soft   stack    10240

    oracle   hard   stack    32768

     修改/etc/pam.d/login

    在 /etc/pam.d/login 文件中添加或编辑下面一行内容:

     session required pam_limits.so

    修改/etc/profile

    [root@rac1 ~]# cat /etc/profile

    if [ $USER = "grid" ] || [ $USER = "oracle" ]; then

    if [ $SHELL = "/bin/ksh" ]; then

    ulimit -p 16384

    ulimit -n 65536

    else

    ulimit -u 16384 -n 65536

    fi

    fi

    4、配置YUM源并安装所需包

    先将默认的yum配置文件删除或者移动,然后创建一个新的

    (以Node1为例,两个节点相同)

    [root@xsyy1 ~]# cd /etc/yum.repos.d

    [root@xsyy1 yum.repos.d]# mkdir bk

    [root@xsyy1 yum.repos.d]# mv public-yum-ol6.repo bk/

    [root@xsyy1 yum.repos.d]# vi gdmsacs.repo

    -- 添加如下内容

    [Oracle]

    name=OEL

    baseurl=file:///mnt/Server

    gpgcheck=0

    enabled=1

    将光驱挂载上

    [root@xsyy1 yum.repos.d]# mount /dev/cdrom /mnt

    mount: block device /dev/sr0 is write-protected, mounting read-only

    下面就可以Yum方式安装所需包了

    [root@xsyy1 yum.repos.d]# yum -y install binutils compat-libstdc++-33 elfutils-libelf

    elfutils-libelf-devel elfutils-libelf-devel-static gcc gcc-c++ glibc glibc-common

    glibc-devel kernel-headers ksh libaio libaio-devel libgcc libgomp libstdc++ libstdc++-devel

    make numactl-devel sysstat unixODBC unixODBC-devel pdksh compat-libcap1

    5、创建用户和组

    (以Node1为例,两个节点相同)

    创建组:

    [root@xsyy1 ~]# /usr/sbin/groupadd -g 54321 oinstall

    [root@xsyy1 ~]# /usr/sbin/groupadd -g 54322 dba

    [root@xsyy1 ~]# /usr/sbin/groupadd -g 54323 oper

    [root@xsyy1 ~]# /usr/sbin/groupadd -g 54324 backupdba

    [root@xsyy1 ~]# /usr/sbin/groupadd -g 54325 dgdba

    [root@xsyy1 ~]# /usr/sbin/groupadd -g 54327 asmdba

    [root@xsyy1 ~]# /usr/sbin/groupadd -g 54328 asmoper

    [root@xsyy1 ~]# /usr/sbin/groupadd -g 54329 asmadmin

    创建用户:

    [root@xsyy1 ~]# /usr/sbin/useradd -u 54321 -g oinstall -G asmadmin,asmdba,asmoper,dba grid

    [root@xsyy1 ~]# /usr/sbin/useradd -u 54322 -g oinstall -G dba,backupdba,dgdba,asmadmin oracle

    设置口令:

    [root@xsyy1 ~]# passwd grid

    [root@xsyy1 ~]# passwd oracle

    6、创建安装目录以及授权

    (以Node1为例,两个节点相同)

    [root@xsyy1 ~]# mkdir -p /u01/app/grid

    [root@xsyy1 ~]# mkdir -p /u01/app/12.1.0/grid

    [root@xsyy1 ~]# mkdir -p /u01/app/oracle/product/12.1.0/db_1

    [root@xsyy1 ~]# chown -R grid.oinstall /u01

    [root@xsyy1 ~]# chown -R oracle.oinstall /u01/app/oracle

    [root@xsyy1 ~]# chmod -R 775 /u01

    7、配置境变量

    节点1: grid用户

    [root@xsyy1 ~]# vi /home/grid/.bash_profile

    export TMP=/tmp

    export TMPDIR=$TMP

    export ORACLE_HOSTNAME=xsyy1.gdmsa.gov.cn

    export ORACLE_BASE=/u01/app/grid

    export ORACLE_HOME=/u01/app/12.1.0/grid

    export ORACLE_SID=+ASM1

    export PATH=/usr/sbin:$PATH

    export PATH=$ORACLE_HOME/bin:$PATH

    export LD_LIBRARY_PATH=$ORACLE_HOME/lib:/lib:/usr/lib

    export CLASSPATH=$ORACLE_HOME/JRE:$ORACLE_HOME/jlib:$ORACLE_HOME/rdbms/jlib

    alias sqlplus="rlwrap sqlplus"

    #Oracle用户

    [root@xsyy1 ~]# vi /home/oracle/.bash_profile

    export PATH

    export TMP=/tmp

    export TMPDIR=$TMP

    export ORACLE_BASE=/u01/app/oracle

    export ORACLE_HOME=$ORACLE_BASE/product/12.1.0/db_1

    export ORACLE_SID=gxshis1

    export PATH=/usr/sbin:$PATH

    export PATH=$ORACLE_HOME/bin:$PATH

    export LD_LIBRARY_PATH=$ORACLE_HOME/lib:/lib:/usr/lib

    export CLASSPATH=$ORACLE_HOME/JRE:$ORACLE_HOME/jlib:$ORACLE_HOME/rdbms/jlib

    节点2:

    [root@xsyy2 ~]# vi /home/grid/.bash_profile

    export TMP=/tmp

    export TMPDIR=$TMP

    export ORACLE_HOSTNAME=xsyy1.gdmsa.gov.cn

    export ORACLE_BASE=/u01/app/grid

    export ORACLE_HOME=/u01/app/12.1.0/grid

    export ORACLE_SID=+ASM1

    export PATH=/usr/sbin:$PATH

    export PATH=$ORACLE_HOME/bin:$PATH

    export LD_LIBRARY_PATH=$ORACLE_HOME/lib:/lib:/usr/lib

    export CLASSPATH=$ORACLE_HOME/JRE:$ORACLE_HOME/jlib:$ORACLE_HOME/rdbms/jlib

    alias sqlplus="rlwrap sqlplus"

    [root@xsyy2 ~]# vi /home/oracle/.bash_profile

    export TMP=/tmp

    export TMPDIR=$TMP

    export ORACLE_BASE=/u01/app/oracle

    export ORACLE_HOME=$ORACLE_BASE/product/12.1.0/db_1

    export ORACLE_SID=gxshis2

    export PATH=/usr/sbin:$PATH

    export PATH=$ORACLE_HOME/bin:$PATH

    export LD_LIBRARY_PATH=$ORACLE_HOME/lib:/lib:/usr/lib

    export CLASSPATH=$ORACLE_HOME/JRE:$ORACLE_HOME/jlib:$ORACLE_HOME/rdbms/jlib

    8、Iscsi挂载磁盘并配置UDEV

    [root@12cr ~]# yum -y install iscsi-initiator-utils

    [root@12cr ~]# service iscsid start

    [root@12cr ~]# chkconfig iscsid on

    [root@12cr ~]# iscsiadm -m discovery -t sendtargets -p 192.168.80.140:3260

    [  OK  ] iscsid: [OK  ]

    192.168.80.140:3260,1 iqn.2006-01.com.openfiler:tsn.3a9cad78121d

    [root@12cr ~]# service iscsi restart

    Stopping iscsi: [  OK  ]

    Starting iscsi: [  OK  ]

    [root@xsyy1 ~]# fdisk -l

    Disk /dev/sda: 53.7 GB, 53687091200 bytes

    255 heads, 63 sectors/track, 6527 cylinders

    Units = cylinders of 16065 * 512 = 8225280 bytes

    Sector size (logical/physical): 512 bytes / 512 bytes

    I/O size (minimum/optimal): 512 bytes / 512 bytes

    Disk identifier: 0x00046ecd

       Device Boot      Start         End      BlocksId  System

    /dev/sda1   *           1          26      204800   83  Linux

    Partition 1 does not end on cylinder boundary.

    /dev/sda2              26         664     5120000   82  Linux swap / Solaris

    Partition 2 does not end on cylinder boundary.

    /dev/sda3             664        6528    47102976   83  Linux

    Disk /dev/sdb: 2147 MB, 2147483648 bytes

    67 heads, 62 sectors/track, 1009 cylinders

    Units = cylinders of 4154 * 512 = 2126848 bytes

    Sector size (logical/physical): 512 bytes / 512 bytes

    I/O size (minimum/optimal): 512 bytes / 512 bytes

    Disk identifier: 0x00000000

    Disk /dev/sdc: 10.5 GB, 10502537216 bytes

    64 heads, 32 sectors/track, 10016 cylinders

    Units = cylinders of 2048 * 512 = 1048576 bytes

    Sector size (logical/physical): 512 bytes / 512 bytes

    I/O size (minimum/optimal): 512 bytes / 512 bytes

    Disk identifier: 0x00000000

    Disk /dev/sdd: 6610 MB, 6610223104 bytes

    204 heads, 62 sectors/track, 1020 cylinders

    Units = cylinders of 12648 * 512 = 6475776 bytes

    Sector size (logical/physical): 512 bytes / 512 bytes

    I/O size (minimum/optimal): 512 bytes / 512 bytes

    Disk identifier: 0x00000000

    Disk /dev/sdf: 8388 MB, 8388608000 bytes

    64 heads, 32 sectors/track, 8000 cylinders

    Units = cylinders of 2048 * 512 = 1048576 bytes

    Sector size (logical/physical): 512 bytes / 512 bytes

    I/O size (minimum/optimal): 512 bytes / 512 bytes

    Disk identifier: 0x00000000

    Disk /dev/sde: 8388 MB, 8388608000 bytes

    64 heads, 32 sectors/track, 8000 cylinders

    Units = cylinders of 2048 * 512 = 1048576 bytes

    Sector size (logical/physical): 512 bytes / 512 bytes

    I/O size (minimum/optimal): 512 bytes / 512 bytes

    Disk identifier: 0x00000000

    Disk /dev/sdg: 5335 MB, 5335154688 bytes

    165 heads, 62 sectors/track, 1018 cylinders

    Units = cylinders of 10230 * 512 = 5237760 bytes

    Sector size (logical/physical): 512 bytes / 512 bytes

    I/O size (minimum/optimal): 512 bytes / 512 bytes

    Disk identifier: 0x00000000

    这里只用了sde、sdf、sdg,其他的是给别的集群使用的。

    //i in b c d e f g h;

    [root@xsyy1 ~]# for i in e f g;do

    > echo "KERNEL=="sd*", BUS=="scsi", PROGRAM=="/sbin/scsi_id --whitelisted --replace-whitespace --device=/dev/$name", RESULT=="`/sbin/scsi_id --whitelisted --replace-whitespace --device=/dev/sd$i`", NAME="asm-disk$i", OWNER="grid", GROUP="asmadmin", MODE="0660""

    > done

    KERNEL=="sd*", BUS=="scsi", PROGRAM=="/sbin/scsi_id --whitelisted --replace-whitespace --device=/dev/$name", RESULT=="14f504e46494c450058444273784d2d64705a6a2d544c4756", NAME="asm-diske", OWNER="grid", GROUP="asmadmin", MODE="0660"

    KERNEL=="sd*", BUS=="scsi", PROGRAM=="/sbin/scsi_id --whitelisted --replace-whitespace --device=/dev/$name", RESULT=="14f504e46494c45003030365263642d32714a702d6866744c", NAME="asm-diskf", OWNER="grid", GROUP="asmadmin", MODE="0660"

    KERNEL=="sd*", BUS=="scsi", PROGRAM=="/sbin/scsi_id --whitelisted --replace-whitespace --device=/dev/$name", RESULT=="14f504e46494c45006c58576a76452d716d50492d71436c76", NAME="asm-diskg", OWNER="grid", GROUP="asmadmin", MODE="0660"

    配置UDEV: 将上述内容拷贝至节点

    (以Node1为例,两个节点相同)

    [root@xsyy1 ~]# vi /etc/udev/rules.d/99-oracle-asmdevices.rules

    -- 添加如下内容:

    KERNEL=="sd*", BUS=="scsi", PROGRAM=="/sbin/scsi_id --whitelisted --replace-whitespace --device=/dev/$name", RESULT=="14f504e46494c45003030365263642d32714a702d6866744c", NAME="asm-data", OWNER="grid", GROUP="asmadmin", MODE="0660"

    KERNEL=="sd*", BUS=="scsi", PROGRAM=="/sbin/scsi_id --whitelisted --replace-whitespace --device=/dev/$name", RESULT=="14f504e46494c450058444273784d2d64705a6a2d544c4756", NAME="asm-fra", OWNER="grid", GROUP="asmadmin", MODE="0660"

    KERNEL=="sd*", BUS=="scsi", PROGRAM=="/sbin/scsi_id --whitelisted --replace-whitespace --device=/dev/$name", RESULT=="14f504e46494c45006c58576a76452d716d50492d71436c76", NAME="asm-crs", OWNER="grid", GROUP="asmadmin", MODE="0660"

    --重启UDEV:

    [root@xsyy1 ~]# /sbin/start_udev

    Starting udev: [  OK  ]

    --检查共享设备的所属关系和权限:

    [root@xsyy1 ~]# ls -l /dev/asm*

    brw-rw---- 1 grid asmadmin 8, 96 Jun 29 21:56 /dev/asm-crs

    brw-rw---- 1 grid asmadmin 8, 64 Jun 29 21:56 /dev/asm-data

    brw-rw---- 1 grid asmadmin 8, 80 Jun 29 21:56 /dev/asm-fra

    9、禁用NTP服务

    (以Node1为例,两个节点相同)

    [root@xsyy1 ~]# chkconfig ntpd off

    [root@xsyy1 ~]# mv /etc/ntp.conf /etc/ntp.conf.bak

    [root@xsyy1 ~]# rm /var/run/ntpd.pid

    10、解压介质

    节点1:

    [root@xsyy1 ~]# chown -R grid.oinstall /install/

    [root@xsyy1 ~]# chown oracle.oinstall /install/linuxamd64_12c_database_*

    [root@xsyy1 ~]# chmod 775 /install

    [root@xsyy1 ~]# su - grid

    [root@xsyy1 ~]# cd /install/

    [grid@xsyy1 install]$ unzip linuxamd64_12c_grid_1of2.zip

    [grid@xsyy1 install]$ unzip linuxamd64_12c_grid_2of2.zip

    [root@xsyy1 ~]# su - oracle

    [oracle@xsyy1 install]$ unzip linuxamd64_12c_database_1of2.zip

    解压之后大小为:

    [oracle@xsyy1 install]$ du -sh grid

    2.1G    grid

    [oracle@xsyy1 install]$ du -sh database/

    2.6G    database/

       

    安装cvu相关rpm包:

    [root@xsyy1 ~]# cd /install/grid/rpm/

    [root@xsyy1 rpm]# rpm -ivh cvuqdisk-1.0.9-1.rpm

    Preparing...########################################### [100%]

    Using default group oinstall to install package

       1:cvuqdisk########################################### [100%]

    拷贝到节点2并安装:

    [root@xsyy1 rpm]# scp cvuqdisk-1.0.9-1.rpm xsyy2:/install

    [root@xsyy2 install]# rpm -ivh cvuqdisk-1.0.9-1.rpm

    Preparing...########################################### [100%]

    Using default group oinstall to install package

       1:cvuqdisk               ########################################### [100%]

    SSH配置

    /usr/bin/ssh-keygen -t rsa

    /usr/bin/ssh-keygen -t dsa

    把4个文件合成一个文件

    注意权限要是是644

    chmod 644 authorized_keys

    在节点RAC2进行的操作

     [oracle@RAC2 ~]$ ls -a

    .   .bash_history  .bash_profile.emacs  .mozilla  .viminfo

    ..  .bash_logout   .bashrc        .kde.ssh      .Xauthority

    [oracle@RAC2 ~]$ cd .ssh

    [oracle@RAC2 .ssh]$ ls

    id_dsa  id_dsa.pub  id_rsaid_rsa.pub

    [oracle@RAC2 .ssh]$ scp id_dsa.pub rac1:/home/oracle/.ssh/id_dsa.pub2

    oracle@rac1's password:

    id_dsa.pub100%  601     0.6KB/s00:00  

    [oracle@RAC2 .ssh]$ ls

    id_dsa  id_dsa.pub  id_rsaid_rsa.pub  known_hosts

    [oracle@RAC2 .ssh]$ scp id_rsa.pub rac1:/home/oracle/.ssh/id_rsa.pub2

    oracle@rac1's password:

    id_rsa.pub

    再到RAC1节点

    cat id_dsa.pubid_dsa.pub2  id_rsa.pub  id_rsa.pub2 > authorized_keys

    scp authorized_keysrac2:/home/oracle/.ssh

    之后要测试一下(如果都不需要密码,得到时间后。就说明ssh 对等性配置成功。)

     

     

    11、校验

    这里只贴失败的项,其中第一个是物理内存不足问题,Oracle推荐每节点至少4GB内存空间,我这里只有2G;第二个问题是配置DNS,这个问题我们可以忽略。

    节点1:

    [grid@xsyy1 grid]$ ./runcluvfy.sh stage -pre crsinst -n xsyy1,xsyy2 -verbose

    Check: Total memory

      Node Name     Available                 Required                  Status   

      ------------  ------------------------  ------------------------  ----------

      xsyy2       1.9567GB (2051748.0KB)    4GB (4194304.0KB)         failed   

      xsyy1       1.9567GB (2051748.0KB)    4GB (4194304.0KB)         failed   

    Result: Total memory check failed

    Result: Default user file creation mask check passed

    Checking integrity of file "/etc/resolv.conf" across nodes

    Checking the file "/etc/resolv.conf" to make sure only one of domain and search entries is defined

    "domain" and "search" entries do not coexist in any  "/etc/resolv.conf" file

    Checking if domain entry in file "/etc/resolv.conf" is consistent across the nodes...

    "domain" entry does not exist in any "/etc/resolv.conf" file

    Checking if search entry in file "/etc/resolv.conf" is consistent across the nodes...

    Checking file "/etc/resolv.conf" to make sure that only one search entry is defined

    More than one "search" entry does not exist in any "/etc/resolv.conf" file

    All nodes have same "search" order defined in file "/etc/resolv.conf"

    Checking DNS response time for an unreachable node

      Node Name                             Status                 

    ------------------------------------------------------------

      xsyy1                               failed                  

      xsyy2                               failed                 

    PRVF-5636 : The DNS response time for an unreachable node exceeded "15000" ms on following nodes: xsyy1,xsyy2

    Check for integrity of file "/etc/resolv.conf" failed

    12、安装

    1)安装GI

    [root@12cr ~]# su - grid

    [grid@12cr ~]$ cd /install/grid/

    我打开Xmanager - Passive,设置DISPLAY,调用runInstaller启动OUI

    [grid@12cr grid]$ export DISPLAY=192.168.1.1:0.0

    [grid@12cr grid]$ ./runInstaller

     

     

    这里有几项校验不过去,这几个都忽略。

     

    脚本输出内容:

    [root@xsyy1 ~]# /u01/app/oraInventory/orainstRoot.sh

    Changing permissions of /u01/app/oraInventory.

    Adding read,write permissions for group.

    Removing read,write,execute permissions for world.

    Changing groupname of /u01/app/oraInventory to oinstall.

    The execution of the script is complete.

    节点1:

    [root@xsyy1 ~]# /u01/app/12.1.0/grid/root.sh

    Performing root user operation for Oracle 12c

    The following environment variables are set as:

        ORACLE_OWNER= grid

        ORACLE_HOME=  /u01/app/12.1.0/grid

    Enter the full pathname of the local bin directory: [/usr/local/bin]:

       Copying dbhome to /usr/local/bin ...

       Copying oraenv to /usr/local/bin ...

       Copying coraenv to /usr/local/bin ...

    Creating /etc/oratab file...

    Entries will be added to the /etc/oratab file as needed by

    Database Configuration Assistant when a database is created

    Finished running generic part of root script.

    Now product-specific root actions will be performed.

    Using configuration parameter file: /u01/app/12.1.0/grid/crs/install/crsconfig_params

    2013/07/01 00:30:25 CLSRSC-363: User ignored prerequisites during installation

    OLR initialization - successful

      root wallet

      root wallet cert

      root cert export

      peer wallet

      profile reader wallet

      pa wallet

      peer wallet keys

      pa wallet keys

      peer cert request

      pa cert request

      peer cert

      pa cert

      peer root cert TP

      profile reader root cert TP

      pa root cert TP

      peer pa cert TP

      pa peer cert TP

      profile reader pa cert TP

      profile reader peer cert TP

      peer user cert

      pa user cert

    2013/07/01 00:31:22 CLSRSC-330: Adding Clusterware entries to file 'oracle-ohasd.conf'

    CRS-4133: Oracle High Availability Services has been stopped.

    CRS-4123: Oracle High Availability Services has been started.

    CRS-4133: Oracle High Availability Services has been stopped.

    CRS-4123: Oracle High Availability Services has been started.

    CRS-2673: Attempting to stop 'ora.drivers.acfs' on 'xsyy1'

    CRS-2677: Stop of 'ora.drivers.acfs' on 'xsyy1' succeeded

    CRS-2672: Attempting to start 'ora.evmd' on 'xsyy1'

    CRS-2672: Attempting to start 'ora.mdnsd' on 'xsyy1'

    CRS-2676: Start of 'ora.mdnsd' on 'xsyy1' succeeded

    CRS-2676: Start of 'ora.evmd' on 'xsyy1' succeeded

    CRS-2672: Attempting to start 'ora.gpnpd' on 'xsyy1'

    CRS-2676: Start of 'ora.gpnpd' on 'xsyy1' succeeded

    CRS-2672: Attempting to start 'ora.cssdmonitor' on 'xsyy1'

    CRS-2672: Attempting to start 'ora.gipcd' on 'xsyy1'

    CRS-2676: Start of 'ora.cssdmonitor' on 'xsyy1' succeeded

    CRS-2676: Start of 'ora.gipcd' on 'xsyy1' succeeded

    CRS-2672: Attempting to start 'ora.cssd' on 'xsyy1'

    CRS-2672: Attempting to start 'ora.diskmon' on 'xsyy1'

    CRS-2676: Start of 'ora.diskmon' on 'xsyy1' succeeded

    CRS-2676: Start of 'ora.cssd' on 'xsyy1' succeeded

    ASM created and started successfully.

    Disk Group RACCRS created successfully.

    CRS-2672: Attempting to start 'ora.crf' on 'xsyy1'

    CRS-2672: Attempting to start 'ora.storage' on 'xsyy1'

    CRS-2676: Start of 'ora.storage' on 'xsyy1' succeeded

    CRS-2676: Start of 'ora.crf' on 'xsyy1' succeeded

    CRS-2672: Attempting to start 'ora.crsd' on 'xsyy1'

    CRS-2676: Start of 'ora.crsd' on 'xsyy1' succeeded

    CRS-4256: Updating the profile

    Successful addition of voting disk d883c23a7bfc4fdcbf418c9f631bd0af.

    Successfully replaced voting disk group with +RACCRS.

    CRS-4256: Updating the profile

    CRS-4266: Voting file(s) successfully replaced

    ##  STATE    File Universal Id                File Name Disk group

    --  -----    -----------------                --------- ---------

     1. ONLINE   d883c23a7bfc4fdcbf418c9f631bd0af (/dev/asm-crs) [RACCRS]

    Located 1 voting disk(s).

    CRS-2791: Starting shutdown of Oracle High Availability Services-managed resources on 'xsyy1'

    CRS-2673: Attempting to stop 'ora.crsd' on 'xsyy1'

    CRS-2677: Stop of 'ora.crsd' on 'xsyy1' succeeded

    CRS-2673: Attempting to stop 'ora.storage' on 'xsyy1'

    CRS-2673: Attempting to stop 'ora.mdnsd' on 'xsyy1'

    CRS-2673: Attempting to stop 'ora.gpnpd' on 'xsyy1'

    CRS-2673: Attempting to stop 'ora.drivers.acfs' on 'xsyy1'

    CRS-2677: Stop of 'ora.storage' on 'xsyy1' succeeded

    CRS-2673: Attempting to stop 'ora.ctssd' on 'xsyy1'

    CRS-2673: Attempting to stop 'ora.evmd' on 'xsyy1'

    CRS-2673: Attempting to stop 'ora.asm' on 'xsyy1'

    CRS-2677: Stop of 'ora.drivers.acfs' on 'xsyy1' succeeded

    CRS-2677: Stop of 'ora.mdnsd' on 'xsyy1' succeeded

    CRS-2677: Stop of 'ora.gpnpd' on 'xsyy1' succeeded

    CRS-2677: Stop of 'ora.evmd' on 'xsyy1' succeeded

    CRS-2677: Stop of 'ora.ctssd' on 'xsyy1' succeeded

    CRS-2677: Stop of 'ora.asm' on 'xsyy1' succeeded

    CRS-2673: Attempting to stop 'ora.cluster_interconnect.haip' on 'xsyy1'

    CRS-2677: Stop of 'ora.cluster_interconnect.haip' on 'xsyy1' succeeded

    CRS-2673: Attempting to stop 'ora.cssd' on 'xsyy1'

    CRS-2677: Stop of 'ora.cssd' on 'xsyy1' succeeded

    CRS-2673: Attempting to stop 'ora.crf' on 'xsyy1'

    CRS-2677: Stop of 'ora.crf' on 'xsyy1' succeeded

    CRS-2673: Attempting to stop 'ora.gipcd' on 'xsyy1'

    CRS-2677: Stop of 'ora.gipcd' on 'xsyy1' succeeded

    CRS-2793: Shutdown of Oracle High Availability Services-managed resources on 'xsyy1' has completed

    CRS-4133: Oracle High Availability Services has been stopped.

    CRS-4123: Starting Oracle High Availability Services-managed resources

    CRS-2672: Attempting to start 'ora.mdnsd' on 'xsyy1'

    CRS-2672: Attempting to start 'ora.evmd' on 'xsyy1'

    CRS-2676: Start of 'ora.mdnsd' on 'xsyy1' succeeded

    CRS-2676: Start of 'ora.evmd' on 'xsyy1' succeeded

    CRS-2672: Attempting to start 'ora.gpnpd' on 'xsyy1'

    CRS-2676: Start of 'ora.gpnpd' on 'xsyy1' succeeded

    CRS-2672: Attempting to start 'ora.gipcd' on 'xsyy1'

    CRS-2676: Start of 'ora.gipcd' on 'xsyy1' succeeded

    CRS-2672: Attempting to start 'ora.cssdmonitor' on 'xsyy1'

    CRS-2676: Start of 'ora.cssdmonitor' on 'xsyy1' succeeded

    CRS-2672: Attempting to start 'ora.cssd' on 'xsyy1'

    CRS-2672: Attempting to start 'ora.diskmon' on 'xsyy1'

    CRS-2676: Start of 'ora.diskmon' on 'xsyy1' succeeded

    CRS-2789: Cannot stop resource 'ora.diskmon' as it is not running on server 'xsyy1'

    CRS-2676: Start of 'ora.cssd' on 'xsyy1' succeeded

    CRS-2672: Attempting to start 'ora.cluster_interconnect.haip' on 'xsyy1'

    CRS-2672: Attempting to start 'ora.ctssd' on 'xsyy1'

    CRS-2676: Start of 'ora.ctssd' on 'xsyy1' succeeded

    CRS-2676: Start of 'ora.cluster_interconnect.haip' on 'xsyy1' succeeded

    CRS-2672: Attempting to start 'ora.asm' on 'xsyy1'

    CRS-2676: Start of 'ora.asm' on 'xsyy1' succeeded

    CRS-2672: Attempting to start 'ora.storage' on 'xsyy1'

    CRS-2676: Start of 'ora.storage' on 'xsyy1' succeeded

    CRS-2672: Attempting to start 'ora.crf' on 'xsyy1'

    CRS-2676: Start of 'ora.crf' on 'xsyy1' succeeded

    CRS-2672: Attempting to start 'ora.crsd' on 'xsyy1'

    CRS-2676: Start of 'ora.crsd' on 'xsyy1' succeeded

    CRS-6023: Starting Oracle Cluster Ready Services-managed resources

    CRS-6017: Processing resource auto-start for servers: xsyy1

    CRS-6016: Resource auto-start has completed for server xsyy1

    CRS-6024: Completed start of Oracle Cluster Ready Services-managed resources

    CRS-4123: Oracle High Availability Services has been started.

    2013/07/01 00:38:01 CLSRSC-343: Successfully started Oracle clusterware stack

    CRS-2672: Attempting to start 'ora.asm' on 'xsyy1'

    CRS-2676: Start of 'ora.asm' on 'xsyy1' succeeded

    CRS-2672: Attempting to start 'ora.RACCRS.dg' on 'xsyy1'

    CRS-2676: Start of 'ora.RACCRS.dg' on 'xsyy1' succeeded

    2013/07/01 00:39:51 CLSRSC-325: Configure Oracle Grid Infrastructure for a Cluster ... succeeded

    节点2:

    [root@xsyy2 ~]# /u01/app/12.1.0/grid/root.sh

    Performing root user operation for Oracle 12c

    The following environment variables are set as:

        ORACLE_OWNER= grid

        ORACLE_HOME=  /u01/app/12.1.0/grid

    Enter the full pathname of the local bin directory: [/usr/local/bin]:

       Copying dbhome to /usr/local/bin ...

       Copying oraenv to /usr/local/bin ...

       Copying coraenv to /usr/local/bin ...

    Creating /etc/oratab file...

    Entries will be added to the /etc/oratab file as needed by

    Database Configuration Assistant when a database is created

    Finished running generic part of root script.

    Now product-specific root actions will be performed.

    Using configuration parameter file: /u01/app/12.1.0/grid/crs/install/crsconfig_params

    2013/07/01 00:42:51 CLSRSC-363: User ignored prerequisites during installation

    OLR initialization - successful

    2013/07/01 00:43:18 CLSRSC-330: Adding Clusterware entries to file 'oracle-ohasd.conf'

    CRS-4133: Oracle High Availability Services has been stopped.

    CRS-4123: Oracle High Availability Services has been started.

    CRS-4133: Oracle High Availability Services has been stopped.

    CRS-4123: Oracle High Availability Services has been started.

    CRS-2791: Starting shutdown of Oracle High Availability Services-managed resources on 'xsyy2'

    CRS-2673: Attempting to stop 'ora.drivers.acfs' on 'xsyy2'

    CRS-2677: Stop of 'ora.drivers.acfs' on 'xsyy2' succeeded

    CRS-2793: Shutdown of Oracle High Availability Services-managed resources on 'xsyy2' has completed

    CRS-4133: Oracle High Availability Services has been stopped.

    CRS-4123: Starting Oracle High Availability Services-managed resources

    CRS-2672: Attempting to start 'ora.evmd' on 'xsyy2'

    CRS-2672: Attempting to start 'ora.mdnsd' on 'xsyy2'

    CRS-2676: Start of 'ora.evmd' on 'xsyy2' succeeded

    CRS-2676: Start of 'ora.mdnsd' on 'xsyy2' succeeded

    CRS-2672: Attempting to start 'ora.gpnpd' on 'xsyy2'

    CRS-2676: Start of 'ora.gpnpd' on 'xsyy2' succeeded

    CRS-2672: Attempting to start 'ora.gipcd' on 'xsyy2'

    CRS-2676: Start of 'ora.gipcd' on 'xsyy2' succeeded

    CRS-2672: Attempting to start 'ora.cssdmonitor' on 'xsyy2'

    CRS-2676: Start of 'ora.cssdmonitor' on 'xsyy2' succeeded

    CRS-2672: Attempting to start 'ora.cssd' on 'xsyy2'

    CRS-2672: Attempting to start 'ora.diskmon' on 'xsyy2'

    CRS-2676: Start of 'ora.diskmon' on 'xsyy2' succeeded

    CRS-2789: Cannot stop resource 'ora.diskmon' as it is not running on server 'xsyy2'

    CRS-2676: Start of 'ora.cssd' on 'xsyy2' succeeded

    CRS-2672: Attempting to start 'ora.cluster_interconnect.haip' on 'xsyy2'

    CRS-2672: Attempting to start 'ora.ctssd' on 'xsyy2'

    CRS-2676: Start of 'ora.ctssd' on 'xsyy2' succeeded

    CRS-2676: Start of 'ora.cluster_interconnect.haip' on 'xsyy2' succeeded

    CRS-2672: Attempting to start 'ora.asm' on 'xsyy2'

    CRS-2676: Start of 'ora.asm' on 'xsyy2' succeeded

    CRS-2672: Attempting to start 'ora.storage' on 'xsyy2'

    CRS-2676: Start of 'ora.storage' on 'xsyy2' succeeded

    CRS-2672: Attempting to start 'ora.crf' on 'xsyy2'

    CRS-2676: Start of 'ora.crf' on 'xsyy2' succeeded

    CRS-2672: Attempting to start 'ora.crsd' on 'xsyy2'

    CRS-2676: Start of 'ora.crsd' on 'xsyy2' succeeded

    CRS-6017: Processing resource auto-start for servers: xsyy2

    CRS-2673: Attempting to stop 'ora.LISTENER_SCAN1.lsnr' on 'xsyy1'

    CRS-2672: Attempting to start 'ora.ons' on 'xsyy2'

    CRS-2677: Stop of 'ora.LISTENER_SCAN1.lsnr' on 'xsyy1' succeeded

    CRS-2673: Attempting to stop 'ora.scan1.vip' on 'xsyy1'

    CRS-2677: Stop of 'ora.scan1.vip' on 'xsyy1' succeeded

    CRS-2672: Attempting to start 'ora.scan1.vip' on 'xsyy2'

    CRS-2676: Start of 'ora.scan1.vip' on 'xsyy2' succeeded

    CRS-2672: Attempting to start 'ora.LISTENER_SCAN1.lsnr' on 'xsyy2'

    CRS-2676: Start of 'ora.ons' on 'xsyy2' succeeded

    CRS-2676: Start of 'ora.LISTENER_SCAN1.lsnr' on 'xsyy2' succeeded

    CRS-6016: Resource auto-start has completed for server xsyy2

    CRS-6024: Completed start of Oracle Cluster Ready Services-managed resources

    CRS-4123: Oracle High Availability Services has been started.

    2013/07/01 00:48:43 CLSRSC-343: Successfully started Oracle clusterware stack

    2013/07/01 00:49:05 CLSRSC-325: Configure Oracle Grid Infrastructure for a Cluster ... succeeded

     

    Oracle Cluster Verification 将在Oracle Database 安装中完成。

    忽略错误,至此GRID 安装完成

    查看状态:

    检查scan配置

    [grid@xsyy1 ~]$ cluvfy comp scan

    [grid@xsyy1 ~]$ crsctl check cluster

    [grid@xsyy1 ~]$ crsctl stat res -t

    --------------------------------------------------------------------------------

    Name           Target  StateServer                   State details      

    --------------------------------------------------------------------------------

    Local Resources

    --------------------------------------------------------------------------------

    ora.LISTENER.lsnr

                   ONLINE  ONLINExsyy1                  STABLE

                   ONLINE  ONLINExsyy2                  STABLE

    ora.RACCRS.dg

                   ONLINEONLINE       xsyy1                  STABLE

                   ONLINE  ONLINExsyy2                  STABLE

    ora.asm

                   ONLINE  ONLINExsyy1Started,STABLE

                   ONLINE  ONLINExsyy2                  Started,STABLE

    ora.net1.network

                   ONLINE  ONLINExsyy1                  STABLE

                   ONLINE  ONLINExsyy2                  STABLE

    ora.ons

                   ONLINE  ONLINExsyy1                  STABLE

                   ONLINEONLINE       xsyy2                  STABLE

    --------------------------------------------------------------------------------

    Cluster Resources

    --------------------------------------------------------------------------------

    ora.xsyy1.vip

      1        ONLINE  ONLINExsyy1                  STABLE

    ora.xsyy2.vip

          1        ONLINE  ONLINExsyy2                  STABLE

    ora.LISTENER_SCAN1.lsnr

          1        ONLINE  ONLINExsyy2                  STABLE

    ora.LISTENER_SCAN2.lsnr

          1        ONLINE  ONLINExsyy1                  STABLE

    ora.LISTENER_SCAN3.lsnr

          1        ONLINE  ONLINExsyy1                  STABLE

    ora.MGMTLSNR

          1        ONLINE  ONLINExsyy1                  169.254.88.173 192.1

    68.80.150,STABLE

    ora.cvu

          1        ONLINE  ONLINExsyy1                  STABLE

    ora.mgmtdb

          1        ONLINE  ONLINExsyy1                  Open,STABLE

    ora.oc4j

          1        ONLINEONLINE       xsyy1                  STABLE

    ora.scan1.vip

          1        ONLINE  ONLINExsyy2                  STABLE

    ora.scan2.vip

          1        ONLINE  ONLINExsyy1                  STABLE

    ora.scan3.vip

          1        ONLINE  ONLINExsyy1                  STABLE

    --------------------------------------------------------------------------------

    如果此前配置了ssh,需要在此配置load ssh key

    [oracle@xsyy1 ~]$ exec /usr/bin/ssh-agent $SHELL

    [oracle@xsyy1 ~]$ /usr/bin/ssh-add

    2)创建ASM磁盘组

    节点1:

    [grid@xsyy1 ~]$ export DISPLAY=192.168.1.1:0.0

    [grid@xsyy1 ~]$ asmca

     

    建立data和fra 磁盘组

     

    3)安装RDBMS软件

    节点1:

    [root@xsyy1 ~]# su - oracle

    [oracle@xsyy1 ~]$ cd /install/database/

    [oracle@xsyy1 database]$ export DISPLAY=192.168.1.1:0.0

    [oracle@xsyy1 database]$ ./runInstaller

     

     

     

      

    执行脚本:

    节点1:

    [root@xsyy1 ~]# /u01/app/oracle/product/12.1.0/dbhome_1/root.sh

    Performing root user operation for Oracle 12c

    The following environment variables are set as:

        ORACLE_OWNER= oracle

        ORACLE_HOME=  /u01/app/oracle/product/12.1.0/dbhome_1

    Enter the full pathname of the local bin directory: [/usr/local/bin]:

    The contents of "dbhome" have not changed. No need to overwrite.

    The contents of "oraenv" have not changed. No need to overwrite.

    The contents of "coraenv" have not changed. No need to overwrite.

    Entries will be added to the /etc/oratab file as needed by

    Database Configuration Assistant when a database is created

    Finished running generic part of root script.

    Now product-specific root actions will be performed.

     

    4)创建数据库

    [oracle@xsyy1 ~]$ dbca

     

    选择server pool为两个节点

     

    勾上此选项,系统自动完善此前grid安装中的未执行的步骤.

    点浏览选择ASM磁盘组

     

     

     

    字符集根据需要进行选择

     

    13、最终结果

    资源状态:

    [grid@xsyy1 ~]$ crsctl stat res -t

    --------------------------------------------------------------------------------

    Name           Target  StateServer                   State details      

    --------------------------------------------------------------------------------

    Local Resources

    --------------------------------------------------------------------------------

    ora.LISTENER.lsnr

                   ONLINE  ONLINExsyy1                  STABLE

                   ONLINE  ONLINExsyy2                  STABLE

    ora.RACCRS.dg

                   ONLINE  ONLINExsyy1                  STABLE

                   ONLINE  ONLINExsyy2                  STABLE

    ora.RACDATA.dg

                   ONLINE  ONLINExsyy1                  STABLE

                   ONLINE  ONLINExsyy2                  STABLE

    ora.RACFRA.dg

                   ONLINE  ONLINExsyy1                  STABLE

                   ONLINE  ONLINExsyy2                  STABLE

    ora.asm

                   ONLINE  ONLINExsyy1Started,STABLE

                   ONLINE  ONLINExsyy2Started,STABLE

    ora.net1.network

                   ONLINE  ONLINExsyy1                  STABLE

                   ONLINE  ONLINExsyy2                  STABLE

    ora.ons

                   ONLINE  ONLINExsyy1                  STABLE

                   ONLINE  ONLINExsyy2                  STABLE

    --------------------------------------------------------------------------------

    Cluster Resources

    --------------------------------------------------------------------------------

    ora.xsyy1.vip

          1        ONLINE  ONLINExsyy1                  STABLE

    ora.xsyy2.vip

          1        ONLINE  ONLINExsyy2                  STABLE

    ora.LISTENER_SCAN1.lsnr

          1        ONLINE  ONLINExsyy2                  STABLE

    ora.LISTENER_SCAN2.lsnr

          1        ONLINE  ONLINExsyy1                  STABLE

    ora.LISTENER_SCAN3.lsnr

          1        ONLINE  ONLINExsyy1                  STABLE

    ora.MGMTLSNR

          1        ONLINE  ONLINExsyy1                  169.254.88.173 192.1

    68.80.150,STABLE

    ora.cvu

          1        ONLINE  ONLINExsyy1                  STABLE

    ora.gdmsacs12c.db

          1        ONLINE  ONLINExsyy1                  Open,STABLE

          2        ONLINE  ONLINExsyy2                  Open,STABLE

    ora.mgmtdb

          1        ONLINE  ONLINExsyy1                  Open,STABLE

    ora.oc4j

          1        ONLINE  ONLINExsyy1                  STABLE

    ora.scan1.vip

          1        ONLINE  ONLINExsyy2                  STABLE

    ora.scan2.vip

          1        ONLINE  ONLINExsyy1                  STABLE

    ora.scan3.vip

          1        ONLINE  ONLINExsyy1                  STABLE

    --------------------------------------------------------------------------------

    RAC数据库配置信息

    [grid@xsyy1 ~]$ srvctl config database -d gdmsacs12c

    Database unique name: gdmsacs12c

    Database name: gdmsacs12c

    Oracle home: /u01/app/oracle/product/12.1.0/dbhome_1

    Oracle user: oracle

    Spfile: +RACDATA/gdmsacs12c/spfilegdmsacs12c.ora

    Password file: +RACDATA/gdmsacs12c/orapwgdmsacs12c

    Domain:

    Start options: open

    Stop options: immediate

    Database role: PRIMARY

    Management policy: AUTOMATIC

    Server pools: gdmsacs12c

    Database instances: gdmsacs12c1,gdmsacs12c2

    Disk Groups: RACFRA,RACDATA

    Mount point paths:

    Services:

    Type: RAC

    Start concurrency:

    Stop concurrency:

    Database is administrator managed

    [grid@xsyy1 ~]$ srvctl status database -d gdmsacs12c

    Instance gdmsacs12c1 is running on node xsyy1

    Instance gdmsacs12c2 is running on node xsyy2

    [grid@xsyy1 ~]$ srvctl status listener

    Listener LISTENER is enabled

    Listener LISTENER is running on node(s): xsyy1,xsyy2

    实例状态:

    sys@GDMSACS12C> select instance_name, status from gv$instance;

    INSTANCE_NAME    STATUS

    ---------------- ------------

    gdmsacs12c1        OPEN

    gdmsacs12c2        OPEN

    sys@GDMSACS12C> col open_time for a25

    sys@GDMSACS12C> col name for a10

    sys@GDMSACS12C> select CON_ID, NAME, OPEN_MODE, OPEN_TIME, CREATE_SCN, TOTAL_SIZE from v$pdbs;

        CON_ID NAME       OPEN_MODE  OPEN_TIME                 CREATE_SCN TOTAL_SIZE

    ---------- ---------- ---------- ------------------------- ---------- ----------

             2 PDB$SEED   READ ONLY  01-JUL-13 04.33.07.302 PM    1720772  283115520

             3 GDMSACS      READ WRITE 01-JUL-13 04.38.41.339 PM    1934986  288358400

    查看归档启用与否:

    sys@GDMSACS12C> archive log list

    Database log mode              No Archive Mode

    Automatic archivalDisabled

    Archive destinationUSE_DB_RECOVERY_FILE_DEST

    Oldest online log sequence     15

    Current log sequence           16

    现在手动开启:

    [oracle@xsyy1 ~]$ srvctl stop database -d gdmsacs12c

    [oracle@xsyy1 ~]$ srvctl start database -d gdmsacs12c -o mount

    节点1:

    [oracle@xsyy1 ~]$ sqlplus / as sysdba

    SQL*Plus: Release 12.1.0.1.0 Production on Mon Jul 1 17:13:58 2014

    Copyright (c) 1982, 2014, Oracle.  All rights reserved.

    Connected to:

    Oracle Database 12c Enterprise Edition Release 12.1.0.1.0 - 64bit Production

    With the Partitioning, Real Application Clusters, Automatic Storage Management, OLAP,

    Advanced Analytics and Real Application Testing options

    idle> alter database archivelog;

    Database altered.

    idle> alter database open;

    Database altered.

    节点2:

    [oracle@xsyy2 ~]$ sqlplus / as sysdba

    SQL*Plus: Release 12.1.0.1.0 Production on Mon Jul 1 17:17:07 2014

    Copyright (c) 1982, 2014, Oracle.  All rights reserved.

    Connected to:

    Oracle Database 12c Enterprise Edition Release 12.1.0.1.0 - 64bit Production

    With the Partitioning, Real Application Clusters, Automatic Storage Management, OLAP,

    Advanced Analytics and Real Application Testing options

    SQL> alter database open;

    Database altered.

    这样RAC已经运行于归档模式了

    sys@GDMSACS12C> archive log list

    Database log mode              Archive Mode

    Automatic archivalEnabled

    Archive destinationUSE_DB_RECOVERY_FILE_DEST

    Oldest online log sequence     15

    Next log sequence to archive   16

    Current log sequence           16

    附录A

     

    【节点1】【节点2】相同操作。

    #yum install smartmontools

    #yum install oracleasm-support

    先对硬盘分区 fdisk /dev/sdb …… (根据需要  sde  sdf sdg 均同样操作)

     

    #oracle asm configure –i

     

    建立 ASM 盘

     

    在每个节点: #oracleasm scandisks

     

  • 相关阅读:
    https进行配置以及http跳转到https配置
    centos7修改系统语言为简体中文
    iptables snat 和dnat说明
    python多线程执行类中的静态方法
    python 磁盘空间操作
    python logging 工具
    python 动态调用函数
    python 读写XML
    js加载json数据成表格
    python 比较两个数据库postgresql
  • 原文地址:https://www.cnblogs.com/liang545621/p/9416881.html
Copyright © 2020-2023  润新知