• Oracle 11gR2 RAC + ASM + Grid安装


    (一)环境准备

    主机操作系统 windows10
    虚拟机平台 vmware workstation 12
    虚拟机操作系统 redhat 5.5 x86(32位) :Linux.5.5.for.x86.rhel-server-5.5-i386-dvd.iso
    grid版本 linux_11gR2_grid.zip (32位)
    oracle版本 linux_11gR2_database_1of2 和  linux_11gR2_database_2of2(32位)
    共享存储 ASM

    (二)操作系统安装

    (2.1)操作系统安装

    操作系统安装相对而言较简单,不做详述。系统的配置信息大致如下,后面还会根据需要添加和删除设备

    image

    (2.2)安装VM Tools

    为了方便在主机与虚拟机之间进行文件拖拽、复制粘贴、使用共享文件夹等功能,我们需要安装VM Tools,VM Tools的安装包已经集合在了vmware workstation里面了。下面一步一步安装VM Tools。

    step 1:虚拟机-> 安装Vmware Tools

    image

    step 2: mount查看是否装载 VMware Tools 虚拟 CD-ROM 映像,如果有红色部分,说明已经将VM tools安装包mount在了/media目录下

    image

    step 3:转到安装目录/tmp,解压VM tools安装包

    image

    step 4:开始安装VM Tools

    [root@Redhat tmp]# cd vmware-tools-distrib
    [root@Redhat vmware-tools-distrib]#./vmware-install.pl

    image

    遇到选项,直接回车即可。

    step 5:安装结束,重启虚拟机

    [root@rac1 ~]# reboot

    step6:测试VM Tools安装是否成功

    从主机拖到一个文档到虚拟机,如果拖动成功,说明VM Tools已经安装成功

    (三)操作系统配置

    (3.1)网络配置

    (3.1.1)主机名配置

    ①节点1:

    [root@rac1 ~]# vim /etc/sysconfig/network
     
    NETWORKING=yes
    NETWORKING_IPV6=no
    HOSTNAME=rac1

    ②节点2:

    [root@rac2 ~]# vim /etc/sysconfig/network
    
    NETWORKING=yes
    NETWORKING_IPV6=no
    HOSTNAME=rac2

    (3.1.2)IP配置

    为了使用静态IP配置数据库,我又为每台虚拟机新增加了一块网卡,将其配置成only-host模式,新增网卡方式如下:

    step1:添加网络

    点击vmware的“编辑”-> “虚拟网络编辑器”-> “更改设置”-> “添加网络”,按下图选择,保存

    image

    step2:在两台虚拟机上添加网卡

    选择虚拟机,“设置”->“添加”->“网络适配器”,选择“自定义”,这个自定义是我们上一步定义的网络,结果如下:

    image

    step3:根据网络设置,我们规划IP地址如下:

    image

    接下来就是配置IP地址了,对于节点1(主机名:rac1),我们:

    ①配置eth1

    --或删除BOOTPROTO

    --不要修改硬件地址

    --设置网卡为开机启动

    --增加IP和MASK

    [root@rac1 ~]# vim /etc/sysconfig/network-scripts/ifcfg-eth1
    #修改下面红色部分
     
    # Advanced Micro Devices [AMD] 79c970 [PCnet32 LANCE]
    DEVICE=eth1
    # BOOTPROTO
    =
    dhcp
    HWADDR=00:0C:29:9C:DF:6A
    ONBOOT
    =yes IPADDR=192.168.19.10
    
    NETMASK
    =255.255.255.0

    ②配置eth2

    [root@rac1 ~]# vim /etc/sysconfig/network-scripts/ifcfg-eth2
    # 修改红色部分
    
    # Advanced Micro Devices [AMD] 79c970 [PCnet32 LANCE]
    DEVICE=eth2
    # BOOTPROTO
    =
    dhcp
    ONBOOT
    =
    yes
    HWADDR
    =00:0C:29
    :6G:8C:5F
    IPADDR
    =192.168.
    15.10
    NETMASK
    =255.255.255.0

    对于节点2(主机名:rac2),我们参照节点1即可:

    ①配置eth1

    [root@rac2 ~]# vim /etc/sysconfig/network-scripts/ifcfg-eth1
    
    # Advanced Micro Devices [AMD] 79c970 [PCnet32 LANCE]
    DEVICE=eth1
    # BOOTPROTO
    =dhcp ONBOOT=
    yes
    HWADDR=00:0c:29:b0:4e:b6
    IPADDR
    =
    192.168
    .
    19.11
    NETMASK
    =255.255.
    255.0

    ②配置eth2

    [root@rac2 ~]# vim /etc/sysconfig/network-scripts/ifcfg-eth2
    
    # Advanced Micro Devices [AMD] 79c970 [PCnet32 LANCE]
    DEVICE=eth2
    # BOOTPROTO
    =
    dhcp
    ONBOOT
    =
    yes
    HWADDR=00:0c:29:b0:4e:c0
    IPADDR
    =192.168.
    15.11
    NETMASK
    =255.255.
    255.0

    (3.1.3)hosts文件配置

    在2个节点上配置hosts文件,以节点1为例

    [root@rac1 ~]# vim /etc/hosts
    # 在文件的最后面加上
    
    #eth1 public
    192.168.19.10 rac1
    192.168.19.11 rac2
    #virtual
    192.168.19.12 rac1-vip
    192.168.19.13 rac2-vip
    #eth2 private
    192.168.15.10 rac1-priv
    192.168.15.11 rac2-priv
    #scan
    192.168.19.14 rac-scan

    配置完成后,重启网卡

    [root@rac1 ~]# service network restart

    重启网卡时,遇到了一个小错误,提示:

    Device eth2 has different MAC address than expected, ignoring.
    [FAILED]

    看了其他人的文章,发现是网卡配置文件里面的MAC地址与实际虚拟机的MAC地址不一样,解决办法如下:

    step1:查看本机实际MAC地址(红色部分)

    [root@rac1 ~]# ifconfig eth2
    eth2      Link encap:Ethernet  HWaddr 00:0C:29:9C:DF:7E  
              inet addr:192.168.15.10  Bcast:192.168.15.255  Mask:255.255.255.0
              inet6 addr: fe80::20c:29ff:fe9c:df7e/64 Scope:Link
              UP BROADCAST RUNNING MULTICAST  MTU:1500  Metric:1
              RX packets:30677 errors:0 dropped:0 overruns:0 frame:0
              TX packets:26377 errors:0 dropped:0 overruns:0 carrier:0
              collisions:0 txqueuelen:1000 
              RX bytes:15839769 (15.1 MiB)  TX bytes:10819637 (10.3 MiB)
              Interrupt:83 Base address:0x2824

    step2:查看我们配置的MAC地址

    [root@rac1 ~]# vim /etc/sysconfig/network-scripts/ifcfg-eth2
    # 修改红色部分
    
    # Advanced Micro Devices [AMD] 79c970 [PCnet32 LANCE]
    DEVICE=eth2
    # BOOTPROTO=dhcp
    ONBOOT=yes
    HWADDR=00:0C:29:6G:8C:5F
    IPADDR=192.168.15.10
    NETMASK=255.255.255.0

    发现我们配置的MAC地址与实际的MAC地址不一样,修改网卡step2里面的MAC地址,重启网卡。

    (3.1.4)测试网络

    测试主机与虚拟机、虚拟机与虚拟机之间通讯是否正常

    ①在主机上执行

    ping 192.168.15.10
    ping 192.168.15.11
    ping 192.168.19.10
    ping 192.168.19.11

    ②在节点1上执行,其中192.168.15.1与192.168.19.1均为主机ip

    ping 192.168.15.1
    ping 192.168.15.11
    ping 192.168.19.1
    ping 192.168.19.11

    ③在节点2执行

    ping 192.168.15.1
    ping 192.168.15.10
    ping 192.168.19.1
    ping 192.168.19.10

    如果都能够ping通,说明网络配置正确。

    (3.2)共享存储配置

    (3.2.1)存储规划

    磁盘功能 个数 每个大小(GB)
    OCR 3 1
    DATA 3 10
    ARCHIVE 1 10

    (3.2.1)创建共享存储

    step1:以管理员方式打开DOS窗口,到vmware workstation安装目录下,执行以下命令创建共享磁盘:

    vmware-vdiskmanager.exe -c -s 1000MB -a lsilogic -t 2 "F:vmwareVirtueMachineRACShareDiskvotingdisk1.vmdk"
    vmware-vdiskmanager.exe -c -s 1000Mb -a lsilogic -t 2 "F:vmwareVirtueMachineRACShareDiskvotingdisk2.vmdk"
    vmware-vdiskmanager.exe -c -s 1000Mb -a lsilogic -t 2 "F:vmwareVirtueMachineRACShareDiskvotingdisk3.vmdk"
    vmware-vdiskmanager.exe -c -s 10000Mb -a lsilogic -t 2 "F:vmwareVirtueMachineRACShareDiskdatadisk1.vmdk"
    vmware-vdiskmanager.exe -c -s 10000Mb -a lsilogic -t 2 "F:vmwareVirtueMachineRACShareDiskdatadisk2.vmdk"
    vmware-vdiskmanager.exe -c -s 10000Mb -a lsilogic -t 2 "F:vmwareVirtueMachineRACShareDiskdatadisk3.vmdk"
    vmware-vdiskmanager.exe -c -s 10000Mb -a lsilogic -t 2 "F:vmwareVirtueMachineRACShareDiskarchdisk1.vmdk"

    (3)修改2个虚拟机的vmx配置文件,在虚拟机的安装目录下的.vmx文件最后加上以下信息:

    scsi1.present = "TRUE"
    scsi1.virtualDev = "lsilogic"
    scsi1.sharedBus = "virtual"
    
    scsi1:1.present = "TRUE"
    scsi1:1.mode = "independent-persistent"
    scsi1:1.filename = "F:vmwareVirtueMachineRACShareDiskvotingdisk1.vmdk"
    scsi1:1.deviceType = "plainDisk"
    
    scsi1:2.present = "TRUE"
    scsi1:2.mode = "independent-persistent"
    scsi1:2.filename = "F:vmwareVirtueMachineRACShareDiskvotingdisk2.vmdk"
    scsi1:2.deviceType = "plainDisk" 
    
    scsi1:3.present = "TRUE"
    scsi1:3.mode = "independent-persistent"
    scsi1:3.filename = "F:vmwareVirtueMachineRACShareDiskvotingdisk3.vmdk"
    scsi1:3.deviceType = "plainDisk"
    
    scsi1:4.present = "TRUE"
    scsi1:4.mode = "independent-persistent"
    scsi1:4.filename = "F:vmwareVirtueMachineRACShareDiskdatadisk1.vmdk"
    scsi1:4.deviceType = "plainDisk"
    
    scsi1:5.present = "TRUE"
    scsi1:5.mode = "independent-persistent"
    scsi1:5.filename = "F:vmwareVirtueMachineRACShareDiskdatadisk2.vmdk"
    scsi1:5.deviceType = "plainDisk"
    
    scsi1:6.present = "TRUE"
    scsi1:6.mode = "independent-persistent"
    scsi1:6.filename = "F:vmwareVirtueMachineRACShareDiskdatadisk3.vmdk"
    scsi1:6.deviceType = "plainDisk"
    
    scsi1:7.present = "TRUE"
    scsi1:7.mode = "independent-persistent"
    scsi1:7.filename = "F:vmwareVirtueMachineRACShareDiskarchdisk1.vmdk"
    scsi1:7.deviceType = "plainDisk"
    
    disk.locking = "false"
    diskLib.dataCacheMaxSize = "0"
    diskLib.dataCacheMaxReadAheadSize = "0"
    diskLib.DataCacheMinReadAheadSize = "0"
    diskLib.dataCachePageSize = "4096"
    diskLib.maxUnsyncedWrites = "0"

    重启虚拟机,2个节点都重启

    [root@rac1 ~]# reboot

    最终我们看到的虚拟机的硬件配置大致是这样的

    image

    (四)Oracle & Grid预配置

    (4.1)创建用户

    在2个节点上创建用户及安装目录:

    /usr/sbin/groupadd -g 1000 oinstall
    /usr/sbin/groupadd -g 1020 asmadmin
    /usr/sbin/groupadd -g 1021 asmdba
    /usr/sbin/groupadd -g 1022 asmoper
    /usr/sbin/groupadd -g 1031 dba
    /usr/sbin/groupadd -g 1032 oper
    useradd -u 1100 -g oinstall -G asmadmin,asmdba,asmoper,oper,dba grid
    useradd -u 1101 -g oinstall -G dba,asmdba,oper oracle
    mkdir -p /u01/app/11.2.0/grid
    mkdir -p /u01/app/grid
    mkdir /u01/app/oracle
    chown -R grid:oinstall /u01
    chown oracle:oinstall /u01/app/oracle
    chmod -R 775 /u01/

    安装目录结构如下:
    [root@rac1 u01]# tree
    .
    `-- app
        |-- 11.2.0
        |   `-- grid
        |-- grid
        `—oracle

    (4.2)配置环境变量

    (4.2.1)配置ORACLE的环境变量,ORACLE_SID需根据主机来填写

    [oracle@rac1 ~]$ cd
    [oracle@rac1 ~]$ vi .bash_profile

    #添加下面的部分
     
    export TMP=/tmp 
    export TMPDIR=$TMP 
    export ORACLE_SID=rac1     
    # 如果是节点2,则: export ORACLE_SID=
    rac2 
    export ORACLE_BASE=/u01/app/oracle 
    export ORACLE_HOME=$ORACLE_BASE/product/11.2.0/db_1
    export TNS_ADMIN=$ORACLE_HOME/network/admin 
    export PATH=/usr/sbin:$PATH 
    export PATH=$ORACLE_HOME/bin:$PATH 
    export LD_LIBRARY_PATH=$ORACLE_HOME/lib:/lib:/usr/lib 
    export CLASSPATH=$ORACLE_HOME/JRE:$ORACLE_HOME/jlib:$ORACLE_HOME/rdbms/jlib 
    umask 022

    (4.2.1)配置GRID的环境变量,ORACLE_SID需根据主机来填写

    [grid@rac1 ~]$ cd 
    [grid@rac1 ~]$ vi .bash_profile 
    # 添加下面的部分
    
    export TMP=/tmp 
    export TMPDIR=$TMP 
    export ORACLE_SID=+asm1  # 如果是节点2,则:export ORACLE_SID=+asm2 
    export ORACLE_BASE=/u01/app/grid
    export ORACLE_HOME=/u01/app/11.2.0/grid
    export PATH=/usr/sbin:$PATH 
    export PATH=$ORACLE_HOME/bin:$PATH 
    export LD_LIBRARY_PATH=$ORACLE_HOME/lib:/lib:/usr/lib 
    export CLASSPATH=$ORACLE_HOME/JRE:$ORACLE_HOME/jlib:$ORACLE_HOME/rdbms/jlib 
    umask 022

    (4.3)配置节点互信

    (4.3.1)配置oracle用户的节点互信

    step1:2个节点执行

    su - oracle
    
    mkdir ~/.ssh
    chmod 755 .ssh 
    /usr/bin/ssh-keygen -t rsa
    /usr/bin/ssh-keygen -t dsa

    step2:将所有的key文件汇总到一个总的认证文件中,在节点1上执行

    su – oracle
     
    ssh rac1 cat ~/.ssh/id_rsa.pub >> authorized_keys
    ssh rac2 cat ~/.ssh/id_rsa.pub >> authorized_keys
    ssh rac1 cat ~/.ssh/id_dsa.pub >> authorized_keys
    ssh rac2 cat ~/.ssh/id_dsa.pub >> authorized_keys

    step3:节点1上存在一份完整的key,拷贝到节点2

    [oracle@rac1 ~]$ cd ~/.ssh/
    [oracle@rac1 .ssh]$ scp authorized_keys rac2:~/.ssh/
    [oracle@rac2 .ssh]chmod 600 authorized_keys

    step4:在2个节点都执行一下命令

    ssh rac1 date
    ssh rac2 date
    ssh rac1-priv date
    ssh rac2-priv date

    step5:检验是否配置成功,在节点1上不用输入密码就可以通过ssh连接节点2,说明配置成功

    [oracle@rac1 ~]$ ssh rac2
    [oracle@rac2 ~]$ id
    uid=1101(oracle) gid=1000(oinstall) groups=1000(oinstall),1021(asmdba),1031(dba),1032(oper)
    [oracle@rac2 ~]$ hostname
    rac2

    (4.3.2)配置grid用户的节点互信

    与Oracle用户执行方式相同,只是执行脚本的用户变成了grid。

    (4.4)配置系统内核参数

    (4.4.1)内核参数设置

    [root@rac1 ~]# vi /etc/sysctl.conf 
    # 在末尾添加
     
    kernel.msgmnb = 65536 
    kernel.msgmax = 65536 
    kernel.shmmax = 68719476736 
    kernel.shmall = 4294967296 
    fs.aio-max-nr = 1048576 
    fs.file-max = 6815744 
    kernel.shmall = 2097152 
    kernel.shmmax = 2002012160 
    kernel.shmmni = 4096 
    kernel.sem = 250 32000 100 128 
    net.ipv4.ip_local_port_range = 9000 65500 
    net.core.rmem_default = 262144 
    net.core.rmem_max = 4194304 
    net.core.wmem_default = 262144 
    net.core.wmem_max = 1048586 
    net.ipv4.tcp_wmem = 262144 262144 262144 
    net.ipv4.tcp_rmem = 4194304 4194304 4194304

    参数说明:

    kernel.shmmni:整个系统共享内存段的最大数目

    fs.file-max:系统中所允许的文件句柄最大数目

    net.core.rmem_default:套接字接收缓冲区大小的缺省值

    net.core.rmem_max:套接字接收缓冲区大小的最大值

    net.core.wmem_default:套接字发送缓冲区大小的缺省值

    net.core.wmem_max:套接字发送缓冲区大小的最大值

    net.ipv4.ip_local_port_range:应用程序可使用的IPv4端口范围

    使得修改内核参数生效

    [root@rac1 ~]# sysctl -p

    (4.4.2)配置oracle、grid用户的shell限制

    [root@rac1 ~]# vi /etc/security/limits.conf 
    # 在末尾添加
     
    grid soft nproc 2047 
    grid hard nproc 16384 
    grid soft nofile 1024 
    grid hard nofile 65536 
    oracle soft nproc 2047 
    oracle hard nproc 16384 
    oracle soft nofile 1024 
    oracle hard nofile 65536

    (4.5)配置共享存储

    step1:查看磁盘

    [root@rac1 ~]# fdisk -l

    通过该命令,可看到我们后面添加上去的磁盘信息,磁盘从sdb到sdh

    step2:分区、格式化磁盘。由于是共享磁盘,所以只需要在一个节点上执行即可

    # 在节点1上格式化,以/dev/sdb为例:
    
    [root@rac1 ~]# fdisk /dev/sdb
    
    The number of cylinders for this disk is set to 3824.
    There is nothing wrong with that, but this is larger than 1024,
    and could in certain setups cause problems with:
    1) software that runs at boot time (e.g., old versions of LILO)
    2) booting and partitioning software from other OSs
       (e.g., DOS FDISK, OS/2 FDISK)
    
    Command (m for help): n
    Command action
       e   extended
       p   primary partition (1-4)
    p
    Partition number (1-4): 1
    First cylinder (1-3824, default 1): 
    Using default value 1
    Last cylinder or +size or +sizeM or +sizeK (1-3824, default 3824): 
    Command (m for help): w
    The partition table has been altered!
    
    Calling ioctl() to re-read partition table.
    Syncing disks.

    step3:在2个节点上添加裸设备,2个节点都要执行

    [root@rac2 ~]# vi /etc/udev/rules.d/60-raw.rules
     # 在后面添加
    
    ACTION=="add", KERNEL=="sdb1", RUN+="/bin/raw /dev/raw/raw1 %N"
    ACTION=="add", KERNEL=="sdc1", RUN+="/bin/raw /dev/raw/raw2 %N"
    ACTION=="add", KERNEL=="sdd1", RUN+="/bin/raw /dev/raw/raw3 %N"
    ACTION=="add", KERNEL=="sde1", RUN+="/bin/raw /dev/raw/raw4 %N"
    ACTION=="add", KERNEL=="sdf1", RUN+="/bin/raw /dev/raw/raw5 %N"
    ACTION=="add", KERNEL=="sdg1", RUN+="/bin/raw /dev/raw/raw6 %N"
    ACTION=="add", KERNEL=="sdh1", RUN+="/bin/raw /dev/raw/raw7 %N"
    
    
    KERNEL=="raw[1]", MODE="0660", OWNER="grid", GROUP="asmadmin"
    KERNEL=="raw[2]", MODE="0660", OWNER="grid", GROUP="asmadmin"
    KERNEL=="raw[3]", MODE="0660", OWNER="grid", GROUP="asmadmin"
    KERNEL=="raw[4]", MODE="0660", OWNER="grid", GROUP="asmadmin"
    KERNEL=="raw[5]", MODE="0660", OWNER="grid", GROUP="asmadmin"
    KERNEL=="raw[6]", MODE="0660", OWNER="grid", GROUP="asmadmin"
    KERNEL=="raw[7]", MODE="0660", OWNER="grid", GROUP="asmadmin"

    启动裸设备,2个节点都执行

    [root@rac2 ~]# start_udev

    查看裸设备,2个节点都要查看

    [root@rac1 ~]# raw -qa
    /dev/raw/raw1:    bound to major 8, minor 17
    /dev/raw/raw2:    bound to major 8, minor 33
    /dev/raw/raw3:    bound to major 8, minor 49
    /dev/raw/raw4:    bound to major 8, minor 65
    /dev/raw/raw5:    bound to major 8, minor 81
    /dev/raw/raw6:    bound to major 8, minor 97
    /dev/raw/raw7:    bound to major 8, minor 113

    (4.6)安装依赖软件包

    (4.6.1)配置共享文件夹(2个节点都要配置)

    step1:关闭虚拟机

    step2: 设置”->“选项”->“共享文件夹”->“总是启用”->“添加 ”。然后添加操作系统安装包,为了方便,我们将grid、oracle安装包也添加进去。

    step3:查看结果

    [root@rac1 ~]# cd /mnt/hgfs/soft/
    [root@rac1 soft]# ls
    database  linux_11gR2_database_1of2.zip  Linux.5.5.for.x86.rhel-server-5.5-i386-dvd.iso
    grid      linux_11gR2_database_2of2.zip

    (4.6.2)配置yum源

    step 1:创建ISO文件的挂载目录并挂载

    [root@rac1 /]# mkdir /yum
    [root@rac1 /]# cd /mnt/hgfs/soft/
    [root@rac1 soft]# mount -o loop Linux.5.5.for.x86.rhel-server-5.5-i386-dvd.iso /yum

    step2:找到rpm包

    [root@rac1 Server]# cd /yum/Server/
    [root@rac1 Server]# ll
    ...
    -r--r--r-- 433 root root   102853 Jan 19  2007 zlib-devel-1.2.3-3.i386.rpm
    -r--r--r-- 368 root root  1787914 May  6  2009 zsh-4.2.6-3.el5.i386.rpm
    -r--r--r-- 368 root root   369147 May  6  2009 zsh-html-4.2.6-3.el5.i386.rpm
    [root@rac1 Server]# ll |wc -l
    2351

    step 3:配置Linux OS介质为yum源

    [root@rac1 ~]# cd /etc/yum.repos.d/
    [root@rac1 yum.repos.d]# ls
    rhel-source.repo
    [root@RedHat yum.repos.d]# rm -rf redhat.repo 
    [root@RedHat yum.repos.d]# vi lijiaman.repo
    
    # 在新的yum源里面添加红色部分
    [dvdinfo]
    name=my_yum
    baseurl=file:///yum/Server
    enabled=1
    gpgcheck=0

    step4:收集安装包的信息

    [root@rac1 yum.repos.d]# yum makecache

    step5:为方便重启虚拟机之后依然可以使用yum,在rc.local里面添加:

    [root@rac1 yum.repos.d]# vim /etc/rc.local
    #添加红色部分
    
    
    mount -o loop Linux.5.5.for.x86.rhel-server-5.5-i386-dvd.iso /yum

    (4.6.3)安装依赖包

    不同的操作系统安装不同的包,具体可看官方教程:

    http://docs.oracle.com/cd/E11882_01/install.112/e41961/prelinux.htm#CWLIN

    我这里安装的包如下:

    yum install -y binutils-*
    yum install -y compat-libstdc++-*
    yum install -y elfutils-libelf-*
    yum install -y elfutils-libelf-*
    yum install -y elfutils-libelf-devel-static-0.125
    yum install -y gcc-4.1.2
    yum install -y gcc-c++-4.1.2
    yum install -y glibc-2.5-24
    yum install -y glibc-common-2.5
    yum install -y glibc-devel-2.5
    yum install -y glibc-headers-2.5
    yum install -y kernel-headers-2.6.18
    yum install -y ksh-20060214
    yum install -y libaio-0.3.106
    yum install -y libaio-devel-0.3.106 
    yum install -y libgcc-4.1.2
    yum install -y libgomp-4.1.2
    yum install -y libstdc++-4.1.2 
    yum install -y libstdc++-devel-4.1.2
    yum install -y make-3.81
    yum install -y sysstat-7.0.2

    (五)安装Grid

    (5.1)安装前的检查

    step1:解压grid安装包

    [root@rac1 ~]# cd /mnt/hgfs/soft/
    [root@rac1 soft]# unzip linux_11gR2_grid.zip

    step2:安装cvuqdisk包,用于检查RAC环境配置

    [root@rac1 soft]# cd grid/
    [root@rac1 grid]# ls
    doc  install  response  rpm  runcluvfy.sh  runInstaller  sshsetup  stage  welcome.html
    [root@rac1 grid]# cd rpm/
    [root@rac1 rpm]# ls
    cvuqdisk-1.0.7-1.rpm
    [root@rac1 rpm]# rpm -ivh cvuqdisk-1.0.7-1.rpm 
    Preparing...                ########################################### [100%]
    Using default group oinstall to install package
       1:cvuqdisk               ########################################### [100%]

    step3:检查环境配置,这里需要一项一项的去核对,确保没有不通过的选项,

    [root@rac1 rpm]# su - grid
    [grid@rac1 ~]$ cd /mnt/hgfs/soft/grid/
    [grid@rac1 grid]$ ls
    doc  install  response  rpm  runcluvfy.sh  runInstaller  sshsetup  stage  welcome.html
    [grid@rac1 grid]$ ./runcluvfy.sh stage -pre crsinst -n rac1,rac2 -fixup -verbose
    
    Performing pre-checks for cluster services setup 
    
    Checking node reachability...

    这里我的检查中有3项目不合格:

    ①用户等效性:

    Checking user equivalence...

    Check: User equivalence for user "grid"
      Node Name                             Comment                
      ------------------------------------  ------------------------
      rac2                                  failed                 
      rac1                                  failed                 
    Result: PRVF-4007 : User equivalence check failed for user "grid"

    解决方法:

    安装前面的方法,重新配置grid用户互信。

    ②补丁包

    Check: Package existence for "unixODBC-2.2.11"
      Node Name     Available                 Required                  Comment  
      ------------  ------------------------  ------------------------  ----------
      rac2          missing                   unixODBC-2.2.11           failed   
    Result: Package existence check failed for "unixODBC-2.2.11"

    Check: Package existence for "unixODBC-devel-2.2.11"
      Node Name     Available                 Required                  Comment  
      ------------  ------------------------  ------------------------  ----------
      rac2          missing                   unixODBC-devel-2.2.11     failed   
    Result: Package existence check failed for "unixODBC-devel-2.2.11"

    解决方法:

    yum install –y unixODBC*

    ③NTP时钟同步

    Starting Clock synchronization checks using Network Time Protocol(NTP)...

    NTP Configuration file check started...
    The NTP configuration file "/etc/ntp.conf" is available on all nodes
    NTP Configuration file check passed

    Checking daemon liveness...

    Check: Liveness for "ntpd"
      Node Name                             Running?               
      ------------------------------------  ------------------------
      rac2                                  no                     
      rac1                                  no                     
    Result: Liveness check failed for "ntpd"
    PRVF-5415 : Check to see if NTP daemon is running failed
    Result: Clock synchronization check using Network Time Protocol(NTP) failed

    解决方法:

    在11gr2 版本的Rac 在安装 GI软件时,需要通过2种途径实现节点间时间同步:

    1.操作系统 NTP 时间同步 

    2.Oracle Cluster Time Synchronization Service

    在oracle11G R2版本中,oracle自己增加了集群中的时间同步服务,如果操作系统没有配置NTP,就会启用oracle 自带的 Cluster Time Synchronization Service 。

    故该错误可以忽略。

    (5.2)安装grid

    安装主需要在一个节点上进行即可,会自动同步到另一个节点,我们在rac1上执行安装工作

    step1:开启安装

    [root@rac1 ~]# xhost +
    xhost:  unable to open display ""
    [root@rac1 ~]# su - grid
    [grid@rac1 ~]$ cd /mnt/hgfs/soft/grid/
    [grid@rac1 grid]$ ls
    doc  install  response  rpm  runcluvfy.sh  runInstaller  sshsetup  stage  welcome.html
    [grid@rac1 grid]$ ./runInstaller

    step2:图形安装

    这里写图片描述

    这里写图片描述

    image

    image

    这一步需要注意SCAN Name,我们在/etc/hosts文件里面设置为什么值,这里就填什么值

    image

    添加另一个节点

    image

    添加节点之后,点击“SSH Connectivity”,填写grid用户的密码,点击“setup”

    image

    配置网络,这里至少需要两个网络,根据前面的IP规划表选择即可

    image

    image

    选择3块1GB的磁盘,组成一个磁盘组OCR

    image

    输入密码,这里有警告信息,看了一下,是密码不够复杂,不符合Oracle的规定,忽略不管

    image

    image

    image

    下面的几个图忘记截图了,从网上找了一张图,在Prerequisite Check中,NTP检查失败,忽略

    image

    开始进行安装…

    image

    安装过程中,会填出如下窗口:

    image

    按照其要求,一个节点一个节点的执行,千万不要2个节点一起执行,以我为例:

    step1:在虚拟机rac1上打开一个窗口,以root用户登录

    执行脚本:

    [root@rac1 ~]# /u01/app/oraInventory/orainstRoot.sh 
    
    # 待上面脚本执行完成,再执行下面脚本
    [root@rac1 ~]# /u01/app/11.2.0/grid/root.sh

    step2:漫长的等待

    step3:确认节点1上的脚本执行完成,在虚拟机rac2上打开一个窗口,以root用户登录

    执行脚本:

    [root@rac2 ~]# /u01/app/oraInventory/orainstRoot.sh 
    
    # 待上面脚本执行完成,再执行下面脚本
    [root@rac2 ~]# /u01/app/11.2.0/grid/root.sh

    step4:漫长的等待

    step5:确认节点2的脚本执行完成。点击“OK”;

    ===================================================

    安装到最后,有报错信息出现:

    image

    原因:

    由于配置了/etc/hosts来解析SCAN,导致未走DNS来进行SCAN的解析,爆出此错误,可以考虑忽略掉
    解决办法:
    ping scan ip,如果可以ping通,忽略,跳过即可。

    [grid@rac1 grid]$ ping rac-scan
    PING rac-scan (192.168.19.14) 56(84) bytes of data.
    64 bytes from rac-scan (192.168.19.14): icmp_seq=1 ttl=64 time=0.021 ms
    64 bytes from rac-scan (192.168.19.14): icmp_seq=2 ttl=64 time=0.026 ms

    至此grid安装完成。

    (5.3)配置磁盘组

    在节点1上执行

    [grid@rac1 ~]$ asmca

    点击“create”

    image

    创建磁盘组DATA

    image

    创建磁盘组ARCH

    image

    最终结果:
    image

    (六)安装Oracle

    (6.1)安装Oracle软件

    解压安装包,调出图形界面

    [oracle@rac1 ~]$ cd /mnt/hgfs/soft/
    [oracle@rac1 soft]$ unzip linux_11gR2_database_1of2.zip 
    [oracle@rac1 soft]$ unzip linux_11gR2_database_2of2.zip 
    [oracle@rac1 soft]$ ls
    database  linux_11gR2_database_1of2.zip  Linux.5.5.for.x86.rhel-server-5.5-i386-dvd.iso
    grid      linux_11gR2_database_2of2.zip
    [oracle@rac1 soft]$ cd database/
    [oracle@rac1 database]$ ls
    doc  install  response  rpm  runInstaller  sshsetup  stage  welcome.html
    [oracle@rac1 database]$ ./runInstall

    image

    选择第二个,只安装数据库软件

    image

    这里节点互信总是出问题,重新手动配置一下就好了,没完全搞明白 -_-

    image

    image

    时钟同步问题,忽略

    image

    漫长的等待

    image

    结束。

    (6.2)创建数据库

    在oracle用户下执行命令

    [oracle@rac1 ~]$ dbca

    部分截图:

    image

    image

    image

    image

    image

    image

    image

    image

    image

    image

    image

    与安装grid一样,执行脚本,一个节点一个节点的执行,先执行完节点1,在执行节点2

    这里写图片描述

    结束安装。

    (6.3)检查安装结果

    发现arch和data磁盘组还是有问题,没有自动启动(这个问题比较奇怪,我安装了3次,在个人电脑上安装遇到2次,在服务器上安装则没有这个问题,怀疑是个人电脑I/O不足,在数据库启动时磁盘mount超时了)

    [grid@rac1 ~]$ crsctl status res -t
    --------------------------------------------------------------------------------
    NAME           TARGET  STATE        SERVER                   STATE_DETAILS       
    --------------------------------------------------------------------------------
    Local Resources
    --------------------------------------------------------------------------------
    ora.ARCH.dg
                   OFFLINE OFFLINE      rac1                                         
                   ONLINE  ONLINE       rac2                                         
    ora.DATA.dg
                   OFFLINE OFFLINE      rac1                                         
                   OFFLINE OFFLINE      rac2                                         
    ora.LISTENER.lsnr
                   ONLINE  ONLINE       rac1                                         
                   ONLINE  ONLINE       rac2                                         
    ora.OCR.dg
                   ONLINE  ONLINE       rac1                                         
                   ONLINE  ONLINE       rac2                                         
    ora.asm
                   ONLINE  ONLINE       rac1                     Started             
                   ONLINE  ONLINE       rac2                     Started             
    ora.eons
                   ONLINE  ONLINE       rac1                                         
                   ONLINE  ONLINE       rac2                                         
    ora.gsd
                   OFFLINE OFFLINE      rac1                                         
                   OFFLINE OFFLINE      rac2                                         
    ora.net1.network
                   ONLINE  ONLINE       rac1                                         
                   ONLINE  ONLINE       rac2                                         
    ora.ons
                   ONLINE  ONLINE       rac1                                         
                   ONLINE  ONLINE       rac2                                         
    ora.registry.acfs
                   ONLINE  ONLINE       rac1                                         
                   ONLINE  ONLINE       rac2                                         
    --------------------------------------------------------------------------------
    Cluster Resources
    --------------------------------------------------------------------------------
    ora.LISTENER_SCAN1.lsnr
          1        ONLINE  ONLINE       rac1                                         
    ora.oc4j
          1        OFFLINE OFFLINE                                                   
    ora.rac.db
          1        OFFLINE OFFLINE                                                   
          2        OFFLINE OFFLINE                                                   
    ora.rac1.vip
          1        ONLINE  ONLINE       rac1                                         
    ora.rac2.vip
          1        ONLINE  ONLINE       rac2                                         
    ora.scan1.vip
          1        ONLINE  ONLINE       rac1

    手动启动arch和data

    [grid@rac1 ~]$ srvctl start diskgroup -g arch
    [grid@rac1 ~]$ srvctl start diskgroup -g data

    再次查看资源状态,已经启动

    [grid@rac1 ~]$ crsctl status res -t
    --------------------------------------------------------------------------------
    NAME           TARGET  STATE        SERVER                   STATE_DETAILS       
    --------------------------------------------------------------------------------
    Local Resources
    --------------------------------------------------------------------------------
    ora.ARCH.dg
                   ONLINE  ONLINE       rac1                                         
                   ONLINE  ONLINE       rac2                                         
    ora.DATA.dg
                   ONLINE  ONLINE       rac1                                         
                   ONLINE  ONLINE       rac2

    接下来,启动数据库

    [grid@rac1 ~]$ srvctl start database -d rac

    查看资源状态,数据库已经启动

    ora.rac.db
          1        ONLINE  ONLINE       rac1                     Open                
          2        ONLINE  ONLINE       rac2                     Open

    进入数据库,

    [grid@rac1 ~]$ su - oracle
    Password: 
    [oracle@rac1 ~]$ sqlplus / as sysdba
    
    SQL*Plus: Release 11.2.0.1.0 Production on Sun Oct 15 19:27:40 2017
    
    Copyright (c) 1982, 2009, Oracle.  All rights reserved.
    
    
    Connected to:
    Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 - Production
    With the Partitioning, Real Application Clusters, Automatic Storage Management, OLAP,
    Data Mining and Real Application Testing options

    查看实例状态

    SQL> col host_name format a10;
    SQL> select host_name,inst_id,instance_name,version from gv$instance;
    
    HOST_NAME     INST_ID INSTANCE_NAME    VERSION
    ---------- ---------- ---------------- -----------------
    rac1            1 rac1           11.2.0.1.0
    rac2            2 rac2           11.2.0.1.0

    一切正常。结束

    其它问题:

    问题1:无法进入asm实例

    在grid下,想要进入asm实例,发现asm实例状态为idle,如下:

    [grid@rac1 ~]$ sqlplus / as sysasm

    SQL*Plus: Release 11.2.0.1.0 Production on Tue Sep 11 18:43:36 2018

    Copyright (c) 1982, 2009, Oracle. All rights reserved.

    Connected to an idle instance.

    [grid@rac1 ~]$ asmcmd
    Connected to an idle instance.
    ASMCMD>

    通过crsctl查看,我的asm实例明明是启动的,那么为什么状态为idle呢?从网上查找资料,发现都是grid用户环境变量的问题。查看监听,确认asm实例名称,发现asm实例名称是大写的,在前面安装grid的时候没注意,导致出现该问题。

    [grid@rac2 ~]$ lsnrctl

    LSNRCTL for Linux: Version 11.2.0.1.0 - Production on 11-SEP-2018 19:20:42

    Copyright (c) 1991, 2009, Oracle. All rights reserved.

    Welcome to LSNRCTL, type "help" for information.

    LSNRCTL> status
    Connecting to (DESCRIPTION=(ADDRESS=(PROTOCOL=IPC)(KEY=LISTENER)))
    STATUS of the LISTENER
    ------------------------
    Alias LISTENER
    Version TNSLSNR for Linux: Version 11.2.0.1.0 - Production
    Start Date 11-SEP-2018 18:18:12
    Uptime 0 days 1 hr. 2 min. 31 sec
    Trace Level off
    Security ON: Local OS Authentication
    SNMP OFF
    Listener Parameter File /u01/app/11.2.0/grid/network/admin/listener.ora
    Listener Log File /u01/app/grid/diag/tnslsnr/rac2/listener/alert/log.xml
    Listening Endpoints Summary...
    (DESCRIPTION=(ADDRESS=(PROTOCOL=ipc)(KEY=LISTENER)))
    (DESCRIPTION=(ADDRESS=(PROTOCOL=tcp)(HOST=10.141.71.102)(PORT=1521)))
    (DESCRIPTION=(ADDRESS=(PROTOCOL=tcp)(HOST=10.141.71.104)(PORT=1521)))
    Services Summary...
    Service "+ASM" has 1 instance(s).
    Instance "+ASM2", status READY, has 1 handler(s) for this service...
    Service "rac" has 1 instance(s).
    Instance "rac2", status READY, has 1 handler(s) for this service...
    Service "racXDB" has 1 instance(s).
    Instance "rac2", status READY, has 1 handler(s) for this service...
    The command completed successfully

    于是将grid用户的.bash_profile文件中的oracle_sid修改为大写即可

    修改前:

    [grid@rac1 ~]$ more .bash_profile 
    # .bash_profile
    
    # Get the aliases and functions
    if [ -f ~/.bashrc ]; then
        . ~/.bashrc
    fi
    
    # User specific environment and startup programs
    
    PATH=$PATH:$HOME/bin
    export TMP=/tmp 
    export TMPDIR=$TMP 
    export ORACLE_SID=+asm1  # 如果是节点2,则:export ORACLE_SID=+asm2 
    export ORACLE_BASE=/u01/app/grid
    export ORACLE_HOME=/u01/app/11.2.0/grid
    export PATH=/usr/sbin:$PATH 
    export PATH=$ORACLE_HOME/bin:$PATH 
    export LD_LIBRARY_PATH=$ORACLE_HOME/lib:/lib:/usr/lib 
    export CLASSPATH=$ORACLE_HOME/JRE:$ORACLE_HOME/jlib:$ORACLE_HOME/rdbms/jlib 
    umask 022

    修改后:

    [grid@rac1 ~]$ more .bash_profile
    # .bash_profile
    
    # Get the aliases and functions
    if [ -f ~/.bashrc ]; then
        . ~/.bashrc
    fi
    
    # User specific environment and startup programs
    
    PATH=$PATH:$HOME/bin
    export TMP=/tmp 
    export TMPDIR=$TMP 
    export ORACLE_SID=+ASM1  # 如果是节点2,则:export ORACLE_SID=+asm2 
    export ORACLE_BASE=/u01/app/grid
    export ORACLE_HOME=/u01/app/11.2.0/grid
    export PATH=/usr/sbin:$PATH 
    export PATH=$ORACLE_HOME/bin:$PATH 
    export LD_LIBRARY_PATH=$ORACLE_HOME/lib:/lib:/usr/lib 
    export CLASSPATH=$ORACLE_HOME/JRE:$ORACLE_HOME/jlib:$ORACLE_HOME/rdbms/jlib 
    umask 022

    问题2:通过scan ip连接不上数据库 (2018.09.12补充)

    在我的电脑上配置了tns文件,然后连接到数据库上时,报错

    ORA-12545: 因目标主机或对象不存在,连接失败

    通过其它IP(public ip和vip)都可以脸上数据库,网上查找相似问题,发现是11.2.0.1的bug,只需要将初始化参数local_listener中的host的值改为vip即可。

    修改方法如下:

    ---------节点1
    SQL> alter system set local_listener='(DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=TCP)(HOST=10.141.71.103)(PORT=1521))))' scope=both sid='racdb1';
    System altered.
    
    SQL> alter system register;
    System altered.
    
    ---------节点2
    SQL> alter system set local_listener='(DESCRIPTION=(ADDRESS_LIST=(ADDRESS=(PROTOCOL=TCP)(HOST=10.141.71.104)(PORT=1521))))' scope=both sid='racdb2';
    System altered.
    
    SQL> alter system register;
    System altered. 

    修改前:

    ----------2个节点配置均如下
    SQL
    > show parameter listener NAME TYPE VALUE ----------------------- ----------- ------------------------------ listener_networks string local_listener string (DESCRIPTION=(ADDRESS_LIST=(AD DRESS=(PROTOCOL=TCP)(HOST=rac1-vip)(PORT=1521)))) remote_listener string rac-scan:1521

    修改后:

    ------------节点1
    SQL> show parameter listener
    
    NAME                      TYPE         VALUE
    ------------------------  -----------  ------------------------------
    listener_networks         string
    local_listener            string     (DESCRIPTION=(ADDRESS_LIST=(AD
                                          DRESS=(PROTOCOL=TCP)(HOST=10.141.71.103)(PORT=1521))))
    remote_listener           string     rac-scan:1521
    
    -------------节点2
    SQL> show parameter listener
    
    NAME                     TYPE         VALUE
    -----------------------  -----------  ------------------------------
    listener_networks        string
    local_listener           string     (DESCRIPTION=(ADDRESS_LIST=(AD
                                         DRESS=(PROTOCOL=TCP)(HOST=10.141.71.104)(PORT=1521))))
    remote_listener          string     rac-scan:1521

    注释:这个错误是另一次安装遇到的,故数据库、服务器信息有所变化。

    【完】

  • 相关阅读:
    gitlab安装配置
    jenkins忘记admin密码的处理方法
    jenkins权限管理插件role-based(二)
    jenkins安装部署(一)
    ERROR 1129 (HY000): Host '192.168.7.210' is blocked because of many connection errors; unblock with 'mysqladmin flush-hosts'
    nginx重试机制proxy_next_upstream
    nginx配置ssi
    python实现查找指定文件
    python 调用系统命令
    Android 调用资源字符串的几种方法
  • 原文地址:https://www.cnblogs.com/lijiaman/p/7673795.html
Copyright © 2020-2023  润新知