我是靠谱客的博主 鲜艳战斗机,这篇文章主要介绍SUSE11 Oracle11gR2 RAC&ASM双机集群环境部署,现在分享给大家,希望可以做个参考。

Oracle RAC集群环境部署

一. 安装前准备

(1) Linux系统版本

SUSE Linux Enterprise Server 11 (x86_64)

(2) Oracle database和Grid安装包

linux.x64_11gR2_database_1of2.zip

linux.x64_11gR2_database_2of2.zip

linux.x64_11gR2_grid.zip

(3) ASMlib安装包

oracleasm-support-2.1.8-1.SLE11.x86_64.rpm

oracleasmlib-2.0.4-1.sle11.x86_64.rpm

oracleasm-2.0.5-7.37.3.x86_64.rpm

oracleasm-kmp-default-2.0.5_3.0.76_0.11-7.37.3.x86_64.rpm

(4) 依赖软件包

gcc gcc-c++ gcc-32bit glibc-devel glibc-devel-32bit libaio libaio-devel libaio-devel-32bit libstdc++43-devel-32bit libstdc++43-devel sysstat libstdc++-devel libcap1 libcap1-32bit libcap2 libcap2-32bit compat* libgomp unixODBC unixODBC-devel

(5) 磁盘分区

名称

分区

容量

/

/dev/sda

100G

Swap

Swap

4G

CRS1

/dev/sdb1

2G

CRS2

/dev/sdb2

2G

CRS3

/dev/sdb3

2G

DATA1

/dev/sdb5

100G

REC1

/dev/sdb6

100G


(6) IP地址划分

Hostname

Node name

IP

Device

Type

 

hmracdb1

hmracdb1

192.168.6.154

eth1

Public

hmracdb1-vip

192.168.6.54

eth1

VIP

hmracdb1-priv

10.17.81.154

eth0

Private

 

hmracdb2

hmracdb2

192.168.6.155

eth1

Public

hmracdb2-vip

192.168.6.55

eth1

VIP

hmracdb2-priv

10.17.81.155

eth0

Private

Rac-scan

hmracdb-scan

192.168.6.66




二. 系统环境配置

(1) 主机hosts文件解析配置,两台主机使用相同配置

# public ip

192.168.6.154    hmracdb1

192.168.6.155    hmracdb2

 

# private ip

10.17.81.154     hmracdb1-priv

10.17.81.155     hmracdb2-priv

 

# vip

192.168.6.54     hmracdb1-vip

192.168.6.55     hmracdb2-vip

 

# scan ip

192.168.6.66     hmracdb-scan


(2) 创建oracle、grig用户和组

1. 创建组和用户

/usr/sbin/groupadd -g 501 oinstall

/usr/sbin/groupadd -g 502 dba

/usr/sbin/groupadd -g 503 oper

/usr/sbin/groupadd -g 504 asmadmin

/usr/sbin/groupadd -g 505 asmdba

/usr/sbin/groupadd -g 506 asmoper

/usr/sbin/useradd -u 501 -g oinstall -Gdba,oper,asmdba,asmadmin oracle -m

/usr/sbin/useradd -u 502 -g oinstall -Gdba,asmadmin,asmdba,asmoper,oper grid -m

 

2. 创建密码

echo oracle | passwd --stdin oracle

echo oracle | passwd --stdin grid

 

# id oracle

uid=501(oracle)gid=501(oinstall)groups=502(dba),503(oper),504(asmadmin),505(asmdba),501(oinstall)

 # id grid

uid=502(grid) gid=501(oinstall)groups=502(dba),503(oper),504(asmadmin),505(asmdba),506(asmoper),501(oinstall)


(3) 创建安装目录

mkdir -p /u01/app/{grid,oracle}

chown -R grid:oinstall /u01/

chown -R grid:oinstall /u01/app/grid/

chown -R oracle:oinstall /u01/app/oracle/

chmod -R 775 /u01/

 

ls -l /u01/app/

total 8

drwxrwxr-x 2 grid   oinstall 4096 Nov 16 19:09 grid

drwxrwxr-x 2 oracle oinstall 4096 Nov 16 19:09 oracle


(4) 配置grid、oracle用户环境变量

-------------------------------------------------------------------------------------------------------

1. 配置hmracdb1节点grid用户的profile,ORACLE_SID为+ASM1,hmracdb2节点改为+ASM2

-------------------------------------------------------------------------------------------------------

export ORACLE_SID=+ASM1

export ORACLE_BASE=/u01/app/oracle

exportORACLE_HOME=/u01/app/grid/product/11.2.0

export PATH=$PATH:$ORACLE_HOME/bin

export TMP=/tmp

export TMPDIR=$TMP

export NLS_DATE_FORMAT="yyyy-mm-ddHH24:MI:SS"

export THREADS_FLAG=native

export CLASSPATH=$ORACLE_HOME/JRE:$ORACLE_HOME/jlib:$ORACLE_HOME/rdbms/jlib

export CVUQDISK GRP=oinstall

if [ $USER = "oracle" ] || [$USER = 'grid' ];then

   if [ $SHELL = "/bin/ksh" ];then

       ulimit -p 16384

       ulimit -n 65536

   else

       ulimit -u 16384 -n 65536

   fi

   umask 022

fi

 

-------------------------------------------------------------------------------------------------------

2. 配置hmracdb1节点oracle用户的profile,ORACLE_SID为rac1,hmracdb2节点改为rac2

-------------------------------------------------------------------------------------------------------

export ORACLE_SID=rac1

export ORACLE_BASE=/u01/app/oracle

exportORACLE_HOME=$ORACLE_BASE/product/11.2.0

exportPATH=$ORACLE_HOME/bin:/usr/sbin:$PATH

export TMP=/tmp

export TMPDIR=$TMP

export ORACLE_TERM=xterm

exportLD_LIBRARY_PATH=$ORACLE_HOME/lib:/lib:/usr/lib:/usr/lib64

exportCLASSPATH=$ORACLE_HOME/JRE:$ORACLE_HOME/jlib:$ORACLE_HOME/rdbms/jlib

export NLS_DATE_FORMAT="yyyy-mm-ddHH24:MI:SS"

export NLS_LANG=AMERICAN_AMERICA.ZHS16GBK

if [ $USER = "oracle" ] || [$USER = 'grid' ];then

   if [ $SHELL = "/bin/ksh" ];then

       ulimit -p 16384

       ulimit -n 65536

   else

       ulimit -u 16384 -n 65536

   fi

   umask 022

fi


(5) 修改内核参数配置

1. 修改/etc/sysctl.conf配置

# vim /etc/sysctl.conf

kernel.shmmax = 68719476736

kernel.shmall = 4294967296

net.core.rmem_default = 262144

net.core.rmem_max = 4194304

net.core.wmem_default = 262144

net.core.wmem_max = 1048576

net.ipv4.ip_local_port_range = 9000 65500

fs.aio-max-nr = 1048576

fs.file-max = 6815744

kernel.sem = 250 32000 100 128

kernel.shmmni = 4096

 # sysctl –p

 

2. 修改/etc/security/limits.conf

# vim /etc/security/limits.conf

oracle          soft    nproc           2047

oracle          hard    nproc          16384

oracle          soft    nofile          1024

oracle          hard    nofile          65536

grid            soft    nproc           2047

grid            hard    nproc           16384

grid            soft    nofile          1024

grid            hard   nofile          65536

# vim /etc/pam.d/login

session  required       pam_limits.so


(5) 配置NTP时间同步

1. 同步时间

# sntp -P no -r 10.10.0.2 &&hwclock –w

 

2. 配置hmracdb1为上层NTP服务器

hmracdb1:~ # vim /etc/ntp.conf    #添加以下配置

restrict hmracdb1 mask 255.255.255.0 nomodify notrap noquery

server hmracdb1

server 127.127.1.0

 

3. 配置hmracdb2节点NTP服务

hmracdb2:~ # vim /etc/ntp.conf

server hmracdb1

 

4. 两个节点修改如下配置,并重启ntp服务

# vim /etc/sysconfig/ntp   #修改以下配置

NTPD_OPTIONS="-x -g -u ntp:ntp"

# service ntp restart


(6) 配置oracle、grid用户ssh密钥

su - grid

grid@hmracdb1:~> ssh-keygen -t rsa

grid@hmracdb2:~> ssh-keygen -t rsa

grid@hmracdb1:~> ssh-copy-id -i~/.ssh/id_rsa.pub grid@hmracdb1

grid@hmracdb1:~> ssh-copy-id -i~/.ssh/id_rsa.pub grid@hmracdb2

grid@hmracdb2:~> ssh-copy-id -i~/.ssh/id_rsa.pub grid@hmracdb1

grid@hmracdb2:~> ssh-copy-id -i~/.ssh/id_rsa.pub grid@hmracdb2

 

su - oracle

oracle@hmracdb1:~> ssh-keygen -t rsa

oracle@hmracdb2:~> ssh-keygen -t rsa

oracle@hmracdb1:~> ssh-copy-id -i~/.ssh/id_rsa.pub oracle@hmracdb1

oracle@hmracdb1:~> ssh-copy-id -i~/.ssh/id_rsa.pub oracle@hmracdb2

oracle@hmracdb2:~> ssh-copy-id -i~/.ssh/id_rsa.pub oracle@hmracdb1

oracle@hmracdb2:~> ssh-copy-id -i ~/.ssh/id_rsa.pub oracle@hmracdb2


三. 配置ASM磁盘

(1 )安装ASM软件(两个节点都安装)

# zypper in -y oracleasmoracleasm-kmp-default

# zypper inoracleasmlib-2.0.4-1.sle11.x86_64.rpm

# zypper in oracleasm-support-2.1.8-1.SLE11.x86_64.rpm

(2) 创建ASM分区(只需在节点1上创建)

复制代码
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
hmracdb1:~ # fdisk /dev/sdb  Device contains neither a valid DOSpartition table, nor Sun, SGI or OSF disklabel Building a new DOS disklabel with diskidentifier 0xd60a0f97. Changes will remain in memory only, untilyou decide to write them. After that, of course, the previouscontent won't be recoverable.   Warning: invalid flag 0x0000 of partitiontable 4 will be corrected by w(rite)   Command (m for help): n Command action   e   extended   p   primary partition (1-4) p Partition number (1-4, default 1):  Using default value 1 First sector (2048-419430399, default2048):  Using default value 2048 Last sector, +sectors or +size{K,M,G}(2048-419430399, default 419430399): +2G   Command (m for help): n Command action   e   extended   p   primary partition (1-4) p Partition number (1-4, default 2):  Using default value 2 First sector (4196352-419430399, default4196352):  Using default value 4196352 Last sector, +sectors or +size{K,M,G}(4196352-419430399, default 419430399): +2G   Command (m for help): n Command action   e   extended   p   primary partition (1-4) p Partition number (1-4, default 3):  Using default value 3 First sector (8390656-419430399, default8390656):  Using default value 8390656 Last sector, +sectors or +size{K,M,G}(8390656-419430399, default 419430399): +2G   Command (m for help): n Command action   e   extended   p   primary partition (1-4) e Selected partition 4 First sector (12584960-419430399, default12584960):  Using default value 12584960 Last sector, +sectors or +size{K,M,G}(12584960-419430399, default 419430399):  Using default value 419430399   Command (m for help): n First sector (12587008-419430399, default12587008):  Using default value 12587008 Last sector, +sectors or +size{K,M,G}(12587008-419430399, default 419430399): +100G   Command (m for help): n First sector (222304256-419430399,default 222304256):  Using default value 222304256 Last sector, +sectors or +size{K,M,G}(222304256-419430399, default 419430399):  Using default value 419430399   Command (m for help): p   Disk /dev/sdb: 214.7 GB, 214748364800bytes 255 heads, 63 sectors/track, 26108cylinders, total 419430400 sectors Units = sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes/ 512 bytes I/O size (minimum/optimal): 512 bytes /512 bytes Disk identifier: 0xd60a0f97     Device Boot      Start         End     Blocks   Id  System /dev/sdb1            2048     4196351    2097152   83  Linux /dev/sdb2         4196352     8390655    2097152   83  Linux /dev/sdb3         8390656    12584959    2097152   83  Linux /dev/sdb4        12584960   419430399  203422720    5  Extended /dev/sdb5        12587008   222302207  104857600   83  Linux /dev/sdb6       222304256   419430398   98563071+  83  Linux   Command (m for help): w The partition table has been altered!   Calling ioctl() to re-read partition table. Syncing disks.     在节点2上查看: hmracdb2:/usr/local/src # fdisk -l /dev/sdb    Disk /dev/sdb: 214.7 GB, 214748364800bytes 255 heads, 63 sectors/track, 26108cylinders, total 419430400 sectors Units = sectors of 1 * 512 = 512 bytes Sector size (logical/physical): 512 bytes/ 512 bytes I/O size (minimum/optimal): 512 bytes /512 bytes Disk identifier: 0xd60a0f97     Device Boot      Start         End      Blocks  Id  System /dev/sdb1            2048     4196351    2097152   83  Linux /dev/sdb2         4196352     8390655    2097152   83  Linux /dev/sdb3         8390656    12584959    2097152   83  Linux /dev/sdb4        12584960   419430399  203422720    5  Extended /dev/sdb5        12587008   222302207  104857600   83  Linux /dev/sdb6       222304256  419430398    98563071+  83 Linux


(3) 初始化ASM配置(两个节点都执行)

复制代码
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
1.加载asm内核模块 hmracdb1:~ # oracleasminit Creating /dev/oracleasm mount point:/dev/oracleasm Loading module "oracleasm":oracleasm Configuring "oracleasm" to usedevice physical block size  Mounting ASMlib driver filesystem:/dev/oracleasm   2.初始化配置 hmracdb1:~ # /etc/init.d/oracleasmconfigure Configuring the Oracle ASM librarydriver.   This will configure the on-bootproperties of the Oracle ASM library driver. The following questions will determine whether the driver is loaded on boot and what permissions itwill have.  The current values will be shown in brackets ('[]').  Hitting <ENTER> without typing an answer will keep that current value.  Ctrl-C will abort.   Default user to own the driver interface[]: grid Default group to own the driver interface[]: asmadmin Start Oracle ASM library driver on boot(y/n) [n]: y Scan for Oracle ASM disks on boot (y/n)[y]: y Writing Oracle ASM library driverconfiguration: done Initializing the Oracle ASMLibdriver:                                                                       done Scanning the system for Oracle ASMLib disks:                                                                 done


(4) 创建ASM盘(只需在节点1上创建)

复制代码
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
1. 在节点1上ASM磁盘 hmracdb1:~ # oracleasm listdisks hmracdb1:~ # oracleasm createdisk CRS1/dev/sdb1  Writing disk header: done Instantiating disk: done hmracdb1:~ # oracleasm createdisk CRS2/dev/sdb2  Writing disk header: done Instantiating disk: done hmracdb1:~ # oracleasm createdisk CRS3/dev/sdb3 Writing disk header: done Instantiating disk: done hmracdb1:~ # oracleasm createdisk DATA1/dev/sdb5 Writing disk header: done Instantiating disk: done hmracdb1:~ # oracleasm createdisk REC1/dev/sdb6 Writing disk header: done Instantiating disk: done hmracdb1:~ # oracleasm listdisks CRS1 CRS2 CRS3 DATA1 REC1   2. 节点2上扫描ASM共享盘 hmracdb2:~ # oracleasm scandisks Reloading disk partitions: done Cleaning any stale ASM disks... Scanning system for ASM disks... Instantiating disk "CRS1" Instantiating disk "CRS2" Instantiating disk "CRS3" Instantiating disk "DATA1" Instantiating disk "REC1" hmracdb2:~ # oracleasm listdisks CRS1 CRS2 CRS3 DATA1 REC1   3. 检查ASM磁盘路径 hmracdb2:~ # oracleasm querydisk/dev/sdb* Device "/dev/sdb" is not markedas an ASM disk Device "/dev/sdb1" is marked anASM disk with the label "CRS1" Device "/dev/sdb2" is marked anASM disk with the label "CRS2" Device "/dev/sdb3" is marked anASM disk with the label "CRS3" Device "/dev/sdb4" is notmarked as an ASM disk Device "/dev/sdb5" is marked anASM disk with the label "DATA1" Device "/dev/sdb6" is marked anASM disk with the label "REC1" hmracdb2:~ # ll /dev/oracleasm/disks/ total 0 brw-rw---- 1 grid asmadmin 8, 17 Nov 2114:19 CRS1 brw-rw---- 1 grid asmadmin 8, 18 Nov 2114:19 CRS2 brw-rw---- 1 grid asmadmin 8, 19 Nov 2114:19 CRS3 brw-rw---- 1 grid asmadmin 8, 21 Nov 2114:19 DATA1 brw-rw---- 1 grid asmadmin 8, 22 Nov 2114:19 REC1 hmracdb2:~ # ll -ltr /dev | grep "8,*17" brw-rw---- 1 root disk      8, 17 Nov 21 14:19 sdb1 hmracdb2:~ # ll -ltr /dev | grep "8,*18" brw-rw---- 1 root disk      8, 18 Nov 21 14:19 sdb2 hmracdb2:~ # ll -ltr /dev | grep "8,*19" brw-rw---- 1 root disk      8, 19 Nov 21 14:19 sdb3 hmracdb2:~ # ll -ltr /dev | grep "8,*21" brw-rw---- 1 root disk      8,  21 Nov 21 14:19 sdb5 hmracdb2:~ # ll -ltr /dev | grep "8,*22" brw-rw---- 1 root disk      8, 22 Nov 21 14:19 sdb6

(5) 安装oracle依赖软件包

hmracdb1:~ # for i in "gcc gcc-c++gcc-32bit glibc-devel glibc-devel-32bit libaio libaio-devel libaio-devel-32bitlibstdc++43-devel-32bit libstdc++43-devel sysstat libstdc++-devel libcap1libcap1-32bit libcap2 libcap2-32bit compat* libgomp unixODBC unixODBC-devel";dorpm -q $i;done

 

hmracdb1:~ # zypper in gcc gcc-c++ gcc-32bit glibc-develglibc-devel-32bit libaio libaio-devel libaio-devel-32bitlibstdc++43-devel-32bit libstdc++43-devel sysstat libstdc++-devel libcap1libcap1-32bit libcap2 libcap2-32bit compat* libgomp unixODBC unixODBC-devel

(6) 安装cvuqdisk软件

解压grid软件包,安装grid目录里的cvuqdisk-1.0.7-1.rpm软件

hmracdb1:/usr/local/src # cd grid/rpm/

hmracdb1:/usr/local/src/grid/rpm # ls

cvuqdisk-1.0.7-1.rpm

hmracdb1:/usr/local/src/grid/rpm # rpm -ivh cvuqdisk-1.0.7-1.rpm

(7) 配置nslookup

复制代码
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
1. 节点1上配置 hmracdb1:~ # mv /usr/bin/nslookup/usr/bin/nslookup.original hmracdb1:~ # vim /usr/bin/nslookup #!/bin/bash HOSTNAME=${1} if [[ $HOSTNAME ="hmracdb-scan" ]]; then    echo "Server: 192.168.6.54 "    echo "Address: 192.168.6.54 #53"    echo "Non-authoritative answer:"    echo "Name: hmracdb-scan"    echo "Address: 192.168.6.66 " else    /usr/bin/nslookup.original $HOSTNAME fi   hmracdb1:~ # chmod 755 /usr/bin/nslookup   2. 节点2上配置 hmracdb2:~ # mv /usr/bin/nslookup/usr/bin/nslookup.original hmracdb2:~ # vim /usr/bin/nslookup #!/bin/bash HOSTNAME=${1} if [[ $HOSTNAME ="hmracdb-scan" ]]; then    echo "Server: 192.168.6.55 "    echo "Address: 192.168.6.55 #53"    echo "Non-authoritative answer:"     echo "Name: hmracdb-scan"    echo "Address: 192.168.6.66 " else    /usr/bin/nslookup.original $HOSTNAME fi   hmracdb2:~ # chmod 755 /usr/bin/nslookup   3. 测试 hmracdb1:~ # nslookup hmracdb-scan Server: 192.168.6.54  Address: 192.168.6.54 #53 Non-authoritative answer: Name: hmracdb-scan Address: 192.168.6.66

(8) Gird安装先决条件检查

1. 修改属主权限为grid用户

hmracdb1:/usr/local/src # chown -R grid.grid

2. 以grid用户执行安装先决条件检查

grid@hmracdb1:/usr/local/src/grid>./runcluvfy.sh stage -pre crsinst -n hmracdb1,hmracdb2 -fixup -verbose


四. 图形化界面安装Grid Infrastructure

安装方法:

方法1:使用VNC软件登陆服务器图形界面安装

方法2:使用Xmanager调用图形界面安装

复制代码
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
ln -s /usr/bin/env /bin/ grid@hmracdb1:/usr/local/src/grid>export DISPLAY=10.18.221.155:0.0 grid@hmracdb1:/usr/local/src/grid>xhost + access control disabled, clients canconnect from any host grid@hmracdb1:/usr/local/src/grid>./runInstaller  Starting Oracle Universal Installer...   Checking Temp space: must be greater than120 MB.   Actual 83913 MB    Passed Checking swap space: must be greater than150 MB.   Actual 4094 MB    Passed Checking monitor: must be configured todisplay at least 256 colors.    Actual16777216    Passed Preparing  to launch Oracle UniversalInstaller from  /tmp/OraInstall2016-11-22_09-16-06AM. Please  wait...grid@hmracdb1:/usr/local/src/grid>  grid@hmracdb1:/usr/local/src/grid>


安装步骤:

wKioL1g25-DDbDrjAAC62B4guwU826.png-wh_50


wKiom1g26KKTHgoqAAC9NrC1PTM267.jpg-wh_50


wKiom1g26M6glo5lAADnDlMX2I8302.png-wh_50


wKioL1g26OLRJIR_AADRJ8AcloU066.png-wh_50


wKiom1g26PXytoe9AADerkraAA0061.png-wh_50


wKioL1g26V_BfKJYAADShDPEUZY329.png-wh_50


wKioL1g26bHyEYbsAADM-7LHrPo678.png-wh_50


wKiom1g26c_Rsk2HAADhbwjqO-4005.png-wh_50


wKioL1g26eWQ2BF9AADgl2eiI50570.png-wh_50


wKiom1g26fywriF9AADJvpFCnKQ981.png-wh_50


wKioL1g26hjyid4jAADT-4EDiio019.png-wh_50


wKiom1g26lmAHEeEAADxT85voUw964.png-wh_50


wKiom1g26nOhiLhDAADWnYFEkXk100.png-wh_50


wKioL1g26ovjkq5UAADohuEiOhc891.png-wh_50


wKiom1g26sLgJDp3AAEv9_I3eGM816.png-wh_50


wKioL1g26uriDpK5AAEh688Jkb4505.png-wh_50


执行脚本

复制代码
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
先执行第一个(节点1执行成功后,再到节点2执行),再执行第二个脚本 hmracdb1:~ #/u01/app/grid/product/11.2.0/root.sh  Running Oracle 11g root.sh script...   The following environment variables areset as:    ORACLE_OWNER= grid    ORACLE_HOME= /u01/app/grid/product/11.2.0   Enter the full pathname of the local bindirectory: [/usr/local/bin]:    Copying dbhome to /usr/local/bin ...   Copying oraenv to /usr/local/bin ...   Copying coraenv to /usr/local/bin ...     Creating /etc/oratab file... Entries will be added to the /etc/oratabfile as needed by Database Configuration Assistant when adatabase is created Finished running generic part of root.shscript. Now product-specific root actions will beperformed. 2016-11-22 14:18:59: Parsing the hostname 2016-11-22 14:18:59: Checking for superuser privileges 2016-11-22 14:18:59: User has super userprivileges Using configuration parameter file:/u01/app/grid/product/11.2.0/crs/install/crsconfig_params Creating trace directory LOCAL ADD MODE  Creating OCR keys for user 'root',privgrp 'root'.. Operation successful.  root wallet  root wallet cert  root cert export  peer wallet  profile reader wallet  pa wallet  peer wallet keys  pa wallet keys  peer cert request  pa cert request  peer cert  pa cert  peer root cert TP  profile reader root cert TP  pa root cert TP  peer pa cert TP  pa peer cert TP  profile reader pa cert TP  profile reader peer cert TP  peer user cert  pa user cert Adding daemon to inittab CRS-4123: Oracle High AvailabilityServices has been started. ohasd is starting ADVM/ACFS is not supported onlsb-release-2.0-1.2.18       CRS-2672: Attempting to start 'ora.gipcd'on 'hmracdb1' CRS-2672: Attempting to start 'ora.mdnsd'on 'hmracdb1' CRS-2676: Start of 'ora.mdnsd' on'hmracdb1' succeeded CRS-2676: Start of 'ora.gipcd' on'hmracdb1' succeeded CRS-2672: Attempting to start 'ora.gpnpd'on 'hmracdb1' CRS-2676: Start of 'ora.gpnpd' on'hmracdb1' succeeded CRS-2672: Attempting to start'ora.cssdmonitor' on 'hmracdb1' CRS-2676: Start of 'ora.cssdmonitor' on'hmracdb1' succeeded CRS-2672: Attempting to start 'ora.cssd'on 'hmracdb1' CRS-2672: Attempting to start'ora.diskmon' on 'hmracdb1' CRS-2676: Start of 'ora.diskmon' on'hmracdb1' succeeded CRS-2676: Start of 'ora.cssd' on'hmracdb1' succeeded CRS-2672: Attempting to start 'ora.ctssd'on 'hmracdb1' CRS-2676: Start of 'ora.ctssd' on'hmracdb1' succeeded   ASM created and started successfully.   DiskGroup CRS created successfully.   clscfg: -install mode specified Successfully accumulated necessary OCRkeys. Creating OCR keys for user 'root',privgrp 'root'.. Operation successful. CRS-2672: Attempting to start 'ora.crsd'on 'hmracdb1' CRS-2676: Start of 'ora.crsd' on'hmracdb1' succeeded CRS-4256: Updating the profile Successful addition of voting disk0cabca1797104f01bfc740460efd9665. Successful addition of voting disk66791219f9f14fc4bf1bb8c8b6bdda6a. Successful addition of voting disk22e396e1b1594ff3bf6f516303e17f28. Successfully replaced voting disk groupwith +CRS. CRS-4256: Updating the profile CRS-4266: Voting file(s) successfullyreplaced ## STATE    File Universal Id                File Name Disk group -- -----    -----------------                --------- ---------  1.ONLINE   0cabca1797104f01bfc740460efd9665(ORCL:CRS1) [CRS]  2.ONLINE   66791219f9f14fc4bf1bb8c8b6bdda6a(ORCL:CRS2) [CRS]  3.ONLINE   22e396e1b1594ff3bf6f516303e17f28(ORCL:CRS3) [CRS] Located 3 voting disk(s). CRS-2673: Attempting to stop 'ora.crsd'on 'hmracdb1' CRS-2677: Stop of 'ora.crsd' on'hmracdb1' succeeded CRS-2673: Attempting to stop 'ora.asm' on'hmracdb1' CRS-2677: Stop of 'ora.asm' on 'hmracdb1'succeeded CRS-2673: Attempting to stop 'ora.ctssd'on 'hmracdb1' CRS-2677: Stop of 'ora.ctssd' on'hmracdb1' succeeded CRS-2673: Attempting to stop'ora.cssdmonitor' on 'hmracdb1' CRS-2677: Stop of 'ora.cssdmonitor' on'hmracdb1' succeeded CRS-2673: Attempting to stop 'ora.cssd'on 'hmracdb1' CRS-2677: Stop of 'ora.cssd' on'hmracdb1' succeeded CRS-2673: Attempting to stop 'ora.gpnpd'on 'hmracdb1' CRS-2677: Stop of 'ora.gpnpd' on'hmracdb1' succeeded CRS-2673: Attempting to stop 'ora.gipcd'on 'hmracdb1' CRS-2677: Stop of 'ora.gipcd' on'hmracdb1' succeeded CRS-2673: Attempting to stop 'ora.mdnsd'on 'hmracdb1' CRS-2677: Stop of 'ora.mdnsd' on'hmracdb1' succeeded CRS-2672: Attempting to start 'ora.mdnsd'on 'hmracdb1' CRS-2676: Start of 'ora.mdnsd' on'hmracdb1' succeeded CRS-2672: Attempting to start 'ora.gipcd'on 'hmracdb1' CRS-2676: Start of 'ora.gipcd' on 'hmracdb1'succeeded CRS-2672: Attempting to start 'ora.gpnpd'on 'hmracdb1' CRS-2676: Start of 'ora.gpnpd' on'hmracdb1' succeeded CRS-2672: Attempting to start'ora.cssdmonitor' on 'hmracdb1' CRS-2676: Start of 'ora.cssdmonitor' on'hmracdb1' succeeded CRS-2672: Attempting to start 'ora.cssd'on 'hmracdb1' CRS-2672: Attempting to start'ora.diskmon' on 'hmracdb1' CRS-2676: Start of 'ora.diskmon' on'hmracdb1' succeeded CRS-2676: Start of 'ora.cssd' on'hmracdb1' succeeded CRS-2672: Attempting to start 'ora.ctssd'on 'hmracdb1' CRS-2676: Start of 'ora.ctssd' on'hmracdb1' succeeded CRS-2672: Attempting to start 'ora.asm'on 'hmracdb1' CRS-2676: Start of 'ora.asm' on'hmracdb1' succeeded CRS-2672: Attempting to start 'ora.crsd'on 'hmracdb1' CRS-2676: Start of 'ora.crsd' on'hmracdb1' succeeded CRS-2672: Attempting to start 'ora.evmd'on 'hmracdb1' CRS-2676: Start of 'ora.evmd' on'hmracdb1' succeeded CRS-2672: Attempting to start 'ora.asm'on 'hmracdb1' CRS-2676: Start of 'ora.asm' on'hmracdb1' succeeded CRS-2672: Attempting to start'ora.CRS.dg' on 'hmracdb1' CRS-2676: Start of 'ora.CRS.dg' on'hmracdb1' succeeded   hmracdb1     2016/11/22 14:25:57    /u01/app/grid/product/11.2.0/cdata/hmracdb1/backup_20161122_142557.olr Configure Oracle Grid Infrastructure fora Cluster ... succeeded Updating inventory properties forclusterware Starting Oracle Universal Installer...   Checking swap space: must be greater than500 MB.   Actual 4094 MB    Passed The inventory pointer is located at/etc/oraInst.loc The inventory is located at/u01/app/oraInventory 'UpdateNodeList' was successful.


五. 安装后的检查

(1) 检查CRS资源组状态

【节点1】

crs_stat -t -v或者crsctl statusresource -t

hmracdb1:~ # su - grid

grid@hmracdb1:~> crsctl statusresource -t

--------------------------------------------------------------------------------

NAME           TARGET  STATE       SERVER                   STATE_DETAILS      

--------------------------------------------------------------------------------

Local Resources

--------------------------------------------------------------------------------

ora.CRS.dg

               ONLINE  ONLINE      hmracdb1                                    

               ONLINE  ONLINE      hmracdb2                                    

ora.LISTENER.lsnr

               ONLINE  ONLINE      hmracdb1                                    

               ONLINE  ONLINE      hmracdb2                                    

ora.asm

               ONLINE  ONLINE      hmracdb1                Started            

               ONLINE  ONLINE      hmracdb2                Started            

ora.eons

               ONLINE  ONLINE      hmracdb1                                    

               ONLINE  ONLINE      hmracdb2                                    

ora.gsd

               OFFLINE OFFLINE      hmracdb1                                    

               OFFLINE OFFLINE      hmracdb2                                    

ora.net1.network

               ONLINE  ONLINE      hmracdb1                                    

               ONLINE  ONLINE      hmracdb2                                    

ora.ons

               ONLINE  ONLINE      hmracdb1                                    

               ONLINE  ONLINE      hmracdb2                                    

--------------------------------------------------------------------------------

Cluster Resources

--------------------------------------------------------------------------------

ora.LISTENER_SCAN1.lsnr

     1        ONLINE  ONLINE      hmracdb1                                    

ora.hmracdb1.vip

     1        ONLINE  ONLINE      hmracdb1                                     

ora.hmracdb2.vip

     1        ONLINE  ONLINE      hmracdb2                                    

ora.oc4j

     1        OFFLINE OFFLINE                                                  

ora.scan1.vip

     1        ONLINE  ONLINE      hmracdb1 

 

【节点2】

hmracdb2:~ # su - grid

grid@hmracdb2:~> crsctl statusresource -t

--------------------------------------------------------------------------------

NAME           TARGET  STATE       SERVER                   STATE_DETAILS      

--------------------------------------------------------------------------------

Local Resources

--------------------------------------------------------------------------------

ora.CRS.dg

               ONLINE  ONLINE      hmracdb1                                     

               ONLINE  ONLINE      hmracdb2                                    

ora.LISTENER.lsnr

               ONLINE  ONLINE      hmracdb1                                    

               ONLINE  ONLINE      hmracdb2                                    

ora.asm

               ONLINE  ONLINE      hmracdb1                Started            

               ONLINE  ONLINE      hmracdb2                Started            

ora.eons

               ONLINE  ONLINE      hmracdb1                                    

               ONLINE  ONLINE      hmracdb2                                    

ora.gsd

               OFFLINE OFFLINE      hmracdb1                                     

               OFFLINE OFFLINE      hmracdb2                                    

ora.net1.network

               ONLINE  ONLINE      hmracdb1                                    

               ONLINE  ONLINE      hmracdb2                                     

ora.ons

               ONLINE  ONLINE      hmracdb1                                    

               ONLINE  ONLINE      hmracdb2                                    

--------------------------------------------------------------------------------

Cluster Resources

--------------------------------------------------------------------------------

ora.LISTENER_SCAN1.lsnr

     1        ONLINE  ONLINE      hmracdb1                                    

ora.hmracdb1.vip

     1        ONLINE  ONLINE      hmracdb1                                    

ora.hmracdb2.vip

     1        ONLINE  ONLINE      hmracdb2                                    

ora.oc4j

     1        OFFLINE OFFLINE                                                  

ora.scan1.vip

     1        ONLINE  ONLINE      hmracdb1 



(2) 检查监听状态

grid@hmracdb1:~> lsnrctl status

 

LSNRCTL for Linux: Version 11.2.0.1.0 -Production on 22-NOV-2016 17:47:35

 

Copyright (c) 1991, 2009, Oracle.  All rights reserved.

 

Connecting to(DESCRIPTION=(ADDRESS=(PROTOCOL=IPC)(KEY=LISTENER)))

STATUS of the LISTENER

------------------------

Alias                     LISTENER

Version                   TNSLSNR for Linux: Version11.2.0.1.0 - Production

Start Date                22-NOV-2016 17:31:00

Uptime                    0 days 0 hr. 16 min. 35 sec

Trace Level               off

Security                  ON: Local OS Authentication

SNMP                      OFF

Listener Parameter File  /u01/app/grid/product/11.2.0/network/admin/listener.ora

Listener Log File        /u01/app/oracle/diag/tnslsnr/hmracdb1/listener/alert/log.xml

Listening Endpoints Summary...

 (DESCRIPTION=(ADDRESS=(PROTOCOL=ipc)(KEY=LISTENER)))

 (DESCRIPTION=(ADDRESS=(PROTOCOL=tcp)(HOST=192.168.6.154)(PORT=1521)))

 (DESCRIPTION=(ADDRESS=(PROTOCOL=tcp)(HOST=192.168.6.54)(PORT=1521)))

Services Summary...

Service "+ASM" has 1instance(s).

 Instance "+ASM1", status READY, has 1 handler(s) for thisservice...

The command completed successfully


六. 创建ASM磁盘组(asmca)

创建数据区和快速恢复区ASM磁盘组

DATA  --- > /dev/oracleasm/disks/DATA1

REC1  --- >  /dev/oracleasm/disks/REC1

以grid用户运行asmca

wKioL1g27yyQ2jLqAADLtGLr-_E651.png-wh_50


wKiom1g27z_T1yhtAADRjq7orvs367.png-wh_50


wKioL1g271DQ5ylLAACiW-w5syM510.png-wh_50


wKiom1g273GzKF69AACihknfRB0794.png-wh_50


wKiom1g2747BM37JAAAlrg-gkqw444.png-wh_50

以同样的方法创建REC磁盘组

wKioL1g275_j6_GrAADbKkFrRks437.png-wh_50


七. Oracle Database 软件安装

(1) 解压database软件

复制代码
1
2
3
hmracdb1:/usr/local/src # unziplinux.x64_11gR2_database_1of2.zip hmracdb1:/usr/local/src # unzip linux.x64_11gR2_database_2of2.zip hmracdb1:/usr/local/src # chown -R oracle. database

(2) 图形安装过程

复制代码
1
oracle@hmracdb1:/usr/local/src/database>./runInstaller

wKioL1g28F6Ra1kxAAC9Twr3_Ao815.png-wh_50


wKiom1g28HCxDSc6AAC7a0c5gmg687.png-wh_50


wKiom1g28IDybfdxAADPcIIXnPo771.png-wh_50


wKioL1g28JDjqc90AADp3u5RrWQ722.png-wh_50


wKioL1g28KfRkHg_AADNrxXh2Iw033.png-wh_50


wKiom1g28L6zBbE0AADQGE36xPA515.png-wh_50


wKiom1g28OSDsLtIAADSqGAs9oY994.png-wh_50


wKiom1g28PuD9g2aAADTDqn_vq8875.png-wh_50


wKioL1g28SOhseLWAAEg-sv1my0160.png-wh_50


wKiom1g28XbSOCEbAAErs4sNlKk869.png-wh_50


执行脚本

复制代码
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
hmracdb1:~ #/u01/app/oracle/product/11.2.0/root.sh Running Oracle 11g root.sh script...   The following environment variables areset as:    ORACLE_OWNER= oracle    ORACLE_HOME= /u01/app/oracle/product/11.2.0   Enter the full pathname of the local bindirectory: [/usr/local/bin]:  The file "dbhome" alreadyexists in /usr/local/bin.  Overwrite it?(y/n)  [n]: n The file "oraenv" alreadyexists in /usr/local/bin.  Overwrite it?(y/n)  [n]: n The file "coraenv" alreadyexists in /usr/local/bin.  Overwrite it?(y/n)  [n]: n   Entries will be added to the /etc/oratabfile as needed by Database Configuration Assistant when adatabase is created Finished running generic part of root.shscript. Now product-specific root actions will beperformed. Finished product-specific root actions.


wKioL1g28aPju4Y0AACo7D5Hucc322.png-wh_50


八. 创建数据库实例(dbca)

复制代码
1
oracle@hmracdb1:~> dbca

wKiom1g28wrRdGCNAADToYRLlzE660.png-wh_50


wKioL1g28wuADujuAADVJZ_Z9Bk880.png-wh_50


wKiom1g28w3hLLNwAADWIzf1y2o330.png-wh_50


wKioL1g28w2SQRNLAADk4ZQ4M_0876.png-wh_50


wKioL1g28w7DHqqaAADVAzuTjTc551.png-wh_50


wKiom1g28w-x7OMoAADFgzfLKM8569.png-wh_50


wKiom1g28xDxDhkRAADjcpsyjiw898.png-wh_50


wKioL1g28xGR_ZCzAADq14cGbNs122.png-wh_50


wKioL1g28xKSOG_sAADTatYeGNA536.png-wh_50


wKiom1g28xOias8nAADbPn6EUPE898.png-wh_50


wKiom1g28xSzb-7qAADtP85invU349.png-wh_50


wKioL1g28xXSMHs0AAD2eJL21uM220.png-wh_50


wKiom1g28xbhJXg-AAB2VpW2TrU027.png-wh_50

表空间可以根据实际情况创建合适的空间

wKioL1g28xfx6zdxAAB9ubyYr0M414.png-wh_50


wKiom1g28xehGi5uAAB-yxNgX5o914.png-wh_50

重组日志

wKioL1g28xiTgQzSAACFcWfWRuk294.png-wh_50


wKioL1g28xjyHPL0AABvm0-3Uj8004.png-wh_50


wKiom1g28xnBuYBhAADLskoGn-I292.png-wh_50


wKiom1g28xqgG2NAAABwleyJQvM025.png-wh_50


wKiom1g28xuzP6MhAAD26JoXPQQ276.png-wh_50


wKioL1g28xyimhSgAACW4Z3T928140.png-wh_50


九. 集群操作命令

(1) 集群启动与关闭

关闭步骤:

1、先关闭数据库实例

hmracdb1:~ # su - oracle

oracle@hmracdb1:~> srvctl stopdatabase -d rac

2、关闭oracle集群

hmracdb1:~ #/u01/app/grid/product/11.2.0/bin/crsctl stop cluster -all

 

启动步骤:(oracle集群默认开机自动启动)

1、  启动oracle集群

hmracdb1:~ #/u01/app/grid/product/11.2.0/bin/crsctl start cluster -all

2、  启动数据库实例

hmracdb1:~ # su - oracle

oracle@hmracdb1:~> srvctl startdatabase -d rac


(2) 检查RAC资源组状态

1. 节点1

grid@hmracdb1:~> crs_stat -t -v


2. 节点2

grid@hmracdb2:~> crs_stat -t -v


3. gsd和oc4j默认不需要启动,设置收到启动并设置自动启动

grid@hmracdb1:~> srvctl enablenodeapps –g

grid@hmracdb1:~> srvctl start nodeapps

grid@hmracdb1:~> srvctl enable oc4j

grid@hmracdb1:~> srvctl start oc4j

wKiom1g289eg1K8hAACK0Azf6C4523.png-wh_50


(3) 检查RAC运行状态

复制代码
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
1. 检查RAC节点 grid@hmracdb1:~> olsnodes -n hmracdb1      1 hmracdb2      2   2. 检查crs集群状态 grid@hmracdb1:~> crsctl check cluster CRS-4537: Cluster Ready Services isonline CRS-4529: Cluster SynchronizationServices is online CRS-4533: Event Manager is online   grid@hmracdb1:~> crsctl check crs CRS-4638: Oracle High AvailabilityServices is online CRS-4537: Cluster Ready Services isonline CRS-4529: Cluster SynchronizationServices is online CRS-4533: Event Manager is online   3. 检查数据库实例状态 grid@hmracdb1:~> srvctl configdatabase -d rac -a Database unique name: rac Database name: rac Oracle home:/u01/app/oracle/product/11.2.0 Oracle user: oracle Spfile: +DATA/rac/spfilerac.ora Domain:  Start options: open Stop options: immediate Database role: PRIMARY Management policy: AUTOMATIC Server pools: rac Database instances: rac1,rac2 Disk Groups: DATA,REC Services:  Database is enabled Database is administrator managed   grid@hmracdb1:~> srvctl statusdatabase -d rac Instance rac1 is running on node hmracdb1 Instance rac2 is running on node hmracdb2   4. 检查ASM磁盘 grid@hmracdb1:~> srvctl config asm ASM home: /u01/app/grid/product/11.2.0 ASM listener: LISTENER   grid@hmracdb1:~> srvctl status asm ASM is running on hmracdb1,hmracdb2   5. 检查监听状态 grid@hmracdb1:~> srvctl status asm ASM is running on hmracdb1,hmracdb2 grid@hmracdb1:~> srvctl statuslistener Listener LISTENER is enabled Listener LISTENER is running on node(s):hmracdb1,hmracdb2 grid@hmracdb1:~> srvctl configlistener Name: LISTENER Network: 1, Owner: grid Home: <CRS home> End points: TCP:1521   6. 检查VIP状态 grid@hmracdb1:~> srvctl status vip -nhmracdb1 VIP hmracdb1-vip is enabled VIP hmracdb1-vip is running on node:hmracdb1 grid@hmracdb1:~> srvctl status vip -nhmracdb2 VIP hmracdb2-vip is enabled VIP hmracdb2-vip is running on node:hmracdb2   grid@hmracdb1:~> srvctl config vip -nhmracdb1 VIP exists.:hmracdb1 VIP exists.:/hmracdb1-vip/192.168.6.54/255.255.255.0/eth0 grid@hmracdb1:~> srvctl config vip -nhmracdb2 VIP exists.:hmracdb2 VIP exists.:/hmracdb2-vip/192.168.6.55/255.255.255.0/eth0   7. 检查SCAN grid@hmracdb1:~> srvctl status scan SCAN VIP scan1 is enabled SCAN VIP scan1 is running on nodehmracdb1 grid@hmracdb1:~> srvctl config scan SCAN name: hmracdb-scan, Network:1/192.168.6.0/255.255.255.0/eth0 SCAN VIP name: scan1, IP:/hmracdb-scan/192.168.6.66   8. 检查表决磁盘 grid@hmracdb1:~> crsctl query cssvotedisk ## STATE    File Universal Id                File Name Disk group -- -----    -----------------                --------- ---------  1.ONLINE   0cabca1797104f01bfc740460efd9665(ORCL:CRS1) [CRS]  2.ONLINE   66791219f9f14fc4bf1bb8c8b6bdda6a(ORCL:CRS2) [CRS]  3.ONLINE   22e396e1b1594ff3bf6f516303e17f28(ORCL:CRS3) [CRS] Located 3 voting disk(s).   9. 检查集群注册表 grid@hmracdb1:~> ocrcheck Status of Oracle Cluster Registry is asfollows :         Version                  :          3         Total space (kbytes)     :    262120         Used space (kbytes)      :      2568         Available space (kbytes) :     259552         ID                       :  858387628         Device/File Name         :      +CRS                                     Device/File integrity checksucceeded                                       Device/Filenot configured                                       Device/Filenot configured                                       Device/Filenot configured                                      Device/File notconfigured           Cluster registry integrity check succeeded           Logical corruption check bypassed due tonon-privileged user     10. 检查ASM磁盘组 grid@hmracdb1:~> asmcmd lsdg State   Type    Rebal  Sector Block       AU  Total_MB Free_MB  Req_mir_free_MB  Usable_file_MB  Offline_disks Voting_files  Name MOUNTED NORMAL  N         512  4096  1048576      6144    5218              309            2454              0             N CRS/ MOUNTED EXTERN  N         512  4096  1048576   102400    83297                0           83297              0             N DATA/ MOUNTED EXTERN  N         512  4096  1048576     96253   94471                0           94471              0             N REC/



(4) OracleDatabase实例检查(oracle查询语句)

复制代码
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
SQL> select instance_name,status fromgv$instance;   INSTANCE_NAME                  STATUS -------------------------------------------------------- rac1                      OPEN rac2                      OPEN   1. asm磁盘检查 SQL> set linesize 100; SQL> select path from v$asm_disk;   PATH ---------------------------------------------------------------------------------------------------- ORCL:CRS1 ORCL:CRS2 ORCL:CRS3 ORCL:DATA1 ORCL:REC1   SQL> show parameter spfile;   NAME                               TYPE                  VALUE ------------------------------------ ---------------------------------------------------- spfile                                string                +DATA/rac/spfilerac.ora SQL> select path from v$asm_disk;   PATH ---------------------------------------------------------------------------------------------------- ORCL:CRS1 ORCL:CRS2 ORCL:CRS3 ORCL:DATA1 ORCL:REC1   2. asm磁盘组空间 SQL> set linesize 500; SQL> selectgroup_number,name,state,type,total_mb,free_mb from v$asm_diskgroup; SQL> selectdisk_number,path,name,total_mb,free_mb from v$asm_disk where group_number=1;   3. 查询表空间名 SQL> select tablespace_name fromdba_tablespaces;   TABLESPACE_NAME ------------------------------------------------------------ SYSTEM SYSAUX UNDOTBS1 TEMP UNDOTBS2 USERS   6 rows selected.   4. 查询指定用户默认表空间 SQL> selectdefault_tablespace,temporary_tablespace from dba_users where username='SYSTEM';   DEFAULT_TABLESPACE                                    TEMPORARY_TABLESPACE ------------------------------------------------------------------------------------------------------------------------ SYSTEM                                                 TEMP     5. 查询表空间使用率 select ff.s tablespace_name, ff.b total,               (ff.b - fr.b)usage,  fr.b free,                round((ff.b - fr.b) / ff.b * 100)|| '% ' usagep from (select tablespace_name s,sum(bytes) / 1024 / 1024 b                     from dba_data_files                   group by tablespace_name) ff,               (select tablespace_name s,sum(bytes) / 1024 / 1024 b                     from dba_free_space                   group by tablespace_name) fr where ff.s = fr.s; 6. 查询表空间数据文件 select name from v$datafile; SQL> select name from v$datafile;   NAME ---------------------------------------------------------------------------------------------------- +DATA/rac/datafile/system.261.928753797 +DATA/rac/datafile/sysaux.262.928753819 +DATA/rac/datafile/undotbs1.263.928753839 +DATA/rac/datafile/undotbs2.265.928753875 +DATA/rac/datafile/users.266.928753895   7. 查询日志文件 select member from v$logfile; SQL> select member from v$logfile;   MEMBER ---------------------------------------------------------------------------------------------------- +DATA/rac/onlinelog/group_1.257.928753771 +REC/rac/onlinelog/group_1.257.928753775 +DATA/rac/onlinelog/group_2.258.928753777 +REC/rac/onlinelog/group_2.258.928753781 +DATA/rac/onlinelog/group_5.259.928753783 +REC/rac/onlinelog/group_5.259.928753787 +DATA/rac/onlinelog/group_6.260.928753791 +REC/rac/onlinelog/group_6.260.928753793 +DATA/rac/onlinelog/group_3.267.928756451 +REC/rac/onlinelog/group_3.261.928756455 +DATA/rac/onlinelog/group_4.268.928756459 +REC/rac/onlinelog/group_4.262.928756461   12 rows selected.   8. 查询归档日志 SQL> show parameterdb_recovery_file_dest_size;   NAME                               TYPE                  VALUE ---------------------------------------------------------- ------------------------------ db_recovery_file_dest_size        big integer         30G







转载于:https://blog.51cto.com/7424593/1876401

最后

以上就是鲜艳战斗机最近收集整理的关于SUSE11 Oracle11gR2 RAC&ASM双机集群环境部署的全部内容,更多相关SUSE11内容请搜索靠谱客的其他文章。

本图文内容来源于网友提供,作为学习参考使用,或来自网络收集整理,版权属于原作者所有。
点赞(82)

评论列表共有 0 条评论

立即
投稿
返回
顶部