내맘대로긍정이 알려주는
Oracle 23ai 신기능
무료 세미나 발표자료
OS환경 : Oracle Linux6.8(64bit)
DB 환경 : Oracle Database 11.2.0.4 3node RAC
호스트 네임: rac1, rac2, rac3
삭제 할 노드: rac3
방법 :
11g R2 RAC는 10g RAC 또는 11g R1 RAC와 비교하여 노드를 클러스터에 삭제하는 프로세스를 매우 단순화되었음
10g RAC 또는 11g R1 RAC pl에서 노드를 삭제하는 절차 링크 => 여기를 클릭해서 확인가능함
이제 SCAN 및 GPNP가 도입 된 후 매우 간단한 단계로 삭제가 가능함
#삭제할 노드(rac3)에 root로 로그인하여 $GRID_HOME/crs/install로 이동한다.
1 | [root@rac3 ~]# cd /oracle/app/11.2.0/grid/crs/install/ |
#노드상의 클러스터 소프트웨어 응용 프로그램 및 응용 프로그램을 비활성화 한다. (rac3)
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 | [root@rac3 install]# ./rootcrs.pl -deconfig -force Using configuration parameter file: ./crsconfig_params Network exists: 1/192.168.137.0/255.255.255.0/eth0, type static VIP exists: /rac1-vip/192.168.137.53/192.168.137.0/255.255.255.0/eth0, hosting node rac1 VIP exists: /rac2-vip/192.168.137.54/192.168.137.0/255.255.255.0/eth0, hosting node rac2 VIP exists: /rac3-vip/192.168.137.55/192.168.137.0/255.255.255.0/eth0, hosting node rac3 GSD exists ONS exists: Local port 6100, remote port 6200, EM port 2016 CRS-2791: Starting shutdown of Oracle High Availability Services-managed resources on 'rac3' CRS-2673: Attempting to stop 'ora.crsd' on 'rac3' CRS-2790: Starting shutdown of Cluster Ready Services-managed resources on 'rac3' CRS-2673: Attempting to stop 'ora.OCR_VOTE.dg' on 'rac3' CRS-2673: Attempting to stop 'ora.racdb.db' on 'rac3' CRS-2677: Stop of 'ora.racdb.db' on 'rac3' succeeded CRS-2673: Attempting to stop 'ora.ORADATA.dg' on 'rac3' CRS-2673: Attempting to stop 'ora.ORAFRA.dg' on 'rac3' CRS-2677: Stop of 'ora.ORADATA.dg' on 'rac3' succeeded CRS-2677: Stop of 'ora.ORAFRA.dg' on 'rac3' succeeded CRS-2677: Stop of 'ora.OCR_VOTE.dg' on 'rac3' succeeded CRS-2673: Attempting to stop 'ora.asm' on 'rac3' CRS-2677: Stop of 'ora.asm' on 'rac3' succeeded CRS-2792: Shutdown of Cluster Ready Services-managed resources on 'rac3' has completed CRS-2677: Stop of 'ora.crsd' on 'rac3' succeeded CRS-2673: Attempting to stop 'ora.crf' on 'rac3' CRS-2673: Attempting to stop 'ora.ctssd' on 'rac3' CRS-2673: Attempting to stop 'ora.evmd' on 'rac3' CRS-2673: Attempting to stop 'ora.asm' on 'rac3' CRS-2673: Attempting to stop 'ora.mdnsd' on 'rac3' CRS-2677: Stop of 'ora.crf' on 'rac3' succeeded CRS-2677: Stop of 'ora.mdnsd' on 'rac3' succeeded CRS-2677: Stop of 'ora.evmd' on 'rac3' succeeded CRS-2677: Stop of 'ora.ctssd' on 'rac3' succeeded CRS-2677: Stop of 'ora.asm' on 'rac3' succeeded CRS-2673: Attempting to stop 'ora.cluster_interconnect.haip' on 'rac3' CRS-2677: Stop of 'ora.cluster_interconnect.haip' on 'rac3' succeeded CRS-2673: Attempting to stop 'ora.cssd' on 'rac3' CRS-2677: Stop of 'ora.cssd' on 'rac3' succeeded CRS-2673: Attempting to stop 'ora.gipcd' on 'rac3' CRS-2677: Stop of 'ora.gipcd' on 'rac3' succeeded CRS-2673: Attempting to stop 'ora.gpnpd' on 'rac3' CRS-2677: Stop of 'ora.gpnpd' on 'rac3' succeeded CRS-2793: Shutdown of Oracle High Availability Services-managed resources on 'rac3' has completed CRS-4133: Oracle High Availability Services has been stopped. Removing Trace File Analyzer Successfully deconfigured Oracle clusterware stack on this node |
#Oracle 클러스터웨어 스택이 성공적으로 구성 해제되었음
#1번노드 alert log
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 | Fri Nov 30 10:57:31 2018 Reconfiguration started (old inc 5, new inc 7) List of instances: 1 2 (myinst: 1) Global Resource Directory frozen * dead instance detected - domain 0 invalid = TRUE Communication channels reestablished Fri Nov 30 10:57:31 2018 * domain 0 not valid according to instance 2 Master broadcasted resource hash value bitmaps Non-local Process blocks cleaned out Fri Nov 30 10:57:31 2018 LMS 0: 0 GCS shadows cancelled, 0 closed, 0 Xw survived Set master node info Submitted all remote-enqueue requests Dwn-cvts replayed, VALBLKs dubious All grantable enqueues granted Submitted all GCS remote-cache requests Fix write in gcs resources Reconfiguration complete |
#2번노드 alert log
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 | Fri Nov 30 10:57:31 2018 Reconfiguration started (old inc 5, new inc 7) List of instances: 1 2 (myinst: 2) Global Resource Directory frozen * dead instance detected - domain 0 invalid = TRUE Communication channels reestablished Fri Nov 30 10:57:31 2018 * domain 0 valid = 0 according to instance 1 Master broadcasted resource hash value bitmaps Non-local Process blocks cleaned out Fri Nov 30 10:57:31 2018 LMS 0: 0 GCS shadows cancelled, 0 closed, 0 Xw survived Set master node info Submitted all remote-enqueue requests Dwn-cvts replayed, VALBLKs dubious All grantable enqueues granted Post SMON to start 1st pass IR Submitted all GCS remote-cache requests Post SMON to start 1st pass IR Fix write in gcs resources Fri Nov 30 10:57:31 2018 Instance recovery: looking for dead threads Reconfiguration complete Beginning instance recovery of 1 threads Started redo scan Completed redo scan read 0 KB redo, 0 data blocks need recovery Started redo application at Thread 3: logseq 2, block 313, scn 289493 Recovery of Online Redo Log: Thread 3 Group 6 Seq 2 Reading mem 0 Mem# 0: +ORADATA/racdb/onlinelog/group_6.267.993490005 Mem# 1: +ORAFRA/racdb/onlinelog/group_6.260.993490007 Completed redo application of 0.00MB Completed instance recovery at Thread 3: logseq 2, block 313, scn 309494 0 data blocks read, 0 data blocks written, 0 redo k-bytes read Thread 3 advanced to log sequence 3 (thread recovery) Fri Nov 30 10:57:32 2018 minact-scn: Master returning as live inst:1 has inc# mismatch instinc:5 cur:7 errcnt:0 minact-scn: Master considers inst:3 dead |
#3번노드 alert log
1 2 3 4 5 6 7 | Fri Nov 30 10:57:30 2018 Shutting down instance (abort) License high water mark = 6 USER (ospid: 7602): terminating the instance Instance terminated by USER, pid = 7602 Fri Nov 30 10:57:31 2018 Instance shutdown complete |
#남아있는 노드(rac1)에서 지운노드(rac3)를 삭제한다.
1 2 | [root@rac1 ~]# crsctl delete node -n rac3 CRS-4661: Node rac3 successfully deleted. |
#남아있는 노드에서 oracle 또는 grid 유저로 로그인하고, $GRID_HOME/oui/bin으로 이동하여 노드 목록을 업데이트한다.
1 2 3 4 5 6 7 8 9 | [root@rac1 bin]# su - oracle [oracle@rac1 ~]$ cd /oracle/app/11.2.0/grid/oui/bin/ [oracle@rac1 bin]$ ./runInstaller -updateNodeList ORACLE_HOME=/oracle/app/11.2.0/grid "CLUSTER_NODES={rac1,rac2}" Starting Oracle Universal Installer... Checking swap space: must be greater than 500 MB. Actual 7999 MB Passed The inventory pointer is located at /etc/oraInst.loc The inventory is located at /oracle/app/oraInventory 'UpdateNodeList' was successful. |
#root 유저로 실행 할 경우 경고 메세지 발생
#1번노드 alert log
1 2 | Fri Nov 30 11:02:44 2018 Stopping background process CJQ0 |
#2번노드 alert log
1 2 | Fri Nov 30 11:02:44 2018 Stopping background process CJQ0 |
#노드1과 노드2만 남아 있는지 확인한다.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 | [oracle@rac1 ~]$ crsctl stat res -t [oracle@rac2 ~]$ crsctl stat res -t -------------------------------------------------------------------------------- NAME TARGET STATE SERVER STATE_DETAILS -------------------------------------------------------------------------------- Local Resources -------------------------------------------------------------------------------- ora.LISTENER.lsnr ONLINE ONLINE rac1 ONLINE ONLINE rac2 ora.OCR_VOTE.dg ONLINE ONLINE rac1 ONLINE ONLINE rac2 ora.ORADATA.dg ONLINE ONLINE rac1 ONLINE ONLINE rac2 ora.ORAFRA.dg ONLINE ONLINE rac1 ONLINE ONLINE rac2 ora.asm ONLINE ONLINE rac1 Started ONLINE ONLINE rac2 Started ora.gsd OFFLINE OFFLINE rac1 OFFLINE OFFLINE rac2 ora.net1.network ONLINE ONLINE rac1 ONLINE ONLINE rac2 ora.ons ONLINE ONLINE rac1 ONLINE ONLINE rac2 -------------------------------------------------------------------------------- Cluster Resources -------------------------------------------------------------------------------- ora.LISTENER_SCAN1.lsnr 1 ONLINE ONLINE rac1 ora.cvu 1 ONLINE ONLINE rac2 ora.oc4j 1 ONLINE ONLINE rac2 ora.rac1.vip 1 ONLINE ONLINE rac1 ora.rac2.vip 1 ONLINE ONLINE rac2 ora.racdb.db 1 ONLINE ONLINE rac1 Open 2 ONLINE ONLINE rac2 Open 3 ONLINE OFFLINE Instance Shutdown ora.scan1.vip 1 ONLINE ONLINE rac1 |
#1번노드, 2번노드에서 동일하게 나온다.
#ora.racdb.db 이부분에 3번이 나오는데 srvctl 명령어로 제거 할 수 있다.
1 2 | [root@rac1 ~]# srvctl remove instance -d racdb -i racdb3 Remove instance from the database racdb? (y/[n]) y |
#노드1과 노드2만 남아 있는지 확인한다.
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 | [root@rac1 ~]# crsctl stat res -t -------------------------------------------------------------------------------- NAME TARGET STATE SERVER STATE_DETAILS -------------------------------------------------------------------------------- Local Resources -------------------------------------------------------------------------------- ora.LISTENER.lsnr ONLINE ONLINE rac1 ONLINE ONLINE rac2 ora.OCR_VOTE.dg ONLINE ONLINE rac1 ONLINE ONLINE rac2 ora.ORADATA.dg ONLINE ONLINE rac1 ONLINE ONLINE rac2 ora.ORAFRA.dg ONLINE ONLINE rac1 ONLINE ONLINE rac2 ora.asm ONLINE ONLINE rac1 Started ONLINE ONLINE rac2 Started ora.gsd OFFLINE OFFLINE rac1 OFFLINE OFFLINE rac2 ora.net1.network ONLINE ONLINE rac1 ONLINE ONLINE rac2 ora.ons ONLINE ONLINE rac1 ONLINE ONLINE rac2 -------------------------------------------------------------------------------- Cluster Resources -------------------------------------------------------------------------------- ora.LISTENER_SCAN1.lsnr 1 ONLINE ONLINE rac1 ora.cvu 1 ONLINE ONLINE rac2 ora.oc4j 1 ONLINE ONLINE rac2 ora.rac1.vip 1 ONLINE ONLINE rac1 ora.rac2.vip 1 ONLINE ONLINE rac2 ora.racdb.db 1 ONLINE ONLINE rac1 Open 2 ONLINE ONLINE rac2 Open ora.scan1.vip 1 ONLINE ONLINE rac1 |
참조 :
http://www.idevelopment.info/data/Oracle/DBA_tips/Oracle11gRAC/CLUSTER_24.shtml
http://oracleinaction.com/delete-node/
'ORACLE > Rac' 카테고리의 다른 글
오라클 11g R2 RAC 환경 운영중에 /etc/passwd 파일이 삭제된다면? (0) | 2020.06.19 |
---|---|
오라클 11g R2 asm rac 환경에서 ocr 백업 확인 및 백업 하기 (0) | 2020.01.04 |
오라클 11g R2 RAC IP 변경 방법(IP Change) (0) | 2018.12.10 |
오라클 11g R2 RAC 제거한 노드 다시 추가하는 방법 (0) | 2018.11.30 |
오라클 11g R2 RAC 환경에서 리스너 관리 (0) | 2018.11.28 |