Oracle GoldenGate
Oracle GoldenGate
Oracle GoldenGate
Oracle Corporation at the end of last year has acquired GoldenGate Software Inc which is one of the leading providers of real time data integration solutions. In the words of Hasan Rizvi, Senior Vice President, Oracle Fusion Middleware Product Development . GoldenGate also enables high availability and disaster tolerance with core functionality that supports real-time data integration, while preserving performance and ensuring scalability. With this capability, business-critical applications such as online banking, claims processing and online shopping continue to operate without any disruption. GoldenGates solutions address all types of outages as a result of upgrades, maintenance, application of patches, and application or platform migrations. The combination of GoldenGate and Oracle is expected to provide customers with a comprehensive platform for real-time data integration that delivers continuous data-change and replication functionality across heterogeneous environments The product is now available for download from the OTN site as well as via the Oracle EDelivery.com web site under the Product Pack titled Oracle Fusion Middleware . I have downloaded and installed the same and have done a bit of testing of my own trying to figure out how the product works and will be sharing the same with the community via a series of simple tutorials. Lets get a brief idea on what products make up the GoldenGate suite: Oracle GoldenGate GoldenGate Director is a multi-tiered client-server application that enables the configuration and management of GoldenGate instances from a remote client which includes a browser based web interface. Oracle GoldenGate Veridata is an online high-speed data comparison solution that identifies and reports data discrepancies between databases and sets of data
Let us know look at the differents components and processes that make up a typical GoldenGate configuration on Oracle.
(source: Oracle GoldenGate Administration Guide) Manager The Manager process must be running on both the source as well as target systems before the Extract or Replicat process can be started and performs a number of functions including monitoring and starting other GoldenGate processes, managing the trail files and also reporting. Extract The Extract process runs on the source system and is the data capture mechanism of GoldenGate. It can be configured both for initial loading of the source data as well as to synchronize the changed data on the source with the target. This can be configured to also propagate any DDL changes on those databases where DDL change support is available. Replicat The Replicat process runs on the target system and reads transactional data changes as well as DDL changes and replicates then to the target database. Like the Extract process, the Replicat process can also be configured for Initial Load as well as Change Synchronization. Collector The Collector is a background process which runs on the target system and is started automatically by the Manager (Dynamic Collector) or it can be configured to stsrt manually (Static Collector). It receives extracted data changes that are sent via TCP/IP and writes then to the trail files from where they are processed by the Replicat process. Trails
Trails are series of files that GoldenGate temporarily stores on disks and these files are written to and read from by the Extract and Replicat processes as the case may be. Depending on the configuration chosen, these trail files can exist on the source as well as on the target systems. If it exists on the local system, it will be known an Extract Trail or as an Remote Trail if it exists on the target system. Data Pumps Data Pumps are secondary extract mechanisms which exist in the source configuration. This is optional component and if Data Pump is not used then Extract sends data via TCP/IP to the remote trail on the target. When Data Pump is configured, the the Primary Extract process will write to the Local Trail and then this trail is read by the Data Pump and data is sent over the network to Remote Trails on the target system. In the absence of Data Pump, the data that the Extract process extracts resides in memory alone and there is no storage of this data anywhere on the source system. In case of network of target failures, there could be cases where the primary extract process can abort or abend. Data Pump can also be useful in those cases where we are doing complex filtering and transformation of data as well as when we are consolidating data from many sources to a central target. Data source When processing transactional data changes, the Extract process can obtain data directly from the database transaction logs (Oracle, DB2, SQL Server, MySQL etc) or from a GoldenGate Vendor Access Module (VAM) where the database vendor (for example Teradata) will provide the required components that will be used by Extract to extract the data changes. Groups To differentiate between the number of different Extract and Replicat groups which can potentially co-exist on a system, we can define processing groups. For instance, if we want to replicate different sets of data in parallel, we can create two Replicat groups. A processing group consisits of a process which could be either a Extract or Replicat process, a corresponding parameter file, checkpoint file or checkpoint table (for Replicat) and other files which could be associated with the process.
GoldenGate software is also available on OTN but for our platform we need to download the required software from the Oracle E-Delivery web site. Select the Product Pack Oracle Fusion Middleware and the platform Linux X86-64. Then select Oracle GoldenGate on Oracle Media Pack for Linux x86-64 and since we are installing this for an Oracle 11g database, we download Oracle GoldenGate V10.4.0.x for Oracle 11g 64bit on RedHat 5.0 $ unzip V18159-01.zip Archive: V18159-01.zip inflating: ggs_redhatAS50_x64_ora11g_64bit_v10.4.0.19_002.tar $tar -xvof ggs_redhatAS50_x64_ora11g_64bit_v10.4.0.19_002.tar $ export PATH=$PATH:/u01/oracle/ggs $ export LD_LIBRARY_PATH=$ORACLE_HOME/lib:/u01/oracle/ggs $ ggsci Oracle GoldenGate Command Interpreter for Oracle Version 10.4.0.19 Build 002 Linux, x64, 64bit (optimized), Oracle 11 on Sep 17 2009 23:51:28 Copyright (C) 1995, 2009, Oracle and/or its affiliates. All rights reserved. GGSCI (redhat346.localdomain) 1> GGSCI (redhat346.localdomain) 1> CREATE SUBDIRS Creating subdirectories under current directory /u01/app/oracle/product/11.2.0/dbhome_1 Parameter files /u01/oracle/ggs/dirprm: created Report files /u01/oracle/ggs/dirrpt: created Checkpoint files /u01/oracle/ggs/dirchk: created Process status files /u01/oracle/ggs/dirpcs: created SQL script files /u01/oracle/ggs/dirsql: created Database definitions files /u01/oracle/ggs/dirdef: created Extract data files /u01/oracle/ggs/dirdat: created Temporary files /u01/oracle/ggs/dirtmp: created Veridata files /u01/oracle/ggs/dirver: created Veridata Lock files /u01/oracle/ggs/dirver/lock: created Veridata Out-Of-Sync files /u01/oracle/ggs/dirver/oos: created Veridata Out-Of-Sync XML files /u01/oracle/ggs/dirver/oosxml: created Veridata Parameter files /u01/oracle/ggs/dirver/params: created
Veridata Report files /u01/oracle/ggs/dirver/report: created Veridata Status files /u01/oracle/ggs/dirver/status: created Veridata Trace files /u01/oracle/ggs/dirver/trace: created Stdout files /u01/oracle/ggs/dirout: created We then need to create a database user which will be used by the GoldenGate Manager, Extract and Replicat processes. We can create individual users for each process or configure just a common user in our case we will create the one user GGS_OWNER and grant it the required privileges. SQL> create tablespace ggs_data 2 datafile /u02/oradata/gavin/ggs_data01.dbf size 200m; SQL> create user ggs_owner identified by ggs_owner 2 default tablespace ggs_data 3 temporary tablespace temp; User created. SQL> grant connect,resource to ggs_owner; Grant succeeded. SQL> grant select any dictionary, select any table to ggs_owner; Grant succeeded. SQL> grant create table to ggs_owner; Grant succeeded. SQL> grant flashback any table to ggs_owner; Grant succeeded. SQL> grant execute on dbms_flashback to ggs_owner; Grant succeeded. SQL> grant execute on utl_file to ggs_owner; Grant succeeded. We can then confirm that the GoldenGate user we have just created is able to connect to the Oracle database
$ ggsci Oracle GoldenGate Command Interpreter for Oracle Version 10.4.0.19 Build 002 AIX 5L, ppc, 64bit (optimized), Oracle 11 on Sep 17 2009 23:54:16 Copyright (C) 1995, 2009, Oracle and/or its affiliates. All rights reserved. GGSCI (devu007) 1> DBLOGIN USERID ggs_owner, PASSWORD ggs_owner Successfully logged into database. We also need to enable supplemental logging at the database level otherwise we will get this error when we try to start the Extract process 2010-02-08 13:51:21 GGS ERROR 190 No minimum supplemental logging is enabled. This may cause extract process to handle key update incorrectly if key column is not in first row piece. 2010-02-08 13:51:21 GGS ERROR 190 PROCESS ABENDING. SQL> ALTER DATABASE ADD SUPPLEMENTAL LOG DATA (ALL) COLUMNS; Database altered
Some other recommended optional parameters are AUTOSTART which automatically start the Extract and Replicat processes when the Manager starts. The USERID and PASSWORD parameter and required if you enable GoldenGate DDL support and this is the Oracle user account that we created for the Manager(and Extract/Replicat) as described in the earlier tutorial. The Manager process can also clean up trail files from disk when GoldenGate has finished processing them via the PURGEOLDEXTRACTS parameter. Used with the USECHECKPOINTS clause, it will ensure that until all processes have finished using the data contained in the trail files, they will not be deleted. The following is an example of a manager parameter file [oracle@redhat346 ggs]$ ./ggsci Oracle GoldenGate Command Interpreter for Oracle Version 10.4.0.19 Build 002 Linux, x64, 64bit (optimized), Oracle 11 on Sep 17 2009 23:51:28 Copyright (C) 1995, 2009, Oracle and/or its affiliates. All rights reserved. GGSCI 2> EDIT PARAMS MGR PORT 7809 USERID ggs_owner, PASSWORD ggs_owner PURGEOLDEXTRACTS /u01/oracle/ggs/dirdat/ex, USECHECKPOINTS The manager can be stopped and started via the GSSCI commands START MANAGER and STOP MANAGER. Information on the status of the Manager can be obtained via the INFO MANAGER command GGSCI (devu007) 4> info manager Manager is running (IP port devu007.7809).
The table PRODUCTS in the SH schema on the source has 72 rows and on the target database the same table is present only in structure without any data. We will be loading the 72 rows in this example from the source database to the target database using GoldenGate Direct Load method. On Source 1) Create the Initial data extract process load1 GGSCI (redhat346.localdomain) 5> ADD EXTRACT load1, SOURCEISTABLE EXTRACT added. Since this is a one time data extract task, the source of the data is not the transaction log files of the RDBMS (in this case the online and archive redo log files) but the table data itself, that is why the keyword SOURCEISTABLE is used. 2) Create the parameter file for the extract group load1 EXTRACT: name of the extract group USERID/PASSWORD: the database user which has been configured earlier for Extract (this user is created in the source database) RMTHOST: This will be the IP address or hostname of the target system MGRPORT: the port where the Manager process is running TABLE: specify the table which is being extracted and replicated. This can be specified in a number of ways using wildcard characters to include or exclude tables as well as entire schemas. GGSCI (redhat346.localdomain) 6> EDIT PARAMS load1 EXTRACT load1 USERID ggs_owner, PASSWORD ggs_owner RMTHOST devu007, MGRPORT 7809 RMTTASK replicat, GROUP load2 TABLE sh.products; On Target 3) Create the initial data load task load2 Since this is a one time data load task, we are using the keyword SPECIALRUN GGSCI (devu007) 1> ADD REPLICAT load2, SPECIALRUN REPLICAT added. 4) Create the parameter file for the Replicat group, load2
REPLICAT: name of the Replicat group created for the initial data load USERID/PASSWORD: database credentials for the Replicat user (this user is created in the target database) ASSUMETARGETDEFS: this means that the source table structure exactly matches the target database table structure MAP: with GoldenGate we can have the target database structure entirely differ from that of the source in terms of table names as well as the column definitions of the tables. This parameter provides us the mapping of the source and target tables which is same in this case GGSCI (devu007) 2> EDIT PARAMS load2 /u01/oracle/software/goldengate/dirprm/rep4.prm [New file] REPLICAT load2 USERID ggs_owner, PASSWORD ggs_owner ASSUMETARGETDEFS MAP sh.customers, TARGET sh.customers; On Source SQL> select count(*) from products; COUNT(*) 72 On Target SQL> select count(*) from products; COUNT(*) 0 On Source 5) Start the initial load data extract task on the source system We now start the initial data load task load 1 on the source. Since this is a one time task, we will initially see that the extract process is runningand after the data load is complete it will be stopped. We do not have to manually start the Replicat process on the target as that is done when the Extract task is started on the source system. On Source GGSCI (redhat346.localdomain) 16> START EXTRACT load1
Sending START request to MANAGER EXTRACT LOAD1 starting GGSCI (redhat346.localdomain) 28> info extract load1 EXTRACT LOAD1 Last Started 2010-02-11 11:33 Status RUNNING Checkpoint Lag Not Available Log Read Checkpoint Table SH.PRODUCTS 2010-02-11 11:33:16 Record 72 Task SOURCEISTABLE GGSCI (redhat346.localdomain) 29> info extract load1 EXTRACT LOAD1 Last Started 2010-02-11 11:33 Status STOPPED Checkpoint Lag Not Available Log Read Checkpoint Table SH.PRODUCTS 2010-02-11 11:33:16 Record 72 Task SOURCEISTABLE On Target SQL> select count(*) from products; COUNT(*) 72
Create a Replicat group Create a parameter file for the online Replicat group Create the GoldenGate Checkpoint table GoldenGate maintains its own Checkpoints which is a known position in the trail file from where the Replicat process will start processing after any kind of error or shutdown. This ensures data integrity and a record of these checkpoints is either maintained in files stored on disk or table in the database which is the preferred option. We can also create a single Checkpoint table which can used by all Replicat groups from the single or many GoldenGate instances. In one of the earlier tutorials we had created the GLOBALS file. We now need to edit that GLOBALS file and add an entry for CHECKPOINTTABLE which will include the checkpoint table name which will be available to all Replicat processes via the EDIT PARAMS command. GGSCI (devu007) 2> EDIT PARAMS ./GLOBALS GGSCHEMA GGS_OWNER CHECKPOINTTABLE GGS_OWNER.CHKPTAB GGSCI (devu007) 4> DBLOGIN USERID ggs_owner, PASSWORD ggs_owner Successfully logged into database. GGSCI (devu007) 6> ADD CHECKPOINTTABLE GGS_OWNER.CHKPTAB Successfully created checkpoint table GGS_OWNER.CHKPTAB.
apex:/u01/oracle/software/goldengate> sqlplus ggs_owner/ggs_owner SQL*Plus: Release 11.1.0.6.0 - Production on Mon Feb 8 09:02:19 2010 Copyright (c) 1982, 2007, Oracle. All rights reserved.
Connected to: Oracle Database 11g Enterprise Edition Release 11.1.0.6.0 - 64bit Production With the Partitioning, OLAP, Data Mining and Real Application Testing options SQL> desc chkptab Name Null? Type ----------------------------------------- -------- --------------------------GROUP_NAME NOT NULL VARCHAR2(8) GROUP_KEY NOT NULL NUMBER(19) SEQNO NUMBER(10) RBA NOT NULL NUMBER(19) AUDIT_TS VARCHAR2(29)
Create the Online Extract Group GGSCI (redhat346.localdomain) 1> ADD EXTRACT ext1, TRANLOG, BEGIN NOW EXTRACT added. Create the Trail We now create a trail note that this path pertains to the GoldenGate software location on the target system and this is where the trail files will be created having a prefix rt which will be used by the Replicat process also running on the target system GGSCI (redhat346.localdomain) 2> ADD RMTTRAIL /u01/oracle/software/goldengate/dirdat/rt, EXTRACT ext1 RMTTRAIL added. Create a parameter file for the online Extract group ext1 GGSCI (redhat346.localdomain) 3> EDIT PARAMS ext1 EXTRACT ext1 USERID ggs_owner, PASSWORD ggs_owner RMTHOST devu007, MGRPORT 7809 RMTTRAIL /u01/oracle/software/goldengate/dirdat/rt TABLE scott.emp; ON TARGET SYSTEM Create the online Replicat group GGSCI (devu007) 7> ADD REPLICAT rep1, EXTTRAIL /u01/oracle/software/goldengate/dirdat/rt REPLICAT added. Note that the EXTTRAIL location which is on the target local system conforms to the RMTTRAIL parameter which we used when we created the parameter file for the extract process on the source system. Create a parameter file for the online Replicat group, rep1 GGSCI (devu007) 8> EDIT PARAMS rep1 REPLICAT rep1 ASSUMETARGETDEFS
USERID ggs_owner, PASSWORD ggs_owner MAP scott.emp, TARGET sh.emp; ON SOURCE Start the Extract process GGSCI (redhat346.localdomain) 16> START EXTRACT ext1 Sending START request to MANAGER EXTRACT EXT1 starting GGSCI (redhat346.localdomain) 17> STATUS EXTRACT ext1 EXTRACT EXT1: RUNNING GGSCI (redhat346.localdomain) 16> INFO EXTRACT ext1 EXTRACT EXT1 Last Started 2010-02-08 14:27 Status RUNNING Checkpoint Lag 00:00:00 (updated 00:00:09 ago) Log Read Checkpoint Oracle Redo Logs 2010-02-08 14:27:48 Seqno 145, RBA 724480 ON TARGET Start the Replicat process GGSCI (devu007) 1> START REPLICAT rep1 Sending START request to MANAGER REPLICAT REP1 starting GGSCI (devu007) 2> INFO REPLICAT rep1 REPLICAT REP1 Last Started 2010-02-08 14:55 Status RUNNING Checkpoint Lag 00:00:00 (updated 00:00:01 ago) Log Read Checkpoint File /u01/oracle/software/goldengate/dirdat/rt000001 2010-02-08 14:27:57.600425 RBA 1045 Note: the trail file has a prefix of rt (which we had defined earlier) LET US NOW TEST ON SOURCE SQL> conn scott/tiger Connected.
SQL> UPDATE emp SET sal=9999 WHERE ename=KING; 1 row updated. SQL> COMMIT; Commit complete. ON TARGET SQL> SELECT SAL FROM emp WHERE ename=KING; SAL 9999
GGSCI (devu007) 3> ADD EXTTRAIL /u01/oracle/software/goldengate/dirdat/lt, EXTRACT ext1 EXTTRAIL added. Create the Data Pump group On the source system create the Data Pump group and using the EXTTRAILSOURCE keywork specify the location of the local trail which will be read by the Data Pump process GGSCI (devu007) 4> ADD EXTRACT dpump, EXTTRAILSOURCE /u01/oracle/software/goldengate/dirdat/lt EXTRACT added. Create the parameter file for the Primary Extract group GGSCI (devu007) 5> EDIT PARAMS ext1 /u01/oracle/software/goldengate/dirprm/ext1.prm [New file] EXTRACT ext1 USERID ggs_owner, PASSWORD ggs_owner EXTTRAIL /u01/oracle/software/goldengate/dirdat/lt TABLE MONITOR.WORK_PLAN; Specify the location of the remote trail on the target system Use the RMTTRAIL to specify the location of the remote trail and associate the same with the Data Pump group as it will be wriiten to over the network by the data pump process GGSCI (devu007) 6> ADD RMTTRAIL /u01/oracle/ggs/dirdat/rt, EXTRACT dpump RMTTRAIL added. Create the parameter file for the Data Pump group Note- the parameter PASSTHRU signifies the mode being used for the Data Pump which means that the names of the source and target objects are identical and no column mapping or filtering is being performed here. GGSCI (devu007) 2> EDIT PARAMS dpump /u01/oracle/software/goldengate/dirprm/dpump.prm [New file] EXTRACT dpump USERID ggs_owner, PASSWORD ggs_owner RMTHOST redhat346, MGRPORT 7809 RMTTRAIL /u01/oracle/ggs/dirdat/rt
PASSTHRU TABLE MONITOR.WORK_PLAN; ON TARGET SYSTEM Create the Replicat group The EXTTRAIL clause indicates the location of the remote trail and should be the same as the RMTTRAIL value that was used when creating the Data Pump process on the source system. GGSCI (redhat346.localdomain) 2> ADD REPLICAT rep1, EXTTRAIL /u01/oracle/ggs/dirdat/rt REPLICAT added. Create the parameter file for the Replicat group GGSCI (redhat346.localdomain) 3> EDIT PARAMS rep1 REPLICAT rep1 ASSUMETARGETDEFS USERID ggs_owner, PASSWORD ggs_owner MAP MONITOR.WORK_PLAN, TARGET MONITOR.WORK_PLAN; ON SOURCE On the source system, now start the Extract and Data Pump processes. GGSCI (devu007) 3> START EXTRACT ext1 Sending START request to MANAGER EXTRACT EXT1 starting GGSCI (devu007) 4> START EXTRACT dpump Sending START request to MANAGER EXTRACT DPUMP starting GGSCI (devu007) 5> info extract ext1 EXTRACT EXT1 Last Started 2010-02-18 11:23 Status RUNNING Checkpoint Lag 00:40:52 (updated 00:00:09 ago) Log Read Checkpoint Oracle Redo Logs 2010-02-18 10:42:19 Seqno 761, RBA 15086096 GGSCI (devu007) 6> INFO EXTRACT dpump
EXTRACT DPUMP Last Started 2010-02-18 11:23 Status RUNNING Checkpoint Lag 00:00:00 (updated 00:00:02 ago) Log Read Checkpoint File /u01/oracle/software/goldengate/dirdat/lt000000 2010-02-18 11:15:10.000000 RBA 5403 Note- the data pump process is reading from the Local Trail file /u01/oracle/software/goldengate/dirdat/lt000000 ON TARGET SYSTEM Start the Replicat process GGSCI (redhat346.localdomain) 4> START REPLICAT rep1 Sending START request to MANAGER REPLICAT REP1 starting GGSCI (redhat346.localdomain) 5> STATUS REPLICAT rep1 REPLICAT REP1: RUNNING
Marker setup table script complete, running verification script... Please enter the name of a schema for the GoldenGate database objects: Setting schema name to GGS_OWNER MARKER TABLE ------------------------------OK MARKER SEQUENCE ------------------------------OK Script complete.
SQL> alter session set recyclebin=OFF; Session altered. SQL> @ddl_setup GoldenGate DDL Replication setup script Verifying that current user has privileges to install DDL Replication... You will be prompted for the name of a schema for the GoldenGate database objects. NOTE: The schema must be created prior to running this script. NOTE: On Oracle 10g and up, system recycle bin must be disabled. NOTE: Stop all DDL replication before starting this installation. Enter GoldenGate schema name:GGS_OWNER You will be prompted for the mode of installation. To install or reinstall DDL replication, enter INITIALSETUP To upgrade DDL replication, enter NORMAL Enter mode of installation:INITIALSETUP Working, please wait ... Spooling to file ddl_setup_spool.txt Using GGS_OWNER as a GoldenGate schema name, INITIALSETUP as a mode of installation. Working, please wait ... RECYCLEBIN must be empty. This installation will purge RECYCLEBIN for all users. To proceed, enter yes. To stop installation, enter no. Enter yes or no:yes DDL replication setup script complete, running verification script...
Please enter the name of a schema for the GoldenGate database objects: Setting schema name to GGS_OWNER DDLORA_GETTABLESPACESIZE STATUS: Line/pos Error ---------- ----------------------------------------------------------------No errors No errors CLEAR_TRACE STATUS: Line/pos Error ---------- ----------------------------------------------------------------No errors No errors CREATE_TRACE STATUS: Line/pos Error ---------- ----------------------------------------------------------------No errors No errors TRACE_PUT_LINE STATUS: Line/pos Error ---------- ----------------------------------------------------------------No errors No errors INITIAL_SETUP STATUS: Line/pos Error ---------- ----------------------------------------------------------------No errors No errors DDLVERSIONSPECIFIC PACKAGE STATUS: Line/pos Error ---------- ----------------------------------------------------------------No errors No errors DDLREPLICATION PACKAGE STATUS: Line/pos Error ---------- ----------------------------------------------------------------No errors No errors DDLREPLICATION PACKAGE BODY STATUS: Line/pos Error ---------- ----------------------------------------------------------------No errors No errors DDL HISTORY TABLE ----------------------------------OK DDL HISTORY TABLE(1) -----------------------------------
OK DDL DUMP TABLES ----------------------------------OK DDL DUMP COLUMNS ----------------------------------OK DDL DUMP LOG GROUPS ----------------------------------OK DDL DUMP PARTITIONS ----------------------------------OK DDL DUMP PRIMARY KEYS ----------------------------------OK DDL SEQUENCE ----------------------------------OK GGS_TEMP_COLS ----------------------------------OK GGS_TEMP_UK ----------------------------------OK DDL TRIGGER CODE STATUS: Line/pos Error ---------- ----------------------------------------------------------------No errors No errors DDL TRIGGER INSTALL STATUS ----------------------------------OK DDL TRIGGER RUNNING STATUS ----------------------------------ENABLED STAYMETADATA IN TRIGGER ----------------------------------OFF DDL TRIGGER SQL TRACING ----------------------------------0 DDL TRIGGER TRACE LEVEL
----------------------------------0 LOCATION OF DDL TRACE FILE ------------------------------------------------------------------------------/u01/app/oracle/diag/rdbms/gavin/gavin/trace/ggs_ddl_trace.log Analyzing installation status... STATUS OF DDL REPLICATION ------------------------------------------------------------------------------SUCCESSFUL installation of DDL Replication software components Script complete. SQL>
SQL> @role_setup GGS Role setup script This script will drop and recreate the role GGS_GGSUSER_ROLE To use a different role name, quit this script and then edit the params.sql script to change the gg_role parameter to the preferred name. (Do not run the script.) You will be prompted for the name of a schema for the GoldenGate database objects. NOTE: The schema must be created prior to running this script. NOTE: Stop all DDL replication before starting this installation. Enter GoldenGate schema name:GGS_OWNER Wrote file role_setup_set.txt PL/SQL procedure successfully completed. Role setup script complete Grant this role to each user assigned to the Extract, GGSCI, and Manager processes, by using the following SQL command: GRANT GGS_GGSUSER_ROLE TO where is the user assigned to the GoldenGate processes.
SQL> @ddl_pin GGS_OWNER PL/SQL procedure successfully completed. PL/SQL procedure successfully completed. PL/SQL procedure successfully completed.
Turn Recyclebin OFF We need to set the parameter recyclebin to OFF via the ALTER SYSTEM SET RECYCLEBIN=OFF command in order to prevent this error which we will see if we try and configure DDL support and then start the Extract process. 2010-02-19 11:13:30 GGS ERROR 2003 RECYCLEBIN must be turned off. For 10gr2 and up, set RECYCLEBIN in parameter file to OFF. For 10gr1, set _RECYCLEBI N in parameter file to FALSE. Then restart database and extract. 2010-02-19 11:13:30 GGS ERROR 190 PROCESS ABENDING. Enable additional logging at the table level Note- We had earlier enabled additional supplemental logging at the database level. Using the ADD TRANDATA command we now enable it at even the table level as this is required by GoldenGate for DDL support. GGSCI (redhat346.localdomain) 5> DBLOGIN USERID ggs_owner, PASSWORD ggs_owner Successfully logged into database. GGSCI (redhat346.localdomain) 6> ADD TRANDATA scott.emp Logging of supplemental redo data enabled for table SCOTT.EMP. Edit the parameter file for the Extract process to enable DDL synchronization We had earlier created a parameter file for an Extract process ext1. We now edit that parameter file and add the entry DDL INCLUDE MAPPED This means that DDL support is now enabled for all tables which have been mapped and in this case it will only apply to the SCOTT.EMP table as that is the only table which is being processed
here. We can also use the INCLUDE ALL or EXCLUDE ALL or wildcard characters to specify which tables to enable the DDL support for. GGSCI (redhat346.localdomain) 1> EDIT PARAM EXT1 EXTRACT ext1 USERID ggs_owner, PASSWORD ggs_owner RMTHOST 10.53.100.100, MGRPORT 7809 RMTTRAIL /u01/oracle/software/goldengate/dirdat/rt DDL INCLUDE MAPPED TABLE scott.emp; Test the same We will now alter the structure of the EMP table by adding a column and we can see that this new table structure is also reflected on the target system. On Source SQL> ALTER TABLE EMP ADD NEW_COL VARCHAR2(10); Table altered. On Target SQL> desc emp Name Null? Type EMPNO NOT NULL NUMBER(4) ENAME VARCHAR2(10) JOB VARCHAR2(20) MGR NUMBER(4) HIREDATE DATE SAL NUMBER(7,2) COMM NUMBER(7,2) DEPTNO NUMBER(2) MYCOL VARCHAR2(10) NEW_COL VARCHAR2(10)
So we can use GoldenGate when the source and database database differ in table structure as well as an ETL tool in a Datawarehouse type environment. We will discuss below two examples to demonstrate this feature column mapping and filtering of data. In example 1, we will filter the records that are extracted on the source and applied on the target only rows where the JOB column value equals MANAGER in the MYEMP table will be considered for extraction. In example 2, we will deal with a case where the table structure is different between the source database and the target database and see how column mapping is performed in such cases. Example 1 Initial load of all rows which match the filter from source to target. The target database MYEMP table will only be populated with rows from the EMP table where filter criteria of JOB=MANAGER is met. On Source GGSCI (redhat346.localdomain) 4> add extract myload1, sourceistable EXTRACT added. GGSCI (redhat346.localdomain) 5> edit params myload1 EXTRACT myload1 USERID ggs_owner, PASSWORD ggs_owner RMTHOST devu007, MGRPORT 7809 RMTTASK replicat, GROUP myload1 TABLE scott.myemp, FILTER (@STRFIND (job, MANAGER) > 0); On Target GGSCI (devu007) 2> add replicat myload1, specialrun REPLICAT added. GGSCI (devu007) 3> edit params myload1 /u01/oracle/software/goldengate/dirprm/myload1.prm [New file] REPLICAT myload1 USERID ggs_owner, PASSWORD ggs_owner ASSUMETARGETDEFS MAP scott.myemp, TARGET sh.myemp; On Source start the initial load extract
GGSCI (redhat346.localdomain) 6> start extract myload1 Sending START request to MANAGER EXTRACT MYLOAD1 starting On SOURCE SQL> select count(*) from myemp; COUNT(*) 14 SQL> select count(*) from myemp where job=MANAGER; COUNT(*) 9 On TARGET SQL> select count(*) from myemp where job=MANAGER; COUNT(*) 9 Create an online change extract and replicat group using a Filter GGSCI (redhat346.localdomain) 10> add extract myload2, tranlog, begin now EXTRACT added. GGSCI (redhat346.localdomain) 11> add rmttrail /u01/oracle/software/goldengate/dirdat/bb, extract myload2 RMTTRAIL added. GGSCI (redhat346.localdomain) 11> edit params myload2 EXTRACT myload2 USERID ggs_owner, PASSWORD ggs_owner RMTHOST 10.53.200.225, MGRPORT 7809 RMTTRAIL /u01/oracle/software/goldengate/dirdat/bb TABLE scott.myemp, FILTER (@STRFIND (job, MANAGER) > 0); On Target
GGSCI (devu007) 2> add replicat myload2, exttrail /u01/oracle/software/goldengate/dirdat/bb REPLICAT added. GGSCI (devu007) 3> edit params myload2 /u01/oracle/software/goldengate/dirprm/myload2.prm [New file] REPLICAT myload2 ASSUMETARGETDEFS USERID ggs_owner, PASSWORD ggs_owner MAP scott.myemp, TARGET sh.myemp; On Source start the online extract group GGSCI (redhat346.localdomain) 13> start extract myload2 Sending START request to MANAGER EXTRACT MYLOAD2 starting GGSCI (redhat346.localdomain) 14> info extract myload2 EXTRACT MYLOAD2 Last Started 2010-02-23 11:04 Status RUNNING Checkpoint Lag 00:27:39 (updated 00:00:08 ago) Log Read Checkpoint Oracle Redo Logs 2010-02-23 10:36:51 Seqno 214, RBA 103988 On Target GGSCI (devu007) 4> start replicat myload2 Sending START request to MANAGER REPLICAT MYLOAD2 starting GGSCI (devu007) 5> info replicat myload2 REPLICAT MYLOAD2 Last Started 2010-02-23 11:05 Status RUNNING Checkpoint Lag 00:00:00 (updated 00:00:08 ago) Log Read Checkpoint File /u01/oracle/software/goldengate/dirdat/bb000000 First Record RBA 989 On Source we now insert two rows into the MYEMP table one which has the JOB value of MANAGER and the other row which has the job value of SALESMAN
On SOURCE
SQL> INSERT INTO MYEMP 2 (empno,ename,job,sal) 3 VALUES 4 (1234,GAVIN,MANAGER,10000); 1 row created. SQL> commit; Commit complete. SQL> INSERT INTO MYEMP 2 (empno,ename,job,sal) 3 VALUES 4 (1235,BOB,SALESMAN,1000); 1 row created. SQL> commit; Commit complete. SQL> select count(*) from myemp; COUNT(*) 16 SQL> select count(*) from myemp where job=MANAGER; COUNT(*) 10 On Target, we will see that even though two rows have been inserted into the source MYEMP table, on the target MYEMP table only one row is inserted because the filter has been applied which only includes the rows where the JOB value equals MANAGER. SQL> select count(*) from myemp; COUNT(*) 10 Example 2 source and target table differ in column structure
In the source MYEMP table we have a column named SAL whereas on the target, the same MYEMP table has the column defined as SALARY. Create a definitions file on the source using DEFGEN utility and then copy that definitions file to the target system GGSCI (redhat346.localdomain) > EDIT PARAMS defgen DEFSFILE /u01/oracle/ggs/dirsql/myemp.sql USERID ggs_owner, PASSWORD ggs_owner TABLE scott.myemp; [oracle@redhat346 ggs]$ ./defgen paramfile /u01/oracle/ggs/dirprm/defgen.prm *********************************************************************** Oracle GoldenGate Table Definition Generator for Oracle Version 10.4.0.19 Build 002 Linux, x64, 64bit (optimized), Oracle 11 on Sep 18 2009 00:09:13 Copyright (C) 1995, 2009, Oracle and/or its affiliates. All rights reserved. Starting at 2010-02-23 11:22:17 *********************************************************************** Operating System Version: Linux Version #1 SMP Wed Dec 17 11:41:38 EST 2008, Release 2.6.18-128.el5 Node: redhat346.localdomain Machine: x86_64 soft limit hard limit Address Space Size : unlimited unlimited Heap Size : unlimited unlimited File Size : unlimited unlimited CPU Time : unlimited unlimited Process id: 14175 *********************************************************************** ** Running with the following parameters ** *********************************************************************** DEFSFILE /u01/oracle/ggs/dirsql/myemp.sql USERID ggs_owner, PASSWORD ********* TABLE scott.myemp; Retrieving definition for SCOTT.MYEMP Definitions generated for 1 tables in /u01/oracle/ggs/dirsql/myemp.sql
If we were to try and run the replicat process on the target without copying the definitions file, we will see an error as shown below which pertains to the fact that the columns in the source and target database are different and GoldenGate is not able to resolve that. 2010-02-23 11:31:07 GGS WARNING 218 Aborted grouped transaction on SH.MYEMP, Database error 904 (ORA-00904: SAL: invalid identifier). 2010-02-23 11:31:07 GGS WARNING 218 SQL error 904 mapping SCOTT.MYEMP to SH.MYEMP OCI Error ORA-00904: SAL: invalid identifier (status = 904), SQL . We then ftp the definitions file from the source to the target system in this case to the dirsql directory located in the top level GoldenGate installed software directory We now go and make a change to the original replicat parameter file and change the parameter ASSUMEDEFS to SOURCEDEFS which provides GoldenGate with the location of the definitions file. The other parameter which is included is the COLMAP parameter which tells us how the column mapping has been performed. The USEDEFAULTS keyword denotes that all the other columns in both tables are identical except for the columns SAL and SALARY which differ in both tables and now we are mapping the SAL columsn in source to the SALARY column on the target. REPLICAT myload2 SOURCEDEFS /u01/oracle/software/goldengate/dirsql/myemp.sql USERID ggs_owner, PASSWORD ggs_owner MAP scott.myemp, TARGET sh.myemp, COLMAP (usedefaults, salary = sal); We now go and start the originall replicat process myload2 which had abended because of the column mismatch (which has now been corrected via the parameter change) and we see that the process now is running without any error. now go and start the process which had failed after table modification GGSCI (devu007) 2> info replicat myload2 REPLICAT MYLOAD2 Last Started 2010-02-23 11:05 Status ABENDED Checkpoint Lag 00:00:03 (updated 00:11:44 ago) Log Read Checkpoint File /u01/oracle/software/goldengate/dirdat/bb000000 2010-02-23 11:31:03.999504 RBA 1225 GGSCI (devu007) 3> start replicat myload2
Sending START request to MANAGER REPLICAT MYLOAD2 starting GGSCI (devu007) 4> info replicat myload2 REPLICAT MYLOAD2 Last Started 2010-02-23 11:43 Status RUNNING Checkpoint Lag 00:00:00 (updated 00:00:03 ago) Log Read Checkpoint File /u01/oracle/software/goldengate/dirdat/bb000000 2010-02-23 11:31:03.999504 RBA 1461
Find the run status of a particular process GGSCI (devu007) 23> status manager Manager is running (IP port devu007.7809). GGSCI (devu007) 24> status extract ext1 EXTRACT EXT1: RUNNING Detailed information of a particular process
GGSCI (devu007) 6> info extract ext1, detail EXTRACT EXT1 Checkpoint Lag Log Read Checkpoint Last Started 2010-02-19 11:19 Status RUNNING 00:00:00 (updated 00:00:02 ago) Oracle Redo Logs 2010-02-26 10:45:18 Seqno 786, RBA 44710400
Target Extract Trails: Remote Trail Name MB /u01/oracle/software/goldengate/dirdat/lt 10 Extract Source /u02/oradata/apex/redo03.log /u02/oradata/apex/redo02.log /u02/oradata/apex/redo02.log Not Available Current directory Report file Parameter file Checkpoint file Process file Stdout file Error log Begin 2010-02-19 11:13 2010-02-19 11:04 2010-02-18 10:42 * Initialized * End 2010-02-26 2010-02-19 2010-02-19 2010-02-18 10:45 11:13 11:04 10:42 2 55644 Seqno RBA Max
Monitoring an Extract recovery GGSCI (devu007) 35> send extract ext1 status Sending STATUS request to EXTRACT EXT1 ... EXTRACT EXT1 (PID 1925238) Current status: Recovery complete: At EOF Sequence #: 786 RBA: 40549888 Timestamp: 2010-02-26 09:59:57.000000 Output trail #1 Current write position: Sequence #: 2 RBA: 55644 Timestamp: 2010-02-26 09:59:54.337574 Extract Trail: /u01/oracle/software/goldengate/dirdat/lt Monitoring processing volume - Statistics of the operations processed GGSCI (devu007) 33> stats extract ext1
Sending STATS request to EXTRACT EXT1 ... Start of Statistics at 2010-02-26 09:58:27. DDL replication statistics (for all trails): *** Total statistics since extract started Operations Mapped operations Unmapped operations Other operations Excluded operations *** 19.00 2.00 9.00 8.00 17.00
Output to /u01/oracle/software/goldengate/dirdat/lt: Extracting from GGS_OWNER.GGS_MARKER to GGS_OWNER.GGS_MARKER: *** Total statistics since 2010-02-19 11:21:03 *** No database operations have been performed. *** Daily statistics since 2010-02-26 00:00:00 *** No database operations have been performed. *** Hourly statistics since 2010-02-26 09:00:00 *** No database operations have been performed. *** Latest statistics since 2010-02-19 11:21:03 *** No database operations have been performed. Extracting from MONITOR.WORK_PLAN to MONITOR.WORK_PLAN: *** Total statistics since 2010-02-19 11:21:03 *** Total inserts Total updates Total deletes Total discards Total operations *** Daily statistics since 2010-02-26 00:00:00 *** Total inserts Total updates Total deletes Total discards Total operations *** Hourly statistics since 2010-02-26 09:00:00 *** No database operations have been performed. *** Latest statistics since 2010-02-19 11:21:03 *** Total inserts 4.00 Total updates 46.00 4.00 46.00 0.00 0.00 50.00 0.00 16.00 0.00 0.00 16.00
View processing rate - can use 'hr','min' or 'sec' as a parameter GGSCI (devu007) 37> stats extract ext2 reportrate hr Sending STATS request to EXTRACT EXT2 ... Start of Statistics at 2010-02-26 10:04:46. Output to /u01/oracle/ggs/dirdat/cc: Extracting from SH.CUSTOMERS to SH.CUSTOMERS: *** Total statistics since 2010-02-26 09:29:48 *** Total inserts/hour: 0.00 Total updates/hour: 95258.62 Total deletes/hour: 0.00 Total discards/hour: 0.00 Total operations/hour: 95258.62 *** Daily statistics since 2010-02-26 09:29:48 *** Total inserts/hour: 0.00 Total updates/hour: 95258.62 Total deletes/hour: 0.00 Total discards/hour: 0.00 Total operations/hour: 95258.62 *** Hourly statistics since 2010-02-26 10:00:00 *** No database operations have been performed. *** Latest statistics since 2010-02-26 09:29:48 *** Total inserts/hour: 0.00 Total updates/hour: 95258.62 Total deletes/hour: 0.00 Total discards/hour: 0.00 Total operations/hour: 95258.62 End of Statistics. View latency between the records processed by Goldengate and the timestamp in the data source GGSCI (devu007) 13> send extract ext2, getlag
Sending GETLAG request to EXTRACT EXT2 ... Last record lag: 3 seconds. At EOF, no more records to process.
GGSCI (devu007) 15> lag extract ext* Sending GETLAG request to EXTRACT EXT1 ... Last record lag: 1 seconds. At EOF, no more records to process. Sending GETLAG request to EXTRACT EXT2 ... Last record lag: 1 seconds. At EOF, no more records to process.
Viewing the GoldenGate error log as well as history of commands executed and other events We can use the editor depending on operating system vi on Unix for example to view the ggserr.log file which is located at the top level GoldenGate software installation directory. We can also use the GGSCI command VIEW GGSEVT as well. View the process report Every Manager, Extract and Replicat process will generate a report file at the end of each run and this report can be viewed to diagnose any problems or errors as well as view the parameters used, the environment variables is use, memory consumption etc For example:
GGSCI (devu007) 2> view report ext1 GGSCI (devu007) 2> view report rep1 GGSCI (devu007) 2> view report mgr Information on Child processes started by the Manager GGSCI (devu007) 8> send manager childstatus Sending CHILDSTATUS request to MANAGER ... Child Process Status - 6 Entries ID ---0 1 2 4 6 7 Group -------EXT1 DPUMP MSSQL1 MYREP MYLOAD2 EXT2 Process ---------1925238 2195496 422034 1302702 1200242 2076844 Retry ----0 0 0 0 0 0 Retry Time -----------------None None None None None None Start Time ----------2010/02/19 11:07:54 2010/02/19 11:08:02 2010/02/22 13:54:59 2010/02/23 09:08:34 2010/02/23 11:05:01 2010/02/26 08:29:22
Oracle GoldenGate Tutorial 10- performing a zero downtime cross platform migration and 11g database upgrade
This note briefly describes the steps required to perform a cross platform database migration (AIX to Red Hat Linux) and also a database upgrade from 10g to 11g Release 2 which is attained with zero downtime using a combination of RMAN, Cross Platform TTS and GoldenGate to achieve the same. This is the environment that we will be referring to in this note: 10..2.0.4 Database on AIX DB10g 10.2.0.4 Duplicate database on AIX Clonedb 11.2 database on Linux DB11g Steps 1) Create the GoldenGate Extract process on source AIX DB10g and start the same. This extract process will be capturing changes as they occur on the 10g AIX database in the remote trail files located on the Linux target system. Since the Replicat process is not running on the target at this time, the source database changes will accumulate in the extract trail files. GGSCI (devu026) 12> add extract myext, tranlog, begin now EXTRACT added. GGSCI (devu026) 13> add rmttrail /u01/oracle/ggs/dirdat/my, extract myext RMTTRAIL added. GGSCI (devu026) 14> edit params myext /u01/rapmd2/ggs/dirprm/myext.prm 7 lines, 143 characters EXTRACT myext USERID ggs_owner, PASSWORD ggs_owner SETENV (ORACLE_HOME = /u01/oracle/product/10.2/rapmd2) SETENV (ORACLE_SID = db10g) RMTHOST 10.1.210.35, MGRPORT 7809 RMTTRAIL /u01/oracle/ggs/dirdat/my DISCARDFILE discard.txt, APPEND TABLE sh.*; TABLE hr.*; TABLE pm.*;
TABLE oe.*; TABLE ix.*; START THE EXTRACT PROCESS NOW GGSCI (devu026) 16> START EXTRACT MYEXT Sending START request to MANAGER EXTRACT MYEXT starting GGSCI (devu026) 17> INFO EXTRACT MYEXT EXTRACT MYEXT Last Started 2010-03-04 08:42 Status RUNNING Checkpoint Lag 00:31:07 (updated 00:00:01 ago) Log Read Checkpoint Oracle Redo Logs 2010-03-04 08:11:26 Seqno 8, RBA 2763280 2) Using RMAN create a duplicate database in the source AIX environment (Clonedb) this database will be used as the source for the export of database structure (no rows export) and tablespace meta data Follow this white paper to get all the steps involved. ***********ON SOURCE UPDATE 1********** SQL> conn sh/sh Connected. SQL> update mycustomers set cust_city=Singapore; 55500 rows updated. SQL> commit; Commit complete. 3) Create a skeleton database on the Linux platform in the 11g Release 2 environment DB11g Note we will then set up the GoldenGate user GGS_OWNER in the database and grant it the required privileges as well as create the checkpoint table. Read one of the earlier tutorials which details the set up of the GGS_OWNER user in the database. 4) Take a full export of the database without any table data to get just the structure of the database this is now taken from the clonedb duplicate database created in step 2
db10g:/u01/oracle> expdp dumpfile=full_norows.dmp directory =dumpdir content=metadata_only exclude=tables,index full=y Export: Release 10.2.0.4.0 64bit Production on Thursday, 04 March, 2010 9:02:44 Copyright (c) 2003, 2007, Oracle. All rights reserved. Username: sys as sysdba Password: Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 64bit Production With the Partitioning, Data Mining and Real Application Testing options FLASHBACK automatically enabled to preserve database integrity. Starting SYS.SYS_EXPORT_FULL_01: sys/******** AS SYSDBA dumpfile=full_norows.dmp directory =dumpdir content=metadata_only exclude=tables,index full=y Processing object type DATABASE_EXPORT/TABLESPACE Processing object type DATABASE_EXPORT/PROFILE Processing object type DATABASE_EXPORT/SYS_USER/USER Processing object type DATABASE_EXPORT/SCHEMA/USER Processing object type DATABASE_EXPORT/ROLE Processing object type DATABASE_EXPORT/GRANT/SYSTEM_GRANT/PROC_SYSTEM_GRANT Processing object type DATABASE_EXPORT/SCHEMA/GRANT/SYSTEM_GRANT Processing object type DATABASE_EXPORT/SCHEMA/ROLE_GRANT . 5) Import the dumpfile into the 11g database DB11g which has the database structure without the table data this will create all the users, roles, synonyms etc We had to create a role and also create the directory before doing the full database import. Ignore he errors during the import as it will pertain to objects which already exist in the scratch database. SQL> create role xdbwebservices; Role created. SQL> create directory dumpdir as /u01/oracle; Directory created. [oracle@redhat346 ~]$ impdp dumpfile=full_norows.dmp directory=dumpdir full=y Import: Release 11.2.0.1.0 Production on Mon Mar 8 13:09:16 2010
Copyright (c) 1982, 2009, Oracle and/or its affiliates. All rights reserved. . 6) On the clonedb database, we now will export the tablespace meta data make the required tablespaces read only. Note that the original source 10g database is in read write mode and is being accessed by the users with no downtime as yet. clonedb:/u01/rapmd2/ggs> expdp dumpfile=tts_meta.dmp directory =dumpdir transport_tablespaces=EXAMPLE,TTS Export: Release 10.2.0.4.0 64bit Production on Monday, 08 March, 2010 13:01:38 Copyright (c) 2003, 2007, Oracle. All rights reserved. Username: sys as sysdba Password: Connected to: Oracle Database 10g Enterprise Edition Release 10.2.0.4.0 64bit Production With the Partitioning, Data Mining and Real Application Testing options Starting SYS.SYS_EXPORT_TRANSPORTABLE_01: sys/******** AS SYSDBA dumpfile=tts_meta.dmp directory =dumpdir transport_tablespaces=EXAMPLE,TTS Processing object type TRANSPORTABLE_EXPORT/PLUGTS_BLK Processing object type TRANSPORTABLE_EXPORT/TYPE/TYPE_SPEC Processing object type TRANSPORTABLE_EXPORT/TABLE Processing object type TRANSPORTABLE_EXPORT/GRANT/OWNER_GRANT/OBJECT_GRANT Processing object type TRANSPORTABLE_EXPORT/INDEX Processing object type TRANSPORTABLE_EXPORT/CONSTRAINT/CONSTRAINT Processing object type TRANSPORTABLE_EXPORT/INDEX_STATISTICS Processing object type TRANSPORTABLE_EXPORT/COMMENT Processing object type TRANSPORTABLE_EXPORT/CONSTRAINT/REF_CONSTRAINT Processing object type TRANSPORTABLE_EXPORT/TRIGGER Processing object type TRANSPORTABLE_EXPORT/INDEX/FUNCTIONAL_AND_BITMAP/INDEX Processing object type TRANSPORTABLE_EXPORT/INDEX/STATISTICS/FUNCTIONAL_AND_BITMAP/INDE X_STATISTICS Processing object type TRANSPORTABLE_EXPORT/TABLE_STATISTICS Processing object type TRANSPORTABLE_EXPORT/DOMAIN_INDEX/TABLE Processing object type TRANSPORTABLE_EXPORT/DOMAIN_INDEX/SECONDARY_TABLE/INDEX Processing object type TRANSPORTABLE_EXPORT/DOMAIN_INDEX/INDEX Processing object type TRANSPORTABLE_EXPORT/MATERIALIZED_VIEW Processing object type
TRANSPORTABLE_EXPORT/POST_INSTANCE/PROCACT_INSTANCE Processing object type TRANSPORTABLE_EXPORT/POST_INSTANCE/PROCDEPOBJ Processing object type TRANSPORTABLE_EXPORT/POST_INSTANCE/PLUGTS_BLK Master table SYS.SYS_EXPORT_TRANSPORTABLE_01 successfully loaded/unloaded ****************************************************************************** Dump file set for SYS.SYS_EXPORT_TRANSPORTABLE_01 is: /u01/oracle/tts_meta.dmp Job SYS.SYS_EXPORT_TRANSPORTABLE_01 successfully completed at 13:02:17 7) Copy the datafiles from the read only tablespaces ( from clonedb) to the target Linux system and using RMAN convert the datafiles from the AIX platform to the Linux platform RMAN> CONVERT DATAFILE /u01/oracle/example01.dbf 2> FROM PLATFORM=AIX-Based Systems (64-bit) 3> FORMAT /u02/oradata/db11g/example01.dbf; Starting conversion at target at 08-MAR-10 using target database control file instead of recovery catalog allocated channel: ORA_DISK_1 channel ORA_DISK_1: SID=11 device type=DISK channel ORA_DISK_1: starting datafile conversion input file name=/u01/oracle/example01.dbf converted datafile=/u02/oradata/db11g/example01.dbf channel ORA_DISK_1: datafile conversion complete, elapsed time: 00:00:03 Finished conversion at target at 08-MAR-10 RMAN> CONVERT DATAFILE /u01/oracle/tts01.dbf 2> FROM PLATFORM=AIX-Based Systems (64-bit) 3> FORMAT /u02/oradata/db11g/tts01.dbf; Starting conversion at target at 08-MAR-10 using channel ORA_DISK_1 channel ORA_DISK_1: starting datafile conversion input file name=/u01/oracle/tts01.dbf converted datafile=/u02/oradata/db11g/tts01.dbf channel ORA_DISK_1: datafile conversion complete, elapsed time: 00:00:01 Finished conversion at target at 08-MAR-10 8) Import the tablespace meta data into the 11g database and plug in the tablespaces -make the tablespaces read write [oracle@redhat346 ~]$ impdp dumpfile=tts_meta.dmp directory=dumpdir transport_datafiles=/u02/oradata/db11g/example01.dbf,/u02/oradata/db11g/tts01.dbf Import: Release 11.2.0.1.0 Production on Mon Mar 8 13:21:37 2010
Copyright (c) 1982, 2009, Oracle and/or its affiliates. All rights reserved. Username: sys as sysdba Password: Connected to: Oracle Database 11g Enterprise Edition Release 11.2.0.1.0 64bit Production With the Partitioning and Real Application Testing options Master table SYS.SYS_IMPORT_TRANSPORTABLE_01 successfully loaded/unloaded Starting SYS.SYS_IMPORT_TRANSPORTABLE_01: sys/******** AS SYSDBA dumpfile=tts_meta.dmp directory=dumpdir transport_datafiles=/u02/oradata/db11g/example01.dbf,/u02/oradata/db11g/tts01.dbf Processing object type TRANSPORTABLE_EXPORT/PLUGTS_BLK Processing object type TRANSPORTABLE_EXPORT/TYPE/TYPE_SPEC ORA-31684: Object type TYPE:PM.ADHEADER_TYP already exists ORA-31684: Object type TYPE:PM.TEXTDOC_TYP already exists ORA-31684: Object type TYPE:IX.ORDER_EVENT_TYP already exists ORA-31684: Object type TYPE:OE.PHONE_LIST_TYP already exists ORA-31684: Object type TYPE:OE.CUST_ADDRESS_TYP already exists ORA-31684: Object type TYPE:PM.TEXTDOC_TAB already exists Processing object type TRANSPORTABLE_EXPORT/TABLE Processing object type TRANSPORTABLE_EXPORT/GRANT/OWNER_GRANT/OBJECT_GRANT Processing object type TRANSPORTABLE_EXPORT/INDEX Processing object type TRANSPORTABLE_EXPORT/CONSTRAINT/CONSTRAINT Processing object type TRANSPORTABLE_EXPORT/INDEX_STATISTICS Processing object type TRANSPORTABLE_EXPORT/COMMENT Processing object type TRANSPORTABLE_EXPORT/CONSTRAINT/REF_CONSTRAINT Processing object type TRANSPORTABLE_EXPORT/TRIGGER Processing object type TRANSPORTABLE_EXPORT/INDEX/FUNCTIONAL_AND_BITMAP/INDEX Processing object type TRANSPORTABLE_EXPORT/INDEX/STATISTICS/FUNCTIONAL_AND_BITMAP/INDE X_STATISTICS Processing object type TRANSPORTABLE_EXPORT/TABLE_STATISTICS Processing object type TRANSPORTABLE_EXPORT/DOMAIN_INDEX/TABLE Processing object type TRANSPORTABLE_EXPORT/DOMAIN_INDEX/SECONDARY_TABLE/INDEX .. .. SQL> alter tablespace tts read write; Tablespace altered. SQL> alter tablespace example read write;
Tablespace altered. ***********ON SOURCE UPDATE 2********** SQL> conn sh/sh Connected. SQL> update mycustomers set cust_city=Hong Kong; 55500 rows updated. SQL> commit; Commit complete. Note: As we make changes in the source database, the trail files on the target start getting populated. These are located in the destination we specified when creating the RMTTRAIL. [oracle@redhat346 dirdat]$ pwd /u01/oracle/ggs/dirdat [oracle@redhat346 dirdat]$ ls -lrt -rw-rw-rw- 1 oracle oinstall 9999950 Mar 8 09:41 gs000000 -rw-rw-rw- 1 oracle oinstall 9999641 Mar 8 09:41 gs000001 -rw-rw-rw- 1 oracle oinstall 9999629 Mar 8 10:00 gs000003 -rw-rw-rw- 1 oracle oinstall 9999724 Mar 8 10:00 gs000002 -rw-rw-rw- 1 oracle oinstall 9999741 Mar 8 10:00 gs000004 -rw-rw-rw- 1 oracle oinstall 2113226 Mar 8 10:00 gs000005 -rw-rw-rw- 1 oracle oinstall 9999791 Mar 8 10:35 rm000000 -rw-rw-rw- 1 oracle oinstall 9999721 Mar 8 10:35 rm000001 -rw-rw-rw- 1 oracle oinstall 9999249 Mar 8 10:49 rm000003 -rw-rw-rw- 1 oracle oinstall 9999309 Mar 8 10:49 rm000002 -rw-rw-rw- 1 oracle oinstall 9999818 Mar 8 10:49 rm000004 -rw-rw-rw- 1 oracle oinstall 9999430 Mar 8 10:49 rm000005 -rw-rw-rw- 1 oracle oinstall 9999412 Mar 8 10:49 rm000006 -rw-rw-rw- 1 oracle oinstall 9999588 Mar 8 10:54 rm000007 -rw-rw-rw- 1 oracle oinstall 9999481 Mar 8 10:54 rm000009 -rw-rw-rw- 1 oracle oinstall 9999399 Mar 8 10:54 rm000008 -rw-rw-rw- 1 oracle oinstall 9999787 Mar 8 10:54 rm000010 -rw-rw-rw- 1 oracle oinstall 9999770 Mar 8 10:57 rm000011 -rw-rw-rw- 1 oracle oinstall 9999941 Mar 8 10:57 rm000012 -rw-rw-rw- 1 oracle oinstall 9999913 Mar 8 10:57 rm000013 -rw-rw-rw- 1 oracle oinstall 9999429 Mar 8 11:09 rm000014 -rw-rw-rw- 1 oracle oinstall 9999812 Mar 8 11:09 rm000015
-rw-rw-rw- 1 oracle oinstall 9999240 Mar 8 11:09 rm000016 -rw-rw-rw- 1 oracle oinstall 9999454 Mar 8 11:09 rm000017 -rw-rw-rw- 1 oracle oinstall 9999914 Mar 8 11:09 rm000018 -rw-rw-rw- 1 oracle oinstall 9999820 Mar 8 11:16 rm000019 -rw-rw-rw- 1 oracle oinstall 9999766 Mar 8 11:16 rm000020 -rw-rw-rw- 1 oracle oinstall 9999706 Mar 8 12:56 rm000021 -rw-rw-rw- 1 oracle oinstall 9999577 Mar 8 12:56 rm000022 -rw-rw-rw- 1 oracle oinstall 9999841 Mar 8 12:56 rm000023 -rw-rw-rw- 1 oracle oinstall 9999890 Mar 8 13:26 rm000024 -rw-rw-rw- 1 oracle oinstall 9999604 Mar 8 13:26 rm000025 -rw-rw-rw- 1 oracle oinstall 9999536 Mar 8 13:26 rm000026 -rw-rw-rw- 1 oracle oinstall 918990 Mar 8 13:26 rm000027 9) On the target Linux environment now we create and start the GoldenGate Replicat process/processes. They will now start reading from the Extract trail files created in Step 1 and will start applying them to the 11g database. GGSCI (redhat346.localdomain) 1> add replicat myrep, extrail /u01/oracle/ggs/dirdat/rm REPLICAT added. GGSCI (redhat346.localdomain) 6> edit params myrep REPLICAT myrep SETENV (ORACLE_HOME = /u01/app/oracle/product/11.2.0/dbhome_1) SETENV (ORACLE_SID = db11g) ASSUMETARGETDEFS USERID ggs_owner, PASSWORD ggs_owner MAP sh.*, TARGET sh.*; MAP pm.*, TARGET pm.*; MAP oe.*, TARGET oe.*; MAP hr.*, TARGET hr.*; MAP ix.*, TARGET ix.*; 10) Once all the changes in the trail files have been applied by the Replicat process and we confirm that both source and target are in sync (we can use another GoldenGate product called Veridata for this), we can now point the users and application to the 11g Linux database with no or minimal downtime which will depend on the infrastructure. We can see the Replicat process going through and reading all the trail files until it has completed processing all the files GGSCI (redhat346.localdomain) 131> info replicat myrep REPLICAT MYREP Last Started 2010-03-08 13:42 Status RUNNING Checkpoint Lag 03:07:37 (updated 00:00:17 ago) Log Read Checkpoint File /u01/oracle/ggs/dirdat/rm000002
2010-03-08 10:35:27.001328 RBA 6056361 . .. GGSCI (redhat346.localdomain) 156> info replicat myrep REPLICAT MYREP Last Started 2010-03-08 13:42 Status RUNNING Checkpoint Lag 02:53:49 (updated 00:00:00 ago) Log Read Checkpoint File /u01/oracle/ggs/dirdat/rm000007 2010-03-08 10:49:39.001103 RBA 2897635 .. GGSCI (redhat346.localdomain) 133> info replicat myrep REPLICAT MYREP Last Started 2010-03-08 13:48 Status RUNNING Checkpoint Lag 00:00:00 (updated 00:00:07 ago) Log Read Checkpoint File /u01/oracle/ggs/dirdat/rm000027 2010-03-08 13:26:43.000861 RBA 918990 GGSCI (redhat346.localdomain) 134> lag replicat myrep Sending GETLAG request to REPLICAT MYREP Last record lag: 1363 seconds. At EOF, no more records to process. TEST! Now check and confirm from the database if second update statement (UPDATE 2) run on the source database has been applied on the target SQL> select distinct cust_city from mycustomers; CUST_CITY Hong Kong