If you're looking for Oracle GoldenGate Interview Questions & Answers for Experienced or Freshers, you are at the right place.
There are a lot of opportunities from many reputed companies in the world. According to research, Oracle GoldenGate has a market share of about 9.3%. So, You still have the opportunity to move ahead in your career in Oracle GoldenGate Development.
Mindmajix offers Advanced Oracle GoldenGate Interview Questions 2021 that helps you in cracking your interview & acquire a dream career as Oracle GoldenGate Developer.
Do you want to master Oracle GoldenGate? Then enrol in "Oracle GoldenGate Training" This course will help you to master Oracle GoldenGate
Here are frequently asked Oracle GoldenGate Interview Questions, let's have a look into them.
Oracle GoldenGate Interview Questions And Answers
Q1) What type of Topology does Goldengate support?
Ans: GoldenGate supports the following topologies.
Q2) What are the main components of the Goldengate replication?
Ans: The replication configuration consists of the following processes.
Q3) What transaction types does Goldengate support for Replication?
Ans: Goldengate supports both DML and DDL Replication from the source to target.
[Related Article: Oracle Exadata Interview Questions]
Q4) What are the supplemental logging pre-requisites?
Ans: The following supplemental logging is required.
1. Database supplemental logging
2. Object-level logging
Q5) Why is Supplemental logging required for Replication?
Integrated Capture (IC):
1. In the Integrated Capture mode, GoldenGate works directly with the database log mining server to receive the data changes in the form of logical change records (LCRs).
2. IC mode does not require any special setup for the databases using ASM, transparent data encryption, or Oracle RAC.
3. This feature is only available for Oracle databases in Version 18.104.22.168 or higher.
4. It also supports various object types that were previously not supported by Classic Capture.
5. This Capture mode supports extracting data from source databases using compression.
Integrated Capture can be configured in an online or downstream mode.
Q6) List the minimum parameters that can be used to create the extract process?
Ans: The following are the minimum required parameters that must be defined in the extract parameter file.
1. EXTRACT NAME
Q7) I want to configure multiple extracts to write to the same extra file? Is this possible?
Ans: Only one Extract process can write to one extra at a time. So you can’t configure multiple extracts to write to the same exttrail.
Q8) What type of Encryption is supported in Goldengate?
Ans: Oracle Goldengate provides 3 types of Encryption.
1. Data Encryption using Blowfish.
2. Password Encryption.
3. Network Encryption.
[Related Article: Oracle Performance Tuning Interview Questions]
Q9) What are the different password encryption options available with OGG?
Ans: You can encrypt a password in OGG using
1. Blowfish algorithm and
2. Advance Encryption Standard (AES) algorithm
Q10) What are the different encryption levels in AES?
Ans: You can encrypt the password/data using the AES in three different keys
a) 128 bit
b) 192 bit and
c) 256 bit
Q11) What are some of the key features of GoldenGate 12c?
Ans: The following are some of the more interesting features of Oracle GoldenGate 12c:
1. Support for Multitenant Database
2. Coordinated Replicat
3. Integrated Replicat Mode
4. Use of Credential store
5. Use of Wallet and master key
6. Trigger-less DDL replication
7. Automatically adjusts threads when RAC node failure/start
8. Supports RAC PDML Distributed transaction
9. RMAN Support for mined archive logs
Q12) What are the installation options available in OGG 12c?
Ans: You can install Oracle GoldenGate 12c using in 2 ways:
1) Interactive Installation with OUI – Graphical interface
2) Silent Installation with OUI – Command Interface
Q13) What is a Credential Store in OGG 12c?
Ans: OGG Credential Store manages Encrypted Passwords and USERIDs that are used to interact with the local database and Associate them with an Alias.
Instead of specifying actual USERID and Password in a command or a parameter file, you can use an alias. The Credential Store is implemented as an auto-login wallet within the Oracle Credential Store Framework (CSF).
Q14) How to configure the Credential Store in OGG 12c?
Ans: Steps to configure Oracle Credential Store are as follows:
1) By Default Credential Store is located under “dircrd” directory.
If you want to specify a different location use can specify the “CREDENTIALSTORELOCATION” parameter in GLOBALS file.
Example: CREDENTIALSTORELOCATION /u01/app/oracle/OGG_PASSWD
2) Goto OGG home and connect to GGSCI.
Q15) What command is used to create the credential store?
Ans: ADD CREDENTIALSTORE
Q16) How do you add credentials to the credential store?
Ans: ALTER CREDENTIALSTORE ADD USER userid,
Example: GGSCI> ALTER CREDENTIALSTORE ADD USER GGS@orcl, PASSWORD oracle ALIAS extorcl DOMAIN OracleGoldenGate
Q17) How do you retrieve information from the Oracle Credential Store?
Ans: GGSCI> INFO CREDENTIALSTORE
GGSCI> INFO CREDENTIALSTORE DOMAIN OracleGoldenGate
Q18) What are the different data encyption methods available in OGG 12c?
Ans: In OGG 12c you can encrypt data with the following 2 methods:
1) Encrypt Data with Master Key and Wallet
2) Encrypt Data with ENCKEYS
Q19) How do you enable Oracle GoldenGate for Oracle database 22.214.171.124?
Subscribe to our youtube channel to get new updates..!
Ans: The database services required to support Oracle GoldenGate capture and apply must be enabled explicitly for an Oracle 126.96.36.199 database.
This is required for all modes of Extract and Replicat.
To enable Oracle GoldenGate, set the following database initialization parameter. All instances in Oracle RAC must have the same setting.
Q20) How does the Replicat works in a Coordinated Model?
In a Coordinated Mode Replicat operates as follows:
1. Reads the Oracle GoldenGate trail.
2. Performs data filtering, mapping, and conversion.
3. Constructs SQL statements that represent source database DML or DDL transactions (in committed order).
4. Applies the SQL to the target through the SQL interface that is supported for the given target database, such as ODBC or the native database interface.
[Related Article: Oracle PL SQL Interview Questions]
Q21) What is the difference between Classic and Coordinated Replicat?
Ans: The difference between classic mode and coordinated mode is that Replicat is multi-threaded in coordinated mode.
Within a single Replicat instance, multiple threads read the trail independently and apply transactions in parallel. Each thread handles all of the filtering, mapping, conversion, SQL construction, and error handling for its assigned workload.
A coordinator thread coordinates the transactions across threads to account for dependencies among the threads.
Q22) How do you create a COORDINATED REPLICATE in OGG 12c?
Ans: You can create the COORDINATED REPLICATE with the following OGG Command:
ADD REPLICATE rfin, COORDINATED MAXTHREADS 50, EXTTRAIL dirt/et
Q23) I have created a Replicat process in OGG 12c and forgot to specify the DISCARDFILE parameter. What will happen?
Ans: Starting with OGG 12c, if you don’t specify a DISCARDFILE OGG process now generates a dicard file with default values whenever a process is started with START command through GGSCI.
Q24) Is it possible to start OGG EXTRACT at a specific CSN?
Ans: Yes, Starting with OGG 12c you can now start Extract at a specific CSN in the transaction log or trail.
START EXTRACT fin ATCSN 12345
START EXTRACT finance AFTERCSN 67890
Q25) List a few parameters which may help improve the replicate performance?
Ans: The parameters below can be used to improve the replicate performance:
Q26) What are the areas to monitor in Goldengate Replication?
Ans: The lag and checkpoint latency of the Extract, pump, and Replicat processes are normally monitored.
Q27) What is the PASSTHRU mode used for?
Ans: In pass-through mode, the Extract process does not look up the table definitions, either from the database or from a data definitions file.
This increases the throughput of the data pump, as the object definition lookup is bypassed.
Q28) What are the most common reasons for an Extract process slowing down?
Ans: Some of the possible reasons are:
1. Long-running batch transactions on a table.
2. Insufficient memory on the Extract side. Uncommitted, long-running transactions can cause the writing of a transaction to a temporary area (dirtmp) on disk. Once the transaction is committed it is read from the temporary location on the file system and converted to trail files.
3. Slow or overburdened Network.
Q29) What are the most common reasons for the Replicat process slowing down?
Ans: Some of the possible reasons are:
1. a large number of transactions on a particular table.
2. Blocking sessions on the destination database where non-Goldengate transactions are also taking place on the same table as the replicate processing.
3. If using DBFS, writing & reading of trail files may be slow if SGA parameters are not tuned.
4. For slow Replicat’s, latency may be due to missing indexes on target.
5. Replicat having to process Update, delete of rows in very large tables.
Q30) My extract was running fine for a long time. All of a sudden it went down. I started the extract processes after 1 hour. What will happen to my committed transactions that occurred in the database during the last 1 hour?
Ans: OGG checkpoint provides the fault tolerance and makes sure that the transaction marked for committed is capture and captured only once.
Even if the extract went down abnormally, when you start the process again it reads the checkpoint file to provide the read consistency and transaction recovery.
Q31) I have configured Oracle GoldenGate integrated capture process using the default values. As the data load increases, I see that extract starts lagging behind by an hour (or more) and database performance degrades. How you will resolve this performance issue?
Ans: When operating in integrated capture mode, you must make sure that you have assigned sufficient memory to STREAMS_POOL_SIZE. An undersized STREAMS_POOL_SIZE or limiting the streams pool to use a specific amount of memory can cause troubles.
The best practice is to allocate STREAMS_POOL_SIZE at the instance level and allocate the MAX. SGA at GG process level as below:
SQL> alter system set STREAMS_POOL_SIZE=3G
TRANLOGOPTIONS INTEGRATEDPARAMS (MAX_SGA_SIZE 2048, PARALLELISM 4)
Q32) Why would you segregate the tables in a replication configuration? How would you do it?
Ans: In OGG you can configure replicate at the data at the schema level or at the table level using the TABLE parameter of extract and MAP parameter of replicate.
For replicating the entire database you can list all the schemas in the database in the extract/replicate parameter file.
Depending on the amount of redo generation you can split the tables in a schema into multiple extracts and replicates to improve the performance of data replication. Alternatively, you can also group a set of tables in the configuration by the application functionality.
Alternatively, you may need to remove tables that have long-running transactions in a separate extract process to eliminate lag on the other tables.
Let’s say that you have a schema named SCOTT and it has 100 hundred tables.
Out of these hundred tables, 50 tables are heavily utilized by the application.
To improve the overall replication performance you create 3 extract and 3 replicates as follows:
Ext_1/Rep_1 –> 25 tables
Ext_2/Rep_2 –> 25 tables
Ext_3/Rep_3 –> 50 tables
Ext_1/Rep_1 and Ext_2/Rep_2 contain 25 tables each which are heavily utilized or generate more redo.
Ext_3/Rep_3 contains all the other 50 tables which are least used.
Q33) How can we report on long-running transactions?
Ans: The WARNLONGTRANS parameter can be specified with a threshold time that a transaction can be open before Extract writes a warning message to the ggs error log.
Example: WARNLONGTRANS 1h, CHECKINTERVAL 10m
Q34) What command can be used to view the checkpoint information for the extract process?
Ans: Use the following command to view the Extract checkpoint information.
GGSCI> info extract, shows
GGSCI> info extract ext_fin, showch
Q35) How is the RESTARTCOLLISION parameter different from HANDLECOLLISIONS?
Ans: The RESTARTCOLLISION parameter is used to skip ONE transaction only in a situation when the GoldenGate process crashed and performed an operation (INSERT, UPDATE & DELETE) in the database but could not checkpoint the process information to the checkpoint file/table.
On recovery, it will skip the transaction and AUTOMATICALLY continue to the next operation in the trail file.
When using HANDLECOLLISION GoldenGate will continue to overwritten and process transactions until the parameter is removed from the parameter files and the processes restarted.
Q36) How do you view the data which has been extracted from the redo logs?
Ans: The log dump utility is used to open the trail files and look at the actual records that have been extracted from the redo or the archive log files.
Q37) What does the RMAN-08147 warning signify when your environment has GoldenGate Capture Processes configured?
Ans: This occurs when the V$ARCHIVED_LOG.NEXT_CHANGE# is greater than the SCN required by the GoldenGate Capture process and RMAN is trying to delete the archived logs.
The RMAN-08147 error is raised when RMAN tries to delete these files.
When the database is open it uses the DBA_CAPTURE values to determine the log files required for mining.
However, if the database is in the mounted state the V$ARCHIVED_LOG. NEXT_CHANGE# value is used.
See MetaLink note: 1581365.1
Q38) How would you look at a trail file using log dump if the trail file is Encrypted?
Ans: You must use the DECRYPT option before viewing data in the Trail data.
List a few useful Logdump commands to view and search data stored in OGG trail files.
Below are few logdump commands used on a daily basis for displaying or analyzing data stored in a trail file.
$ ./logdump – to connect to the logdump prompt
logdump> open /u01/app/oracle/dirdat/et000001 – to open a trail file in logdump
logdump> fileheader on – to view the trail file header
logdump> ghdr on – to view the record header with data
logdump> detail on – to view column information
logdump> detail data – to display HEX and ASCII data values to the column list
logdump> reclen 200 – to control how much record data is displayed
logdump> pos 0 – To go to the first record
logdump> next (or simply n) – to move from one record to another in sequence
logdump> count – counting records in a trail
Q39) Why should I upgrade my GoldenGate Extract processes to Integrated Extract?
Ans: Oracle is able to provide faster integration of the new database features by moving the GoldenGate Extraction processes into the database.
Due to this, the GoldenGate Integrated Extract has a number of features like Compression which are not supported in the traditional Extract. You can read more about how to upgrade to Integrated Extract and more about Integrated Delivery.
Going forward preference should be given to creating new extracts as Integrated Extracts and also to upgrade existing traditional Extracts.
Q40) What is the minimum Database version which supports Integrated Delivery?
Ans: Oracle 188.8.131.52 is the minimum required database version that supports both Integrated extract and Integrated Replica.
Q41) What databases supports GoldenGate Integrated Delivery?
Ans: Oracle Integrated Delivery is only available for Oracle Databases.
Q42) With Integrated Delivery, where can we look for the performance stats?
Ans: Yes with 12c, performance statistics are collected in the AWR repository and the data is available via the normal AWR reports.
Q43) What are the steps required to add a new table to an existing replication setup?
Ans: The steps to be executed would be the following:
1. Include the new table in the Extract & pump process.
2. Obtain starting database SCN and Copy the source table data to the target database
3. Start Replicat on target at the source SCN database point.
Q44) What does the GoldenGate CSN equate to, in the Oracle Database?
Ans: It is equivalent to the Oracle database SCN transaction number.
Q45) How do you create a CSV file from the Extracted data?
Ans: You will have to use the CSV Flat File Adaptor to create CSV files. The source would be the extract trail files which use the configuration of the adaptor settings to generate CSV files.
Q46) What is the purpose of the DEFGEN utility?
Ans: When the source and the target schema objects are not the same (different DDL’s) the Replicat process needs to know the source definition of the objects. The output from the DEFGEN utility is used in conjunction with the trail data to determine which column value in the trail belongs to which column.
Q47) We want to setup one-way data replication for my online transaction processing application. However, there are compressed tables in the environment. Please suggest how I can achieve it.
Ans: You must use OGG 11.2 and configure the GoldenGate Integrated Capture process to extract data from compressed tables.
Note: Pre OGG 11.2 doesn’t support extracting data from compressed tables
Q48) We want to setup one-way data replication using Integrated Extract for my Oracle database running 10.2.0.4. Please suggest the best architecture to achieve it.
Ans: Oracle GoldenGate Integrated Capture process supports Oracle databases 10.2 and higher. But if you are running Oracle database 10.2 and want to you Oracle GoldenGate Integrated Capture process then you must configure downstream topology.
Q49) I am migrating my Oracle database from non-exadata to Exadata but my source OGG software and target OGG software versions are different (Say Source is running OGG 11.1 and target is running OGG 11.2). How can I configure my OGG process to work in such a configuration?
Ans: It is recommended that all instances of Oracle GoldenGate be the same version to take advantage of the new functionality, but this is not possible all the time and is not required.
In this scenario, OGG provides a parameter called ‘FORMAT RELEASE’ which allows customers to use different versions of Oracle GoldenGate Extract, trail files, and Replicat together.
Example: RmtTrail /u01/app/oracle/dirdat, Format Release 11.1
Note: The input and output trails of a data pump must have the same trail file version.
Q50) What are the different OGG Initial load methods available?
Ans: OGG has 2 functionalities, one it is used for Online data Replication and the second for Initial Loading.
If you are replicating data between 2 homogeneous databases then the best method is to use a database-specific method (Exp/Imp, RMAN, Transportable tablespaces, Physical Standby, and so on). Database specific methods are usually faster than the other methods.
—If you are replicating data between 2 heterogeneous databases or your replicate involves complex transformations, then the database-specific method can’t be used. In those cases, you can always use Oracle GoldenGate to perform the initial load.
Within Oracle GoldenGate you have 4 different ways to perform initial load.
1. Direct Load – Faster but doesn’t support LOB data types (12c include support for LOB)
2. Direct Bulk Load – Uses SQL*LOAD API for Oracle and SSIS for MS SQL SERVER
3. File to the replica – Fast but the rmtfile limit is 2GB. If the table can’t be fit in 1 rmtfile you can use maxfiles but the replica needs to be registered on the target OGG home to read the rmtfiles from the source.
4. File to Database utility – depending on the target database, use SQL*LOAD for Oracle and SSIS for MS SQL SERVER and so on.
Oracle GoldenGate initial loading reads data directly from the source database tables without locking them. So you don’t need downtime but it will use database resources and can cause performance issues. Take extra precaution to perform the initial load during the non-peak time so that you don’t run into resource contention.
Q51) I have a table called ‘TEST’ on source and target with same name, structure and data type but in a different column order. How can you set up replication for this table?
Ans: OGG by default assumes that the sources and target tables are identical. A table is said to be identical if and only if the table structure, data type and column order are the same on both the source and the target.
If the tables are not identical you must use the parameter ‘SOURCEDEFS’ pointing to the source table definition and ‘COLMAP’ parameter to map the columns from source to target.
Q52) What is the best practice to delete the extract files in OGG?
Ans: Use the manager process to delete the extract files after they are consumed by the extract/replicat process
PURGEOLDEXTRACTS /u01/app/oracle/dirdat/et*, USECHECKPOINTS, MINKEEPHOURS 2
Q53) What parameters can be used to configure Oracle GoldenGate to extract data from Archived Redo log files only?
Ans: Use the TRANLOGOPTIONS ARCHIVEDLOGONLY option in the parameter file.
Q54) What are the basic resources required to configure Oracle GoldenGate high availability solution with Oracle Clusterware?
Ans: There are 3 basic resources required:
1. Virtual IP
2. Shared storage
3. Action script
Q55) I have a one-way replication setup. The system administration team wants to apply an OS patch to both the OGG source host and the target servers. Provide the sequence of steps that you will carry before and after applying this patch.
1. Check to make sure that the Extract has processed all the records in the data source (Online Redo/archive logs)
GGSCI> send extract , logend
(The above command should print YES)
2. Verify the extract, pump and replicat has zero lag.
GGSCI> send extract , getlag
GGSCI> send extract , getlag
GGSCI> send replicat , getlag
(The above command should pring “At EOF, no more records to process.”)
3. Stop all application and database activity.
4. Make sure that the primary extract is reading the end of the redolog and that there is no LAG at all for the processes.
5. Now proceed with stopping the processes:?
1. Stop the primary extract
2. Stop the pump extract
3. Stop the manager process
4. Make sure all the processes are down.
1. Stop replicat process
2. Stop mgr
3. Make sure that all the processes are down.
4. Proceed with the maintenance
5. After the maintenance, proceed with starting up the processes:
1. Start the manager process
2. Start the primary extract
3. Start the pump extract
(Or simply all the extract processes as GGSCI> start extract *)
4. Make sure all that the processes are up.
1. Start the manager process
2. Start the replicat process.
3. Make sure that all the processes are up.