If you're looking for ETL Testing Interview Questions & Answers for Experienced or Freshers, you are in the right place. There are a lot of opportunities from many reputed companies in the world. According to research ETL Testing has an extension market share. So, You still have the opportunity to move ahead in your career in ETL Testing Analytics. Mindmajix offers Advanced ETL Testing Interview Questions 2021 that helps you in cracking your interview & acquire a dream career as an ETL Testing Analyst.
Are you interested in taking up the ETL Testing Certification Course? Enroll for Free Demo on ETL Testing Training
ETL Testing Interview Questions And Answers
Q) ETL Testing Vs DB Testing
|Compare ETL Testing and DB Testing|
|ETL Testing||DB Testing|
|Business Intelligence reporting||The goal is to integrate data|
|Business flow environment based on earlier data||Applicable to business flow systems|
|Informatica, Cognos, and QuerySurge can be used||QTP and Selenium tools for automation|
|Analyzing data may have a potential impact||Architectural implementation involves high impact.|
|Dimensional model||Entity-relationship model|
|Analytics are processed||Transactions are processed|
|Denormalized data is used||Data used is normalized|
Q1) What exactly do you mean by ETL?
Ans: ETL stands for Extract Transform Load and is widely regarded as one of the essential Testing tools in the data warehousing architecture. Its main task is to handle data management for the business process which is complex and is useful to the business in many ways. Extracting simply means reading the data from a database. Transformation means converting the data into a form that is suitable for analysis and reporting. The load on the other side handles the process of writing and managing the data into the database which users want to target simply.
Q2) There is a group of parameters that direct the server regarding the movement of the data from the source to the target. What it is called?
Ans: It is called a Session.
Q3) Do you have any idea about the ETL testing and the operations that are a part of the same?
Ans: Well, there are certain important tasks that are opted in this. It simply means verifying the data in terms of its transformation in the correct or the right manner as per the needs of a business. It also includes the verification of the projected data. The users can also check whether the data is successfully loaded in the warehouse or not without worrying about the loss of data. The improvement in scalability, as well as the performance, can also be assured from this directly. In addition to this, the ETL simply replaces the default values which are not always useful to the users.
Q4) How can you put Power Center different from the Power Mart?
Ans: Power Mart is good to be considered only when the data processing requirements are low. On the other side, the Power Center can simply process bulk time in a short span of time. Power Center can easily support ERP such as SAP while no support of the same is available on the ERP. The local repository can be supported by the Mart while the center cannot perform this task reliably.
Q5) What is partitioning in ETL?
Ans: The transactions are always needed to be divided for better performance. The same processes are known as Partitioning. It simply makes sure that the server can directly access the sources through multiple connections.
Q6) Name a few tools that you can easily use with the ETL?
Ans: There are many tools that can be considered. However, it is not always necessary that a user needs all of them at the same time. Also, which tool is used simply depends on the preference and the task that needs to be accomplished. Some of the commonly used ones are Oracle Warehouse Builder, Congos Decision Stream, SAS Enterprise ETL server, and SAS Business warehouse.
Q7) What do you understand by the term fact in the ETL and what are the types of the same?
Ans: Basically, it is regarded as a federal component that generally belongs to a model that has multiple dimensions. The same can also be used when it comes to considering the measures that belong to analyzation. The facts are generally useful for providing the dimensions that largely maters in the ETL. The commonly used types of facts in ETL are Semi-additive facts, Additive Facts, as well as Non-additive Facts.
Q8) What exactly do you know about the tracing level and the types of the same?
Ans: There are file logs and there is a limit on them when it comes to storing data in them. The Tracing level is nothing but the amount of data that can be easily stored on the same. These levels clearly explain the tracing levels and in a manner that provides all the necessary information regarding the same. There are two types of same and they are:
Q9) Do you think data warehousing and data mining is different from one another. How the same can be associated with the warehousing applications?
Ans: Well, the warehousing applications that are important generally include the analytical process, Information processing, as well as Data mining. There are actually a very large number of predictive that needs to be extracted from the database with a lot of data. The warehousing sometimes depends on the mining for the operations involved. Data mining is useful for the analytical process while the other doesn’t. The data can simply be aggregated from the sources that are different through the warehousing approach while the same is not possible in case of mining.
Q10) Do you have any information regarding the Grain of Fact?
Ans: The fact information can be stored at a level that is known as grain fact. The other name of this is Fact Granularity. It is possible for the users to change the name of the same when the need for the same is realized. There are multiple files that are associated with the same and the users can use this for changing the name of all of them directly.
Q11) It is possible to load the data and use it as a source?
Ans: Yes, in ETL it is possible. This task can be accomplished simply by using the Cache. The users must make sure that the Cache is free and is generally optimized before it is used for this task. At the same time, the users simply make sure that the desired outcomes can simply be assured without making a lot of effort.
Q12) What is a Factless fact table in ETL?
Subscribe to our youtube channel to get new updates..!
Ans: It is defined as the table without measures in the ETL. There are a number of events that can be managed directly with the same. It can also record events that are related to the employees or with the management and this task can be accomplished in a very reliable manner.
Q13) What exactly do you mean by the Transformation? What are the types of same?
Ans: It is basically regarded as the repository object which is capable to produce the data and can even modify and pass it in a reliable manner. The two commonly used transformations are Active and Passive.
Q14) What is the exact purpose of an ETL according to you?
Ans: It is actually very beneficial for the extracting of data and from the systems that are based on legacy.
Q15) Can you define measures in a simple statement?
Ans: Well, they can be called the number data which is generally based on the columns and is generally present in a fact table by default.
Q16) When you will make use of the Lookup Transformation?
Ans: It is one of the finest and in fact one of the very useful approaches in the ETL. It simply makes sure that the users can get a related value from a table and with the help of a column value that seems useless. In addition to this, it simply makes sure of boosting the performance of a dimensions table which is changing at a very slow rate. Also, there are situations when the records already remain present in the table. Dealing with such issues can be made possible with the help of Lookup transformation.
Q17) What do you understand by Data Purging?
Ans: There are needs and situations when the data needs to be deleted from the data warehouse. It is a very daunting task to delete the data in bulk. The Purging is an approach that can delete multiple files at the same time and enable users to maintain speed as well as efficiency. A lot of extra space can be created simply with this.
Q18) Can you tell me something about Bus Schema?
Ans: Dimension identification is something that is very important in the ETL and the same is largely handled by the Bus Schema.
Q19) Are you familiar with the Dynamic and the Static Cache?
Ans: When it comes to updating the master table, the dynamic cache can opt. Also, the users are free to use it for changing the dimensions. On the other side, the users can simply manage the flat files through the Static Cache. It is possible to deploy both the Dynamic and the Static Cache at the same time depending on the task and the overall complexity of the final outcome.
Q20) What do you mean by staging area?
Ans: It is an area that is used when it comes to holding the information or the data temporary on the server that controls the data warehouse. There are certain steps that are included and the prime one among them is Surrogate assignments.
Q21) What are the types of Partitioning you are familiar with?
Ans: There are two types of Partitioning that are common in ETL and they are:
Q22) Can you tell a few benefits of using the Data Reader Destination Adapter?
Ans: There are ADO recordsets that generally consist of columns and records. When it comes to populating them in a simple manner, the Data Reader Destination Adapter is very useful. It simply exposes the data flow and let the users impose various restrictions on the data which is required in many cases.
Q23) Is it possible to extract the SAP data with the help of Informatica?
Ans: There is a power connect option that simply lets the users perform this task and in a very reliable manner. It is necessary to import the source code in the analyzer before you accomplish this task.
Q24) What do you mean by the term Mappet?
Ans: This is actually an approach that is useful for creating or arranging the different sets in the transformation. It simply let user accomplish other tasks also that largely matters and are related to the data warehouse.
Q25) What are commercial ETL tools?
- Ab Initio
- Adeptia ETL
- Business Objects Data Services
- Informatica PowerCenter
- Business Objects Data Integrator (BODI) Confluent
Q26) Can you explain the fact table?
Ans: The data warehouse fact table is the central table of the star schema.
Q27) What are the types of measures?
Ans: There are 3 types of measures:
- Additive Measures - Can be joined across any dimensions of the fact table.
- Semi Additive Measures - Can be joined across only some of the dimensions of the fact table.
- Non-Additive Measures - Cannot be joined across any dimensions of the fact table.
Q28) Give a brief on Grain of Fact?
Ans: Grain fact functionality defined as a level/stage where the fact information will be stored. Also called Fact Granularity.
Q29) Define Transformation?
Ans: In ETL, Transformation involves, data cleansing, Sorting the data, Combining or merging, and applying the business rules to the data for improving the data for quality and accuracy in the ETL process.
Q30) What is Lookup Transformation?
Ans: The Lookup transformation accomplished lookups by joining information in input columns with columns in a reference dataset. You utilize the lookup to get extra data in a related table that depends on values in common columns.
Q31) Is it possible to update the table using the SQL service?
Ans: Yes, it is actually possible and the users can perform this task simply and in fact without worrying about anything. The users generally have several options to accomplish this task easily. The methods that can be helpful in this matter are using a staging cable, using a SQL command, using the MSSQL, as well as using the Cache.
Q32) How can you define a Workflow?
Ans: It is basically a group that contains instructions that let the server perform the executions related tasks.
Q33) Tell one basic difference between the Connected Lookup and Unconnected ones?
Ans: Mapping is common in the connected lookups while it is not so common in the latter. It is only used in the Unconnected lookup only when the function is already defined. . There are several values that can be returned from the Connected Lookup while the Unconnected Lookup has a strict upper limit on the same.
Q34) Tell something about the Data source view and how it is significant?
Ans: There are several analysis services databases that largely depend on the relational schema and the prime task of the Data source is to define the same. They are also helpful in creating the cubes and dimensions with the help of which the users can set the dimensions in a very easy manner.
Q35) What are objects in the Schema?
Ans: These are basically considered as the logical structures that are related to the data. They generally contain tables, views, synonyms, clusters as well as function packages. In addition to this, there are several database links that are present in them.
Q36) What are Cubes in the ETL and how they are different from that of OLAP?
Ans: There are things on which the data processing depends largely and cubes are one among them, they are generally regarded as the units of the same that provide useful information regarding the dimensions and fact tables. When it comes to multi-dimensions analysis, the same can simply be assured from this. On the other side, the Online Analytics Processing stores a large data in the dimensions that are more than one. This is generally done for making the reporting process more smooth and reliable. All the facts in it are categorized and this is exactly what makes them easy to understand.
Q37) Unconnected Vs Connected Lookups?
|Connected Lookup||Unconnected Lookup|
|Either dynamic or Static Cache can be used.||Can use only Static Cache.|
|We can return multiple rows from the same row||Can return only one output port|
|It supports user-defined values||It won't support user-defined values|
|We can pass any number of values to another transformation.||Can pass one output value to one transformation|
|Cache has all lookup columns that are used in the mapping.||Cache has all the lookups or output ports of lookup conditions and returns ports.|
Q38) Define Bus Schema?
Ans: In a data warehouse, BUs Schema is used for identifying the most common dimensions in the business process. In one word its is a definite dimension and a standardized equivalent of facts.
Q39) What data purging means?
Ans: Data Purging - Common word that used in data warehousing for deleting or erasing data from the storage.
Q40) Schema Objects mean?
Ans: Schema objects can be defined as logical structures, whereas DB stores the schema object logically within a database tablespace. Schema Objects can be the tables, clusters or views, sequence or indexes, functions packages and db links.
Q41) Can you brief the terms Mapplet, Session, Workflow and Worklet?
- Mapplet: Reusable object that contains a set of transformations.
- Worklet: Represents set of workflow tasks
- Workflow: Customs the tasks for each record that need to execute.
- Session: a set of instructions that instructs how to flow the data to the target.