IDQ Reference

Download as docx, pdf, or txt
Download as docx, pdf, or txt
You are on page 1of 31

1.

The address Validation transformation can operate in which modes:


- Country Recognition/ Parse/ Suggestion List/ batch/ Certified

2. Where you can create a DB conn


- Window > Preferences > Informatica > Connections

3. To use AV in Suggestion list mode


- Doesn’t matter what input you use
- Discrete Fields for Input
- Hybrid
- Multiline

4. With regard to Custom templates in AV, which is not true


- Custom templates require inputs from the basic model only

5. Which is NOT true when exporting Ref. data tables for use by DQ mappings to PC
- Set the ref. data location to the services level not lower
- Select the data service
- The correct code page is automatically assigned and can’t be modified by the user
- During export select export ref. data

6. DQ mapping exported to PC, exported as


- Only as PC mapplets

7. Utilize DQ accelerators in PC
- Install accelerators in MR and export accelerator mappings and/or mapplets to PC

8. IDQ objects are integrated with PC, how is ref. content handled?
- Ref. content needs to be exported into a flat file and placed in the appropriate location

9. Which parameter function is supported in IDQ


- Ref. table location/path
- Ref. table name
- Relational and flat-file connection
- All the above – [Correct]

10. Exception for duplicate that score above the threshold, which one is true –
- Close enough to use automatic consolidation and can be output to standard output
11. Not a consolidation function within simple consolidation strategies
- Most Data

12. Association Transformation – which one is false


- Used ahead of matching to organize the record into group

13. Exception management which one is false


- Bad record management and consolidation data are stored in the same table in MR

14. Extension of Data matching process and a precursor of data consolidation process
- Association

15. Automatically consolidates matched records


- Consolidation

16. Project export, typical output file


- XML and Zip for ref. data

17. False with regard to import/export


- In Basic method, user must import entire project with a generic conflict resolution

18. Match mapplet over match transformation


- Reusable

19. Anomaly does a fuzzy key account for


- Scoring filters

20. Match strategy when position of data is important


- Hamming Distance

21. Configuration for dual source match mapping


- 2 source DO, 2 key generator, 1 match transformation and 2 target objects

22. Pattern based parser takes input from


- Labeller

23. Why mid-stream profiling helpful


- To determine if transformation is working correctly
24. Can’t perform mid-stream profiling on
– Data object used as target

25. Required info. to connect to developer for first time


- Domain name, domain host name, port no, MR name, id & pwd
26. Supported OS for Client installation
- MS Windows only
27. Service used to manage content
28. Version control in IDQ
- Integrated with 3rd party version control system
29. Merger transformation – Analytical questions
30. IDQ Services
31. IDQ Architecture – components
32. How many columns must be marked as valid in ref table
33. Accelerator related questions
34. Not a DQ accelerator currently offered by Informatica
- Telecom
35. Content installer executable file to install
- Address Ref. data, Identity population ref data and accelerator demonstrate data
36. Application deployed, where can be located
- DIS
37. Tags can be applied to all MR objects
38. Which will not allow to find tags
- Window > reset perspective
39. Analytical question on standardizer
40. Scorecarding can be performed on
- Per column basis
1. Which of the following is stored in the Staging Repository
A. Reference tables
B. Mappings
C. Reference Tables and Audit Information
D. Authentication Details

2. Which of these statements are true about Data Rules


A. Data Rules must be run against the entire data set they are defined upon
B. Once created, a Data Rule cannot be changed. It can only be deleted and recreated.
C. A Data Rule created in the Analyst will appear as a mapplet with an Expression
Transformation in the Developer
D. Data Rules can only be run on data in the project in which they were created

3. Where does Scorecarding fit into the DQ Process?


A. Not at all – Scorecarding is a separate process
B. At the start of the process only to assess current data quality levels
C. At the start and during regular intervals of the process to measure on-going data quality
levels
D. At the end of the process only to prove the value of the DQ tool.

4. Accelerators provide which of the following? Choose 4 answers


A. Rules
B. Reference Tables
C. Demonstration Mappings
D. Demonstration Data Objects
E. Profiles

5. When applying a reference table in the Labeler in token mode, which of the following
statements is correct?
A. The order of the reference tables defined will not impact your results displayed
B. Selecting Inclusive mode will output items that are present in the reference table under the
label
C. Selecting Exclusive mode will exclude items that are present in the reference table
D. Any items that do not appear in the reference table are not displayed in the output

6. What is the best approach to implementing Address Verification in Version 9.1?


A. Run the address into AV Transformation without modification
B. Take an iterative approach and review AV output. Determine why address did not validate
and attempt to resolve the issue prior to running them in the AV Transformation.
C. Canada Post defines the SERP certification standard for software applications that validate
Canadian addresses
D. Remove all punctuation before sending the address to the AV Transformation.

7. Which is true with regard to Profiles?


A. The Developer can only review Profiles created by the Analyst but cannot create any profiles
themselves
B. The Developer can take Analyst built profiles with rules and generate a mapping ensuring
lossless information between the Specification and Implementation
C. Comments, Tags and Rules can only be applied to profiles in the Analyst Tool.
D. Only one Tag can be applied to each object in a project.

8. Select an item that is not a common symptom of DQ problems.


A. Lack of trust and agreement in reporting
B. High Customer Satisfaction
C. Supply-chain ‘out of stock’ situations
D. Failure of IT projects

9. Should data be standardized before it reaches the Address Validator?


A. Yes, street and directional abbreviations should always be expanded to their unabbreviated
forms
B. No, any change may cause problems to the AV Transformation
C. Yes, if after review, specific problems have been identified that are causing address to not
validate than otherwise would
D. Yes, they should all be in upper case

10. Which of the following is not a function of Data Standardization?


A. Standardization converts data into Standardized formats
B. Standardization can remove “noise”
C. Standardization corrects and automatically updates the data in the source system
D. Standardization can remove or replace bad or inconsistent data

11. Which of the following will NOT work when adding transformation to mappings?
A. Double click on the transformation in the transformation palette
B. Click and drag the transformation onto the mapping
C. From the Mapping menu, select Add transformation
D. Right click and choose Add transformation on the mapping

12. Which two IDQ transformations can use parameters?


A. Comparison
B. Decision
C. Case Converter
D. Address Validator
E. Labeller

13. During a project export, typically what will the output be from the export?
A. An xml file containing all the mapping/mapplet/object definition
B. An xml file containing all the mapping/mapplet/object definition and a zip file containing all
reference data if required
C. The contents of each Project (including the dependent reference tables) in a zip file.
D. Projects will be exported to an .irob file

14. Which of the following is NOT a required step for using DQ for Excel?
A. In Excel, choose to Add a service and either enter the URL address or select the .wsdl or .xml
imported
B. Build and test a Web Services Mapping in the Developer Tool
C. Export the Mapping to PowerCenter and test
D. Create and Deploy the mapping as an Application

15. Which is the best combination of functions to effectively implement a DQ project?


A. Company Executives & IT
B. IT & Business
C. IT & Administration
D. Sales & Marketing

16. What happens to the records when corrections are made and a status of Accept is assigned?
A. The corrections are committed to the Staging DB and an UPDATED_STATUS of Accepted is
assigned to the record
B. The corrections are committed to the Staging DB and record is automatically pushed into
another DB
C. Records are corrected and sent to an Accepted DB table
D. Records are accepted and are pushed out to Target/Production DB

17. In Version 9.1, what does the Content Management Service do?
A. It manages the dictionaries created in the Analyst Tool
B. It provides the DIS with the information on the location of the address validation data as
well as AV Configuration Settings
C. It runs Identity Matching
D. It makes the Profile Repository available to outside reporting applications like Jaspersoft

18. Which of the following statements are true regarding DQ for Excel?
A. DQ for Excel allows .xls files be used as source and target objects in mappings
B. DQ for Excel enables non Informatica users to benefit from DQ Rules by consuming and
applying DQ rules directly from MS Excel data/worksheets
C. DQ for Excel can be used in any version of MS Excel
D. In order to be able to use DQ for Excel, both PowerCenter and Data Quality Developer must
be installed

19. What types of profiling can be performed in the Developer Tool in Data Quality?
A. Column profiling, Primary key inference, Dependency inference
B. Column and Join Profiling only
C. Column, Join Analysis, Mid Stream and Comparitive Profiling
D. Column, Join Analysis, Mid Stream, Comparitive, Primary and Foreign Key and Overlap
Profiling.

20. IDQ Parameters can be defined at what levels?


A. Mappings
B. Mapplets
C. Re-Usable Transformations
D. Rules

21. When would you use the DQA?


A. On suspect/exception records that fall out of a profile, standardization or matching
B. On records that could not be matched in a matching mapping
C. On records that are output from the Consolidation Transformation
D. On records that are output from the Association Transformation

22. The Pattern Parser is used in conjunction with which transformation?


A. Token Parser
B. Standardizer
C. Exception
D. Labeler

23. In the Case Converter transformation, which of the following is not a valid Case Transform?
A. Upper Case
B. Lower Case
C. Toggle merge Case
D. Sentence Case

24. Which DQ Transformations can be used to automatically consolidate matched records?


A. Consolidation
B. Match and Association
C. Key Gen and Consolidation
D. Key Gen, Association and Consolidation

25. If the inputs to Edit Distance Strategy below are ‘john’ and NULL, what will be the resulting
match score?
A. 0
B. 1
C. 0.5
D. 1.5

26. Which transformations can be used in DQ for Excel?


A. Only mappings using a Case Converter and a Standardizer can be used
B. Only mappings using non-active transformations can be used
C. All DQ Transformations, including Matching and Consolidation can be used
D. All DQ and Core Transformations can be used in DQ for Excel, there is no limitation.

27. Which strategy pairs characters together as its match algorithm?


A. Edit Distance
B. Bigram
C. Hamming Distance
D. Jaro Distance

28. Where are IDQ Mappings saved?


A. In the location specified in your Source object
B. In the Staging Database
C. In the Model Repository
D. On the Object Explorer tab in the Developer

29. Which informatica option provides solutions to common data quality issues in a country,
region or industry?
A. Populations
B. Accelerators
C. Dictionaries
D. Mapplets

30. Which are correct statements for IDQ Grouping and Matching? Chose 2 answers
A. IDQ Field Level matching does not utilize grouping
B. When field level matching is performed, the records within each group will be compared
against each other
C. When field level matching is performed, matching will be performed across multiple
groups in a single match transformation
D. When field level matching is performed, matching will not be performed across groups,
therefore it is imperative grouping is performed on a complete and accurate field(s)

31. How much data can DQA handle?


A. 2000 rows and 20 columns
B. Million rows and 100 columns
C. There are no built-in restrictions. Large deployments will require more hardware and
database tuning.
D. Maximum 1 million rows and upto 10,000 columns.

32. Which of the following cannot be executed in BOTH the Analyst and the Developer Tool?
A. A rule created in the Analyst Tool
B. A mapplet created in the Developer Tool and validated as a rule
C. A mapplet created in the Developer Tool and validated as a mapplet
D. A reusable rule created in the Analyst tool

33. How is Accelerator Content installed?


A. Using the IDQ Content Installer
B. It is imported through the Developer
C. The file extracted onto the Server and it is automatically picked up after restarting the server
D. It is installed using the importer available in the Content Management service

34. When is sampling for Profiling a good idea? Select the best answer.
A. When the data set contains sensitive information like credit card number
B. When the probability of data quality issues is small
C. When the client provides assurances that the quality of data is good
D. When the volume of data is large and there are storage and/or time constraints.
35. The IDQ and PowerCenter repositories all share the same schema.
A. TRUE
B. FALSE

36. Select the transformation that will not allow a DQ mapplet to validate as a rule
A. Expression Transformation
B. Address Validation Transformation
C. Aggregator Transformation
D. Labeler Transformation

37. With respect to field matching in IDQ, which best describes the approach to address
matching?
A. No standardization is necessary, map input fields to the match transformation
B. Standardize address fields and use address line, locality, region and postcode fields as input
to the match transformation
C. Standardize address fields and use the discrete address component fields, locality, region
and postcode input to the match transformation
D. Standardize address fields, create a group key field which contains parts of address fields
and on group key for address matching.

38. Users would export DQ mappings to Power Center for which of the following reasons?
(Choose 4)
A. Performance
B. Scalability
C. Batch access
D. DQ as part of the ETL Process

4 options were there and they asked to choose 4

39. Which definition describes Column Profiling?


A. Column Profiling automatically recognizes all functional and other dependencies in all tables
with foreign and primary keys
B. Column profiling is the process of looking at every distinct value to determine the true
minimum and maximum values of a column, data type, pattern and other important
information about each attribute
C. Column profiling identifies the attributes for referential integrity
D. Column profiling is the process of creating normalized schema for your target system

40. When changes are made to a mapplet in IDQ 9.1 that has been integrated into PowerCenter,
how are the changes propagated to the integrated mapplet?
A. Those changes are automatically propagated and managhed through domain settings
B. The user must manually re-export the IDQ mapplet to PowerCenter
C. The IDQ versioning system automatically synchs with PowerCenter and updates the mapplet
to current version.
D. None of the above.

41. When matching, is it always necessary to group data?


A. Only if the customer thinks it’s a good idea
B. Only if the quality of the data is poor
C. Not always. If the dataset is small enough, grouping is not required
D. Yes, it’s always necessary

42. What Address Validator output field should you use to determine the quality of address
validation?
A. Locality/City
B. ElementInputStatus
C. Match Code
D. MSA

43. Which phrase best describes how organizations should view their approach to DQ?
A. As a one-off project addressing the issues in their Data Warehouse in Batch
B. As on ongoing process addressing issues over time with a combination of fit-for-purpose
approach
C. As a real-time process addressing issues as they are introduced
D. As a batch process to be implemented to run on a monthly basis

44. What data quality transformation utilizes Identity Populations?


A. Key Generator Transformation
B. Match Transformation
C. Edit Distance Transformation
D. Comparison Transformation

45. Which of the following is NOT TRUE when it comes to creating and updating Reference tables?
A. Reference Tables can be updated from a Value Frequency list in the Developer
B. All except Un-Managed Reference tables can be updated through the Reference Table
manager in the Analyst tool
C. Reference Tables maintained in the Analyst and Developer tools are stored in separate
repositories
D. Reference Tables can be imported and exported using the Developer Tool

46. When using field matching which is the best approach if using First Name as a match field?
A. Never use First Name as it is not necessary
B. No standardization of First Name field is necessary
C. Use the Edit Distance Strategy
D. Create Soundex or NYSIIS value for the First Name and use it in match
47. Which statements describe Consolidation Transformation? Choose 2 answers.
A. A passive transformation
B. An active transformation
C. A transformation which is found only in PowerCenter
D. Creates a single, consolidated record from records identified as possible duplicates by the
Match transformation.

48. When profiling, why is it good to sort the value lists in both ascending and descending order?
A. There is no specific reason. It is simply a step of an ordered best practice approach
B. It is good because it is easier to determine if a field is numeric or not
C. Values with leading spaces or punctuation will typically sort to the top or bottom making
them easier to identify
D. None of the above

49. What versions of PowerCenter can an IDQ 9.1 mapping be exported to?
A. PC 8.6.1 and higher
B. PC 8.51 and higher
C. All PC Versions
D. PC 8.11, PC 8.51 and PC 8.6

50. What statement best describes IDQ’s Web Service capabilities?


A. Mappings can only be created as a Web Service through PC integration
B. DQ has both native web service functionality and allows integration through PC
C. DQ mappings can only be made accessible through native means
D. DQ has no web service support

51. When sharing projects across teams which of the following is True?
A. The same icon is used to indicate a shared and non-shared project
B. Comments that are added to a profile in the Analyst can be viewed by the Developer in the
Developer tool
C. Permissions need to be assigned to each folder within the project so users of shared
projects can access objects
D. Rules that were created in the Analyst tool will be visible in Developer but Developer built
tools will not be made available in the Analyst.

52. Which two are index key levels in the Identity Matching? Choose 2 answers
A. Narrow
B. Limited
C. Typical
D. Extended
53. Which of the following options is TRUE? The Advanced Import Wizard differs from the Basic
Import Wizard in the following ways
A. The Advanced Method allows conflict resolution and the Basic doesn’t
B. The Advanced Import Wizard allows objects to be imported into multiple target projects
with resolutions to dependencies.
C. Import steps can be undone in the Advanced Wizard but not in Basic one.
D. Users have a choice regarding which project objects are imported using Basic and Advanced
imports.

54. Which parameter function is supported in IDQ 9.1?


A. Set parameter for RTM Location / Path
B. Set parameter for RTM Name
C. Ability to parameterize relational and flat file connections
D. All of the above

55. How is data loaded into DQA and how is it configured?


A. By running the DQ client installer
B. The Analyst defines the table in the Analyst Tool and the mappings in the Developer tool
create and populate data in the appropriate tables
C. The Analyst populates and creates the table in the Analyst tool
D. The Administrator needs to create a DQA Service and associate it with the staging database.

56. When importing a flat file into a project, Developer and Analyst tool work exactly the same?
A. TRUE
B. FALSE

57. A customer wants to be able to review data on the overall quality of each record, which
technique should be used?
A. Scorcecarding
B. Matching
C. Grading
D. Standardization

58. A Customer has provided a number of delimited flat files that req profiling but not the file
definition what should you do first?
A. Allow the wizard bin developer, to automatically select the datatype when creating the data
object
B. Set all of the data types to text and length longer than specified. The profiler will tell you
what they should be
C. Import the file into a database table before profiling
D. Change the code page of the file to UTF-8.

59. You want to derive the gender from a firstname, how would you do this?
A. Use a ref table with the FN as the valid col and the gender as the 2 nd col. Then choose to
replace ref tab matches with the valid values when applying the ref tab in a strategy in the
standardizer. Use the FN as i/p.
B. Use a ref tab with the gender as the valid column and the FN in the next col. Then choose n
replace ref tab matches with the valid values when applying the ref tab in a strategy in the
standardizer. Use FN as i/p
C. Use decision and create an if then else rule
D. Use a ref tab with the gender as the valid column and the FN in the next col. Then choose n
replace ref tab matches with Custom strings when applying the ref tab in a strategy in the
standardizer. Use the FN as i/p.

60. What transformation would u use to standardize data using a ref table
A. Labeler
B. Decision
C. Standardizer
D. Parser

61. You have just standardized values in a field and u would like to review how the modified data
appears alongside the original data. Which profiling methods best be suited for this?
A. Col profiling
B. Mid stream
C. Join analysis
D. Comparative

62. What functionality does DQA provide?


A. Consolidate duplicate into master record
B. Bad record management
C. Automatically consolidates data
D. Regroup records for matching

63. With regard to a basic import which of the following is false?


A. Once added the target will display resolution of each folder imported
B. A conflict resolution can be selected to apply to all imported object
C. Default is replace options include rename reuse
D. The import can consist of an xml file and a zip containing the ref tab

64. Which two of the following values can be parameterized in the match transformation?
A. Threshold
B. Scoring methods
C. Match field i/p
D. Weights

65. What is required for customers to access and use IDQ Accelerator content?
A. Customer must license each Accelerator pack individually to download and use the content
B. No additional steps required
C. IDQ Customers simply need to request an IPS resource to download and install the content
D. Accelerators content is available only with PC license

66. What must you do in order to utilize Data Quality Accelerators in PC?
A. Import accelerator directly to PC
B. Install accelerator directly to PC
C. Install accelerator directly to Developer
D. Export accelerators mapplets and/or mapping to PC

67. When you select Suggestion list mode which statement is true?
A. The AV transformation searches the address ref data and returns all possible address
matches with the i/p address
B. The AV transformation searches the address ref data and returns most likely candidates
C. The AV transformation parses data into suggested address field without performing
validation
D. The AV transformation will return cleansed standardized versions of the input records for
the user to choose from

68. What is the primary purpose of grouping data?


A. Standardize data in preparation for matching
B. Normalize data in preparation for matching
C. Improve match performance
D. Improve match result

69. Which transformation would you typically use to remove noise, symbols or words?
A. Decision
B. Standardizer
C. Comparison
D. Parser

70. Which of the following will impact the total amount of time it takes to execute a column
profiling process?
A. Number of Columns in the file
B. Number of Rows in the source data
C. Amount of memory and speed of the processor on the server
D. All of the above
1: List of application services part of the Data Quality Standard Edition:

 Analyst Service
 Content Management Service
 Data Integration Service
 Model Repository Service
 Search Service

2: What are the different databases required while installing IDQ

 Model repository for the Model Repository Service.


 Data object cache database to cache logical data objects and virtual tables.
 Profiling warehouse to perform data profiling and discovery.
 Workflow database to store run-time metadata for workflows.
 Reference data warehouse to store reference table data for the Content Management Service.

3: What are the different types of Data Profiling you can do in IDQ

 Basic profile
 Rule Based profile
 Enterprise Data Domain Discovery
 Multiple profile
 Pkey and Fkey analysis
 Join Analysis

4: What is the difference between Live and Stage drill down

 Live drill down fetches data from source


 Stage drill down fetches data from profile warehouse

5: What are the strategies used in Standardizer

 Replace reference table matches with valid values


 Replace reference table matches with custom string
 Remove values with reference table matches
 Remove custom string

6: Can we use a labeler to standardize records ?

 Yes, cause we can use reference table in Labeler with the “Replace reference table matches with valid values” option

7: Is Update strategy Active or Passive transformation and why

 Active transformation cause it changes the rowtype

8: Can we use a mapplet in profile ?

 Yes only after validating it as a rule

9: Mapplet contains the following transformations : Expression, Standardizer, sorter, labler.


Can this be used in a data profile ?

 No, cause you can’t validate a mapplet as rule if it contains active transformations

10: What is the difference between Link score and Driver score in Match Transformation
 Link score : Each of the records are matched with the other records, in the cluster. Records having greater match
with each other are assigned high scores whereas outliers are given low scores
o Use Case : Link score is used when you are trying to find how many records are near similar in a cluster.
 Driver Score : The record with highest sequence ID is given as highest score (1) and all other records in the cluster are
scored w.r.t to that record.
o Use Case : When you have identified your reference record and you are trying to find the nearest close
match to it in other records of the cluster.

11: Can we use an expression instead of key gen before match

 No, it has to be KeyGen transformation as Match expects Groupkey value from KeyGen transformation only. Mapping
will fail

12: What all are the fuzzy match algorithms in IDQ ?

 Edit Distance : Use Edit Distance to compare words or short text strings such as a name.
 Hamming distance : Use Hamming Distance when the position of characters in a string is a critical factor, such as in
product or ZIP
 Bigram : To compare long text strings, such as freeform address lines.
 Jaro Distance : to compare words or short text strings where similarity of initial characters is a priority
 Reverse Hamming : Reverse Hamming Distance analyzes data in the same manner as Hamming Distance but reads
input field characters from right to left.

13: What is consolidation transformation and it's utility


 Used to create a best copy within a Match cluster

14: What is merge transformation


 Used to concatenate values of multiple columns into one column

15: What is IDQ Analyst


 Web based thin client of IDQ where users can create/view Profiles, rules and scorecards.

16: What is Midstream profiling


 A mid-stream profile is a profile that executes directly from a mapping by doing a right-click on a transformation and
"profile now".

17: What are the input templates in Address validator?


 Discrete
 Hybrid
 Multiline

18: What are the different modes on which you can use Address Validator
 Parse : Parses address elements, does not do validation or enrichment
 Batch : focuses on address completion and deliverability
 Interactive : Checks for validtity and completes an incomplete valid address
 Certified : Performs validation in adherence to certification standards of respective countries
 Suggestion List : for partial valid address match , returns all valid addresses for user to choose.
 Address Code Lookup : returns partial or complete valid address based on input address code.
 Country Recognition : Determines destination country for postal address, doesn’t validate address

19: What are the metrics of Address validator


 Mailability Score
 Match code

20: What is an IDQ scorecard. Explain the steps of creating one.


 Select a metric from Profile
 Right click on it and opt “Add to scorecard”
 Select “Create a new scorecard”
 Select all the metrics you want to include in the scorecard
 Identify the valid and invalid values
 Set the DQ threshold
 Create metric group if you want and specify the weightage if required.
 Finish and run the scorecard

21: What is a Content Management Service (CMS)


 The Content Management Service is an application service that manages reference data. The service
provides reference data properties to the Data Integration Service and to the Developer tool. The service
also generates mapplets from rule specifications. You can create rule specifications and generate mapplets
from rule specifications in the Analyst tool.

22: Can address validator run when CMS is offline and why ?
 No, cause address doctor requires the postal reference files to function which is managed by CMS

23: What is an LDO and its use


 LDO stands for Logical data object, it is similar to a DB view where we can write elaborate sqls and create a view
which is a representation of a table. Here also inside LDO we can create elaborate mappings and use the LDO as a
data object which will give us the output dataset of the mapping.

24: Can we use run profile in an LDO ?


 Yes we can.

25: What is a Human Task in IDQ and where is it used ?


 Human task is an IDQ workflow component which can be used to establish a data stewardship workflow via IDQ
Analyst. It can trigger notifications as emails to data stewards whenever some DQ reject happens and is captured by
the exception transformation

26: Suppose you have 100 records in source and you are required to load the records whose ordinality is a multiple of 5.
Explain the logic.
 Use variable ports to number the records. Then take the number and do a modulo division, if the result is zero then
load the record otherwise filter it out.

27: How would you transpose a record set from Column to rows and vice versa.
 Column to rows : Normalizer
 Row to column: use variable port, concatenate , aggregator filter and expression to parse

28: What are the strategies that you can find in Key Generator
 Substring, Soundex and NYSIIS

29 : What are the different types of Lookup Cache


 Static
 Dynamic
 Persistant

30 : In a Joined suppose you are joining Dataset A with 100 records and Dataset B with 10000 which will you take as Master and
Detail and why
 Smaller dataset as Master

31: What is the difference between the PowerCenter Repository Service and the Model Repository Service?
 The PowerCenter application services and PowerCenter application clients use the PowerCenter Repository
Service. The PowerCenter repository has folder-based security.
The other application services, such as the Data Integration Service, Analyst Service, Developer tool, and
Analyst tool, use the Model Repository Service. The Model Repository Service has project-based security.
You can migrate some Model repository objects to the PowerCenter repository.
32: What is the difference between a mapplet and a rule? 
 You can validate a mapplet as a rule. A rule is business logic that defines conditions applied to source data
when you run a profile. You can validate a mapplet as a rule when the mapplet meets the following
requirements:
It contains an Input and Output transformation. 
The mapplet does not contain active transformations. 
It does not specify cardinality between input groups. 

33: Define what do you mean by a workflow?


 Workflow can be defined as a set of instructions which are intended to communicate to the server and
letting it know on how to implement the tasks.

34 : What is the difference between active transformations and a passive transformation in Informatica? Give example
transformations for each?
 Active transformation:
It is a process it changes the number of rows that have gone through the mapping. This process is called as
Active transformation. Some of the Active transformations are:
o Sorter transformations
o Filter transformations
o Joiner transformations
o Rank transformations
o Router transformations
 Passive transformation:
It is a process where it doesn’t change the number of rows that have gone through the mapping. This process is
called as Passive transformation. Some of the Passive transformations are:
o Expression transformation
o Sequence Generator transformation
o Lookup transformation
o External procedure transformation
o Output transformation
o Input transformation, Etc.

 
 35 : How would you ingest multiple flat files of same structure through one physical data object
 Use indirect mode of source data ingestion

36: Can we export an object from IDQ to Powercenter tool. if yes then how ?
 Yes, we can export an object from IDQ to Powercenter tool.
o Connect to Repositary Service
o Locate your Project Folder in Developer tool
o Expand Mapping tab
o Choose your mapping(Needs to be exported)
o Expand Informatica Folder
o Click Object Export File
o Locate under your project folder select the Mapping/mapplets
o Click Browse and select the Location where you want to export it

37: Define Informatica repository?


 Informatica Repository:The Informatica repository is at the center of the Informatica suite. You create a set of
metadata tables within the repository database that the Informatica application and tools access. The
Informatica client and server access the repository to save and retrieve metadata.

38: What is the difference between a connected lookup and unconnected lookup?
 Connected lookup takes input values directly from other transformations in the pipeline.
Unconnected lookup doesn't take inputs directly from any other transformation, but it can be used in any
transformation (like expression) and can be invoked as a function using :LKP expression. So, an unconnected
lookup can be called multiple times in a mapping.

39 : What is Update Strategy transformation?


 Update strategy transformation is an active and connected transformation. It is used to update data in target
table, either to maintain history of data or recent changes. You can specify how to treat source rows in table,
insert, update, delete or data driven.

40: What is the difference between Labeler and parser transformation


 Labeller id used to label or mask certain strings based on regex match whereas parser is used to identify and
segregate records in respective columns based on regex evaluation
 Labeller can use reference table and can be used as standardizer, parser cannot be used to standardize
records.
All the answers are highlighted in yellow, and links are provided for necessary topics -

1. IDQ supports integration with all the big data engines.


True false
2. We can define own conformity criteria to carry out primary key analysis.
True false
https://docs.informatica.com/data-quality-and-governance/data-quality/10-1/data-
discovery-guide/data-discovery-with-informatica-developer/data-object-profiles/primary-
key-discovery.html

3. We can perform consolidation without match transformation with IIR input


True false
https://docs.informatica.com/data-integration/powercenter/10-2/transformation-guide/
identity-resolution-transformation/identity-resolution-transformation-overview.html

https://network.informatica.com/thread/10949

4. We can make analyst and developer collaborate with each other


True false
https://docs.informatica.com/data-quality-and-governance/data-quality/10-1-1/profiling-
getting-started-guide/getting-started-overview/informatica-developer-user-interface/
informatica-developer-tasks.html
5. What is the most accurate filed level algorithm?
https://docs.informatica.com/data-integration/data-services/10-1/developer-
transformation-guide/match-transformations-in-field-analysis/field-match-strategies.html
Bigram
6. What is the use of column profiling?
i. Depends on the need one or more or all columns analyzed
ii. None
iii. To analyze selected columns
iv. To analyze one column
v. To analyze all the columns
7. What is the difference between rule and mapplet?
i. Mapplet can be converted to rule and vice versa
ii. Both are different
iii. All
iv. Both are same in functionality
v. Rule created by analyst
8. What is the difference between custom reference data and free reference data?
i. Both are same
ii. None
iii. Free reference data obtained from informatica and custom reference data created
by analyst.
iv. Both are different
User defined and Informatica Reference data –
https://docs.informatica.com/data-engineering/data-engineering-quality/10-2/reference-
data-guide/introduction-to-reference-data/informatica-reference-data.html

9. What is the default port number of analyst service?


8085
https://docs.informatica.com/data-integration/powercenter/10-2-hotfix-2/security-guide/
domain_security/application_services_and_ports.html

10. We can export custom reference data from idq to power center
True false
https://docs.informatica.com/data-integration/data-services/10-2/developer-mapping-
guide/export-to-powercenter/rules-and-guidelines-for-exporting-to-powercenter.html
11. Where is the profiling data stored?
i. All
ii. Staging database
iii. Model repository
iv. Profiling data warehouse
https://docs.informatica.com/data-integration/powercenter/10-2/developer-tool-guide/
informatica-developer/informatica-developer-overview/informatica-data-quality-and-
profiling.html
https://docs.informatica.com/data-integration/powercenter/10-2/developer-tool-guide/
informatica-developer/informatica-developer-overview/informatica-data-quality-and-
profiling.html

12. We can use expression transformation instead of decision transformation to identify bad
records and get it communicated to analyst.
True false
13. Where we can change the default location of target directory?
i. DIS\processes\targetdir
ii. …

In the target instance – Properties – Output File Directory

14. How many ways to create custom reference table?


i. 4
ii. None
iii. 3
iv. 2
v. 1
https://docs.informatica.com/data-integration/data-services/10-2/_reference-data-
guide_data-services_10-2_ditamap/reference_tables_in_the_analyst_tool/
creating_a_reference_table_in_the_reference_table_editor.html

15. What is the use of column profiling?


https://docs.informatica.com/data-integration/data-services/10-0/profile-guide/profiling-
with-informatica-analyst/column-profiles-in-informatica-analyst/column-profiling-
process.html

16. What is the default threshold for match transformation?


0.9
17. Default location for idq content files?
https://docs.informatica.com/data-quality-and-governance/data-quality/10-1-1/content-
installation-guide/content-installation-overview/data-quality-content-installer.html

18. We can drill down venn diagram in idq


True false
https://docs.informatica.com/data-quality-and-governance/data-quality/10-4-0/getting-
started-guide/getting-started-with-informatica-developer/lesson-3--run-a-profile-on-source-
data/task-2--view-join-analysis-results.html

19. We can have limit on adding match strategies


True false
20. We can process unstructured data in idq
True false
21. We can export an object from idq to powercenter tool.
True false
22. What is a custom reference table?
23. What is standardization used for?
24. What is rule?
25. Role of idd in idq?
26. Weight is same as threshold value in match
True false
27. Who fix bad or duplicate records?
i. Manager
ii. Analyst
iii. Developer
iv. None
v. Data steward
28. Standard output port after passing regular expression?
29. Idq supports runtime column propagation
True false
30. What is the use of data quality table?
i. To fix bad records
ii. To fix duplicate records
iii. None
iv. To fix duplicate records, to fix any kind of data
v. To fix any kind of data
31. Pre requisite for data integration services?
i. Mrs
32. Best way to perform standardization?
33. Any restriction for custom reference table size.
True false
34. Labeler and standardizer are similar.
True false
35. We can convert manual profiling to automated.
True false
36. What is the repository that we use for idq?
37. What is the limit on adding match strategies?
38. How many standard output ports in email aafter parsing using regular expression?
i. 4
ii. 3
iii. 2
iv. 5

39. What is match score formula for bigram distance


i. Number of matching pairs/total pairs
40. We can use expression transformation instead of decision transformation to identify bad
data analyst.
True false
41. What is a custom reference table?
42. We can build common logic for all the column based on the type.
True false
43. Types of input file for parameterization in idq?
i. Xml
ii. Txt
iii. All
iv. Json
44. What is the indirect port number of tomcat?
45. How many filed level algorithms are supported by idq?
46. We can configure push down optimization at DE in idq.
True false
47. We can define more than one user defined output ports with only token set without using
regular expression in idq?
True false
48. We can create physical data objects based on more than two input files.
True false
49. The logical data object follow the linear data model in idq.
True false
50. How to define any kind of logic in consolidation in idq?
51. Reading of multiple files in parallel is possible in idq.
True false
52. Idq data element to lics…
53. Which transformations that are used to identify and collaborate bad records to another list?
54. Is customer accelerate in idq?
IDQ MCQ

All the answers are highlighted in yellow, and links are provided for necessary topics -
1. What is difference between custom reference data and free reference data?
a. Both are same
b. Both are different
c. Free reference data obtained from informatica and custom reference data created by analyst
d. None
User defined and Informatica Reference data –
https://docs.informatica.com/data-engineering/data-engineering-quality/10-2/reference-data-guide/
introduction-to-reference-data/informatica-reference-data.html
2. How many ways to create custom reference table?
a. 4
b. None
c. 3
d. 2
e. 1
https://docs.informatica.com/data-integration/data-services/10-2/_reference-data-guide_data-services_10-
2_ditamap/reference_tables_in_the_analyst_tool/
creating_a_reference_table_in_the_reference_table_editor.html
3. Where we can change the default location of target directory?
a. DIS\processes\targetdir
b. Define own profile
c. Client
d. Dis\client
e. None
In the target instance – Properties – Output File Directory
https://docs.informatica.com/data-integration/powercenter/10-5/advanced-workflow-guide/parameters-and-variables-in-
sessions/working-with-session-parameters/changing-the-target-file-and-directory.html

4. We can use expression TF instead of decision transformation to identify bad records and get it to communicated to
analyst
a. True
b. False

https://docs.informatica.com/data-integration/data-services/10-2/developer-transformation-guide/bad-record-exception-
transformation/bad-record-exception-transformation-overview.html

5. Where is the profiling data stored?


a. All of the listed
b. Staging database
c. Model repository
d. Profiling data warehouse
https://docs.informatica.com/data-integration/powercenter/10-2/developer-tool-guide/informatica-
developer/informatica-developer-overview/informatica-data-quality-and-profiling.html
https://docs.informatica.com/data-integration/powercenter/10-2/developer-tool-guide/informatica-
developer/informatica-developer-overview/informatica-data-quality-and-profiling.html

6. We can export custom reference data from IDQ to power Center.


a. True
b. False
https://docs.informatica.com/data-integration/data-services/10-2/developer-mapping-guide/export-to-
powercenter/rules-and-guidelines-for-exporting-to-powercenter.html
7. IDQ supports integration with all the big engines.
a. True
b. False

8. What is default port number for analyst service?


a. 8080
b. 8085
c. 8089
d. ....
https://docs.informatica.com/data-integration/powercenter/10-2-hotfix-2/security-guide/
domain_security/application_services_and_ports.html

9. What is difference between a rule and mapplet?


a. Mapplet can be converted to rule and vice-versa
b. Both are different
c. All of the listed
d. Both are same in functionality
e. Rule created by analyst
10. What is use of column profiling?
a. Depends on the need one or more or all columns are analyzed
b. None
c. To analyze selected columns
d. To analyze one column
e. To analyze all the columns
11. We can make analyst and developer collaborate with each other
a. True
b. False
12. What is the most accurate field level algorithm?
a. Hamming distance
b. Bigram distance
c. Jaro distance
d. Edit distance
e. Reverse hamming distance
Go through all other match algorithms with example –
https://docs.informatica.com/data-integration/data-services/10-1/developer-transformation-guide/comparison-
transformation/field-matching-strategies/reverse-hamming-distance.html
13. We can define our own conformity criteria to carry out primary key analysis.
a. True
b. False
https://docs.informatica.com/data-quality-and-governance/data-quality/10-1/data-discovery-guide/data-
discovery-with-informatica-developer/data-object-profiles/primary-key-discovery.html
14. We can perform consolidation without match transformation with IIR input
a. True
b. False
https://docs.informatica.com/data-integration/powercenter/10-2/transformation-guide/identity-resolution-
transformation/identity-resolution-transformation-overview.html
https://network.informatica.com/thread/10949
15. What is the Type of input for parameterization in IDQ?
a. Xml
b. Flat/text file
c. Json
d. All of the listed

https://docs.informatica.com/data-quality-and-governance/data-quality/10-1/developer-mapping-guide/mapping-parameters/
parameter-files/parameter-file-structure.html

16. How many standard output ports in email after parsing using regular expression?
a. 3 ------ pattern <email (Part-1)>@<host (Part-2)>.<domain(part-3)>
b. 5
c. 4
d. 2
17. Who fixes Bad/duplicate Records?
a. Analyst
b. SME
c. DEVELOPER
d. Manger
e. All of the listed
https://docs.informatica.com/data-quality-and-governance/data-quality/10-2/exception-management-guide/
introduction-to-exception-management/exception-management-process-flow/bad-records-example.html
18. What is the indirect port no of apache tomcat?
a. 8080
b. 8443
c. None
d. 8089
19. We can configure push down optimization at DIS in informatica Developer.
a. True
b. False
https://docs.informatica.com/data-quality-and-governance/data-quality/10-4-1/performance-tuning-guide/mapping-
optimization/pushdown-optimization.html
20. Weight is same as threshold value in match?
a. True
b. False
https://docs.informatica.com/data-integration/data-services/10-0/developer-transformation-guide/weighted-average-
transformation/weighted-match-scores-example.html
21. What are IDQ transformations that are used to identify and collaborate bad record to analyst.
a. Decision
b. All of the listed
c. Exception
22. What are the prerequisites to configure consolidation transformation?
a. None
b. Group key only
c. Group key and column profiling
https://docs.informatica.com/data-integration/data-services/10-0/developer-transformation-guide/consolidation-
transformation/consolidation-transformation-ports.html
23. What is standardization?
a. Both
b. To unify the data
c. Convert n representation to one
24. Labeler transformation and standardizer are similar
a. True
b. False
25. We can build common logic for all the columns based on type
a. True
b. False
26. What is use of data quality table?
a. To fix bad record
b. To fix duplicate records
c. None
d. To fix duplicate records, to fix any kind of data
e. To fix any kind of data
27. can we export an object from IDQ to power center tool?
a. True
b. False
https://docs.informatica.com/data-integration/data-services/10-2/developer-mapping-guide/export-to-powercenter/
exporting-an-object-to-powercenter.html
28. What is the role of idd in idq ?
a. None of the listed
b. Data Governance
c. For reference data
d. nothing
https://network.informatica.com/thread/18326
29. we have provision to make analyst data that would become as input to the developer.
a. True
b. False
30. we can integrate idq meta elements to IICS.
a. True
b. False – need cloud data integration or cloud data quality for IICS
31. what is the repository that we use for idq.
a. Model Repository
b. Power Center Repository
c. Model+Pc
d. none
https://docs.informatica.com/data-integration/data-services/10-0/installation-and-configuration-guide/before-you-
install-the-services/plan-the-domain/plan-the-application-services/model-repository-service.html
32. we can define more than one user defined output ports with only token set but no regular expression
a. True
b. False
Regular expression we can set number of output ports, but token set its already defined
33. how many field level algorithms are supported by match transformation?
a. 5
b. 4
c. 1
d. .......
e. ...........
34. we can process unstructured data in idq
a. True – data processor transformation
b. False

https://docs.informatica.com/data-integration/data-services/10-1/developer-transformation-guide/data-processor-
transformation/data-processor-transformation-overview.html

35. Does Idq supports runtime column propagation


a. True
b. False
36. we can have customer accelerators in idq
a. True
b. False
37. we can convert manual profiling to automated
a. True
b. False
38. we can make analyst and developer collaborate with each other
a. True
b. False
39. what is scorecard in informatica?
a. to identify valid or invalid data
b. both the listed options
c. to measure data quality progress
40. Driver score and link score are same
a. True
b. False
41. we can create physical data object based on more than 2 input files
a. True
b. False
42. we can drill down venn diagram in idq
a. True
b. False
43. The logical data object follows linear data model.
a. True
b. False
44. what is address doctor in informatica
a. address validation
b. address verification
c. Phone number verification
d. Identity verification
45. Best Way to perform standardization?
a. With token Set
b. Manual
c. Reference table
d. All
46. What is rule in idq?
a. Created by analyst
b. Virtual Column
c. All of the listed options
d. Math expressionobke
47. What is match score formula for bigram distance?
a. Number of matching characters/maximun row length
b. Number of unmatched pairs/total pairs
c. None
d. Number of matching pairs/total pairs
48. What is the difference Custom reference data vs free reference data ?
a. Both are same
b. Both are different
c. Free reference data obtained from informatica and custom reference data created by analyst
d. None of the listed
49. Reading of identify files in parallel is possible in IDQ
a. True
b. False
50. We have a limit on adding match strategies in IDQ.
a. True
b. False
51. What is the default location of IDQ content files?
a. Server\DQContent
b. All of the listed
c. Third party store
d. Client\DQContent
52. What is the best option to define any kind of logic in consolidation?
a. Row based
b. advanced
c. all of the listed
d. default
53. There is restriction for custom reference table size
a. True
b. False
https://docs.informatica.com/data-integration/data-services/10-2/_reference-data-guide_data-services_10-
2_ditamap/reference_tables_in_the_analyst_tool/rules_and_guidelines_for_reference_tables.html
54. Can we use run profile in a logical data object?
a. True
b. False
55. Transformation used to figure out bad records
a. Exception
b. Both
c. Decision
56. We define our own conformity criteria to carry out primary key analysis
a. True
b. False
57. Best match transformation algorithm
a. BiGram
b. Hamming Distance
c. Reverse Hamming
d. Edit
58. We can remove junk characters using standardization
a. True
b. False
59. What is the prerequisite/s for the creation of data integration service?
a. mrs
b. mrs,pcrep
c. all of the listed
d. mrs and as
MRS – DIS – Analyst Service (AS) – Content Management Service (CMS) (Dependency)
60. What is the default threshold set in match transformation?
a. 1
b. 0.95
c. 0.8
d. 0.75
e. None – 0.9
61. We can export custom reference data from IDQ to power center
a. True
b. False

You might also like