IDQ Reference
IDQ Reference
IDQ Reference
5. Which is NOT true when exporting Ref. data tables for use by DQ mappings to PC
- Set the ref. data location to the services level not lower
- Select the data service
- The correct code page is automatically assigned and can’t be modified by the user
- During export select export ref. data
7. Utilize DQ accelerators in PC
- Install accelerators in MR and export accelerator mappings and/or mapplets to PC
8. IDQ objects are integrated with PC, how is ref. content handled?
- Ref. content needs to be exported into a flat file and placed in the appropriate location
10. Exception for duplicate that score above the threshold, which one is true –
- Close enough to use automatic consolidation and can be output to standard output
11. Not a consolidation function within simple consolidation strategies
- Most Data
14. Extension of Data matching process and a precursor of data consolidation process
- Association
5. When applying a reference table in the Labeler in token mode, which of the following
statements is correct?
A. The order of the reference tables defined will not impact your results displayed
B. Selecting Inclusive mode will output items that are present in the reference table under the
label
C. Selecting Exclusive mode will exclude items that are present in the reference table
D. Any items that do not appear in the reference table are not displayed in the output
11. Which of the following will NOT work when adding transformation to mappings?
A. Double click on the transformation in the transformation palette
B. Click and drag the transformation onto the mapping
C. From the Mapping menu, select Add transformation
D. Right click and choose Add transformation on the mapping
13. During a project export, typically what will the output be from the export?
A. An xml file containing all the mapping/mapplet/object definition
B. An xml file containing all the mapping/mapplet/object definition and a zip file containing all
reference data if required
C. The contents of each Project (including the dependent reference tables) in a zip file.
D. Projects will be exported to an .irob file
14. Which of the following is NOT a required step for using DQ for Excel?
A. In Excel, choose to Add a service and either enter the URL address or select the .wsdl or .xml
imported
B. Build and test a Web Services Mapping in the Developer Tool
C. Export the Mapping to PowerCenter and test
D. Create and Deploy the mapping as an Application
16. What happens to the records when corrections are made and a status of Accept is assigned?
A. The corrections are committed to the Staging DB and an UPDATED_STATUS of Accepted is
assigned to the record
B. The corrections are committed to the Staging DB and record is automatically pushed into
another DB
C. Records are corrected and sent to an Accepted DB table
D. Records are accepted and are pushed out to Target/Production DB
17. In Version 9.1, what does the Content Management Service do?
A. It manages the dictionaries created in the Analyst Tool
B. It provides the DIS with the information on the location of the address validation data as
well as AV Configuration Settings
C. It runs Identity Matching
D. It makes the Profile Repository available to outside reporting applications like Jaspersoft
18. Which of the following statements are true regarding DQ for Excel?
A. DQ for Excel allows .xls files be used as source and target objects in mappings
B. DQ for Excel enables non Informatica users to benefit from DQ Rules by consuming and
applying DQ rules directly from MS Excel data/worksheets
C. DQ for Excel can be used in any version of MS Excel
D. In order to be able to use DQ for Excel, both PowerCenter and Data Quality Developer must
be installed
19. What types of profiling can be performed in the Developer Tool in Data Quality?
A. Column profiling, Primary key inference, Dependency inference
B. Column and Join Profiling only
C. Column, Join Analysis, Mid Stream and Comparitive Profiling
D. Column, Join Analysis, Mid Stream, Comparitive, Primary and Foreign Key and Overlap
Profiling.
23. In the Case Converter transformation, which of the following is not a valid Case Transform?
A. Upper Case
B. Lower Case
C. Toggle merge Case
D. Sentence Case
25. If the inputs to Edit Distance Strategy below are ‘john’ and NULL, what will be the resulting
match score?
A. 0
B. 1
C. 0.5
D. 1.5
29. Which informatica option provides solutions to common data quality issues in a country,
region or industry?
A. Populations
B. Accelerators
C. Dictionaries
D. Mapplets
30. Which are correct statements for IDQ Grouping and Matching? Chose 2 answers
A. IDQ Field Level matching does not utilize grouping
B. When field level matching is performed, the records within each group will be compared
against each other
C. When field level matching is performed, matching will be performed across multiple
groups in a single match transformation
D. When field level matching is performed, matching will not be performed across groups,
therefore it is imperative grouping is performed on a complete and accurate field(s)
32. Which of the following cannot be executed in BOTH the Analyst and the Developer Tool?
A. A rule created in the Analyst Tool
B. A mapplet created in the Developer Tool and validated as a rule
C. A mapplet created in the Developer Tool and validated as a mapplet
D. A reusable rule created in the Analyst tool
34. When is sampling for Profiling a good idea? Select the best answer.
A. When the data set contains sensitive information like credit card number
B. When the probability of data quality issues is small
C. When the client provides assurances that the quality of data is good
D. When the volume of data is large and there are storage and/or time constraints.
35. The IDQ and PowerCenter repositories all share the same schema.
A. TRUE
B. FALSE
36. Select the transformation that will not allow a DQ mapplet to validate as a rule
A. Expression Transformation
B. Address Validation Transformation
C. Aggregator Transformation
D. Labeler Transformation
37. With respect to field matching in IDQ, which best describes the approach to address
matching?
A. No standardization is necessary, map input fields to the match transformation
B. Standardize address fields and use address line, locality, region and postcode fields as input
to the match transformation
C. Standardize address fields and use the discrete address component fields, locality, region
and postcode input to the match transformation
D. Standardize address fields, create a group key field which contains parts of address fields
and on group key for address matching.
38. Users would export DQ mappings to Power Center for which of the following reasons?
(Choose 4)
A. Performance
B. Scalability
C. Batch access
D. DQ as part of the ETL Process
40. When changes are made to a mapplet in IDQ 9.1 that has been integrated into PowerCenter,
how are the changes propagated to the integrated mapplet?
A. Those changes are automatically propagated and managhed through domain settings
B. The user must manually re-export the IDQ mapplet to PowerCenter
C. The IDQ versioning system automatically synchs with PowerCenter and updates the mapplet
to current version.
D. None of the above.
42. What Address Validator output field should you use to determine the quality of address
validation?
A. Locality/City
B. ElementInputStatus
C. Match Code
D. MSA
43. Which phrase best describes how organizations should view their approach to DQ?
A. As a one-off project addressing the issues in their Data Warehouse in Batch
B. As on ongoing process addressing issues over time with a combination of fit-for-purpose
approach
C. As a real-time process addressing issues as they are introduced
D. As a batch process to be implemented to run on a monthly basis
45. Which of the following is NOT TRUE when it comes to creating and updating Reference tables?
A. Reference Tables can be updated from a Value Frequency list in the Developer
B. All except Un-Managed Reference tables can be updated through the Reference Table
manager in the Analyst tool
C. Reference Tables maintained in the Analyst and Developer tools are stored in separate
repositories
D. Reference Tables can be imported and exported using the Developer Tool
46. When using field matching which is the best approach if using First Name as a match field?
A. Never use First Name as it is not necessary
B. No standardization of First Name field is necessary
C. Use the Edit Distance Strategy
D. Create Soundex or NYSIIS value for the First Name and use it in match
47. Which statements describe Consolidation Transformation? Choose 2 answers.
A. A passive transformation
B. An active transformation
C. A transformation which is found only in PowerCenter
D. Creates a single, consolidated record from records identified as possible duplicates by the
Match transformation.
48. When profiling, why is it good to sort the value lists in both ascending and descending order?
A. There is no specific reason. It is simply a step of an ordered best practice approach
B. It is good because it is easier to determine if a field is numeric or not
C. Values with leading spaces or punctuation will typically sort to the top or bottom making
them easier to identify
D. None of the above
49. What versions of PowerCenter can an IDQ 9.1 mapping be exported to?
A. PC 8.6.1 and higher
B. PC 8.51 and higher
C. All PC Versions
D. PC 8.11, PC 8.51 and PC 8.6
51. When sharing projects across teams which of the following is True?
A. The same icon is used to indicate a shared and non-shared project
B. Comments that are added to a profile in the Analyst can be viewed by the Developer in the
Developer tool
C. Permissions need to be assigned to each folder within the project so users of shared
projects can access objects
D. Rules that were created in the Analyst tool will be visible in Developer but Developer built
tools will not be made available in the Analyst.
52. Which two are index key levels in the Identity Matching? Choose 2 answers
A. Narrow
B. Limited
C. Typical
D. Extended
53. Which of the following options is TRUE? The Advanced Import Wizard differs from the Basic
Import Wizard in the following ways
A. The Advanced Method allows conflict resolution and the Basic doesn’t
B. The Advanced Import Wizard allows objects to be imported into multiple target projects
with resolutions to dependencies.
C. Import steps can be undone in the Advanced Wizard but not in Basic one.
D. Users have a choice regarding which project objects are imported using Basic and Advanced
imports.
56. When importing a flat file into a project, Developer and Analyst tool work exactly the same?
A. TRUE
B. FALSE
57. A customer wants to be able to review data on the overall quality of each record, which
technique should be used?
A. Scorcecarding
B. Matching
C. Grading
D. Standardization
58. A Customer has provided a number of delimited flat files that req profiling but not the file
definition what should you do first?
A. Allow the wizard bin developer, to automatically select the datatype when creating the data
object
B. Set all of the data types to text and length longer than specified. The profiler will tell you
what they should be
C. Import the file into a database table before profiling
D. Change the code page of the file to UTF-8.
59. You want to derive the gender from a firstname, how would you do this?
A. Use a ref table with the FN as the valid col and the gender as the 2 nd col. Then choose to
replace ref tab matches with the valid values when applying the ref tab in a strategy in the
standardizer. Use the FN as i/p.
B. Use a ref tab with the gender as the valid column and the FN in the next col. Then choose n
replace ref tab matches with the valid values when applying the ref tab in a strategy in the
standardizer. Use FN as i/p
C. Use decision and create an if then else rule
D. Use a ref tab with the gender as the valid column and the FN in the next col. Then choose n
replace ref tab matches with Custom strings when applying the ref tab in a strategy in the
standardizer. Use the FN as i/p.
60. What transformation would u use to standardize data using a ref table
A. Labeler
B. Decision
C. Standardizer
D. Parser
61. You have just standardized values in a field and u would like to review how the modified data
appears alongside the original data. Which profiling methods best be suited for this?
A. Col profiling
B. Mid stream
C. Join analysis
D. Comparative
64. Which two of the following values can be parameterized in the match transformation?
A. Threshold
B. Scoring methods
C. Match field i/p
D. Weights
65. What is required for customers to access and use IDQ Accelerator content?
A. Customer must license each Accelerator pack individually to download and use the content
B. No additional steps required
C. IDQ Customers simply need to request an IPS resource to download and install the content
D. Accelerators content is available only with PC license
66. What must you do in order to utilize Data Quality Accelerators in PC?
A. Import accelerator directly to PC
B. Install accelerator directly to PC
C. Install accelerator directly to Developer
D. Export accelerators mapplets and/or mapping to PC
67. When you select Suggestion list mode which statement is true?
A. The AV transformation searches the address ref data and returns all possible address
matches with the i/p address
B. The AV transformation searches the address ref data and returns most likely candidates
C. The AV transformation parses data into suggested address field without performing
validation
D. The AV transformation will return cleansed standardized versions of the input records for
the user to choose from
69. Which transformation would you typically use to remove noise, symbols or words?
A. Decision
B. Standardizer
C. Comparison
D. Parser
70. Which of the following will impact the total amount of time it takes to execute a column
profiling process?
A. Number of Columns in the file
B. Number of Rows in the source data
C. Amount of memory and speed of the processor on the server
D. All of the above
1: List of application services part of the Data Quality Standard Edition:
Analyst Service
Content Management Service
Data Integration Service
Model Repository Service
Search Service
3: What are the different types of Data Profiling you can do in IDQ
Basic profile
Rule Based profile
Enterprise Data Domain Discovery
Multiple profile
Pkey and Fkey analysis
Join Analysis
Yes, cause we can use reference table in Labeler with the “Replace reference table matches with valid values” option
No, cause you can’t validate a mapplet as rule if it contains active transformations
10: What is the difference between Link score and Driver score in Match Transformation
Link score : Each of the records are matched with the other records, in the cluster. Records having greater match
with each other are assigned high scores whereas outliers are given low scores
o Use Case : Link score is used when you are trying to find how many records are near similar in a cluster.
Driver Score : The record with highest sequence ID is given as highest score (1) and all other records in the cluster are
scored w.r.t to that record.
o Use Case : When you have identified your reference record and you are trying to find the nearest close
match to it in other records of the cluster.
No, it has to be KeyGen transformation as Match expects Groupkey value from KeyGen transformation only. Mapping
will fail
Edit Distance : Use Edit Distance to compare words or short text strings such as a name.
Hamming distance : Use Hamming Distance when the position of characters in a string is a critical factor, such as in
product or ZIP
Bigram : To compare long text strings, such as freeform address lines.
Jaro Distance : to compare words or short text strings where similarity of initial characters is a priority
Reverse Hamming : Reverse Hamming Distance analyzes data in the same manner as Hamming Distance but reads
input field characters from right to left.
18: What are the different modes on which you can use Address Validator
Parse : Parses address elements, does not do validation or enrichment
Batch : focuses on address completion and deliverability
Interactive : Checks for validtity and completes an incomplete valid address
Certified : Performs validation in adherence to certification standards of respective countries
Suggestion List : for partial valid address match , returns all valid addresses for user to choose.
Address Code Lookup : returns partial or complete valid address based on input address code.
Country Recognition : Determines destination country for postal address, doesn’t validate address
22: Can address validator run when CMS is offline and why ?
No, cause address doctor requires the postal reference files to function which is managed by CMS
26: Suppose you have 100 records in source and you are required to load the records whose ordinality is a multiple of 5.
Explain the logic.
Use variable ports to number the records. Then take the number and do a modulo division, if the result is zero then
load the record otherwise filter it out.
27: How would you transpose a record set from Column to rows and vice versa.
Column to rows : Normalizer
Row to column: use variable port, concatenate , aggregator filter and expression to parse
28: What are the strategies that you can find in Key Generator
Substring, Soundex and NYSIIS
30 : In a Joined suppose you are joining Dataset A with 100 records and Dataset B with 10000 which will you take as Master and
Detail and why
Smaller dataset as Master
31: What is the difference between the PowerCenter Repository Service and the Model Repository Service?
The PowerCenter application services and PowerCenter application clients use the PowerCenter Repository
Service. The PowerCenter repository has folder-based security.
The other application services, such as the Data Integration Service, Analyst Service, Developer tool, and
Analyst tool, use the Model Repository Service. The Model Repository Service has project-based security.
You can migrate some Model repository objects to the PowerCenter repository.
32: What is the difference between a mapplet and a rule?
You can validate a mapplet as a rule. A rule is business logic that defines conditions applied to source data
when you run a profile. You can validate a mapplet as a rule when the mapplet meets the following
requirements:
It contains an Input and Output transformation.
The mapplet does not contain active transformations.
It does not specify cardinality between input groups.
34 : What is the difference between active transformations and a passive transformation in Informatica? Give example
transformations for each?
Active transformation:
It is a process it changes the number of rows that have gone through the mapping. This process is called as
Active transformation. Some of the Active transformations are:
o Sorter transformations
o Filter transformations
o Joiner transformations
o Rank transformations
o Router transformations
Passive transformation:
It is a process where it doesn’t change the number of rows that have gone through the mapping. This process is
called as Passive transformation. Some of the Passive transformations are:
o Expression transformation
o Sequence Generator transformation
o Lookup transformation
o External procedure transformation
o Output transformation
o Input transformation, Etc.
35 : How would you ingest multiple flat files of same structure through one physical data object
Use indirect mode of source data ingestion
36: Can we export an object from IDQ to Powercenter tool. if yes then how ?
Yes, we can export an object from IDQ to Powercenter tool.
o Connect to Repositary Service
o Locate your Project Folder in Developer tool
o Expand Mapping tab
o Choose your mapping(Needs to be exported)
o Expand Informatica Folder
o Click Object Export File
o Locate under your project folder select the Mapping/mapplets
o Click Browse and select the Location where you want to export it
38: What is the difference between a connected lookup and unconnected lookup?
Connected lookup takes input values directly from other transformations in the pipeline.
Unconnected lookup doesn't take inputs directly from any other transformation, but it can be used in any
transformation (like expression) and can be invoked as a function using :LKP expression. So, an unconnected
lookup can be called multiple times in a mapping.
https://network.informatica.com/thread/10949
10. We can export custom reference data from idq to power center
True false
https://docs.informatica.com/data-integration/data-services/10-2/developer-mapping-
guide/export-to-powercenter/rules-and-guidelines-for-exporting-to-powercenter.html
11. Where is the profiling data stored?
i. All
ii. Staging database
iii. Model repository
iv. Profiling data warehouse
https://docs.informatica.com/data-integration/powercenter/10-2/developer-tool-guide/
informatica-developer/informatica-developer-overview/informatica-data-quality-and-
profiling.html
https://docs.informatica.com/data-integration/powercenter/10-2/developer-tool-guide/
informatica-developer/informatica-developer-overview/informatica-data-quality-and-
profiling.html
12. We can use expression transformation instead of decision transformation to identify bad
records and get it communicated to analyst.
True false
13. Where we can change the default location of target directory?
i. DIS\processes\targetdir
ii. …
All the answers are highlighted in yellow, and links are provided for necessary topics -
1. What is difference between custom reference data and free reference data?
a. Both are same
b. Both are different
c. Free reference data obtained from informatica and custom reference data created by analyst
d. None
User defined and Informatica Reference data –
https://docs.informatica.com/data-engineering/data-engineering-quality/10-2/reference-data-guide/
introduction-to-reference-data/informatica-reference-data.html
2. How many ways to create custom reference table?
a. 4
b. None
c. 3
d. 2
e. 1
https://docs.informatica.com/data-integration/data-services/10-2/_reference-data-guide_data-services_10-
2_ditamap/reference_tables_in_the_analyst_tool/
creating_a_reference_table_in_the_reference_table_editor.html
3. Where we can change the default location of target directory?
a. DIS\processes\targetdir
b. Define own profile
c. Client
d. Dis\client
e. None
In the target instance – Properties – Output File Directory
https://docs.informatica.com/data-integration/powercenter/10-5/advanced-workflow-guide/parameters-and-variables-in-
sessions/working-with-session-parameters/changing-the-target-file-and-directory.html
4. We can use expression TF instead of decision transformation to identify bad records and get it to communicated to
analyst
a. True
b. False
https://docs.informatica.com/data-integration/data-services/10-2/developer-transformation-guide/bad-record-exception-
transformation/bad-record-exception-transformation-overview.html
https://docs.informatica.com/data-quality-and-governance/data-quality/10-1/developer-mapping-guide/mapping-parameters/
parameter-files/parameter-file-structure.html
16. How many standard output ports in email after parsing using regular expression?
a. 3 ------ pattern <email (Part-1)>@<host (Part-2)>.<domain(part-3)>
b. 5
c. 4
d. 2
17. Who fixes Bad/duplicate Records?
a. Analyst
b. SME
c. DEVELOPER
d. Manger
e. All of the listed
https://docs.informatica.com/data-quality-and-governance/data-quality/10-2/exception-management-guide/
introduction-to-exception-management/exception-management-process-flow/bad-records-example.html
18. What is the indirect port no of apache tomcat?
a. 8080
b. 8443
c. None
d. 8089
19. We can configure push down optimization at DIS in informatica Developer.
a. True
b. False
https://docs.informatica.com/data-quality-and-governance/data-quality/10-4-1/performance-tuning-guide/mapping-
optimization/pushdown-optimization.html
20. Weight is same as threshold value in match?
a. True
b. False
https://docs.informatica.com/data-integration/data-services/10-0/developer-transformation-guide/weighted-average-
transformation/weighted-match-scores-example.html
21. What are IDQ transformations that are used to identify and collaborate bad record to analyst.
a. Decision
b. All of the listed
c. Exception
22. What are the prerequisites to configure consolidation transformation?
a. None
b. Group key only
c. Group key and column profiling
https://docs.informatica.com/data-integration/data-services/10-0/developer-transformation-guide/consolidation-
transformation/consolidation-transformation-ports.html
23. What is standardization?
a. Both
b. To unify the data
c. Convert n representation to one
24. Labeler transformation and standardizer are similar
a. True
b. False
25. We can build common logic for all the columns based on type
a. True
b. False
26. What is use of data quality table?
a. To fix bad record
b. To fix duplicate records
c. None
d. To fix duplicate records, to fix any kind of data
e. To fix any kind of data
27. can we export an object from IDQ to power center tool?
a. True
b. False
https://docs.informatica.com/data-integration/data-services/10-2/developer-mapping-guide/export-to-powercenter/
exporting-an-object-to-powercenter.html
28. What is the role of idd in idq ?
a. None of the listed
b. Data Governance
c. For reference data
d. nothing
https://network.informatica.com/thread/18326
29. we have provision to make analyst data that would become as input to the developer.
a. True
b. False
30. we can integrate idq meta elements to IICS.
a. True
b. False – need cloud data integration or cloud data quality for IICS
31. what is the repository that we use for idq.
a. Model Repository
b. Power Center Repository
c. Model+Pc
d. none
https://docs.informatica.com/data-integration/data-services/10-0/installation-and-configuration-guide/before-you-
install-the-services/plan-the-domain/plan-the-application-services/model-repository-service.html
32. we can define more than one user defined output ports with only token set but no regular expression
a. True
b. False
Regular expression we can set number of output ports, but token set its already defined
33. how many field level algorithms are supported by match transformation?
a. 5
b. 4
c. 1
d. .......
e. ...........
34. we can process unstructured data in idq
a. True – data processor transformation
b. False
https://docs.informatica.com/data-integration/data-services/10-1/developer-transformation-guide/data-processor-
transformation/data-processor-transformation-overview.html