Professional Documents
Culture Documents
INFORMATICA PowerCenter
INFORMATICA PowerConnect
INFORMATICA Power Mart
INFORMATICA Power Exchange
INFORMATICA Power Analysis
INFORMATICA Power Quality
We need INFORMATICA while working with data systems which contain data to perform
certain operations along with a set of rules. INFORMATICA facilitates operations line
cleaning and modifying data from structured and unstructured data systems.
Only the rows that meet the Filter Condition pass through the Filter transformation to the next
transformation in the pipeline. TRUE and FALSE are the implicit return values from any
filter condition we set. If the filter condition evaluates to NULL, the row is assumed to be
FALSE.
The numeric equivalent of FALSE is zero (0) and any non-zero value is the equivalent of
TRUE.
As an ACTIVE transformation, the Filter transformation may change the number of rows
passed through it. A filter condition returns TRUE or FALSE for each row that passes
through the transformation, depending on whether a row meets the specified condition. Only
rows that return TRUE pass through this transformation. Discarded rows do not appear in the
session log or reject files.
7. What Informatica product line does Master Informatica Question and Answer set™
cover?
Answer: We cover Informatica PowerCenter® Mapping Designer, Workflow Manager,
Repository Manager, Workflow Monitor and PowerCenter® Administration Console
(version in
Loading important data into a data warehouse from any source known as Target.
Extracting data from a data warehouse from any sources such as database tables or files.
Transforming the data received from different sources in an organized way. Some of the
notable sources where data are received include SAP solutions, Teradata, or web services.
13. State the limitations where we cannot use Joiner in the mapping pipeline?
Answer: The Joiner transformation accepts input from most transformations. However, the
following are the limitations:
Joiner transformation cannot be used when either of the input pipelines contains an Update
Strategy transformation.
Joiner transformation cannot be used if we connect a Sequence Generator transformation
directly before the Joiner transformation.
18. In informatics server Which files are created during the session rums?
Answer: Errors log, Bad file, Workflow low and session log namely files are created during
the session rums.Company.
If you want to save all of the log files created by every run, and then select the option Save
session log for these runs –> Session TimeStamp
22. Suppose we configure Sorter transformations in the master and detail pipelines with
the following sorted ports in order: ITEM_NO, ITEM_NAME, PRICE. When we
configure the join condition, what are the guidelines we need to follow to maintain the
sort order?
Answer: If we have sorted both the master and detail pipelines in order of the ports say
ITEM_NO, ITEM_NAME and PRICE we must ensure that:
Use ITEM_NO in the First Join Condition.
If we add a Second Join Condition, we must use ITEM_NAME.
If we want to use PRICE as a Join Condition apart from ITEM_NO, we must also use
ITEM_NAME in the Second Join Condition.
If we skip ITEM_NAME and join on ITEM_NO and PRICE, we will lose the input sort order
and the Integration Service fails the session.
25. What are the performance considerations when working with Aggregator
Transformation?
Answer:
1. Filter the unnecessary data before aggregating it. Place a Filter transformation in the
mapping before the Aggregator transformation to reduce unnecessary aggregation.
2. Improve performance by connecting only the necessary input/output ports to subsequent
transformations, thereby reducing the size of the data cache.
3. Use Sorted input which reduces the amount of data cached and improves session
performance.
26. What will happen if the SELECT list COLUMNS in the Custom override SQL
Query and the OUTPUT PORTS order in SQ transformation do not match?
Answer: Mismatch or Changing the order of the list of selected columns to that of the
connected transformation output ports may result in session failure.
27. What are the different ways to implement parallel processing in Informatica?
Answer: We can implement parallel processing using various types of partition algorithms:
Database partitioning: The Integration Service queries the database system for table partition
information. It reads partitioned data from the corresponding nodes in the database.
Hash Auto-Keys Partitioning: The Powercenter Server uses a hash function to group rows of
data among partitions. When the hash auto-key partition is used, the Integration Service uses
all grouped or sorted ports as a compound partition key. You can use hash auto-keys
partitioning at or before Rank, Sorter, and unsorted Aggregator transformations to ensure that
rows are grouped properly before they enter these transformations.
Hash User-Keys Partitioning: Here, the Integration Service uses a hash function to group
rows of data among partitions based on a user-defined partition key. You can individually
choose the ports that define the partition key.
Key Range Partitioning: With this type of partitioning, you can specify one or more ports to
form a compound partition key for a source or target. The Integration Service then passes
data to each partition depending on the ranges you specify for each port.
Pass-through Partitioning: In this type of partitioning, the Integration Service passes all rows
from one partition point to the next partition point without redistributing them.
28. Mention a few design and development best practices for Informatica?
Answer: Mapping design tips: Standards – sticking to consistent standards is beneficial in the
long run. This includes naming conventions, descriptions, environment settings, parameter
files, documentation, among others.
Reusability – in order to react quickly to potential changes, use Informatica components like
applets, worklets, and reusable transformations.
Scalability – when designing and developing mappings, it is a good practice to keep volumes
in mind. This is caching, queries, partitioning, initial vs incremental loads.
Simplicity – it is recommended to create multiple mappings instead of a few complex ones.
Use Staging Area and try to keep the processing logic as clear and simple as possible.
Modularity – use the modular design technique (common error handling, reprocessing).
Mapping development best practices
Source Qualifier – use shortcuts, extract only the necessary data, limit the reading of columns
and rows on the source. Try to use the default query options (User Defined Join, Filter)
instead of using SQL Query override which may impact database resources and make unable
to use partitioning and push-down.
Expressions – use local variables to limit the number of redundant calculations, avoid
datatype conversions, reduce invoking external scripts (coding outside of Informatica),
provide comments, use operators (||, +, /) instead of functions. Keep in mind that numeric
operations are generally faster than string operations.
Filter – use the Filter transformation as close to the source as possible. If multiple filters need
to be applied, usually it’s more efficient to replace them with Router.
Aggregator – use sorted input, also use as early (close to the source) as possible and filter the
data before aggregating.
Joiner – try to join the data in Source Qualifier wherever possible, and avoid outer joins. It is
good practice to use a source with fewer rows, such as a Master source.
Lookup – relational lookup should only return ports that meet the condition. Call
Unconnected Lookup in expression (IIF). Replace large lookup tables with joins whenever
possible. Review the database objects and add indexes to database columns when possible.
Use Cache Calculator in session to eliminate paging in lookup cache.
35. What are the transformations that cannot be placed between the sort origin and the
Joiner transformation so that we do not lose the input sort order?
Answer: The best option is to place the Joiner transformation directly after the sort origin to
maintain sorted data. However, do not place any of the following transformations between
the sort origin and the Joiner transformation:
Custom
UnsortedAggregator
Normalizer
Rank
Union transformation
XML Parser transformation
XML Generator transformation
Mapplet [if it contains any one of the above-mentioned transformations]
36. What is the format of INFORMATICA objects in a repository? What are the
databases that INFORMATICA can connect to Windows?
Answer: INFORMATICA objects can be written in XML format.
SQL Server
Oracle
MS Access
MS Excel
DB2
Sybase
Teradata.
37. How does the Aggregator Transformation handle NULL values?
Answer: By default, the aggregator transformation treats null values as NULL in aggregate
functions. But we can specify to treat null values in aggregate functions is NULL or zero.
38. What differs when we choose the Sorted Input for Aggregator Transformation?
Answer: Integration Service creates the index and data caches files in memory to process the
Aggregator transformation. If the Integration Service requires more space as allocated for the
index and data cache sizes in the transformation properties, it stores overflow values in cache
files i.e. paging to disk. One way to increase session performance is to increase the index and
data cache sizes in the transformation properties. But when we check Sorted Input the
Integration Service uses memory to process an Aggregator transformation it does not use
cache files.
39. Under what conditions selecting Sorted Input in aggregator will still not boost
session performance?
Answer
1. Incremental Aggregation, session option is enabled.
2. The aggregate expression contains nested aggregate functions.
3. Source data is data-driven.
40. Suppose we have a Source Qualifier transformation that populates two target tables.
How do you ensure TGT2 is loaded after TGT1?
Answer: In the Workflow Manager, we can Configure Constraint-based load ordering for a
session. The Integration Service orders the target load on a row-by-row basis. For every row
generated by an active source, the Integration Service loads the corresponding transformed
row first to the primary key table, then to the foreign key table.
Hence if we have one Source Qualifier transformation that provides data for multiple target
tables having primary and foreign key relationships, we will go for Constraint-based load
ordering.
Constraint-based loading
Revisiting Filter Transformation
After creating session we can use server manager or command-line program pmcmd to stop
or start the session.
Master Source
Detail Source
Following joins can be created using Joiner transformation as in SQL.
Normal Join
Full Outer Join
Master outer join(Right Outer Join)
Detail outer join(Left Outer Join)
46. What is Aggregator Transformation in INFORMATICA?
Answer: Aggregator Transformation can be active or connected. It works as the GROUP BY
clause in SQL. It is useful to perform aggregate calculations on groups in INFORMATICA
PowerCenter. It performs an aggregate calculation on data using aggregate type function viz.
SUM, AVG, MAX, and MIN.