Professional Documents
Culture Documents
Introduction To Informatica Powercenter
Introduction To Informatica Powercenter
Informatica PowerCenter
Data Warehousing
Data warehousing is the entire
process of data extraction,
transformation, and loading of data to
the warehouse and the access of the
data by end users and applications
Data Mart
A data mart stores data for a limited
number of subject areas, such as
marketing and sales data. It is used
to support specific applications.
An independent data mart is created
directly from source systems.
A dependent data mart is populated
from a data warehouse.
Data Sources ETL Software Data Stores Data Analysis Users
Tools and
Applications
Transaction Data S
T
IBM A
Prod G
I
N SQL
G ANALYSTS
Mkt IMS
A
R Cognos
Ascential E
HR VSAM Dat a Mart s
A Teradata SAS
IBM MANAGERS
O Finance
Oracle P
Fi n E Load
R Dat a Essbase Queries,Reporting,
Extract A Warehouse DSS/EIS,
T Informatica Data Mining
Acctg Syba se Marketing
I
O EXECUTIVES
Other Internal Data N Micro Strategy
A Meta
L Dat a Sales
ERP SAP Sagent
D
A Microsoft Siebel
Web Data T
A Business OPERATIONAL
Infor mix Objects PERSONNEL
Clickstream S
T
SAS O
External Data R Web
E Browser
Demographic Harte-
Hanks Clean/Scrub CUSTOMERS/
Trans form SUPPLIERS
Fi rst logic
Need For ETL Tool
Data Extraction
Repository Manager
Designer
Workflow Manager
Informatica provides the following integrated
components:
File. Fixed and delimited flat file, COBOL file, and XML.
Repository Client
Repository Server
----------------------------
Repository Agent
Repository Database
Creating a Repository
To create Repository
Shared Folders
When you create a folder, you can configure it as a shared
folder. Shared folders allow users to create shortcuts to objects
in the folder. If you have reusable transformation that you want
to use in several mappings or across multiple folders, you can
place the object in a shared folder.
Adding Repository
Creation of mapping
Tools provided by Designer
Source Analyzer: Importing Source definitions for
Flat file, XML, COBOL and relational Sources.
Choose Mappings-Create.
In the Mapping Name dialog box, enter <Mapping Name> as the name
of the new mapping and click OK.
Passive transformations
Expression performs simple calculations
Lookup looks up values and passes to other objects
Sequence generator generates unique ID values
Stored procedure calls a stored procedure and captures return values
Update strategy allows for logic to insert, update, delete, or reject
data
Transformations Contd..
Create the transformation. Create it in the Mapping
Designer as part of a mapping, in the Mapplet Designer as
part of a Mapplet, or in the Transformation Developer as
a reusable transformation.
The Source Qualifier represents the records that the Informatica Server reads when it runs a
session. You can use the Source Qualifier to perform the following tasks:
Join data originating from the same source database. You can join two or more tables
with primary-foreign key relationships by linking the sources to one Source Qualifier.
Filter records when the Informatica Server reads source data. If you include a filter
condition, the Informatica Server adds a WHERE clause to the default query.
Specify an outer join rather than the default inner join. If you include a user-defined
join, the Informatica Server replaces the join information specified by the metadata in the
SQL query.
Specify sorted ports. If you specify a number for sorted ports, the Informatica Server
adds an ORDER BY clause to the default SQL query.
Select only distinct values from the source. If you choose Select Distinct, the
Informatica Server adds a SELECT DISTINCT statement to the default SQL query.
Create a custom query to issue a special SELECT statement for the Informatica Server to
read source data. For example, you might use a custom query to perform aggregate
calculations or execute a stored procedure
Configuring Source Qualifier Transformation
Option Description
Defines a custom query that replaces the default query the Informatica Server uses
SQL Query
to read data from sources represented in this Source Qualifier
User-Defined Specifies the condition used to join data from multiple sources represented in the
Join same Source Qualifier transformation
Source Filter Specifies the filter condition the Informatica Server applies when querying records.
Indicates the number of columns used when sorting records queried from relational
Number of sources. If you select this option, the Informatica Server adds an ORDER BY to
Sorted the default query when it reads source records. The ORDER BY includes the
Ports number of ports specified, starting from the top of the Source Qualifier.
When selected, the database sort order must match the session sort order.
Sets the amount of detail included in the session log when you run a session
Tracing Level
containing this transformation.
Select Distinct Specifies if you want to select only unique records. The Informatica Server includes a
SELECT DISTINCT statement if you choose this option.
Joiner Transformation
While a Source Qualifier transformation can join data originating from a common source database,
the Joiner transformation joins two related
heterogeneous sources residing in different locations or file systems. The combination of sources
can be varied. You can use the following sources:
If two relational sources contain keys, then a Source Qualifier transformation can easily join the
sources on those keys. Joiner transformations typically combine information from two
different sources that do not have matching keys, such as flat file sources.
The Joiner transformation allows you to join sources that contain binary data.
Creating a Joiner Transformation
To create a Joiner Transformation:
Case-Sensitive
If selected, the Informatica Server uses case-sensitive string
String
comparisons when performing joins on string columns.
Comparison
Specifies the directory used to cache master records and the index to
these records. By default, the caches are created in a directory
Cache Directory specified by the server variable $PMCacheDir. If you override the
directory, be sure there is enough disk space on the file system. The
directory can be a mapped or mounted drive.
Specifies the type of join: Normal, Master Outer, Detail Outer, or Full
Join Type
Outer.
Lookup Transformation
1) Receives input values directly from of a Receives input values from the result of
the pipe line transformation. LKP expression within other
transformation.
2) U can use a dynamic or static cache U can use a static cache.
3) Cache includes all lookup columns used Cache includes all lookup out put ports.
in the mapping.
4) Support user defined default values Does not support user defined default
values
Diff between Static & Dynamic Cache
1) U can not insert or update the U can insert rows into the cache as u pass
cache to the target
Setting Description
Treat all records as inserts. If inserting the record violates a primary or
Insert foreign key constraint in the database, the Informatica Server rejects the
record.
Treat all records as deletes. For each record, if the Informatica Server finds a
corresponding record in the target table (based on the primary key value),
Delete
the Informatica Server deletes it. Note that the primary key constraint must
exist in the target definition in the repository.
Treat all records as updates. For each record, the Informatica Server looks for
a matching primary key value in the target table. If it exists, the Informatica
Update
Server updates the record. Again, the primary key constraint must exist in
the target definition.
The Informatica Server follows instructions coded into Update Strategy
transformations within the session mapping to determine how to flag records
for insert, delete, update, or reject.
Data If the mapping for the session contains an Update Strategy transformation,
Driven this field is marked Data Driven by default.
If you do not choose Data Driven setting, the Informatica Server ignores all
Setting Use To
Populate the target tables for the first time, or maintaining a historical
Insert data warehouse. In the latter case, you must set this strategy for the
entire data warehouse, not just a select group of target tables.
Update target tables. You might choose this setting whether your data
warehouse contains historical data or a snapshot. Later, when you
Update configure how to update individual target tables, you can determine
whether to insert updated records as new records or use the updated
information to modify existing records in the target.
Exert finer control over how you flag records for insert, delete, update,
or reject. Choose this setting if records destined for the same table
Data
need to be flagged on occasion for one operation (for example, update),
Driven
or for a different operation (for example, reject). In addition, this
setting provides the only way you can flag records for reject.