Professional Documents
Culture Documents
You specify the target loadorder based on source qualifiers in a maping.If you have the
multiple source qualifiers connected to the multiple targets,You can designatethe order in
which informatica server loads data into the targets.
How do you configure mapping in informatica?
You should configure the mapping with the least number of transformations and
expressions to do the most amount of work possible. You should minimize the amount of
data moved by deleting unnecessary links between transformations.
For transformations that use data cache (such as Aggregator, Joiner, Rank, and Lookup
transformations), limit connected input/output or output ports. Limiting the number of
connected input/output or output ports reduces the amount of data the transformations
store in the data cache.
You can also perform the following tasks to optimize the mapping:
Configure single-pass reading.
Optimize datatype conversions.
Eliminate transformation errors.
Optimize transformations.
Optimize expressions. You should configure the mapping with the least number of
transformations and expressions to do the most amount of work possible. You should
minimize the amount of data moved by deleting unnecessary links between
transformations.
For transformations that use data cache (such as Aggregator, Joiner, Rank, and Lookup
transformations), limit connected input/output or output ports. Limiting the number of
connected input/output or output ports reduces the amount of data the transformations
store in the data cache.
You can also perform the following tasks to optimize the mapping:
Configure single-pass reading.
Optimize datatype conversions.
Eliminate transformation errors.
Optimize transformations.
Optimize expressions.
when conventional datawarehousing is able to handle and answer complex queries
than wat is the need of OLTP and OLAP?
Designed for analysis of business measures by categories and attributes
Optimized for bulk loads and large, complex, unpredictable queries that access many
rows per table.
Loaded with consistent, valid data; requires no real time validation
Supports few concurrent users relative to OLTP
Again, all these are covered in the tutorials, why and how you should select.
How can we join 3 database like Flat File, Oracle, Db2 in Informatrica.?
You have to use two joiner transformations.fIRST one will join two tables and the next
one will join the third with the resultant of the first joiner.
How do you transfert the data from data warehouse to flatfile?
You can write a mapping with the flat file as a target using a DUMMY_CONNECTION.
A flat file target is built by pulling a source into target space using Warehouse Designer
tool.
How can u work with remote database in informatica?did u work directly by using
remote connections?
You can work with remote,
But you have to
Configure FTP
Connection details
IP address
User authentication
What is difference between IIF and DECODE function?
You can use nested IIF statements to test multiple conditions. The following example
tests for various conditions and returns 0 if sales is zero or negative:
IIF( SALES > 0, IIF( SALES < 50, SALARY1, IIF( SALES < 100, SALARY2, IIF(
SALES < 200, SALARY3, BONUS))), 0 )
You can use DECODE instead of IIF in many cases. DECODE may improve readability.
The following shows how you can use DECODE instead of IIF :
SALES > 0 and SALES < 50, SALARY1,
SALES > 49 AND SALES < 100, SALARY2,
SALES > 99 AND SALES < 200, SALARY3,
SALES > 199, BONUS)
What are the new features of the server manager in the informatica 5.0?
You can use command line arguments for a session or batch.This allows you to change
the values of session parameters,and mapping parameters and maping variables.
Parallel data processing: This feature is available for powercenter only.If we use the
informatica server on a SMP system, You can use multiple CPU’s to process a session
concurently.
Process session data using threads: Informatica server runs the session in two processes.
Can we use aggregator/active transformation after update strategy transformation?
You can use aggregator after update strategy. The problem will be, once you perform the
update strategy, say you had flagged some rows to be deleted and you had performed
aggregator transformation for all rows, say you are using SUM function, then the deleted
rows will be subtracted from this aggregator transformation.
what is the best way to show metadata(number of rows at source, target and each
transformation level, error related data) in a report format?
You can select these details from the repository table. you can use the view
REP_SESS_LOG to get these data
Can you start a batches with in a batch?
You can not. If you want to start batch that resides in a batch,create a new independent
batch and copy the necessary sessions into the new batch.
How can you create or import flat file definition in to the warehouse designer?
You can not create or import flat file defintion in to warehouse designer directly.Instead
you must analyze the file in source analyzer,then drag it into the warehouse
designer.
When you drag the flat file source defintion into warehouse desginer workspace,the
warehouse designer creates a relational target defintion not a file defintion.If you want to
load to a file,configure the session to write to a flat file.When the informatica server runs
the session,it creates and loads the flat file.
In my source table 1000 rec's r there.I want to load 501 rec to 1000 rec into my
Target table ?how can u do this ?
You can overide the sql Query in Wofkflow Manager. LIke
select * from tab_name where rownum<=1000
minus
select * from tab_name where rownum<=500;
This will work fine. Try it and get back to me if u have any issues about the same.
How to Generate the Metadata Reports in Informatica?
You can generate PowerCenter Metadata Reporter from a browser on any workstation,
even a workstation that does not have PowerCenter tools installed.
What r the methods for creating reusable transforamtions?
You can design using 2 methods
using transformation developer
create normal one and promote it to reusable
What are the diffrence between joiner transformation and source qualifier
transformation?
You can join hetrogenious data sources in joiner transformation which we can not achieve
in source qualifier transformation.
You need matching keys to join two relational sources in source qualifier
transformation.Where as you doesn’t need matching keys to join two sources.
Two relational sources should come from same datasource in sourcequalifier.You can join
relatinal sources which are coming from diffrent sources also.
Can any body write a session parameter file which will change the source and
targets for every session. i.e different source and targets for each session run.
You are supposed to define a parameter file. And then in the Parameter file, you can
define two parameters, one for source and one for target.
Give like this for example:
$Src_file = c:\program files\informatica\server\bin\abc_source.txt
$tgt_file = c:\targets\abc_targets.txt
Then go and define the parameter file:
[folder_name.WF:workflow_name.ST:s_session_name]
$Src_file =c:\program files\informatica\server\bin\abc_source.txt
$tgt_file = c:\targets\abc_targets.txt
If its a relational db, you can even give an overridden sql at the session level...as a
parameter. Make sure the sql is in a single line.
In a sequential batch can you run the session if previous session fails?
Yes.By setting the option always runs the session.
Can you use the maping parameters or variables created in one maping into any
other reusable transformation?
Yes.Because reusable tranformation is not contained with any maplet or maping.
Can you generate reports in Informatcia?
Yes. By using Metadata reporter we can generate reports in informatica.
There are 3 depts in dept table and one with 100 people and 2nd with 5 and 3rd with
some 30 and so. i want to diplay those deptno where more than 10 people exists
Yes! the answer provided is absolutely right. by an SQL application(Oracle).
If you want to perform it thru informatica, the Fire the same query in the SQL Override
of Source qualifier transformation and make a simple pass thru mapping.
Other wise, you can also do it by using a Filter.Router transformation by giving the
condition there deptno>=10.
Can you copy the session to a different folder or repository?
Yes. By using copy session wizard You can copy a session in a different folder or
repository. But that target folder or repository should consists of mapping of that session.
If target folder or repository is not having the maping of copying session ,
You should have to copy that maping first before you copy the session.
Can Informatica be used as a Cleansing Tool? If Yes, give example of
transformations that can implement a data cleansing routine.
Yes, we can use Informatica for cleansing data. some time we use stages to cleansing the
data. It depends upon performance again else we can use expression to cleasing data.
For example an feild X have some values and other with Null values and assigned to
target feild where target feild is notnull column, inside an expression we can assign space
or some constant value to avoid session failure.
The input data is in one format and target is in another format, we can change the format
in expression.
we can assign some default values to the target to represent complete set of data in the
target.
Can batches be copied/stopped from server manager?
Yes, we can stop the batches using server manager or pmcmd command
Can Informatica load heterogeneous targets from heterogeneous sources?
Yes it can. For example...Flat File and Relations sources are joined in the mapping, and
later, Flat File and relational targets are loaded.
What is tracing level?
Ya its the level of information storage in session log.
The option comes in the properties tab of transformations. By default it remains
"Normal". Can be
Verbose Initialisation
Verbose Data
Normal
or Terse.
Can we run a group of sessions without using workflow manager
ya Its Posible using pmcmd Command with out using the workflow Manager run the
group of session.
as per my knowledge i give the answer.
How do you create single lookup transformation using multiple tables?
Write a override sql query. Adjust the ports as per the sql query.
How can we join the tables if the tables have no primary and forien key relation and
no matchig port to join?
without common column or common data type we can join two sources using dummy
ports.
1.Add one dummy port in two sources.
2.In the expression trans assing '1' to each port.
2.Use Joiner transformation to join the sources using dummy port(use join conditions).
hope this will help.
Describe two levels in which update strategy transformation sets?
Within a session: When you configure a session, yoYou can instruct the Informatica
Server to either treat all records in the same way (for example, treat all records as inserts),
or use instructions coded into the session mapping to flag records for different database
operations.
Within a mapping: Within a mapping, you use the Update Strategy transformation to flag
records for insert, delete, update, or reject.
How do you handle decimal places while importing a flatfile into informatica?
While importing flat file definetion just specify the scale for a neumaric data type. in the
mapping, the flat file source supports only number datatype(no decimal and integer). In
the SQ associated with that source will have a data type as decimal for that number port
of the source.
source ->number datatype port ->SQ -> decimal datatype.Integer is not supported. hence
decimal is taken care.
How do you handle decimal places while importing a flatfile into informatica?
while importing flat file definetion just specify the scale for a neumaric data type. in the
mapping, the flat file source supports only number datatype(no decimal and integer). In
the SQ associated with that source will have a data type as decimal for that number port
of the source.
source ->number datatype port ->SQ -> decimal datatype.Integer is not supported. hence
decimal is taken care.
What is parameter file?
When you start a workflow, you can optionally enter the directory and name of a
parameter file. The Informatica Server runs the workflow using the parameters in the file
you specify.
For UNIX shell users, enclose the parameter file name in single quotes:
-paramfile '$PMRootDir/myfile.txt'
For Windows command prompt users, the parameter file name cannot have beginning or
trailing spaces. If the name includes spaces, enclose the file name in double quotes:
-paramfile ”$PMRootDir\my file.txt”
Note: When you write a pmcmd command that includes a parameter file located on
another machine, use the backslash (\) with the dollar sign ($). This ensures that the
machine where the variable is defined expands the server variable.
pmcmd startworkflow -uv USERNAME -pv PASSWORD -s SALES:6258 -f east -w
wSalesAvg -paramfile '\$PMRootDir/myfile.txt'
What is aggregate cache in aggregator transforamtion?
When you run a workflow that uses an Aggregator transformation, the Informatica Server
creates index and data caches in memory to process the transformation. If the Informatica
Server requires more space, it stores overflow values in cache files.
Why you use repository connectivity?
When you edit,schedule the sesion each time,informatica server directly communicates
the repository to check whether or not the session and users are valid.All the metadata of
sessions and mappings will be stored in repository.
Briefly explian the Versioning Concept in Power Center 7.1.
When you create a version of a folder referenced by shortcuts, all shortcuts continue to
reference their original object in the original version. They do not automatically update to
the current folder version.
For example, if you have a shortcut to a source definition in the Marketing folder, version
1.0.0, then you create a new folder version, 1.5.0, the shortcut continues to point to the
source definition in version 1.0.0.
Maintaining versions of shared folders can result in shortcuts pointing to different
versions of the folder. Though shortcuts to different versions do not affect the server, they
might prove more difficult to maintain. To avoid this, you can recreate shortcuts pointing
to earlier versions, but this solution is not practical for much-used objects. Therefore,
when possible, do not version folders referenced by shortcuts.
What is source qualifier transformation?
When you add a relational or a flat file source definition to a mapping, you need to
connect it to a Source Qualifier transformation. The Source Qualifier represents the rows
that the Informatica Server reads when it executes a session.
Join data originating from the same source database. You can join two or more tables
with primary-foreign key relationships by linking the sources to one Source Qualifier.
Filter records when the Informatica Server reads source data. If you include a filter
condition, the Informatica Server adds a WHERE clause to the default query.
Specify an outer join rather than the default inner join. If you include a user-defined join,
the Informatica Server replaces the join information specified by the metadata in the SQL
query.
Specify sorted ports. If you specify a number for sorted ports, the Informatica Server adds
an ORDER BY clause to the default SQL query.
Select only distinct values from the source. If you choose Select Distinct, the Informatica
Server adds a SELECT DISTINCT statement to the default SQL query.
Create a custom query to issue a special SELECT statement for the Informatica Server to
read source data. For example, you might use a custom query to perform aggregate
calculations or execute a stored procedure.
What is a source qualifier?-
When you add a relational or a flat file source definition to a mapping, you need to
connect it to a Source Qualifier transformation. The Source Qualifier represents the rows
that the Informatica Server reads when it executes a session.
What is source qualifier transformation?
When you add a relational or a flat file source definition to a maping,U need to connect it
to a source qualifer transformation.The source qualifier transformation represnets the
records that the informatica server reads when it runs a session.
What is incremantal aggregation?
When using incremental aggregation, you apply captured changes in the source to
aggregate calculations in a session. If the source changes only incrementally and you
can capture changes, you can configure the session to process only those changes. This
allows the Informatica Server to update your target incrementally, rather than
forcing it to process the entire source and recalculate the same calculations each time you
run the session.
If you are workflow is running slow in informatica. Where do you start trouble
shooting and what are the steps you follow?
When the work flow is running slowly u have to find out the bottlenecks
in this order target source mapping session system
What is exact use of 'Online' and 'Offline' server connect Options while defining
Work flow in Work flow monitor? . The system hangs when 'Online' server connect
option. The Informatica is installed on a Personal laptop.
When the repo is up and the PMSERVER is also up, workflow monitor always will be
connected on-line.
When PMserver is down and the repo is still up we will be prompted for an off-line
connection with which we can just monitor the workflows
Explain about perform recovery?
When the Informatica Server starts a recovery session, it reads the
OPB_SRVR_RECOVERY table and notes the row ID of the last row committed to the
target database.
The Informatica Server then reads all sources again and starts processing from the next
row ID. For example, if the Informatica Server commits 10,000 rows before the
session fails, when you run recovery, the Informatica Server bypasses the rows up to
10,000 and starts loading with row 10,001.
By default, Perform Recovery is disabled in the Informatica Server setup. You must
enable Recovery in the Informatica Server setup before you run a session so the
Informatica Server can create and/or write entries in the OPB_SRVR_RECOVERY table.
How the informatica server sorts the string values in Ranktransformation?
When the informatica server runs in the ASCII data movement mode it sorts session data
using Binary sortorder.If you configure the seeion to use a binary sort order,the
informatica server caluculates the binary value of each string and returns the specified
number of rows with the higest binary values for the string.
In which circumstances that informatica server creates Reject files?
When it encounters the DD_Reject in update strategy transformation.
Violates database constraint
Filed in the rows was truncated or overflowed.
How the informatica server sorts the string values in Ranktransformation?
When Informatica Server runs in UNICODE data movement mode ,then it uses the sort
order configured in session properties.
What are the joiner caches?
When a Joiner transformation occurs in a session, the Informatica Server reads all the
records from the master source and builds index and data caches based on the
master rows.
After building the caches, the Joiner transformation reads records from the detail source
and perform joins.
When do u we use dynamic cache and when do we use static cache in an connected
and unconnected lookup transformation
We use dynamic cache only for connected lookup. We use dynamic cache to check
whether the record already exists in the target table are not. And depending on that, we
insert,update or delete the records using update strategy. Static cache is the default cache
in both connected and unconnected. If u select static cache on lookup table in infa, it
own't update the cache and the row in the cache remain constant. We use this to check the
results and also to update slowly changing records
What are variable ports and list two situations when they can be used?
We have mainly tree ports Inport, Outport, Variable port. Inport represents data is flowing
into transformation. Outport is used when data is mapped to next transformation. Variable
port is used when we mathematical caluculations are required. If any addition i will be
more than happy if you can share.
How to load time dimension?
We can use SCD Type 1/2/3 to load any Dimensions based on the requirement.
where do we use MQ series source qualifier, application multi group source
qualifier. just give an example for a better understanding
We can use a MQSeries SQ when we have a MQ messaging system as source(queue).
When there is need to extract data from a Queue, which will basically have messages in
XML format, we will use a JMS or a MQ SQ depending on the messaging system. If you
have a TIBCO EMS Queue, use a JMS source and JMS SQ and an XML Parser, or if you
have a MQ series queue, then use a MQ SQ which will be associated with a Flat file or a
Cobal file.
In a sequential Batch how can we stop single session?
We can stop it using PMCMD command or in the monitor right click on that perticular
session and select stop.this will stop the current session and the sessions next to it.
Can you start a session inside a batch idividually?
We can start our required session only in case of sequential batch.in case of concurrent
batch we cant do like this.
What is a view? How it is related to data independence?And what are the different
types of views,and what is Materialize view
views
view is a combination of one or more table.view does not stores the data,it just store the
query in file format.If we excutes the query the query will fetch the data from the tables
and just make it to view for us. Types views materilized view
What are various types of Aggregation?
Various types of aggregation are SUM, AVG, COUNT, MAX, MIN, FIRST, LAST,
MEDIAN, PERCENTILE, STDDEV, and VARIANCE.
What is mystery dimention?
using Mystery Dimension ur maitaining the mystery data in ur Project.
what is the look up transformation?
Using it we can access the data from a relational table which is not a source in the
mapping.
For Ex:Suppose the source contains only Empno, but we want Empname also in the
mapping.Then instead of adding another tbl which contains Empname as a source ,we
can Lkp the table and get the Empname in target.
How do you create a mapping using multiple lookup transformation?
Use unconnected lookup if same lookup repeats multiple times.
How can we eliminate duplicate rows from flat file?
Use Sorter Transformation. When you configure the Sorter Transformation to treat output
rows as distinct, it configures all ports as part of the sort key. It therefore discards
duplicate rows compared during the sort operation
How can you improve session performance in aggregator transformation?
Use sorted input.
What is the look up transformation?
Use lookup transformation in u’r mapping to lookup data in a relational
table,view,synonym.
Informatica server queries the look up table based on the lookup ports in the
transformation.It compares the lookup transformation port values to lookup table column
values based on the look up condition.
How to get two targets T1 containing distinct values and T2 containing duplicate
values from one source S1.
Use filter transformation for loading the target with no duplicates. and for the other
transformation load it directly from source.
How to delete duplicate rows in flat files source is any option in informatica
Use a sorter transformation , in that u will have a "distinct" option make use of it .
When we create a target as flat file and source as oracle.. how can i specify first rows
as column names in flat files...
use a pre sql statement....but this is a hardcoding method...if you change the column
names or put in extra columns in the flat file, you will have to change the insert statement
Why did you use stored procedure in your ETL Application?
usage of stored procedure has the following advantages
1checks the status of the target database
2drops and recreates indexes
3determines if enough space exists in the database
4performs aspecilized calculation
why did u use update stategy in your application?
Update Strategy is used to drive the data to be Inert, Update and Delete depending upon
some condition. You can do this on session level tooo but there you cannot define any
condition.For eg: If you want to do update and insert in one mapping...you will create
two flows and will make one as insert and one as update depending upon some
condition.Refer : Update Strategy in Transformation Guide for more information
What r the options in the target session of update strategy transsformatioin?
Update as Insert:
This option specified all the update records from source to be flagged as inserts in the
target. In other words, instead of updating the records in the target they are inserted as
new records.
Update else Insert:
This option enables informatica to flag the records either for update if they are old or
insert, if they are new records from source.
How do we do unit testing in informatica?
How do we load data in informatica ?
Unit testing are of two types
1. Quantitaive testing
2.Qualitative testing
Steps.
1.First validate the mapping
2.Create session on themapping and then run workflow.
Once the session is succeeded the right click on session and go for statistics tab.
There you can see how many number of source rows are applied and how many number
of rows loaded in to targets and how many number of rows rejected.This is called
Quantitative testing.
If once rows are successfully loaded then we will go for qualitative testing.
Steps
1.Take the DATM(DATM means where all business rules are mentioned to the
corresponding source columns) and check whether the data is loaded according to the
DATM in to target table.If any data is not loaded according to the DATM then go and
check in the code and rectify it.
This is called Qualitative testing.
This is what a devloper will do in Unit Testing.
How can u complete unrcoverable sessions?
Under certain circumstances, when a session does not complete, you need to truncate the
target tables and run the session from the beginning. Run the session from the
beginning when the Informatica Server cannot run recovery or when running recovery
might result in inconsistent data.
What is the difference between connected and unconnected stored procedures.
Unconnected:
The unconnected Stored Procedure transformation is not connected directly to the flow of
the mapping. It either runs before or after the session, or is called by an expression in
another transformation in the mapping.
connected:
The flow of data through a mapping in connected mode also passes through the Stored
Procedure transformation. All data entering the transformation through the input ports
affects the stored procedure. You should use a connected Stored Procedure transformation
when you need data from an input port sent as an input parameter to the stored procedure,
or the results of a stored procedure sent as an output parameter to another transformation.
Can we eliminate duplicate rows by using filter and router transformation ?if so
explain me in detail .
U can use SQL query for uniqness if the source is Relational
But if the source is Flat file then u should use Shorter or Aggregatot transformation
Can u start a batches with in a batch?
U can not. If u want to start batch that resides in a batch,create a new independent batch
and copy the necessary sessions into the new batch.
in the concept of mapping parameters and variables, the variable value will be saved
to the repository after the completion of the session and the next time when u run
the session, the server takes the saved variable value in the repository and starts
assigning the next value of the saved value. for example i ran a session and in the
end it stored a value of 50 to the repository.next time when i run the session, it
should start with the value of 70. not with the value of 51.
how to do this.
u can do onething after running the mapping,, in workflow manager
start-------->session.
right clickon the session u will get a menu, in that go for persistant values, there u will
find the last value stored in the repository regarding to mapping variable. then remove it
and put ur desired one, run the session... i hope ur task will be done
How can U create or import flat file definition in to the warehouse designer?
U can create flat file definition in warehouse designer.in the warehouse designer,u can
create new target: select the type as flat file. save it and u can enter various columns for
that created target by editing its properties.Once the target is created, save it. u can import
it from the mapping designer.
What r the different types of Type2 dimension maping?
Type2
1. Version number
2. Flag
3.Date
What are the mapings that we use for slowly changing dimension table?
Type1: Rows containing changes to existing dimensions are updated in the target by
overwriting the existing dimension. In the Type 1 Dimension mapping, all rows contain
current dimension data.
Use the Type 1 Dimension mapping to update a slowly changing dimension table when
you do not need to keep any previous versions of dimensions in the table.
Type 2: The Type 2 Dimension Data mapping inserts both new and changed dimensions
into the target. Changes are tracked in the target table by versioning the primary
key and creating a version number for each dimension in the table.
Use the Type 2 Dimension/Version Data mapping to update a slowly changing dimension
table when you want to keep a full history of dimension data in the table. Version
numbers and versioned primary keys track the order of changes to each dimension.
Type 3: The Type 3 Dimension mapping filters source rows based on user-defined
comparisons and inserts only those found to be new dimensions to the target. Rows
containing changes to existing dimensions are updated in the target. When updating an
existing dimension, the Informatica Server saves existing data in different columns
of the same row and replaces the existing data with the updates.
How many ways you create ports?
Two ways:-
1.Drag the port from another transforamtion
2.Click the add buttion on the ports tab.
How many ways you can update a relational source defintion and what are they?
Two ways:-
1. Edit the definition
2. Reimport the defintion.
What are the basic needs to join two sources in a source qualifier?
Two sources should have primary and Foreign key relation ships.
Two sources should have matching data types.
What are the different options used to configure the sequential batches?
Two options
Run the session only if previous session completes sucessfully. Always runs the session.
What are the methods for creating reusable transforamtions?
Two methods:-
1.Design it in the transformation developer.
2.Promote a standard transformation from the mapping designer.After you add a
transformation to the mapping , You can promote it to the status of reusable
transformation.
Once you promote a standard transformation to reusable status,You can demote it to a
standard transformation at any time.
If you change the properties of a reusable transformation in mapping,You can revert it to
the original reusable transformation properties by clicking the revert button.
What r the active and passive transforamtions?
Transformations can be active or passive. An active transformation can change the
number of rows that pass through it, such as a Filter transformation that removes rows
that do not meet the filter condition.
A passive transformation does not change the number of rows that pass through it, such
as an Expression transformation that performs a calculation on data and passes all rows
through the transformation.
What is tracing level and what r the types of tracing level?
Tracing level represents the amount of information that informatcia server writes in a log
file.
Types of tracing level:-
Normal
Verbose
Verbose init
Verbose data
Compare Data Warehousing Top-Down approach with Bottom-up approach
Top down
ODS-->ETL-->Datawarehouse-->Datamart-->OLAP
Bottom up
ODS-->ETL-->Datamart-->Datawarehouse-->OLAP
How can u work with remote database in informatica?did u work directly by using
remote connections?
To work with remote datasource u need to connect it with remote connections.But it is
not preferable to work with that remote source directly by using remote connections
.Instead u bring that source into U r local machine where informatica server resides.If u
work directly with remote source the session performance will decreases by passing less
amount of data across the network in a particular time.
Why use the lookup transformation?
To perform the following tasks:-
Get a related value. For example, if your source table includes employee ID, but you
want to include the employee name in your target table to make your summary data
easier to read.
Perform a calculation. Many normalized tables include values used in a calculation, such
as gross sales per invoice or sales tax, but not the calculated value (such as net
sales).
Update slowly changing dimension tables. You can use a Lookup transformation to
determine whether records already exist in the target.
In certain mapping there are four targets tg1,tg2,tg3 and tg4.
tg1 has a primary key,tg2 foreign key referencing the tg1's primary key,tg3 has
primary key that tg2 and tg4 refers as foreign key,tg2 has foreign key referencing
primary key of tg4 ,the order in which the informatica will load the target?
2]How can I detect aggregate tranformation causing low performance?
To optimize the aggregator transformation, you can use the following options.
Use incremental aggregation
Sort the ports before you perform aggregation
Avoid using aggregator transformation after update strategy, since it might be confusing.
Explain use of update strategy transformation
To flag source records as INSERT, DELETE, UPDATE or REJECT for target database.
Default flag is Insert. This is must for Incremental Data Loading.
How do you load the time dimension.
ime Dimension will generally load manually by using PL/SQL , shell scripts, proc C
etc......
What are the types of data that passes between informatica server and stored
procedure?
Three types of data:-
Input/Out put parameters
Return Values
Status code.
What is update strategy transformation?
This transformation is used to maintain the history data or just most recent changes in to
target table.
Why sorter transformation is an active transformation?
This is type of active transformation which is responsible for sorting the data either in the
ascending order or descending order according to the key specifier. the port on which the
sorting takes place is called as sortkeyport
properties
if u select distinct eliminate duplicates
case sensitive valid for strings to sort the data
null treated low null values are given least priority
How to append the records in flat file(Informatica) ? Where as in Datastage we have
the options
i) overwrite the existing file
ii) Append existing file
This is not there in Informatica v 7. but heard that its included in the latest version 8.0
where u can append to a flat file. Its about to be shipping in the market.
In which condtions we can not use joiner transformation(Limitaions of joiner
transformation)?
This is no longer valid in version 7.2
Now we can use a joiner even if the data is coming from the same source.
If you have four lookup tables in the workflow. How do you troubleshoot to improve
performance?
There r many ways to improve the mapping which has multiple lookups.
1) we can create an index for the lookup table if we have permissions(staging area).
2) divide the lookup mapping into two (a) dedicate one for insert means: source - target,,
these r new rows . only the new rows will come to mapping and the process will be fast .
(b) dedicate the second one to update : source=target,, these r existing rows. only the
rows which exists allready will come into the mapping.
3)we can increase the chache size of the lookup.
How many types of dimensions are available in informatica?
There r 3 types of dimensions
1.star schema
2.snowflake schema
3.glaxy schema
Where should U place the flat file to import the flat file defintion to the designer?
There is no such restrication to place the source file. In performance point of view its
better to place the file in server local src folder. if you need path please check the server
properties availble at workflow manager.
It doesn't mean we should not place in any other folder, if we place in server src folder by
default src will be selected at time session creation.
What is difference b/w Informatica 7.1 and Abinitio
There is a lot of diffrence between informatica an Ab Initio
In Ab Initio we r using 3 parllalisim
but Informatica using 1 parllalisim
In Ab Initio no scheduling option we can scheduled manully or pl/sql script
but informatica contains 4 scheduling options
Ab Inition contains co-operating system
but informatica is not
Ramp time is very quickly in Ab Initio campare than Informatica
Ab Initio is userfriendly than Informatica
How many types of facts and what are they?
There are
Factless Facts:Facts without any measures.
Additive Facts:Fact data that can be additive/aggregative.
Non-Additive facts: Facts that are result of non-additon
Semi-Additive Facts: Only few colums data can be added.
Periodic Facts: That stores only one row per transaction that happend over a period of
time.
Accumulating Fact: stores row for entire lifetime of event.
How to use the unconnected lookup i.e., from where the input has to be taken and
the output is linked?
What condition is to be given?
The unconnected lookup is used just like a function call. in an expression output/variable
port or any place where an expression is accepted(like condition in update strategy etc..),
call the unconnected lookup...something like :LKP.lkp_abc(input_port).......(lkp_abc is
the name of the unconnected lookup...(plz check the exact syntax)).....give the input value
just like we pass parameters to functions, and it'll return the output after looking up.
What r the circumstances that infromatica server results an unreciverable session?
The source qualifier transformation does not use sorted ports.
If u change the partition information after the initial session fails.
Perform recovery is disabled in the informatica server configuration.
If the sources or targets changes after initial session fails.
If the maping consists of sequence generator or normalizer transformation.
If a concuurent batche contains multiple failed sessions.
Normal Load: Normal load will write information to the database log file so that if any
recorvery is needed it is will be helpful. when the source file is a text file and loading
data to a table,in such cases we should you normal load only, else the session will be
failed.
Bulk Mode: Bulk load will not write information to the database log file so that if any
recorvery is needed we can't do any thing in such cases.
compartivly Bulk load is pretty faster than normal load.
informatica server connects source data and target data using native
odbc drivers
again it connect to the repository for running sessions and retriveing metadata
information
source------>informatica server--------->target
REPOSITORY
I was working in SQL server, now i got an golden opertunity to work in
INFORMATICA. I have lots of (silly) Questions to build my Career, so pls guide
me properly. I will ask lots of questions....
In the fact table contain measurable data and less columns and meny rows,
In the dimensions table contain textual descrption of data and also contain meny
columns,less rows
Import the exported mapping in to the production repository with replace options.
When do u use a unconnected lookup and connected lookup....
what is the difference between dynamic and static lookup...y and when do v use
these types of lookups ( ie...dynamic and static )
In static lookup cache, you cache all the lookup data at the starting of the session. in
dynamic lookup cache, you go and query the database to get the lookup value for each
record which needs the lookup. static lookup cache adds to the session run time....but it
saves time as informatica does not need to connect to your databse every time it needs to
lookup. depending on how many rows in your mapping needs a lookup, you can decide
on this...also remember that static lookup eats up space...so remember to select only those
columns which are needed
With out using Updatestretagy and sessons options, how we can do the update our
target table?
In session properties, There is an option
insert
update
insert as update
update as update
like that
1. But we have data sampling. set the ata sampling properties for session in workflow
manager for specified number of rows and test the mapping.
2. Use the debugger and test the mapping for sample records.
What is hash table informatica?
In hash partitioning, the Informatica Server uses a hash function to group rows of data
among partitions. The Informatica Server groups the data based on a partition key.Use
hash partitioning when you want the Informatica Server to distribute rows to the
partitions by group. For example, you need to sort items by item ID, but you do not know
how many items have a particular ID number.
In Dimensional modeling fact table is normalized or denormalized?in case of star
schema and incase of snow flake schema?
In Dimensional modeling, Star Schema: A Single Fact table will be surrounded by a
group of Dimensional tables comprise of de- normalized data Snowflake Schema: A
Single Fact table will be surrounded by a group of Dimensional tables comprised of
normalized dataThe Star Schema (sometimes referenced as star join schema) is the
simplest data warehouse schema, consisting of a single "fact table" with a compound
primary key, with one segment for each "dimension" and with additional columns of
additive, numeric facts.The Star Schema makes multi-dimensional database (MDDB)
functionality possible using a traditional relational database. Because relational databases
are the most common data management system in organizations today,
implementing multi-dimensional views of data using a relational database is very
appealing. Even if you are using a specific MDDB solution, its sources likely are
relational databases. Another reason for using star schema is its ease of understanding.
Fact tables in star schema are mostly in third normal form (3NF), but dimensional tables
in de-normalized second normal form (2NF). If you want to normalize dimensional
tables, they look like snowflakes (see snowflake schema) and the same problems of
relational databases arise - you need complex queries and business users cannot
easily understand the meaning of data. Although query performance may be improved by
advanced DBMS technology and hardware, highly normalized tables make reporting
difficult and applications complex.The Snowflake Schema is a more complex data
warehouse model than a star schema, and is a type of star schema. It is called a snowflake
schema because the diagram of the schema resembles a snowflake.Snowflake schemas
normalize dimensions to eliminate redundancy. That is, the dimension data has been
grouped into multiple tables instead of one large table. For example, a product dimension
table in a star schema might be normalized into a products table, a Product-category
table, and a product-manufacturer table in a snowflake schema. While this saves space, it
increases the number of dimension tables and requires more foreign key joins. The result
is more complex queries and reduced query performance.
We are using Update Strategy Transformation in mapping how can we know
whether insert or update or reject or delete option has been selected during running
of sessions in Informatica.
In Designer while creating Update Strategy Transformation uncheck "forward to next
transformation". If any rejected rows are there automatically it will be updated to the
session log file.
Update or insert files are known by checking the target file or table only.
How to use mapping parameters and what is their use
In designer u will find the mapping parameters and variables options.u can assign a value
to them in designer. comming to there uses suppose u r doing incremental extractions
daily. suppose ur source system contains the day column. so every day u have to go to
that mapping and change the day so that the particular data will be extracted . if we do
that it will be like a layman's work. there comes the concept of mapping parameters and
variables. once if u assign a value to a mapping variable then it will change between
sessions.
What is difference between stored procedure transformation and external
procedure transformation?
In case of storedprocedure transformation procedure will be compiled and executed in a
relational data source.U need data base connection to import the stored procedure
in to u’r maping.Where as in external procedure transformation procedure or function
will be executed out side of data source.Ie u need to make it as a DLL to access in u r
maping.No need to have data base connection in case of external procedure
transformation.
Can u copy the session to a different folder or repository?
In addition, you can copy the workflow from the Repository manager. This will
automatically copy the mapping, associated source,targets and session to the target folder.
Discuss the advantages & Disadvantages of star & snowflake schema?
In a STAR schema there is no relation between any two dimension tables, whereas in a
SNOWFLAKE schema there is a possible relation between the dimension tables.
What is a time dimension? give an example.
In a relational data model, for normalization purposes, year lookup, quarter lookup,
month lookup, and week lookups are not merged as a single table. In a dimensional data
modeling(star schema), these tables would be merged as a single table called TIME
DIMENSION for performance and slicing data.
This dimensions helps to find the sales done on date, weekly, monthly and yearly basis.
We can have a trend analysis by comparing this year sales with the previous year or this
week sales with the previous week.
1)What are the various test procedures used to check whether the data is loaded in
the backend, performance of the mapping, and quality of the data loaded in
INFORMATICA.
2) What are the common problems developers face while ETL development
If you want to know the performance of a mapping at transformation level, then select the
option in the session properties-> collect performance data. At the run time in the monitor
you can see it in the performance tab or you can get it from a file.
The PowerCenter Server names the file session_name.perf, and stores it in the same
directory as the session log. If there is no session-specific directory for the session log,
the PowerCenter Server saves the file in the default log files directory.
Quality of the data loaded depends on the quality of data in the source. If cleansing is
required then have to perform some data cleansing operations in informatica. Final
data will always be clean if followed.
Explain about Recovering sessions?
If you stop a session or if an error causes a session to stop, refer to the session and error
logs to determine the cause of failure. Correct the errors, and then complete the
session. The method you use to complete the session depends on the properties of the
mapping, session, and Informatica Server configuration.
Use one of the following methods to complete the session:
· Run the session again if the Informatica Server has not issued a commit.
· Truncate the target tables and run the session again if the session is not recoverable.
· Consider performing recovery if the Informatica Server has issued at least one commit.
How can u recover the session in sequential batches?
If you configure a session in a sequential batch to stop on failure, you can run recovery
starting with the failed session. The Informatica Server completes the session and
then runs the rest of the batch. Use the Perform Recovery session property
To recover sessions in sequential batches configured to stop on failure:
1.In the Server Manager, open the session property sheet.
2.On the Log Files tab, select Perform Recovery, and click OK.
3.Run the session.
4.After the batch completes, open the session property sheet.
5.Clear Perform Recovery, and click OK.
If you do not clear Perform Recovery, the next time you run the session, the Informatica
Server attempts to recover the previous session.
If you do not configure a session in a sequential batch to stop on failure, and the
remaining sessions in the batch complete, recover the failed session as a standalone
session.
Which is better among connected lookup and unconnected lookup transformations
in informatica or any other ETL tool?
f you are having defined source you can use connected, source is not well defined or from
different database you can go for unconnected
We are using like that only
How u will create header and footer in target using informatica?
If you are focus is about the flat files then one can set it in file properties while creating a
mapping or at the session level in session properties
What are mapping parameters and varibles in which situation we can use it
If we need to change certain attributes of a mapping after every time the session is run, it
will be very difficult to edit the mapping and then change the attribute. So we use
mapping parameters and variables and define the values in a parameter file. Then we
could edit the parameter file to change the attribute values. This makes the process
simple.
Mapping parameter values remain constant. If we need to change the parameter value
then we need to edit the parameter file .
But value of mapping variables can be changed by using variable function. If we need to
increment the attribute value by 1 after every session run then we can use mapping
variables .
In a mapping parameter we need to manually edit the attribute value in the parameter file
after every session run.
What will happen if you are using Update Strategy Transformation and your session
is configured for "insert"?
If you have rank index for top 10. However if you pass only 5 records, what will be
the output of such a Rank Transformation?
if u r using a update strategy in any of ur mapping, then in session properties u have to set
treat source rows as Data Driven. if u select insert or udate or delete, then the info
server will not consider UPD for performing any DB operations.
ELSE
u can use the UPD session level options. instead of using a UPD in mapping just select
the update in treat source rows and update else insert option. this will do the same job as
UPD. but be sure to have a PK in the target table.
for teradata:tpump,mload.
3) if u pass only 5 rows to rank, it will rank only the 5 records based on the rank port.