Sl. No.

1 2 3

Questions Can we have multiple conditions in a Filter? How the flags are called in Update strategy? What is diff. Things u can do using PMCMD? What kind of Test plan? What kind of validation you do? What is the usage of unconnected/connected look up? What is the difference between Connected and Unconnected Lookups ? Yes. (We can place multiple conditions in one filter condition, we can not give multiple conditions in separate groups) DD_UPDATE-1,DD_INSERT-0,DD_DELETE-2,DD_REJECT-3 pmcmd command is a command line prompt to contact the informatica server. In previous versions it is called as pmrepserver. using pmcmd command you can run the mapping without using the workflow manager.

4

5

6

Lookup is used to get the related value from the source or target.Two types of lookups are there: Connected and Unconnected. If the return value is single we need to go for unconnected lookup. If we want to return multiple columns then we need to use connected lookup. Dynamic cache, user defined default values will supoort in connected lookup. Connected transformation receives input from the pipeline. We can use dynamic or static cache return multiple columns as output if there is no match for the lookup condition, the powe center server returns the default value. We can define the user defined value Unconnected transformation receives input from the result of :LKP expression in another transformation It uses static cache If there is no match for the lookup condition, the power center server returns the null value. We cannot define the user defined value

7

8

If u have data coming from For hetrogenous sources we need to use join transformation diff. sources what transformation will u use in your designer? What are different ports in Input, output,input/output,variable,Group by,Rank port,Lookup port Informatica?

9

What is a Variable port? Why it is used? Diff between Active and passive transormation ?

Variable port will use with in the transformation. It is local to the particular transformation. We can store temporary results to perform the calculations. The values will maintain for the entire mapping. Active transformation will change the number of rows that pass through it. Eg: Source Qualifier, Filter, Router, Joiner,Aggregator,Union ,Update strategy etc. Passive transformation will not change the number of rows that pass through it.Eg: Lookup, Sequence generator, Stored procedure, external procedure etc.

10

11 12

What are Mapplet? What is Aggregate transformation

13

What is Router Transformation? How is it different from Filter transformation?

Mapplet is set of transformations that can be reusable. Aggregaotor transformation allows us to perform the calculations such as averages and sums etc. The aggregator transformation unlike the expression transformation in that you can use aggregate functions and groups but difference is expression transformaion will do calculations on row by row basis. The power center server performs aggregate calculations as it reads and stores necessary data group and row data in an aggreate cache. We can improve the performance by using the sorted input. if we place the sorted input the data should come in sorted order otherise session will fail Router transformation is similar to the filter transformation because both transformations will use to test the condition. A filter transformation tests the data for one condition and drops the rows of data that do not meet conditon. Router transformation can tests data for one or more conditions and gives you the option to route rows that do not meet any of the condition to a default output group. Router transformation contains one input group , one or more output groups and one default group.

14

What are connected and Connected transformation connected to other transformation in a pipeline. unconnected transformations? Unconnected transformation is not connected to other transformation in the mapping. A connected transformation is called within another transformation and returns a value to that transformation. What is Normalizer transformation? Normalizer transformation is the process of organizing data. In database terms, this includes creating normalized tables and establishing relationships between those tables according to rules designed to both protect the data and make the database more flexible by eliminating redundancy. Normalizer transformation normalizes records from cobol and relational sources. Use normalizer transformation instead of source qualifier transformation for cobol sources. 1. We can call in source qualifer transformation (sequencename.nextval) 2. By using stored procedure transformation we can get the value of sequence

15

16

How to use a sequence created in Oracle in Informatica?

17

What are source qualifier transformations? What are cache and their types in Informatica?

18

19

What is an incremental aggregation?

When you add a relational or a flat file source definition to a mapping, you need to connect it to a Source Qualifier transformation. The Source Qualifier transformation represents the rows that the Integration Service reads when it runs a session The power center server builds a cache in memory for the rank,aggregator,joiner,lookup,sorter transformations in a mapping. It allocates memory for the cache based on the amount we defined in the transformation or session properties. Types of caches : Index and Data cache, lookup caches (Static,dynamic,shared,persistent,recache from source) Incremental aggregation is used for aggregator transformation. Once the aggregate transformation is placed in mapping, in session properties we need to check the increment aggregation property. So that the data will aggregate for incrmentallly. The first time you run an incremental aggregation session, the Integration Service processes the source. At the end of the session, the Integration Service stores the aggregated data in two cache files, the index and data cache files. The Integration Service saves the cache files in the cache file directory. The next time you run the session, the Integration Service aggregates the new rows with the cached aggregated values in the cache files. When you run a session with an incremental Aggregator transformation, the Integration Service creates a backup of the Aggregator cache files in $PMCacheDir at the beginning of a session run. The Integration Service promotes the backup cache to the initial cache at the beginning of a session recovery run. The Integration Service cannot restore the backup cache file if the session aborts

20

What is Reject loading?

By default, the Integration Service process creates a reject file for each target in the session. The reject file contains rows of data that the writer does not write to targets. The writer may reject a row in the following circumstances: It is flagged for reject by an Update Strategy or Custom transformation. It violates a database constraint such as primary key constraint. A field in the row was truncated or overflowed, and the target database is configured to reject truncated or overflowed data. By default, the Integration Service process saves the reject file in the directory entered for the service process variable $PMBadFileDir in the Workflow Manager, and names the reject file target_table_name.bad. Note: If you enable row error logging, the Integration Service process does not create a reject file.

21

WHAT IS SESSION and BATCHES?

Session - A Session Is A set of instructions that tells the Informatica Server How And When To Move Data From Sources To Targets. After creating the session, we can use either the server manager or the command line program pmcmd to start or stop the session. Batches - It Provides A Way to Group Sessions For Either Serial Or Parallel Execution By The Informatica Server. There Are Two Types Of Batches :

22

Significance of Source Qualifier Transformation

23

What are 2 modes of data movement in Informatica Server?

Sequential - Run Session One after the source Other.concurrent - Run Session you At The Same Time. it to a Source When you add a relational or a flat file definition to a mapping, need to connect Qualifier transformation. The Source Qualifier transformation represents the rows that the Integration Service reads when it runs a session. The following tasks we can do by using the SQ transformation: 1. Join data coming from the same database 2. Filter the rows when the informatica server reads the source data 3. We can use outer join instead of normal join ASCII UNICODE

24

Why we use lookup transformations?

By using lookup transformation we can do the following things: 1. Get a related value 2. Perform Calculation 3. Update slowly changing dimension tables - We can use lookup transformation to determine whether the records already exist in the target or not.

25 26

What are confirmed Confirmed dimensions can be shared by multiple facts dimensions What is Data warehousing Datawarehouse is a relational database and it is designed for query and analysis purpose. It contains the historical data derived from the transaction data and also it include data from other sources. What is a reusable transf.. What is a mapplet . Explain diff. Bet them Reusable tranformation is a single transformation it can be used in any other mappings. Mapplet is a reusable object it contains the multiple transformations. (The set of transformation logc embed in mapplet we can use this as a reusable logic in any number of mappings)

27

28

29 30

31

32

33 34

What happens when u use the delete or update or reject or insert statement in your update strategy? Where do u define users and privileges in Informatica when u run the session does debugger loads the data to target ? Can u use flat file and table (relational) as source together ? suppose I need to separate the data for delete and insert to target depending on the codition, which transformation u use ? What is the difference between lookup Data cache and Index cache. What is an indicator file and how it can be used.

It will perform the certain action what we specified in update strategy transformation and it depends on treat source rows option in session properties Repository manager it will load the data but if we select the target discard option it will not load the data.

Yes.

Router or filter

Data cache - Output columns data other than condition columns Index cache - Condition columns This is one of the output file of informatica when the session runs it will generate the indicator file.If you use a flat file as a target, you can configure the Integration Service to create an indicator file for target row type information. For each target row, the indicator file contains a number to indicate whether the row was marked for insert, update, delete, or reject. The Integration Service process names this file target_name.ind and stores it in the same directory as the target file. Filter transformation is used to filter the data based on the condition specified in filter condition value.

35

36

What is an Filter Transformation? or what options u have in Filter Transformation? What happens to the discarded rows in Filter Transformation.

The discarded rows will ignore by the informatica server

40 41 What is Mapplet and how do u create Mapplet? If data source is in the form of Excel Spread sheet then how do use? Mapplet is reusable transformation logic. Integration Service logs each row that passes into the mapping. the Integration Service writes row data for all rows in a block when it processes a transformation. The Designer helps you create source definitions. You can use information provided through the Server Manager to troubleshoot sessions and improve session performance. import the excel into source analyser . and provides detailed transformation statistics. stop the Informatica Server. 4. but not at the level of individual rows. names of index and data files used. errors encountered. and transformations to build the mappings. The Designer has tools to help you build mappings and mapplets so you can specify how to move and transform data between sources and targets. target definitions. Create data source for the driver.Install the microsoft excel odbc driver 2. Summarizes session results. 3.37 What are the two programs that communicate with the Informatica Server? Informatica provides Server Manager and pmcmd programs to communicate with the Informatica Server: Server Manager. and verify if the Informatica Server is running. pmcmd. When you configure the tracing level to verbose data. and detailed transformation statistics Verbose data: In addition to verbose initialization tracing. A command-line program that allows you to start and stop sessions and batches. 1. A client application used to create and manage sessions and batches. Integration Service logs additional initialization details. Terse:Integration Service logs initialization information and error messages and notification of rejected data Verbose initialize:In addition to normal tracing. and skipped rows due to transformation row errors. Define ranges for the excel sheet and set the datatypes for all the columns. 38 What u can do with Designer ? 39 What are different types of Tracing Levels u hv in Transformations? Normal :Integration Service logs initialization and status information. We will create mapplet in mapplet designer. and to monitor and stop the Informatica Server.

Use the Expression transformation to perform any non-aggregate calculations. nested aggregates Transformation? filter . you might need to adjust employee salaries. we can use same lookup in multiple transformations. For Transformations. Connected lookup transformation is part of mapping pipeline. Explain example. by using this we can receive multiple return values. You can also use the Expression transformation to test conditional statements before you output the results to target tables or other transformations. we can call this in expression transformation by using the :LKP qualifier. filter the records Expressions in Use the Expression transformation to calculate values in a single row before you write to the target.call the oracle procedure Source qualifier .42 When do u use connected lookup n when do u use unconnected lookup? 43 44 45 How many values it (informatica server) returns when it passes thru Connected Lookup n Unconncted Lookup? What kind of modifications u Expression . Unconnected lookup transforamtion is a separate from the flow. or convert briefly how do u use? strings to numbers.Filter the records Router . concatenate first and last names.Filter the multiple conditions Stored procedure .Find the aggregate values.performs calculations can do/perform with each Aggregator .we can override the query. Unconnected lookup will return single return port connect lookup will return one or more output values 46 In case of Flat files (which comes thru FTP as source) has not arrived then what happens We will get the fatar error and session will fail .

50 51 52 53 54 Bad file is a file that contains the rejected information. I want to join the cache performance will increase them. We can use oracle sequence 2. Create the session log file. For each master row the detail record will process. 100 records and S2 having The cache will create for the master table rows.1-update.47 48 49 What does a load manager do A component of the Integration Service that dispatches Session.Output columns data other than condition columns Index cache . Record and row indicator. Here 10000 records. The load Manager is the Primary informatica Server Process. Locks the session and read session properties. declare one variable port and increment by 1 for every row processing. What is a cache It stores the temporary results while running the session. Command.Condition columns . But I don’t want to use sequence generator transformation.S2) should be master to improve my performance? Why? I have a source and I want to generate sequence numbers using mappings in informatica.In general. Create the Data Transformation Manager which execute the session. Row indicator (0-insert.3-reject) Data cache . using joiner transformation. It Performs the following tasks Manages session and batch scheduling. in expression transformation. Reads the parameter file.2-delete. Is there any other way to do it? What is a bad file? What is the first column of the bad file? What are the contents of the cache directory in the server 1. Which of these two sources (S1. Expand the server and session variables and parameters. Master table contains less rows and detail table contains the more rows . and predefined Event-Wait ? tasks across nodes in a grid. Verify permissions and privileges. Validate source and target code pages. What is an Expression Expression transformation is passive transformation transformation? I have two sources S1 having Master should be S1.

When the Integration Service runs a session. Defines the target table or file. If the 3 target are using the nextval output port from the same transformation the value of 3 targets are same. Its like multiple of 3. Abort can be used to stops trasforming at the row. Describes the characteristics of a source table or file. Passive transformation can not change the number of rows that pass through it. and write data. If the nextval outport port coming from 3 different flows the value of 3 target will be different.55 56 Is lookup a Active transformation or Passive transformation ? What is a Mapping? Passive A mapping is a set of source and target definitions linked by transformation objects that define the rules for data transformation. Target definition. Active transformation can change the number of rows that pass through it. Transformation. 57 58 59 What are the types of transformations If a sequence generator (with increment of 1) is connected to (say) 3 targets and each target uses the NEXTVAL port. Mappings represent the data flow between sources and targets. Modifies data before writing it to targets. Links. Every mapping must contain the following components: Source definition. you use ABORT within an IIF or DECODE function to set rules for aborting a session . Connect sources. Use different transformation objects to perform different functions. it uses the instructions configured in the mapping to read. transform. Generally. targets. and transformations so the Integration Service can move the data as it transforms it. what value will each target get? Have you used the Abort. Decode functions? Active and Passive.

This is also known as buffer memory.The DTM allocates process memory for the session and divide it into buffers..Date -1/1/1753 . it creates the DTM process. Transformer 61 62 What are the default values for variables? How many ways you can filter Filter or router or source qualifier or rank or update startagy the records? Load Manager is the first process started when the session runs. creates session logs. Writer. which is called the master thread. DTM. numeric -0. the DTM creates a set of threads for each partition to allow concurrent processing. The DTM process is the second process associated with the session run.If we partition a session. It creates the main thread. creates DTM process. The primary purpose of the DTM process is to create and manage threads that carry out the session tasks. lock the session and read the session. Reader.empty. It will manage the session andbatch scheduling. The master thread creates and manages all other threads.60 What do you know about the Informatica server architecture? Load Manager. After the load manager performs validations for the session. validate the parameter value and session variables. read the parameter file. String .

You can identify target bottlenecks by configuring the session to write to a flat file target. Identifying Source Bottlenecks If the session reads from flat file source. you probably do not have a target bottleneck. 3. Use the read test mapping in a test session. If your session already writes to a flat file target. If the session performance increases significantly when you write to a flat file. you can improve the session performance by setting the number of bytes the informatica server reads per line if you read from a flat file source. 2. In the copied mapping.63 · How do you identify the bottlenecks in Mappings? We should look the performance bottelnecks in the following order: Target Source Mapping Session System Identifying Target Bottlenecks The most common performance bottleneck occurs when the Informatica Server writes to a target database. a read test mapping or database query to identify source bottelnecks. Using a Read Test Session: Use the following steps to create a read test mapping: 1. small database network packet size. or problems during heavy loading operations. you have a target bottleneck. Make a copy of the original mapping. 4. You can optimize session performance by writing to a flat file target local to the Informatica Server. you have a source bottleneck. Connect the source qualifiers to a file target. and any custom joins or queries. Causes for a target bottleneck may include small check point intervals. Remove all transformations. keep only the sources. source qualifiers. we have to use filter transformation. If the session reads from relational source. . we probably do not have a source bottelneck . If the test session performance is similar to the original session.

reduce the error records. Manually edit the definition by adding new columns/update the existing columns. we can implement partition concept at session level 2.64 How to improve the Session performance? If we do not have a source. How many ways U can update a relational source definition and what r they? 1. 65 66 What is Business components Business components component will be available in folder ? Where it exists ? What are Short cuts ? Where it is used ? By dragging from one folder to another folder we can create shortcuts. target. low buffer memory. constraints and database name source definition from database. All transformations have some basic counters that indicate the number of input rows. by optimizing the transformations 5. 67 While importing the relational Extension name. output rows. We can reimport the source definition 2. and small commit intervals can cause session bottlenecks. commit intervals 3. column names. 68 . The Integration Service creates performance details when you enable Collect Performance Data in the Performance settings on the session properties. by running concurrent sessions 4.datatypes. You can identify a session bottleneck by using the performance details. we can increase the cache size. you may have a session bottleneck. 1. what are the meta data of source U import? . Small cache size. and error rows. After dragging the object we can rename it but we can not change anything else. or mapping bottleneck. If we want to create shortcuts the folder should be shared. Performance details display information about each transformation. When ever the main object changes it will inherit all the changes to shortcut also.

XML Transformation 3. Joiner transformation 8. When u use the maping parameter .What r the unsupported repository objects for a mapplet? 70 What r the mapping parameters and mapping variables? 1. XML Targets 4. non reusable sequence generators Maping parameter represents a constant value that you can define before running a session. Yes. We can use when the same mapping parameters and variables are created in other mapping. No.U declare and use the parameter in a maping or maplet.a maping variable represents a value that can change throughout the session. 71 Can U use the mapping parameters or variables created in one mapping into another mapping? Can u use the mapping parameters or variables created in one mapping into any other reusable transformation? How can U improve session performance in aggregator transformation? 72 73 By using sorted input option . A mapping parameter retains the same value throughout the entire session. COBOL sources 5. Then define the value of parameter in a parameter file for the session.The informatica server saves the value of maping variable to the repository at the end of session run and uses that value next time U run the session. other mapplets 7. pre and post session stored procedures 8. Unlike a mapping parameter. Normalizer transformation 2. other targets 6.69 .

if the join source is coming from the Update starategy transformation. master outer.full outer to configure the joiner join condition transformation? Master and detail source check What r the join types in joiner transformation? How the informatica server sorts the string values in Rank transformation? Normal. The power center server uses the rank index port to stored the ranking postion for each row in a group. 2.74 . It is output port only.full outer If the data movement mode is ASCII format the string sort will use the binary sort order If the data movement mode is UNICODE format the string sort will use the sort order defined in session properties.What r the difference Joiner transformation used to join the hetrogenous sources between joiner transformation and source qualifier Source qualifier used to join the tables in same database transformation? 75 In which conditions we can not use joiner transformation(Limitations of joiner transformation)? 1.master outer. .detail outer. Sequence generator 76 What r the settings that u use Join type : Normal. 77 78 79 What is the Rank index in Rank transformation? the designer automatically creates rank index port for each rank transformation.detail outer.

You configure options in the Workflow Manager to continue or stop the session in the event of a stored procedure error. What r the types of data that We can send the data to stored procedure and also we can get the data from the stored procedure. 84 What is the status code? Status codes provide error handling for the Integration Service during a workflow. You cannot see this value. The stored procedure issues a status code that notifies whether or not the stored procedure completed successfully.default group (we cannot change the default group) 82 83 Why we use stored procedure Stored procedures run in either connected or unconnected mode.80 What is the Router transformation? Router transformation is similar to the filter transformation because both transformations will use to test the condition.e user defined groups . Router transformation can tests data for one or more conditions and gives you the option to route rows that do not meet any of the condition to a default output group. one or more output groups and one default group. Input group Router transformation? 2. The Integration Service uses it to determine whether to continue running the session or stop. Output groups : two types of output groups i. return value. Input/output server and stored procedure? values. passes between informatica There are three types of data that passes between informatica server and stored procedure. 81 What r the types of groups in 1. A filter transformation tests the data for one condition and drops the rows of data that do not meet conditon. . We can call the stored procedures from transformation? the database. status code. Router transformation contains one input group .

Otherwise. Join data coming from the same database 2.Inset each row flagged for update. you need to determine whether to maintain all the historic data or just the most recent changes Mapping level and session level Mapping level . Update else Insert . The following tasks we can do by using the SQ transformation: 1. We can use outer join instead of normal join 4. table . the Workflow Manager displays a warning. Update .Type2. Delete . Filter the rows when the informatica server reads the source data 3. If you do not choose Data Driven when a mapping contains an Update Strategy. The Source Qualifier transformation represents the rows that the Integration Service reads when it runs a session.we will use the update strategy transformation to flag rows for insert . the Integration Service does not follow instructions in the Update Strategy in the mapping to determine how to flag rows.data driven. insert it. As part of the target table design.update.we need to set the treat all rows property as inserts. this field is marked Data Driven by default.Select this option to truncate the target table before loading data.85 What r the tasks that source qualifier performs? 86 87 What is the default join that source qualifier provides? . you need to decide what type of information to store in targets.Type3 Mappings use for slowly changing dimension table? . What r the basic needs to join two sources in a source qualifier? what is update strategy transformation ? Describe two levels in which update strategy transformation sets? What is Data driven? When you add a relational or a flat file source definition to a mapping. Update as Insert . When you design a data warehouse. When you run the session.update. specify sorted order 5.You have the following options in this situation: Update as Update .Update the row if it exists.delete.Select this option to delete a row from a table. 88 89 90 91 What r the options in the target session of update strategy transformation? 92 Insert . For handling changes in existing rows we will go for update strategy. you need to connect it to a Source Qualifier transformation.Select this option to insert a row into a target table.delete or reject Session level . If the mapping for the session contains an Update Strategy transformation. What r the mappings that we Truncate Type1 .Update each row flagged for update if it exists in the target table. specify select distinct option normal the two sources should come from same database two sources join condition datatypes are same.

or pass-through partitioning. and load for each partition in parallel. It dispatches the task to an Integration Service process running on the node. We can check by using version or flag or effective date of the particular records. use the following session and server configuration: ♦ Configure the session with multiple partitions. it uses the instructions configured in the mapping to read. key-range. Mappings represent the data flow between sources and targets. and that informatica runs the predefined Event-Wait tasks within the workflow. the PowerCenter Server can achieve high performance by partitioning the pipeline and performing the extract. database partitioning. . To accomplish this. When the Integration Service runs a session. The PowerCenter Server can partition data using round-robin. When running sessions. or not the newly added rows in the source r gets insert in the target ? What r two types of processes Load Balancer:When you run a workflow. hash. and write data. transformation. You can configure the partition type at most transformations in the pipeline. Command. It may dispatch tasks to a single node or across nodes DTM Processes:The Integration Service starts a DTM process to run each Session and Command task Can u generate reports in Yes Informatica? What is metadata reporter? we can generate reports by using the metadata reporter. ♦ Install the PowerCenter Server on a machine with multiple CPUs. Define mapping and sessions? Mapping :A mapping is a set of source and target definitions linked by transformation objects that define the rules for data transformation. The Load Balancer matches task requirements with session? resource availability to identify the best node to run a task.93 What r the different types of Type2 dimension mapping? 94 Type2 Version Data Mapping: Type2 Flag current Mapping: Type2 Effective Date Range Mapping: How can u recognize whether 1. transform. Which tool U use to create and manage sessions and batches and to monitor and stop the informatica server? Why we use partitioning the session in informatica? To achieve the session partition what r the necessary tasks u have to do? Informatica server manager 95 96 97 98 99 100 101 To increase the session perofrmance. the Load Balancer dispatches the Session.

It creates the main thread. Create the Data Transformation Manager which execute the session. Create the workflow log file. After the load manager performs validations for the session. and predefined Event-Wait tasks across nodes in a grid. Why u use repository connectivity? When you run a session that partitions relational or Application sources. Expand the server and workflow variables and parameters.102 How the informatica server increases the session performance through partitioning the source? . Locks the workflow and reads workflow properties. schedule the session each time. The master thread creates and manages all other threads. It Performs the following tasks Manages worflow and batch scheduling. it creates the DTM process. Reads the parameter file. All the metadata of sessions and mappings will be stored in repository. This is also known as buffer memory. You can also override the SQL query for each source partition using the Transformations view on the Mapping tab. The DTM process is the second process associated with the session run. It then creates an SQL query for each partition. You can customize the query for each source partition by entering filter conditions in the Transformation view on the Mapping tab. The load Manager is the Primary informatica Server Process.If we 103 104 What r the tasks that Load manger process will do? 105 What is DTM process? . When u edit. which is called the master thread. The primary purpose of the DTM process is to create and manage threads that carry out the session tasks. Command. the Integration Service creates a separate connection to the source database for each partition.The DTM allocates process memory for the session and divide it into buffers. A component of the Integration Service that dispatches Session. Verify permissions and privileges. Validate source and target code pages. informatica server directly communicates the repository to check whether or not the session and users are valid.

In which circumstances that It is flagged informatica server creates Reject files? It violates a database constraint such as primary key constraint. 107 108 109 .One Thread Each To Perform Pre And Post Session Operations.One Thread for Each Partition If Target Exist In The Source pipeline Write To The Target. MAPPING THREAD . What r the data movement ASCII modes in informatica? UNICODE What r the out put files that the informatica server creates during the session running? The PowerCenter Server creates the following output files: PowerCenter Server log Workflow log file Session log file Session details file Performance details file Reject files Row error logs Recovery tables and files Control file Post-session email Output file Cache files for reject by an Update Strategy or Custom transformation. TRANSFORMATION THREAD . and the target database is configured to reject truncated or overflowed data.One Thread to Each Session. Pre And Post Session Thread .Main thread of the DTM process. WRITER THREAD .One Thread for Each Partition for Each Source Pipeline. A field in the row was truncated or overflowed.One or More Transformation Thread For Each Partition. READER THREAD .106 What r the different threads in MASTER THREAD DTM process? . Fetches Session and Mapping Information. Creates and manages all other threads.

concurrent.reject filenames 111 112 113 114 115 116 117 118 119 120 121 122 123 What is parameter file? Parameter file contains the mapping parameters. no any number of sessions When the session fails (the session property indicates run if the previous session successful) pmcmd command 1. Always run the session yes.session parameters. Database connections. You can also create workflow variables in the Workflow Properties and define the values in a parameter file.110 Can u copy the session to a yes different folder or repository? . Run the session if the previous session successful 2. such as database connections or source and target files. we need to create the database connection for the relational target source into Ur session? 2.mapping varaiables. by setting the always run the session property NO Yes By using pmcmd command Session parameters represent values you can change between session runs. How can u access the remote 1. Two types of batches sequential. source file names.How many number of sessions that u can create in a batch? When the informatica server marks that a batch is failed? . we need to create FTP connection for the flat files . What is batch and describe about types of batches? Can u copy the batches? .target file names. You use a session parameter in session or workflow properties and define the parameter value in a parameter file. What is a command that used to run a batch What r the different options used to configure the sequential batches? In a sequential batch can u run the session if previous session fails? Can u start batches with in a batch? Can u start a session inside a batch individually? How can u stop a batch? What r the session parameters? Grouping sessions is called a batch. Cachefile dir.

minimize aggreate function calls. If the target is flat file. it will create multiple threads to write the target. Drop indexes and key constraints. allocate buffer memory. If we create the partition on file target. remove staging areas. Use bulk loading. it will create the multiple database connections for each partition and it will run the sql query and write it in to target. run sessions and workflows concurrently. 125 Normalizer transformation. . we need to move the file to local server. increase the network packet size Optimizing the Mapping: Use Source qualifier/filter transformations to filter the data. remove unnecessary datatype conventions. increase cache size for index and data. Minimize deadlocks. Use external loading. Increase checkpoint intervals.use pushdown optimization.Tune the tansformations Optimizing theSession: Reduce error tracing. Optimize Oracle target databases. Performance tuning in Informatica? If we create the partition on relational target. use conditional filters.124 What is difference between partioning of relational target and partitioning of file targets? what r the transformations that restricts the partitioning of sessions? . Join transformation for master data 126 After finding the bottlenecks for the mapping: For optimizing the target we need to go for the following options: If it relation target. Optimizing the Source: Optimize the query. Increase database network packet size. increase commit interval.XML Target.

transformation and flow is stored as meta data in the repository.The Repository Manager connects to the repository database and runs the code needed to create the repository tables. How can u work with remote We have to create remote connection for this.127 Define informatica repository? The Informatica repository is a relational database that stores information.source definitions. and connect strings for sources and targets. 128 129 What r the types of metadata Following r the types of metadata that stores in the repository that stores in repository? Database connections Global objects Mappings Mapplets Multidimensional metadata Reusable transformations Sessions and batches Short cuts Source definitions Target defintions Transformations What is power center Power Center repository is used to store informatica's meta data . we need to database in informatica? did u import the source/target objects into local machine.client tools use. But this is not suggestable in informatica. The repository also stores administrative information such as usernames and passwords.target definitions. Use repository manager to create the repository. repository? Information such as mapping name. and product version.location.Thsea tables stores metadata in specific format the informatica server. or metadata. work directly by using remote 130 . Metadata can include information such as mappings describing how to transform source data. permissions and privileges. sessions indicating when you want the Informatica Server to perform the transformations. used by the Informatica Server and Client tools.

At the end of the session. in session properties we need to check the increment aggregation property. The Integration Service runs a scheduled workflow as configured. the Integration Service stores the aggregated data in two cache files. The next time you run the session. the Integration Service creates a backup of the Aggregator cache files in $PMCacheDir at the beginning of a session run. the index and data cache files. and skipped r the types of tracing level? rows due to transformation row errors. The Integration Service cannot restore the backup cache file if the session aborts 132 . but not at the level of individual rows. Summarizes session results. Once the aggregate transformation is placed in mapping. the Integration Service aggregates the new rows with the cached aggregated values in the cache files.131 what is incremental aggregation? Incremental aggregation is used for aggregator transformation. The Integration Service promotes the backup cache to the initial cache at the beginning of a session recovery run. and detailed transformation statistics Verbose data: In addition to verbose initialization tracing. Run on demand . the Integration Service processes the source. Integration Service logs additional initialization details. When you run a session with an incremental Aggregator transformation. By default. errors encountered. 133 Scheduling options: Run on server initialization . the Integration Service reschedules the workflow according to the new settings. What r the scheduling options to run a session? You can schedule a workflow to run continuously. repeat at a given time or interval. Integration Service logs each row that passes . the workflow runs on demand. The Integration Service saves the cache files in the cache file directory. Terse:Integration Service logs initialization information and error messages and notification of rejected data Verbose initialize:In addition to normal tracing. You can change the schedule settings by editing the scheduler.the integaration service run othe workflow when we start the workflow manually What is tracing level and what Normal :Integration Service logs initialization and status information. The first time you run an incremental aggregation session. So that the data will aggregate for incrmentallly.the integration service runs the workflow as soon as the service is initialized. If you change schedule settings. or you can manually start a workflow. names of index and data files used.

If a session fails after loading Consider By using Perform recovery option setting in session properties of 10. and then complete the session. the Informatica Server bypasses the rows up to 10. performing recovery if the Informatica Server has issued at least one commit. it reads the OPB_SRVR_RECOVERY table and recovery? notes the row ID of the last row committed to the target database. session. Perform Recovery is disabled in the Informatica Server setup. Truncate the target tables and run the session again if the session is not recoverable.Where as in external procedure transformation procedure or function will be executed out side of data source. You must enable Recovery in the Informatica Server setup before you run a session so the Informatica Server can create and/or write entries in the OPB_SRVR_RECOVERY table . when you run recovery. 135 136 137 Use one of the following methods to complete the session: Run the session again if the Informatica Server has not issued a commit. Correct the errors. The method you use to complete the session depends on the properties of the mapping.000 records in to the target.No need to have data base connection in case of external procedure transformation.Iet we need to make it as a DLL to access in maping. refer to the session and error logs to determine the cause of failure. If you stop a session or if an error causes a session to stop. How can u load the records from 10001st record when u run the session next time? Explain about perform When the Informatica Server starts a recovery session. if the Informatica Server commits 10.000 and starts loading with row 10.You need data base connection to import the stored procedure in to maping. The Informatica Server then reads all sources again and starts processing from the next row ID.001.134 What is difference between stored procedure transformation and external procedure transformation? Explain about Recovering sessions? In case of storedprocedure transformation procedure will be compiled and executed in a relational data source. By default.000 rows before the session fails. For example. and Informatica Server configuration.

These options are not available for batched sessions.Clear Perform Recovery. 3. the next time you run the session. and the remaining sessions in the batch complete. The Informatica Server completes the session and then runs the rest of the batch. 2. If you do not clear Perform Recovery. 2. 139 How can u recover the session in sequential batches? . If a standalone session fails. Use the Perform Recovery session property To recover sessions in sequential batches configured to stop on failure: 1. you can run recovery starting with the failed session. From the command line. and click OK. To recover sessions using the menu: 1. open the session property sheet.After the batch completes. With the failed session highlighted. If you configure a session in a sequential batch to stop on failure. 5.From the command line. In the Server Manager. recover the failed session as a standalone session.138 How to recover the standalone session? A standalone session is a session that is not nested in a batch. If you do not configure a session in a sequential batch to stop on failure. Select Server Requests-Stop from the menu. select Server Requests-Start Session in Recovery Mode from the menu. 2.In the Server Manager. open the session property sheet.Run the session. highlight the session you want to recover. start recovery. select Perform Recovery. stop the session. and click OK.On the Log Files tab. you can run recovery using a menu command or pmcmd. 3. To recover sessions using pmcmd: 1. 4. the Informatica Server attempts to recover the previous session.

What r the circumstances that The source qualifier transformation does not use sorted ports. Run the session from the beginning when the Informatica Server cannot run recovery or when running recovery might result in inconsistent data. If i've done any modifications It will not reflect automatically. you need to truncate the target tables and run the session from the beginning. If the sources or targets changes after initial session fails. you might want to truncate all targets and run the batch again.Delete the standalone copy. three sources(sql server. If a concuurent batch contains multiple failed sessions. it reflect in informatica warehouse or mapping designer or source analyzer? After dragging the ports of If we join the three sources into source qualifier then we can map the ports to target.140 How to recover sessions in concurrent batches? If multiple sessions in a concurrent batch fail.oracle.$PMWorkflowLogDir 143 144 145 . However.$PMBadFileDir. 141 How can u complete unrecoverable sessions? 142 Under certain circumstances. 4. you can recover the session as a standalone session.informix) to a single source qualifier.$PMSourceFileDir. informatica server results an If u change the partition information after the initial session fails. when a session does not complete. If the maping consists of sequence generator or normalizer transformation. 2. To recover a session in a concurrent batch: 1.Copy the failed session using Operations-Copy Session. 3.$PMTargetFileDir. we need to reimport the objects or manually we need to add the for my table in back end does changes.$PMSessionLogDir. can u map these three ports directly to target? Server Variables $PMCacheDir.Follow the steps to recover a standalone session. unrecoverable session? Perform recovery is disabled in the informatica server configuration. if a session in a concurrent batch fails and the rest of the sessions complete successfully.Drag the copied session outside the batch to be a standalone session.

to help you logically organize the repository. you can create a folder that allows users to share objects within the folder. but not to edit them. Each folder has a set of configurable properties that help you define how users access the folder. Folders are designed to be flexible. you can create a folder that allows all repository users to see objects within the folder. schemas.146 Folders 147 Multiple Servers Folders provide a way to organize and store all metadata in the repository. including mappings. For example. You can create shared and non-shared folders. and sessions. Or. .

.

.

.

.

.

.

.

.

.

.

.

.

.

.

No. Questions 1 what is a Data Warehouse? What is the conventional definition of a DWH? Explain 2 each term. What are the goals of the Data 4 warehouse? 5 What are the approaches in constructing a Datawarehouse and the datamart? Data Mart 6 7 Can a datamart be independent? . 3 Draw the architecture of a Datawarehousing system.Sl.

What are the sources for a 8 datawarehouse? 9 What the difference is between a database.DOLAP? . MOLAP . a data warehouse and a data mart? OLAP (On-Line Analytical Processing) 10 11 what do you mean by Multidimensional Analysis? What is the difference between 12 OLAP. ROLAP.

15 Which is the suitable data model for a datawarehouse? Why? Star Schema 16 17 What are the benefits of STAR SCHEMA? What are Additive Facts? Or what 18 is meant by Additive Fact? 19 Snowflake Schema 20 What is Galaxy schema? .13 Difference between OLAP & OLTP? What are the different types of 14 OLAP? Give an eg.

of the fact.21 What is Dimension & Fact ? 22 Different types of Dimensions 23 Are the dimensional tables normalized? If so when? What is Transaction fact table & 24 Centipede Fact table? 25 Different types of Facts? What are the types of Factless 26 fact tables? What is Granularity? 28 Is the Fact table normalized? 29 Can 2 Fact Tables share same dimensions Tables? Give egs. dimensional tables. 27 . Explain what data 30 each contains. datamarts/DWH used in your project.

DSS (Decision support System). 37 What are surrogate keys? Name a few data modelling 38 tools. Materialized views? Can you insert into materialized 40 views? 41 39 Definition of Adhoc Queries? What is ODS (Operational Data Store). 42 Data Presentation Area. Data Staging Area.31 What are fact constellations? What is a Fact less fact table 32 ? 33 What is metadata? What is data quality? 34 35 How do you achieve data quality? 36 What is Parsing & Data Mining. 43 What is Market-Basket analysis? .

SCD Types 44 46 45 what is a Hypercube? 47 Explain the performance improvement techniques in DW? 48 Explain slice and dice ? .

2. make sure the data is accurate and can be trusted. characterstics of datawarehouse: Subject-oriented The data in the data warehouse is organized so that all the data elements relating to the same real-world event or object are linked together.Datawarehouse is a relational database and it is designed for query and analysis purpose. labeling it consistently across the enterprise 5. actionable subset of the OLTP data to present in the data warehouse. Data marts contain atomic data and.once committed. 4. Time-variant The changes to the data in the data warehouse are tracked and recorded so that reports can be produced showing changes over time. job responsibilities. that may be distributed to support business needs. It contains the historical data derived from the transaction data and also it include data from other sources. though in the bottom-up data warehouse design methodology the data warehouse is created from the union of organizational data marts. read-only. The main goals of our Data Warehouse are: 1. Integrated The data warehouse contains data from most or all of an organization's operational systems and this data is made consistent. continuously monitor the accuracy of the data and the content of the delivered reports 6. understand the users needs by business area. So it can be independent. usually oriented to a specific purpose or major data subject. choose the most effective. These data marts can eventually be unioned together to create a comprehensive data warehouse A data mart is a subset of an organizational data store. publish the data on a regular basis Top down approach Bottom up approach: data marts are first created to provide reporting and analytical capabilities for specific business processes. the data is static. and retained for future reporting. summarized data. Non-volatile Data in the data warehouse is never over-written or deleted . if necessary.[1] Data marts are analytical data stores designed to focus on specific business functions for a specific community within an organization. Datamarts can use for small organizations. Data marts are often derived from subsets of data in a data warehouse. . determine the decisions the business users want to make with the help of the data warehouse 3.

1.Online Analytical processing. Partially normalized/denormalized tables 4. Online Analytical 2. stored in a Relational database(RDBMS).( Online Transaction processing). Read only data 3. Processing Less joins to retrieve data 1. HOLAP . records are based on surrogate field 6. Normalized Form 3. cannot delete the records 7.Relational Online Analytical Process that provides multidimensional analysis of data. .Hybrid OLAP a combination of both ROLAP and MOLAP can provide multidimensional analysis simultaneously of data stored in a multidimensional database and in a relational database(RDBMS). Complex joins 4. query and analytical purpose 2.OLTP and operations sources Database . Holds current data Datawarehouse -1. provides the analysis of data stored in a multi-dimensional data cube. It is collection of data. It is normalized data. simplified datamodel multidimensional analysis is a data analysis process that groups data into two basic categories: data dimensions and measurements OLAP . MOLAP -Multidimensional OLAP. To retrieve data complex joins has to perform. More DML operations 5. DOLAP . partially normalized/denormalized form 3. 2. ROLAP . It is relational database .Desktop OLAP or Database OLAP. provide multidimensional analysis locally in the client machine on the data collected from relational or multidimensional database servers. holds current and historical data 5.

. The main difference is the dimensional tables in a snowflake schema are normalized. This schema is more complex than start and snowflake schema. Hybrid OLAP (HOLAP) refers to technologies that combine MOLAP and ROLAP Star schema. Star schema is the simplest datawarehouse design. Online Transaction processing 2. holds current data 5. These are used when the dimension table becomes very big and star schema can't represent the complexity of data structure. Fact tables tables are in a star schema are in database are 3rd normal form. Queries results fastly.OLAP 1. Dimensions are denormalized form It is a demoralized model. The main feature of schema is a table at the center called fact table and surrounding dimension tables. No need to use complicated joins. It is very hard to manage. continuosuly updates data 3. It is a demoralized model. which is because it contains multiple fact tables. so they have a typical relational database design. cannot delete the records 7. records are based on surrogate field 6. there are mainly two different types: Multidimensional OLAP (MOLAP) and Relational OLAP (ROLAP). Online Analytical Processing 2. simplified datamodel OLTP 1. Additive facts are fact that can be summed up through all of the dimensions in the fact table Snowflake schema is more complex variation of a star schema design. Queries results fastly. complex data model In the OLAP. No need to use complicated joins. Partially normalized/denormalized tables 4. Read only data 3. This allows dimension tables to be shared amongest many fact tables. records are maintained on primary key field 6. normalized form 4. holds current and historical data 5. delete the table or records 7.

In a schema less number of fact than dimension table. Additive Fact Semi additive fact Non Additive fact A fact table which doesn't contains any facts then called as fact less fact table. values of columns are always in numeric. then dimensions are normalized form. cluttering your database with these mini ‘identifier’ tables. instrument_dim are dimensions . In a schema more number of dimensions are presented than fact table. date_dim. By yes. that will have few records in them. The first type of factless fact table is a table that records an event. For ex Time dimension.1.The dimension used by multiple facts. values of columns are in numerica nd test representation 1.Date dimension Degenerated Dimension Slowly Changing Dimensions If the schema is snowflake . Fact contains measurements. Size of dimension table is smaller than fact. Fact table will be in 3rd normal form Yes Rating identity instr fact table is fact table. size of fact table is larger than dimension table. Junk Dimension -When you consolidate lots of small dimensions and instead of having 100s of small dimensions. all_org_dim. all records from all these small dimension tables are loaded into ONE dimension table and we call this dimension table Junk dimension table Confirmed Dimension . Surrogate key is used to prevent the primary key violation. Many event-tracking tables in dimensional data warehouses turn out to be factless A second kind of factless fact table is called a coverage table Designing a fact table is to determine the granularity of the fact table. Dimension contains the descriptive information for a fact table.

DSS applications are systems and subsystems that help people make decisions based on data that is pulled from a wide range of sources. It is always numeric data type. but behaves like a view. This data used for analytical and reporting purpose. Query that is not predefined or anticipated.For each star schema it is possible to construct fact constellation schema(for example by splitting the original star schema into more star schemes each of them describes facts on another level of dimension hierarchies).RATIONAL ROSE A materialized view is a table that actually contains rows. particularly in a data warehouse. Data quality assurance (DQA) is the process of verifying the reliability and effectiveness of data. Data from ODs is staged. Data Staging Area: The Data Warehouse Staging Area is temporary location where data from source systems is copied . The fact constellation architecture contains multiple fact tables that share many dimension tables. typically for business purposes. we will not insert records into materialized views but certain query refresh will be there in certain period of times then materialized view will get the data from the source tables. Moreover. the data in the table changes when the data in the underlying tables changes it will done in periodical basis. transformed and then moved to datawarehouse. To achieve good quality information a database must have good system design accurate and complete user friendly interface data validation Surroage keys are sequence generated numbers. ERWIN. The main shortcoming of the fact constellation schema is a more complicated design because many variants for particular kinds of aggregation must be considered and selected. dimension tables are still large A fact table which doesn't contains any facts then called as fact less fact table. These are typical in the dataware housing environment. Maintaining data quality requires going through the data periodically and scrubbing it. DSS:Gathers and presents data from a wide range of sources. ODS:ODS is the Operational Data Source which is also called transactional data ODS is the source of a warehouse. metadata is information about data Data quality is the reliability and effectiveness of data. usually just runs once are called Ad-hoc queries. That is. No.

the new information overrides the current information. By using partitions. No history kept. query tuning. we need to find the bottleneck of the mapping then we need to proceed to improve the performance of the warehouse. TYPE2 SCD. .Three types of Slowly Changing Dimension. TYPE3 SCD: This will maintain the partial history. the new record is added to the table. TYPE1 SCD . indexing. History will avialable.