Professional Documents
Culture Documents
By PenchalaRaju.Yanamala
don't know
Latest Answer: We can say "Surrogate key" is a User defined primary key..
Data Mart is a segment of a data warehouse that can provide data for reporting
and analysis on a section, unit, department or operation in the company, e.g.
sales, payroll, production. Data marts are sometimes
Latest Answer: A Data Mart is the subset of the data warehouse that caters the
needs of a specific functional domain.examples of functional domains can be
given as Sales, Finance, Maketing, HR etc. ...
Latest Answer: There are four methods in which one can build a
datawarehouse.1. Top-Down (Emphasizes the DW. )2. Bottom-Up (Emphasizes
data marts.)3. Hybrid (Emphasizes DW and data marts; blends “top-down” and
“bottom-up” methods.)4. Federated (Emphasizes the need to ...
Latest Answer: By far, the best ETL tool on the market is Hummingbird
Genio.Hummingbird is a division of OpenText, they make, among other things,
connectivity and ETL software. ...
What is ODS
Latest Answer: ODS means Operational Data store. ODS & Staging layer are the
two layers between the source and the target datbases in the data
warehouse..ODS is used to store the recent data. ...
Latest Answer: A dimension which can be shared with multiple fact tables such
dimensions are know as conformed dimension. ...
What is ER Diagram
Answered by Puneet on 2005-05-07 04:21:07: ER - Stands for entitity
relationship diagrams. It is the first step in the design of data model which will
later lead to a physical database design of possible
Latest Answer: Entity Relationship Diagrams are a major data modelling tool and
will help organize the data in your project into entities and define the relationships
There are three basic elements in ER models: Entities are the "things" about
which we seek ...
What is ETL
Latest Answer: E-R Modeling is a model for OLTP, optimized for Operational
database, namely insert, update, delete data and stressing on data relational
integrity.Dimensional Modeling is a model for OLAP, optimized for retrieving data
because it's uncommon to update ...
Latest Answer: The Fact table is central table in Star schema, Fact table is kept
Normalized because its very bigger and so we should avoid redundant data in it.
Thats why we make different dimensions there by making normalized star
schema model which helps in query ...
Latest Answer: The definition of slowly changing dimension is in its name only.
The dimension which changes slowly with time. A customer dimension table
represents customer. When creating a customer, normal assumption is it is
independent of time. But what if address ...
Difference between Snow flake and Star Schema. What are situations where
Snow flake Schema is better
Difference between Snow flake and Star Schema. What are situations where
Snow flake Schema is better than Star Schema to use and when the opposite is
true?
A cube can be stored on a single analysis server and then defined as a linked
cube on other Analysis servers. End users connected to any of these analysis
servers can then access the cube. This arrangement
Latest Answer: Hi All,Could you please let me know what is Replicate Cube &
Transparent Cube?Thanks & regards,Amit Sagpariya ...
For 80GB Datawarehouse How many records are there in Fact Table There are
25 Dimension and 12 Fact
For 80GB Datawarehouse How many records are there in Fact Table There are
25 Dimension and 12 Fact Tables
How data in datawarehouse stored after data has been extracted and
transformed from hetrogeneous sources
How data in datawarehouse stored after data has been extracted and
transformed from hetrogeneous sources and where does the data go from
datawarehouse.
What is the role of surrogate keys in data warehouse and how will u generate
them?
Latest Answer: A surrogate key is a substitution for the natural primary key. We
tend to use our own Primary keys (surrogate keys) rather than depend on the
primary key that is available in the source system. When integrating the data,
trying to work with ...
What is Normalization, First Normal Form, Second Normal Form , Third Normal
Form
Latest Answer: Non additive facts are the facts that do not participate in
arithmetic caliculations. for example in stock fact table there will be opening and
closing balances along with qty sold and amt etc. but opening and closing
balances were never used in arithmetic ...
What is VLDB
Answered by Kiran on 2005-05-06 20:12:19: The perception of what constitutes a
VLDB continues to grow. A one terabyte database would normally be considered
to be a VLDB.
Latest Answer: Very Large Database (VLDB)it is sometimes used to describe
databases occupying magnetic storage in the terabyte range and containing
billions of table rows. Typically, these are decision support systems or
transaction processing applications serving large ...
Latest Answer: SCD1, SCD2, SCD3 are also called TYPE1, TYPE2, TYPE3
dimensions Type1: It never maintains history in the target table. It keeps the most
recent updated record in the data base. Type2: It maintains full history in the
target. It maintains history by ...
Why are OLTP database designs not generally a good idea for a Data
Warehouse
-
Latest Answer: meta data is stored in repository only not in dataware house .. but
we r placing our repository in database in that way ur correct ,,but not directly
stored in the dataware house plz check it mam ...
Wht r the data types present in bo?n wht happens if we implement view in the
designer n report
What is the definition of normalized and denormalized view and what are the
differences between them
What is the definition of normalized and denormalized view and what are the
differences between them
What is the main difference between Inmon and Kimball philosophies of data
warehousing?
What is the need of surrogate key;why primary key not used as surrogate key
Latest Answer: Datawarehousing depends on the surrogate key not primary key,
for suppose if u r taking the product price it will change over the time, but product
no. will not change but price will change over the time to maintain the full
hystorical data ...
How do you connect two fact tables ? Is it possible ?
Latest Answer: The only way to connect two fact tables is by using conformed
dimension. ...
Explain the flow of data starting with OLTP to OLAP including staging
,summary tables,Facts and dimensions.
Explain the flow of data starting with OLTP to OLAP including staging ,summary
tables,Facts and dimensions.
Latest Answer: There is one more data modelling tool available in the market and
that is "KALIDO".This is end to end data warehousing tool. Its a unique and
user friendly tool. ...
What are Semi-additive and factless facts and in which scenario will you use
such kinds of fact tables
What are Semi-additive and factless facts and in which scenario will you use
such kinds of fact tables
Latest Answer: Hybrid SCDs are combination of both SCDÂ 2 and SCD
3.Whatever changes done in source for each and every record there is a new
entry in target side, whether it may be UPDATE or INSERT. There is new
column added to provide the previous record info (generally ...
What is the difference between star schema and snow flake schema ?and
when we use those schema's?
What is the difference between star schema and snow flake schema ?and when
we use those schema's?
Can you convert a snowflake schema in to star schema?
Latest Answer: Star ----->Snow Flake also vice versa is possibleIn Star
SchemaWhen we try to access many attributes or few attributes from a single
dimension table the performance of the query falls. So we denormalize this
dimension table into two or sub dimensions. ...
Latest Answer: A snowflake schema is a way to handle problems that do not fit
within the star schema. It consists of outrigger tables which relate to dimensions
rather than to the fact table.The amount of space taken up by dimensions is so
small compared to the ...
Which columns go to the fact table and which columns go the dimension table
Why should you put your data warehouse on a different system than your
OLTP system
Latest Answer: An DW is typically used most often for intensive querying . Since
the primary responsibility of an OLTP system is to faithfully record on going
transactions (inserts/updates/deletes), these operations will be considerably
slowed down by the heavy querying ...
the ODS is technically designed to be used as the feeder for the DW and other
DM's -- yes. It is to be the source of truth.Read the complete thread at
http://asktom.oracle.com/pls/ask/f?
p=4950:8:16165205144590546310::NO::F4950_P8_DISPLAYID,F4950_P8_CRI
TERIA:30801968442845,
Latest Answer: Hi According to Bill Inmon's paradigm an enterprize can have one
datware house and datamarts source their information from the datawarehouse.
In the dataware house, information is stroed in 3rd Normalization. This Dataware
house is build on ODS. You ...
Latest Answer: Source qualifier is a transformation which extracts data from the
source. Source qualifier acts as SQL query when the source is a relational
database and it acts as a data interpreter if the source is a flatfile. ...
Latest Answer: It also means that we can have (for example) data agregated for
a year for a given product as well as the data can be drilled down to Monthly,
weekl and daily basis...teh lowest level is known as the grain. going down to
details is Granularity ...
Latest Answer: Find where data for this dimension are located. Figure out how to
extract this data. Determine how to maintain changes to this dimension (see
more on this in the next section). Change fact table and DW population routines.
...
Latest Answer: ODS: this is operational data stores, which means the real time
transactional databases. In data warehouse, we extract the data from ODS,
transform in the stagging area and load into the target data warehouse.I think,
earlier comments on the ODS is little ...
What is the purpose of "Factless Fact Table"? How it is involved in Many to many
relationship?
Explain the advanatages of RAID 1, 1/0, and 5. What type of RAID setup would
you put your TX logs
Latest Answer: Raid 0 - Make several physical hard drives look like one hard
drive. No redundancy but very fast. May use for temporary spaces where loss of
the files will not result in loss of committed data. Raid 1- Mirroring. Each hard
drive in the ...
Latest Answer: Hi, Slicing and Dicing is a feature that helps us in seeing the
more detailed information about a particular thing. For eg: You have a report
which shows the quarterly based performance of a particular product. But you
want to see it ...
What is data validation strategies for data mart validation after loading process
Latest Answer: Data validation is to make sure that the loaded data is accurate
and meets the business requriments.Strategies are different methods followed to
meet the validation requriments ...
Latest Answer: No Tool testing in done in DWH, only manual testing is done.
Latest Answer: Data Mining is used for the estimation of future. For example, if
we take a company/business organization, by using the concept of Data Mining,
we can predict the future of business interms of Revenue (or) Employees (or)
Cutomers (or) Orders ...
Latest Answer: static cache stores overloaded values in the memory and it wont
change throught the running of the session where as dynamic cache stores the
values in the memory and changes dynamically duirng the running of the session
used in scd types -- where target ...
What is cube and why we are crating a cube what is diff between etl and olap
cubes any budy ans
What is cube and why we are crating a cube what is diff between etl and olap
cubes any budy ans plz?
What are the various attributes in time dimension, If this dimension has to
consider only date of birth
What are the various attributes in time dimension, If this dimension has to
consider only date of birth of a citizen of a country?
What are late arriving Facts and late arriving dim ? How does it impacts DW?
Latest Answer: ER modelling is the first step for any Database project like
Oracle, DB2.1. Conceptual Modelling2. Logical Modelling3. Physical Modelling ...
Bill Inmon vs Ralph Kimball In the data warehousing field, we often hear about
discussions on where a person / organization's philosophy falls into Bill Inmon's
camp or into Ralph Kimball's
Latest Answer: Bill inmon : Data warehouse à Data martRalph Kimbol : Data
mart à Data warehouseCheers,Sithu, sithusithu@Hotmail.com ...
I want to know how to protect my data over networ.which software will be use
Data warehouse is relational database used for query analysis and reporting. By
definition data warehouse is Subject-oriented, Integrated, Non-volatile, Time
variant.
Data warehouse is maintaining the total organization of data. Multiple data marts
used in data warehouse. where as data mart is maintained only particular subject.
5) Explain ODS?
Operational data store is a part of data warehouse. This is maintained only current
transactional data. ODS is subject oriented, integrated, volatile, current data.
Power center receive all product functionality including ability to multiple register
servers and metadata across the repository and partition data.
One repository multiple informatica servers. Power mart received all features except
multiple register servers and partition data.
Staging area is a temporary storage area used for transaction, integrated and rather
than transaction processing.
When ever your data put in data warehouse you need to clean and process your data.
Star Schema consists of one or more fact table and one or more dimension tables
that are related to foreign keys.
Dimension tables are De-normalized, Fact table-normalized
If both data marts use same type of dimension that is called confirm
dimension.If you have same type of dimension can be used in multiple fact that
is called confirm dimension.
Type2: The Type2 Dimension data mapping inserts both new and changed
dimensions into the target.Changes are tracked in the target table by versioning
the primary key and creating a version number for each dimension in the table.
Type3: The type 3 dimension mapping filters source rows based on user-
defined comparisions and inserts only those found to be new dimensions to the
target.Rows containing changes to existing dimensions are updated in the
target. When updating an existing dimension the informatica server saves
existing data in different columns of the same row and replaces the existing
data with the updates.
Your target table is also look up table then you go for dynamic cache .In
dynamic cache multiple matches return an error.use only = operator.
Override the default SQL statement.You can join multiple sources use lookup
override.By default informatica server add the order by clause.
You specify the target load order based on source qualifiers in a mapping.if u
have the multiple source qualifiers connected to the multiple targets you can
designate the order in which informatica server loads data into the targets.
23) what are the difference between joiner transformation and source qualifier
transformation?
Two relational sources should come from same data source in source
qualifier.You can join relational sources, which are coming from different
sources in source qualifier.You can join relational sources which are coming
from different sources also.
Whenever you create the target table whether you are store the historical data
or current transaction data in to target table.
Data driven.
28) what are the options in the trarget session of update strategy transformation?
Insert
Delete
Update
Update as update
Update as insert
Truncate table.
Source filter is filtering the data only relational sources. Where as filter
transformation filter the data any type of source.
-- can you connect multiple ports from one group to multiple transformations?
Yes
31) can you connect more than one group to the same target or transformation?
NO
Two methods
When you use the mapping parameter , you declare and use the
parameter in a mapping or mapplet.Then define the value of parameter in
a parameter file for the session.
35)Can you are the mapping parameters or variables created in one mapping
into any other result transformation.
36)How the informatica server sorts the string values in rank transformation?
When the informatica server runs in the ASCII data movement mode it
sorts session data using binary sort order.If you configures the session to
use a binary sort order, the informatica server calculates the binary value
of each string and returns the specified number of rows with the highest
binary values for the string.
WORKFLOW MANAGER
41)what is a server?
The power center server moves data from source to targets based on a
workflow and mapping metadata stored in a repository.
-- what is session?
A session is a set of instructions that describes how to move data from source to
target using a mapping.
Use the work flow monitor work flows and stop the power center server.
The power center server uses both process memory and system shared
memory to perform these tasks.
Load manager process: stores and locks the workflow tasks and start the
DTM run the sessions.
Mapping thread.
Transformation thread.
Reader thread.
Writer thread.
1) Task developer.
3) Worklet designer.
You can sehedule a work flow to run continuously, repeat at given time or
interval or you manually start a work flow.By default the workflow runs on
demand.
47)Explain stopping or aborting a session task?
If the power center is executing a session task when you issue the stop
the command the power center stop reading data. If continuous
processing and writing data and committing data to targets.
If the power center can’t finish processing and committing data you issue
the abort command.
You can also abort a session by using the Abort() function in the mapping
logic.
48)What is a worklet?
A worklet is an object that represents a set of taske.It can contain any task
available in the work flow manager. You can run worklets inside a
workflow. You can also nest a worklet in another worklet.The worklet
manager does not provide a parameter file for worklets.
The power center server writes information about worklet execution in the
workflow log.
Target Based commit: The power center server commits data based on
the number of target rows and the key constraints on the target table. The
commit point also depends on the buffer block size and the commit
interval.
Source-based commit:---------------------------------------------
User-defined commit:----------------------------------------------
You can use bulk loading to improve performance of a session that inserts
a large amount of data to a db2,sysbase,oracle or MS SQL server
database.
When bulk loading the power center server by passes the database
log,which speeds performance.
With out writing to the database log, however the target database can’t
perform rollback.As a result you may not be perform recovery.
When you select this option the power center server orders the target load
on a row-by-row basis only.
Edit tasks->properties->select treat source rows as insert.
Use the constraint based loading only when the session option treat rows
as set to insert.
You can capture new source data.use incremental aggregation when you
can capture new source data much time you run the session.Use a stored
procedure on filter transformation only new data.
The first time u run an incremental aggregation session the power center
server process the entire source.At the end of the session the power
center server stores aggregate data from the session runs in two files, the
index file and the data file .The power center server creates the files in a
local directory.
Transformations.
1) active
2) passive.
Active transformation can change the number of rows that pass through it.No
of output rows less than or equal to no of input rows.
Filter transformation to filter the data only one condition and drop the rows
don’t meet the condition.
Drop rows does not store any ware like session log file..
Router transformation to filter the data based on multiple conditions and give
yiou the option to route rows that don’t match to a default group.
Expression transformation calculate the single row values before writes the
target.Expression transformation executed by row-by-row basis only.
Normal
Master outer
Detail outer
Full outer
Both input pipelines originate from the same source qualifier transformation.
63)what are the settings that u use to configure the joiner transformation?
Type of join
Receives input values directly Receives input values from the result of
from the pipe line. a clkp expression in a another
transformation.
Cache
The informatica server stores conditions values in the index cache and
output values in the data cache.
Persistent cache: U can save the look up cache files and reuse them the next
time the informatica server processes a lookup transformation to use the
cache.
Dynamic cache: If you want to cache the target table and insert new rows into
cache and the target you can create a look up transformation to use dynamic
cache.The informatica server dynamically inserts data into the target table.
Shared cache: You can share the lookup cache between multiple
transformations.You can share unnamed cache between transformation in the
same mapping.
You cannot insert or update You can insert rows into the cache
the cache as you pass rows to the target
ORACLE:
The set of redo log files for a database is collectively know as the databases redo
log.
A data base is divided into logical storage unit called table space.A table space is
used to grouped related logical structures together.
Self-join,equi-join,outer join.
One of which rows that don’t match those in the commen column of another table.
Select * from emp e where 5>(select count(*) from emp where sal>e.sal)
80) --------------------------------
81)
Select distinct (a.sal) from emp a where &n=select count(distinct(b.sal) from emp b
where a.sal<=b.sal
4. Briefly explain the difference between first ,second ,third and fourth
normal forms?
First Normal form : Attribute should be atomic.
Second Normal Form : Non-Key attribute should be fully functionally dependent
on key Attribute.
Third normal Form : There is no transitivity dependency between attribute.
Suppose 'y' is dependent on 'x' i.e. x->y and 'z' is dependent on 'y' i.e. y->z this
is transitivity dependency So we can split table on to two tables os that result will
be x->z.
Forth Normal Form : A determinant is any attribute (simple or composite) on
which some other attribute is fully functionally dependent.
A relation is in BCNF is, and only if, every determinant is a candidate key.
6. There is a eno & gender in a table. Eno has primary key and gender has a
check constraints for the values 'M' and 'F'.
While inserting the data into the table M was misspelled as F and F as M.
What is the update statement to replace F with M and M with F?
CREATE TABLE temp(
eno NUMBER CONSTRAINTS pk_eno PRIMARY KEY,
gender CHAR(1) CHECK (gender IN( 'M','F')));
COMMIT;
UPDATE temp SET gender =DECODE(gender,'M','F','F','M');
Commit;
8. How to find out the database name from SQL*PLUS command prompt?
SELECT INSTANCE_NAME FROM V$INSTANCE;
SELECT * FROM V$DATABASE;
SELECT * FROM GLOBAL_NAME;
9. What is Normalization?
Normalization is the process of removing redundant data from your tables in
order to improve storage efficiency, data integrity and scalability.
11. What is the difference between Single row sub-Query and Scalar sub-
Query
Single row sub-queries returns only one row of results. A single row sub query
uses a single row operator; the common operator is the equality operator(=).
A Scalar sub-query returns exactly one column value from one row. Scalar sub-
queris can be used in most places where you would use a column name or
expression, such as inside a single row function as an argument, in insert, order
by clause, where clause, case expressions but not in group by or having clause.
12. TRUNCATE TABLE EMP; DELETE FROM EMP; Will the outputs of the
above two commands
Delete Command:
1. It’s a DML Command
2. Data can be rolled back.
3. Its slower than Truncate command b’coz it logs each row deletion.
4. With delete command trigger can be fire.
Truncate Command:
1. It’s a DDL Command
2. Data Can not be rolled back.
3. Its is faster than delete b’coz it does not log rows.
With Truncate command trigger can not be fire.
both cases only the table data is removed, not the table structure.
13. What is the use of the DROP option in the ALTER TABLE command
Drop option in the ALTER TABLE command is used to drop columns you no
longer need from the table.
The column may or may not contain data
Using alter column statement only one column can be dropped at a time.
The table must have at least one column remaining in it after it is altered.
Once a column is dropped, it cannot be recovered.
16. How can one transfer LOB and user defined data from oracle to
warehouse using ETL informatica because whenever you select the source
data in informatica it shows it can take only character data.
LOB can be trasferred as text in informatica 7.1.2
17. what is data validation strategies for data mart validation after loading
process
Data validation strategies are often heavily influenced by the architecture for the
application. If the application is already in production it will be significantly harder
to build the optimal architecture than if the application is still in a design stage. If
a system takes a typical architectural approach of providing common services
then one common component can filter all input and output, thus optimizing the
rules and minimizing efforts.
There are three main models to think about when designing a data validation
strategy.
Accept Only Known Valid Data
Reject Known Bad Data
Sanitize Bad Data
We cannot emphasize strongly enough that "Accept Only Known Valid Data" is
the best strategy. We do, however, recognize that this isn't always feasible for
political, financial or technical reasons, and so we describe the other strategies
as well.
All three methods must check:
Data Type
Syntax
Length
19. what is the difference between ETL tool and OLAP tools?
ETL tools are used to extract, transformation and loading the data into data
warehouse / data mart
OLAP tools are used to create cubes/reports for business analysis from data
warehouse / data mart
Within a hierarchy, each level is logically connected to the levels above and
below it. Data values at lower levels aggregate into the data values at higher
levels. A dimension can be composed of more than one hierarchy. For example,
in the product dimension, there might be two hierarchies--one for product
categories and one for product suppliers.
Dimension hierarchies also group levels from general to granular. Query tools
use hierarchies to enable you to drill down into your data to view different levels
of granularity. This is one of the key benefits of a data warehouse.
Levels
A level represents a position in a hierarchy. For example, a time dimension might
have a hierarchy that represents data at the month, quarter, and year levels.
Levels range from general to specific, with the root level as the highest or most
general level. The levels in a dimension are organized into one or more
hierarchies.
Level Relationships
Level relationships specify top-to-bottom ordering of levels from most general
(the root) to most specific information. They define the parent-child relationship
between the levels in a hierarchy.
Hierarchies are also essential components in enabling more complex rewrites.
For example, the database can aggregate an existing sales revenue on a
quarterly base to a yearly aggregation when the dimensional dependencies
between quarter and year are known.
21. what are the data types present in BO? what happens if we implement
view in the designer and report?
Three different data types: Dimensions, Measure and Detail.
View is nothing but an alias and it can be used to resolve the loops in the
universe.
It is just a unique identifier or number for each row that can be used for the
primary key to the table. The only requirement for a surrogate primary key is that
it is unique for each row in the table.
It is useful because the natural primary key (i.e. Customer Number in Customer
table) can change and this makes updates more difficult.
On the 1st of January 2002, Employee 'E1' belongs to Business Unit 'BU1' (that's
what would be in your Employee Dimension). This employee has a turnover
allocated to him on the Business Unit 'BU1' But on the 2nd of June the Employee
'E1' is muted from Business Unit 'BU1' to Business Unit 'BU2.' All the new
turnover have to belong to the new Business Unit 'BU2' but the old one should
Belong to the Business Unit 'BU1.'
If you used the natural business key 'E1' for your employee within your
datawarehouse everything would be allocated to Business Unit 'BU2' even what
actualy belongs to 'BU1.'
If you use surrogate keys, you could create on the 2nd of June a new record for
the Employee 'E1' in your Employee Dimension with a new surrogate key.
This way, in your fact table, you have your old data (before 2nd of June) with the
SID of the Employee 'E1' + 'BU1.' All new data (after 2nd of June) would take the
SID of the employee 'E1' + 'BU2.'
25. What is the main difference between schema in RDBMS and schemas in
Data Warehouse....?
RDBMS Schema
* Used for OLTP systems
* Traditional and old schema
* Normalized
* Difficult to understand and navigate
* Cannot solve extract and complex problems
* Poorly modeled
DWH Schema
* Used for OLAP systems
* New generation schema
* De Normalized
* Easy to understand and navigate
* Extract and complex problems can be easily solved
* Very good model
1. Reduce or eliminate the time taken to get new and changed data out of
your source systems.
2. Eliminate, or reduce as much as possible, the time required to cleanse,
transform and load your data.
3. Reduce as much as possible the time required to update your aggregates.
Starting with version 9i, and continuing with the latest 10g release, Oracle has
gradually introduced features into the database to support real-time, and near-
real-time, data warehousing. These features include:
29. What type of Indexing mechanism do we need to use for a typical data-
warehouse?
On the fact table it is best to use bitmap indexes. Dimension tables can use
bitmap and/or the other types of clustered/non-clustered, unique/non-unique
indexes.
Snowflake schema
It is the normalised form of Star schema.
Contains in-depth joins, because the tables are splitted in to many pieces. We
can easily do modification directly in the tables.
We have to use complicated joins, since we have more tables .
There will be some delay in processing the Query .
51. What does the Metadata Application Programming Interface (API) allow
you to do?
A. Repair damaged data dictionary entries.
B. Delete data dictionary information about database objects you no longer need.
C. Extract data definition commands from the data dictionary in a variety of
formats.
D. Prepare pseudocode modules for conversion to Java or PL/SQL programs
with a Metadata code generator.
53. If I done any modifications for my table in back end does it reflect in
informatca warehouse or mapping designer or source analyzer?
NO. Informatica is not at all concern with back end data base. It displays all the
information that is to be stored in repository. If want to reflect back end changes
to informatica screens, again we have to import from back end to informatica by
valid connection and you have to replace the existing files with imported files.
54. What’s the diff between Informatica, powercenter server, repository
server and repository?
Powercenter server contains the scheduled runs at which time data should load
from source to target
Repository contains all the definitions of the mappings done in designer.
55. What are the tasks that Load manger process will do?
Manages the session and batch scheduling: When you start the informatica
server the load manager launches and queries the repository for a list of
sessions configured to run on the informatica server. When you configure the
session the load manager maintains list of list of sessions and session start
times. When you start a session load manger fetches the session information
from the repository to perform the validations and verifications prior to starting
DTM process.
Locking and reading the session: When the informatica server starts a session
load manager locks the session from the repository. Locking prevents you
starting the session again and again.
Reading the parameter file: If the session uses a parameter files, load manager
reads the parameter file and verifies that the session level parameters are
declared in the file
Verifies permission and privileges: When the session starts load manger
checks whether or not the user have privileges to run the session.
Creating log files: Load manger creates log file contains the status of session.
62. What is the difference between Power Center & Power Mart?
Power Center : we can connect to single and multiple Repositories, generally
used in big Enterprises.
Power Mart : we can connect to only a single Repository. ERP support.
70. What is a staging area? Do we need it? What is the purpose of a staging
area?
Staging area is place where you hold temporary tables on data warehouse
server. Staging tables are connected to work area or fact tables. We basically
need staging area to hold the data , and perform data cleansing and merging ,
before loading the data into warehouse.