project architecture We have to start with .Our projects are mainly onsite and offshore model projects.

In this project we have one staging area in between source to target databases. In some project they won t use staging area s. Staging area simplify the process..

Architecture Analysis Requirement Gathering Design Development Testing Production

Analysis and Requirement Gathering: Output :Analysis Doc, Subject Area 100% in onsite, Business Analyst, project manager. Gather the useful information for the DSS and indentifying the subject areas, identify the schema objects and all.. Design: Output: Technical Design Doc s, HLD, UTP ETL Lead, BA and Data Architect 80% onsite .( Schema design in Erwin and implement in database and preparing the technical design document for ETL. 20% offshore: HLD & UTP Based on the Technical specs.. developers has to create the HLD(high level design) it will have he Informatica flow chart. What are the transformation required for that mapping. In some companies they won t have HLD. Directly form technical specs they will create mappings. HLD will cover only 75% of requirement. UTP Unit Test Plan.. write the test cases based on the requirement. Both positive and negative test cases.

Development : output : Bugs free code, UTR, Integration Test Plan ETL Team and offshore BA 100% offshore

For each repository access permissions and location will be different. Based on the review comments u have to updated the mapping.. Migrating to testing repository Integration test plan has to prepare by the senior people. Deployment Doc and User Guide Testing Team. Mapping Bugs needs to fix by Development team. After that code review and code standard review will be done by another team member. U have to fill the UTP and Enter the expected values and name it as UTR (Unit Test Results). this is last phase of the etl project. next deployment in production environment. If client satisfy with the product . Production: 50% offshore 50% onsite Work will be distributed between offshore and onsite based on the run time of the application. Again testing starts till the bugs free code.. In ETL projects Three Repositorys. U have to create the mappings. Development : E1 Testing : E2 Prduction : E3 In Everyphase i given the Out come of the phase as OUTPUT after that .. Business Analyst and Client. UAT. Unit testing based on the UTP. Developers will fix the bugs in the development repository and again migrated to testing repository.Based on the HLD. Testing : Output: ITR. 80% offshore Based on the integration test plan testing the application and gives the bugs list to the developer. Client will do the UAT. 2 times code review and 2 times unit testing will be conducted in this phase. Development team will support for warranty period of 90 days or based on agreement days. 20% Onsite UAT User Accept Testing.

Its nothing to prepare mapping docs it easy based on the source structure and target structure. if some fields hav no related fields in the source for few target coloumns we will intract with Architech .. Customer.. because now a days informatica is using for automations of any kind of process. Based on ur knowledge and confidence level u can say i involved in that.. Involveparty. etc... we will mapp the matching fields. In my project . we will mentioned on the related coloumn in the doc. converting from flat file to xml . my requirement is to validate the data and populate the target with valid and useful data to the clients... Facts are relation between all those tables. simply using for some validations.. we used to automate some process of the client.. Distributor. IP_X_IP Its is not mendatory that ur project had dimensions and fact tables.. we willl called as standard tables( Dimensions) and relationship tables( facts tables). May be its snowflake schema..... etc..... IP_X_ACCT. dimensions are Account. This is first time i'm hearing level 4 and snowflake dimensions Did you do mapping document? If not who does that? How to do it? A) Mapping documents(technical design document ) will developed by Onsite ETL Lead.. In interviews u people can tel that my project is not Datawarehousig project. Q) Did you ever load a SnowFlake Dimension? If Yes can i know the process of doing ETL Mapping for a Level4 SnowFlake Dimension A) Snowflake Dimensions????? a Level4 SnowFlake Dimension ????? there is no such term as per my knowledge.Roles involved percentage of involvment. we won't call as dimensions and facts. Business Analyst and sometimes offshore Lead and Sr Software... because if clients are purchasing informatica software. they are trying to utilize the tool for another projects also.. this logics will be provided by the Business Analysts.. some fields needs to validate and lookup on some other tables etc.. ACCT_X_CUST..

.wht is that fields represents for and wht kind of data needs to populate to those fields. & All attributes of dimensions are updated & source system is having only latest information for populating dimension A) Inital Load is loading all the historical data from the starting to current date.. Worklet: . Mapplet restrictions: 1. we cant use non reusable sequence generator transformation in mapplet. In OLTP's they will keeep only few years data and they will load the old data into some flat files and keep it as historical data. 6.e from day one. u need to hav all the previous data from the starting i. 2. normalizer Transformation 4. Q) How do you do the inital loads (ex: 5 Years data). Traget defination 8. Sometimes hardcode values or nulls based on the design he will confirm. 3. First of all its not valid data. It contains a set of transformations and allows us to reuse the transformation logic in multiple mappings. once its finished all the loading till date. If u want 5 yrs data but only had 3 yrs data. Other mapplets. In tat case we hav to load flat files as initial loads and after that we have to load those OLTP data. Q) A) Mapplet: Mapplet is a reusable object that we create in mapplet desinger. XML Source qualifier transformation. we will Load Incremental Load from next day onwards on daily. while we have only 3 years data in OLTP. Cobal sources 5. we cant use external stored procedure transformation. XML Sources 7.

. The repository uses version numbers to differnciate versions. ex: sal return_value 10000 10000 15000 25000 13000 38000 in the first row return_value is 10000. second row is 25000(10000+15000) and so on. redundincies and incomplete information. it groups the set of tasks and allows us to reuse the tasks logic in differnet workflows.. Version: Any one of multiple copies of versioned objects stored in repository. CUME Function: Ruturns a running total. .Worklet is a reusable object that we create in worklet desinger. inspect data for errors inconsistences. A running total means CUME returns a total each time it adds a value .. Profiling: It is a process of comparing data and check wheather it meets the standards..

1) Local Variable 2) Global Variable Local Variable ----------------Local Variable are used in mapping itself.(Expression. Aggregator. Workflow Variables: Workflow variables use within workflow to write expressions referencing values that change from workflow to workflow. This means create mapping parameters for use within a mapping or mapplet to reference values that remain constant throughout a session. The Scope of the variable is LOCAL) Global Variable ------------------- . Creating mapping variables in mapping or mapplet to write exprdssions referencing values that change from session to session. Q) Abt Variables Two types of variables are there. this ports are local to the particular tranformation.Parameters and variables: Mapping parameters and variables: Mapping Parameter is a constant value between the session runs. Rank having the variable ports.

COUNT).interval point It is independent of base table eg: used in reporting tool wen u create cubes and dimenssions. In mapping variables three aggregrate functions are there(MAX. This includes matching the source and reflecting an accurate. The scope of the variable is GLOBAL. It stores the data nadlso data definition It occupies space.commit type. .Global Variable are used in both mapping. Mapping variables are called as global variable. It can be run on demand. Q)What is the difference between View and Materialized View?Explain both. normal viewl store the data but not the data definition normal view is depended on base table Materialized VIew:: Materialized view gives the access to view set of attributes in a table.In this case materialized will be used on that table Q) what is data reconciliation? A) Data reconciliation is element level checking where each element is valid.MIN.. A> Normal View :: It is a virtual table.It will not occupy any space. which shouldnt be shard like CREDICARD NUMBER. session.Sme info is very imp. workflow. valid value.

Sign up to vote on this title
UsefulNot useful