This action might not be possible to undo. Are you sure you want to continue?
There are two types in Informatica.. 1)Web based client tool - We need browser to Access --- Admin Console 2)Window Based Client tool - We dont need browser to Access ---Repository Manager,Designer,Workflow Manager,Workflow Monitor... and Admin Console-- Its A web based client tool n administrating the all informatica client tools such as Repository Manager,Desiger etc. And Administrating Users and Groups. Responsible for creating Repository services and Integration Services and Access by Administrator.
-----------------------------------------------------------------------------------------------------------------------------------------2) can any one explain clearly what do u mean by dirty dimension and junk dimension with
example? ANS: Junk Dim: Dimension table containing flags, gender, text,«. Which is not useful to generate reports? Then we say this is junk dimension. Dirty Dim: If a dimension contains records those records are maintaining duplication.
------------------------------------------------------------------------------------------------------------------------------------3) I need your help for faster loading from flat file to oracle table
Flatfile contains 20 million records please suggest me best to way load and better session performance
1)Use Expernal Loading Concept in informatic at session level.For this u need configure "SQL Loader". 2)Split ur single Flat-File in unix around 10 Files based on the size and Enable Partition in session and select target load type "Bulk" (or) I know there is feature sqlloader in informatica..., there is anything i have to do at db level .. i meant table at db as currently table is at Oracle database..
4) Can anybody explain the difference between
1) session variable and mapping variable. 2)Session parameter and mapping parameter.
ANS: session variable: Yes they are same But suppose if we define Same variable in Session and mapping with some value then Informatica will take Value which has been defined in Session mapping variable: mapping variable represent value that can be changed during mapping run mapping variable is reqiresd to define a incremantal extratoin and source qualifier t/r to per form incremental extraction
5) 1)how to load first half records to one target and remaining half records to another target
2)where we can use status code 3)der r 2 workflows wkf1,wkf2 if wkf1 sucess then only wkf2 executed how 4)what is vi editor 5) what operations performed on materalized view 5)when we go for connected and unconnected lookup 6)how target table gets refreshed daily 7)session variables and workflow variables 8)how to delete particular row i.e 2nd in a file unix 9) what r the properties set to improve session performance
6) unconnected lookup can return multiple values?
plz give me with brief explination? (I) (II) No it cannot. It will can retrun only one value. Google it you will get better answers. Any Lookup can't return more than one value. if lookup find multiple matches there are few options u can select 1) custom 2) first
3) Last If question related multiple return port fields You can get multiple fields out of unconnected lookup. To achieve it In Lookup SQL override write concate(field1, Field2) In Expression u can split that field again into two fields and use it as two separate fields. (III) Using Unconnected lookup also we can return Multilpe Column Values . Strictly Speaking Un connected Lookup Returns only one Value.So here i am concating all the column values using the || operator and returning this column from lookupup.So after that i am spliting the entire column into multiple values using Substr and Instr. Ex : Table (emp---> No,Name,Sal,Loc,Dname). So here i am passing only No to EMP i want to all other columns. Note : Enable the Lookup override : select name|| '~' || sal ||'~' || loc|| '~' ||Dname from emp (Give the dateype as string and Lenght in the Lookup as 10000 for eno it retruns all the columns based on '~' we will split the single column into multiple columns
7) How to load the filename of the flat file into the session statisitc table that is audit table
(I) (II) Which flat file name do you want to upload. I mean is it a source or a target? or anything else From Informatica 8 Onwords its ver Simple. 1) Click on on source defenition and edit the properties tab 2)Select the option 'Currently add Processing File Name" (III) Yes i know that option ... Session1 is flatfile to targettable(this is general business requirement load ) ....here we are not loading filename . Source filename is loaded into different table that is audit table this means after session1 is completed i will be running session2 (this is to load into audit table) Here I am thinking how can we get 'Currently add Processing File Name" from the session1 into session2 . (IV) decalre one variable in mapping variables & parameters. and hardcode the required source table in initial value or in parameter file.
. whether condition is true or false whatever it may be it processes all the records( if the condition is true it sends corresponding value and if the condition is false sends null ) where as Joiner is active and it processes the records which are satisfied by the condition. Mapplets are prefered to use repeated logic and re use that mapplets in different mappints. Both had its own advantages and disadvantages. It should use the cache for joiner. 2) Instead of multiple transformation we won't use mapplets.. but mapplet will not show the transformations. Performance won't differ by changing multiple trans into mapplets..wkf2 if wkf1 sucess then only wkf2 executed how 4)what is vi editor 5) what operations performed on materalized view 5)when we go for connected and unconnected lookup 6)how target table gets refreshed daily 7)session variables and workflow variables . Joiner is nice performer here. ------------------------------------------------------------------------------------------------------------ 9) 1)how to load first half records to one target and remaining half records to another target 2)where we can use status code 3)der r 2 workflows wkf1.8) Joiner and look up t\r which one gives better performance ? 1)how can we tell particular t/r will give good performance (lookup. joiner) ? 2) i have multiple transformations instead of multiple transformations we will take mapplet which one gives good performance??? Ans) my answer is both are same but insted of using multiple transformations we will use mapplet.. but we can't say which one is good that depends upon the requirement.. based on requirement performance differs.. (I) 1) Lookup and joiner can be used for the same purpose. Joiner will take master table into cache by default. ANS: As per my knowledge. Lookup also uses cache but u can minimize the cache usage in lookup and u can write sql override in lookup filter only required fields and records by mentioned some conditions in override. (II) Lookup is passive t/r that means it processes all the records.
e 2nd in a file unix 9) what r the properties set to improve session performance ----------------------------------------------------------------------------------------------------------------- 10) need solution for this i/p1: a b c d i/p2: e f g h o/p should be: ae bf cg dh ANS: (I) give the clarity of this question Rashmi.8)how to delete particular row i. ANS: Regarding UNIX Usage in Informatica: First of all . I am generating a seqence numbers for the first file and second file. what i am asking is i/p 1 and 2 are different files or two ports are in same file ? and o/p is two columns or same column ? Please find the solution below for your requirement : Sourc1---SQ---->SEQ---> Souce2--SQ---SEQ-------> Joiner [It will Join the 2 pilenines into one(Condition is Source1-->SEQ=Source2-->SEQ)--->Load the records into target Here. You know need to know unix commands when ur working in Windows environment. If Sources are databases and target database then u hardly use Unix environment... In Informatica development projects Unix is not mandatory. Some projects use windows based environment. If anybody related documents. Even if you are working in Unix environment. (II) 10) please anybody explain unix commands which are frequently in Informatica development. So based on the seq numbers i am joining the two files. .
. 14) how can i lookup the data without using lookup transformation? ANS: with join t.get the file from unix cp-.search for the content in the file These are some basic commands ------------------------------------------------------------------------------------------------------- 11) how to split one half to one target and another half to another target using unconnected lookup how to do this can any one explain me..which situetion..connecting to unix for files transfer or download put -. while null values on the scan it should take default only. ANS: null comparision in lookup with dynamic chace when we specified a condition in lookup tm. for null values the integration service take default value instead of taking the null values....m 15) table A: col1 bangalore#chennai#Hyd Delhi#bombay .. 12) regarding update stretagy? without using update strategy how can i update the table? with simple mapping? ANS: 1)Enable the Update Override in Target in the Mapping 2)Select Treat Rows as "Update" 3)Target Session Properties "Enable only Update as Update" 13) comparison null operetor what is comparison null operator in lookup.....copy file to different file rm-..remove file vi-.. in un-connected look up tm it should give null only..UNIX Basic commands using in dev FTP -..vi editor to modify the file content Grep-..where we use it in dynamic lookup.place the file in unix get -.
16) flat file hi can u please help on this i have flat file. ANS: (I) (II) (III) (IV) Yes you can use Normalizer Transformation Use instring and substring in oracle to acheive the result. how to remove the unwanted data and how can we achieve this. .. and source is connected to target it is one to one mapping only. using informatica and also want sql query for that(oracle).. user defined join..Required format Col1 col2 col3 bangalore chennai hyd delhi bomby How can divide one row to multiple rows... now i want to find the max(sal) from two tables and needs to compare tables like(max(sal) of emp1 >max(sal) of emp2) if emp1 having the max(sal) i need to pass one set of records and emp2 has max(sal) i need to pass another set of records. eather data coming from database means we can use substring and instring concepts... the content of file is huge and it contains lot of unwanted data. and we should not use any transformations..please help me ANS: (I) 1) Please some filter condition in the source qualifier like Deptno = 10 2) Use some pre UNIX scripts to validate the flat file and copy those records from the flat file to the other file and then trigger the workflow Hope this may help (II) (III) can you please tell me where we have to implement unix script In the informatica server it self Please move/copy your file into a particular directory of unix server where Informatica is installed and from there using unix script do all the data validation and format correction then trigger your Informatica job in sq trns write a query in sql override.... (IV) 17) i have two employee tables with the same structures but with differnet data . source filter and select distinct we can restrict the unwanted data. when ur importing the flat file take a delimiter as # then directly map to target we get output what ever ur asking..
when u want to restrict to access ur mappings from others use version controlling. I think LRF is another concept bcoz. it is compared with input ports. we have 4 techniques. checkin used for to the mapping.e. 1. there we have 2 options checkin. adding the data) merging (merge the data diff sources) aggregations (loading the summarized data) 3. 19) Do you have any idea about LRF ( Load Ready File) in Informatica? What is LRF ?. associated port is default port when we are enabling the dynamic cache in the lookup trans.e. versioning is good concept of informatica. data validation means when source is coming to the staging area we have to validate the i. 18) What is associated port in informatica?????? What is data validation???????? What is versioning in Informatica??? ANS: 1. 20) Which transformation is used instead of lookup to check whether the record alredy exist or not??? ANS: . checkout used for edit mapping. we will wait for the file to trigger the Jobs/Sessions. 3. 4. 2. i. data data data data cleansing (remove unwanted data) scrubing ( adding the ports i. and compared result will give the "NewLookupRow" port. indicator file will give each every record different id's.e. 2. checkout.Using Even Wait Task in informatica. weather it is inserts means give one id updates means give another id like that. in this informatica server will give one version number starts with 1. applying the some changes.please help me asap. group of members accessing the same repository. one type of output file. When it comes in to the picture ?. how we deal with it ? ANS: (I) (II) LRF is "Indicator File". "indicator file" maintains the id's and it is one of the session log statistics.
. $Target User Defined Variables: ---------------------------- . (III) taking 3 instances of target is working but full records are not comming.. but i guess it works Let me know if i am wrong. @shashank No need of Union transformation here..(I) (II) Stored Procedure Joiner tm 21) how to print singele record three times for example i have emp table i have the king record only once in the table [i.. Its defined by the Informatica itself.k also when i took union t/r they are getting successfully loaded with 3 duplicates of every record... Create a target instance three times.e no duplicate records] but my requirement is to print the same record three times ANS: (I) I guess u can take union transformation and do it.. U will get the expect output..... They are 1) Pre Defined Variable 2) User Defined Variable 3) System Variable Pre Defined Variables: ---------------------------It denoted by Single Dollar ($). i have removed the p. Take 3 emp source instances and union it in union transformation and u get 3 duplicates for each record in target. Example : $Source. Connect all the output port the three instances.. Never tried it.. (II) 22) diff between $ and $$ in informatica If any body know the information about this post me ANS: (I) $ is Mapping Parameter & $$ is Session Parameter (II) Three types of variables are available in Informatica..
25) mapping lock .e. then we want to use sql t. Example : $$Currdate.. while session is running). 3... then session parameters come under predefined variables??? examples of session parameters $session parameter $Inputfile ---------------------------------------------------------------------------------------------------- 23) TCS question 1... it will show a window of the speed that inf. What is the use of sql t/r?? when we are using sql t/r?? Wt r the situations are used sql t/r in project?? ANS: 1. when u want to create a table dynamically (i. wt is pipeline partition.wt is throughput in informatica 2. Its defined by the informatica developers depends on the logic. it wil be displayed when u double click on session in workflow monitor . wht u r saying is not the exact answer... and also depends on client requirements. 2...r mandatory.r not supported to ddl commands. ANS: THROUGHPUT is nothing but the speed that informatica server reads the data from source and writs the data to the target /sec..It denoted by double dollar ($$). 24) SQL Transformation Hi Dudes.. check it out .sql t.server read and writes..... that is on depends on LLd preparations. $$deptno System Variable: --------------------It denoted by trible dollar ($$$) Example : $$$Sessionsystime @Chocsweet: Pls chk the things above. For mapping parameter and mapping variables are using the user defined variables.. (III) if $ is used for predefined variable.. where as other t.r will supports the ddl and dml commands..
this lock can be released by disconnecting the respective connection id(user specific) for that object in Admin console I think from Infa 8.Under what cicumstances a mapping gets locked. If you directly close the window. second person will prompt a msg " m_test mapping is locked by user(pavan). (III) Its a Write intent lock. Locks will be on corresponding mapping or workflow. 3) if two persons are working on the same mapping / workflow.6. may be it will be acheaved by using aggregator t/r but i dont know exactly. But earlier version all the locks can be managed by Repositry Manager. hope this will help you to understand locks in informatica.------.-------1 maths 60 1 science 72 2 maths 67 2 science 82 now i need output like this Rno subject Marks subject Marks ----. There are different situations mappings or workflows are locked by other users. (IV) 26) i have scenario like this Rno subject Marks ----. if some user has opened it. i think Normalizer T/R will giv us the Input in this example as its output when the output of this particular example is used as the Normalizers Input 27) in which scenario we go for mapplets?pls tell me in real time with example ANS: (I) Definition and Limits .--------.1 onwards this activity is pushed to Admin console to manage. If you reconnect to repository and if you open the mapping which you are working on sometimes it will says mappings is locked by User(Pavan) 2) If you are not closed the designer or workflow manager is not disconnected properly.--------. And how to realease the lock from the mapping?? ANS (I) Locks in informatica Hi shashank.--------. 1) Repository is disconnected when u r working on any of the mappings. If you want to release the locks u have to go to repository manager check user locks and u can release the locks.-------1 maths 60 science 72 2 maths 67 science 82 ANS: (I) (II) i think normalizer will not work here.
You can create a non-reusable instance of a reusable transformation. If you make changes to the original. the copy does not inherit your changes You can use a single mapplet.. it expands the mapplets.. we have error stretagy which is applicable for all the mappings.ERROR Default value for variables . what are the scenarios we thought errors.. even more than once on a mapping.. example one row had three errors we are creating three records out of it and showing three errors for the same record. We are routing the error records based on the key coloumns null or null values.Contains one or more active transformations (b) Passive Mapplets . Mapplet Objects: (a) Input transformation (b) Source qualifier (c) Transformations.Contains only passive transformation Copied mapplets are not an instance of original mapplets. In Mapplet. changes to these can invalidate the mapplet and every mapping using the mapplet. this the logic common for all the mappings.. so we implemented this logic in MAPPLETS this example any one and any domain can tell as example of mapplet. empno name date sal 101 praveen 12/12/1999 10000 Null pavan 33/13/1990 afdsfdsfds .Mapplets When the server runs a session using a mapplets. as you need (d) Output transformation Mapplet Won¶t Support: Joiner Normalizer Pre/Post session stored procedure Target definitions XML source definitions Types of Mapplets: (a) Active Mapplets .. we capture those as errors and sending the only error records to the Mapplet input. Ports Default value for I/P port .Does not support default values Example In one our my projects. That will capture the error records and flag them and attach error msg to the error records and writing to the error table. we are converting single row into multiple rows if that records contains more then one error... The server then runs the session as it would any other sessions.. If you use a reusable transformation in a mapplet.NULL Default value for O/P port . passing data through each transformations in the mapplet. example.
. Error Details :- .. sal is not valid. so our output in the error file like this rowid tablename error field error msg 2 emp empno empno null 2 emp date date invalid 2 emp sal invalid salary. u can use connected also. OS is Vista and database is Oracle 10G vista compatible .in second record.. we have to mention occurs option based on the max number of possible errors.. i'm just explaning till the last to load data into error table ater that flow will go to mapplet out---> Normalizer --> error table... In Mapplet.6 on my PC. In normalizer. if ur using connnected lookup it will create the cache for each coloumn in the lookup. Mapplet Input---> Expression---> UnconnectedLookup--> Mapplet output actuall mapplet will do only segregate the errors and assign the error code and lookup on error msg table based on the error code and assign the error messages to the each error and also error count for each record in expressoin based on tat we will mention occurs option in normalizer to split single row into multiple rows. Please let me know if ur required more abt this. so it might hit performance so we used unconnected. date is invalid date. hope it cleared ur doubts. based on the code we will lookup on the lkp file. I am getting the error while installing infa 8. empno is null. (I) can u explaine me how maplet doesnot support joiner T/R joiner transformation is supported in mapplets. that is old version.....6 x In latest versions Joiner will work. Sorry i just copied that matter from material.in the above example we have three errors. if not reply back... we required only one return port from the lookup so we used unconnected where we will get the performance improvement also.. In the Expression we will validate what kind of error and assign the code for each type of error. Its out of mapplet.. there is no specific reason to use unconnected lookup. so what will do is we generate row number for each input records to identify which record had error. (II) (III) No logs have been created. 28) Hi All.
Found few errors with respect to configuration in Admin console and Client let you know later. 3)While the installer was pinging the domain I stopped and re-started the service... Stopped the service and disabled it then after saving re-enabled it and restarted. but my qtn is just by chking that option how informatica filters out only the new records from the source? do we need to do some kind of look up in mappings? or its enough if we just click this incremental aggregation option? ANS: (I) (II) (III) when u send data second time and you have new values to be aggregated then you go for incremental aggregation. etc.. 4)Used infacmd commands to try to add a domain and/or ping the domain externally... when u select incremental aggregation in session properties the data is copied into cache and checked in cache for the new values and aggregation is performed. Thanks Shashank . (I) 1)Be sure that java is installed as its own installation not included with other software..0 cannot start Informatica Services.. property in sessions.... Use the error below and catalina. (II) Thanks a Lot Aparna...out and node. Select Retry to continue the installation. 5)Handle Compatibility issues.... etc. I did find one time I had to manually create a domain using the infacmd. 2)Put the Informatica service bin path in the environment path of the system. changed the service ID back to a domain user ID..log in the server/tomcat/logs directory on the current machine to get more information.6. ANS: Check the following. change user service ID from a network domain user to the system account.. 29) I am new to informatica . .As per your suggestions infa Server got successfully Installed.. u mean to say the source data as is copied into the cache? and it compares the new source with the cache and processes only the new records for aggregation? Its is not source data which is going into cache it is target data and source data is compared with this cache. I knw we need to chk incremental agg... etc.Informatica PowerCenter 8.. EXIT CODE: S Please Help me out regarding this issue.. Let me know if it works or no. can any one explain me abt incremental aggregation property.
The property(checkbox) specified in the sesssion tab would NOT be responsible to filter out the new records and aggregate it...Jan Feb and Mar Target may be Qtr? How it compares? As per my knowledge. Ya thts correct. (V) (VI) . * telecom domain in postpaid connection to find out the total balance what have you talked till today for this month will be calculated in incremental aggregation.(IV) okey . *Insurance domain: To sum up ur premium amt from d start effective date to till date they will be using incremental aggregation.explained by the shiva prasad.infa creates two set of caches for data and index.. but the target is aggregated . The option is used only for the incremental aggregation or also called as running aggregation.the target table would look like below: -------------------------------------------------------------------empid allowances date 1 100 20sep 2 200 20sep 1 300 21sep 2 500 21sep 3 400 21sep Note the aggregation of the allowances that has taken place above and also the new record that been inserted into the table... The onus would be on the user to filter the delta records(changed/New) at the source side and pass it on to the aggregator......(original set of cache and backup set of caches which is used for recovery purposes).. When you check the option in the session properties.. for example Source .. eg: existing data in the emp table as on 20sep ------------------------------------------------empid allowances date 1 100 20sep 2 200 20sep incoming records on sep21st ---------------------------------1 200 21sep 2 300 21sep 3 400 21sep After running the incremental aggregation enabled session.In realtime these incremental aggregation will be used in telecom industry and insurance domain.
.Name || decode (row_number() over(partition by id order by id).rank) Disable the Outputport in the exp for the columns (name.1...1.null. -------------------------------------------------------------------------31) please tell about ETL testing using informatica. At what stage ETL testing people will come in the project. Roles of ETL testing experts..null.update strategy to filter out the new/changed records just like SCD ?? and pass it to the aggregator ? 30) scenario id name 1 abc 2def 3ghi 1 abc 2 def 2 def 3 ghi ---------Target ----------id name 1 abc 1 abc1 2 def 2 def2 2 def3 3 ghi1 3 ghi2 how to acheive this logic ANS: (I) (II) (III) select id. ANS: (I) basically ETL testing is 4 types:1 unit test: for developed mappings 2.rank) Create One O/P Port named as Named_rn give this formula name || (decode(rank.name. So we need use lookup. If this is a flat file how do solve this Source(Flat-File)--SQ-->Rank Trans(Group by ID and select no of ranks say 10000)-->Exp (Drag all the fileds from Rank Trans-id.rank-1) Now Connect the id & Names_rn from expresiion to Target.Any explaination needed reply in this forum itself.integrate test: gathering all mappings to one location this will do superiors .(VII) Thse are the some small examples in realtime.row_number() over(partition by id order by id)-1) from Table Use the above query in Source Qualifier and load the records into your target.
..Existing data testing 2).Editing/Deletion of existing data..3. I guess even if the parameter file goes missin and if you have assigned the initial values for that parameter @ the mapping level.... how it is possible in informatica? ANS: (II) ..You shouldn't be getting the error. How UI reacts to the changes 4).UAT:user acceptance test: this will show to our client this is the final test.system test: this is all so one of the impotent testing 4. 33) I don't have source data but i have to test my mapping.. test procedure: test case: expecting result: actual result: status: pass or fail Apart from mappings....the following test scenarios should be considered 1).New data creation 3).. testing is different scenarios are there : test spec will provide your org...What are the filters/SQ's/Target connections/Jobs and Tasks 32) This is tcs interview question if parameter file miss what eror will come?plz answer ANS: (I) (II) PETL_24049 failed to get the initialization properties from the master service process for the prepare phase [Session task instance[Session_name] Unable to read variable definition from parameter file[file_name]] with error code  This error i got n if its wrong let me know the error.
. this is also having different options.. Hope this will help you to understand migration methods . 1) Repository Dump and create new repository with different name in Prod 2) Export and import workflows from Repository Manager.(I) Hi Guys. 3) Unix Scripts to migrate workflows from folder to new folder Who's responsible for this migration. A mapping is not valid without a source or target. Siva is right here . 1) Some project will have separate Admin team will handle this migration.. It will be different from project to project. Deepthi might have confused here. A debugger can not be used in a invalid mapping. without loading to target table u can test with sample data using debugger or test load options. -----------------------------------------------34) How do you migrate from one environment to another like development to production in Informatica ANS: (I) (II) Migration code from dev --> test --> Prod is having different methods. 2) In team experienced people will do the migrate 3) sometimes development is handle by one company and Separate vendor's(different company) will maintained environment in this case we have raise migration request to respective team to migrate the code. Without sample data u can't test the mapping.
else I could help you out.. this work just like copy workflow from Repo to Repo...1 onwards. Infa 7. PMERR_DATA..... PMERR_MSG. Check workflow administration guide for further explanation.these tables log the error details on row by row basis. But I think you can get the error code and description in some error table as well. This errors are related to only informatica functions or validations or technical errors not data related errors.. Trust me this works amezing. but I think 8.. 'ERROR_CODE'..... --------------------------------------- (III) 34) ERROR TABLE IN INFORMATICA REPOSITORY ANY BODY KNOW ABOUT THE " ERROR TABLE " IN INFORMATICA REPOSITORY.....1. Not sure from which version onwards it comes... PMERR_TRANS.......1 onwards provides 4 inbuitl error tables. In session properties you have option to log the errors in error tables. These OPB* tables are specific to Informatica.4 onwards provided separate repository for error tables.... Sorry I do not have any OPB table access....6... IT HAS A UNIQUE 'SEQUENCE_NUMBER'. PMERR_SESS.(II) import & Export through repository manager import & Export through pmrep command line Keep both the Repositories open in a Infa Repository manager client Copy Paste Drag & Drop Deployment groups Another way is called Deployment Group.... ---------------------------------------------------------------------------------------------- (II) .... Informatica will log Technical issues related to informatica process.. But make sure you use right way of Infa version control. Informatica build 4 separate error tables to log the errors. PLS. ANY BODY HAVE THE KNOWLEDGE OF THIS TASK... You need not to worry about the dumps or export import... Thats right Pawan. Informatica log is built based on that repository information..'ERROR_DATA'.... PLS EXPLAIN US ANS: (I) In Informatica from 7.. For error info..
router is no use here. (II) Ramakrishna.... no need of router. you can directly map first 3 ports to Target1 and remaining 3 ports to target 2.. first three coloumns needs to load to target 1 and next three columns need to load to target 2.35) Splitting source columns I have flat file source with 6 columns I want to load 1st 3 columns in target1 2n 3 columns in target2 How? ANS: (I) From source qualifier directly connect first 3 columns to first target and remaining 3 columns to second target.. yes . Could any one tell me what is transformation error? . (III) i don't think so you need to go for sorter and ranker for this solution. Siva answer is correct u can directly connect the columns to different targets based on requirement..delete in session in target mapping tab n target one option is there IN TARGET properties Upadate override option is there using this one we can update the target (III) 37) what is Transformation Error hello friends.. check the problem clearly it is simply loading from source to target... RamaKrishna: Router concept will not come into picture here.. Update else Insert (II) updating target by using data driven select update. Abdull Wrong answer.. he is not talking about records only columns.etc (IV) 36) I got a Question like With out using Update strategy transformation can we Update the target in informatica? If it is possible how ? waiting for reply ANS: (I) By Using Session Properites.insert .....
.ANS: (I) transformation errors are sql override in the source qualifier transformation some date conversions and lookup override in the lookup transformation 38) Please Provide Large Databases? please i want to gain more knowledge on informatica so while taking source database it will be better if we have some database which has more fact nd dimension tables along with more records. Ya that may be possible. next like that only But you should have fallowing : 1) Two DB users 2) oracle port number(default is 1521) .i have faced d question in DELL interviews...do any 1 have large databases just like emp database ANS: (I) in cognos u will find gosales database with lacks of records (II) we have SH ( sales History ) and OE (Order Entry) schemas are there in Oracle so those may helpful to your requirement 38) WEIRD QUESTION FACED IN INTERVIEW I have d source file with 3GB memory. Oracle SID ( difault is ORCL) ANS: (I) .6 installation It is just next .3)=2) Pass this to T1 Default Group : Pass this to T2. Source(Flat-File/DB)--->SQ-->EXP(Pass all the columns and have a sequence number by using seq generator)--->RTR(two Group )--->Targets (T1. So here you want to load the 75% of your input data into first target and remaining 25% of data into the second target..I need to load d data in 2 target tabls according to following condition... T1: should have d data for 2GB T2: It should have d remaing 1 GB data... ANS: (I) Hi.but he asked me to implement the mapping for above scenario. ----------------------------------------------------------------------------- (II) (III) 39) informatica 8.3)=1 or MOD(seq. 'split' unix command can be used to split the files. plz answer me for dis question.the only solution that i may think as of now is to split the file into two based on the size and then load them into there respective targets.T2) First Group Cond : IIF(MOD(seq.
the data got loaded successfully but when i checked in session log file . Hi Prasad. Create an event-wait task and use it as the first task in wf2.txt and in wkf2 use even wait task and use file watch event on that flat file..once all the tasks are executed successfully in workflow1. (III) 41) how to delete dublicates in flat files ANS: (I) using sorter t/r Check the distinct option in property tab 42) help me pls in this data issue hi friends While loading from flat file(fixed length) to oracle table . the pmcmd command in command task will trigger the workflow2 automatically. after the execution of wk1 we have to execute wkf2. the file created can be used as a file-watcher in wf2. Use the post session command for the wf1 to create a null/zero file. In this approach we can't achieve all the requirements. You can select the option of Delete File-watch in the event wait task(depends on your requirement)....And other part of the installations are 1) creating the Repository 2) creating Integration service After this you can start using Informatica I hope it will be usefull after the installing integrating server/service tne integrating service & rep not working change the operating mode from exclusive to normal and click on enable then ur informatica repository service will be running 40) i have 2 work flows namely wkf1. i got the below error . Such that it starts the preceding sessions of wf2 & after the creation of 0 file(wf1 completes). in this solution if you run the wkf2 it will wait for the flat file to create if not it won't run the wkf2. if wkf1 is not executed means you should not execute wkf2 we have to do it automatically ANS: (I) (II) (III) (II) use the command task in workflow1 as the final task and start the workflow2 using pmcmd command.. Hope this will be ur solution.. Severity: ERROR Timestamp: 9/16/2010 1:31:45 PM Node: NODE_02 Thread: TRANSF_1_1_1 .wkf2.. Solution for In this case u have to create a flat file at the end of wkf1 by using touch xyz. any time u can start the wkf2 in ur case but requirement is wkf2 shouldn't run if wkf1 is not succeeded.
Use condition something like this. Error message is [<<Expression Error>> [TO_DATE]: invalid string for converting to Date . t:TO_DATE(s:LTRIM(s:RTRIM(s:' '. If you trying to do to_date function in null values it will throw an error like this. Solution To resolve this use the convert the date format of the SYSDATE to the formate used by the TO_DATE function. ANS: (I) Hi Karthik. I got this question in an interview His requirement is he has two individual source qualifiers and he want to copy the port from two SQ to Expression with out using any other (II) . to_date('19991231'.'yyyymmdd'..s:' '). (III) 43) source qualifier to expression Hi friends i have 2 source qualifiers ( EMP & DEPT) and i want to use expression transformation and in which i want to copy port from both the transformation.'DD-MON-YYYY'). Hi aparna Thanks for reply.Process ID: 6749 Message Code: TT_11132 Message: Transformation [e_Donnelly] had an error evaluating output column [v_DATE2]. Example: TO_DATE(TO_CHAR(SYSDATE. so i did it and port are copied from both the table SQ but links are forming from only one source qualifier only why like that if you don't get i will explain again ANS: (I) You have to define the Join Condition in the Source qualifier. This error will occur when using incompatible date formats...'yyyymmdd') ) Try this logic it will resolve ur issue. to_date(in_date. Let me know if it works or not.s:'YYYY-MM-DD')]. The default date format in PowerCenter is MM/DD/YYYY HH24:MI:SS and hence the SYSDATE is converted to MM/DD/YYYY HH24:MI:SSbut in this case the TO_DATE is expecting the input in the DD-MON-YYYY format.s:' '). IIF( ISNULL(IN_DATE).'DD-MON-YYYY') (II) data is in string format like 20070802 this data is getting loaded but some data has no value and it is creating error for those data the file is fixed limit If some records having null u have to filter those records or send default date to those records.
.DEPT_NAME.... and create one source quailifier transformation . it has duplicate records I want to send 1.TPJ Your EXPRESSION T/R PORT will be like this PORT NAME...CHE 10.. now your scenario is to compare the record by the use of DEPTNO Table data: DEPTNO..dept) because they have common columns.....Distinct records in target 1 2...AAA..transformation like source1----source1_SQ ---------> Expression Trans source2----source2_SQ Like above is it Possible (III) (IV) You can use one SQ->Lookup->Expression hi ratna kumar i got you but after SQ no other Transformation . source qualifier is a active transfor mation.BGR 20.. then remove one two sourc quailfiers ..DEPT_NAME 10. only multi input group tm accepts the two inputs .BBB. after that connect the ports to expression transformation. direct Expression only like this he asked me if the two tables coming frm same database..CCC. then combain two tables(emp..I. read transformation guid for assistence.KMU 30.. Duplicates in flat file For ex: Assume you have Dept table (DEPTNO. EXP DEPTNO (will be I/P port) DEPT_NAME(will be I/P Port) LOC(will be I/P Port) (II) .Duplicate records in target 2 How to do this? ANS: (I) Use Exp Tx compare the row and set the flag then use router transformation.. (V) 44) Duplicates in flat file I have a flat file. so expressin never act lick that..DDD.O...V.......EEE....HYD 30.LOC) are the fields in the table..
'Y'.sal) how it get sorted whether it will create a virtual table to sort or what it do actually what what is the need of count please help me ANS: (I) Its querying the emp table inorder to find the Nth Highest salary..When the workflow runs for the third time only the third target should be populated and the rest two(first and second) should not be populated.. Before executing this query you can give this command.sal>=e. hope this helps you. Then use router ----------------------------------------------------45) Please explan this query Select * from emp e where &N=(select count(distinct(sal)) from emp f where f. posting the thought. Check PLUSTRACE role is enabled SP2-0611 Error enabling STATISTICS report (III) (IV) Which version of oracle DB are u using? IM using oracle 9i You have to run these scripts before you set autotrace on.Now Create new ports in this order PRE_RECORD(Variable port) TEMP(in expression) TEMP(Variable port) DEPTNO(in expression) FLAG_DUP(O/P port) iif(PRE_RECORD=DEPTNO.When the workflow runs for the second time only the second target should be populated and the rest two(first and last) should not be populated.sql $ORACLE_HOME/rdbms/admin/utlxplan. for eg: 5th Highest Salary. "set autotrace on".When the workflow runs for the first time only the first target should be populated and the rest two(second and last) should not be populated. take sequence generator in exp and generate the numbers starting from 0 (II) .'N') This is use to compare the records and it will flag Y if it is duplicate.sql 46) router is an active or passive transformation ANS: Router is a Active tx because of Update strategy property and it will change the Rownum of the table --------------------------------------------------------------------------------------------------------------47) Suppose I have one source which is linked into 3 targets.Could any one ANS: by my first look.. $ORACLE_HOME/sqlplus/admin/plustrce. Thanks for ur reply but it showing following errors Could you help me SQL>set autotrace on SP2-0613:Unable to verify PLAN_TABLE format or existence SP2-0611:Error enabling EXPLAIN report SP2-0618:Cannot find the session Identifier.. Then you will know how the SQL got executed Hi Aparna.
BTECH 102.BCOM (II) (III) .3)=0 Table2 : MOD(SEQ_VALUE.use a Router t/r which is connected to the Target table.EDUCATION 101.PRAMODH.SNAME.3)=1 Table3 : MOD(SEQ_VALUE. define three group conditions as Table1 : MOD(SEQ_VALUE.3)=2 48) can anybody solve this scenario please my source is id 1 1 1 1 2 2 2 2 3 3 3 3 then my targets are like target1: id 1 2 3 target2: id 1 1 1 2 2 2 3 3 3 ANS: (I) This question is similar to past post only to do this SQ----> sorter--->expression---->router------>target1 and target 2 routing of unique values into one target hi frens i have got one source table(here source is of type flat file)like this SNO.VIJAY.
PAYAL.BTECH 104.MCA 108.RAJESH.MSC 105.. may be u got confused with Jai answer using aggregator.SHANTI.MCA 110.MOHAN. here i'm giving you the coding check it out. now aggregator is not required.PRAMODH.SNEHA.SNAME.KISHORE.PAYAL.EDUCATION 103.BCOM 109.BCOM 106. ENAME. 1.BTECH 108.MOHAN. Source Qualifier -> sorter -> Expressino ->router -.MSC 107. EDUCATION Take two variable and one output port v_FLAG : IIF(V_TEMP = EDUCATION.RAJESH. the way you are proceeding seems doubtful to me «. hi friends correct me if any wrong in that code. Din & Ramesh is correct.SNEHA.0) V_TEMP : EDUCATION O_flag : v_flag In router : Create two groups unique values group : o_flag = 0 duplicate values group : o_flag=1 map those values to corresponding group. thanks pavan for ur detailed explanation here i followed another approach and got it i used source qualifier -> rank -> router ->target in rank i grouped by education and gave ranking from bottom(say ranking upto 5) and in router i used o/p grp condition rank=1 which obviously routes the unique values to one target and default grp routes dummy values to another target i have a doubt Shankar .MCA 110.BCOM 106.EDUCATION 101.MSC Hi shankar.BCOM 104.KAMALINI.SHANKAR.103.VIJAY.KISHORE.MSC and i need o/p in such a way tht routing of unique values into one target and dummy values into other with in a given mapping only SNO.SHANKAR. Here is one example suppose in your target table you have rows as below (IV) (V) (VI) ..BTECH 102.2 targets in sorter : sort on education in expression : INPUT & OUTPUT PORTS ARE ENO.BTECH 105.SNAME.MCA and in the other tagrget table i need like this SNO.BTECH 107.BCOM 109.SHANTI.KAMALINI.
Second Difference is Whenever we perform join Using Joiner It contains Must Primary key- .BCOM 104. may be u got confused with the question.SNAME.BTECH 105.orkut.PAYAL. tanx for sharing another approach also. We can have a look up to the existing table « and can give a condition «.SHANKAR.SNO.SHANTI.MSC Now if next day a row comes with the data SNO. first day all the records won't be there in the target table so all the records go to one target.SNAME. 49) interview question Hi How to extract original records at one target & Duplicate records at one target? Thanks in Advance.co.KISHORE. unique record into two different targets in the same day.VIJAY..MCA and in the other tagrget table SNO.If you use the lookup on the target table first day ur approach won't workout..BCOM 106.MOHAN..AMIT..EDUCATION 101. in informatica we can do it in N number of ways. Check the similar question in the old postshttp://www. ANS: You are solution is not completed/Correct.BTECH Now tell me the row will get inserted to which table? So what I want to suggest is «.MSC 107.RAJESH.BCOM 109.BTECH 102.EDUCATION 103.PRAMODH.SNEHA. i can understand ur point but shankar requirement is not that..KAMALINI.in/Main#CommMsgs?cmm=791012&tid=2543660813794278811&kw=v_co unt --------------------------------------------------------------------------------------------------------------50) what is source commit and target commit intervals where we can find rejected records and how to reload that records what operations we can performed on materalized views exact difference between lookup and joiner trs ANS: (I) The main differencee between lookup and joiner is Lookup is performs Non-equijoins and joiner performs only EquiJoins. If that found put into table 1 if not put into table 2.SNAME. u are thinking abt next day also.MCA 110. (VII) Shankar u approach is rite.EDUCATION 111. it will work for ur requirement. Shankar wants to findout the duplicate .BTECH 108.
. eg after sening 1000 records . The PowerCenter Server commits data based on the number of target rows and the key constraints on the target table... By using Lookup Transformation Just We Need a Matching Port. You can also configure some commit and rollback options in the session properties... to improve the session performence you have to increase the commit interval . The commit point is the commit interval you configure in the session properties..Foreign key Relation ship. Can anybody clarify me these questions. If any diff is there Tell me Commit Is The Wat amount of data is loaded in to target during session runs main we have 3 types of commit points is there (II) Target-based commit... The PowerCenter Server commits data based on transactions defined in the mapping properties.. The commit point also depends on the buffer block size the commit interval and the PowerCenter Server configuration for writer timeout . The PowerCenter Server commits data based on the number of source rows. User-defined commit....input(unchecked) and write 'New York' in the expression part 51) Hi All. 8)Why to choose oracle or sql server for data warehouse? ANS: ..... Source-based commit.. Rejected Records Are Saved in Session log in the form .. these are just like savepoint in oracle ok.bad File (III) source commit is after how many records you wan load to target . 1)Surrogate comes under which category? 2)Where we use surrogate key?(technically) 3)Whether first we should use and why? i)Expression Transformation ii)Filter Transformation 4)What is basic diffrence between star schema and snowflake schema? 5)How to do performance tuning in Informatica? 6)What is diffrence between data migration and data warehouse.even if i 'hard code' any value in output port its not coming to next transformation. 7)What is dynamic and static lookup transformation..... 51) In expression transformation no values are getting passed from output port. ANS: (I) if the value is coming into the expression then its not hard coded hard coded means that the value remains the same for tht port like suppose if we want a port STATE to have a value as 'New York' then v make that port as output (checked) . commit interval and target commit point after reaching 1000 rec to target.
You can use the test load option to run sessions when you tune session performance. Eliminate session bottlenecks. Primary Key is a natural identifier for an entity. But as per the requirement if necessary u can use Expression Transformation first. You can use system monitoring tools to view percent CPU usage. Studying performance details. Performance details provide information such as buffer input and output efficiency. Eliminate mapping bottlenecks. 5)How to do performance tuning in Informatica? Excerpts from Infomatica Manuals: The goal of performance tuning is to optimize session performance by eliminating performance bottlenecks. Once you determine the location of a performance bottleneck. Monitoring system performance. or configuring index and key constraints. You can optimize the session strategy and use performance details to help tune session configuration. Fine tune the pipeline logic and transformation settings and options in mappings to eliminate mapping bottlenecks. I/O waits. The most common performance bottleneck occurs when the PowerCenter Server writes to a target database. you have to set some flag and based on those flag you want to filter records. (II) 3)Whether first we should use and why? i)Expression Transformation ii)Filter Transformation As per the informatica documents for bottlenecks in informatica. (III) . we should use Filter Transformation first to remove as much as unwanted records and then allow Expression Transformation to process records based on user requirement. There will be no repeatition of data. you can eliminate the Bottleneck by following these guidelines: Eliminate source and target database bottlenecks. In Primary keys all the values are entered manually by the user which are uniquely identified. increasing the database network packet size. For example. To tune the performance of a session. So in such situation you have to use Expression first. You can configure a test session to read from a flat file source or to write to a flat file target to identify source and target bottlenecks. You can identify performance bottlenecks by the following methods: Running test sessions. eliminate it. You can create a set of information called performance details to identify session bottlenecks..In surrogate key values are generated by the system sequentially(Like Identity property in SQL Server and Sequence in Oracle). first you identify a performance bottleneck. and then identify the next performance bottleneck until you are satisfied with the session performance. They do not describe anything. Need for surrogate key not Primary Key If a column is made a primary key and later there needs a change in the datatype or the length for that column then all the foreign keys that are dependent on that primary key should be changed making the database Unstable Surrogate Keys make the database more stable because it insulates the Primary and foreign key relationships from changes in the data types and length. and paging to identify system bottlenecks.(I) 1)Surrogate comes under which category? 2)Where we use surrogate key?(technically) Surrogate Key is an artificial identifier for an entity. Have the database administrator optimize database performance by optimizing the query.
The center of the star schema consists of a large fact table and it points towards the dimension tables. Hierarchies for the dimensions are stored in the dimensional table itself in star schema Whereas hierachies are broken into separate tables in snow flake schema. the PowerCenter Server returns a value from the lookup table or cache. When the condition is not true. last but not least Star schema is more common than snow flakes 6)What is diffrence between data migration and data warehouse. Uncache You cannot insert or update the cache. a dimension table will have one or more parent tables. i. You cannot use a flat file lookup. You can insert or update rows in the cache as you pass rows to the target. depending on the row type. Static Cache When the condition is true. Have the system administrator analyze information from system monitoring tools and improve CPU and network performance. Star schema utilises less joins than snow flakes so the performance is faster. interviewer confuses candidate by that way (V) 7)What is dynamic and static lookup transformation. In a star schema. It is called a star schema because the entityrelationship diagram between dimensions and fact tables resembles a star where one fact table is connected to multiple dimensions.(IV) Eliminate system bottlenecks. business analysis or tren analysis. When the condition is true. You cannot insert or update the cache. Dyanamic Cache When the condition is true. Remember that DW can be relational but can not be OLTP. Data Warehouse: It is a store or warehouse where we manage the unoperational data (non transactional) for historic usage. You can use a relational or a flat file lookup. a dimension table will not have any parent table while in a snow flake schema. In a star schema every dimension will have a primary key. Snowflake Schema : A snowflake schema is a term that describes a star schema structure normalized through the use of outrigger tables. You can pass updated rows to . the PowerCenter Server either updates rows in the cache or leaves the cache unchanged. 4)What is basic diffrence between star schema and snowflake schema? Star Schema : Star Schema is a relational database schema for representing multimensional data. It is the simplest form of data warehouse schema that contains one or more dimensions and fact tables. the PowerCenter Server returns the default value for connected transformations and NULL for unconnected transformations.e dimension table hierachies are broken into simpler tables. the PowerCenter Server returns a value from the lookup table or cache. the PowerCenter Server returns the default value for connected transformations and NULL for unconnected transformations. The advantage of star schema are slicing down. When the condition is not true. Data Migration: It is a process of migration of data from one database location either relational or non relational to another database location. This indicates that the row is in the cache and target table. These hierachies helps to drill down the data from topmost hierachies to the lowermost hierarchies. You can use a relational lookup only. performance increase and easy understanding of data.
2.. 8)Why to choose oracle or sql server for data warehouse? This is a debatable topic. Use Normalizer with OCCURS as 10.Whenever Counter1 becomes 1 increment Counter2. connect 2 SQs. 5. ---------------------------------------------------------------------------------------- 52) ) i hav 1table containing 1000records i want to load first five records in one target and next 5 records in second target alternatley up2 1000recordss ??????????? 2)i hav 1 table contaning 10records and dat table contaning sum duplicates now i want to load duplicates and original records in two targets??????? 3) how to load the same table 10 times in2 target in one mapping???? plzz help out vth ds quests? ANS: (I) Q1 Use two counter variables in Expression. depending on the row type. I have worked in both SQL and ORACLE. You can use aggregator 2. One should give the reason by analysing all the logics. You can pass inserted rows to the target table. Counter 2 . the PowerCenter Server either inserts rows into the cache or leaves the cache unchanged. From your Source Table. When the condition is not true. both are pros and cons. user requirement etc. financial.(VI) the target table. depends on what kinda requirement you have 1. At Workflow level also you should be able to use a workflow variable and a decision task I guess. 4.Increment it from 1 to 5 and then cycle it back.2) = 0 Hope this serves the requirement There could be multiple approaches to handle this 1. ----------------------------------------------------------------------------------------------53) comparing two tables (II) (III) ..2) = 1 For Target2 . You can connect 10 SQs from your source.If mod(Counter2.one person can not say easily that ORACLe is better than SQL SErver. Counter1 .If mod(Counter2. In first SQ. Depends whats the actual business requirement. Use Unix Script 3. use DISTINCT query to get the unique records and load into first target In 2nd SQ. You can also schedule your workflow to run for 10 times one after another. This indicates that the row is not in the cache or target table. put to query to get the duplicates and load into second table.For this also there are multiple ways . So for Target 1 . Q3 . there can be many reasons like technical.
b.. there could be other ways as well ----------------------------------------------------------------------------------------------------- 54) hi.... After that you can use SQL transformation and write your SELECT query there.txt...C2.txt..txt .i have two employee tables with the same structures but with differnet data . briefly i want to load the data into target is all the data from source(a.3...C2.. ANS: (I) I guess SQL transformation can come to your rescue here.C3 1..txt(source file name) 2 3 4 b............ this is one approach......4 like this up to n.C3 2...txt( " ) . .. please help me asap. now i want to find the max(sal) from two tables and needs to compare tables like(max(sal) of emp1 >max(sal) of emp2) if emp1 having the max(sal) i need to pass one set of records and emp2 has max(sal) i need to pass another set of records..txt ------------C1...n......2..n.txt i want to insert my data into a target as C1 C2 C3 C4 1 2 3 a..txt....txt in a... In the source qualifier...3 in b....... use SELECT COUNT(*) FROM TABLE1 UNION SELECT COUNT(*) FROM TABLE 2 In the filter check the condition count1 > count2.txt ----------C1. i have n number of flat files like a.txt) as well as source file name and if i want to add another n+1 file in source system that will add in the target ANS: .b..
Pull : In this case the ETL server pulls data i.(I) (II) we have option in source flat file " Add currently processed file name as port" you check the option then u will be the achieve the ur desired output.(gets data) from the source system. The junk dimension is simply a structure that provides a convenient place to store the junk attributes.for indirect option check with informatica help 55) what are Push and Pull ETL strategies? 2) Can anyone explain factless fact table with example? 3) what is a Junk Dimension can you explain it with example? ANS: (I) Push and Pull strategies determine how data comes from source system to ETL server.(sends data) to the ETL server. first read ur input files using indirect option in session.then in flat file source properties "add currently procesed files" port u have achive tje desired output.e. Fact table without any measure is called factless fact table.e. Push : In this case the Source system pushes data i. flags and/or text attributes that are unrelated to any particular dimension. dropping and creating indexes 57) source filter condition what is the main difference between source filter which we will give in the source qualifier and filter condition which we will add by overriding the source qualifier query? ANS: (I) Filter . A "junk" dimension is a collection of random transactional codes. hi. A good example would be a trade fact in a company that brokers equity trades (II) (III) 56) Pre and post QSL Overrides Hi Guys can some one help me on what condition we prefer the pre and post SQL Overrides And whats its advantages and dis advantages ANS: (I) In case if you want to do anything on the table before data loading or after data loading we use pre and post sql's for ex.
..Hi. His Question was Source qualifier filter and SQL over ride in source with filter condition. Paluri was not talking about diff between source filter in source qualifier transformation and filer in filter transformation. what r the steps we follow when we do this. It means source Qualifier reads all the data from Database and then based on filter condition it passes the records (V) if you observe the session log . so it will increase the session time. try to run a mapping with SQL over ride with where clause and don't put and any filter.. To import. select the xml which u saved and follow the steps. though the log shows default sql query without any customizations (II) 58) mapping xml doubt how to export total mapping to a xml file. click on export it will save the file in XML format. how to use that xml file in other system. (III) is there any difference in the way how they will execute? i mean both the filter conditions we execute in the database level? Purpose of both is the same but executes in diff levels..(where clause will be the condition which we added in the source filter of the sql transformation) As per my understanding source fileter condition will be add to the select statement first and it will be issued to the database. If you set the filter at SQ level.. go to repository. In source filter u can only give filter condition to filter out unwanted records or for testing purpose... (VIII) if we don't give any filter condition or sql over ride. where as just filter condition executes in Informatica level. select query which is issued to the database will be with the "where clause". where as SQL over ride source qualifer is to customize the default sql which is generated by Source Qualifier. (VI) for your clarity. go to repository. and again run the same with just filter condition ( no sql over ride) then u will understand the diff (VII) If we run with just filter condition ( no sql over ride) then observe the log for the sql queiry which is issued on database. ANS: (I) In the designer... (IV) SQL override execute in Database level. it limits the source data( means it filter the data at source lever so it increase the performance) if you use the filter condition with the filter tranceformation it limits the target rows( means it retrive all rows from the source then it will apply the filter condition and load the data in to the target) it will take some time to retrive the records from source. click on import.. .. Both are the.it will be adding filter condition only. if we over ride the query then it will take the over rided query..
In order to get the output as you mentioned in the table.(II) (III) First you want expand the navigator window. sname... when i import it is saying not a valid xml 59) how to maintain history data in oracle Hi Friends.---------... my source table is having like this sno sname comm ------. sum(comm) from table_name group by sno. Here is the query select sno....---------. sname . click on ok exported as xml file..----------1 anand 2500 2 reddy 3000 3 vinod 1000 Note : This is only in oracle not for Informatica.----------1 anand 500 2 reddy 1000 1 anand 2000 2 reddy 2000 3 vinod 1000 i want target table like this sno sname comm ------. select the which mapping you want export go to menu's and select Export option.. ANS: (I) your question and wat you want to achieve are totally different.
60) Why Union is Active transformation
Hi guys can any body explain why Union is Active transformation? I tried Union but it doesn't delete any duplication records.
(I) UNION I S ACTIVE SOLUTION
FIRST WE know a transformation is said to be active if it changes the rows to pass through it. it combines the two tables jof similar stru. so there is a change in records so it is active. suppose u want to delete duplicate records we use only option . otherwise we use union all for displaying with duplicate records.
1) What is Direct and Indirect in flat file property 2) What are the types of Caches 3) Scenario: I have 3 tables 1st table Emp ID and 2nd table Telephone Number, Address , Location and 3rd table Bank Name , Account num .I want all the 3 tables in one target table (all columns converted to rows .Note table is in Denormalized form) ** Empid is Common in 3 tables Other part of Question Scenario: Can i solve the above problem with Unconnected Lookup. Other part Scenario: If we are using Joiner what is the join condition 4) Scenario : I want to update the target table only, without using Dynamic Cache 5) Scenario : I have used Sql override in lookup ,there are 5 ports from Col A to Col D But I have used override for first 3 columns Col A, Col B, Col C order by Col B and mapping is validated .when i run the mapping the session throws error and Sql overide is not valid. 6) Scenario : I have to join table is it better using the Sql override or lookup or Joiner. Performance wise which is better 7) Doing overrides in SrcQua is it a better performance 8) Scenario: In workflow I have a reusable Session, the same session is reused across other workflow. Any change made in either of the Session does reflect in the other Sessions. 9) Can we make any changes in reusable & Non reusable Session often
10) Scenario : In my mapping the Update Strategy is not Updating 11) Scenario : Is it better filtering the rows in filter or in SrcQua
1> Direct : When we want to load data from 1 flat file only. Indirect: when we want to load data from 2 or more flat files of same data structure. 2> Static Cache, Dynamic Cache, Persistent Cache, Recache Cache, Shared Cache 3> Explain more plz. is there any column common??? 4> Use a filter after dynamic cache and give filter condition NewLookUprow=0 5> After SQL override give "--" without quotes else override wont work. 6> If tables are from same database use Source Qualifier else joiner. Lookup is not a good option. 7> Yes, override in SQL is better. 8> When you make changes to a reusable task definition in the Task Developer, the changes reflect in the instance of the task in the workflow only if you have not edited the instance. 9> Yes, we can make. 10> Check whether you have set the option Treat Source Row as Update in session properties or not. Set it to Update. 11> Better in source quailfier
3> We need to join the three tables using Joiner. We need 2 joiner transformations to join them. Join condition would be empid of one table=empid of another table. We do not need to use unconnected lookup as it is used to update Slowly changing dimension tables.
4> We can use simple lookup to know whether record exists in target table or not. We can the use update strategy. We can also use Unconnected lookup. Other option is to set the Treat Source Row as Update in session properties.
1) How to delete header and footer in flat file. 2) In source we have 1000 rows and i have 3 tragets . The 1st 100 rows have to go in 1st target and the next 200 rows in 2nd traget and the rest of the rows in the 3rd tgt. 3) I have some duplicate rows in the Source table , i have 2 targtes the unique records has to be loaded in the 1st target and the duplicate records in 2nd target. 4) I have Empno,Name ,Loc in the source table , we have two targets the 1st target is Tgt_ india and Tgt_USA. when the employee moves from india to USA the row of that employee must be inserted in USA tgt and must be deleted in Indian tgt and vice versa.
1> Skip the first row to delete header. Not sure for footer. 2> Use an expression t/f after source and use a mapping variable of count aggreation type. Then use a router to filter the records in 3 groups and load to 3 target tables. ( REFER TO FLATFILES FOLDER MAPPING) 3> Already discussed in thread. 4> Insert table using a router to appropriate table and use Post SUCCESS COMMAND to delete the row from other table. (REFER TO FLATFILES MAPPING)
SQL STAEMENT ::
Aim of Informatica :: Informatica is a ETL tool used to extract data from OLTP sources .2 / 7. 2 >delete from emp_india where empno =(select empno from emp_usa).1.1> delete from emp_india a where empno =(select empno from emp_usa b where a= b).1 . applied transforamtion to cleanse the data before applying the business logic and then loaded into the star tables (warehouse) ARCHITECTURE Data Warehouse Architecture Informatica Architecture Version 6.
Clients Server Repository Repository Admin Console Server Repository Informatica Manager Server Designer Workflow Manager Workflow Monitor SCHEMA Sources Source Target / DataWarehouse Users Repository Data warehouse/Datamart 1Star Schema 2Snow Flake Schema 3Fact Constellation or Galaxy Schema DWH CONCEPTS | TYPES OF SCHEMA STAR SCHEMA .
drill-downs and specifying criteria. most of the fact tables in a star schema are in database third normal form. The fact table typically contains records that are ready to explore.to enhance query response times.usually grouped by month and loaded once at the end of a year. Records in the fact table are often referred to as events. Sometimes the values for historical orders are stored in a sales fact table.usually grouped by month and loaded daily. called the fact table and the dimension tables which allow browsing of specific categories. weekly or monthly. while dimensional tables are de-normalized (second normal form). Typical dimension tables in a data warehouse are: time dimension table customers dimension table products dimension table key account managers (KAM) dimension table sales office dimension table Star schema example An example of a star schema architecture is depicted below.report stocks. usually refreshed daily Dimension table Nearly all of the information in a typical fact table is also present in one or more dimension tables. summarizing. inventory fact table .in some cases the table can be split into open orders and historical orders.Star schema architecture is the simplest data warehouse design. The main purpose of maintaining Dimension Tables is to allow browsing the categories quickly and easily. due to the time-variant nature of a data warehouse environment. . budget fact table . Fact table The fact table is not a typical relational database table as it is de-normalized on purpose . there may be some company or business specific fact tables): sales fact table . The primary key for the fact table is a composite of all the columns except numeric values / scores (like QUANTITY. TURNOVER. Typically. The primary keys of each of the dimension tables are linked together to form the composite primary key of the fact table. forecast fact table . The main feature of a star schema is a table at the center. Typical fact tables in a global enterprise data warehouse are (apart for those.contains all details regarding sales orders fact table . In a star schema design. there is only one de-normalized table for a given dimension. usually with ad hoc queries. exact invoice date and time).
Snowflake schemas are generally used when a dimensional table becomes very big and when a star schema can¶t represent the complexity of a data structure.SNOWFLAKE SCHEMA Snowflake schema architecture is a more complex variation of a star schema design. the more complicated SQL joins must be issued to query them. many tables need to be joined and aggregates generated. The main difference is that dimensional tables in a snowflake schema are normalized. the use of snowflake schemas should significantly improve performance by moving out some data to other table (with BRANDS for instance). For example if a PRODUCT dimension table contains millions of rows. This is because in order for a query to be answered. . The problem is that the more normalized the dimension table is. so they have a typical relational database design. An example of a snowflake schema architecture is depicted below.
different fact tables are explicitly assigned to the dimensions. which are for given facts relevant. however it may be hard to manage and support. In that case using two different fact tables on a different level of grouping is realized through a fact constellation model. That solution is very flexible. In a fact constellation schema. The main disadvantage of the fact constellation schema is a more complicated design because many variants of aggregation must be considered. Use of that model should be reasonable when for example. client id and product id. there is a sales fact table (with details down to the exact date and invoice header id) and a fact table with sales forecast which is calculated based on month. This may be useful in cases when some facts are associated with a given dimension level and other facts with a deeper dimension level. which is because it contains multiple fact tables.GALAXY SCHEMA For each star schema or snowflake schema it is possible to construct a fact constellation schema. Normalization . This allows dimension tables to be shared amongst many fact tables. This schema is more complex than star or snowflake architecture.
Ensuring data dependencies. First Normal Form First Normal form(1 NF) sets the very basic rules for an organized database. y y Meet all the requirements of the second normal form.What is Normalization? Normalization is the process of efficiently organizing data in a database. Remove subsets of data that apply to multiple rows of a table and place them in seperate tables. y y Eliminate duplicative columns from the same table Create seperate tables for each group of related data and identify each row with a unique column or set of columns(the promary key) Second Normal Form Second Normal form(2 NF) further addresses the concept of removing duplicative data. Remove columns that are not dependent upon the primary key. y y y Meet all the requirements of teh first normal form. Create relationships between these new tables and their predecessors through the use of foreign keys. There are two goals of the normalizaton process:: Eliminating redundant data. Third Normal Form Third Normal form(3 NF) remove columns which are not dependent upon the primary key. Next Install y y Keep the INFORMATICA CD into the CD Rom Click on CD-Drive .
. . Click on OK Here you added the data source successfully. Set the "Server Port Number" any value between 5002 to 65535.y Click on the icon y y y y Now it is showing prompt box with option Now click on " Power Center for windows" button. Start Services y y y y y Clcik on Control Panel Now click on Administrator Tool> Services You see the prompt display on the screen. Set the "administrative password" of your own choice.We will tell how to cofigure it in the next few lines. Configure Repository Server y y y y y y Go to Program files Informatica power center 7.1.Select the right data source. Add Data Source y y y y y y y Click on Control panel Click on Administrator tool >Data source (ODBC) Now click on add button. Right click on "Informatica" and click on start.You can name it as you want for eg.That you can do it now or later on.It is case sensitive.It shows set of drivers of differnt data sources.For eg. ora_data_src. You see the prompt screen of Repository server with differnt options.1 > Repository Server > Repository Setup. Now we successfully configured Repository Server. Right click on "Informatica Repository Server" and click on start. Successfully installed.. Cick on OK It ask for description of the Data Source. while in the process it asks you to configure Repository Server and Informatica Server . If your database is Oracle then select respected oracle driver.
1. Right click on Informatica Repository Sever and click on new server registration. Create two users one for Repository and other for Target Database. Previous Create Repository (Version 7.1 Informatica Power Center -client > Repository Admin Console Under console root . For eg. Click on the Server (represented with your host name) and clcik on connect. Repository (Username :: Rep_one . Password :: r) and Target (username::trg_wh .1) y y y y y y y y y y y y y y ALl programs > Informatica Power Center 7. Now Right click on the repository folder and click on New Repository .. Prompt comes with a port number. Host name :: your computer name .Make sure the port number and password must be same as the Repository server. It asks you to enter the password . Now your repository keeps running. Create Temporary Server ( Workflow Manager) . Now click on LICENSE tab -. This are necessary to secure the data in the repositories.y successfully started the services.1. Port number :: Server Port number(Repository) Now your system connected with the Repository server. After creating the users. Click on OK It takes some time to create repository.Test it by connecting to the database.enter the keys in the following order Product License Key. Enter the Repository name and select Global Repository (optional) . Create two users y y y y y Login to your database. Password :: t). Opton Key and Connectivity License Key.
(Be careful wen you enter login information) Successfully server created. Back up repository y y y y y y Move the copy of the Repository to a different Server Disable the Repository. click on server and choose "Add server" Now enter the new server name and the host name must be "Computer name" click on OK On menu bar click on "Connections" and choose the relational option. Configure Informatica Server y y y y y y y y y Go to All Programs Informatica Power Center 7. click on update Click on ok REPOSITORY ADMIN CONSOLE Actions Create Local or Global Repository Start Repositories. Now you find a bigger prompt with set of tabs with different options.1 >Informatica Power center -Server > Informatica Server Setup It prompts with a box .y y y y y y y y y Clck on workflow manager (Informatica Client) Start the Infromtica Services Open the reepository On the menu bar . In the first tab "Server" tab Server Name :: Temporary server name TCP/IP Host Name :: Computer Name Click on "Repository" tab Repository Name:: Name of the repository created at console place Repository User :: Username of Repository Repository Password :: Password of Repository Repository Server host Name :: Computer Name Now move to "License" tab Enter Option key .1.click on Continue button.. Now we have to create teo relationals one for Source and the other for Target Database. Notificy Users :: Notification message can be send to all the users connected to the Repository Propagate Register Repositories . click on update Enter Connectivity . Export connection information.
Notificy Users :: Notification message can be send to all the users connected to the Repository Propagate Register Repositories Rstore Repository Upgrade Repository DESIGNER MENU DESIGNER OVERVIEW y y y y y y Design Mappings Represent how to move data from source to target. Views .Versioning etc. Excel ) Relational Data .. Synonyms(Oracle .y y Rstore Repository Upgrade Repository REPOSITORY MANAGER Actions y y y y y y y y y y y Create Local or Global Repository Start Repositories. Debugger . Back up repository Move the copy of the Repository to a different Server Disable the Repository. Many more features like Data Profiling . Sql Server . Export connection information. Design Mapplets Create Reusable and Non-Reusable Transformation Access multiple Repositories and folders templates/tables at a time. Access) XMl Data (XML) COBOL Data DESIGNER MENU Important Designer Tools . Different Data Providers/Sources y y y y Flat File (Note pad . Propagate .
Mapplet Input Transformation is used only when you dont want to use the Source Definition in Mapplet Designer. Mapplet Output Transformation is always used when ever you create Mapplet.Import data from Flat Files/ Relational/ Application/ XML/ COBOL y Warehouse Designer Create Target Definition y Transformation Developer Create Reusable Transformation y Mapplet Designer Create Mapplets (Group of transformation which can be reused in different mappings) y Mapping Designer Create Mappings . 3. Mapplets . .So that you can reuse the group of transformation in multiple mappings. You can use multiple source definitions and source qualifiers to provide source data for a mapping.It consists of Source Definition . Example Mapplet Flows Source > Sorter > Expression > Mapplet Output Mapplet Input > Sorter > Expression > Mapplet Output Advantages y Include source definitions. 4. MAPPLETS When you have to create a Mapplet ? If in a need of particular set of transformations which uses same logic in multiple mappings. You can create Mapplets in Mapplet Designer Tool.y Source Analyzer Create Source Definition . Transformations and Target Definition. Create Mapplet 1.Represents how the data to move from Source to Target Table. 2.
You do not have to connect all mapplet input and output ports in a mapping. Inserts new and changed dimensions. (Shows Current Data) Inserts new and changed dimensions. Contain unused ports. Include multiple transformations. Slowly Changed Dimension Mapping SCD Type 1 Type Slowly Changing Dimension Slowly Changing Dimension Description Inserts new dimensions. Normal Stored Procedure transformation alone can be used. If you want the mapplet to receive data from the mapping. SCD Type 2 /Version Data SCD Type 2 /Flag Current Slowly Changing Dimension SCD Type 2 /Date Range Slowly Changing Dimension SCD Type 3 Slowly Changing Dimension . Overwrites existing dimensions with changed dimensions. You can create a mapplet to feed data to multiple transformations. Creates an effective date range to track changes.y y y y Accept data from sources in a mapping. Pass data to multiple transformations. Flags the current version and increments the primary key to track changes. Inserts new and changed dimensions. Inserts new dimensions. Updates changed values in existing dimensions. Each Output transformation in a mapplet represents one output group in a mapplet. A mapplet can contain as many transformations as you need. Optionally uses the load date to track changes. Limitations y y y Must use a reusable Sequence Generator transformation. Creates a version number and increments the primary key to track changes. you can use an Input transformation to receive source data. Cannot use o Normalizer transformations o COBOL sources o XML Source Qualifier transformations o XML sources o Target definitions o Other mapplets VERSIONING Slowly Changed Dimension y It is a Dimension which slowly changes over a time.
You can use Data Profiling to analyze source data in the following situations: y y During mapping development During production to maintain data quality VERSIONING How to Enable Version Control? y y y y y In Repository Admin Console . and determine information. Select the repository for which you want to enable version control. detect data patterns. It stores multiple versions of an object as you check out. such as implicit datatype. Advantage of Versioning ? A repository enabled for version control maintains an audit trail of version history. You may want to do this for the following reasons: y y Determine what versions are obsolete and no longer necessary to store in the repository.Data Profiling Data profiling is a technique used to analyze source data. you may want to view the object version history. PowerCenter lets you profile source data to suggest candidate keys. and check it in. Choose Action-Properties. As the number of versions of an object grows. modify. Workflows. Click on Ok button. PowerCenter Data Profiling can help you evaluate source data and detect patterns and exceptions. evaluate join criteria. WORKFLOW MANAGER Actions y Create Reusable tasks . Workflow A workflow is a set of instructions that describes how and when to run tasks related to extracting. y Configure tasks. y Schedule Workflows. transforming. Worklets . Troubleshoot changes in functionality between different versions of metadata. and loading data. . Select the Supports Version Control option.
You can choose to run the Command task only if the previous task in the workflow completes. Stops or aborts the workflow. Waits for a specified period of time to run the next task. Once the event occurs. TASKS There are many tasks available . Types of Tasks Task Name Session Command Reusable Yes Yes Description Set of instructions to run a mapping Specifies shell commands to run during the workflow. The workflow that contains the worklet is called the parent workflow. Represents the location of a user-defined event. Specifies a condition to evaluate in the workflow. Waits for a user-defined or a pre-defined event to occur. which are used to create workflows and worklets. the PowerCenter Server completes the rest of the workflow.Worklets A worklet is an object that represents a set of tasks. Use the Decision task to create branches in a workflow. Where to use Worklets? You can run worklets inside a workflow. Use the Worklet Designer to create and edit worklets. You can also nest a worklet in another worklet. Email Control Decision Yes No No Event-Raise No Event-Wait No Timer No . The Event-Raise task triggers the userdefined event when the PowerCenter Server runs the Event-Raise task. When to create Worklets? Create a worklet when you want to reuse a set of workflow logic in several workflows. Sends email during the workflow.
It performs many tasks such as override default sql query . Top or bottom 'N' analysis . Similar to the UNION ALL statement. stored procedures to automate tasks that are too complicated for standard SQL statements. When you add an XML source definition to a mapping. sum . or synonym. abort. You can use the Expression transformation to calculate values in a single row before you write to the target. . min . Split a single row into multiple rows. you need to connect it to an XML Source Qualifier transformation. stop. View details about a workflow or task in Gantt Chart view or Task view. avg (mainly to perform calculation or multiple rows or group) Reads cobol files ( denormalized format). To perform group calculation such as count . view. You can view the log file and Performance Data TRANSFORMATIONS Transformation SORTER RANK JOINER FILTER ROUTER __ __ __ AGGREGATOR __ NORMALIZER __ SOURCE QUALIFIER UNION __ __ __ Active Passive __ Description sorting the tables in ascending or descending and aslo to obtain Distinct records.WORKFLOW MONITOR You can monitor workflows and tasks in the Workflow Monitor. the Union transformation does not remove duplicate rows. filters the rows that do not meet the condition. Join two different sources cmng from different and same location . and resume workflows from the Workflow Monitor. __ EXPRESSION __ LOOK UP __ STORED PROCEDURE __ XML SOURCE QUALIFIER It merges data from multiple sources similar to the UNION ALL SQL statement to combine the results from two or more SQL statements. Actions You can run. It is useful to test multiple conditions . join data from two or more table etc Represents the flatfile or relational data. Use a Lookup transformation in a mapping to look up data in a flat file or a relational table. filtering records .You can call by using Stored Procedure Transformation. max .
You can Reuse a mapping by varing the parameters and Variables. After you save a mapping. Default Values of Mapping Parameter and Variables Data String Numeric Datetime DEBUGGER Actions y Default Value Empty String 0 1/1/1753 A. A mapping parameter retains the same value throughout the entire session. . o After you run a session. you can run the Debugger against the session. Variables A mapping variable represents a value that can change through the session. You can debug a valid mapping to gain troubleshooting information about data and error conditions. The PowerCenter Server saves the value of a mapping variable to the repository at the end of each successful session run and uses that value the next time you run the session. delete. .. Represntation $$parametername/$$variablename Parameters A mapping parameter represents a constant value that you can define before running a session.__ UPDATE STRATEGY To flag rows for insert. If a session fails or if you receive unexpected results in your target. you can run some initial tests with a debug session before you create and configure a session in the Workflow Manager. update.D. Why we have to create Mapping Parameters or Variables ? you can use mapping parameters and variables to make mappings more flexible. or reject. Situation to run the Debugger o Before you run a session.
sal. What is Difference between E-R Modeling and Dimensional Modeling? . less number of joins while retrieving the fact. ename. A fact table consists of measurements of business requirements and foreign keys of dimensions tables as per business rules.Define Data Warehouse ? ³A subject-oriented . OLTP uses the Normalization process Denormalization is the process of allowing redundancies.Emp table has empno. more granularity is achieved with less coding i.e.1. What is junk dimension? What is the difference between junk dimension and degenerated dimension? A "junk" dimension is a collection of random transactional codes. job. integrated . Being in normal form. Why fact table is in normal form? A fact table consists of measurements of business requirements and foreign keys of dimensions tables as per business rules. 5. deptno But We are talking only the column empno. There can just be SKs within a Star schema. which itself is de-Normalized. Now.Differnce between Normalization and Denormalization? Normalization is the process of removing redundancies. The junk dimension is simply a structure that provides a convenient place to store the junk attributes. these columns are formed a dimension is called junk dimension Degenerative dimension: the column which we use in dimension are degenerative dimension Ex. OLAP/DW uses the denormalized process to capture greater level of detailed data (each and every transaction) 4. I would agree.where as A degenerate dimension is data that is dimensional in nature but stored in a fact table. if there were then FKs on the dimensions as well. ename from the EMP table and forming a dimension this is called degenerative dimension 3. flags and/or text attributes that are unrelated to any particular dimension. Junk dimension: the column which we are using rarely or not used. time-variant and non-volatile collection of data in support of management's decision making process´ 2.
which can be used across multiple Data Marts in combination with multiple facts tables accordingly Conformed facts are allowed to have the same name in separate tables and can be combined and compared mathematically. by which system performance will improve. 8. There will b no need to change the metadata of these tables and they can go along with any number of facts in that application without any changes Dimension table. 1. Dimensional modeling is used for de-normalizing the ROLAP/MOLAP design.Basic difference is E-R modeling will have logical and physical model. 6. E-R modeling is used for normalizing the OLTP database design. What are the methodologies of Data Warehousing? They are mainly 2 methods. Conformed dimensions are those tables that have a fixed structure. .Up approach :: Data Mart --> Data Warehouse) Inmon model structured as Normalized structure. In ER modeling the data is in normalized form. What is conformed fact? Conformed dimensions are the dimensions. Strategies are different methods followed to meet the validation requirements. which may adversely affect the system performance. What are data validation strategies for data mart validation after loading process? Data validation is to make sure that the loaded data is accurate and meets the business requirements. Inmon Model. Dimensional model will have only physical model. Whereas in Dimensional Modeling the data is denormalized.Down approach :: Data Warehouse --> Data Mart) Kimball model always structured as Denormalized structure. which is used. Ralph Kimbell Model (Top . 2. 7. by more than one fact table is known as a conformed dimension. Dimensional model / Multidimensional Modeling revolves around Dimensions (point of analysis) for decision-making and not to capture the process. (Bottom . So more number of Joins. Adding to the point: E-R modeling revolves around the Entities and their relationships to capture the overall process of the system. so less number of joins.
data dictionaries. Business Analyst or data modeler usually capture information about data . What are the possible data marts in Retail sales? Product information. attribute/property descriptions. It is useful because the natural primary key (i. Metadata synchronization joins these differing elements together in the data warehouse to allow for easier access.e. varchar. or SQL Server Identity values for the surrogate key. relating and synchronizing data elements with the same or similar meaning from different systems. existance. length. range/domain descriptions. Metadata Synchronization The process of consolidating. subsets.9.k. data type descriptions. Meta data is the data about data. or Oracle sequence. 11. (also known as artificial or identity key). nullable. Metadata is also presented at the Data mart level. What is the main difference between schema in RDBMS and schemas in Data Warehouse? RDBMS Schema * Used for OLTP systems * Traditional and old schema * Normalized * Difficult to understand and navigate * Cannot solve extract and complex DWH Schema * Used for OLAP systems * New generation schema * De Normalized * Easy to understand and navigate * Extract and complex problems can be easily . Metadata is stored in a data dictionary and repository. and navigation services. 10. valid values etc) and behavior of data (how it is modified / derived and the life cycle) in data dictionary a. What is surrogate key? Surrogate key is the primary key for the Dimensional table. nature of data (char. and description of a data element. The repository environment encompasses all corporate metadata resources: database catalogs. Examples of metadata include data element descriptions. Data warehouses typically use a surrogate. Metadata includes things like the name. sales information 12. What is meant by metadata in context of a Data warehouse and how it is important? Metadata or Meta data is data about data. valid values. For a DW user. key for the dimension tables primary keys. In context of a Data warehouse metadata is meant the information about the data. metadata provides vital information for analysis / DSS. They can use Infa sequence generator. This information is stored in the designer repository. Surrogate key is a substitution for the natural primary key. It insulates the data warehouse from changes in the schema of operational systems.a metadata. and process/method descriptions. Customer Number in Customer table) can change and this makes updates more difficult and also used in SCDs to preserve historical data. ODS etc.the source (where and how the data is originated). fact and dimensions.
.What is Dimensional Modeling? solved * Very good model In Dimensional Modeling. let us look at geographical level of granularity. Without careful planning you may create a database that omits data required to create critical reports. primary and foreign keys.. Why is Data Modeling Important? The data model is also detailed enough to be used by the database developers to use as a "blueprint" for building the physical database. These dimensions don't have it's own dimensions. product description etc. which is stored. Level of granularity means the upper/lower level of hierarchy. In this design model all the data is stored in two types of tables . profit etc. 17. in fact table is called degenerate dimensions. 16. Data is stored in two kinds of tables: Fact Tables and Dimension tables. What does level of Granularity of a fact table signify? It describes the amount of space required for a database. level of granularity defines the extent of detail.. In this case. More Granularity implies more aggregation potential and vice-versa. stored procedures. revenue. A poorly designed database will require more time in the long-term. As an example. and triggers. Level of Granularity indicates the extent of aggregation that will be permitted to take place on the fact data.. We may analyze data at the levels of COUNTRY. Dimension table contains dimensional data such as Product Id.. Fact table contains the facts/measurements of the business and the dimension table contains the context of measurements i. What is degenerate dimension table? The values of dimension... 15. 14. In simple terms. sales.. and is unable to accommodate changes in the user's requirements.problems * Poorly modelled 13. up to which we can see/drill the data in the fact table. Fact Table contains fact data e. up to which we can see/drill the data in the fact table.g. How do you load the time dimension? . REGION.Facts table and Dimension table. TERRITORY. produces results that are incorrect or inconsistent.e. product name.. The information contained in the data model will be used to define the relational tables. we say the highest level of granularity is STREET. Level of granularity means the upper/lower level of hierarchy. CITY and STREET. the dimensions on which the facts are calculated. Dimensional Modeling is a design concept used by many data warehouse designers to build their data warehouse.
every month. because the tables are splited in to many pieces. 18. Snowflake schema y y y y It is the normalized form of Star schema. Why do you need Snowflake schema? . Why do you need Star schema? 1) Less joiners contains 2) Simply database 3) Support drilling up options 20. No need to use complicated joins. It would be at the most granular level at which the business runs at (ex: week day. The nature/purpose of the data that is to be feed to the model is the key to your question as to which is better. Star schema y y y y contains the dimension tables mapped around one or more fact tables. Contains in depth joins. Difference between Snowflake and Star Schema. Weekly process gets updated every week and monthly process. The whole set-up is totally denormalized. since we have more tables. It is a denormalized model.In Data warehouse we manually load the time dimension. Every Data warehouse maintains a time dimension. Depending on the data loads. 19. Star schema is a dimensional model with a fact table (large) and a set of dimension tables (small). What are situations where Snow flake Schema is better than Star Schema to use and when the opposite is true? Star schema and snowflake both serve the purpose of dimensional modeling when it comes to data warehouses. Queries results fastly. There will be some delay in processing the Query. We have to use complicated joins. day of the month and so on). We can easily do modification directly in the tables. However in cases where the dimension tables are split to many tables that are where the schema is slightly inclined towards normalization (reduce redundancy and dependency) there comes the snowflake schema. these time dimensions are updated.
D4 are the Facts and Dimensions here D1. Conformed dimensions are those tables that have a fixed structure. For example. 22. What are conformed dimensions They are dimension tables in a star schema data mart that adhere to a common structure.D1. Dimension tables contain the data from which dimensions are created.Some times we used to provide separate dimensions from existing dimensions that time we will go to snowflake Disadvantage Of snowflake: Query performance is very low because more joiners is there 21. regardless of what data mart it is used in your organization. For example: Current Balance and Profit Margin are the facts. A fact table in data ware house is it describes the transaction data.D2. the Calendar dimension is commonly needed in most data marts. What are Semi-additive and faceless facts and in which scenario will you use such kinds of fact tables Semi-Additive: Semi-additive facts are facts that can be summed up for some of the dimensions in the fact table. as it makes sense to add them up for all accounts (what's the total current balance for all accounts in . which is used. you can query by date/time from one data mart to another to another. Current Balance is a semi-additive fact. What is Fact table A table in a data warehouse whose entries describe data in a fact table.D3 and Cube-2 contains F2. What is conformed fact? Conformed dimensions are the dimensions. By making this Calendar dimension adhere to a single structure. 24. but not the others. and therefore allow queries to be executed across star schemas.(cubes are the schemas contains facts and dimension tables) Consider Cube-1 contains F1. Conformed dimentions are dimensions which are common to the cubes.D1. There will b no need to change the metadata of these tables and they can go along with any number of facts in that application without any changes Dimension table.D2 are the Conformed Dimensions 23.D2. which can be used across multiple Data Marts in combination with multiple facts tables accordingly Conformed facts are allowed to have the same name in separate tables and can be combined and compared mathematically. It contains characteristics and key figures. by more than one fact table is known as a conformed dimension.
Data will be loaded directly. Direct load:(Faster Loading) All the Constraints will be disabled. Common examples of factless fact tables include: . it contains only the textual attributes.What are Aggregate tables Aggregate tables contain redundant data that is summarized from other data in the warehouse.Please share with us Click Here. E. These tables will be used to reduce the query execution time. facility.the bank?). What are the Different methods of loading Dimension tables Conventional Load: Before loading the data. monthly sales information. What is a dimension table A dimensional table is a collection of hierarchies and categories along which the user can drill down and drill up. They are often used to record events or coverage information.Questions related to Informatica/ business Objects /Database etc 25.g Yearly.Identifying product promotion events (to determine promoted products that didn't sell) .Retrieving the required data from the actual table. but it does not make sense to add them up through time (adding up all current balances for a given account for each day of the month does not give us any useful information A factless fact table captures the many-to-many relationships between dimensions. Conventional and Direct load method are applicable for only oracle.. which have millions of records will take more time and also affects the server performance. Why are OLTP database designs not generally a good idea for a Data Warehouse . The naming convension is not general one applicable to other RDBMS like DB2 or SQL server. all the Table constraints will be checked against the data. 26. 27.Tracking student attendance or registration events . These are the tables which contain aggregated / summarized data.Mention the name of the company . and equipment schedules for a hospital or university 25.To avoid this we can aggregate the table to certain required level and can use it.This tables reduces the load in the database server and increases the performance of the query and can retrieve the result very fastly. but contains no numeric or textual facts.Later the data will be checked against the table constraints and the bad data won't be indexed.Tracking insurance-related accident events .Identifying building. Previous Next Send ::If you have any Data Warehouse related interview questions of a particular company. Aggregate table contains the summary of existing warehouse data which is grouped to certain levels of dimensions.
a product price changes over time.In surrogate key values are generated by the system sequentially(Like Identity property in SQL Server and Sequence in Oracle). out-of-date.. These are a few examples of Slowly Changing Dimensions since some changes are happening to them over a period of time 29. For Example : You are extracting Customer Information from OLTP Source and after ETL process. incomplete. People change their names for some reason. Yes you can use Primary Key of Source CustomerID as Primary Key in Dimension Table. For instance. What are slowly changing dimensions Dimensions that change over time are called Slowly Changing Dimensions. why primary key not used as surrogate key Surrogate Key is an artificial identifier for an entity. Type 2. data that is incorrect.e. so that indexing and retrieval is much faster. Then you need another unique no apart from CustomerID. redundant. Another reason : If you have AlphaNumeric as a CustomerID. What are Data Marts . What is data cleaning? how is it done? Data Cleansing: the act of detecting and removing and/or correcting a database's dirty data (i. There you have to use Surrogate Key. loading customer information in a dimension table (DW). It is used for storing the details of daily transactions while a datawarehouse is a huge storage of historical information obtained from different datamarts for making intelligent decisions about the organization. 27. What is the need of surrogate key. But if you would like to preserve history of customer in Dimension table i. Need for surrogate key not Primary Key If a column is made a primary key and later there needs a change in the datatype or the length for that column then all the foreign keys that are dependent on that primary key should be changed making the database Unstable Surrogate Keys make the database more stable because it insulates the Primary and foreign key relationships from changes in the data types and length.e. In Primary keys all the values are entered manually by the user which are uniquely identified. 28. Then you have to use surrogate key in Dimension Table. If you take SCD Type 1. They do not describe anything.OLTP cannot store historical information about the organization. or formatted incorrectly) It can be done by using the exisitng ETL tools or using third party tools like Trivillium etc. Primary Key is a natural identifier for an entity. There will be no repeatition of data. 26. It is advisable to have system generated small integer number as a surrogate key in the dimension table. Country and State names may change over time..
Data Mart of Maketing. payroll. Data marts are sometimes complete individual data warehouses which are usually smaller than the corporate data warehouse. production. which is not having any source in oltp It is generated at the time of transaction Like invoice no this is generated when the invoice is raised It is not used in linking and it is also not a fkey But we can refer these degenerated dimensions as a primary key of the fact table A Degenerate dimension is a Dimension which has only a single attribute. 31. Can a dimension table contains numeric values? No. Data Mart of Finance. Data Mart of HR etc. The data items thar are not facts and data items that do not fit into the existing dimensions are termed as Degenerate Dimensions. then it may be divided into the following individual Data Marts. eg: Data Mart of Sales. Data Mart: a data mart is a small data warehouse. if we take a Data Warehouse of an organization.g. Only Fact Table having Numeric Fields. Degenerate Dimensions are used when fact tables represent transactional data. This dimension is typically represented as a single field in a fact table. 30. sales. In general. Degenerated dimension is a dimension. Explain degenerated dimension in detail. 32. unit. e. we have: . for example. Give examples of degenerated dimensions Degenerated Dimension is a dimension key without corresponding dimension.Data Mart is a segment of a data warehouse that can provide data for reporting and analysis on a section. a data warehouse is divided into small units according the busness requirements. Example: In the PointOfSale Transaction Fact table. department or operation in the company. Data Marts are used to improve the performance during the retrieval of data. Degenerate Dimensions are the fastest way to group similar transactions.
In a traditional parent-child database. design the Dimensional tables with the Lower level Grains. 2. Therefore.Therefore.Gathering bussiness requiremnts Identifying Sources Identifying Facts Defining Dimensions Define Attribues Redefine Dimensions & Attributes Organise Attribute Hierarchy & Define Relationship Assign Unique Identifiers Additional convetions:Cardinality/Adding ratios Understand the bussiness requirements. POS Transaction Number is a degenerated dimension. What are the steps to build the data warehouse 1. 5. Product Key (FK). But in this dimensional model. POS Transaction Number looks like a dimension key in the fact table but does not have the corresponding dimension table.Once the business requirements are clear then Identify the Grains(Levels). 3. such as the transaction date and store identifier.Date Key (FK). Store Key (FK). In logical phase data base design looks like Star Schema design so it is named as Star Schema Design 34. Production Dimension corresponds to Production Key. Promotion Key (FP).Once the dimensions and Fact tables are designed define the relation ship between the tables by using primary key and Foreign Key.Once the Dimensions are designed. POS Transactional Number would be the key to the transaction header record that contains all the info valid for the transaction as a whole. 33. What is the different architecture of data warehouse . we have already extracted this info into other dimension.Grains are defined. and POS Transaction Number Date Dimension corresponds to Date Key. design the Fact table With the Key Performance Indicators (Facts). 4.
Architecture: Source --> Staging Area -->Warehouse --> End Users Data warehouse with staging area and data marts Architecture: Customization of warehouse architecture for different groups in the organization then data marts are added and used. Architecture: Source --> Staging Area --> Warehouse --> Data Marts --> End Users Q1>How do u change parameter when u move it from development to production. Ans :: UNIQ <filename> Q5>How to find no. (Posted :: Vijay ) Q3>How do u reset sequence generator value u move it from development to production Ans :: Keep the sequence value as 1 in dev and move the code to prod.(bill Inmon) 2. (Posted :: Vijay ) Q2>How do u retain variable value when u move it from development to production.(Ralph kimbol) There are three types of architectures. Top down . Ans :: We can manually move the parameter file and save in prod server. Architecture: Source --> Warehouse --> End Users Data warehouse with staging area Architecture: Whenever the data that is derived from sources need to be cleaned and processed before putting it into warehouse then staging area is used.of rows commited to the target. Ans :: while moving the variables to prod make sure that you assign the default value while creating the variables in dev environment.1. Ans :: Log file . Date warehouse Basic Architecture: In this architecture end users access data that is derived from several sources through the data warehouse. when a session fails.when the code was moved it will check the repository and see if there is any value for that variable if there is no value then it takes the default value. Bottom up . (Posted :: Vijay ) Q4>How to delete duplicate values from UNIX.
17.(with sequence generator we can generate upto 2 billion values only) Ans :: Create a Stored Procedure in database level and call it using storedprocedure transformation. 2. battery. Nextval>900 and nextval<=1000. 1 . Nextval>100 and Nextval<=200.59 101. 2.19 101.Q6>How to remove the duplicate records from flat file (other than using sorter trans. decode function) > target Expression :: decode(nexval<=100. 2. How can we do using informatica? Ans :: develop a mapping flow Source > sorter > sequencegenerator (generate numbers)> expression (batchnumber . and mapping variables) Ans :: (i)Dynamic Lookup (ii) sorter and aggregator Q7>How to generate sequence of values in which target has more than 2billion records.0 ) Q9>Lets take that we have a Flat File in the Source System and It was in the correct path then when we ran the workflow and we got the error as "File Not Found". 3. 3. battery. Q8>I have to generate a target field in Informatica which doesn exist in the source table. battery. what might be the reson? Ans :: Not entered ³source file name´ properly at the session level Q10>How to load 3 unstructured flat files into single target file? Ans :: Indirect file option (configure at session level) Q11>There are 4 columns in the table target Definition :: Store_id. 2. 2. Nextval>200 and nextval<=300. 3. Qty. battery. nextval . 1 . battery.59 Source Definition:: 101. Price 101.nextval+10. battery. The first 100 rows should have the same batch number 100 and the next 100 would have the batch numbe 101 and so on. battery.99 101.nextval+2 . Item. It is the batch number. 2.34 How can we do this using Aggregator? .Nextval+1.99 101. 3. 2. There are 1000 rows altogether.19 101.
Ans :: Source > aggregator (group by on store_id .? how to slove it? Ans :: (i)if source is flat file you cannot use this feature in source qualifier (ii)In case if u are using the realational file as source and if u forget make the connection to the next transformation from source qulaifier .. Q12> in the source qualifer if the default query is not generated.u cannot gerate SQL query ABBREVIATIONS ASCII BI BO BPM C/S DBA DBMS DDL DM DSN DSS DW ERD ERP ETL American Standard Code for Information Interchange Business Intelligence Business Object Business Process Modeling Client/Server Database Administrator Database Management System Data Definition Language Data Modeling Data Source Name Decision Support System Data Warehouse Enterprise Relationship Diagram Enterprise Relationship Planning Extract Transformation Loading .. what is the reason... qty ) > target Tip :: aggregator will sort the data in descending order if u dnt use sorted input. item .
GB GUI HOLAP HTML JDBC MB MDBMS MOLAP ODBC ODS OLAP OLTP OS PCS QA RDBMS ROLAP SCD SQA SRS TB TCP/IP VPN XML Giga Bytes Graphical User Interface Hybrid OnLine Analytical processing Hyper Text Markup Language Java Database Connectivity Mega Bytes Multi-dimensional Data Base Management System Multi-dimensional On Line Analytical Processing Open Data Base Connectivity Operational Data Store On Line Analytical Processing On Line Transaction Processing Operating System Power Center Server Quality Assurance Relational Data Base Management System Relational On lIne Analytical Processing Slowly Changing Dimension Software Quality Assurance Software Requirement Specification Tera Bytes Transmission Control Protocol/Internet Protocol Virtual Private Network eXtensible Markup Language .
Creating log files: Loadmanger creates logfile contains the status of session. Reading the parameter file: If the session uses a parameter files. DTM means data transformation manager.in informatica this is main back ground process.When you configure the session the loadmanager maintains list of list of sessions and session start times.loadmanager reads the parameter file and verifies that the session level parematers are declared in the file Verifies permission and privelleges: When the sesson starts load manger checks whether or not the user have privelleges to run the session.it run after complition of load manager. Informatica integration server maintains two types executing process.in this process informatica server search source and tgt connection in repository if it correct then informatica server fetch the data from source and load it to target. Manages the session and batch scheduling: Whe you start the informatica server the load maneger launches and queries the repository for a list of sessions configured to run on the informatica server.I creates the master thread. Locking and reading the session: When the informatica server starts a session lodamaager locks the session from the repository.Locking prevents you starting the session again and again.When you sart a session loadmanger fetches the session information from the repository to perform the validations and verifications prior to starting DTM process.it creates the DTM process.Master thread creates and manges all the other threads. What is DTM process? ANS: After the loadmanger performs validations for session.Data transfer Manager DTM is to read from Source by using the Readers thread and loading the data into target by using writer thread What are the different threads in DTM process? ANS: . Load manager 2.What are the tasks that Loadmanger process will do? ANS.DTM is to create and manage the threads that carry out the session tasks. 1.
It reads data from source. Reader thread: One thread will be created for each partition of a source. Session detail file: This file contains load statistics for each targets in mapping. Transformation thread: It will be created to tranform data.Session detail include information such as table name.U can view this file by double clicking on the session in monitor window Performance detail file: This file contains information known as session performance details which helps you where performance can be improved. Control file: Informatica server creates control file and a target file when you run a session that uses the external loader. What are the out put files that the informatica server creates during the session running? ANS: Informatica server log: Informatica server(on unix) creates a log for all status and error messages(default name: pm. Two types of datamovement modes avialable in informatica:ASCII mode Uni code mode. Writer thread: It will be created to load data to the target.U choose the datamovement in the informatica server configuration settings.number of rows written or rejected.creation of sql commands for reader and writer threads. Reject file: This file contains the rows of data that the writer does notwrite to targets.Fectchs session and maping information.Master thread: Creates and manages all other threads Maping thread: One maping thread will be creates for each session. What are the data movement modes in informatcia? ANS: Datamovement modes determines how informatcia server handles the charector data.errors encountered and load summary.The control file contains the information about the .log).server.It writes information about session into log files such as initialization process.The amount of detail in session log file depends on the tracing level that you set. These files will be created in informatica home directory:Session log file: Informatica server creates session log file for each session. Pre and post session threads: This will be created to perform pre and post session operations.To genarate this file select the performance detail option in the session property sheet. It also creates an error log for error messages.
What is polling? ANS: It displays the updated information about the session in the monitor window.You can configure the informatica server to create indicator file. ----------------------------------------------------------------------------------------------------------------------------------------Can you copy the session to a different folder or repository? ANS:Yes. Indicator file: If you use the flat file as a target. output file: If session writes to a target file.One if the session completed sucessfully the other if the session fails. Post session email: Post session email allows you to automatically communicate information about a session run to designated recipents.the indicator file contains a number to indicate whether the row was marked for insert.target flat file such as data format and loading instructios for the external loader. For the following circumstances informatica server creates index and datacache files:Aggreagtor transformation Joiner transformation Rank transformation Lookup transformation In which circumstances that informatica server creates Reject files? ANS: When it encounters the DD_Reject in update strategy transformation. .delete or reject. The monitor window displays the status of each session when you poll the informatica server. Violates database constraint Filed in the rows was truncated or overflowed.the informatica server creates the target file based on file prpoerties entered in the session property sheet. By using copy session wizard You can copy a session in a different folder or repository. If target folder or repository is not having the maping of copying session .update. But that target folder or repository should consists of mapping of that session. Cache files: When the informatica server creates memory cache it also creates cache files. You should have to copy that maping first before you copy the session.For each target row.You can create two different messages.
Following are user defined session parameters:Database connections Source file names: use this parameter when you want to change the name or location of session source file between session runs.represent values you might want to change between sessions such as database connections or source files. What are the session parameters? ANS: Session parameters are like maping parameters. Can you copy the batches? ANS: NO.If you have several independent sessions You can use concurrent batches which runs all the sessions at the same time. What is parameter file? ANS: Parameter file is to define the values for parameters and variables used in a session. Server manager also allows you to create userdefined session parameters. You can define the following values in parameter file:Maping parameters Maping variables session parameters. Reject file name : Use this parameter when you want to change the name or location of session reject files between session runs.A parameter file is a file created by text editor such as word pad or notepad. If you have sessions with source-target dependencies you have to go for sequential batch to start the sessions one after another. .What is batch and describe about types of batches? ANS: Grouping of session is known as batch. Batches are two types:Sequential: Runs sessions one after the other Concurrent: Runs session at same time. Target file name : Use this parameter when you want to change the name or location of session target file between session runs.
choose server configure database connections.To do this go to server manger . If u are target consists key constraints and indexes u slow the loading of data. Flat files: If u?r flat files stored on a machine other than the informatca server.Increase the session performance by following. Run the informatica server in ASCII datamovement mode improves the session performance. Partittionig the session improves the session performance by creating multiple connections to sources and targets and loads data in paralel pipe lines. move those files to the machine that consists of informatica server.Unicode mode takes 2 bytes to store a character. In some cases if a session contains a aggregator transformation . .targets and informatica server to improve session performance.So concurent batches may also increase the session performance.Moving target database into server system may improve session performance. The performance of the Informatica Server is related to network connections. Running a parallel sessions by using concurrent batches will also reduce the time of loading the data. You can run the multiple informatica servers againist the same repository.To improve the session performance in this case drop constraints and indexes before u run the session and rebuild them after completion of session. Removing of staging areas may improve session performance.Performance tuning in Informatica? ANS:The goal of performance tuning is optimize session performance so sessions run during the available load window for the Informatica Server.You can use incremental aggregation to improve session performance. If the sessioin containd lookup transformation You can improve the session performance by enabling the look up cache. Aviod transformation errors to improve the session performance.Distibuting the session load to multiple informatica servers may improve session performance. Thus network connections ofteny affect on session performance. We can improve the session performance by configuring the network packet size. Data generally moves across a network at less than 1 MB per second. single table select statements with an ORDER BY or GROUP BY clause may benefit from optimization such as adding indexes. Staging areas: If u use staging areas u force informatica server to perform multiple datapasses. whereas a local disk moves data five to twenty times faster. If U?r session contains filter transformation . Also. If a session joins multiple source tables in one Source Qualifier. optimizing the query may improve performance.create that filter transformation nearer to the sources or You can use filter condition in source qualifier. Relational datasources: Minimize the connections to sources .which allows data to cross the network at one time.So aviod netwrok connections.Because ASCII datamovement mode stores a character value in one byte.
Let me know if you required any further information. how can i generate list file. check the below answers. 4) In mapping we can get the file name including field values. 1) In a single flat file. and must be different from the escape ." in the field.then how can i handle this case. 3)List of files will be appended in file list file using UNIX Script. 1) Not possible to load multiple delimiters.if am having n number of flat files of similar structure? 4)While loading multiple flat files using Indirect loading.Increase the temporary database space also improves the performance. Once all the files are FTP to source Directory. Informatica can't handle this case.Because they must group data before processing it. 2)Informatica can't consider that field value with '. " delimited flat file and while loading data one of my field or attribute is having this ".i want to load data into one target and the file names into one target? How u can do this? ANS: 1) According to my knowledge.x Onwards it possible 1) Enable the below property Column Delimiters : One or more characters used to separate columns of data. 2) From Informatica 8.Rank and joiner transformation may oftenly decrease the session performance . Request the source data team to send in only single delimiter.Aggreagator. if am having multiple delimiters. Delimiters can be either printable or single-byte unprintable characters.To improve session performance in this case use sorted ports option..? 3)while loading multiple flat files using Indirect loading method. connect that file name field into sorter or aggregator to get the distinct file names and load into filename target table or file. Either request source system to change the delimiter to some other delimiter. remaining fields other than filename connect to other target.'.then how can i load the flat file? 2)If a flat file is " . Unix script will generate the file list with all the files exists in the source file directory..
6. In the Delimiters dialog box. the Integration Service ignores the commas within the quotes and reads the row as four fields. a source file uses a comma as a delimiter and contains the following row: 342-3849. the Integration Service ignores delimiter characters within the quote characters. . Maximum number of delimiters is 80. the Integration Service uses quote characters to escape the delimiter. Single Quote. To enter a single-byte unprintable character.character and the quote character (if selected). If you select the optional single quote character. Smith. For example. Therefore. If you select a quote character. 4)Enable the Below Property Add Currently Processed Flat File Name Port. 2) Enable the Below Property Optional Quotes Select No Quotes. The Designer adds the CurrentlyProcessedFileName port as the last column on the Columns tab. or Double Quotes. The CurrentlyProcessedFileName port is a string port with default precision of 256 characters. Jenna . select an unprintable character from the Insert Delimiter list and click Add. click the Browse button to the right of this field. You cannot select unprintable multibyte characters as delimiters. MD . Rockville.