wt is Dynamic lookup Transformation?
when we use?how we use?
Answers were Sorted based on User's Feedback
Answer / ashis
When we use dynamic lookup the process is as follows:
1. When the first record is read from the lookup
table/flatfile, server keeps the value in the RAM also the
same information is written into two auto generated files
(having random names, forgot the type, hope it is something
related to index and cache).
2. When the next record comes for lookup, first the record
is searched in the RAM or created file based on the
conditions used in lookup condition.
3. If not found then searched in the Lookup table.
4. In case of dynamic lookup once value is fetched from the
source with modified fields it will be updated in the RAM
or autogenerated file for further lookup for same key
values, before the completion of the session.
(Example:- in case if emp_id 101 got modified more than one
time between the two runs).
5. In case of static lookup. It will always look into the
already existing value in the lookup table.
| Is This Answer Correct ? | 24 Yes | 6 No |
Answer / padmavathi
What Ashis told is Exactly correct apart of this Actually
when lookup table contains less data then we will go for
the Dynamic Lookup cache, performance increases.If large
data is there means its better to go for static cache.
| Is This Answer Correct ? | 20 Yes | 6 No |
in realtime which situations u can use unconnected lookup transformation
What is the command used to run a batch?
What are junk dimensions?
Hi experts, can anyone tell how much we use plsql in real time
Hi, Can someone send me the DWH and Informatica FAQ's at vanibv6@gmail.com Thanks in Advance, Vani
what is the difference between lookupoveride and joiner?
Where is the cache stored in informatica?
how to run the batch using pmcmd command
There are 2 files, Master and User. We need to compare 2 files and prepare a output log file which lists out missing Rolename for each UserName between Master and User file. Please find the sample data- MASTER.csv ---------- Org|Tmp_UsrID|ShortMark|Rolename ---|---------|----------|------------ AUS|0_ABC_PW |ABC PW |ABC Admin PW AUS|0_ABC_PW |ABC PW |MT Deny all GBR|0_EDT_SEC|CR Edit |Editor GBR|0_EDT_SEC|CR Edit |SEC MT103 GBR|0_EDT_SEC|CR Edit |AB User USER.csv -------- Org|UserName|ShortMark|Rolename ---|--------|---------|------------ AUS|charls |ABC PW |ABC Admin PW AUS|amudha |ABC PW |MT Deny all GBR|sandya |CR Edit |Editor GBR|sandya |CR Edit |SEC MT103 GBR|sandya |CR Edit |AB User GBR|sarkar |CR Edit |Editor GBR|sarkar |CR Edit |SEC MT103 Required Output file: --------------------- Org|Tmp_UsrID|UserName|Rolename |Code ---|---------|--------|------------|-------- AUS|0_ABC_PW |charls |ABC Admin PW|MATCH AUS|0_ABC_PW |charls |MT Deny all |MISSING AUS|0_ABC_PW |amudha |ABC Admin PW|MISSING AUS|0_ABC_PW |amudha |MT Deny all |MATCH GBR|0_EDT_SEC|sandya |Editor |MATCH GBR|0_EDT_SEC|sandya |SEC MT103 |MATCH GBR|0_EDT_SEC|sandya |AB User |MATCH GBR|0_EDT_SEC|sarkar |Editor |MATCH GBR|0_EDT_SEC|sarkar |SEC MT103 |MATCH GBR|0_EDT_SEC|sarkar |AB User |MISSING Both the files are mapped through Organization, Shor_mark. So, based on each Organization, Short_Mark, for each UserName from User.csv, we need to find the Matching and Missing Rolename. I am able to bring Matching records in the output. But really I don't find any concept or logic to achieve "MISSING" records which are present in Master and not in User.csv for each UserName. Please help out guys. Let me know if you need any more information. Note:- In User.csv file, there are n number of Organization, under which n number Shortmark comes which has n number of UserName.
What's the layout of parameter file (what does a parameter file contain?)?
How to eliminate 1st and last rows from the source and load the inbetween rows.
why normalizer transformation can not used in mapplet? plzzzz answer