from Source 100 rows are coming, on target there are 5 m
rows which options is better to match data
1. Joiner
2 No cache
3. Static
4. Dynamic
Answers were Sorted based on User's Feedback
Answer / coolboy
Hi,
Here we will use joiner for better performance.We will join
the two sources making source table as master source.so
only 100 comparisons will be done.So it will be very faster.
Whereas in static and dynamic we have to look up on the
target which is very large 5m rows.So caching will take
more time.
| Is This Answer Correct ? | 28 Yes | 3 No |
Answer / arnab
Lookup with No cache will make it faster , personal opinion
though , there are other ways do it also , if you have a
common key between the source and the target and they are
on the same database , it would be easier and faster to do
a join in the source qualifier itself and let the database
handle the query rather than Informatica server
| Is This Answer Correct ? | 6 Yes | 3 No |
Answer / vijaykumar
Answer 7# is wrong because here we have only 1 source. It is
correct in case of 2 sources.
| Is This Answer Correct ? | 2 Yes | 0 No |
Answer / jaybrata
Unless you want to join in Database, the only option considering performance is - USE NO CACHE.
BUT, you have to have proper index defined on your target database table, this is crucial as otherwise there will be 100 full table scans on 5 million records.
Static Cache is the worst option as it will always cache 5 million records just to compare 100 records.
Dynamic cache is not an option assuming 100 source records do not contain duplicate.
Informatica side joining is also not good, better to join in DB side instead (To know why and to check the comparison between Infa side and DB side joins, check www.dwbiconcepts.com)
| Is This Answer Correct ? | 1 Yes | 0 No |
Answer / jaspreet banga
Its better to use connected dynamic non-persistent lookup
transformation, to lookup into the traget and if record
exist , update it else insert the new using unconnected
lookup tranmsformation..
| Is This Answer Correct ? | 1 Yes | 0 No |
Answer / ajay bhosal
Uncached lookup can give you good performance here but only when your target is properly indexed. Moreover, it's not a good idea to hit the database 100 times and scan huge data volume in the database.
Similar or better performance can be achieved by using Joiner with Sorted Input, but ensure that 100 rows are in master pipeline and the detail pipeline is sorted.
There is no question of using cached lookup.
Answer will be Joiner.
| Is This Answer Correct ? | 1 Yes | 0 No |
Answer / srik
5m => 5 million recors.
1) Since target side records are very huge compared to
the source, caching is not preffred. so No caching.
2) Again with no caching we can accomplsh this in 2 ways
a) By Joiner
b) Unconnected lookup
but as unconnected lookup is not in answers is in the
option, Joiner is the answer
| Is This Answer Correct ? | 2 Yes | 2 No |
Answer / prakash
USE NOCACHE,IT WILL GIVE BETTER PERFORMANCE RATHER THAN
JOINER
| Is This Answer Correct ? | 1 Yes | 1 No |
What are the components of the workflow manager?
I want skip first 5 rows to load in to target? what will be the logic at session level ??
Write the advantages of partitioning a session?
write a query to get maximum salary from the employers table without duplicates....kindly help me
why we are using level option in normalizer transformation
How do you handle two sessions in Informatica
How can yoU improve session performance in aggregator transformation?
What happen when you enable grid option avilable at session level ? (Ans found: Scalabily . A single session Parallelization) But how can one session is sharable among different nodes at the same time while running ?
What is the role of informatica in project ?Like i showed financial projects?
My source data like... Empid Name 10 chandra 10 sekhar I am expecting result is Empid Ename 10 Chandrasekhar How can we solve this prob?
i have oracle table A and target B. i don't know how many records. i want get get last record in table A as first record in target table B. write a sql query?
I HAVE 10 records in source. I want to store i record in target 1, second record in target 2, third record in target 3, 4 th record again in target 11, 5th again in target 2, 6th again in target3 etc.... how it is possible?