to improve the performance of aggregator we use sorted input option and use sorter t/r befor aggregator. But here we are increasing one more cache in our mapping i.e; sorter. So how can u convince that you are increasing the performance.?
Answers were Sorted based on User's Feedback
Answer / koti
To do aggr. caliculations function it needs some time.. we
can reduce that time by providing sorted I/P. The time is
less which can takes to forward the rows from sorter to
aggregator rather than the time to do aggr. calculations
without sorted data.....
| Is This Answer Correct ? | 2 Yes | 0 No |
Answer / bhupal
if u wont use the sorted input the aggregator transformation will reads the records and keep until all the calculations are done.
if we provided the sorted input it reads the records and release the records reaching next first record of new group.
the integration service will creates the cache for only the sorter keys and only one cache in sorter t/r. so by providing the sorted input we get the performance advantage.
| Is This Answer Correct ? | 1 Yes | 0 No |
Answer / james
Sorted input in an aggregator transformation does not use
cache instead it uses the memory. Reading from memory is
faster than reading from cache.
| Is This Answer Correct ? | 0 Yes | 1 No |
What is the capacity of power cube?
what is dynamic cashe
Examples of Fatal & Non-Fatal error?
How to delete the (flat file) data in the target table after loaded.
we have 30 wf with 3 batches means 1batch=10 wf 2 batch=10-20 wf 3batch=20-30wf through unix First you have to complete batch ofter second batch have to run how can you do in unix?
How to convert a row into column and a column into rows? Name all DTM threads. What all threads stop when we issue STOP or ABORT? How to pass the value of a data(variable kind of) from one session ( generated in mapping) to another session in the same workflow... What are the tyoes of partitioning you know and how to apply them in real time ... Can partitioning be applied to expression transformation and how
what are the limitations for bulk loading in informatica for all kind of databases and transformations?
can u explain me banking domain projects ? and in banking domain projects how many dimension tables and how many fact tables and how to load source and targets in mapping level plese expain give me one example?
1 Answers Accenture, Cognizant, HCL, IBM, iGate, L&T, TCS, Virtusa,
if we are extracting 600000 records from a source table so how much volume of records frequently we load in warehouse. (Iknow its depend on client requirement) still i want to know how much volume of records store in warehouse.
What is the use of an aggregator cache file?
What is informatica?
Quickly characterize reusable change?