what is the difference between normal & bulk loading? and
which one is recommended ? why?
Answer Posted / murthy
normalloding:
this is the optin which we to set in the session under
mapping tab.this will represents the loading of data into
target.this is slower when compared to bulk load.this can
load the data even if u have constraints built on u r
target.this will create the session logs
bulloading
it is the faster loading method.we can' load if u have
constraint defined.this will not create sesion logs
generally normal loading is preffered
| Is This Answer Correct ? | 7 Yes | 2 No |
Post New Answer View All Answers
where are the log files generally stored.Can you change the path of the file.what can the path be?
Just Completed the informatica Certification ,for Dumps please email me on infamaddy@gmail.com i will send it to youÂ
Suppose I am loaded with questions as I am an experienced etl coder, but not an analytical report builder. I am using analysis services to build a cube but am trying to choose the reporting architecture. Can someone please confirm whether or not reporting services (using the business intelligence report) will allow for slicing and dicing, or is it only a static report builder. Also, if I am using the cube browser in analysis services, where can I put the option of non empty so that I dont see records that are blank.?
What is virtual Data Warehousing?
Hi Guys, my requirement is to load 1 yr data from oracle to teradata database through abinitio etl graph.i need to schedule this as a monthly job.but i want to load 12 month data for the first time and from second time it should load last (recent) month data and delete 13th month data. For example For the first time it should load 201208 to 201108 data, second time it should delete 201108 month data and load 201209 month data. I developed garph to load 12 month data ,but i don't know how to handle above requirement as am new to abinitio. i greatly appreciate you help.Thank You.
How can you use an Oracle sequences in Informatica ? You have an Informatica sequence generator transformation also. which one is better to use?
how do u really know that paging to disk is happening while u are using a lookup transform?
how would u estimate the size of Lookup transform data and index cache?
Explain about round-robi?
5. How many repositories can we create in Informatica? 6. What is surrogate key? 7. What is difference between Mapplet and reusable transformation? 8. What is aggregate awareness? 9. Explain reference cursor? 10. What are parallel querys and query hints? 11. DWH architecture? 12. What are cursors? 13. Advantages of de normalized data? 14. What is operational data source (ODS)? 15. What is meta data and system catalog? 16. What is factless fact schema? 17. What is confirmed dimension? 18. What is the capacity of power cube? 19. Difference between PowerPlay transformer and power play reports? 20. What is IQD file? 21. What is Cognos script editor? 22. What is difference macros and prompts? 23. What is power play plug in? 24. Which kind of index is preferred in DWH? 25. What is hash partition? 26. What is DTM session? 27. How can you define a transformation? What are different types of transformations in Informatica? 28. What is mapplet? 29. What is query panel? 30. What is a look up function? What is default transformation for the look up function? 31. What is difference between a connected look up and unconnected look up? 32. What is staging area? 33. What is data merging, data cleansing and sampling? 34. What is up date strategy and what are th options for update strategy? 35. OLAP architecture? 36. What is subject area? 37. Why do we use DSS database for OLAP tools?
Identifying the bottlenecks in various components of Informatica and resolving them.
how many steps are there in etl process?
What are the different Lookup methods used in Informatica?
what is the meant by Normalization and de-normalization?
what are the concerns of OLTP and DSS systems?