explain the scenario for bulk loading and the normal
loading option in Informatica Work flow manager ???
Answers were Sorted based on User's Feedback
Answer / rekha
NORMAL LOAD : IT LOADS THE RECORD ONE BY ONE AND WRITES LOG
FOR EACH FILE . IT TAKES MORE TIME TO COMPLETE
BULK LOAD : I LOAD THE NUMBER OF RECORDS AT A TIME ,IT WONT
FALLOW ANT LOG FILES OR TRACE LEVELS . IT TAKES LESS TIME
USE THE BULK MODE FOR IMPROVING THE SESSION PERFORMANCE
| Is This Answer Correct ? | 49 Yes | 4 No |
Answer / addy
Hi
I would like to mention- apart from logging differences
there is a crucial diffrence - if we are running the session
in bulk target load mode we cant recover a session from the
same point when we run it for the next time.
whereas if we are running the session in normal target load
mode we can recover the session if the output being loaded
to the target is deterministic in nature - this is relay a
helpful feature with the applications handling RealTime data.
-addy
| Is This Answer Correct ? | 31 Yes | 4 No |
Answer / shashank
in the normal loading the taget write all the row on the
database log , while laoading the bulk loading the database
log is not come in the picture (that mean its skip the
property )so when the session got failed we can easily find
recover the seesion by the help of data base log.but in
case of bulk loading we can do .
but normaol loading is very slow as compare to bulk laoding.
| Is This Answer Correct ? | 14 Yes | 3 No |
Answer / sujith
Above 5,6&7 are the correct for this questions....
| Is This Answer Correct ? | 8 Yes | 1 No |
Answer / thiru
In normal loading it will creates the log file before
loading target.it will takes the time but in this sessiom
recovery is available.
In bulk loading the integration service bypasses the log
file,direct load into the target.inthis there is no session
recovery available but performance increase.
| Is This Answer Correct ? | 7 Yes | 2 No |
Answer / sankar
NORMAL LOADING:THE INTEGRATION SERVUCE CREATE THE DATA BASE
LOG BEFORE LOADING DATA INTO THE THE TARGET DATA BASE.SO
--THE INTEGRATION SERVICE PERFORM ROLL BACK AND SESSION
RECOVERY.
BULK LOADING:THE INTEGRATION SERVICE IN WORK THE BULK
UTILITY WHEN BYPASS THE DATA BASE LOG
--THIS IS IMPROVES THE PERFORMANCE DATA LOADING
--THIS IS NOT PERFORM ROLLBACK
| Is This Answer Correct ? | 6 Yes | 2 No |
1)Bulkload & Narmal load
Normal: In this case server manager allocates the
resources(Buffers) as per the parameter settings. It creates
the log files in database.
Bulk: In this case server manager allocates maximum
resources(Buffers) available irrespective of the parameter
settings. It will not create any log files in database.
In first case data loading process will be time taking
process but other applications are not affected. While in
bulk data loading will be much faster but other application
are affected.
| Is This Answer Correct ? | 12 Yes | 13 No |
Answer / jyothsna katakam
when you select the normal, it will check the p.k and f.k
relation ship while running the mapping but when you select
the bulk it wont check any p.k and f.k relation ship
| Is This Answer Correct ? | 7 Yes | 27 No |
Explain the use of aggregator cache file?
In a mapping i have three dimensions. If i want to pass a same surrogate key value to all the three dimensions by using one sequence generator is possible?If the mapping is containing single flow? And in the same case if the mapping is contaning 3 flows for the three dimensions then by using one sequence generator can we populate surrogate key (same value) to all the three dimensions?
if the column name in sql override is diff from the column name in source qualifier,what is the impact?
State the limitations where we cannot use joiner in the mapping pipeline?
How can u stop a batch?
suppose we will take flatfile target and load type is bulkload the session fails why please give me the answer
What is the format of informatica objects in a repository?
Draw a workflow diagram for this situation. If s3 is true session s4 and s5 should concurreently Run.if any one s4 or s5 true it should run s1.If s1 sucess it should execute s2 otherwise exit.can any one give me suggestion.Thanks in advance.
following source name gender ramya female ram male deesha female david male kumar male i want the target male female ram ramya david deesha kumar any body give solution above question?
Can we update the data in flat file using update strategy?
What is aggregate awareness?
The structure of source file is as below: Source structure(two fields) Name, Card NUmber A, 111111111(SSN) A, 01010101(Creditcard number) A, 34343434(Debit card number) B, 55555555(Creditcard number) C, 77777777(Debit card number) Target Structure(4 fields) Name,Credit card,SSN,Debit card A,01010101,111111111, 34343434 B,55555555,, C,,,77777777 Corresponding to one name there can be maximum 3 rows and minimum zero rows. Given that I do not know which record might have a particular type of number. How can I handle above requirement with informatica transformations?