I have source table 100 records after run session successfully but loads the target table 50 records only what happened that 50 records and how can i load remaining 50 records
Answers were Sorted based on User's Feedback
Answer / shiva
just truncate table once and run the session again (or)
in session properties just checke the recovery stratagy as
resume from last check point
if i am wrong plese currecct me
| Is This Answer Correct ? | 4 Yes | 1 No |
"Restart the session" . if the integration service has not issued at least one commit.
(or)
perform "Session Recovery" if the integration service has issued atleast one commit.
the I.S reads row id of last row commited from OPB_SRVR from repository table.
the I.S reads all the data from source , start processing from next row id..
we can perform session recovery when only target load type is normal. it can not support in bulk loading..
| Is This Answer Correct ? | 4 Yes | 2 No |
Answer / guest
use dynamic lookup cache so that u can load the data after
load 50 records
| Is This Answer Correct ? | 1 Yes | 0 No |
Answer / sunny
Did you check if there is any filter in between ?
| Is This Answer Correct ? | 0 Yes | 0 No |
Answer / lakshmi
If no errors, check if you have set "test rows" to 50 in session property.
| Is This Answer Correct ? | 0 Yes | 0 No |
How to create Target definition for flat files?
Can a joiner be used in a mapplet.
If you want to create indexes after the load process which transformation you choose?
what is the size of u r project? and what is the size of source file?
What is the difference between stop and abort in informatica
in flatfile target how can u create header and footer plese give me the comands how will u write in session properties
What is the procedure to load the fact table.Give in detail?
In development project what is the process to follow for an etl developer from day1
What happens if the info. Server doesn't find the session parameter in the parameter file?
What are variable ports and list two situations when they can be used?
generate Unique sequence numbers for each partition in session with Unconnected Lookup ? Hi All, Please help me to resolve the below issue while Applying partitioning concept to my Session. This is a very simple mapping with Source, Lookup , router, and target. I need to Lookup on the target and compare with the source data, if any piece of data is new then Insert, and If any thing change in the existed data then Update. while Inserting the new records to the target table I'm generating sequence numbers with Unconnected lookup, by calling the maximum PK ID from the target table. The above flow is working fine from last one year. Now I wish to apply the Partitioning concept to the above floe(session) At source I used 4 pass through partitions.(For Each partition different filter conditions to pull the data from source) at Target I used 4 passthrough Partitions. it is working fine for some data, but for some rows for Insert Operation , it is throwing Unique key errors, because while Inserting the data it is generating the same sequence key twice. In detail : 1st row is coming from 1st partition and generated the sequence number 1 for that row. 2nd row is coming from 1st partition and generated the sequence number 2 for that row 3rd row is coming from the 2nd partition generated the sequence number 2 again for that row. (it must generate 3 for this row) the issue is becuase of generating the same sequence numbers twice for different partitions. Can any one Please help me to resolve this issue. While Applying partitions how can I generate a Unique Sequence numbers from Unconnected lookup for Each partitioned data. Regrads, N Kiran.
What is mapping debugger?