How to LOG 'unmatched Master' records and 'Reject Updates' in log files using MERGE stage?
Answer / subbuchamala
by setting below option in MERGE stage, we can log 'unmatched Master' records and 'Reject Updates' in log files:
1. Warn on Rejected Updates=TRUE
--->True to generate a warning when bad records from any update links are rejected.
2. Warn on Unmatched Masters=TRUE.
--->True to generate a warning when there are unmatched rows from the master link.
Is This Answer Correct ? | 7 Yes | 0 No |
Out of 4 mill records only 3 mill records are loaded to target and then job aborted. How to load only those 1 mill(not loaded records) for next run. This job is not sequential job, it is stand alone parallel job.What are the possibilities available in datastage8.1?
I am running a job with 1000 records.. If the job gots aborted after loading 400 records into target... In this case i want to load the records in the target with 401 record... How will we do it??? This scenario is not for sequence job it's only in the job Ex: Seq file--> Trans--> Dataset..
9 Answers Cognizant, IBM, TCS, Virtusa,
DataStage Scenario based Interview Questions
if a column contains data like ram,rakesh,madhan,suraj,pradeep,bhaskar then I want to place names separated by commas in another columns how can we do?
How can we move a DATASTAGE JOB from Development to Testing environment with the help of a datastage job using unix commands.
How much data u can get every day? 2)which data ur project contains? 3) what is the source in ur project?what is the biggest table & size in ur schema or in ur project?
how can u handle null values in transformer stage.
Display all files which have size more 3KB in given directory/folder.
What is job control?
How can we perform the 2nd time extraction of client database without accepting the data which is already loaded in first time extraction?
INPUT file 'A' contains: 1 2 3 4 5 6 7 8 9 10 input file 'B' contains: 6 7 8 9 10 11 12 13 14 15 Output file 'X' contains: 1 2 3 4 5 Output file 'Y' contains: 6 7 8 9 10 Output file 'Z' contains: 11 12 13 14 15 How can we implement this in a single ds job?
With out using Funnel Stage, how to populate the data from different sources to single target