Out of 4 mill records only 3 mill records are loaded to
target
and then job aborted. How to load only those 1 mill(not
loaded records) for next run.
This job is not sequential job, it is stand alone parallel
job.What are the possibilities available in datastage8.1?
Answers were Sorted based on User's Feedback
Answer / prasanna
Use already loaded records TGT as reference lookup,and use
lookup stage,with DROP option..and design the job.u ll get
req answer. refer
.
.
sour.......>lookup......>tgt
| Is This Answer Correct ? | 7 Yes | 4 No |
Answer / nish
there are plenty of options available.
just carefully study the scenario.
Source: 4 Mil records (doesn't change)
Target: 3 Million already loaded
Option1:
you just need to identify those 1Million pool them and then load to target.
This is clearly a case for Change Data Capture (CDC) stage.
use the Source As Before Table and Target as After.
Write those 1 Mil records based on change_code() (Deleted) to a file.
Move the contents of this file to target.
Option 2: This scenario also hints at updating the target with merge stage.use the DROP option to gather into a file and then update the target.
Option 3: Look Up being a performance concern should be your alternative.
| Is This Answer Correct ? | 2 Yes | 0 No |
Answer / venkatesh
first load the data using with surrogate stage
after that use filter stage and take that key of
surrogateand in lookup maintain the Drop option
| Is This Answer Correct ? | 2 Yes | 8 No |
I am defining one varaible parameter date in job parameters.I want use this variable date in where clause in source query.
What is ibm datastage flow designer?
How will you move hashed file from one location to another location?
What are the functionalities of link partitioner and link collector?
How many nodes supported by a one cpu in parallel jobs?
Hi guys, Design job sequence, we have 3 sources, in that 1st source in abort then only run the remaining sources.. How please design the job. Thanks.
SOURCE LIKE I_D,F1,F2 --------- 100,N,Y 100,N,N 100,Y,N 101,Y,Y 101,N,Y 102,Y,N 103,N,N 104,Y,Y 105,N,N 106,N,Y 102,N,Y 105,Y,Y O/P LIKE ID flag1 flag2 101 Y Y 101 N Y 102 Y N 102 N Y 104 Y Y 106 N Y
my source is sequencial file and my target is dataset. i am running the job in two node configuration file. my source having 10 records how the data move to target?
In my previous project we get data from mainframe and load it into datastage DB2 tables.Sometimes we get data as flat file or a mainframe tables itself directly we fetch the data.Is this a migration project?
how to load meta data
what is stage is used for below Input columns: dept|mgr|employee|salary Output columns: mgr|count of employee per mgr|avg salary per dept note: each dept has one mgr and each mgr has many employees
If you want to use a same piece of code in different jobs, how will you achieve this?