Hi guys,
In sequencer job,
we have 3 sources, in that 1st source have some records,
Here requirement is 1st source records are 100 then only run
the job otherwise total job will abort...
How to calculate this. please design the job.
Thanks.
Answer Posted / anil
Hi,
We can do it with one seq job and one parallel job
First create parallel job (p1) that calculates aggregate count. If aggregate count is less than 100 then abort the parallel job
Paralled Job Design(P1)
1. Take the data from source.
2. Add a column generator with column name as Temp and its value as A
3. Add aggregrator stage, where group by column is Temp and calculation is count rows. So after aggregation, count will be calculated and stored in Total Count column.
4. In the transformer, add a constraint for output link to a peak stage saying totalcount < 100 and abort after 1 row.
SO if count is greater than 100, parallel job (P1) is successful, Else failed
Sequence Job
1. Add parallel job(p1) to the sequence and add exception handler and terminator activity stage.
2. After P1 you can create whatever sources you want to as a each parallel job.
3. Add automatically handle failed activities in the sequence job properties
Now if P1 fails, sequence fails else rest will run.
| Is This Answer Correct ? | 2 Yes | 0 No |
Post New Answer View All Answers
1)How will u implement SCD2 by using surrogate key. 2)What are the disadvantages with surrogate key. 3)How will you handle nulls in your project for the varchar, integer data types. 4)Can I use two fact tables in star schema. 5)3 jobs are running on the 2 nodes after I added one more node so can I compile those jobs to run on three nodes.
Define data aggregation?
What is datastage?
How to manage date conversion in Datastage?
Difference between sequential file and data set?
State the difference between an operational datastage and a data warehouse?
how to connect source to db?generally what r stages u used? how to find the data is having delimiter format?
Can you explain repository tables in datastage?
how to sort two columns in single job in datastage.
How to implement complex jobs in data stage?
What are constraints and derivations?
How do you import and export the datastage jobs?
How we can covert server job to a parallel job?
what is repositery?
How to read the length of word in unix?