6) my sorce
d1
d2
d1
d2
d1
d3
d2
d1
my target
tartget
1 d1
2 d1
3 d1
4 d1
target2
1 d2
2 d2
3 d3
Target 3
1 d3
Can any body send me answer this is argent requierment
Answers were Sorted based on User's Feedback
1. SRC--->FILTER---->TGT1,TGT2,TGT3
In source query,
SELECT COL1,
RANK() OVER(PARTITION BY COL1 ORDER BY COL1) SNO
FROM TABLE
in filter,
col1='d1'------>TGT1
col1='d2'------>TGT2
col1='d3'------>TGT3
2. SRC---->TFM---->TGT1,TGT2,TGT3
in TFM, do the HASH partitioning on col1 and use the stage
varibles as below:
if col1 =prevCol then SV1=S1+1
else SV1=1
prevCol---->col1
In the link constraints:
TGT1 link, COL1='d1'
map SV1, COL1 to the tgt.
TGT2 link, COL1='d2'
map SV1, COL1 to the tgt.
TGT3 link, COL1='d3'
map SV1, COL1 to the tgt.
| Is This Answer Correct ? | 11 Yes | 0 No |
Answer / ankit gosain
Hi,
The solution to the problem can be achieved by the below
job structure & settings:
Source---->SortStage---->Transformer---->Tgt1,Tgt2,Tgt3
1. In Sort Stage, sort in assending mode w.r.t. source
column (say col1) & Create a change key column (say
ChangeKey).
2. In Transformer stage, create a stage variable (say
count) & it's derivation would be:
if ChangeKey=1 then 1 else count+1
3. Constraint for Tgt1 => col1=d1
Constraint for Tgt2 => col1=d2
select otherwise constraint for Tgt3.
Cheers,
Ankit :)
| Is This Answer Correct ? | 6 Yes | 1 No |
Answer / srinivas
Hii guys use following stages to get output
sequecial file->copy->aggregator and Joinstage then use filter to get required output
by using copy stage take two links one for aggregator and other for Join then you will get
required output
| Is This Answer Correct ? | 1 Yes | 0 No |
Can you filter data in hashed file?
Can you explain kafka connector?
What is the purpose of interprocessor stage in server jobs?
There are two file are there .1st file contains 5 records and 2nd file contain 10 records in target they want 50 records.how can achieve this
What are the partitioning techniques available in link partitioner?
what is datastage job Monitoring
Difference between the Sequence and Sequencer in DS?
How do you load dimension data and fact data? Which is first
1)s.key generate 1 to 700 records today. tomorrow another 400 will updated how to update the records using s.key generator? 2)source is like :-- DB --> T/F stage1 --> seq1file T/f 1 is linking with T/F2 ---> seq 2 how to load the data? in source i given some conditions those r going in seq1. The another data will going to seq2 how to do this ?
How to transfer file from one system to another system in UNIX?which command to be use?
how can we create rank using datastage?what is the meaning of rank?
What is Cleanup Resources and when do you use it?