Hi
If i had source like unique & duplicate records like 1,1,2,3,3,4
then i want load unique records in one target like 2,4 and i want
load duplicate records like 1,1,3,3 then can any body please send
me what is th scnario. my mail i
shek.inform@gmail.com
Answers were Sorted based on User's Feedback
Answer / shridhar kasat
For Unique and Duplicate , we can go for same flow :
Source-->SrcQualifier-->Aggregator(group
by this column and find count also)--> then use joiner (use
the column from Source qualifier and join the data from
Aggregator as source qualifier.col=aggregator.col) -->
Router (insert into t1 for count>1 and for count=1 insert
into t2). t1 is for duplicate records and it will cintain
1,1,3,3 and t2 will have 2,4.
Records in Source Q :
col
1,1,2,3,3,4
The records after agregator will look like this :
col count
1 2
2 1
3 2
4 1
Records after Joiner : source q.col=aggregator.col
col count
1 2
1 2
2 1
3 2
3 2
4 1
These records will flow in router and we define two groups;
one for count>1 (for duplicates record) and second for
count=1 ( for unique record)
Hope this answer suffice your requirements.
Regards
Shri
| Is This Answer Correct ? | 23 Yes | 0 No |
Answer / suman chattopadhyay
Shirish,
Part 1 of your answer is not correct. If you use distinct
option in sorter you will get all the distinct values even
for the duplicate. i.e if the values are 1,1,2,3,3,4, you
will get 1,2,3,4 in target.
Part 2 of your answer will do all the trick for duplicate
and unique ones. If you use group by and count the number
of occurences you will get 1 as the count for unique values
and more than one as count for duplicate values. Pass those
records with count 1 to 1 target and the other to the 2nd
target.
Hope this helps.
Suman
| Is This Answer Correct ? | 10 Yes | 1 No |
Answer / sanju_yu
source> dynamic lookup>router,2 conditions in router
1. condition if column_lkp port is null then insert into
target1(unqie)
2. condtion if COLumn_lkp port is not null then insert into
target2(duplicates
Plz let me know if am wrong
Sanjay
| Is This Answer Correct ? | 7 Yes | 4 No |
Answer / shirish
1.For unique: Source-->SrcQualifier--> sorter (use distinct
option)--> Target
2. For duplicate : Source-->SrcQualifier-->Aggregator(group
by this column and find count also)-->filter(use count
greater than 1 )--> left outer join with source qualifier --
> target 2.
Hope i am clear.
| Is This Answer Correct ? | 3 Yes | 3 No |
Answer / rahul
hi sanjay,
will this logic work when we are loading for the 1st time
| Is This Answer Correct ? | 0 Yes | 1 No |
Please explain in detail with example about 1.Confirmed Dimension. 2.Junk Dimension. 3.Degenerated Dimension. 4.Slowly changing Dimensions
Explain joiner transformation in informatica
in which situations do u go for starflake schema ?
Plz can anybody send Informatica and Cognos FAQ Plz To thid ID anil643@gmail.com (or) kumar643@gmail.com
You have defined the following: - Commit Type = 'Target' - Commit Interval = 10000 - writer buffer block can hold multiple 7,500 rows - you are loading 40,000 records into the target After how many records will the Informatica Server issue commit commands? a)7500, 15000, 22500, 30000, 40000 b)15000, 22500, 30000, 37500, 40000 c)15000, 22500, 30000, 40000 d)15000, 30000, 40000
is there any way to read the ms excel datas directly into informatica?like is there any possibilities to take excel file as target?
if i have source with 100 records target with 100 records and we lookup on another database table and it has 10 million record so what is the method of limiting that much record in lookup table?
how lookup is acting as both active and passive ? how is it possitble?
Define maping and sessions?
how do u move the code from development to production?
 Code page compatibly –Where all changes are made to handle the scenario.
Target increases every day even though no new record was inserted