I have a file with data comming as
1,x1,x2,x3
2,a1,a2
3,b1,b2,b3,b4
1,y1,y2,y3
2,c1,c2,c3
3,d1,d2
my out put should be as follows
x1,x2,x3,a1,a2,<null>,b1,b2,b3,b4
y1,y2,y3,c1,c2,c3,d1,d2,<null>
Please let me know how can we acheive this in informatica
Thanks in advance
Answers were Sorted based on User's Feedback
Answer / siva
Import the file with comma delimited, so that, it creates
with 5 columns c1,c2,c3,c4,c5
take the mapping as follows:
SQ --> Exp -->filter--> target_
in the expression take a variable port and assign the value
as iif(c1 = 1,concatenate all the c2,c3,c4,c5 otherwise
concatenate varibale post value with c2,c3,c4,c5)
then you will get as follows:
c1, v_p
1, 'x1,x2,x3,<null>'
2, 'x1,x2,x3,<null>,a1,a2,<null>,<null>'
3, 'x1,x2,x3,<null>,a1,a2,<null>,<null>,b1,b2,b3,b4'
1, 'y1,y2,y3,<null>
2, 'y1,y2,y3,<null>,c1,c2,c3,<null>
3, 'y1,y2,y3,<null>,c1,c2,c3,<null>,d1,d2,<null>,<null>'
you can lso ingnore the unncessary columns where ever
filter out only c1 = 3 to the target
It helps you to succeed..
| Is This Answer Correct ? | 1 Yes | 0 No |
Answer / pandu
u can use sq transformation .
start value 1,
incremental value 1,
reset value:slect,
and use the rank transformation.
| Is This Answer Correct ? | 1 Yes | 0 No |
Answer / rahul
use an exp t/r,
take an variable port as count = 0
keep inc the count by 1 and place a conditin if count > 4 ,
than set count = 0
and in other variable port ypu concate the columns value.
| Is This Answer Correct ? | 1 Yes | 2 No |
Answer / bebo
use comma delimiter..infa will identify the null values
while importing the flat file
| Is This Answer Correct ? | 0 Yes | 2 No |
I have the input file as col1 col2 col3 3 2 1 7 6 8 I should get the output as col1 col2 col3 1 2 3 6 7 8 ....What is the logic to get this? Is there any transformation which sorts row wise ? If not how to sort the incoming records row wise?
Explain what are the different types of transformation available in informatica. And what are the mostly used ones among them?
We can achieve better perfomance using sorter before an aggragator. Then what is the significance of 'Sorted Input' check in aggregator ?? Please let me know the difference ??
What is the commit type if you have a transaction control transformation in the mapping?
major difference between normal loading and bulk loading?
Which transformation should u need while using the cobol sources as source defintions?
Source and Target are flat files, Source table is as below ID,NAME 1,X 1,X 2,Y 2,Y On Target flat file i want the data to be loaded as mentioned below ID,NAME,REPEAT 1,X,2 1,X,2 2,Y,2 2,Y,2 How to achieve this, Can i get a map structure
Can you start a session inside a batch individually?
For stage table data processing, suppose in first run we processed 8 records out of 10 records then in 2nd run we should consider only not processed records (here total no of records =2) along with new records which got loaded in stage table thru real time mapping. Note : In this example, 8 records are those records for which we got transaction number after lookup on trn_no_cod table and 2 records for which lookup returns trn_no as NULL
Which is the best place to learn Informatica Powercenter Online Training?
what is update strategy transformation ?
what is pre and post sql and what is diff between source presql and target pre sql