Hi,
I have one flatfile as target in a mapping . When i am
trying to load data second time into it.
The records already is in flatfile is getting override. I
don't want to override existing records.
Note : we can do this by implementing CDC / Incremental pool
logic if target is relational .
But this is flatfile. So, even i use this same technique it
will override only
So what is the solution ?
Is there any option at session level for flatfile target ?
Advance Thanks
Answers were Sorted based on User's Feedback
Answer / krish
there is an option like APPEND IF EXISTS in session
properties
| Is This Answer Correct ? | 11 Yes | 1 No |
Answer / gvmahesh
It's very Simple We have one option at session level.
Double click on session-->Mapping tab-->>Target properties--
>Append if exists(check this option).
| Is This Answer Correct ? | 5 Yes | 0 No |
Answer / jz
At session level goto target check only insert, thus every
time session runs it inserts the existing target file.
| Is This Answer Correct ? | 0 Yes | 0 No |
What will happen if we have a filter transformation used in a mapping and in the filter condition we specify 1. Yes nothing else, only the number 1. In other words, assuming we have 10 rows coming to this filter transformation as an input, how many rows will be output by this transformation when we have only 1 in the filter condition?
What is informatica etl tool?
what is diff between IN and EXIT commands
Hi, In a mapping I have 3 targets and one fixed width file as source. Total 193 records are there . I connected one port in aggregator to all 3 targets. The same value need to be load into these 3 targets . It is loaded like that only but in different order. Why? The order of insertion should be same know for all 3 targets ? Then why the order is changed ? Any one please help me. Advance thanks.
What is the use of an aggregator cache file?
There are 2 files, Master and User. We need to compare 2 files and prepare a output log file which lists out missing Rolename for each UserName between Master and User file. Please find the sample data- MASTER.csv ---------- Org|Tmp_UsrID|ShortMark|Rolename ---|---------|----------|------------ AUS|0_ABC_PW |ABC PW |ABC Admin PW AUS|0_ABC_PW |ABC PW |MT Deny all GBR|0_EDT_SEC|CR Edit |Editor GBR|0_EDT_SEC|CR Edit |SEC MT103 GBR|0_EDT_SEC|CR Edit |AB User USER.csv -------- Org|UserName|ShortMark|Rolename ---|--------|---------|------------ AUS|charls |ABC PW |ABC Admin PW AUS|amudha |ABC PW |MT Deny all GBR|sandya |CR Edit |Editor GBR|sandya |CR Edit |SEC MT103 GBR|sandya |CR Edit |AB User GBR|sarkar |CR Edit |Editor GBR|sarkar |CR Edit |SEC MT103 Required Output file: --------------------- Org|Tmp_UsrID|UserName|Rolename |Code ---|---------|--------|------------|-------- AUS|0_ABC_PW |charls |ABC Admin PW|MATCH AUS|0_ABC_PW |charls |MT Deny all |MISSING AUS|0_ABC_PW |amudha |ABC Admin PW|MISSING AUS|0_ABC_PW |amudha |MT Deny all |MATCH GBR|0_EDT_SEC|sandya |Editor |MATCH GBR|0_EDT_SEC|sandya |SEC MT103 |MATCH GBR|0_EDT_SEC|sandya |AB User |MATCH GBR|0_EDT_SEC|sarkar |Editor |MATCH GBR|0_EDT_SEC|sarkar |SEC MT103 |MATCH GBR|0_EDT_SEC|sarkar |AB User |MISSING Both the files are mapped through Organization, Shor_mark. So, based on each Organization, Short_Mark, for each UserName from User.csv, we need to find the Matching and Missing Rolename. I am able to bring Matching records in the output. But really I don't find any concept or logic to achieve "MISSING" records which are present in Master and not in User.csv for each UserName. Please help out guys. Let me know if you need any more information. Note:- In User.csv file, there are n number of Organization, under which n number Shortmark comes which has n number of UserName.
tell me the rules and responsblites in our project(my project is development)
There are 3 groups in router transformation in which there are no conditions mentioned, so what happens if I connect these groups to the 3 targets? How many records gets into each target.(suppose if we are getting 10 records from the source)
What is lookup change?
Write the advantages of partitioning a session?
Hi All, i have question regarding Change Data Capture in Informatica. As i know we can do this using Incremental Aggregation(i never worked on Informatica but i read) and i think we can do the same with SQL Over write and with variables. Can someone tell me how do we do this using other two techniques.. using SQL Over write and variables. Thanks for the help and greatly appreciated.
Can one use mapping parameter or variables created in one mapping into any other reusable transformation?