When will u go for (which means situation) Connected Lookup and Connected Lookup? Pls explain with an example?
Answers were Sorted based on User's Feedback
Answer / rascal
connected lookup is used when we need to use dynamic cache,we can return multiple o/p&need to use default values...otherwise if we want to use same lookup multiple times in the mapping then we go for un-connected lookup..In this case we can use only static cache..
| Is This Answer Correct ? | 0 Yes | 0 No |
Answer / poluraju
u return multiple ports you can choose connected lookup.if u want to return only one port u can choose connected or unconnected lkp.but if we want to call the same lookup multiple time with different input for single row then we choose unconnected lkp...
| Is This Answer Correct ? | 0 Yes | 0 No |
Answer / sandy
Connected Lookup :- When we require more than one row to be
output from the lookup.
UnConnected Lookup:- When we need only one port as output
from the lookup
| Is This Answer Correct ? | 1 Yes | 2 No |
When we want to return multiple ports from a mapping we wil
use connected lookup.
An connected lookup returns only one port from the t/r. If
we want to use lookup multiple times in a mapping and if we
want to lookup data only when the condition is true we will
use unconnected lookup. (if we dont want to look up all the
records and only some records)
regards,
rayudu
| Is This Answer Correct ? | 2 Yes | 3 No |
if suppose i have 3 columns namely 1,2,3 in table a and 1,2,3 in table b Now in table a 1,2 and table b 1,2 columns are having same data.now based on this two columns either data should get inserted or updated. How do i write a sql query to check whether the data is populated correctly or not
what r the values tht r passed between informatics server and stored procedure?
I've an excel sheet, in which one column contains data about the location like "City, State" and few columns contains only "State". Now I wanna load into my target City and State into two different columns. Pls help me out...
rank() over (partition by opt2.dim_plat_site_id, opt2.dim_site_opt_sid order by case when opt2.dm_market_flg in ('Y', 'U') then 1 else 2 end, lkp.contact_rank) as rank1, case opt2.contact_type when 'Buyer' then row_number() over (partition by opt2.dim_plat_site_id, opt2.dim_site_opt_sid, lkp.contact_rank order by has_name_flg desc, ship_to_flg desc , last_order_dt desc) when 'Decision Maker' then row_number() over (partition by opt2.dim_plat_site_id, opt2.dim_site_opt_sid, lkp.contact_rank order by has_name_flg desc , last_quote_dt desc , mailability_score desc , source_ranking desc) when 'Influencer' then row_number() over (partition by opt2.dim_plat_site_id, opt2.dim_site_opt_sid, lkp.contact_rank order by has_name_flg desc, mailability_score desc, source_ranking desc) when 'Payer' then row_number() over (partition by opt2.dim_plat_site_id, opt2.dim_site_opt_sid, lkp.contact_rank order by has_name_flg desc, mailability_score desc, source_ranking desc) --elu 05/28/2013 else row_number() over (partition by opt2.dim_plat_site_id, opt2.dim_site_opt_sid, lkp.contact_rank order by has_name_flg desc, mailability_score desc, source_ranking desc) end rank2 row_number() over (partition by opt3.dim_plat_site_id, opt3.dim_site_opt_sid order by rank1,rank2) as "rank", case when "rank"<= opt3.maximum_value then 'Y' else 'N' end as include_flg
What is different between the data warehouse and data mart?
hi real timers . iam waiting for ur reply regarding ETL TESTING
Do joiner transformation use cache? If so describe?
Different circumstance which drives informatica server to expel records?
how to import multiple flat files in to single target where there is no common column in the flat files
can you avoid static cache in the lookup transformation? I mean can you
What is substr in informatica?
hi,there is a scenario like. there are three columns empid,salmonth, sal contains the values 101,jan,1000 101 feb 1000 like twelve rows are there then my required out put is like contains 13 columns empid jan feb marc.......dec and the velues are 101 1000 1000 1000 like this . thank you