Data transformed successfully from Source table to target table. Now how you will ensure that the data in Target table is proper. I answered will verify one or two records and check Then question was that the Development is doing (1-2 record verification)but as a tester you have to verify the complete data, how you will do?? Please answer
Answers were Sorted based on User's Feedback
Answer / abhinaw prakash
You can write the query using both the source and target tables.
You can use the NOT IN and MINUS keyword.
Also you can make a mapping using both the tables and
generate a report on the missing records.
| Is This Answer Correct ? | 5 Yes | 0 No |
Target is zero file and how can you send the email?
Why you use repository connectivity?
what is mapping parameter and mapping variables
Why use the lookup transformation ?
my source is like this id,name sal 10 abc 1000,10 pqr 2000, 10 xyz 3000 ,10 jkl 4000 and my requirement is like this 10 abc,pqr,xyz,jkl 2000 .... i have try for this by using expression transformatin its ok of the concatenation of second column but the thing is that on third column if u group by using agg t/r the last value will com i.e 4000 but i asked by a interviewer that i dont want the first or last column i want the middle column i.e 2000 .plz reply for the same
what is the flow?
how to return multiple columes through through un-connect lookup?
I have Flat file like the data, sal have 10,000. I want to load the data in the same format as sal as 10,000. Can anybody know the answer means please mail me. Thanks in advance.. My mail id is chandranmca2007@gmail.com
100 records in source and i have 3 targets, my requirement is first record insert into 1st target,2 record into 2nd target,3 record into 3rd target and again 4th record into 1st target like dat..how to implement dis?
If u r using dynamic cache lookup port will be thier in thet which option u will select
List the transformation used for the following: (1) Heterogeneous Sources (2) Homogeneous Sources (3) Find the 5 highest paid employees within a dept. (4) Create a Summary table (5) Generate surrogate keys
from Source 100 rows are coming, on target there are 5 m rows which options is better to match data 1. Joiner 2 No cache 3. Static 4. Dynamic