Golgappa.net | Golgappa.org | BagIndia.net | BodyIndia.Com | CabIndia.net | CarsBikes.net | CarsBikes.org | CashIndia.net | ConsumerIndia.net | CookingIndia.net | DataIndia.net | DealIndia.net | EmailIndia.net | FirstTablet.com | FirstTourist.com | ForsaleIndia.net | IndiaBody.Com | IndiaCab.net | IndiaCash.net | IndiaModel.net | KidForum.net | OfficeIndia.net | PaysIndia.com | RestaurantIndia.net | RestaurantsIndia.net | SaleForum.net | SellForum.net | SoldIndia.com | StarIndia.net | TomatoCab.com | TomatoCabs.com | TownIndia.com
Interested to Buy Any Domain ? << Click Here >> for more details...


A flat file contains 200 records. I want to load first 50
records at first time running the job, second 50 records at
second time running and so on, how u can develop this job?



A flat file contains 200 records. I want to load first 50 records at first time running the job, se..

Answer / subhash

1st Way:
1. Add 'row number' column in Seq File stage, so that each
record has a number associated with it.
2. Add a job param with which we can provide the number of
record from where we want to run the job. We can pass this
either using Sequence Start LOOP(List type variables-
50,100,150,200) or by shell script.
3. In the tfm, use a stage variable to run only from the
record number till 50 records by counting each record.

2nd way:
Design the job like this:
1. Add 'row number' column in Seq File stage, so that each
record has a number associated with it.
2. Use filter stage and write the conditions like this:
a. row number column<=50(in 1st link to load the records
in target file/database)
b. row number column>50 (in 2nd link to load the records
in the file with the same name as input file name, in
overwrite mode)


So, first time when your job runs first 50 records will be
loaded in the target and same time the input file records
are overwritten with records next first 50 records i.e. 51
to 200.
2nd time when your job runs first 50 records(i.e. 51-100)
will be loaded in the target and same time the input file
records are overwritten with records next first 50 records
i.e. 101 to 200.
And so on, all 50-50 records will be loaded in each run to
the target

Is This Answer Correct ?    8 Yes 1 No

Post New Answer

More Data Stage Interview Questions

I have 100 records how can I load at a time from the single time

1 Answers  


Thanks to all people who are posting their comments...

1 Answers   CTS,


iam new to datastage...now i want to know what are fact tables, dimension tables in bank domain...if any body knows plz tell me asap..

4 Answers   Wipro,


Have you used Unstructured data?

0 Answers   CTS,


how can i get 2nd highest salary in datastage?can u send me ,thanQ 2)if i had source has 2 records 1st record ie 1st column contains 1,2,3 and 2nd coulmn contains 10,10,10 i have to get target as 2nd columns as 20,30,40 how can i?

3 Answers   Accenture,


What are the different options associated with dsjob command?

0 Answers  


How can remove duplicates in a file using UNIX?

3 Answers   Cap Gemini,


What are the partition techniques available in your last project?

1 Answers  


file1 1 2 3 4 file2 3 4 5 6 output should be in three targets T1 T2 T3 1 3 5 2 4 6 how to do this? can any one help? Thanks

5 Answers   Cap Gemini,


how can u handle null values in transformer stage.

7 Answers  


What is exact difference between Parallel Jobs and server Jobs..

8 Answers   DELL, IBM,


Hi friends If anybody completed datastage 8.5 certification.Could please guide me what are the topics new in 8.5 certification.

2 Answers  


Categories