i have seq file that contents 10 million records load to target any data base..
in that case it takes lot of time for loading..how do performance tuning in that situation...?
Answers were Sorted based on User's Feedback
1.Increase the 'Numbers Readers per Node' to '2' in source file.
2. Use the Bulk Load Option in Target Oracle Connector.
3. set option Re-Build Index option, so that it will drop indexes before loading and It will create the indexes once load is completed.
4. Increase the 'Record count'(set 10,000, should be multiple of Array size) in TGT.
5. Increase 'Array size'(set 5,000, default is 2,000) in TGT.
6. SRC---->COPY---->TGT,
use COPY stage, so, it work like buffer whenever there is diff in number records processing at source side and at target side.
| Is This Answer Correct ? | 10 Yes | 0 No |
Answer / krishna
1) In sequential stage we have stage properties on that
select options and increase the readers
2) In datastage admin we have on environmental variable like
logical reader that set as one
| Is This Answer Correct ? | 2 Yes | 1 No |
Answer / bharath
It can be achieved by DEGREE OF PARALLELISM.
i.e, we can configure the multiple nodes in the configuration
file. In configuration file we can specify multiple nodes..
| Is This Answer Correct ? | 2 Yes | 3 No |
Answer / prasad
In Seq file, we have one property called no. of readers/node(by default is one)=4 and
in Target table use "Bulk load" and also use Round Robin partition.
first drop index before loading data and once data loaded then re-create index again.
Plz correct me, if i am wrong.....
| Is This Answer Correct ? | 0 Yes | 1 No |
Explain briefly scd type2 in datastage7.5x2(parallel)
Unix Qn asked in datastage interview: I have diff type(.txt, .tmp, .bat etc) of file in 4 diff directories, I want move all '.txt' file from 4 directories to other folder. And need to delete all the files except which are created TODAY?
how do u capture duplicates through sort & transformer
Hi This is Vijay, How can your remove the duplicates in sequential File?
15 Answers HCL, Scope International,
I have oracle enterpaise stage, oracle bulkloader stage, ODBC stage and I want to load data in to target, which of the above stages do I need to use and why?
Anyone has Datastage certification free dumps for 000-418 , 000-421 codes, mail me @ manik.dwh@gmail.com 000-418 : InfoSphere DataStage v8.0 000-421 : InfoSphere DataStage v8.5
Hi, My source is oracle(eno,ename,sal,commision,...), my requirement is like this, if there is a null values in commission col i want to keep it as null,and for the remaining first two characters of the value in my target. Plz help me
How do you load dimension data and fact data? Which is first
How to RD using transformer?
Drop duplicate records ... SOURCE LIKE .......... ID flag1 flag2 100 N Y 100 N N 100 Y N 101 Y Y 101 N Y 102 Y N 103 N N 104 Y Y 105 N N 106 N Y 102 N Y 105 Y Y in above file if any id having both the flags as "N" then that corresponding id records should be dropped, in above case o/p should be as ID flag1 flag2 101 Y Y 101 N Y 102 Y N 102 N Y 104 Y Y 106 N Y Steps to do : 1) Identified the id’s that got duplicated (both the flag values having vales “N”) 2) Look up with these id’s to existing id’s to drop .
What is the diff between sort performed at sort stage and the stream sort performed at the input of few stages in DS Enterprise edition?
1.How to read multiple file from sequential stage. 2.If a file doesn't arrive or doesn't exists in sequential stage how do u handle this. 3.What do you do before taking data from source to staging area. 4. I have a remove duplicate stage and transformer stage.what will u do for optimizeing the performance of the job.