what r the stages mostly used in realtime scenarios
Answers were Sorted based on User's Feedback
Answer / bietl
if you are working on server jobs then this are the stages
we can use commonly..
1.seq.file stage
2.transformer stage
3.hashfile stage
4.sort stage and agg.stage
if you are working on parallel jobs then this are the
stages we can use commonly..
1.Datset stage
2.transformer stage
3.lookupstage
4.Join stage
5.funnel and merge stage
mostly it's depends on the transformation logic and complex
of the logic..
I hope this will help..
| Is This Answer Correct ? | 26 Yes | 1 No |
Answer / balaji
these are mostly used in real time scenario:
database stages are:
1.oracle enterprise
2.tera data
3.db2
4.odbc
file stage are:
1.sequencial file stage
2.dataset
3.file set
4.hash file stage
processing stage are:
1.transformation
2.sort
3.modify
4.copy
5.funnel
6.lookup
7.surrogate key
8.join
9.aggregater
10 funnel stage
11.merge stage.
debugging stage are:
1.head stage
2.tail stage
3.peek stage
| Is This Answer Correct ? | 12 Yes | 5 No |
these are mostly used in real time scenario:
database stages are:
1.oracle enterprise
2.tera data
3.db2
4.odbc
file stage are:
1.sequencial file stage
2.dataset
3.file set
4.hash file stage
processing stage are:
1.transformation
2.sort
3.modify
4.copy
5.funnel
6.lookup
7.surrogate key
8.join
9.aggregater
10 funnel stage
11.merge stage
12.Change Capture
13.Filter
debugging stage are:
1.head stage
2.tail stage
3.peek stage
| Is This Answer Correct ? | 6 Yes | 0 No |
Answer / kishore reddy
database stages are:
1.oracle enterprise
2.tera data
3.db2
4.odbc
file stage are:
1.sequencial file stage
2.dataset
3.file set
4.hash file stage
processing stage are:
1.transformation
2.sort
3.modify
4.copy
5.funnel
6.lookup
7.surrogate key
8.join
9.aggregater
10 funnel stage
11.merge stage.
| Is This Answer Correct ? | 2 Yes | 0 No |
what is the function in Transformer stage can act as Rank function in Oracle
Explain ibm infosphere information server and highlight its main features?
Create a job to load all unique products in one table and the duplicate rows in to another table. The first table should contain the following output A D The second target should contain the following output B B B C C Q2. Create a job to load each product once into one table and the remaining products which are duplicated into another table. The first table should contain the following output A B C D The second table should contain the following output B B C
What is the difference b/w Hash file and Lookup Fileset stages?
Can you explain repository tables in datastage?
source file having the data like aabbccc, i want target file result like a1a2b1b2c1c2c3.
What are the different database update actions available?
There are two file are there .1st file contains 5 records and 2nd file contain 10 records in target they want 50 records.how can achieve this
we run a job in 4 nodes,i want to run same job on 8 nodes is it possible? job recompling is mandtory?why?
what is meta data? Explain? Where it is used?
What are operators in datastage?
What is process model?