if 3 table having different columes. like first table
having 4 columns , second table having 3 columns and third
table having 2 columns then how to capture the data by
using funnel stage in parallel jobs...srinu.thadi
Answer Posted / kalyanreddy
Hi Vyas,
Please ask the question properly....
Other wise you wont get right ans...
There are two posibilities
1.
table 1 has 4 col..
like
a,b,c,d
table 2 has 3 col..
like
a,b,c
table 3 has 2 col..
a,b
In this case funnel stage give only two output columns
(consider same metadata)
which are common in all the table ( in funnel stage no
properties are there for selecting common columns from
input).
2.
If Metadata is not same in above table change metadata in
transformer stage before using funnel based on requirement
only and consider datatypes of the columns.
Is This Answer Correct ? | 12 Yes | 1 No |
Post New Answer View All Answers
client know skid info?
What are stage variables, derivations and constants?
What are the types of jobs we have in datastage?
How complex jobs are implemented in datstage to improve performance?
Have you used Unstructured data?
What are the enhancements made in datastage 7.5 compare with 7.0?
What is meta stage?
Hi, what is use of Macros,functions and Routines..? At what situation you are used. If you know the answer please explain it. Thanks.
Explain entity, attribute and relationship in datastage?
What are the differences between datastage and informatica?
create a job that splits the data in the Jobs.txt file into
four output files. You will direct the data to the
different output files using constraints. • Job name:
JobLevels
• Source file: Jobs.txt
• Target file 1: LowLevelJobs.txt
− min_lvl between 0 and 25 inclusive.
− Same column types and headings as Jobs.txt.
− Include column names in the first line of the output file.
− Job description column should be preceded by the
string “Job
Title:” and embedded within square brackets. For example, if
the job description is “Designer”, the derived value
is: “Job
Title: [Designer]”.
• Target file 2: MidLevelJobs.txt
− min_lvl between 26 and 100 inclusive.
− Same format and derivations as Target file 1.
• Target file 3: HighLevelJobs.txt
− min_lvl between 101 and 500 inclusive.
− Same format and derivations as Target file 1.
• Rejects file: JobRejects.txt
− min_lvl is out of range, i.e., below 0 or above 500.
− This file has only two columns: job_id and reject_desc.
− reject_desc is a variable-length text field, maximum
length
100. It should contain a string of the form: “Level out of
range:
I have a few records just I want to store data in to targets cycling way how?
What is job control?
What a datastage macro?
What are the features of datastage flow designer?