Hi Every one,
I have a scenario plz suggest me
1)On daily we r getting some huge files data so all files
metadata is same we have to load in to target table how we
can load?
2) One column having 10 records at run time we have to send
5th and 6th record to target at run time how we can send?
Hi plz help me for above scenarios and If any one is having
JobSequence kindly send me one example and the scenario to
my mail ID(nrvdwh@gmail.com)
Answer Posted / ankit gosain
Hi All,
Ans1: In the source sequential file, you can specify the
read method from specific file to file pattern, so that it
can take the file on daily basis.
like:-you are getting the files ALSSM_12122012.dat,
ALSSM_13122012.dat, ALSSM_14122012.dat and so on which is
just appended by the date with file name. for this, you can
specify a file pattern like <path>/ALSSM_*.dat
Ans2: SeqFile--->Transformer--->Tgt
create two parameters of integer type in your job (say Val1
& Val2) &
In transformer stage, define a constraint:
@INROWNUM=Val1 OR @INROWNUM=Val2 (Where @INROWNUM is a
System Variable). Now you can assign the value of Val1 &
Val2 at the run time.
For more queries, mail me on ankitgosain@gmail.com
Cheers,
Ankit :)
Is This Answer Correct ? | 2 Yes | 2 No |
Post New Answer View All Answers
How do you schedule or monitoring the job?
Difference between IBM DATA STAGE8.5 and DATA STAGE9.1 ?
Describe link sort?
Which is the best institute in Hyderabad to learn OBIEE and DataStage?Please tell me faculty name also
if i have two tables table1 table2 1a 1a,b,c,d 1b 2a,b,c,d,e 1c 1d 2a 2b 2c 2d 2e how can i get data as same as in tables? how can i implement scd typ1 and type2 in both server and in parallel? field1 field2 field3 suresh , 10,324 , 355 , 1234 ram , 23,456 , 450 , 456 balu ,40,346,23 , 275, 5678 how to remove the duplicate rows,inthe fields?
Hi, what is use of Macros,functions and Routines..? At what situation you are used. If you know the answer please explain it. Thanks.
how to get sum of sal based on dept_no and then sum of all sal irrespective of dept_no in same sql. output:- 10, 200(sum of sal for dept_no 10), 5000(sum of all sal)
create a job that splits the data in the Jobs.txt file into
four output files. You will direct the data to the
different output files using constraints. • Job name:
JobLevels
• Source file: Jobs.txt
• Target file 1: LowLevelJobs.txt
− min_lvl between 0 and 25 inclusive.
− Same column types and headings as Jobs.txt.
− Include column names in the first line of the output file.
− Job description column should be preceded by the
string “Job
Title:” and embedded within square brackets. For example, if
the job description is “Designer”, the derived value
is: “Job
Title: [Designer]”.
• Target file 2: MidLevelJobs.txt
− min_lvl between 26 and 100 inclusive.
− Same format and derivations as Target file 1.
• Target file 3: HighLevelJobs.txt
− min_lvl between 101 and 500 inclusive.
− Same format and derivations as Target file 1.
• Rejects file: JobRejects.txt
− min_lvl is out of range, i.e., below 0 or above 500.
− This file has only two columns: job_id and reject_desc.
− reject_desc is a variable-length text field, maximum
length
100. It should contain a string of the form: “Level out of
range:
How we can covert server job to a parallel job?
What are the different options associated with dsjob command?
What is the difference between hashfile and sequential file?
What are the functionalities of link collector?
What are the functionalities of link partitioner and link collector?
What is the difference between Datastage 7.5 and 7.0?
Define orabulk and bcp stages?