how to define satge variables in transformer stage
Answers were Sorted based on User's Feedback
Answer / dilip anand k
The stage variables are those variables that can be used
and declared on our own inside the transformer stage. These
are accessible only from the transformer stage in which
they are declared.
Any Stage variables we declare are shown in a table in the
right pane of the links area.
Defining the Stage Variable:
Select Insert New Stage variable from the stage variable
shortcut menu. A New variable is added in the table showing
the stage variables in the links area. The default name
given would be StageVar with the default datatype Varchar
(255). These properties can be edited using the Transformer
stage properties dialog box.
To Open the Transformer Stage Properties box
*Click on the stage properties button on the transformer
toolbar.
*Select the stage properties from the shortcut menu
*Select Stage variable properties from the stage variable
shortcut menu
The Properties box contains the Variable Name, Initial
Value, SQL data type, Precision, Scaling and Description
(optional).
Edit the variable as required and Click OK.
| Is This Answer Correct ? | 17 Yes | 0 No |
Answer / mandru naganji
open transformer stage and in stage variable tab right click and add sv and define datatype and length
| Is This Answer Correct ? | 0 Yes | 0 No |
I have a few records all are same structures data, I want to store data in multiple targets how?
if 3 table having different columes. like first table having 4 columns , second table having 3 columns and third table having 2 columns then how to capture the data by using funnel stage in parallel jobs...srinu.thadi
17 Answers IBM, TCS,
Wat is pre-load in Hashed file
What could be a data source system?
In aggregator stage,to find the sum of the entire group of columns,it displays in binary format. How can i solve this problem.
when u face critical situation in ur project, how to solve them?
source file is having 5 records while moving into target it want to be 10 records
if we using two sources having same meta data and how to check the data in two sources is same or not? and if the data is not same i want to abort the job ?how we can do this?
What is the version control how can i apply this in DataStage can any one tell me the anser
difference between server shared container and parallel shared container
How can we perform 2nd time extraction of client database without accepting the data which is already loaded in first time extraction?
what is the best stage in datastage parller jobs to use full out join and why