CFF vs Seq File stage

Post questions here relative to DataStage Enterprise/PX Edition for such areas as Parallel job design, Parallel datasets, BuildOps, Wrappers, etc.

Moderators: chulett, rschirm, roy

Post Reply
mydsworld
Participant
Posts: 321
Joined: Thu Sep 07, 2006 3:55 am

CFF vs Seq File stage

Post by mydsworld »

Please let me know when the use of CFF is absolutely required and can not be done using Seq File stage.
ray.wurlod
Participant
Posts: 54607
Joined: Wed Oct 23, 2002 10:52 pm
Location: Sydney, Australia
Contact:

Post by ray.wurlod »

When you have any structured data (repeating groups, redefines), when you have packed decimal fields, just to name a couple.
IBM Software Services Group
Any contribution to this forum is my own opinion and does not necessarily reflect any position that IBM may hold.
mydsworld
Participant
Posts: 321
Joined: Thu Sep 07, 2006 3:55 am

Post by mydsworld »

For packed decimals,
can not we do that in seq file stage using format option for Decimal, Packed = yes and setting EBCDIC
John Smith
Charter Member
Charter Member
Posts: 193
Joined: Tue Sep 05, 2006 8:01 pm
Location: Australia

Post by John Smith »

yes for packed decimals you can use the seq file stage. CFF is mainly for files you get from mainframes i,e structured data.
ray.wurlod
Participant
Posts: 54607
Joined: Wed Oct 23, 2002 10:52 pm
Location: Sydney, Australia
Contact:

Post by ray.wurlod »

I meant packed decimals within structured data. Sorry for any confusion.
IBM Software Services Group
Any contribution to this forum is my own opinion and does not necessarily reflect any position that IBM may hold.
mydsworld
Participant
Posts: 321
Joined: Thu Sep 07, 2006 3:55 am

Post by mydsworld »

Can we Convert a file from EBCDIC format to ASCII format with a design like this

Seq File with EBCDIC -> Seq File with ASCII

Also, do we need to specify 'Data Format = Binary' for the 'Seq File with EBCDIC'.


Thanks.
bcarlson
Premium Member
Premium Member
Posts: 772
Joined: Fri Oct 01, 2004 3:06 pm
Location: Minnesota

Post by bcarlson »

About 90% of our input data from 100+ sources are from the mainframe and we do not use CFF anywhere, only SEQ file stage. There are some circumstances where CFF would have been handy, but its functionality in PX is not complete compared to Server - mainly in the handling of redefines or multiple record types in single file.
  • If the data has occurences (repeating groups) we use vectors in the import schema.
  • We remove (or ignore) column level redefines or have the source remove them.
  • If the redefines are used to represent multiple record types, then we import the file by only defining the first X bytes - enough to provide natural keys and record types. The rest of the record is left a one big raw field. Then we split the output with a filter based on record type and the output of each of those streams uses the column import to apply a record-type specific layout to the raw field.
Hope this helps.

Brad
It is not that I am addicted to coffee, it's just that I need it to survive.
vkhandel
Participant
Posts: 35
Joined: Wed Oct 04, 2006 12:12 am
Location: Pune

Post by vkhandel »

Hi Brad

We are doing exactly as the last point of your's (providing enough natural keys and record types), but we are facing an issue due to the presence of "packed decimal" values in the raw field of some record type.

Could you please suggest, if there is any property in Column Import stage which will unpack the "packed decimal" values.
bcarlson
Premium Member
Premium Member
Posts: 772
Joined: Fri Oct 01, 2004 3:06 pm
Location: Minnesota

Post by bcarlson »

You unpack a decimal in the column import the same as you would with an import. Your record should already be specified as binary EBCDIC. Set your import field to Decimal with the approriate length and scale and then set Packed to Yes (Packed is an option for the Decimal datatype).

Brad.
It is not that I am addicted to coffee, it's just that I need it to survive.
Post Reply