Huge data

Post questions here relative to DataStage Server Edition for such areas as Server job design, DS Basic, Routines, Job Sequences, etc.

Moderators: chulett, rschirm, roy

Post Reply
sheema
Premium Member
Premium Member
Posts: 204
Joined: Sat Jul 22, 2006 10:20 am

Huge data

Post by sheema »

We have a huge table with around 100000000 rows,now we get weekly file,which has to be updated or inserted to this table.What would be the best approach to get this done.
We load the file into a temprary table, should we have a procedure to delete all the rows in the file and again load the rows in a file to the table.
Which method would be performance efficient.

Thanks in advance
chulett
Charter Member
Charter Member
Posts: 43085
Joined: Tue Nov 12, 2002 4:34 pm
Location: Denver, CO

Post by chulett »

Not sure I'd call that "huge". :wink:

What database? What's the typical ratio of inserts v. updates, any clue?
-craig

"You can never have too many knives" -- Logan Nine Fingers
sheema
Premium Member
Premium Member
Posts: 204
Joined: Sat Jul 22, 2006 10:20 am

Post by sheema »

Oracle is the database,not sure of the inserts vs updates.

Thanks
Nagaraj
Premium Member
Premium Member
Posts: 383
Joined: Thu Nov 08, 2007 12:32 am
Location: Bangalore

Full Refresh

Post by Nagaraj »

For Temp Table Use Full Refresh, Write a small script to truncate the temp tables, and run the full refresh job, configure all this in a sequence.

For Upsert user Upsert Mode in oracle EE stage target.

i think it should be very simple jobs.

Please let me know if you need further information.
chulett
Charter Member
Charter Member
Posts: 43085
Joined: Tue Nov 12, 2002 4:34 pm
Location: Denver, CO

Post by chulett »

Server job, so no "EE" stage.

How big is this weekly file? That size and the ratio of inserts to updates (which is important to know) will drive the design.
-craig

"You can never have too many knives" -- Logan Nine Fingers
sheema
Premium Member
Premium Member
Posts: 204
Joined: Sat Jul 22, 2006 10:20 am

Post by sheema »

The weekly file is around 28 million,the no of updates are more than the updates.(the ratio is around 90:10).


Thanks
sheema
Premium Member
Premium Member
Posts: 204
Joined: Sat Jul 22, 2006 10:20 am

Post by sheema »

The weekly file is around 28 million,the no of updates are more than the updates.(the ratio is around 90:10).


Thanks
chulett
Charter Member
Charter Member
Posts: 43085
Joined: Tue Nov 12, 2002 4:34 pm
Location: Denver, CO

Post by chulett »

"the no of updates are more than the updates". :?
-craig

"You can never have too many knives" -- Logan Nine Fingers
sheema
Premium Member
Premium Member
Posts: 204
Joined: Sat Jul 22, 2006 10:20 am

Post by sheema »

sorry,the no of updates are more than inserts.
sheema
Premium Member
Premium Member
Posts: 204
Joined: Sat Jul 22, 2006 10:20 am

Post by sheema »

can someone shed some light on this.

Thanks
chulett
Charter Member
Charter Member
Posts: 43085
Joined: Tue Nov 12, 2002 4:34 pm
Location: Denver, CO

Post by chulett »

Your DBA. Have a chat with them and then let us know what approach they recommend and we'll see about helping you implement that in DataStage.
-craig

"You can never have too many knives" -- Logan Nine Fingers
sheema
Premium Member
Premium Member
Posts: 204
Joined: Sat Jul 22, 2006 10:20 am

Post by sheema »

Thank you, I will contact the DBA and get back.
Post Reply