Please start any new threads on our new site at https://forums.sqlteam.com. We've got lots of great SQL Server experts to answer whatever question you can come up with.

 All Forums
 General SQL Server Forums
 New to SQL Server Programming
 Efficient Processing of Millions of Records??

Author  Topic 

mclusker
Starting Member

6 Posts

Posted - 2009-07-12 : 02:41:44
Hello Everyone,

I need to manipulate 43 million records. I will be querying off this data with joins. So the data can exponentially grow. What is the most efficient approach to handle this amount of data? Thank you ahead of time.

Matt

Matt

SwePeso
Patron Saint of Lost Yaks

30421 Posts

Posted - 2009-07-12 : 02:50:01
It depends on what your definition of "manipulate" is...

DELETE? UPDATE? INSERT?


N 56°04'39.26"
E 12°55'05.63"
Go to Top of Page

khtan
In (Som, Ni, Yak)

17689 Posts

Posted - 2009-07-12 : 02:55:35
quote:
I will be querying off this data with joins

It will depends on what is the query doing. What are the indexes on the related tables.


KH
[spoiler]Time is always against us[/spoiler]

Go to Top of Page

mclusker
Starting Member

6 Posts

Posted - 2009-07-12 : 03:56:10
My definition of manipulate is 'Update' 'Delete' and 'Insert'. I need to manage edits constantly. I will use a base index (I think, still learning about indexes) that consists of 43 million + records. With that base record I need to join it with other tables that consist of less records. Then periodically I will have to update, delete, insert into the base record to keep it accurate. SQL is the right tool? Do I need to have dedicated server? If so, what are the minimum specs to have it run fast with cost savings in mind?



Matt
Go to Top of Page

visakh16
Very Important crosS Applying yaK Herder

52326 Posts

Posted - 2009-07-12 : 04:14:09
yup sql is right tool. however whats your rule of manipulate? if more than one tables have reference to single record, which data you want to update? or is it like one table contain details of records to be deleted, one for insert etc.? perhaps you could explain your scenario with some sample data and table stuctures.
Go to Top of Page

mclusker
Starting Member

6 Posts

Posted - 2009-07-12 : 04:30:34
Hey visakh16,

Thank you for the reply. I understand the joins. The nitty gritty is that I have a massive amount of data to deal with/manipulate...What is the best way to process this amount of data? Buy a powerful server for myself? Rent a dedicated server? (I don't know maintenance so I'm figuring I need to do this) or....any suggestion is welcome.
Go to Top of Page

visakh16
Very Important crosS Applying yaK Herder

52326 Posts

Posted - 2009-07-12 : 04:51:16
what would be approx amount of data? i think you need a moderate server as long as you use proper indexes as well as write optimised query
Go to Top of Page

mclusker
Starting Member

6 Posts

Posted - 2009-07-12 : 05:34:23
Thank you very much visakh16! I am gaining an understanding of indexes as I research so I know I can break this data stream into more manageable pieces. I will have to look more into optimizing my queries by testing them against the gigs of data I will be analyzing. I will conduct this research and be back definitely to this forum because of resources like yourself. I have a great start. Thanks again.
Go to Top of Page

visakh16
Very Important crosS Applying yaK Herder

52326 Posts

Posted - 2009-07-12 : 08:26:24
welcome
Go to Top of Page
   

- Advertisement -