Re: [GENERAL] Huge number of raws
On Fri, 19 Mar 2004, Anton Nikiforov wrote: > >Or were you talking about something else like storing different data in > >different media speeds? (Like Hierarchical Storage Management) > > > > > I do not exactly know how to deal wth such a huge amount of data. The disk subsytem > is the must and i do undrstand this. SMP architecture is the must also. > I was asking is there any way that data will migrate from fast disk subsystem to > slower but relyible automaticaly. Like in Nivell Netware (i used to work with it 7-8 > years ago) you could ask the system if the file is untached for a month - then move > it from one disk to magnetic-optical or tape but if this file is requested OS could > move it back to the operational volume. Ok so you were talking about HSM (Hierarchical Storage Management). I don't believe it's possible at this time anything like that with PostgreSQL. Also I don't think, personally, that anything like that will ever be implemented. The main object of a database is to allow access to data quickly. HSM is a way to keep cost of storage low. There are, however, ways you could do this with programs. If there is data that your users will not need often you could write programs to move that particular data from one server to another. ---(end of broadcast)--- TIP 1: subscribe and unsubscribe commands go to [EMAIL PROTECTED]
Re: [GENERAL] Huge number of raws
On Friday 19 March 2004 08:10, Anton Nikiforov wrote: > > I do not exactly know how to deal wth such a huge amount of data. The disk > subsytem is the must and i do undrstand this. SMP architecture is the must > also. I was asking is there any way that data will migrate from fast disk > subsystem to slower but relyible automaticaly. Like in Nivell Netware (i > used to work with it 7-8 years ago) you could ask the system if the file is > untached for a month - then move it from one disk to magnetic-optical or > tape but if this file is requested OS could move it back to the operational > volume. Won't work for a database - it would have to sit there waiting while someone went to the vault, got a tape out and restored a few GB of data. You could separate out old data, say monthly into archive tables and then backup/drop those. -- Richard Huxton Archonet Ltd ---(end of broadcast)--- TIP 9: the planner will ignore your desire to choose an index scan if your joining column's datatypes do not match
Re: [GENERAL] Huge number of raws
Francisco Reyes ?: On Thu, 18 Mar 2004, Anton Nikiforov wrote: But i'm worry about mentioned centeral database that should store 240 millions of records daily and should collect this data for years. I have not worked with anything even remotely so big. A few thougths.. I think this is more of a hardware issue than a PostgreSQL issue. I think a good disk subsystem will be a must. Last time I was looking for my ex employer at large disk subsystems I think the one we were leaning towards was an IBM disk subsystem. I think it was in the $100,000 range. Regardless of architecture (ie PC, SUN, etc..) SMP may be of help if you have concurrent users. Lots and lots of memory will help too. And the data migration problem is still an opened issue for me - how to make data migration from fast devices (RAID ARRAY) to slower devices (MO Library or something like this) still having access to this data? Don't follow you there. You mean backup? You can make a pg_dump of the data while the DB is running and then back that up. Or were you talking about something else like storing different data in different media speeds? (Like Hierarchical Storage Management) I do not exactly know how to deal wth such a huge amount of data. The disk subsytem is the must and i do undrstand this. SMP architecture is the must also. I was asking is there any way that data will migrate from fast disk subsystem to slower but relyible automaticaly. Like in Nivell Netware (i used to work with it 7-8 years ago) you could ask the system if the file is untached for a month - then move it from one disk to magnetic-optical or tape but if this file is requested OS could move it back to the operational volume. Anton smime.p7s Description: S/MIME Cryptographic Signature
Re: [GENERAL] Huge number of raws
Anton Nikiforov пишет: Dear All! I have a question about how the PostgreSQL will manage a huge number of raws. I have a projet where each half an hour 10 millions of records will be added to the database and they should be calculated, summarized and managed. I'm planning to have a few servers that will receive something like a million records per server and then they will store this data into the centeral server in report-ready format. I know that one million records could be managed by postgres (i have a database with 25 millions of records and it is working just fine) But i'm worry about mentioned centeral database that should store 240 millions of records daily and should collect this data for years. I cannot even imagine the needed hardware to collect monthly statistics. And my question is - is this task is for postgres, or i should think about Oracle or DB2? I'm also thinking about replication of data between two servers for redundancy, what you could suggst for this? And the data migration problem is still an opened issue for me - how to make data migration from fast devices (RAID ARRAY) to slower devices (MO Library or something like this) still having access to this data? And one more question - is there in postgress something like table partitioning in Oracle to store data according to the some rules, like a group of data source (IP network or something)? -- Best regads, Anton Nikiforov smime.p7s Description: S/MIME Cryptographic Signature
[GENERAL] Huge number of raws
Dear All! I have a question about how the PostgreSQL will manage a huge number of raws. I have a projet where each half an hour 10 millions of records will be added to the database and they should be calculated, summarized and managed. I'm planning to have a few servers that will receive something like a million records per server and then they will store this data into the centeral server in report-ready format. I know that one million records could be managed by postgres (i have a database with 25 millions of records and it is working just fine) But i'm worry about mentioned centeral database that should store 240 millions of records daily and should collect this data for years. I cannot even imagine the needed hardware to collect monthly statistics. And my question is - is this task is for postgres, or i should think about Oracle or DB2? I'm also thinking about replication of data between two servers for redundancy, what you could suggst for this? And the data migration problem is still an opened issue for me - how to make data migration from fast devices (RAID ARRAY) to slower devices (MO Library or something like this) still having access to this data? -- Best regads, Anton Nikiforov smime.p7s Description: S/MIME Cryptographic Signature