I am busy reading Gregory Smith' s PostgreSQL 9.0 High Performance and when the book was written he seemed to me a bit sceptical about SSD's. I suspect the reliability of the SSD's has improved significantly since then.
Our present server (128Gb RAM and 2.5 Tb disk space and 12 CPU cores - RAID 10) will become a development server and we are going to buy a new server.
At the moment the 'base' directory uses 1.5Tb of disk space and there is still more data to come.
The database contains blbliometric data that receive updates on a weekly basis but not much changes other than that except for cleaning of data by a few persons.
Some of the queries can take many hours to finish.
On our present system there are sometimes more than 300GB in temporary files which I suspect will not be the case on the new system with a much larger RAM.
Analysis or the SAR-logs showed that there were too much iowait in the CPU's on the old system which has a lower spec CPU than the ones considered for the new system.
We are looking possibly the following hardware:
CPU: 2 x Ivy Bridge 8C E5-2667V2 3.3G 25M 8GT/s QPI - 16 cores
RAM: 24 x 32GB DDR3-1866 2Rx4 LP ECC REG RoHS - 768Gb
with enough disk space - about 4.8 Tb on RAID 10.
My question is about the possible advantage and usage of SSD disks in the new server. At the moment I am considering using 2 x 200GB SSD' s for a separate partion for temporary files and 2 x 100GB for the operating system.
So my questions:
1. Will the SSD's in this case be worth the cost?
2. What will the best way to utilize them in the system?
Regards
Johann
--
Because experiencing your loyal love is better than life itself,
my lips will praise you. (Psalm 63:3)