Andy, IN2P3 is getting this cluster for free from Dell,
the decisions were made well many many month ago, by
Dell people. We didn't have many choices/options.
We are very grateful for the 1 TB size disks, there
are plenty of them!
Jacek
On 01/16/2015 05:33 PM, Andrew Hanushevsky wrote:
> Hmmm, why are you going with 1TB disks? Certainly 2TB disks would be
> just as good and probably more cost effective. Frankly, I'd price out
> each size (1TB, 2TB, 3TB and even 4TB -- they should be all available
> now in the 2.5" form factor).
>
> Andy
>
> On Fri, 16 Jan 2015, Yvan Calas wrote:
>
>> Hi,
>>
>>> On 15 Jan 2015, at 00:16, Fabrice Jammes <[log in to unmask]>
>>> wrote:
>>>
>>> Here's what Qserv team would like to have on the cluster, in addition
>>> to what we have defined previously:
>>>
>>> - Scientific Linux on all node, in order to get C++11 support
>>
>> Do you need SL7 only because of C++11, or is there any other reason?
>> Is it possible to have full C++11 support on SL6 nodes actually?
>>
>> As I already told you, it might take time to install SL7 on servers at
>> CC-IN2P3 (probably 2 months or more).
>>
>>> - 10TB shared storage available for all nodes and able to support a
>>> large amount of io during data-loading
>>
>> The features of the new Dell machines are as follow:
>>
>> - DELL PowerEdge R620
>> + 2 x Processors Intel Xeon E5-2603v2 1.80 Ghz 4 cores, 10 Mo cache,
>> 6.4 GT/s , 80W
>> + RAM: 16 Go DDR-3 1600MHz (2x8Go)
>> + 10 x 1 TB disk Nearline SAS 6 Gbps 7200 Tpm 2,5" - hotplug
>> + 1 x RAID card H710p with 1 GB nvram
>> + 1 x 1 GbE with 4 ports Broadcom® 5720 Base-T card
>> + 1 x iDRAC 7 Enterprise card
>> + redundant power supply
>>
>>> and some questions:
>>>
>>> - what will be the disk architecture (which kind of RAID, or
>>> something else, or nothing?)
>>
>> Since there is 10TB on each server, we plan to configure them in
>> RAID-6 (2 parity disks - 7.4 TB available), or in RAID-5 (one parity
>> disk - 8.4TB available) in you need more space on each node. If you
>> are thinking of another better RAID configuration for qserv, please
>> let us know ;)
>>
>> Note that we plan to install the 25 first machines int the computing
>> center at the beginning of week 5 (26-27/01/2015).
>>
>>> - we don't know a lot about Puppet and would like to know which kind
>>> of feature it offers (system monitoring, service restart, ...)?
>>
>> qserv admins at CC-IN2P3 (mainly myself) will write a puppet module in
>> order to:
>>
>> - deploy the qserv software automatically,
>> - tune the OS and qserv parameters.
>>
>> Moreover, since the soft will run as qserv user (as it was the case
>> last year on the 300+ nodes), I guess that you will be able to restart
>> the service, change qserv configuration files if needed, etc. using sudo.
>>
>> The monitoring will be based on Nagios (probes to define and write)
>> and collectd/smurf mainly. However plots generated by smurf will be
>> only accessible from inside CC-IN2P3. If some extra monitoring is
>> needed, we will deploy it.
>>
>>
>>> Would it be possible to talk to a Puppet/monitoring expert when these
>>> kind of questions occurs?
>>
>> I am not an expert of puppet, but I can try to answer to all of your
>> questions ;) If I don't know the answer, I will ask to Mattieu, and if
>> really needed you can contact him directly.
>>
>> Cheers,
>>
>> Yvan
>>
>>
>> ---
>> Yvan Calas
>> CC-IN2P3 -- Storage Group
>> 21 Avenue Pierre de Coubertin
>> CS70202
>> F-69627 Villeurbanne Cedex
>> Tel: +33 4 72 69 41 73
>>
>>
>> ########################################################################
>> Use REPLY-ALL to reply to list
>>
>> To unsubscribe from the QSERV-L list, click the following link:
>> https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=QSERV-L&A=1
>>
>
> ########################################################################
> Use REPLY-ALL to reply to list
>
> To unsubscribe from the QSERV-L list, click the following link:
> https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=QSERV-L&A=1
########################################################################
Use REPLY-ALL to reply to list
To unsubscribe from the QSERV-L list, click the following link:
https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=QSERV-L&A=1
|