Print

Print


Hmmm, why are you going with 1TB disks? Certainly 2TB disks would be just 
as good and probably more cost effective. Frankly, I'd price out each size 
(1TB, 2TB, 3TB and even 4TB -- they should be all available now in the 
2.5" form factor).

Andy

On Fri, 16 Jan 2015, Yvan Calas wrote:

> Hi,
>
>> On 15 Jan 2015, at 00:16, Fabrice Jammes <[log in to unmask]> wrote:
>>
>> Here's what Qserv team would like to have on the cluster, in addition to what we have defined previously:
>>
>> - Scientific Linux on all node, in order to get C++11 support
>
> Do you need SL7 only because of C++11, or is there any other reason? Is it possible to have full C++11 support on SL6 nodes actually?
>
> As I already told you, it might take time to install SL7 on servers at CC-IN2P3 (probably 2 months or more).
>
>> - 10TB shared storage available for all nodes and able to support a large amount of io during data-loading
>
> The features of the new Dell machines are as follow:
>
> - DELL PowerEdge R620
>  + 2 x Processors Intel Xeon E5-2603v2 1.80 Ghz 4 cores, 10 Mo cache, 6.4 GT/s , 80W
>  + RAM: 16 Go DDR-3 1600MHz (2x8Go)
>  + 10 x 1 TB disk Nearline SAS 6 Gbps 7200 Tpm 2,5" - hotplug
>  + 1 x RAID card H710p with 1 GB nvram
>  + 1 x 1 GbE with 4 ports Broadcom® 5720 Base-T card
>  + 1 x iDRAC 7 Enterprise card
>  + redundant power supply
>
>> and some questions:
>>
>> - what will be the disk architecture (which kind of RAID, or something else, or nothing?)
>
> Since there is 10TB on each server, we plan to configure them in RAID-6 (2 parity disks - 7.4 TB available), or in RAID-5 (one parity disk - 8.4TB available) in you need more space on each node. If you are thinking of another better RAID configuration for qserv, please let us know ;)
>
> Note that we plan to install the 25 first machines int the computing center at the beginning of week 5 (26-27/01/2015).
>
>> - we don't know a lot about Puppet and would like to know which kind of feature it offers (system monitoring, service restart, ...)?
>
> qserv admins at CC-IN2P3 (mainly myself) will write a puppet module in order to:
>
> - deploy the qserv software automatically,
> - tune the OS and qserv parameters.
>
> Moreover, since the soft will run as qserv user (as it was the case last year on the 300+ nodes), I guess that you will be able to restart the service, change qserv configuration files if needed, etc. using sudo.
>
> The monitoring will be based on Nagios (probes to define and write) and collectd/smurf mainly. However plots generated by smurf will be only accessible from inside CC-IN2P3. If some extra monitoring is needed, we will deploy it.
>
>
>> Would it be possible to talk to a Puppet/monitoring expert when these kind of questions occurs?
>
> I am not an expert of puppet, but I can try to answer to all of your questions ;) If I don't know the answer, I will ask to Mattieu, and if really needed you can contact him directly.
>
> Cheers,
>
> Yvan
>
>
> ---
> Yvan Calas
> CC-IN2P3 -- Storage Group
> 21 Avenue Pierre de Coubertin
> CS70202
> F-69627 Villeurbanne Cedex
> Tel: +33 4 72 69 41 73
>
>
> ########################################################################
> Use REPLY-ALL to reply to list
>
> To unsubscribe from the QSERV-L list, click the following link:
> https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=QSERV-L&A=1
>

########################################################################
Use REPLY-ALL to reply to list

To unsubscribe from the QSERV-L list, click the following link:
https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=QSERV-L&A=1