Print

Print


Hi Alessandra,

He needs supervisors because he has more than 64 data nodes. You will need 
them too. Everything is explained in

http://xrootd.slac.stanford.edu/doc/olb_config_050601/olb_config.htm

in the introduction section.

Andy

----- Original Message ----- 
From: "Alessandra Forti" <[log in to unmask]>
To: <[log in to unmask]>
Cc: <[log in to unmask]>
Sent: Wednesday, May 24, 2006 7:12 AM
Subject: Re: New xrootd development version 20060523-1741 available


> Hi,
>
> I see you have a 3 tiers hierarchy: redirector, supervisors and 
> workernodes. What is a supervisor? Other configurations don't seem to have 
> this layer. Can supervisors be specialised worker nodes? What is olbd? and 
> finally is the software for redirector supervisors and data servers the 
> same (i.e. I install the same rpm) but configured differently? Why are you 
> using 2 xrootd instances per WN? Have you bound each with a different 
> partition? Or do they both serve all the available partitions.
>
> Apologies for all these questions you'll have to bear with me until I get 
> up o speed.
>
> thanks
>
> cheers
> alessandra
>
> Pavel Jakl wrote:
>> Hello,
>>
>> Yes, it is. But the plot is showing sum of usage at production and 
>> development cluster. I have to split the statistics into 2 plots, sorry 
>> about not giving you a proper plot.
>> I can say, that I couldn't see what Fabrizio is talking about because we 
>> still don't have so many users using the xrootd. We wanted to move from 
>> our old solution very slowly and see the behavior of xrootd with 
>> increasing number of users as time goes. Users are switching from rootd 
>> to xrootd by their choice and I can say that this process in completely 
>> unclear to me :-)  But the number of users is higher and higher as xrootd 
>> gives more advantages.
>>
>> Cheers
>> Pavel
>>
>> Alessandra Forti wrote:
>>> Hi Pavel,
>>>
>>> thanks this is very useful. Is this the STAR production cluster? I read 
>>> also that talk given at chep.
>>>
>>> cheers
>>> alessandra
>>>
>>>
>>> Pavel Jakl wrote:
>>>> Hello Alessandra,
>>>>
>>>> Andy talked mostly about scalability and performance tests of xrootd 
>>>> and trying to show the behavior of xrootd at upper limits I believe. I 
>>>> don't know nothing about the fact that xrootd uses a lots of memory.
>>>> We are using xrootd on WNs and we have deployed it on 320 nodes with 
>>>> about 130TB. These nodes are heavily loaded by user's jobs, mostly ROOT 
>>>> jobs.
>>>> For your imagination, I am sending you plots being produced by really 
>>>> home-made monitoring with Ganglia toolkit.
>>>> I also need to mention, that you have to divide these numbers by the 
>>>> factor of two, because we are running two instances of xrootd per node 
>>>> (we have a development and production cluster). You can see that memory 
>>>> usage is really, really small.
>>>>
>>>> Cheers
>>>> Pavel
>>>>
>>>> Alessandra Forti wrote:
>>>>> Apologies that worries me in using xrootd on the WNs....
>>>>> I'm a bit confused right now :(
>>>>>
>>>>> Alessandra Forti wrote:
>>>>>> Hi Peter,
>>>>>>
>>>>>> after yesterday I also remember that the second thing that worried me 
>>>>>> in using dcache on the WN is a comment Andy made about the use of 
>>>>>> memory in his talk at CHEP I believe. He said it xrootd uses a lot of 
>>>>>> memory but there weren't number specified.
>>>>>>
>>>>>> My system is double cpu with 2GB of memory per cpu. Considering that 
>>>>>> an atlas job can use more than 1 GB we are now at 1.1 I thik. Will it 
>>>>>> be enough? I think so but I just wanted to check.
>>>>>>
>>>>>> cheers
>>>>>> alessandra
>>>>>>
>>>>>>
>>>>>> Peter Elmer wrote:
>>>>>>>    Hi All,
>>>>>>>
>>>>>>>   There is now a new xrootd development version: xrootd 
>>>>>>> 20060523-1741, please see:
>>>>>>>   http://xrootd.slac.stanford.edu/download/20060523-1741/
>>>>>>>
>>>>>>> for downloads.
>>>>>>>
>>>>>>>   Relative to the last development build (20060418-0404) this 
>>>>>>> includes a variety of small bug fixes, plus one important one for 
>>>>>>> the redirector. See
>>>>>>> the xrootd.History file for more details. I've included a link to 
>>>>>>> the
>>>>>>> SL3 debuginfo rpm on the rpm page. (Although I've not tried it 
>>>>>>> myself,
>>>>>>> so I have no idea if it works! Feedback is welcome.)
>>>>>>>
>>>>>>>   Gerri, if it is still possible, you could add this to the next 
>>>>>>> ROOT
>>>>>>> build, too.
>>>>>>>
>>>>>>>    For the full set of changes and links to rpms/tarballs to 
>>>>>>> download see the
>>>>>>>  the xrootd web page and/or version history:
>>>>>>>
>>>>>>>     http://xrootd.slac.stanford.edu
>>>>>>>     http://xrootd.slac.stanford.edu/xrootd.History
>>>>>>>
>>>>>>>  Let us know if there are problems.
>>>>>>>
>>>>>>>                                    Pete
>>>>>>>
>>>>>>> -------------------------------------------------------------------------
>>>>>>>
>>>>>>> Peter Elmer     E-mail: [log in to unmask]      Phone: +41 (22) 
>>>>>>> 767-4644
>>>>>>> Address: CERN Division PPE, Bat. 32 2C-14, CH-1211 Geneva 23, 
>>>>>>> Switzerland
>>>>>>> -------------------------------------------------------------------------
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>>
>>>> ------------------------------------------------------------------------
>>>>
>>>>
>>>> ------------------------------------------------------------------------
>>>>
>>>>
>>>> ------------------------------------------------------------------------
>>>>
>>>>
>>>> ------------------------------------------------------------------------
>>>>
>>>
>>
>
> -- 
> *******************************************
> * Dr Alessandra Forti   *
> * Technical Coordinator - NorthGrid Tier2 *
> * http://www.hep.man.ac.uk/u/aforti   *
> *******************************************
>