Print

Print


    Would it be useful if we provide our configurations as examples to
be added to the Xrootd site (??!!). 
http://xrootd.slac.stanford.edu/examples/
We could make a picture and text and pass to Pete something like
http://xrootd.slac.stanford.edu/examples/slac/index.html


Andrew Hanushevsky wrote:
> Hi Alessandra,
>
> He needs supervisors because he has more than 64 data nodes. You will 
> need them too. Everything is explained in
>
> http://xrootd.slac.stanford.edu/doc/olb_config_050601/olb_config.htm
>
> in the introduction section.
>
> Andy
>
> ----- Original Message ----- From: "Alessandra Forti" 
> <[log in to unmask]>
> To: <[log in to unmask]>
> Cc: <[log in to unmask]>
> Sent: Wednesday, May 24, 2006 7:12 AM
> Subject: Re: New xrootd development version 20060523-1741 available
>
>
>> Hi,
>>
>> I see you have a 3 tiers hierarchy: redirector, supervisors and 
>> workernodes. What is a supervisor? Other configurations don't seem to 
>> have this layer. Can supervisors be specialised worker nodes? What is 
>> olbd? and finally is the software for redirector supervisors and data 
>> servers the same (i.e. I install the same rpm) but configured 
>> differently? Why are you using 2 xrootd instances per WN? Have you 
>> bound each with a different partition? Or do they both serve all the 
>> available partitions.
>>
>> Apologies for all these questions you'll have to bear with me until I 
>> get up o speed.
>>
>> thanks
>>
>> cheers
>> alessandra
>>
>> Pavel Jakl wrote:
>>> Hello,
>>>
>>> Yes, it is. But the plot is showing sum of usage at production and 
>>> development cluster. I have to split the statistics into 2 plots, 
>>> sorry about not giving you a proper plot.
>>> I can say, that I couldn't see what Fabrizio is talking about 
>>> because we still don't have so many users using the xrootd. We 
>>> wanted to move from our old solution very slowly and see the 
>>> behavior of xrootd with increasing number of users as time goes. 
>>> Users are switching from rootd to xrootd by their choice and I can 
>>> say that this process in completely unclear to me :-)  But the 
>>> number of users is higher and higher as xrootd gives more advantages.
>>>
>>> Cheers
>>> Pavel
>>>
>>> Alessandra Forti wrote:
>>>> Hi Pavel,
>>>>
>>>> thanks this is very useful. Is this the STAR production cluster? I 
>>>> read also that talk given at chep.
>>>>
>>>> cheers
>>>> alessandra
>>>>
>>>>
>>>> Pavel Jakl wrote:
>>>>> Hello Alessandra,
>>>>>
>>>>> Andy talked mostly about scalability and performance tests of 
>>>>> xrootd and trying to show the behavior of xrootd at upper limits I 
>>>>> believe. I don't know nothing about the fact that xrootd uses a 
>>>>> lots of memory.
>>>>> We are using xrootd on WNs and we have deployed it on 320 nodes 
>>>>> with about 130TB. These nodes are heavily loaded by user's jobs, 
>>>>> mostly ROOT jobs.
>>>>> For your imagination, I am sending you plots being produced by 
>>>>> really home-made monitoring with Ganglia toolkit.
>>>>> I also need to mention, that you have to divide these numbers by 
>>>>> the factor of two, because we are running two instances of xrootd 
>>>>> per node (we have a development and production cluster). You can 
>>>>> see that memory usage is really, really small.
>>>>>
>>>>> Cheers
>>>>> Pavel
>>>>>
>>>>> Alessandra Forti wrote:
>>>>>> Apologies that worries me in using xrootd on the WNs....
>>>>>> I'm a bit confused right now :(
>>>>>>
>>>>>> Alessandra Forti wrote:
>>>>>>> Hi Peter,
>>>>>>>
>>>>>>> after yesterday I also remember that the second thing that 
>>>>>>> worried me in using dcache on the WN is a comment Andy made 
>>>>>>> about the use of memory in his talk at CHEP I believe. He said 
>>>>>>> it xrootd uses a lot of memory but there weren't number specified.
>>>>>>>
>>>>>>> My system is double cpu with 2GB of memory per cpu. Considering 
>>>>>>> that an atlas job can use more than 1 GB we are now at 1.1 I 
>>>>>>> thik. Will it be enough? I think so but I just wanted to check.
>>>>>>>
>>>>>>> cheers
>>>>>>> alessandra
>>>>>>>
>>>>>>>
>>>>>>> Peter Elmer wrote:
>>>>>>>>    Hi All,
>>>>>>>>
>>>>>>>>   There is now a new xrootd development version: xrootd 
>>>>>>>> 20060523-1741, please see:
>>>>>>>>   http://xrootd.slac.stanford.edu/download/20060523-1741/
>>>>>>>>
>>>>>>>> for downloads.
>>>>>>>>
>>>>>>>>   Relative to the last development build (20060418-0404) this 
>>>>>>>> includes a variety of small bug fixes, plus one important one 
>>>>>>>> for the redirector. See
>>>>>>>> the xrootd.History file for more details. I've included a link 
>>>>>>>> to the
>>>>>>>> SL3 debuginfo rpm on the rpm page. (Although I've not tried it 
>>>>>>>> myself,
>>>>>>>> so I have no idea if it works! Feedback is welcome.)
>>>>>>>>
>>>>>>>>   Gerri, if it is still possible, you could add this to the 
>>>>>>>> next ROOT
>>>>>>>> build, too.
>>>>>>>>
>>>>>>>>    For the full set of changes and links to rpms/tarballs to 
>>>>>>>> download see the
>>>>>>>>  the xrootd web page and/or version history:
>>>>>>>>
>>>>>>>>     http://xrootd.slac.stanford.edu
>>>>>>>>     http://xrootd.slac.stanford.edu/xrootd.History
>>>>>>>>
>>>>>>>>  Let us know if there are problems.
>>>>>>>>
>>>>>>>>                                    Pete
>>>>>>>>
>>>>>>>> ------------------------------------------------------------------------- 
>>>>>>>>
>>>>>>>>
>>>>>>>> Peter Elmer     E-mail: [log in to unmask]      Phone: +41 
>>>>>>>> (22) 767-4644
>>>>>>>> Address: CERN Division PPE, Bat. 32 2C-14, CH-1211 Geneva 23, 
>>>>>>>> Switzerland
>>>>>>>> ------------------------------------------------------------------------- 
>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>>
>>>>> ------------------------------------------------------------------------ 
>>>>>
>>>>>
>>>>>
>>>>> ------------------------------------------------------------------------ 
>>>>>
>>>>>
>>>>>
>>>>> ------------------------------------------------------------------------ 
>>>>>
>>>>>
>>>>>
>>>>> ------------------------------------------------------------------------ 
>>>>>
>>>>>
>>>>
>>>
>>
>> -- 
>> *******************************************
>> * Dr Alessandra Forti   *
>> * Technical Coordinator - NorthGrid Tier2 *
>> * http://www.hep.man.ac.uk/u/aforti   *
>> *******************************************
>>

-- 

             ,,,,,
            ( o o )
         --m---U---m--
             Jerome