Hi Jerome,
Pavel was already very helpful yesterday, having also your configuration
(files) at hand would also be very useful. :)
cheers
alessandra
Jerome LAURET wrote:
>
> Would it be useful if we provide our configurations as examples to
> be added to the Xrootd site (??!!).
> http://xrootd.slac.stanford.edu/examples/
> We could make a picture and text and pass to Pete something like
> http://xrootd.slac.stanford.edu/examples/slac/index.html
>
>
> Andrew Hanushevsky wrote:
>> Hi Alessandra,
>>
>> He needs supervisors because he has more than 64 data nodes. You will
>> need them too. Everything is explained in
>>
>> http://xrootd.slac.stanford.edu/doc/olb_config_050601/olb_config.htm
>>
>> in the introduction section.
>>
>> Andy
>>
>> ----- Original Message ----- From: "Alessandra Forti"
>> <[log in to unmask]>
>> To: <[log in to unmask]>
>> Cc: <[log in to unmask]>
>> Sent: Wednesday, May 24, 2006 7:12 AM
>> Subject: Re: New xrootd development version 20060523-1741 available
>>
>>
>>> Hi,
>>>
>>> I see you have a 3 tiers hierarchy: redirector, supervisors and
>>> workernodes. What is a supervisor? Other configurations don't seem to
>>> have this layer. Can supervisors be specialised worker nodes? What is
>>> olbd? and finally is the software for redirector supervisors and data
>>> servers the same (i.e. I install the same rpm) but configured
>>> differently? Why are you using 2 xrootd instances per WN? Have you
>>> bound each with a different partition? Or do they both serve all the
>>> available partitions.
>>>
>>> Apologies for all these questions you'll have to bear with me until I
>>> get up o speed.
>>>
>>> thanks
>>>
>>> cheers
>>> alessandra
>>>
>>> Pavel Jakl wrote:
>>>> Hello,
>>>>
>>>> Yes, it is. But the plot is showing sum of usage at production and
>>>> development cluster. I have to split the statistics into 2 plots,
>>>> sorry about not giving you a proper plot.
>>>> I can say, that I couldn't see what Fabrizio is talking about
>>>> because we still don't have so many users using the xrootd. We
>>>> wanted to move from our old solution very slowly and see the
>>>> behavior of xrootd with increasing number of users as time goes.
>>>> Users are switching from rootd to xrootd by their choice and I can
>>>> say that this process in completely unclear to me :-) But the
>>>> number of users is higher and higher as xrootd gives more advantages.
>>>>
>>>> Cheers
>>>> Pavel
>>>>
>>>> Alessandra Forti wrote:
>>>>> Hi Pavel,
>>>>>
>>>>> thanks this is very useful. Is this the STAR production cluster? I
>>>>> read also that talk given at chep.
>>>>>
>>>>> cheers
>>>>> alessandra
>>>>>
>>>>>
>>>>> Pavel Jakl wrote:
>>>>>> Hello Alessandra,
>>>>>>
>>>>>> Andy talked mostly about scalability and performance tests of
>>>>>> xrootd and trying to show the behavior of xrootd at upper limits I
>>>>>> believe. I don't know nothing about the fact that xrootd uses a
>>>>>> lots of memory.
>>>>>> We are using xrootd on WNs and we have deployed it on 320 nodes
>>>>>> with about 130TB. These nodes are heavily loaded by user's jobs,
>>>>>> mostly ROOT jobs.
>>>>>> For your imagination, I am sending you plots being produced by
>>>>>> really home-made monitoring with Ganglia toolkit.
>>>>>> I also need to mention, that you have to divide these numbers by
>>>>>> the factor of two, because we are running two instances of xrootd
>>>>>> per node (we have a development and production cluster). You can
>>>>>> see that memory usage is really, really small.
>>>>>>
>>>>>> Cheers
>>>>>> Pavel
>>>>>>
>>>>>> Alessandra Forti wrote:
>>>>>>> Apologies that worries me in using xrootd on the WNs....
>>>>>>> I'm a bit confused right now :(
>>>>>>>
>>>>>>> Alessandra Forti wrote:
>>>>>>>> Hi Peter,
>>>>>>>>
>>>>>>>> after yesterday I also remember that the second thing that
>>>>>>>> worried me in using dcache on the WN is a comment Andy made
>>>>>>>> about the use of memory in his talk at CHEP I believe. He said
>>>>>>>> it xrootd uses a lot of memory but there weren't number specified.
>>>>>>>>
>>>>>>>> My system is double cpu with 2GB of memory per cpu. Considering
>>>>>>>> that an atlas job can use more than 1 GB we are now at 1.1 I
>>>>>>>> thik. Will it be enough? I think so but I just wanted to check.
>>>>>>>>
>>>>>>>> cheers
>>>>>>>> alessandra
>>>>>>>>
>>>>>>>>
>>>>>>>> Peter Elmer wrote:
>>>>>>>>> Hi All,
>>>>>>>>>
>>>>>>>>> There is now a new xrootd development version: xrootd
>>>>>>>>> 20060523-1741, please see:
>>>>>>>>> http://xrootd.slac.stanford.edu/download/20060523-1741/
>>>>>>>>>
>>>>>>>>> for downloads.
>>>>>>>>>
>>>>>>>>> Relative to the last development build (20060418-0404) this
>>>>>>>>> includes a variety of small bug fixes, plus one important one
>>>>>>>>> for the redirector. See
>>>>>>>>> the xrootd.History file for more details. I've included a link
>>>>>>>>> to the
>>>>>>>>> SL3 debuginfo rpm on the rpm page. (Although I've not tried it
>>>>>>>>> myself,
>>>>>>>>> so I have no idea if it works! Feedback is welcome.)
>>>>>>>>>
>>>>>>>>> Gerri, if it is still possible, you could add this to the
>>>>>>>>> next ROOT
>>>>>>>>> build, too.
>>>>>>>>>
>>>>>>>>> For the full set of changes and links to rpms/tarballs to
>>>>>>>>> download see the
>>>>>>>>> the xrootd web page and/or version history:
>>>>>>>>>
>>>>>>>>> http://xrootd.slac.stanford.edu
>>>>>>>>> http://xrootd.slac.stanford.edu/xrootd.History
>>>>>>>>>
>>>>>>>>> Let us know if there are problems.
>>>>>>>>>
>>>>>>>>> Pete
>>>>>>>>>
>>>>>>>>> -------------------------------------------------------------------------
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Peter Elmer E-mail: [log in to unmask] Phone: +41
>>>>>>>>> (22) 767-4644
>>>>>>>>> Address: CERN Division PPE, Bat. 32 2C-14, CH-1211 Geneva 23,
>>>>>>>>> Switzerland
>>>>>>>>> -------------------------------------------------------------------------
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>>
>>>>>> ------------------------------------------------------------------------
>>>>>>
>>>>>>
>>>>>>
>>>>>> ------------------------------------------------------------------------
>>>>>>
>>>>>>
>>>>>>
>>>>>> ------------------------------------------------------------------------
>>>>>>
>>>>>>
>>>>>>
>>>>>> ------------------------------------------------------------------------
>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>> --
>>> *******************************************
>>> * Dr Alessandra Forti *
>>> * Technical Coordinator - NorthGrid Tier2 *
>>> * http://www.hep.man.ac.uk/u/aforti *
>>> *******************************************
>>>
>
--
*******************************************
* Dr Alessandra Forti *
* Technical Coordinator - NorthGrid Tier2 *
* http://www.hep.man.ac.uk/u/aforti *
*******************************************
|