Would it be useful if we provide our configurations as examples to be added to the Xrootd site (??!!). http://xrootd.slac.stanford.edu/examples/ We could make a picture and text and pass to Pete something like http://xrootd.slac.stanford.edu/examples/slac/index.html Andrew Hanushevsky wrote: > Hi Alessandra, > > He needs supervisors because he has more than 64 data nodes. You will > need them too. Everything is explained in > > http://xrootd.slac.stanford.edu/doc/olb_config_050601/olb_config.htm > > in the introduction section. > > Andy > > ----- Original Message ----- From: "Alessandra Forti" > <[log in to unmask]> > To: <[log in to unmask]> > Cc: <[log in to unmask]> > Sent: Wednesday, May 24, 2006 7:12 AM > Subject: Re: New xrootd development version 20060523-1741 available > > >> Hi, >> >> I see you have a 3 tiers hierarchy: redirector, supervisors and >> workernodes. What is a supervisor? Other configurations don't seem to >> have this layer. Can supervisors be specialised worker nodes? What is >> olbd? and finally is the software for redirector supervisors and data >> servers the same (i.e. I install the same rpm) but configured >> differently? Why are you using 2 xrootd instances per WN? Have you >> bound each with a different partition? Or do they both serve all the >> available partitions. >> >> Apologies for all these questions you'll have to bear with me until I >> get up o speed. >> >> thanks >> >> cheers >> alessandra >> >> Pavel Jakl wrote: >>> Hello, >>> >>> Yes, it is. But the plot is showing sum of usage at production and >>> development cluster. I have to split the statistics into 2 plots, >>> sorry about not giving you a proper plot. >>> I can say, that I couldn't see what Fabrizio is talking about >>> because we still don't have so many users using the xrootd. We >>> wanted to move from our old solution very slowly and see the >>> behavior of xrootd with increasing number of users as time goes. >>> Users are switching from rootd to xrootd by their choice and I can >>> say that this process in completely unclear to me :-) But the >>> number of users is higher and higher as xrootd gives more advantages. >>> >>> Cheers >>> Pavel >>> >>> Alessandra Forti wrote: >>>> Hi Pavel, >>>> >>>> thanks this is very useful. Is this the STAR production cluster? I >>>> read also that talk given at chep. >>>> >>>> cheers >>>> alessandra >>>> >>>> >>>> Pavel Jakl wrote: >>>>> Hello Alessandra, >>>>> >>>>> Andy talked mostly about scalability and performance tests of >>>>> xrootd and trying to show the behavior of xrootd at upper limits I >>>>> believe. I don't know nothing about the fact that xrootd uses a >>>>> lots of memory. >>>>> We are using xrootd on WNs and we have deployed it on 320 nodes >>>>> with about 130TB. These nodes are heavily loaded by user's jobs, >>>>> mostly ROOT jobs. >>>>> For your imagination, I am sending you plots being produced by >>>>> really home-made monitoring with Ganglia toolkit. >>>>> I also need to mention, that you have to divide these numbers by >>>>> the factor of two, because we are running two instances of xrootd >>>>> per node (we have a development and production cluster). You can >>>>> see that memory usage is really, really small. >>>>> >>>>> Cheers >>>>> Pavel >>>>> >>>>> Alessandra Forti wrote: >>>>>> Apologies that worries me in using xrootd on the WNs.... >>>>>> I'm a bit confused right now :( >>>>>> >>>>>> Alessandra Forti wrote: >>>>>>> Hi Peter, >>>>>>> >>>>>>> after yesterday I also remember that the second thing that >>>>>>> worried me in using dcache on the WN is a comment Andy made >>>>>>> about the use of memory in his talk at CHEP I believe. He said >>>>>>> it xrootd uses a lot of memory but there weren't number specified. >>>>>>> >>>>>>> My system is double cpu with 2GB of memory per cpu. Considering >>>>>>> that an atlas job can use more than 1 GB we are now at 1.1 I >>>>>>> thik. Will it be enough? I think so but I just wanted to check. >>>>>>> >>>>>>> cheers >>>>>>> alessandra >>>>>>> >>>>>>> >>>>>>> Peter Elmer wrote: >>>>>>>> Hi All, >>>>>>>> >>>>>>>> There is now a new xrootd development version: xrootd >>>>>>>> 20060523-1741, please see: >>>>>>>> http://xrootd.slac.stanford.edu/download/20060523-1741/ >>>>>>>> >>>>>>>> for downloads. >>>>>>>> >>>>>>>> Relative to the last development build (20060418-0404) this >>>>>>>> includes a variety of small bug fixes, plus one important one >>>>>>>> for the redirector. See >>>>>>>> the xrootd.History file for more details. I've included a link >>>>>>>> to the >>>>>>>> SL3 debuginfo rpm on the rpm page. (Although I've not tried it >>>>>>>> myself, >>>>>>>> so I have no idea if it works! Feedback is welcome.) >>>>>>>> >>>>>>>> Gerri, if it is still possible, you could add this to the >>>>>>>> next ROOT >>>>>>>> build, too. >>>>>>>> >>>>>>>> For the full set of changes and links to rpms/tarballs to >>>>>>>> download see the >>>>>>>> the xrootd web page and/or version history: >>>>>>>> >>>>>>>> http://xrootd.slac.stanford.edu >>>>>>>> http://xrootd.slac.stanford.edu/xrootd.History >>>>>>>> >>>>>>>> Let us know if there are problems. >>>>>>>> >>>>>>>> Pete >>>>>>>> >>>>>>>> ------------------------------------------------------------------------- >>>>>>>> >>>>>>>> >>>>>>>> Peter Elmer E-mail: [log in to unmask] Phone: +41 >>>>>>>> (22) 767-4644 >>>>>>>> Address: CERN Division PPE, Bat. 32 2C-14, CH-1211 Geneva 23, >>>>>>>> Switzerland >>>>>>>> ------------------------------------------------------------------------- >>>>>>>> >>>>>>>> >>>>>>> >>>>>> >>>>> >>>>> >>>>> ------------------------------------------------------------------------ >>>>> >>>>> >>>>> >>>>> ------------------------------------------------------------------------ >>>>> >>>>> >>>>> >>>>> ------------------------------------------------------------------------ >>>>> >>>>> >>>>> >>>>> ------------------------------------------------------------------------ >>>>> >>>>> >>>> >>> >> >> -- >> ******************************************* >> * Dr Alessandra Forti * >> * Technical Coordinator - NorthGrid Tier2 * >> * http://www.hep.man.ac.uk/u/aforti * >> ******************************************* >> -- ,,,,, ( o o ) --m---U---m-- Jerome