Print

Print


Hi Alessandra,

  typically the memory needs of a server are proportional to the number 
of clients and to the ongoing activity rates. For a small number of 
clients (let's say 1-100) the memory needs are quite low. I've done many 
many stress tests in a pIII machine with 256 megs with no problems at 
all. The memory becomes an important parameter when you plan thousands 
of clients per server.

Fabrizio

Alessandra Forti wrote:
> Apologies that worries me in using xrootd on the WNs....
> I'm a bit confused right now :(
> 
> Alessandra Forti wrote:
>> Hi Peter,
>>
>> after yesterday I also remember that the second thing that worried me 
>> in using dcache on the WN is a comment Andy made about the use of 
>> memory in his talk at CHEP I believe. He said it xrootd uses a lot of 
>> memory but there weren't number specified.
>>
>> My system is double cpu with 2GB of memory per cpu. Considering that 
>> an atlas job can use more than 1 GB we are now at 1.1 I thik. Will it 
>> be enough? I think so but I just wanted to check.
>>
>> cheers
>> alessandra
>>
>>
>> Peter Elmer wrote:
>>>    Hi All,
>>>
>>>   There is now a new xrootd development version: xrootd 
>>> 20060523-1741, please see:
>>>   http://xrootd.slac.stanford.edu/download/20060523-1741/
>>>
>>> for downloads.
>>>
>>>   Relative to the last development build (20060418-0404) this 
>>> includes a variety of small bug fixes, plus one important one for the 
>>> redirector. See
>>> the xrootd.History file for more details. I've included a link to the
>>> SL3 debuginfo rpm on the rpm page. (Although I've not tried it myself,
>>> so I have no idea if it works! Feedback is welcome.)
>>>
>>>   Gerri, if it is still possible, you could add this to the next ROOT
>>> build, too.
>>>
>>>    For the full set of changes and links to rpms/tarballs to download 
>>> see the
>>>  the xrootd web page and/or version history:
>>>
>>>     http://xrootd.slac.stanford.edu
>>>     http://xrootd.slac.stanford.edu/xrootd.History
>>>
>>>  Let us know if there are problems.
>>>
>>>                                    Pete
>>>
>>> ------------------------------------------------------------------------- 
>>>
>>> Peter Elmer     E-mail: [log in to unmask]      Phone: +41 (22) 
>>> 767-4644
>>> Address: CERN Division PPE, Bat. 32 2C-14, CH-1211 Geneva 23, 
>>> Switzerland
>>> ------------------------------------------------------------------------- 
>>>
>>
>