Hi Alessandra, typically the memory needs of a server are proportional to the number of clients and to the ongoing activity rates. For a small number of clients (let's say 1-100) the memory needs are quite low. I've done many many stress tests in a pIII machine with 256 megs with no problems at all. The memory becomes an important parameter when you plan thousands of clients per server. Fabrizio Alessandra Forti wrote: > Apologies that worries me in using xrootd on the WNs.... > I'm a bit confused right now :( > > Alessandra Forti wrote: >> Hi Peter, >> >> after yesterday I also remember that the second thing that worried me >> in using dcache on the WN is a comment Andy made about the use of >> memory in his talk at CHEP I believe. He said it xrootd uses a lot of >> memory but there weren't number specified. >> >> My system is double cpu with 2GB of memory per cpu. Considering that >> an atlas job can use more than 1 GB we are now at 1.1 I thik. Will it >> be enough? I think so but I just wanted to check. >> >> cheers >> alessandra >> >> >> Peter Elmer wrote: >>> Hi All, >>> >>> There is now a new xrootd development version: xrootd >>> 20060523-1741, please see: >>> http://xrootd.slac.stanford.edu/download/20060523-1741/ >>> >>> for downloads. >>> >>> Relative to the last development build (20060418-0404) this >>> includes a variety of small bug fixes, plus one important one for the >>> redirector. See >>> the xrootd.History file for more details. I've included a link to the >>> SL3 debuginfo rpm on the rpm page. (Although I've not tried it myself, >>> so I have no idea if it works! Feedback is welcome.) >>> >>> Gerri, if it is still possible, you could add this to the next ROOT >>> build, too. >>> >>> For the full set of changes and links to rpms/tarballs to download >>> see the >>> the xrootd web page and/or version history: >>> >>> http://xrootd.slac.stanford.edu >>> http://xrootd.slac.stanford.edu/xrootd.History >>> >>> Let us know if there are problems. >>> >>> Pete >>> >>> ------------------------------------------------------------------------- >>> >>> Peter Elmer E-mail: [log in to unmask] Phone: +41 (22) >>> 767-4644 >>> Address: CERN Division PPE, Bat. 32 2C-14, CH-1211 Geneva 23, >>> Switzerland >>> ------------------------------------------------------------------------- >>> >> >