Print

Print


Hi Derek,

Derek Feichtinger wrote:
> Hi,
> 
> This is slightly off-topic, but nontheless important for the setup of large 
> direct attached storage systems typically used with xrootd. Maybe some of you 
> have good suggestions or experiences.
> 

  Well, I don't know exactly your requirements, but wouldn't it be 
sufficient to look at the traffic by making an average of the data seen 
by each client after the file close ?

  Another (better) way could be to setup XrdMon. Why not ?


Fabrizio


> For our next upgrade of our Tier2 I would need a benchmark with which I can 
> measure whether I can satisfy an I/O requirement per worker node (WN, or CPU 
> core). This has to be tested while all WNs are reading in parallel from all 
> file servers. I just want to assume that the clients from the WNs are reading 
> in a nicely distributed fashion from the file servers, e.g. in the case of 10 
> file servers and 150 WNs, I would assume that in average 15 WNs are reading 
> at the same time from any file server. But any combination of 15 WNs must be 
> able to yield the desired bandwidth.
> 
> Naturally, this benchmark is targeted at mimicking a cluster running analysis 
> applications.
> 
> A primitive test (but not exactly matching the use case) could be using 
> netperf or iperf in UDP mode. E.g. the file servers would receive packets 
> from the required fraction of worker nodes (The sending intervals and packet 
> sizes can be set for netperf). One would gradually increase the sending rate 
> per worker node until UDP packet loss is observed. 
> 
> I'm glad for any suggestions.
> 
> Cheers,
> Derek
>