Hello,
I've been reading the documentation on xrootd, and I'm wondering how it
might work in practice for a particular workload I'm trying to optimize.
xrootd performs very well to provide high aggregate read performance for
one or many clients. Can someone provide information, or point me
towards published results, on write intensive and concurrent read/write
intensive applications? How are large data sets normally loaded into
xrootd?
I'm trying to understand how well xrootd would perform if there was say
a stream of data (about 25GB over 8 hours each day, peak write event
rates of 125,000 messages a second) which had to be written while
simultaneously there was a number of clients attempting to analysis the
data - either large historical analyses or perhaps just interested in
the last 10 minutes worth of events.
Any pointers?
Regards,
Niall
|