I don't know the reason why there is a 16M limit and it clearly is not ideal for Ceph or HDFS type storage systems. Matevz, do you know? Can the limit be removed? Andy On Fri, 6 Sep 2019, Sam Skipsey wrote: > Hm. Adding that line to the config prevents the xrootd cache service from > staying up (and removing it makes it work again). > > 190906 19:03:37 19779 XrdFileCache_a2x: get block size 64M may not be > greater than 16777216 > 190906 19:03:37 19779 XrdFileCache_Manager: error Cache::Config() error in > parsing > > Setting it to the maximum, > > pfc.blocksize 16M > > does significantly improve the rate though (I get about 80 to 90% of the > direct rate - around 300MB/s). > > thanks. > > Is it possible to have the arbitrary limit on pfc.blocksize removed in > later releases? > > Sam > > > > On Fri, 6 Sep 2019 at 18:03, Yang, Wei < > [log in to unmask]> wrote: > >> Hi Sam, >> >> >> >> One thing that maybe useful for Xcache to talk to a CEPH storage is to set >> the Xcache "*pfc.blocksize*" to the same as the CEPH native bucket size. >> So if the CEPH bucket size is 64M, then you can add the following to >> cephc02's config >> >> >> >> pfc.blocksize 64M >> >> >> >> The default is 1M >> >> >> >> regards, >> >> -- >> >> Wei Yang | [log in to unmask] | 650-926-3338 (O) >> >> >> >> >> >> *From: *<[log in to unmask]> on behalf of Sam Skipsey < >> [log in to unmask]> >> *Date: *Friday, September 6, 2019 at 6:20 AM >> *To: *xrootd-l <[log in to unmask]> >> *Subject: *Help debugging slow Xrootd Proxy Cache <-> xrootd server >> transfers >> >> >> >> Hello everyone: >> >> >> >> I'm currently seeing some odd behaviour, and would appreciate some insight >> from people more deeply aware of xrootd than I am. >> >> >> >> Currently we are testing an xrootd setup using internal Xrootd disk >> caching proxies to improve performance against a Ceph object store. >> >> >> >> The configuration is: >> >> >> >> Xrootd server [with ceph plugin] on cephs02.beowulf.cluster >> >> >> >> Xrootd cache on cephc02.beowulf.cluster >> >> Cache is backed by a software raid-0 array of 6 SSDs. The SSD array >> achieves > 1.5GB/s transfer rates when tested, and is not an I/O bottleneck. >> >> Cephc02 is configured as a direct proxy for cephs02. >> >> >> >> Firewall is configured so all nodes are in the trusted zone relative to >> each other, so no ports blocked. >> >> >> >> The problem is that connections proxied through cephc02's xrootd server >> are extremely slow (10x to 15x slower) than direct connections to the >> xrootd server on cephs02. >> >> >> >> From cephc02, directly copying from cephs02: >> >> >> >> [root@cephc02 ~]# time xrdcp root://cephs02:1095/ecpool:testfile2GB >> testfile2GB-in-1 >> >> [1.863GB/1.863GB][100%][==================================================][381.5MB/s] >> >> >> >> versus connecting via the proxy on cephc02 (cold cache): >> >> [root@cephc02 ~]# time xrdcp -v -v -v root:// >> 10.1.50.12:1094/ecpool:testfile2GB test-from-cache2 >> >> [1.863GB/1.863GB][100%][==================================================][20.51MB/s] >> >> >> >> (once the cache is warm, fetching from the cache itself is v fast, at > >> 1GB/s) >> >> >> >> >> >> Whilst I'd expect some caching overhead, this seems unuseably slow. >> >> What am I doing wrong here? >> >> >> >> Any help appreciated, >> >> >> >> Sam Skipsey >> >> University of Glasgow >> >> >> >> >> >> The Cache and Server authenticate with a shared secret, and their relevant >> configs are: >> >> >> >> cephc02: >> >> >> >> [root@cephc02 ~]# cat /etc/xrootd/xrootd-cache.cfg >> ofs.osslib libXrdPss.so >> pss.cachelib libXrdFileCache.so >> pfc.ram 16g >> pfc.trace info >> pfc.diskusage 0.90 0.95 >> oss.localroot /cache/ >> >> all.export /xroot:/ >> all.export /root:/ >> all.export * >> pss.origin 10.1.50.2:1095 >> xrd.allow host *.beowulf.cluster >> >> #inbound security protocol, for authenticating to the xrootd-ceph >> #sec.protocol >> xrootd.seclib /usr/lib64/libXrdSec.so >> sec.protocol sss -s /etc/gateway/xrootd/sss.keytab.grp -c >> /etc/gateway/xrootd/sss.keytab.grp >> sec.protbind cephs02.beowulf.cluster:1095 only sss >> >> #outside security protocol, for authenticating users wanting to use the >> proxy >> #sec.protocol >> sec.protbind localhost only none >> >> xrd.report 127.0.0.1:9527 every 5s all >> >> >> >> - >> >> >> >> cephs02: >> >> >> >> # The export directive indicates which paths are to be exported. While the >> # default is '/tmp', we indicate it anyway to show you this directive. >> # >> all.export *? >> all.export / >> >> # The adminpath and pidpath variables indicate where the pid and various >> # IPC files should be placed >> # >> all.adminpath /var/spool/xrootd >> all.pidpath /var/run/xrootd >> xrootd.async segsize 67108864 >> xrd.buffers maxbsz 67108864 >> >> # Configure sss security - this is a shared secret between the xrootd-ceph >> and the xrootd-proxies, so the proxies are trusted to talk to the >> xrootd-ceph >> # >> xrootd.seclib /opt/xrootd/lib64/libXrdSec.so >> sec.protocol sss -s /etc/gateway/xrootd/sss.keytab.grp -c >> /etc/gateway/xrootd/sss.keytab.grp >> sec.protbind * only sss >> >> xrootd.seclib /opt/xrootd/lib64/libXrdSec.so >> #sec.protocol host >> #sec.protbind localhost none >> >> # Configure rados connection >> # /this needs to be configured for the right stripe width >> ofs.osslib +cksio /opt/xrootd/lib64/libXrdCeph.so admin@ecpool >> ,1,8388608,83886080 >> ofs.xattrlib /opt/xrootd/lib64/libXrdCephXattr.so >> xrootd.chksum adler32 >> >> # Configure the port >> # >> xrd.port 1095 >> >> >> >> >> ------------------------------ >> >> Use REPLY-ALL to reply to list >> >> To unsubscribe from the XROOTD-L list, click the following link: >> https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=XROOTD-L&A=1 >> >> ------------------------------ >> >> Use REPLY-ALL to reply to list >> >> To unsubscribe from the XROOTD-L list, click the following link: >> https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=XROOTD-L&A=1 >> > > ######################################################################## > Use REPLY-ALL to reply to list > > To unsubscribe from the XROOTD-L list, click the following link: > https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=XROOTD-L&A=1 > ######################################################################## Use REPLY-ALL to reply to list To unsubscribe from the XROOTD-L list, click the following link: https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=XROOTD-L&A=1