Print

Print


Hi Matevz (including xrootd list again which i forgot in the last reply),

> Well, if for some reason more new files are placed on a single disk,
> those files will be "newer" and purge would preferentially wipe data
> off other disks.
Mhhhh - then i have an idea how i may have triggered this. As mentioned 
in my first email the issue started after i updated my container image 
and had to change the xrootd user ids. This changes the Access time of 
the files - if that is used by xrootd to determine which files are newer 
than it could just be that the chown process walked this directory last 
and therefore will purge it last.
When i then deleted it when the disk ran full i made the problem even 
worse since now all the files that end up there are recently accessed.

So deleting the whole cache should fix it?

Cheers,
Nikolai

On 2/16/23 10:50, Matevz Tadel wrote:
> Hi Andy, Nikolai,
> 
> On 2/15/23 23:51, Andrew Hanushevsky wrote:
>> Hi Nikolai,
>>
>> Hmm, this sounds like an off by one problem in Xcache.
> 
> How? XCache does not do disks, it just uses oss API to a pool.
> 
>> The question is what is
>> the "one". It does seem that ity consistently does not purge files from a
>> particular disk but then again it doesn't know about disks. So, there is some
>> systematic issue that resolves to ignoring a disk. Matevz?
> 
> Well, if for some reason more new files are placed on a single disk, those files
> will be "newer" and purge would preferentially wipe data off other disks.
> 
> That's why I asked in the first email how disks are selected for new files and
> if we could inject some debug printouts there.
> 
> Perhaps a coincidence, but the full disk is the one that is listed first by df.
> 
> The docs say default for oss.alloc fuzz = 0 and that this "forces oss to always
> use the partition with the largest amount of free space" -- so the fuller one
> should never get selected for new files. And xcache does pass the appropriate
> oss.asize opaque parameter to open.
> 
> https://xrootd.slac.stanford.edu/doc/dev56/ofs_config.htm#_Toc116508676
> 
> Matevz
> 
>> Andy
>>
>>
>> On Thu, 16 Feb 2023, Nikolai Hartmann wrote:
>>
>>> Hi Andy,
>>>
>>> The behavior seems to be that it purges all the disks except one. After the
>>> other disks now again surpassed the threshold of 95% it seemed to trigger the
>>> cleanup and now i have this:
>>>
>>> Filesystem                 Type      Size  Used Avail Use% Mounted on
>>> /dev/sdb                   btrfs     5,5T  5,3T  215G  97% /srv/xcache/b
>>> /dev/sda                   btrfs     5,5T  5,0T  560G  90% /srv/xcache/a
>>> /dev/sdh                   btrfs     5,5T  4,9T  588G  90% /srv/xcache/h
>>> /dev/sdj                   btrfs     5,5T  4,9T  584G  90% /srv/xcache/j
>>> /dev/sdf                   btrfs     5,5T  4,9T  580G  90% /srv/xcache/f
>>> /dev/sdm                   btrfs     5,5T  5,0T  535G  91% /srv/xcache/m
>>> /dev/sdc                   btrfs     5,5T  5,0T  553G  91% /srv/xcache/c
>>> /dev/sdg                   btrfs     5,5T  4,9T  612G  90% /srv/xcache/g
>>> /dev/sdi                   btrfs     5,5T  4,9T  596G  90% /srv/xcache/i
>>> /dev/sdl                   btrfs     5,5T  5,0T  518G  91% /srv/xcache/l
>>> /dev/sdn                   btrfs     5,5T  4,9T  570G  90% /srv/xcache/n
>>> /dev/sde                   btrfs     5,5T  4,9T  593G  90% /srv/xcache/e
>>> /dev/sdk                   btrfs     5,5T  4,8T  677G  88% /srv/xcache/k
>>> /dev/sdd                   btrfs     5,5T  4,9T  602G  90% /srv/xcache/d
>>>
>>> Cheers,
>>> Nikolai
>>>
>>> On 2/14/23 21:52, Andrew Hanushevsky wrote:
>>>> Hi Matevz & Nikolai,
>>>>
>>>> The allocation should favor the disk with the most free space unless it's
>>>> atered using the oss.alloc directive:
>>>> https://urldefense.com/v3/__https://xrootd.slac.stanford.edu/doc/dev54/ofs_config.htm*_Toc89982400__;Iw!!Mih3wA!AsisYxoXis_6IdoiqK-BwdMsHfHTB41Z4-GEjaMqvO0PQHh6TqU8Sn79JUgDeJDLCvO63yQiG63Zu6syVA$
>>>> I don't think Nikolai specifies that and I don't think the pfc alters it in
>>>> any way. So, I can't explain why we see that difference other than via an
>>>> uneven purge.
>>>>
>>>> Andy
>>>>
>>>>
>>>> On Tue, 14 Feb 2023, Matevz Tadel wrote:
>>>>
>>>>> Hi Nikolai, Andy,
>>>>>
>>>>> I saw this a long time back, 2++ years. The thing is that xcache does oss df on
>>>>> the whole space and then deletes files without any knowledge of the usage on
>>>>> individual disks themselves. Placement of new files should prefer the more
>>>>> empty
>>>>> disks though, iirc.
>>>>>
>>>>> I remember asking Andy about how xcache could be made aware of individual disks
>>>>> and he prepared something for me but it got really complicated when I was
>>>>> trying
>>>>> to include this into the cache purge algorithm so I think I dropped this.
>>>>>
>>>>> Andy, could we sneak some debug printouts into oss new file disk selection to
>>>>> see if something is going wrong there?
>>>>>
>>>>> Nikolai, how fast does this happen? Is it a matter of days, ie, over many purge
>>>>> cycles? Is it always the same disk?
>>>>>
>>>>> Cheers,
>>>>> Matevz
>>>>>
>>>>> On 2/13/23 23:21, Nikolai Hartmann wrote:
>>>>>> Hi Andy,
>>>>>>
>>>>>> The config is the following:
>>>>>>
>>>>>> https://urldefense.com/v3/__https://gitlab.physik.uni-muenchen.de/etp-computing/xcache-nspawn-lrz/-/blob/086e5ade5d27fc7d5ef59448c955523e453c091f/etc/xrootd/xcache.cfg__;!!Mih3wA!DfZZQn5-SZKaGYvPW97K8SD5gDYYTy0wuUgMgQCUMhwQehl01yhKQdErjCRUz3BoZYL_nKVipwRIRYyR$
>>>>>> The directories for `oss.localroot` and `oss.space meta` are on the system
>>>>>> disk.
>>>>>> The `/srv/xcache/[a-m]` are individually mounted devices.
>>>>>>
>>>>>> Best,
>>>>>> Nikolai
>>>>>>
>>>>>> On 2/14/23 00:34, Andrew Hanushevsky wrote:
>>>>>>> Hi Nikolai,
>>>>>>>
>>>>>>> Hmmm, no it seems you are the first one. Then again, not many people have a
>>>>>>> multi-disk setup.  So, could you send a link to your config file? It might be
>>>>>>> the case that all of the metadata files wind up on the same disk and that is
>>>>>>> the source of the issue here.
>>>>>>>
>>>>>>> Andy
>>>>>>>
>>>>>>> On Mon, 13 Feb 2023, Nikolai Hartmann wrote:
>>>>>>>
>>>>>>>> Dear xrootd-l,
>>>>>>>>
>>>>>>>> I'm seeing the issue that one of the disks on one of our xcache servers
>>>>>>>> fills
>>>>>>>> up disproportionally - that means it runs completely full until i get "no
>>>>>>>> space left on device" errors without xcache running cleanup, while the other
>>>>>>>> disks still have plenty of space left. My current df output:
>>>>>>>>
>>>>>>>> /dev/sdb                   btrfs     5,5T  5,2T  273G  96% /srv/xcache/b
>>>>>>>> /dev/sda                   btrfs     5,5T  4,9T  584G  90% /srv/xcache/a
>>>>>>>> /dev/sdh                   btrfs     5,5T  5,0T  562G  90% /srv/xcache/h
>>>>>>>> /dev/sdj                   btrfs     5,5T  5,0T  551G  91% /srv/xcache/j
>>>>>>>> /dev/sdf                   btrfs     5,5T  4,9T  579G  90% /srv/xcache/f
>>>>>>>> [...]
>>>>>>>>
>>>>>>>> If you look at the first line you see that disk is 96% full while the others
>>>>>>>> are around 90%. The issue occurred the first time after i built a new
>>>>>>>> container for running xrootd. That change involved switching the container
>>>>>>>> from centos7 to almalinux8 and changing the xrootd user id (ran chown and
>>>>>>>> chgrp afterwards on the cache directories which are bind mounted). The
>>>>>>>> xrootd
>>>>>>>> version stayed the same (5.4.2). The high/low watermark configuration is the
>>>>>>>> following:
>>>>>>>>
>>>>>>>> pfc.diskusage 0.90 0.95
>>>>>>>>
>>>>>>>> I already tried clearing the misbehaving disk (after it ran full to 100%),
>>>>>>>> but now the issue is reappearing. Has anyone seen similar issues or does it
>>>>>>>> ring any bells for you?
>>>>>>>>
>>>>>>>> One thing i checked is the size that xrootd reports in the log for the total
>>>>>>>> storage and that at least matches what i get when i sum the entries from
>>>>>>>> `df`.
>>>>>>>>
>>>>>>>> Cheers,
>>>>>>>> Nikolai
>>>>>>>>
>>>>>>>> ########################################################################
>>>>>>>> Use REPLY-ALL to reply to list
>>>>>>>>
>>>>>>>> To unsubscribe from the XROOTD-L list, click the following link:
>>>>>>>> https://urldefense.com/v3/__https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=XROOTD-L&A=1__;!!Mih3wA!DfZZQn5-SZKaGYvPW97K8SD5gDYYTy0wuUgMgQCUMhwQehl01yhKQdErjCRUz3BoZYL_nKVip_SnON6x$
>>>>>
>>>
> 

########################################################################
Use REPLY-ALL to reply to list

To unsubscribe from the XROOTD-L list, click the following link:
https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=XROOTD-L&A=1