Print

Print


Hello all,

Just wanted to let the list know that the solution to this was to add 
the user_xattr option in the fstab entry for all the filesystems on my 
data server nodes.  We switched to ext4, from ext3, but they still 
needed the user_xattr option.

The docs describe this here: 
http://xrootd.org/doc/prod/frm_migr.htm#_Toc284350516

Though I found that the command to test is:

  frm_admin -c /path/to/cfile convert -t old2new

Cheers,
ksb

On 02/17/12 16:53, Keith Beattie wrote:
> Here it is:
>
> -----
> all.role server
> all.manager santa 3121
>
> all.adminpath /export/data/xrd/admin
> all.pidpath /export/data/xrd/admin
>
> all.export / nostage
> oss.localroot /export/data/xrd/ns
>
> oss.space public /export/data/xrd/data
> oss.space public /export/data/xrd/data
> oss.space public /export/data1/xrd/data
> oss.space public /export/data1/xrd/data
> oss.space public /export/data2/xrd/data
> oss.space public /export/data2/xrd/data
> oss.space public /export/data3/xrd/data
> oss.space public /export/data3/xrd/data
> ------
>
> On 02/17/12 16:46, Andrew Hanushevsky wrote:
>> Hi Keith,
>>
>> Please send me your config file.
>>
>> Andy
>>
>> On Fri, 17 Feb 2012, Keith Beattie wrote:
>>
>>> ok, I spoke too soon, this problem seems to be back.
>>>
>>> I've added data server nodes, now 13 of them, each with 3 drives and
>>> only the first drive is getting populated with any data - except for 1
>>> of them. After loading the cluster with lots of files, the first drive
>>> on all of them are at 100% usage with the rest of the drives at 3-5%
>>> usage. The one outlier's 2nd drive is at 20% usage. All these drives
>>> are 6-700 GB total size.
>>>
>>> Digging deeper on each data server node, on all of them (except for
>>> that outlier) oss.localroot (/export/data/xrd/ns/) is not getting
>>> populated with symlinks into the 'oss.space public' locations
>>> (/export/data*/xrd/data/) but the data files are directly there.
>>> Further, there *are* lots of files under the oss.space public
>>> locations (e.g. /export/data2/xrd/data/public/1B/271B3B4FDB000000176%)
>>> but they are all empty!
>>>
>>> On the one outlier there are symlinks to the 2nd drive, but that only
>>> accounts for ~200 of the ~1000 entries under oss.localroot there.
>>>
>>> The config files for each of the data server nodes are identical.
>>>
>>> I tried duplicating the oss.space public lines in the configs, as was
>>> previously suggested but that didn't seem to change anything.
>>>
>>> I'll try some other test, like using xrootd 3.0.5, different configs,
>>> but these test take a while to complete, so I'm hoping this rings some
>>> bells here as to what I might be missing.
>>>
>>> Thanks,
>>> ksb
>>>
>>> On 01/30/12 12:47, Keith Beattie wrote:
>>>> to follow up here...
>>>>
>>>> I've tried it again, with the same configs, and now it works. To the
>>>> best of my knowledge nothing is different, except the absence of this
>>>> problem - both drives on all the clients are receiving data.
>>>>
>>>> I'd rather know what fixed it, but I'll take it and move on.
>>>>
>>>> Thanks for your help,
>>>> ksb
>>>>
>>>> On 12/20/11 5:08 PM, Keith Beattie wrote:
>>>>> Here you go, attached. Both from xrootd and cmsd.
>>>>>
>>>>> Thanks,
>>>>> ksb
>>>>>
>>>>> On 12/16/11 11:55 PM, Wilko Kroeger wrote:
>>>>>>
>>>>>> Hello Keith
>>>>>>
>>>>>> I am using the release 3.1.0 at slac with cache systems (using
>>>>>> oss.space
>>>>>> public ...) and files are placed in all cache systems.
>>>>>> Could you maybe post the startup message from the xrdlog?
>>>>>>
>>>>>> Cheers,
>>>>>> Wilko
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Fri, 16 Dec 2011, Keith Beattie wrote:
>>>>>>
>>>>>>> Hi Andy,
>>>>>>>
>>>>>>> I'm using 3.1.0, built from source. I tried doubling the oss.cache
>>>>>>> lines and the servers failed to start. When using oss.space rather
>>>>>>> than oss.cache, doubling those lines isn't fatal, but regardless I
>>>>>>> get
>>>>>>> the same behavior - data only to the first entry.
>>>>>>>
>>>>>>> Thanks,
>>>>>>> ksb
>>>>>>>
>>>>>>> On 12/16/11 6:06 PM, Andrew Hanushevsky wrote:
>>>>>>>> Hi Keith,
>>>>>>>>
>>>>>>>> You may not be missing anything, depending on what version you are
>>>>>>>> using. There was a bug that exhibited exactly this behaviour. The
>>>>>>>> bypass
>>>>>>>> (other than upgrading to atleast 3.0.2) is to list the oss.cache
>>>>>>>> entries
>>>>>>>> twice for each path.
>>>>>>>>
>>>>>>>> Andy
>>>>>>>>
>>>>>>>> On Fri, 16 Dec 2011, Keith Beattie wrote:
>>>>>>>>
>>>>>>>>> Hello all,
>>>>>>>>>
>>>>>>>>> This seems like I'm missing something simple but can't seem to
>>>>>>>>> find
>>>>>>>>> it.
>>>>>>>>>
>>>>>>>>> On my data servers nodes, only the first 'oss.cache' entry is
>>>>>>>>> getting
>>>>>>>>> data, the following disks don't seem to ever get written to.
>>>>>>>>> Below is
>>>>>>>>> an example of what the config looks like on the data nodes (not
>>>>>>>>> santa
>>>>>>>>> where the manager is running). What am I missing?
>>>>>>>>>
>>>>>>>>> ----
>>>>>>>>> all.role server
>>>>>>>>> all.manager santa 3121
>>>>>>>>>
>>>>>>>>> all.export /
>>>>>>>>> oss.localroot /data/ns
>>>>>>>>> oss.cache public /data/xrddata
>>>>>>>>> oss.cache public /data1/xrddata
>>>>>>>>> ----
>>>>>>>>>
>>>>>>>>> i.e. /data/xrddata gets filled, /data1/xrddata does not.
>>>>>>>>>
>>>>>>>>> Thanks,
>>>>>>>>> ksb
>>>
>>> ########################################################################
>>> Use REPLY-ALL to reply to list
>>>
>>> To unsubscribe from the XROOTD-L list, click the following link:
>>> https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=XROOTD-L&A=1
>>>
>
> ########################################################################
> Use REPLY-ALL to reply to list
>
> To unsubscribe from the XROOTD-L list, click the following link:
> https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=XROOTD-L&A=1

########################################################################
Use REPLY-ALL to reply to list

To unsubscribe from the XROOTD-L list, click the following link:
https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=XROOTD-L&A=1