Hi Patrick, We have a new development release on the web page. This release should address all of the problems you raised. Specifically, look at the "xa" option in the 'oss.cache' directive. Andy ----- Original Message ----- From: "Patrick McGuigan" <[log in to unmask]> To: "Andrew Hanushevsky" <[log in to unmask]> Cc: <[log in to unmask]> Sent: Monday, March 03, 2008 4:01 PM Subject: Re: oss.cache is too limited, any advice > Hi Andy, > > I'll let you know how the r/o migration attempt works when I try it later > tonight. > > I think it would be useful to remove the pathname limit under caching, > although I am biased ;) I expect that ATLAS Tier3 sites will take a long > look at Scalla because of the integration with root and the more flexible > the software is, the more likely it will fit the perceived need. I am > about to help setup a Tier3 site where the existing hardware provides > three disks per data server and would be willing to test implementations > there. > > Patrick > > Andrew Hanushevsky wrote: >> Hi Patrick, >> >> Yes, you can do what you want to do. The most generc way is to export the >> paths as r/o (read/only). See the xrootd.export directive. You can do >> what you want, though we have never tried it. That is, having a server >> exporting a read/only path that you essentially want to duplicate on a >> read/write server(s). It probably will work. If not, there are a couple >> of other things you can do to force it to work. BTW to keep a single >> config file, simply bracket the exceptional export in an if/fi construct >> rereferencing that host. For instance, >> >> if <problem host> >> xrootd.export /thepath r/o >> else >> xrootd.export /thepath >> fi >> >> I understand your concern about granularity. That problem seems to plague >> many LVMs. The newest ones (i.e., zfs) try to address that issue. So, >> does that mean you think it's worth spending time removing the 255 path >> limit? >> >> Andy >> >> >> On Mon, 3 Mar 2008, Patrick McGuigan wrote: >> >>> Hi Andy, >>> >>> Is there a way to to enforce a read-only rule for particular data >>> servers? If this is possible, I can ensure that newly written data >>> avoids the systems that need to be "LVM'ed", while allowing the existing >>> data to be read. I am curious if I can replicate data under this >>> scenario using xrdcp, or will I still have to take the read-only systems >>> off-line to move the data? >>> >>> I am concerned about the granularity of recovery actions in our systems >>> under LVM, but I need to support larger pathnames now. The pathnames >>> are being driven by the physics users and the use of metadata in the >>> path components. >>> >>> Patrick >>> >>> Andrew Hanushevsky wrote: >>>> Hi Patrick, >>>> >>>> You are quite right that the design of the cache does impose limits on >>>> file names. The oss was designed almost 10 years ago when there were no >>>> integrated LVM's and few that worked really well and users kept path >>>> names to less that 200 characters. Over the years, as LVM's became >>>> common, the oss turned into the "poor man's LVM". In general, we don't >>>> recommend using it if you have ready access to a suitable LVM. While, >>>> yes, you do give up some features (like fine-grained recoverability and >>>> application-directed partition selection) the other >>>> limitations may be even more annoying. We'll make sure that this >>>> restriction is prominently mentioned in the manual. >>>> >>>> The path you've chosen is about the only one that will work (i.e., >>>> copying off the data and creating a single filesystem using an LVM). >>>> >>>> Now, we do have some ideas on how to remove the pathname length >>>> restriction but wonder if it's really worth the trouble of doing it, >>>> given that LVM's provide practically the same basic features. Any >>>> thoughts? >>>> >>>> Andy >>>> >>>> P.S. What's the driving force for very long path names at your site? >>>> >>>> >>>> On Sat, 1 Mar 2008, Patrick McGuigan wrote: >>>> >>>>> Hi, >>>>> >>>>> Some of our data servers have two disks and I am using the oss.cache >>>>> directive to use both disks to support a single namespace. However, >>>>> it looks like the users of our applications have already run into a >>>>> problem. All of the files are stored in one directory (for a single >>>>> cache) and the filename is the full namespace path with "/" >>>>> substituted with "%". Our problem arises from the fact that full >>>>> namespace path is now limited to the leafname length of the filesystem >>>>> (255 characters) when writing to the cache directory. >>>>> >>>>> I see a couple of ways to mediate the problem; removing one disk or >>>>> using and LVM to create one drive in the OS. I am curious if there >>>>> are other alternatives? >>>>> >>>>> If I have to move to one disk, I would like to migrate the data in the >>>>> existing caches to other data servers while I rework the existing >>>>> system. What is the best way to migrate this data? I am planning on >>>>> taking the "problem" data server off-line and use xrdcp to move the >>>>> data to the other servers. >>>>> >>>>> Regards, >>>>> >>>>> Patrick >>>>> >>> >