Hi Jan, at the moment such a primitive is not part of the protocol. The simpler way of doing it is to call Stat for each file, but this reduces the per-file overhead only by a small amount, with respect to a Open call. In fact, both primitives actually drive the client to the final endpoint (the file), so you cannot avoid the overhead (mainly communication latencies) of being redirected to other servers. Since you say it's critical for you, my suggestion is to open as many files as you can in the parallel way. Doing so, all the latencies are parallelized, and you can expect a much higher performance. To do this, just call TFile::AsyncOpen(fname) for each file you need to open (a cycle), and then, later, you can call the regular TFile::Open (another cycle). The async call is non-blocking and very fast. You can find an example of its ROOT-based usage here: http://root.cern.ch/root/Version512.news.html The ugly thing is that doing this you are using a lot of resources, so, if you have really a lot of files to open (let's say, 5000) and the resources are a problem, maybe you can find a workaround by opening them in bunches of fixed size. Fabrizio Jan Iwaszkiewicz wrote: > Hi, > > In PROOF we realized that we need a possibility to query exact locations > of a set of files. As far as I have seen in the xrootd protocol, there > is no way to ask for locations of a vector of files. > > At the beginning of a query, we want to check exact locations of all the > files form a data set. The current implementation does it by opening all > the files, one by one. > The speed is about 30 files/sec. For many queries, the lookup takes much > longer than the processing. > It is a critical problem for us. > > The bool XrdClientAdmin::SysStatX(const char *paths_list, kXR_char > *binInfo) method can check multiple files but it only verifies whether > the files exist. > I imagine that it would be best for us to have something similar but > returning file locations. Is such an extension to the protocol > possible/reasonable to implement? > > Cheers, > Jan