Mario thanks, forwarding to the qserv mailing list Jacek On 09/24/2013 02:26 PM, Mario Juric wrote: > Jacek, > Here's something for you and the team to think about in November: how > would you modify qserv to download and cache chunks on-demand? > > Imagine the following scenario: scientist M at university H has access > to 200+ nodes w. petabytes of storage. They install qserv, and configure > it in a mode where it initially downloads only the metadata from the > LSST Archive (e.g., where the chunks are, how many of them are there, > which tables are where, etc.). When a query is fired against that > database, it downloads the chunks touched by the query and caches them > for future use. Note that if no query ever touched a particular table, > those chunks would never get downloaded (e.g., a stellar astronomer > would never download the table with bdSamples). After an initial period > of downloading/caching, the database would asymptote to a steady state > mode where all frequently queried tables are cached locally (thus > offloading our database). Bonus points for allowing bootstrapping with > mass-downloaded data (or sneakernet delivered disks), or expiration of > rarely accessed chunks. > > This was a killer feature for LSD [*] (combined with on-the-fly > cross-matching against remote catalogs, which is a simple extension). > I'm willing to bet people would use it if we supplied it. It may also > solve our issue with the number of replicas needed to guard against > failure, since we could configure our Archive center database to fetch > any chunks that it doesn't have (e.g., because the nodes have failed) > from the Chilean or French site. > > [*] http://mwscience.net/trac/wiki/ReleaseNotes_0_5_4 > > Have fun :), > ######################################################################## Use REPLY-ALL to reply to list To unsubscribe from the QSERV-L list, click the following link: https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=QSERV-L&A=1