Hello,
Osman Aidel, a CC-IN2P3 expert in databases administration, try
currently to load in MySQL the 3TB dataset produced during last data
challenge.
Osman and Dominique Boutigny succeeded in loading the whole dataset in
MySQL, but some post-processing steps of this dataset (like removal of
duplicates) take a infinite time.
Please remark that these issues are good news for Qserv as it validates
its distributed data model ;-).
Christian Arnault, french manager for LSST computing, thinks that some
of the tools developed by Qserv team could help Osman, and CC-IN2P3.
Indeed, the partitioning algorithm developed by Serge could be used to
partition the DC dataset in a collection of chunks.
Osman could then load a part of contiguous chunks of this collection
into a single-node MySQL server.
Do you think this proposal could be attainable soon ? Indeed, French
physicists are interested in studying a representative sample of the DC
dataset and this solution would help them a lot.
Thanks,
Fabrice
########################################################################
Use REPLY-ALL to reply to list
To unsubscribe from the QSERV-L list, click the following link:
https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=QSERV-L&A=1
|