Print

Print


Hello,

Osman Aidel, a CC-IN2P3 expert in databases administration, try 
currently to load in MySQL the 3TB dataset produced during last data 
challenge.
Osman and Dominique Boutigny succeeded in loading the whole dataset in 
MySQL, but some post-processing steps of this dataset (like removal of 
duplicates) take a infinite time.

Please remark that these issues are good news for Qserv as it validates 
its distributed data model ;-).

Christian Arnault, french manager for LSST computing, thinks that some 
of the tools developed by Qserv team could help Osman, and CC-IN2P3.
Indeed, the partitioning algorithm developed by Serge could be used to 
partition the DC dataset in a collection of chunks.
Osman could then load a part of contiguous chunks of this collection 
into a single-node MySQL server.

Do you think this proposal could be attainable soon ? Indeed, French 
physicists are interested in studying a representative sample of the DC 
dataset and this solution would help them a lot.

Thanks,

Fabrice

########################################################################
Use REPLY-ALL to reply to list

To unsubscribe from the QSERV-L list, click the following link:
https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=QSERV-L&A=1