Hello, Osman Aidel, a CC-IN2P3 expert in databases administration, try currently to load in MySQL the 3TB dataset produced during last data challenge. Osman and Dominique Boutigny succeeded in loading the whole dataset in MySQL, but some post-processing steps of this dataset (like removal of duplicates) take a infinite time. Please remark that these issues are good news for Qserv as it validates its distributed data model ;-). Christian Arnault, french manager for LSST computing, thinks that some of the tools developed by Qserv team could help Osman, and CC-IN2P3. Indeed, the partitioning algorithm developed by Serge could be used to partition the DC dataset in a collection of chunks. Osman could then load a part of contiguous chunks of this collection into a single-node MySQL server. Do you think this proposal could be attainable soon ? Indeed, French physicists are interested in studying a representative sample of the DC dataset and this solution would help them a lot. Thanks, Fabrice ######################################################################## Use REPLY-ALL to reply to list To unsubscribe from the QSERV-L list, click the following link: https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=QSERV-L&A=1