As we just talked, my numbers are for data chunks,
index is up to 2x larger, so we can use 2x larger
numbers. Data+index come in separate files, so
they can be transferred in parallel, so I think
it'd be unfair to assume 3x my numbers though
Jacek
On 9/24/2013 3:07 PM, Jacek Becla wrote:
>
>> Chunks are expected to be multiple terabytes in size, which
>> means that downloads are hours long.
>
>
> K-T,
>
> Based on the baseline, which assumes flat 20K chunks per tables,
> the largest chunk will be 255 GB. The numbers are (in GB,
> DR1 --> DR11)
> - Object: 2 --> 4
> - ObjExtra: 25 --> 69
> - Source: 9 --> 255
> - ForcedSrc: 2 --> 98
>
> This is in LDM-141, dbL2, L141 (and nearby)
>
> And, that is before compression.
>
> We talked about keeping chunk size const rather than #chunks
> constants, which will probably make us go with DR1-size chunk
> sizes, thus keeping chunk size closer to 25 GB than 1/4 TB)
>
> Jacek
>
########################################################################
Use REPLY-ALL to reply to list
To unsubscribe from the QSERV-L list, click the following link:
https://listserv.slac.stanford.edu/cgi-bin/wa?SUBED1=QSERV-L&A=1
|