Print

Print


Hi Rene,
 
The mini-PetaCache at SLAC consisted of 10 * Sun v20z's each with dual 1.8GHz AMD Opteron cpus with 2GB main memory with a Chelsio or Neterion 10GE NIC and a 36 or 73GByte SCSI system disk; plus 4 * Sun v20z's each with dual 2.6GHz AMD Opteron cpus 4GB of main memory, with a Neterion or Chelsio 10GE NIC, a 73GB SCSI system disk and a dual 2 Gbits/s fibre channel connection to a Sun 3510 fibre channel 12 disk tray.

The mini-Petacache at SC05 had 10 Sun v20z's with dual 2.6GHz AMD Opteron cpus with 4GB main memory, a Neterion 10GE NIC and a 73GB SCSI system disk; plus 4 Sun v20z's with dual 1.8GHz AMD Opteron cpus with 2GB main memory a Chelsio 10GE NIC and a SCSI 36 GB SCSI system disk.  Each 2.6GHz cpu also had dual 2 Gbits/s fibre channel HBA's connected to Storcloud for a total of 20TBytes.  

We ran xrootd with 125 clients per host, and had 3 pairs of machines for each (of two) 10Gbits/s waves from SLAC to the SLAC/FNAL booth at SC05. The waves were provided by ESnet, one was a shared routed wave, the second was a dedicated layer 2 wave. Using standard (Linux 2.6.12 New Reno) TCP, we achieved about 9.8Gbits/s single direction (3 hosts to 3 hosts), and over 16Gbits/s peak (for 5 mins) for both directions simultaneously. 

For more details see: http://www-iepm.slac.stanford.edu/monitoring/bulk/sc2005/hiperf.html
-----Original Message-----
From: Andrew Hanushevsky [mailto:[log in to unmask]] 
Sent: Sunday, November 20, 2005 9:38 PM
To: Cottrell, Les
Subject: Re: FW: SC2005 Bandwidth Challenge Result (fwd)

Hi Les,

Could you give Rene the details (pleasew copy me); thanks.

Andy


---------- Forwarded message ----------
Date: Sun, 20 Nov 2005 08:14:40 +0100 (MET)
From: Rene Brun <[log in to unmask]>
To: Andrew Hanushevsky <[log in to unmask]>
Cc: Peter Elmer <[log in to unmask]>, Rene Brun <[log in to unmask]>,
     Fabrizio Furano <[log in to unmask]>,
     Gerardo Ganis <[log in to unmask]>,
     Fons Rademakers <[log in to unmask]>,
     Jean-Yves Nief <[log in to unmask]>, [log in to unmask]
Subject: Re: FW: SC2005 Bandwidth Challenge Result  (fwd)

Hi Andy,

Congratulations!

Could you give more details on teh setup on both sides, in particular the mini-Petacache at SLAC?

Rene

On Sat, 19 Nov
2005, Andrew Hanushevsky wrote:

> See the last paragraph where the SC05 bandwidth challenge was won 
> using xrootd. Seems like the server is the fastest thing out there today.
>
> Andy
>
> ---------- Forwarded message ----------
> Date: Thu, 17 Nov 2005 15:24:25 -0800
> From: "Cottrell, Les" <[log in to unmask]>
> To: [log in to unmask]
> Cc: "Rao, Nageswara S." <[log in to unmask]>, W. R. Wing <[log in to unmask]>,
>     scs-l <[log in to unmask]>, William E. Johnston <[log in to unmask]>,
>     Kevin Oberman <[log in to unmask]>,
>     "Calder, Neil" <[log in to unmask]>
> Subject: FW: SC2005 Bandwidth Challenge Result
>
> Attached is email from Harvey Newman concerning the Bandwidth Challenge.
>
> We (SLAC) need to especially thank UltraScienceNet and ESnet for the timely provision of the two ESnet waves (one routed, the other layer 2 via UltraScienceNet). Nagi Rao, Steven and Bill Wing from UltraScienNet and Kevin Oberman from ESnet provided invaluable support whenever requested.
>
> With each of these 10Gbits/s (per direction) we were able to successfully read and send ~ 15Gbits/sec over long periods (~8.5Gbits/s in one direction and ~6.5Gbits/s simultaneously in the other.), see sea01-05.jpg for the Esnet router layer 3 wave and sea06-10.jpg for the USN layer 2 wave. We also simultaneoulsy wrote about 3 Tbytes/hour to StorCloud. The aggregate from the SLAC/FNAL booth was about 45-55Gbits/s (see alllinks.jpg). The aggregate from SLAC and FNAL peaked around 150Gbits/s (see bw.jpg, note the read outs were at 20 second intervals but the display is averaged over a longer interval), and during the challenge we easily exceeded last years record of about 101 Gbits/s (which was suatained for about 100 seconds) during most of the challenge time.  BTW I just heard we did win this year's bandwidth challenge, awards this afternoon.
>
> With a single direction we were able to get up to 9.8Gbits/s.
>
> The main application we used was Andy Hanushevsky's xrootd between a mini-PetaCache cluster built in the SLAC booth and a mini-PetaCache cluster built at SLAC.
>
> -----Original Message-----
> From: Harvey Newman [mailto:[log in to unmask]]
> Sent: Thursday, November 17, 2005 10:21 AM
> To: Harvey Newman
> Cc: ultralight; Conrad Steenberg; Iosif LeGrand; Julian Bunn; Rick 
> Wilkinson; Suresh Singh; Xun Su; Saima Iqbal; Michael Thomas; 
> Frank.Van.Lingen; Yang Xia; Dan Nae; Bradley, W. Scott; Philippe 
> Galvez; ICFA SCIC; Michael Stanton; Doug Walsten; Philippe Levy; US 
> CMS Collaboration Board; US CMS Advisory Board; US CMS Level1/Level2; 
> Harvey Newman
> Subject: Re: SC2005 Bandwidth Challenge Result
>
>
> Dear Colleagues,
>
> Congratulations for a great job yesterday and throughout the week
>   by the HEP team.
>
> The SC2005 BWC from HEP was designed to preview the scale and complexity of data operations among many sites interconnected with many 10 Gbps links. We had 22 10 Gbps optical "waves" connected to the Caltech/CACR and FNAL/SLAC booths.
>
> We reached a measured peak of 150.7 Gbps, and sustained more than 100 
> Gbps for several hours using multiple applications based on TCP and in 
> many cases FAST: bbcp, xrootd, gridftp and dcache. We transported 470 
> Terabytes of physics data in under
> 24 hours.
>
> The SCInet Sc2005 network team assigned taps to monitor 17 of the waves at our booths and recorded a peak of 131 Gbps during a 15 minute measurement period last evening.
>
> We are awaiting official word but discussions with the judges last night indicated that we outpaced the competition by a wide margin.
>
> The exercise was not at all trivial. We needed to work through repeated  system and/or network interface crashes under stress.
> A great number of kernel, configuration and routing issues had to be worked out in the days before the BWC itself. It is a tribute to the team that these were all worked through successfully.
>
> The result was a great learning experience, and it had lasting value in several areas (a partial list):
>
> TCP optimization and Linux kernel building (including FAST); Performance optimization and tuning of applications:
>   bbcp and xrootd from SLAC gave surprisingly good results
>    for example
> Use of production and test clusters at FNAL reaching
>   more than 20 Gbps of network throughput  Stability limits of server and network interfaces (and heating)
>    under heavy loads.W. R. Wing [[log in to unmask]]
>
> We also were very pleased with the participation of our international
>   partners from Brazil, Japan and Korea who worked hard in the
>   days to weeks before the competition to be able to participate
>   effectively.
>
> More as the day progresses. A BWC award session will take place  at 15:30 - 17:00 Pacific time.
>
> Best regards
> Harvey
>
>