[Beowulf] evaluating FLOPS capacity of our cluster
tom.elken at qlogic.com
Mon May 11 16:52:18 EDT 2009
> On Behalf Of Rahul Nabar
> Rmax/Rpeak= 0.83 seems a good guess based on one very similar system
> on the Top500.
> Thus I come up with a number of around 1.34 TeraFLOPS for my cluster
> of 24 servers. Does the value seem reasonable ballpark? Nothing too
> accurate but I do not want to be an order of magnitude off. [maybe a
> decimal mistake in math! ]
You're in the right ballpark. I recently got 0.245 Tflops on HPL on a 4-node version of what you have (with Goto BLAS), so 6x that # is in the same ballpark as your 1.34 TF/s estimate. My CPUs were 2.3 GHz Opteron 2356 instead of your 2.2 GHz.
Greg is also right on the memory size being a factor allowing larger N to be used for HPL. I used a pretty small N on this HPL run since we were running it as part of a HPC Challenge suite run, and a smaller N can be better for PTRANS if you are interested in the non-HPL parts of HPCC (as I was).
> All 64 bit machines with a dual channel
> bonded Gigabit ethernet interconnect. AMD Quad-Core AMD Opteron(tm)
> Processor 2354.
As others have said, 50% is a more likely HPL efficiency for a large GigE cluster, but with your smallish cluster (24 nodes) and bonded channels, you would probably get closer to 80% than 50%.
> PS. The Athelon was my typo, earlier sorry!
> Beowulf mailing list, Beowulf at beowulf.org sponsored by Penguin
> To change your subscription (digest mode or unsubscribe) visit
Beowulf mailing list, Beowulf at beowulf.org sponsored by Penguin Computing
To change your subscription (digest mode or unsubscribe) visit http://www.beowulf.org/mailman/listinfo/beowulf
This message has been scanned for viruses and
dangerous content by MailScanner, and is
believed to be clean.
More information about the Beowulf