[Beowulf] QDR InfiniBand interconnect architectures ... approaches ...

Craig Tierney Craig.Tierney at noaa.gov
Thu Apr 8 14:42:49 EDT 2010


richard.walsh at comcast.net wrote:
> 
> All, 
> 
> 
> What are the approaches and experiences of people interconnecting 
> clusters of more than128 compute nodes with QDR InfiniBand technology? 
> Are people directly connecting to chassis-sized switches? Using multi-tiered 
> approaches which combine 36-port leaf switches? What are your experiences? 
> What products seem to be living up to expectations? 
> 
> 
> I am looking for some real world feedback before making a decision on 
> architecture and vendor. 
> 
> 

We have been telling our vendors to design a multi-level tree using
36 port switches that provides approximately 70% bisection bandwidth.
On a 448 node Nehalem cluster, this has worked well (weather, hurricane, and
some climate modeling).  This design (15 up/21 down) allows us to
scale the system to 714 nodes.

Craig






> Thanks, 
> 
> 
> rbw 
> 
> 
> ------------------------------------------------------------------------
> 
> _______________________________________________
> Beowulf mailing list, Beowulf at beowulf.org sponsored by Penguin Computing
> To change your subscription (digest mode or unsubscribe) visit http://www.beowulf.org/mailman/listinfo/beowulf

_______________________________________________
Beowulf mailing list, Beowulf at beowulf.org sponsored by Penguin Computing
To change your subscription (digest mode or unsubscribe) visit http://www.beowulf.org/mailman/listinfo/beowulf



More information about the Beowulf mailing list