[Beowulf] recommendation on crash cart for a cluster room: full cluster KVM is not an option I suppose?
gerry.creager at tamu.edu
Wed Sep 30 09:53:01 EDT 2009
Hearns, John wrote:
> I like the shared socket approach. Building a separate IPMI network
> seems a lot of extra wiring to me. Admittedly the IPMI switches can be
> configured to be dirt cheap but it still feels like building a extra
> tiny road for one car a day when a huge highway with spare capacity
> exists right next door carrying thousands of cars. (Ok, cheesy
> Errrr.... you missed all my Beowulf posts about the clashes with the
> IPMI ports
> and the ports used for 'rsh' connections on a cluster then? And all the
> with setting sunrpc.min_resvport etc.?
> Having a separate, simple IPMI network which comes up when you power the
> racks up
> has a lot of advantages. 10/100 Netgear switches cost almost nothing,
> and getting
> another loom of Cat5 cables configured when the racks are being built is
> relatively easy.
> By the way, which hardware do you use?
We've been down both paths. On our recent acquisition, we ended up with
separate, dedicated IPMI ports, despite our spec stating we wanted
shared socked ports. I bought 4 Netgear switches and added
infrastructure cabling. Having been down both paths, now, in the last
year (nothing is too old to have the memory clear in my mind) I
definitely have decided the completely separate IPMI network plan is
superior overall. I wish I could retrofit the Dell cluster to
accomplish this, but it ain't gonna happen.
It's a much cleaner (from a cluster management view) approach, IMNSHO.
Beowulf mailing list, Beowulf at beowulf.org sponsored by Penguin Computing
To change your subscription (digest mode or unsubscribe) visit http://www.beowulf.org/mailman/listinfo/beowulf
More information about the Beowulf