[Beowulf] integrating node disks into a cluster filesystem?

Joshua Baker-LePain jlb17 at duke.edu
Fri Sep 25 18:32:36 EDT 2009


On Fri, 25 Sep 2009 at 6:09pm, Mark Hahn wrote

> but since 1U nodes are still the most common HPC building block, and most of 
> them support 4 LFF SATA disks with very little added cost (esp using the 
> chipset's integrated controller), is there a way to integrate them into a 
> whole-cluster filesystem?

This is something I've considered/toyed-with/lusted after for a long 
while.  I haven't pursued it as much as I could have because the clusters 
I've run to this point have generally run embarrassingly parallel jobs, 
and I train the users to cache data-in-progress to scratch space on the 
nodes.  But there's a definite draw to a single global scratch space that 
scales automatically with the cluster itself.

> - obviously want to minimize the interference of remote IO to a node's jobs.
>  for serial jobs, this is almost moot.  for loosely-coupled parallel jobs
>  (whether threaded or cross-node), this is probably non-critical.  even for
>  tight-coupled jobs, perhaps it would be enough to reserve a core for
>  admin/filesystem overhead.

I'd also strongly consider a separate network for filesystem I/O.

> - distributed filesystem (ceph?  gluster?  please post any experience!)  I
>  know it's possible to run oss+ost services on a lustre client, but not
>  recommended because of the deadlock issue.

I played with PVFS1 a bit back in the day.  My impression at the time was 
they they were focused on MPI-IO, and the POSIX layer was a bit of an 
afterthought -- access with "regular" tools (tar, cp, etc) was pretty 
slow.  I don't know what the situation is with PVFS2.  Anyone?

> - this is certainly related to more focused systems like google/mapreduce.
>  but I'm mainly looking for more general-purpose clusters - the space would
>  be used for normal files, and definitely mixed read/write with something
>  close to normal POSIX semantics...

It seems we're after the same thing.

-- 
Joshua Baker-LePain
QB3 Shared Cluster Sysadmin
UCSF
_______________________________________________
Beowulf mailing list, Beowulf at beowulf.org sponsored by Penguin Computing
To change your subscription (digest mode or unsubscribe) visit http://www.beowulf.org/mailman/listinfo/beowulf



More information about the Beowulf mailing list