Actually, the DC crowd here is really focused on distributed client computing, not clusters (at least that's how I read 95% of their posts). The best site is probably a site like clustermonkey.net, which is an offshoot of clusterworld.com (which got sold to a linux magazine a few months ago and has since deceased).
To briefly answer your question: there are a LOT of reasons to build a cluster, and the details of how you fit it out very much depend upon it's use. Is it for scientific computing (High Performance Computing, or HPC) use? Then you will care a lot about inter-node connectivity (using Infiband or GigaEthernet) to get the nodes communicating between themselves as fast as possible, as many scientific needs have to pass partial results sets between nodes. Think about weather forecasting - the results of one "cell" of the weathermap obviously affect the results in the neighboring areas, and as the forecast is run each node has to pass that data on to it's surrounding nodes at each step in the simulation. So high-speed node connects are important for a lot of those scientific uses.
If you are using your cluster for a database server, the emphasis is different. Internode communication isn't that important (GigaE is more than enough usually, as the only information that is transmitted between nodes is usually data lock information), but connectivity from the nodes to the disk arrays is critical. In most database clusters, either high-performance RAID arrays are directly attached to the individual nodes, or the entire cluster is wired to a SAN. The former gives higher performance but is a pain to administer, the latter can have a SAN-connectivity bottleneck but is easier to build and administer.
Some clusters may go for iSCSI to wire the nodes to the NAS/SAN, but this is usually restricted to non-disk intensive applications (such as HPC), due to the overhead inherent with iSCSI.
If rolling your own cluster, again, depending upon what you want to do with it, I might suggest a linux variant that runs cluster apps well. Many commercial cluster apps (such as Oracle 10g) require Red Hat or Suse to work out of the box, as they are intended for commercial sites, but I know of people that have had success running CentOS (a freeware Red Hat distro) as a basis for Oracle 10g installs. Fedora has just enough differences that many RH-specific applications are a pain to get working, due to different directory structures, etc.
Lastly, from a software side, I suggest you take a look at globus.org, a commercial/academic effort to standardize grid/cluster middleware. Interesting reading. Also check out the hardware vendor's sites - I know hp.com has a lot of VERY detailed information as to how to configure their own kit on their site - check out their line of blade processors (BL-series) and look for the technical white papers for download.
If you have specific questions, PM me and I can try to at least point you in the right direction...
Future Shock