WARNING: This is the _old_ Lustre wiki, and it is in the process of being retired. The information found here is all likely to be out of date. Please search the new wiki for more up to date information.

Difference between revisions of "Guidelines for Setting Up a Cluster"

From Obsolete Lustre Wiki
Jump to navigationJump to search
Line 1: Line 1:
 
Listed below are some guidelines for setting up a cluster to make it easier to manage and debug.
 
Listed below are some guidelines for setting up a cluster to make it easier to manage and debug.
  
*  '''Set up shared home directories.''' A shared namespace is useful for bringing up Lustre builds, collecting logs, using a ''blat'' command-line utility to email configuration files, etc.  The most commonly shared namespace is /home.
+
*  '''Set up shared home directories.''' A shared namespace is useful for bringing up Lustre builds, collecting logs, using a ''blat'' command-line utility to email configuration files, etc.  The most commonly shared namespace is ''/home''.
 
*  '''Use ''pdsh''.''' This parallel-distributed, multithreaded remote shell enables efficient execution of commands on multiple remote hosts in parallel.
 
*  '''Use ''pdsh''.''' This parallel-distributed, multithreaded remote shell enables efficient execution of commands on multiple remote hosts in parallel.
 
*  '''Use a regular node naming scheme.''' A node naming scheme consisting of a short prefix combined with regularly incremented decimal node numbers (e.g., n0001, n0002, etc.) works well with an automated tool like ''pdsh''.  Also, machines tend to be used for different roles in a cluster over time, so hostnames based on roles in the Lustre file system (mds, ost, etc) are not always practical. However, documenting how hostnames map to Lustre functions is useful.
 
*  '''Use a regular node naming scheme.''' A node naming scheme consisting of a short prefix combined with regularly incremented decimal node numbers (e.g., n0001, n0002, etc.) works well with an automated tool like ''pdsh''.  Also, machines tend to be used for different roles in a cluster over time, so hostnames based on roles in the Lustre file system (mds, ost, etc) are not always practical. However, documenting how hostnames map to Lustre functions is useful.
 
*  '''Use serial consoles.''' A serial console enables output to be logged for later retrieval in case a problem occurs. It can be provided with a useful front end like ''conman'' or ''conserver''. A front end that can send breaks to the kernel's ''sysrq'' facility over the serial console is preferable.   
 
*  '''Use serial consoles.''' A serial console enables output to be logged for later retrieval in case a problem occurs. It can be provided with a useful front end like ''conman'' or ''conserver''. A front end that can send breaks to the kernel's ''sysrq'' facility over the serial console is preferable.   
  
:In 2.6 kernels, reliable network-based consoles allow sending (nearly) all kernel messages to a remote system, even ''oops'' messages.  In 2.6.5, ''netconsole'' is provided. In 2.6.9 and later,  ''netdump'' supercedes ''netconsole''.  The ''netdump'' code also supports kernel crash dumps over the network to another host, which can be invaluable for debugging node-crashing problems.
+
:In 2.6 kernels, reliable network-based consoles allow sending (nearly) all kernel messages to a remote system, even ''oops'' messages.  In 2.6.5, ''netconsole'' is provided. In 2.6.9 and later,  ''netdump'' supercedes ''netconsole''.  The ''netdump'' code also supports kernel crash dumps over the network to another host, which can be invaluable for debugging node-crashing problems. [[Is this paragraph up to date?  Elsewhere we say netdump has been superceded by kdump.]]
  
 
* '''Collect syslogs in one place.''' In addition to collecting logs on a per node basis, collecting syslogs in one location lets an administrator monitor a single log for errors reported to ''syslog'' from across the cluster.
 
* '''Collect syslogs in one place.''' In addition to collecting logs on a per node basis, collecting syslogs in one location lets an administrator monitor a single log for errors reported to ''syslog'' from across the cluster.

Revision as of 11:34, 19 November 2009

Listed below are some guidelines for setting up a cluster to make it easier to manage and debug.

  • Set up shared home directories. A shared namespace is useful for bringing up Lustre builds, collecting logs, using a blat command-line utility to email configuration files, etc. The most commonly shared namespace is /home.
  • Use pdsh. This parallel-distributed, multithreaded remote shell enables efficient execution of commands on multiple remote hosts in parallel.
  • Use a regular node naming scheme. A node naming scheme consisting of a short prefix combined with regularly incremented decimal node numbers (e.g., n0001, n0002, etc.) works well with an automated tool like pdsh. Also, machines tend to be used for different roles in a cluster over time, so hostnames based on roles in the Lustre file system (mds, ost, etc) are not always practical. However, documenting how hostnames map to Lustre functions is useful.
  • Use serial consoles. A serial console enables output to be logged for later retrieval in case a problem occurs. It can be provided with a useful front end like conman or conserver. A front end that can send breaks to the kernel's sysrq facility over the serial console is preferable.
In 2.6 kernels, reliable network-based consoles allow sending (nearly) all kernel messages to a remote system, even oops messages. In 2.6.5, netconsole is provided. In 2.6.9 and later, netdump supercedes netconsole. The netdump code also supports kernel crash dumps over the network to another host, which can be invaluable for debugging node-crashing problems. Is this paragraph up to date? Elsewhere we say netdump has been superceded by kdump.
  • Collect syslogs in one place. In addition to collecting logs on a per node basis, collecting syslogs in one location lets an administrator monitor a single log for errors reported to syslog from across the cluster.
  • Set up remote power management. If a machine wedges, it must be possible to reboot it without physically flipping a switch. Various vendors offer serial-controlled power widgets. Power widgets that work with powerman are the most useful. Remote power management is a requirement for doing automated failover (STONITH).
  • Automate node provisioning. Although infrequently used, it's convenient to be able to reimage a node via netbooting and network software installs.
  • Boot quickly. To be able to boot quickly, do the following:
    • Disable non-essential services from starting at boot-time.
    • Minimize hardware checks made by the BIOS.
    • Avoid starting utilities at boot-time that ask for user input before proceeding.