Connection lost among nodes
 
I am having trouble scaling up a cluster to handle multiple samples.  A
test with one 1Tb attached EBS drive and 30 nodes worked great.  Moving to
10 EBS drives and 300 nodes crashed at about 70-80 nodes.  There are no
errors in the STDOUT or STDERR output of jobs.  Looking the in node
messages (/opt/sge6/spool/...), most nodes have a commlib error got read
error , followed by the job failing causing a shepherd error.  The cluster
is set up on a private subnet and as I mentioned I have 10 EBS drives
attached which may be too much?  I was also pushing the nodes with intense
work, with an nload_average of 3-4.  My guess is the load is increased by
having a larger cluster and causing the issue.  Does anyone know if that
could be the problem or have any other ideas?
Thanks,
  Nick
Received on Fri Mar 27 2015 - 10:49:38 EDT
This archive was generated by
hypermail 2.3.0.