[StarCluster] FW: commlib error
Amanda Joy Kedaigle
mandyjoy at mit.edu
Tue Sep 23 14:56:16 EDT 2014
Thanks for that plugin suggestion! I tried avoiding the master node through "qsub -l", and so far that seems to be helping!
As for using a t2.micro or small, those don't seem to be an option in starcluster? They are not listed as instance options in my config file, and when I previously tried an option outside of those listed, I got an error. If anyone knows a way around that, I'd be interested!
Thanks to both of you
Amanda
________________________________
From: MacMullan, Hugh [hughmac at wharton.upenn.edu]
Sent: Tuesday, September 23, 2014 2:50 PM
To: Amanda Joy Kedaigle
Cc: starcluster at mit.edu
Subject: RE: [StarCluster] FW: commlib error
Amanda:
I agree with Rajat's t2 suggestion … even just a t2.micro will help over a t1.micro … and it's cheaper!
And are you running jobs on the master as well as the nodes? If so, you could disable that:
[cluster mycluster]
DISABLE_QUEUE=True
PLUGINS = sge
[plugin sge]
setup_class = starcluster.plugins.sge.SGEPlugin
master_is_exec_host = False
That might help with stability a good bit.
You can also use spot pricing for the master to get a beefier master for a much lower price … but of course with the risk of losing the whole cluster if you are outbid.
Good luck with the project!
-Hugh
From: starcluster-bounces at mit.edu [mailto:starcluster-bounces at mit.edu] On Behalf Of Rajat Banerjee
Sent: Tuesday, September 23, 2014 1:11 PM
To: Amanda Joy Kedaigle
Cc: starcluster at mit.edu
Subject: Re: [StarCluster] FW: commlib error
HI Amanda,
I googled your error and found a few pages that suggest that sge service on the master node went down:
http://verahill.blogspot.com/2012/08/sun-gridengine-commlib-error-got-select.html
https://supcom.hgc.jp/english/utili_info/manual/faq.html
http://comments.gmane.org/gmane.comp.clustering.gridengine.users/17283
If your OpenBLAS command is killing the process on master that could cause your issues according to those authors. Sorry I don't have anything more helpful, but the t2.small is still less than $.03 per hour now. That may not increase your costs too much.
Raj
On Tue, Sep 23, 2014 at 12:55 PM, Amanda Joy Kedaigle <mandyjoy at mit.edu<mailto:mandyjoy at mit.edu>> wrote:
Thanks, Raj. I can communicate with the master node, it just looks like SGE is failing. I restarted the cluster and everything seemed to be working, but then it just failed in the same way again.
> starcluster listclusters (should list status of all your active clusters and running nodes)
-----------------------------------------------------
fraenkelcluster (security group: @sc-fraenkelcluster)
-----------------------------------------------------
Launch time: 2014-09-23 11:59:43
Uptime: 0 days, 00:45:58
VPC: vpc-c71f0fa5
Subnet: subnet-e6b8c8ce
Zone: us-east-1c
Keypair: fraenkel-keypair
EBS volumes:
vol-5e75ba11 on master:/dev/sdz (status: attached)
Cluster nodes:
master running i-acc76242 ec2-54-164-81-80.compute-1.amazonaws.com<http://ec2-54-164-81-80.compute-1.amazonaws.com>
node001 running i-5177ddbf ec2-54-164-98-38.compute-1.amazonaws.com<http://ec2-54-164-98-38.compute-1.amazonaws.com>
node002 running i-9976c077 ec2-54-164-88-184.compute-1.amazonaws.com<http://ec2-54-164-88-184.compute-1.amazonaws.com>
node003 running i-9e76c070 ec2-54-164-38-146.compute-1.amazonaws.com<http://ec2-54-164-38-146.compute-1.amazonaws.com>
node004 running i-1776c0f9 ec2-54-86-252-119.compute-1.amazonaws.com<http://ec2-54-86-252-119.compute-1.amazonaws.com>
node005 running i-1676c0f8 ec2-54-165-66-3.compute-1.amazonaws.com<http://ec2-54-165-66-3.compute-1.amazonaws.com>
Total nodes: 6
> starcluster sshmaster <your cluster name>
works just fine, I am ssh'd into master under root user.
Some more details: I am wondering if this is because my master node is a t1.micro - either it is an older generation and not updated, or doesn't have enough memory to run the queue? When doing my initial tests, running thousands of simple jobs, it worked fine, and the load balancer added and deleted nodes as expected. However, when running slightly more intensive jobs, including the python module networkx, the jobs give this error and then SGE dies:
OpenBLAS : Your OS does not support AVX instructions. OpenBLAS is using Nehalem kernels as a fallback, which may give poorer performance.
Killed
I would really like to have a very cheap master node since I expect to keep it running 24/7, but only use the cluster in bursts.
On Mon, Sep 22, 2014 at 5:13 PM, Amanda Joy Kedaigle <mandyjoy at mit.edu<mailto:mandyjoy at mit.edu>> wrote:
Hi,
I am trying to run starcluster's loadbalancer to keep only one node running until jobs are submitted to the cluster. I know it's an experimental feature, but I'm wondering if anyone has run into this error before, or has any suggestions. The cluster has been whittled down to 1 node after a weekend of inactivity, and now it seems that when jobs are submitted to the queue, instead of adding nodes, SGE fails.
>>> Loading full job history
*** WARNING - Failed to retrieve stats (1/5):
Traceback (most recent call last):
File "/net/dorsal/apps/python2.7/lib/python2.7/site-packages/StarCluster-0.95.5-py2.7.egg/starcluster/balancers/sge/__init__.py", line 552, in get_stats
return self._get_stats()
File "/net/dorsal/apps/python2.7/lib/python2.7/site-packages/StarCluster-0.95.5-py2.7.egg/starcluster/balancers/sge/__init__.py", line 522, in _get_stats
qhostxml = '\n'.join(master.ssh.execute('qhost -xml'))
File "/net/dorsal/apps/python2.7/lib/python2.7/site-packages/StarCluster-0.95.5-py2.7.egg/starcluster/sshutils.py", line 578, in execute
msg, command, exit_status, out_str)
RemoteCommandFailed: remote command 'source /etc/profile && qhost -xml' failed with status 1:
error: commlib error: got select error (Connection refused)
error: unable to send message to qmaster using port 63231 on host "master": got send error
Thanks for any help!
Amanda
_______________________________________________
StarCluster mailing list
StarCluster at mit.edu<mailto:StarCluster at mit.edu>
http://mailman.mit.edu/mailman/listinfo/starcluster
_______________________________________________
StarCluster mailing list
StarCluster at mit.edu<mailto:StarCluster at mit.edu>
http://mailman.mit.edu/mailman/listinfo/starcluster
-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://mailman.mit.edu/pipermail/starcluster/attachments/20140923/d3e01020/attachment.htm
More information about the StarCluster
mailing list