since 'qstat -j 2' shows me
error reason 8: 08/22/2016 15:31:35 [1000:44925]: unable to find job file "/opt/sge6/default/spool/exec_spool_local/mynew1-node001/job_scripts/2"
error reason 9: 08/22/2016 15:31:35 [1000:44926]: unable to find job file "/opt/sge6/default/spool/exec_spool_local/mynew1-node001/job_scripts/2"
this sounds a *lot* like the race condition described at
but adding '-b yes' doesn't seem to fix the problem. (there were no
embedded SGE options in my scriptfile)
Has anyone else encountered this? Found a work around?
08/22/2016 15:31:36| main|mynew1-node001|E|shepherd of job 2.8 exited with exit status = 11
08/22/2016 15:31:36| main|mynew1-node001|C|exec of mailer "/bin/mail" failed: "No such file or directory"
08/22/2016 15:31:36| main|mynew1-node001|E|shepherd of job 2.9 exited with exit status = 11
From: starcluster-bounces_at_mit.edu <starcluster-bounces_at_mit.edu> on behalf of Mike Cariaso <mike.cariaso_at_keygene.com>
Sent: Tuesday, August 23, 2016 12:13 AM
Subject: [StarCluster] workers go idle until a new worker is added ... ?
using the latest version from
I start a master node, and zero workers, and put an array job into the queue. I then then gradually add workers nodes. A new worker accepts as many tasks as the slots allow, but after they complete it never picks up additional work. When I add a new worker machine, it accepts some tasks and runs them successfully, but never goes back for more. Usually during this time one of the idle previous machines will also pickup some more tasks, but once those are finished it again sits waiting.
qstat -j 1.19 shows me 'unable to find job file "/opt/sge6/default/spool/exec_spool_local/mynew1-node002/job_scripts/1"'
and it's true that no file is there. When I add a new machine, the job appears, suggesting this isn't a file permission issue.
some nodes remain out of action.
starcluster addnode -x -a nodename clustername
doesn't seem to help.
Received on Mon Aug 22 2016 - 13:15:03 EDT