[StarCluster] StarCluster Digest, Vol 49, Issue 1

Lyn Gerner schedulerqueen at gmail.com
Sun Sep 1 13:08:52 EDT 2013


Hey Jacob,

Have you looked at the qsub man page -hold_jid option?  You can give it
multiple jids or jobnames (or regex patterns).
http://gridscheduler.sourceforge.net/htmlman/manuals.html
http://gridscheduler.sourceforge.net/htmlman/htmlman1/sge_types.html<http://gridscheduler.sourceforge.net/htmlman/htmlman1/sge_types.html?pathrev=V62u5_TAG>

This preso has a couple of examples:
http://www.bioteam.net/wp-content/uploads/2011/03/02-SGE-SimpleWorkflow.pdf

Best of luck,
Lyn



On Sun, Sep 1, 2013 at 6:26 AM, <starcluster-request at mit.edu> wrote:

> Send StarCluster mailing list submissions to
>         starcluster at mit.edu
>
> To subscribe or unsubscribe via the World Wide Web, visit
>         http://mailman.mit.edu/mailman/listinfo/starcluster
> or, via email, send a message with subject or body 'help' to
>         starcluster-request at mit.edu
>
> You can reach the person managing the list at
>         starcluster-owner at mit.edu
>
> When replying, please edit your Subject line so it is more specific
> than "Re: Contents of StarCluster digest..."
>
> Today's Topics:
>
>    1. SGE priorities and job dependency (Jacob Barhak)
>
>
> ---------- Forwarded message ----------
> From: Jacob Barhak <jacob.barhak at gmail.com>
> To: "starcluster at mit.edu" <starcluster at mit.edu>
> Cc:
> Date: Sun, 1 Sep 2013 02:19:12 -0500
> Subject: [StarCluster] SGE priorities and job dependency
> Hello,
>
> Does someone have experience with the SGE scheduler that comes with
> StarCluster? Experienced enough to figure out how make a dependent job
> launch before other jobs once its dependencies are gone?
>
> I have been trying to give the dependent job a high priority, yet it seems
> the scheduler ignores this and launches the jobs in FIFO order.
>
> Here is a simplified description of my problem. Lets say I have 100k jobs.
> The first 50k are generating files in a shared directory and the last 50k
> are processing those files and deleting them.
>
> Jobs 1-50000 are independent while job 50001 is dependent on job 50000 and
> job 50000+n is dependent on job n.
>
> I tried lowering the priority of the first 50k jobs using qsub -p -100. I
> was hoping to get the effect of job 1 completing and since job 50001
> dependency was satisfied and has the highest priority then it would be
> launched next. The idea is to perform cleanup of the file after each job -
> otherwise too many files can accumulate in a shared directory and may slow
> down the NFS I/O significantly - not to mention disk space.
>
> However, I cannot get this behavior from SGE on a single test machine
> outside StarCluster. So I assume this needs some special configuration.
>
> I am trying to avoid the I/O bottleneck I experienced on the cloud due to
> too many files in a shared directory.
>
> Can someone help with this without changing the order if the jobs when
> being launched?
>
> I hope there is a simple one line / qsub option for this.
>
>        Jacob
>
>
>
> Sent from my iPhone
>
>
> _______________________________________________
> StarCluster mailing list
> StarCluster at mit.edu
> http://mailman.mit.edu/mailman/listinfo/starcluster
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://mailman.mit.edu/pipermail/starcluster/attachments/20130901/527b5248/attachment.htm


More information about the StarCluster mailing list