-
Notifications
You must be signed in to change notification settings - Fork 129
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Dynamically scale clustermq workers #751
Comments
Fortunately, |
Thanks! maybe I misinterpreted #561, I was under the impression |
You interpreted the beginning of #561 correctly. Farther along in that thread, though, I realized we needed |
Would it be possible to kill the extra workers once |
Both parallel backends actually do this already, even library(drake)
plan <- drake_plan(A = 1, B = Sys.sleep(12))
options(clustermq.scheduler = "your_scheduler", clustermq.template = "your_template.tmpl")
make(plan, parallelism = "clustermq", jobs = 2) When I run the example above in SGE, I see that both workers start at first, but one of the workers quickly terminates after target |
I encounter an issue where I'm using more resources then needed. I have a set of simulations that need to be summarized before starting the next set. This loop iterates a few times. Now I switched to clustermq to future proof my simulation. The problem is that all the workers are persistent and thus wait until the last simulation is finished. Given the variation in simulation length this wastes resources. Is there a way to make the workers non persistent? I also will encounter issues when the total simulation takes longer then the maximally allowed duration. Non persistent workers would also address this.
The text was updated successfully, but these errors were encountered: