问题描述
我了解使用 subprocess 是调用外部命令的首选方式.
I understand using subprocess is the preferred way of calling external command.
但是,如果我想并行运行多个命令,但要限制生成的进程数怎么办?困扰我的是我无法阻止子进程.例如,如果我调用
But what if I want to run several commands in parall, but limit the number of processes being spawned? What bothers me is that I can't block the subprocesses. For example, if I call
subprocess.Popen(cmd, stderr=outputfile, stdout=outputfile)
然后该过程将继续,无需等待 cmd 完成.因此,我无法将其包装在 multiprocessing 库的工作人员中.
Then the process will continue, without waiting for cmd to finish. Therefore, I can't wrap it up in a worker of multiprocessing library.
例如,如果我这样做:
def worker(cmd):
subprocess.Popen(cmd, stderr=outputfile, stdout=outputfile);
pool = Pool( processes = 10 );
results =[pool.apply_async(worker, [cmd]) for cmd in cmd_list];
ans = [res.get() for res in results];
然后每个工作人员将在生成子进程后完成并返回.所以我不能通过使用Pool 来真正限制subprocess 生成的进程数.
then each worker will finish and return after spawning a subprocess. So I can't really limit the number of processes generated by subprocess by using Pool.
限制子进程数量的正确方法是什么?
What's the proper way of limiting the number of subprocesses?
推荐答案
如果要等待命令完成,可以使用subprocess.call.有关详细信息,请参阅 pydoc 子进程.
You can use subprocess.call if you want to wait for the command to complete. See pydoc subprocess for more information.
您也可以调用 Popen.wait 你的工人中的方法:
You could also call the Popen.wait method in your worker:
def worker(cmd):
p = subprocess.Popen(cmd, stderr=outputfile, stdout=outputfile);
p.wait()
这篇关于在python中控制用于调用外部命令的子进程数的文章就介绍到这了,希望我们推荐的答案对大家有所帮助,也希望大家多多支持跟版网!


大气响应式网络建站服务公司织梦模板
高端大气html5设计公司网站源码
织梦dede网页模板下载素材销售下载站平台(带会员中心带筛选)
财税代理公司注册代理记账网站织梦模板(带手机端)
成人高考自考在职研究生教育机构网站源码(带手机端)
高端HTML5响应式企业集团通用类网站织梦模板(自适应手机端)