mirror of
https://github.com/zebrajr/pytorch.git
synced 2026-01-15 12:15:51 +00:00
Increase the number of DAG execution worker threads.
Reviewed By: akyrola Differential Revision: D5158414 fbshipit-source-id: add377aec5588076db881a2a3750101710f29732
This commit is contained in:
committed by
Facebook Github Bot
parent
73a8a49c7e
commit
f0f4c2fc5d
@@ -259,25 +259,27 @@ def Parallelize_GPU_BMUF(
|
||||
master_gpu_opt = core.DeviceOption(caffe2_pb2.CUDA, master_gpu)
|
||||
|
||||
num_workers = len(devices)
|
||||
num_worker_threads = 4 * len(devices)
|
||||
loss_scale = 1.0 / num_workers
|
||||
if block_momentum is None:
|
||||
block_momentum = 1.0 - 1.0 / num_workers
|
||||
|
||||
model_helper_obj.net.Proto().num_workers = num_workers
|
||||
model_helper_obj.net.Proto().num_workers = num_worker_threads
|
||||
model_helper_obj.net.Proto().type = net_type
|
||||
|
||||
# A net for initializing global model parameters. Its called once in the
|
||||
# same step as net parameters initialization.
|
||||
model_helper_obj._global_model_init_net = core.Net('global_model_init')
|
||||
model_helper_obj._global_model_init_net.Proto().type = net_type
|
||||
model_helper_obj._global_model_init_net.Proto().num_workers = num_workers
|
||||
model_helper_obj._global_model_init_net.Proto().num_workers = \
|
||||
num_worker_threads
|
||||
|
||||
# A net for computing final parameter updates. Its will run once after
|
||||
# running net (local models updates) for `num_local_iterations` times.
|
||||
model_helper_obj._global_model_param_updates_net = core.Net('global_model')
|
||||
model_helper_obj._global_model_param_updates_net.Proto().type = net_type
|
||||
model_helper_obj._global_model_param_updates_net.Proto().num_workers = \
|
||||
num_workers
|
||||
num_worker_threads
|
||||
|
||||
def _v(param):
|
||||
return "{}_v".format(param)
|
||||
|
||||
Reference in New Issue
Block a user