Libraries

PS-Lite Documents - ps-lite 1.0 documentation

Allreduce: Collective Reduction

Technologies behind Distributed Deep Learning: AllReduce

Interface: result = allreduce(float buffer[size])

grad = gradient(net, w)

for epoch, data in enumerate(dataset):
	g = net.run(grad, in=data)
	gsum = comm.allreduce(g, op=sum)
	w -= lr * gsum / num_workers

Tree Shape Reduction

o

o

Ring based Reduction

time complexity:

Parameter Server

Interface: key-value store

ps.push(index, gradient) && ps.pull(index)

Data Consistency: BSP