Decentralised learning with distributed gradient descent and random features
We investigate the generalisation performance of Distributed Gradient Descent with implicit regularisation and random features in the homogenous setting where a network of agents are given data sampled independently from the same unknown distribution. Along with reducing the memory footprint, random...
Main Authors: | , , |
---|---|
格式: | Conference item |
语言: | English |
出版: |
Proceedings of Machine Learning Research
2020
|