From c5bcd9722627a4bfa4caeccdbb32e694d9d328f2 Mon Sep 17 00:00:00 2001 From: =?UTF-8?q?Aur=C3=A9lien?= <aurelien.bellet@inria.fr> Date: Fri, 2 Apr 2021 09:02:41 +0200 Subject: [PATCH] small fixes in refs --- main.bib | 24 ++++++------------------ main.tex | 2 +- 2 files changed, 7 insertions(+), 19 deletions(-) diff --git a/main.bib b/main.bib index 4050742..52b6cbd 100644 --- a/main.bib +++ b/main.bib @@ -605,12 +605,11 @@ pages={211-252} year={2010} } -@incollection{lian2017d-psgd, +@inproceedings{lian2017d-psgd, title = {{Can Decentralized Algorithms Outperform Centralized Algorithms? A Case Study for Decentralized Parallel Stochastic Gradient Descent}}, author = {Lian, Xiangru and Zhang, Ce and Zhang, Huan and Hsieh, Cho-Jui and Zhang, Wei and Liu, Ji}, - booktitle = {Advances in Neural Information Processing Systems}, - year = {2017}, - publisher = {Curran Associates, Inc.}, + booktitle = {NIPS}, + year = {2017} } @article{nedic2016sgp, @@ -674,8 +673,7 @@ pages={211-252} title = {{$D^2$: Decentralized Training over Decentralized Data}}, author = {Tang, Hanlin and Lian, Xiangru and Yan, Ming and Zhang, Ce and Liu, Ji}, booktitle = {ICML}, - year = {2018}, - publisher = {PMLR} + year = {2018} } @article{xiao2007distributed, @@ -721,7 +719,7 @@ pages={211-252} title={Small worlds: The dynamics of networks between order and randomness}, author={Watts, Duncan J}, year={2000}, - publisher={Princeton University Press Princeton} + publisher={Princeton University Press} } % Random Model Walk !!! @@ -790,15 +788,6 @@ pages={211-252} primaryClass={cs.CV} } -@misc{kong2021consensus, - title={Consensus Control for Decentralized Deep Learning}, - author={Lingjing Kong and Tao Lin and Anastasia Koloskova and Martin Jaggi and Sebastian U. Stich}, - year={2021}, - eprint={2102.04828}, - archivePrefix={arXiv}, - primaryClass={cs.LG} -} - @article{krizhevsky2009learning, title={{Learning Multiple Layers of Features from Tiny Images}}, author={Krizhevsky, Alex}, @@ -832,8 +821,7 @@ pages={211-252} title = {On the importance of initialization and momentum in deep learning}, author = {Ilya Sutskever and James Martens and George Dahl and Geoffrey Hinton}, booktitle = {ICML}, - year = {2013}, - publisher = {PMLR} + year = {2013} } @article{lecun1998gradient, diff --git a/main.tex b/main.tex index 682998c..c5c3f98 100644 --- a/main.tex +++ b/main.tex @@ -807,7 +807,7 @@ We have proposed D-Cliques, a sparse topology that recovers the convergence spee \begin{itemize} \item Clustering does not seem to make a difference in MNIST, even when using a higher-capacity model (LeNet) instead of a linear model. (Fig.\ref{fig:d-cliques-mnist-comparison-to-non-clustered-topologies}) - \item Except for the random 10 topology, convergence speed seems to be correlated with scattering in CIFAR-10 with LeNet model (Fig.\ref{fig:d-cliques-cifar10-linear-comparison-to-non-clustered-topologies}). There is also more difference between topologies both in convergence speed and scattering than for MNIST (Fig.~\ref{fig:d-cliques-mnist-comparison-to-non-clustered-topologies}). Scattering computed similar to Consensus Control for Decentralized Deep Learning~\cite{kong2021consensus}. + \item Except for the random 10 topology, convergence speed seems to be correlated with scattering in CIFAR-10 with LeNet model (Fig.\ref{fig:d-cliques-cifar10-linear-comparison-to-non-clustered-topologies}). There is also more difference between topologies both in convergence speed and scattering than for MNIST (Fig.~\ref{fig:d-cliques-mnist-comparison-to-non-clustered-topologies}). Scattering computed similar to Consensus Control for Decentralized Deep Learning~\cite{consensus_distance}. \end{itemize} \end{document} -- GitLab