{"title":"The convergence of stochastic gradient descent in asynchronous shared memory","quality_controlled":"1","status":"public","oa_version":"Preprint","oa":1,"date_created":"2019-02-13T09:58:58Z","user_id":"c635000d-4b10-11ee-a964-aac5a93f6ac1","publisher":"ACM Press","scopus_import":"1","date_published":"2018-07-23T00:00:00Z","publication_status":"published","department":[{"_id":"DaAl"}],"main_file_link":[{"url":"https://arxiv.org/abs/1803.08841","open_access":"1"}],"abstract":[{"text":"Stochastic Gradient Descent (SGD) is a fundamental algorithm in machine learning, representing the optimization backbone for training several classic models, from regression to neural networks. Given the recent practical focus on distributed machine learning, significant work has been dedicated to the convergence properties of this algorithm under the inconsistent and noisy updates arising from execution in a distributed environment. However, surprisingly, the convergence properties of this classic algorithm in the standard shared-memory model are still not well-understood. In this work, we address this gap, and provide new convergence bounds for lock-free concurrent stochastic gradient descent, executing in the classic asynchronous shared memory model, against a strong adaptive adversary. Our results give improved upper and lower bounds on the \"price of asynchrony'' when executing the fundamental SGD algorithm in a concurrent setting. They show that this classic optimization tool can converge faster and with a wider range of parameters than previously known under asynchronous iterations. At the same time, we exhibit a fundamental trade-off between the maximum delay in the system and the rate at which SGD can converge, which governs the set of parameters under which this algorithm can still work efficiently.","lang":"eng"}],"type":"conference","day":"23","month":"07","article_processing_charge":"No","date_updated":"2023-09-19T10:42:53Z","isi":1,"year":"2018","citation":{"ama":"Alistarh D-A, De Sa C, Konstantinov NH. The convergence of stochastic gradient descent in asynchronous shared memory. In: Proceedings of the 2018 ACM Symposium on Principles of Distributed Computing  - PODC ’18. ACM Press; 2018:169-178. doi:10.1145/3212734.3212763","chicago":"Alistarh, Dan-Adrian, Christopher De Sa, and Nikola H Konstantinov. “The Convergence of Stochastic Gradient Descent in Asynchronous Shared Memory.” In Proceedings of the 2018 ACM Symposium on Principles of Distributed Computing  - PODC ’18, 169–78. ACM Press, 2018. https://doi.org/10.1145/3212734.3212763.","apa":"Alistarh, D.-A., De Sa, C., & Konstantinov, N. H. (2018). The convergence of stochastic gradient descent in asynchronous shared memory. In Proceedings of the 2018 ACM Symposium on Principles of Distributed Computing  - PODC ’18 (pp. 169–178). Egham, United Kingdom: ACM Press. https://doi.org/10.1145/3212734.3212763","ista":"Alistarh D-A, De Sa C, Konstantinov NH. 2018. The convergence of stochastic gradient descent in asynchronous shared memory. Proceedings of the 2018 ACM Symposium on Principles of Distributed Computing  - PODC ’18. PODC: Principles of Distributed Computing, 169–178.","short":"D.-A. Alistarh, C. De Sa, N.H. Konstantinov, in:, Proceedings of the 2018 ACM Symposium on Principles of Distributed Computing  - PODC ’18, ACM Press, 2018, pp. 169–178.","mla":"Alistarh, Dan-Adrian, et al. “The Convergence of Stochastic Gradient Descent in Asynchronous Shared Memory.” Proceedings of the 2018 ACM Symposium on Principles of Distributed Computing  - PODC ’18, ACM Press, 2018, pp. 169–78, doi:10.1145/3212734.3212763.","ieee":"D.-A. Alistarh, C. De Sa, and N. H. Konstantinov, “The convergence of stochastic gradient descent in asynchronous shared memory,” in Proceedings of the 2018 ACM Symposium on Principles of Distributed Computing  - PODC ’18, Egham, United Kingdom, 2018, pp. 169–178."},"conference":{"start_date":"2018-07-23","name":"PODC: Principles of Distributed Computing","location":"Egham, United Kingdom","end_date":"2018-07-27"},"_id":"5962","language":[{"iso":"eng"}],"author":[{"last_name":"Alistarh","first_name":"Dan-Adrian","id":"4A899BFC-F248-11E8-B48F-1D18A9856A87","full_name":"Alistarh, Dan-Adrian","orcid":"0000-0003-3650-940X"},{"full_name":"De Sa, Christopher","first_name":"Christopher","last_name":"De Sa"},{"first_name":"Nikola H","last_name":"Konstantinov","full_name":"Konstantinov, Nikola H","id":"4B9D76E4-F248-11E8-B48F-1D18A9856A87"}],"publication":"Proceedings of the 2018 ACM Symposium on Principles of Distributed Computing - PODC '18","doi":"10.1145/3212734.3212763","page":"169-178","external_id":{"isi":["000458186900022"],"arxiv":["1803.08841"]},"publication_identifier":{"isbn":["9781450357951"]}}