File Download
Supplementary

postgraduate thesis: Efficient parameter update strategy for distributed deep learning system

TitleEfficient parameter update strategy for distributed deep learning system
Authors
Advisors
Advisor(s):Wang, CL
Issue Date2021
PublisherThe University of Hong Kong (Pokfulam, Hong Kong)
Citation
Zhang, Z. [張兆瑞]. (2021). Efficient parameter update strategy for distributed deep learning system. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR.
AbstractDeep neural networks (DNNs) have made a great breakthrough and demonstrated remarkable success for learning tasks in multiple domains including autonomous driving, natural language processing, computer vision, and speed recognition in recent years. With the scaling up of the DNN model with respect to parameter size and rapid increase of the training dataset, DNN training becomes a time-consuming task and usually takes several days even weeks. High-performance computing clusters that consist of multiple high-density multi-GPU servers connected by the high-speed network (HDGib) promote the blooming of distributed deep learning, which has been promising to shorten the training time. Parameter server architecture with data parallelism and asynchronous parameter update strategy has been identified as an efficient approach for scaling DNNs training on HDGib clusters. However, the performance improvement gained by the high-performance computing resources is often limited and far from the theoretical value. Communication and synchronization overhead has been widely considered as the bottleneck when deploying large-scale DNN training on HDGib clusters. They often occupy a significant portion of overall training time for each worker during the training procedure, which can be as high as 90%. To improve the training performance, emerging related works are proposed to reduce the communication and synchronization overhead including the local SGD, gradient compression, and communication_computation overlapping approaches. We identify five research issues based on the previous works: the communication period determination of the asynchronous local SGD, unbounded delay problem in asynchronous local SGD, the significant gradient identification, the gradient compensation for gradient sparsification, gradient quantization based on probability distribution. We provide five solutions to address the above research issues in this thesis. First, we modeled the asynchronous local-SGD as a Global Variable Consensus problem and defined a trigger according to the trigger definition in Multi-Agent-Systems (MASs) to determine when the worker updates the parameter with the server for the asynchronous local-SGD. Second, we calculate a weight for the gradient according to the most recent global loss value to solve the unbounded delay problem. Third, we identify the significant gradient according to the significance of their corresponding parameters defined by the model interpretability to prioritize higher updates for the significant parameters. Fourth, we propose using the exponential smoothing prediction approach to compensate the gradient for reducing the gradient error produced by the gradient sparsification. Fifth, we provide a layer-wise gradient quantization algorithm based on their probability distribution to minimize the quantization error and accelerate the DNN training. Furthermore, we indicate that when deploying the distributed DNN training on the HDGib clusters, besides the communication overhead, the gradient error caused by the random sampling of the training data for distributed SGD and magnified stale gradient problem also affects the convergence of the DNN training extremely. We further explore a momentum-driven adaptive synchronization algorithm to constrain the gradient error and address such problems. Extensive evaluation results generated based on the HDGib clusters and most popular DNN model and training dataset indicate that our proposed algorithms significantly improve the communication efficiency and reduce the gradient error for the distributed DNN training on the HDGib clusters, while also guaranteeing robust convergence of achieving high accuracy for various DNN models.
DegreeDoctor of Philosophy
SubjectMachine learning
Dept/ProgramComputer Science
Persistent Identifierhttp://hdl.handle.net/10722/311674

 

DC FieldValueLanguage
dc.contributor.advisorWang, CL-
dc.contributor.authorZhang, Zhaorui-
dc.contributor.author張兆瑞-
dc.date.accessioned2022-03-30T05:42:21Z-
dc.date.available2022-03-30T05:42:21Z-
dc.date.issued2021-
dc.identifier.citationZhang, Z. [張兆瑞]. (2021). Efficient parameter update strategy for distributed deep learning system. (Thesis). University of Hong Kong, Pokfulam, Hong Kong SAR.-
dc.identifier.urihttp://hdl.handle.net/10722/311674-
dc.description.abstractDeep neural networks (DNNs) have made a great breakthrough and demonstrated remarkable success for learning tasks in multiple domains including autonomous driving, natural language processing, computer vision, and speed recognition in recent years. With the scaling up of the DNN model with respect to parameter size and rapid increase of the training dataset, DNN training becomes a time-consuming task and usually takes several days even weeks. High-performance computing clusters that consist of multiple high-density multi-GPU servers connected by the high-speed network (HDGib) promote the blooming of distributed deep learning, which has been promising to shorten the training time. Parameter server architecture with data parallelism and asynchronous parameter update strategy has been identified as an efficient approach for scaling DNNs training on HDGib clusters. However, the performance improvement gained by the high-performance computing resources is often limited and far from the theoretical value. Communication and synchronization overhead has been widely considered as the bottleneck when deploying large-scale DNN training on HDGib clusters. They often occupy a significant portion of overall training time for each worker during the training procedure, which can be as high as 90%. To improve the training performance, emerging related works are proposed to reduce the communication and synchronization overhead including the local SGD, gradient compression, and communication_computation overlapping approaches. We identify five research issues based on the previous works: the communication period determination of the asynchronous local SGD, unbounded delay problem in asynchronous local SGD, the significant gradient identification, the gradient compensation for gradient sparsification, gradient quantization based on probability distribution. We provide five solutions to address the above research issues in this thesis. First, we modeled the asynchronous local-SGD as a Global Variable Consensus problem and defined a trigger according to the trigger definition in Multi-Agent-Systems (MASs) to determine when the worker updates the parameter with the server for the asynchronous local-SGD. Second, we calculate a weight for the gradient according to the most recent global loss value to solve the unbounded delay problem. Third, we identify the significant gradient according to the significance of their corresponding parameters defined by the model interpretability to prioritize higher updates for the significant parameters. Fourth, we propose using the exponential smoothing prediction approach to compensate the gradient for reducing the gradient error produced by the gradient sparsification. Fifth, we provide a layer-wise gradient quantization algorithm based on their probability distribution to minimize the quantization error and accelerate the DNN training. Furthermore, we indicate that when deploying the distributed DNN training on the HDGib clusters, besides the communication overhead, the gradient error caused by the random sampling of the training data for distributed SGD and magnified stale gradient problem also affects the convergence of the DNN training extremely. We further explore a momentum-driven adaptive synchronization algorithm to constrain the gradient error and address such problems. Extensive evaluation results generated based on the HDGib clusters and most popular DNN model and training dataset indicate that our proposed algorithms significantly improve the communication efficiency and reduce the gradient error for the distributed DNN training on the HDGib clusters, while also guaranteeing robust convergence of achieving high accuracy for various DNN models.-
dc.languageeng-
dc.publisherThe University of Hong Kong (Pokfulam, Hong Kong)-
dc.relation.ispartofHKU Theses Online (HKUTO)-
dc.rightsThe author retains all proprietary rights, (such as patent rights) and the right to use in future works.-
dc.rightsThis work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.-
dc.subject.lcshMachine learning-
dc.titleEfficient parameter update strategy for distributed deep learning system-
dc.typePG_Thesis-
dc.description.thesisnameDoctor of Philosophy-
dc.description.thesislevelDoctoral-
dc.description.thesisdisciplineComputer Science-
dc.description.naturepublished_or_final_version-
dc.date.hkucongregation2022-
dc.identifier.mmsid991044494006303414-

Export via OAI-PMH Interface in XML Formats


OR


Export to Other Non-XML Formats