IDEAS home Printed from https://ideas.repec.org/a/gam/jmathe/v11y2023i15p3403-d1210489.html
   My bibliography  Save this article

A Derivative-Incorporated Adaptive Gradient Method for Federated Learning

Author

Listed:
  • Huimin Gao

    (College of Information Engineering, Henan University of Science and Technology, Luoyang 471023, China)

  • Qingtao Wu

    (College of Information Engineering, Henan University of Science and Technology, Luoyang 471023, China)

  • Hongyan Cao

    (China Research Institute of Radiowave Propagation, Qingdao 266107, China)

  • Xuhui Zhao

    (College of Information Engineering, Henan University of Science and Technology, Luoyang 471023, China)

  • Junlong Zhu

    (College of Information Engineering, Henan University of Science and Technology, Luoyang 471023, China)

  • Mingchuan Zhang

    (College of Information Engineering, Henan University of Science and Technology, Luoyang 471023, China)

Abstract

As a new machine learning technique, federated learning has received more attention in recent years, which enables decentralized model training across data silos or edge intelligent devices in the Internet of Things without exchanging local raw data. All kinds of algorithms are proposed to solve the challenges in federated learning. However, most of these methods are based on stochastic gradient descent, which undergoes slow convergence and unstable performance during the training stage. In this paper, we propose a differential adaptive federated optimization method, which incorporates an adaptive learning rate and the gradient difference into the iteration rule of the global model. We further adopt the first-order moment estimation to compute the approximate value of the differential term so as to avoid amplifying the random noise from the input data sample. The theoretical convergence guarantee is established for our proposed method in a stochastic non-convex setting under full client participation and partial client participation cases. Experiments for the image classification task are performed on two standard datasets by training a neural network model, and experiment results on different baselines demonstrate the effectiveness of our proposed method.

Suggested Citation

  • Huimin Gao & Qingtao Wu & Hongyan Cao & Xuhui Zhao & Junlong Zhu & Mingchuan Zhang, 2023. "A Derivative-Incorporated Adaptive Gradient Method for Federated Learning," Mathematics, MDPI, vol. 11(15), pages 1-21, August.
  • Handle: RePEc:gam:jmathe:v:11:y:2023:i:15:p:3403-:d:1210489
    as

    Download full text from publisher

    File URL: https://www.mdpi.com/2227-7390/11/15/3403/pdf
    Download Restriction: no

    File URL: https://www.mdpi.com/2227-7390/11/15/3403/
    Download Restriction: no
    ---><---

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jmathe:v:11:y:2023:i:15:p:3403-:d:1210489. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    We have no bibliographic references for this item. You can help adding them by using this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.