Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

Efficient Federated Learning with Pre-Trained Large Language Model Using Several Adapter Mechanisms

Full metadata record
DC Field Value Language
dc.contributor.authorKim, Gyunyeop-
dc.contributor.authorYoo, Joon-
dc.contributor.authorKang, Sangwoo-
dc.date.accessioned2023-12-15T15:08:18Z-
dc.date.available2023-12-15T15:08:18Z-
dc.date.issued2023-11-
dc.identifier.issn2227-7390-
dc.identifier.issn2227-7390-
dc.identifier.urihttps://scholarworks.bwise.kr/gachon/handle/2020.sw.gachon/89516-
dc.description.abstractRecent advancements in deep learning have led to various challenges, one of which is the issue of data privacy in training data. To address this issue, federated learning, a technique that merges models trained by clients on servers, has emerged as an attractive solution. However, federated learning faces challenges related to data heterogeneity and system heterogeneity. Recent observations suggest that incorporating pre-trained models into federated learning can mitigate some of these challenges. Nonetheless, the main drawback of pre-trained models lies in their typically large model size, leading to excessive data transmission when clients send these models to the server. Additionally, federated learning involves multiple global steps, which means transmitting a large language model to multiple clients results in too much data exchange. In this paper, we propose a novel approach to address this challenge using adapters. Adapters demonstrate training efficiency by training a small capacity adapter layer alongside a large language model. This unique characteristic reduces the volume of data transmission, offering a practical solution to the problem. The evaluation results demonstrate that the proposed method achieves a reduction in training time of approximately 20-40% and a transmission speed improvement of over 98% compared to previous approaches.-
dc.language영어-
dc.language.isoENG-
dc.publisherMDPI-
dc.titleEfficient Federated Learning with Pre-Trained Large Language Model Using Several Adapter Mechanisms-
dc.typeArticle-
dc.identifier.wosid001100441000001-
dc.identifier.doi10.3390/math11214479-
dc.identifier.bibliographicCitationMATHEMATICS, v.11, no.21-
dc.description.isOpenAccessY-
dc.identifier.scopusid2-s2.0-85176608751-
dc.citation.titleMATHEMATICS-
dc.citation.volume11-
dc.citation.number21-
dc.type.docTypeArticle-
dc.publisher.location스위스-
dc.subject.keywordAuthorfederated learning-
dc.subject.keywordAuthordeep learning-
dc.subject.keywordAuthortransfer learning-
dc.subject.keywordAuthoradapter transformer-
dc.relation.journalResearchAreaMathematics-
dc.relation.journalWebOfScienceCategoryMathematics-
dc.description.journalRegisteredClassscie-
dc.description.journalRegisteredClassscopus-
Files in This Item
There are no files associated with this item.
Appears in
Collections
ETC > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher Yoo, Joon photo

Yoo, Joon
College of IT Convergence (Department of Software)
Read more

Altmetrics

Total Views & Downloads

BROWSE