Repository logo
 

LLM tuning: neural language persistence through adaptive mixture

dc.contributor.authorBanik, Mridul, author
dc.contributor.authorACM, publisher
dc.date.accessioned2025-12-22T19:09:11Z
dc.date.available2025-12-22T19:09:11Z
dc.date.issued2025-12-09
dc.description.abstractThis paper presents a novel architectural paradigm addressing knowledge degradation in large language models during continual fine-tuning. The framework leverages a Mixture-of-Experts-style approach, integrating multiple low-rank adapters governed by an intelligent routing mechanism. By freezing core model parameters and dynamically allocating task-specific expertise, this method preserves inherent world knowledge while enhancing performance across diverse downstream applications. The proposed Dynamic LoRA-Experts with Prototype-Ensemble Matching (DLEPM) framework demonstrates superior performance on sequential NLP benchmarks, achieving 89.2% average accuracy with only 5.4% forgetting—outperforming existing continual learning methods. Empirical evaluations validate the framework's efficacy in maintaining large language model fidelity during continuous adaptation.
dc.format.mediumborn digital
dc.format.mediumarticles
dc.identifier.bibliographicCitationMridul Banik. 2025. LLM Tuning: Neural Language Persistence through Adaptive Mixture. In 2025 International Conference on Artificial Intelligence and its Applications (ICARTI 2025), December 09-10, 2025, Port Louis, Mauritius. ACM, New York, NY, USA, 6 pages. https://doi.org/10.1145/3774791.3774803
dc.identifier.doihttps://doi.org/10.1145/3774791.3774803
dc.identifier.urihttps://hdl.handle.net/10217/242555
dc.languageEnglish
dc.language.isoeng
dc.publisherColorado State University. Libraries
dc.relation.ispartofPublications
dc.relation.ispartofACM DL Digital Library
dc.rights.licenseThis work is licensed under a Creative Commons Attribution 4.0 International License.
dc.rights.urihttps://creativecommons.org/licenses/by/4.0
dc.subjectcontinual learning
dc.subjectcatastrophic forgetting
dc.subjectparameter-efficient finetuning
dc.subjectlarge language models
dc.subjectlow-rank adaptation
dc.titleLLM tuning: neural language persistence through adaptive mixture
dc.typeText

Files

Original bundle

Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
FACF_ACMOA_3774791.3774803.pdf
Size:
574.62 KB
Format:
Adobe Portable Document Format

Collections