LLM tuning: neural language persistence through adaptive mixture
| dc.contributor.author | Banik, Mridul, author | |
| dc.contributor.author | ACM, publisher | |
| dc.date.accessioned | 2025-12-22T19:09:11Z | |
| dc.date.available | 2025-12-22T19:09:11Z | |
| dc.date.issued | 2025-12-09 | |
| dc.description.abstract | This paper presents a novel architectural paradigm addressing knowledge degradation in large language models during continual fine-tuning. The framework leverages a Mixture-of-Experts-style approach, integrating multiple low-rank adapters governed by an intelligent routing mechanism. By freezing core model parameters and dynamically allocating task-specific expertise, this method preserves inherent world knowledge while enhancing performance across diverse downstream applications. The proposed Dynamic LoRA-Experts with Prototype-Ensemble Matching (DLEPM) framework demonstrates superior performance on sequential NLP benchmarks, achieving 89.2% average accuracy with only 5.4% forgetting—outperforming existing continual learning methods. Empirical evaluations validate the framework's efficacy in maintaining large language model fidelity during continuous adaptation. | |
| dc.format.medium | born digital | |
| dc.format.medium | articles | |
| dc.identifier.bibliographicCitation | Mridul Banik. 2025. LLM Tuning: Neural Language Persistence through Adaptive Mixture. In 2025 International Conference on Artificial Intelligence and its Applications (ICARTI 2025), December 09-10, 2025, Port Louis, Mauritius. ACM, New York, NY, USA, 6 pages. https://doi.org/10.1145/3774791.3774803 | |
| dc.identifier.doi | https://doi.org/10.1145/3774791.3774803 | |
| dc.identifier.uri | https://hdl.handle.net/10217/242555 | |
| dc.language | English | |
| dc.language.iso | eng | |
| dc.publisher | Colorado State University. Libraries | |
| dc.relation.ispartof | Publications | |
| dc.relation.ispartof | ACM DL Digital Library | |
| dc.rights.license | This work is licensed under a Creative Commons Attribution 4.0 International License. | |
| dc.rights.uri | https://creativecommons.org/licenses/by/4.0 | |
| dc.subject | continual learning | |
| dc.subject | catastrophic forgetting | |
| dc.subject | parameter-efficient finetuning | |
| dc.subject | large language models | |
| dc.subject | low-rank adaptation | |
| dc.title | LLM tuning: neural language persistence through adaptive mixture | |
| dc.type | Text |
Files
Original bundle
1 - 1 of 1
Loading...
- Name:
- FACF_ACMOA_3774791.3774803.pdf
- Size:
- 574.62 KB
- Format:
- Adobe Portable Document Format
