计算机科学
联合学习
采样(信号处理)
分布式学习
分布式计算
人工智能
计算机视觉
心理学
教育学
滤波器(信号处理)
作者
Sheng Yun,Md Zakirul Alam Bhuiyan,Md. Taufiq Al Hasib Sadi,Shen Su
标识
DOI:10.1109/ispa-bdcloud-socialcom-sustaincom59178.2023.00101
摘要
Recently, Large Language Models (LLMs) have been a phenomenal trend in the Artificial intelligence field. However, training and fine-tuning can be challenging because of privacy concerns and limited computing resources. Federated Learning (FL) has emerged as a novel machine learning framework offering privacy protection. The challenges in applying FL to real-world applications include dealing with heterogeneous data, poor client updates, and client selection. This paper introduces Privacy-preserving Federated Learning through Clustered Sampling on LLMs (FCLM), a framework that clusters models by their distribution similarity. It helps the model group similar models to improve text data heterogeneity handling and privacy concerns in distributed machine-learning environments. The FCLM framework is implemented and evaluated using popular Language models and text data. The framework shows a robust performance over the heterogeneous text data, which can further extend to the use of more complex LLMs.
科研通智能强力驱动
Strongly Powered by AbleSci AI