Federated Learning (FL) is a Machine Learning (ML) setting in which several clients (e.g., mobile devices) train a model cooperatively under the direction of a central server (e.g., cloud server), while training data is decentralized. Due to the fact that FL clients frequently have restricted transmission capacity, communication among clients and servers needs to be reduced to enhance presentation. FL clients frequently employ Wi-Fi and must interact in Unstable Network Environments (UNE). Existing FL aggregation techniques send and receive a huge number of weights, which dramatically reduces the accuracy of the UNE. In this paper, we propose a Federated Sine Cosine Algorithm (FedSCA) to reduce data communication by transferring score principles rather than all client models' weights and utilizing the Sine Cosine Algorithm (SCA) mechanism as a weight updating technique to improve the clients' models. This paper reveals that using FedSCA significantly decreases the quantity of data utilized in network communication and increases the global model's accuracy by an average of 9.87% over FedAvg and 2.29% over Federated Particle Swarm Optimization (FedPSO). Moreover, in studies conducted on an unstable network, it demonstrated a 4.3% improvement in comparison to accuracy loss in existing algorithms.
- Convolutional Neural Network (CNN),
- Deep Learning,
- Federated Learning,
- Sine Cosine Algorithm (SCA)