ISSN :2582-9793

Quantifying Cloud Cost Efficiency in Federated Learning: An Empirical Comparison with Centralized Training

Original Research (Published On: 23-Jan-2026 )
DOI : https://doi.org/10.54364/AAIML.2026.61270

Avijit Bose, Pradyut Sarkar, Sabyasachi Saha and Premanada Jana

Adv. Artif. Intell. Mach. Learn., 6 (1):4894-4908

1. Avijit Bose: Department of Computer Science & Engineering, MCKV Institute of Engineering, Liluah, Howrah-711204

2. Pradyut Sarkar: Department of Computer Science & Engineering, MAKAUT, Kalyani, West Bengal, India.

3. Sabyasachi Saha: Chief AI Scientist, Techno Exponent, Miami, Floria, USA.

4. Premanada Jana: Director, Netaji Subhash Open University, Kalyani, West Bengal, India -7112006

Download PDF Here

DOI: 10.54364/AAIML.2026.61270

Article History: Received on: 12-Oct-25, Accepted on: 16-Jan-26, Published on: 23-Jan-26

Corresponding Author: Avijit Bose

Email: avijit.bose@mckvie.edu.in

Citation: Avijit Bose, et al. Quantifying Cloud Cost Efficiency in Federated Learning: An Empirical Comparison with Centralized Training. Advances in Artificial Intelligence and Machine Learning. 2026;6(1):270. https://dx.doi.org/10.54364/AAIML.2026.61270


Abstract

    

Machine learning has become an important component of data driven applications. Most large-scale machine learning jobs are now executed on cloud platforms. These deployments often face high bandwidth and storage cost issues which can overside computation expenses. The present study investigates whether FL (Federated Learning) can overcome such costs without compromising predictive performance. A controlled experiment was conducted using two clients and ten communication rounds on the California Housing dataset to compare a centralized configuration with a federated one. In the FL set up, only model parameters were exchanged thus eliminating repeated data transfer. Realistic cloud pricing models for data transfer, compute and temporary storage were applied to recorded execution statistics. Results showed that centralized training consumed nearly 5,000 × more network bandwidth than FL while producing similar accuracy (MSE = 0.542 vs 0.556; R² = 0.587 vs 0.576). Under limited-bandwidth conditions and when local data are available, FL substantially lowers operating expenses without compromising on model quality. The study also highlights deployment variables such as round budgeting, update sizing, and client selection that influence overall efficiency. These findings indicate that federated learning which was popular for its privacy preservation is a practical, economically feasible solution for cloud-based machine learning systems.

Statistics

   Article View: 866
   PDF Downloaded: 5