A distributed machine learning approach where models are trained across multiple devices or organizations without centralizing the raw data. Federated learning enables privacy-preserving multimodal AI training on sensitive data that cannot be shared.
In federated learning, a central server coordinates training across multiple participants (clients). Each client trains a local model on their private data and sends only the model updates (gradients or weights) to the server. The server aggregates updates from all clients into a global model and sends it back. Raw data never leaves the client, preserving privacy while enabling collaborative model improvement.
The FedAvg algorithm averages client model weights proportional to local dataset size. Communication rounds alternate between local training (multiple SGD steps) and global aggregation. Differential privacy can be added by clipping and noising gradients before sharing. Challenges include non-IID data distributions across clients, communication efficiency, and handling stragglers. Frameworks include TensorFlow Federated, PySyft, and Flower.