Explore the essentials of federated learning and its privacy-preserving techniques for Edge AI. Assess your understanding of distributed model training, data privacy, security concerns, and the role of local devices in decentralized machine learning systems.
Which statement best describes federated learning in the context of Edge AI?
Explanation: Federated learning aggregates model updates from devices without gathering raw data, preserving user privacy. Sending raw data to a central server, as stated in another option, contradicts privacy goals. Devices do share some information, unlike the option suggesting no sharing. Training on one device at a time sequentially is inefficient and not how federated learning works.
In federated learning, what main privacy benefit is provided to users compared to traditional centralized training?
Explanation: Keeping user data on local devices reduces privacy risks by minimizing exposure. Encrypting and sending data to the cloud still involves transmission and possible exposure. Deleting user data before training would prevent any learning. Broadcasting user data conflicts with privacy preservation.
What is the primary function of an edge device in the federated learning process?
Explanation: Edge devices perform local model training using their data and share only improved model updates, preserving privacy. Running models independently without sharing updates impedes learning. Collecting data from other devices undermines decentralization. Broadcasting raw data defeats the privacy intent.
Which method is commonly used in federated learning to combine updates from multiple devices?
Explanation: Averaging model weights is a standard approach to aggregate distributed model updates. Random selection fails to use all learned knowledge. Using only the largest update ignores most devices' contributions. Summing raw data and re-training violates privacy.
How does federated learning address the challenge of different data distributions (heterogeneity) on edge devices?
Explanation: By collecting diverse updates, federated learning helps the global model adapt to different data conditions. Forcing identical datasets is impractical and not privacy-preserving. Ignoring data differences reduces effectiveness, while transferring data to the cloud undermines privacy.
In federated learning, what does a 'communication round' refer to?
Explanation: A communication round involves local training followed by the transfer of model updates, not raw data. Devices do not exchange all their data with each other, nor do they continuously stream data. Server password requests are unrelated to federated learning.
Which security threat may affect federated learning during the transmission of model updates?
Explanation: Eavesdropping intercepts model updates and poses a risk, even if data is not sent directly. Stealing raw data is less likely since it is not transmitted. Sharing updates on paper is unrealistic and irrelevant. Sending encrypted updates is actually a mitigation strategy, not a threat.
How can differential privacy be applied in federated learning?
Explanation: Adding noise helps protect individual data contributions, enhancing privacy. Broadcasting unmodified updates reduces privacy. Deleting all updates would impede learning, while sending complete datasets contradicts the privacy-preserving goals.
Which of the following is an example where federated learning is beneficial?
Explanation: Federated learning allows models like predictive text to learn from user data without transferring it off the device. Storing all data centrally undermines privacy. Training on just one device is less effective and not federated. Sharing passwords is unsafe and irrelevant.
What is an important requirement for edge devices participating in federated learning?
Explanation: Edge devices need enough computing power to update models with local data. Permanent connection is unnecessary as updates can be sent intermittently. Collecting data from others contradicts federated learning's decentralized design. Some storage is essential for model and temporary data.