A change in the machine learning landscape

By David Linthicum

Federated learning marks a milestone in enhancing collaborative model AI training. It is shifting the main approach to machine learning, moving away from the traditional centralized training methods towards more decentralized ones. Data is scattered, and we need to leverage it as training data where it exists.

This paradigm is nothing new. I was playing around with it in the 1990s. What’s old is new again… again. Federated learning allows for the collaborative training of machine learning models across multiple devices or servers, harnessing their collective data without needing to exchange or centralize it. Why should you care? Security and privacy, that’s why.

[ Also on InfoWorld: 14 popular AI algorithms and their uses ]

Here are the core principles of federated learning:

These factors are useful for AI, offering better security and privacy. Also, we’re not storing the same data in two different places, which is the common practice today in building new AI systems, such as generative AI.

The RoPPFL framework

Federated learning offers the promising prospect of collaborative model training across multiple devices or servers without needing to centralize the data. However, there are still security and privacy concerns, primarily the risk of local data set privacy leakage and the threat of AI model poisoning attacks by malicious clients.

What will save us? Naturally, when a new problem comes along, we must create unique solutions with cool names and acronyms. Let me introduce you to the Robust and Privacy-Preserving Federated Learning (RoPPFL) framework, a solution to address the inherent challenges associated with federated learning in edge computing environments.

The RoPPFL framework introduces a blend of local differential privacy (LDP) and similarity-based Robust Weighted Aggregation (RoWA) techniques. LDP protects data privacy by adding calibrated noise to the model updates. This makes it exceedingly difficult for attackers to infer individual data points, which is a common security attack against AI systems.

RoWA enhances the system’s resilience against poisoning attacks by aggregating model updates based on their similarity, mitigating the impact of any malicious interventions. RoPPFL uses a hierarchical federated learning structure. This structure organizes the model training process across different layers, including a cloud server, edge nodes, and client devices (e.g., smartphones).

Improved privacy and security

RoPPFL represents a step in the right direction for a cloud architect who needs to deal with this stuff all the time. Also, 80% of my work is generative AI these days, which is why I’m bringing it up, even though it’s borderline academic jargon.

This model addresses the simultaneous challenges of security and privacy, including the use of edge devices, such as smartphones and other personal devices, as sources of training data for data-hungry AI systems. The model can combine local differential privacy with a unique aggregation mechanism. The RoPPFL framework paves the way for the collaborative model training paradigm to exist and thrive without compromising on data protection and privacy, which is very much at risk with the use of AI.

The authors of the article that I referenced above are also the creators of the framework. So, make sure to read it if you’re interested in learning more about this topic.

I bring this up because we need to think about smarter ways of doing things if we’re going to design, build, and operate AI systems that eat our data for breakfast. We need to figure out how to build these AI systems (whether in the cloud or not) in ways that don’t do harm.

Given the current situation where enterprises are standing up generative AI systems first and asking the important questions later, we need more sound thinking around how we build, deploy, and secure these solutions so they become common practices. Right now, I bet many of you who are building AI systems that use distributed data have never heard of this framework. This is one of many current and future ideas that you need to understand.

© Info World