Skip to content

Latest commit

 

History

History

FederatedLearning

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
 
 
 
 
 
 

Federated Learning

Promo banner for the two-part course series

Dear learner,

Addressing security and privacy in applications is vital. Applications built on LLMs pose special challenges, especially regarding private data.

Introducing Federated Learning, a two-part course series built in collaboration with Flower Labs and instructed by its Co-Founder & CEO, Daniel Beutel, and Co-founder and Chief Scientific Officer Nic Lane.

This series is designed to help you learn how to use Flower, a popular open source framework, to build a federated learning system, and implement federated fine-tuning of LLMs with private data. Federated learning allows models to be trained across multiple devices or organizations without sharing data, improving privacy and security.

In the first course, called Intro to Federated Learning, you'll learn about the federated training process, how to tune and customize it, how to increase data privacy, and how to manage bandwidth usage in federated learning.

In the second course, Federated Fine-tuning of LLMs with Private Data, you’ll learn to apply federated learning to LLMs. You’ll explore challenges like data memorization and the computational resources required by LLMs, and explore techniques for efficiency and privacy enhancement, such as Parameter-Efficient Fine-Tuning (PEFT) and Differential Privacy (DP).

This two-part course series is self-contained. If you already know what federated learning is, you can start directly with part two of the course.

Details

  • Explore the components of federated learning systems and learn to customize, tune, and orchestrate them for better model training.

  • Leverage federated learning to enhance LLMs by effectively managing key privacy and efficiency challenges.

  • Learn how techniques like parameter-efficient fine-tuning and differential privacy are crucial for making federated learning secure and efficient.

In detail, here’s what you’ll do in part one:

  • Learn how federated learning is used to train a variety of models, ranging from those for processing speech and vision all the way to the large language models, across distributed data while offering key data privacy options to users and organizations.
  • Learn how to train AI on distributed data by building, customizing, and tuning a federated learning project using Flower and PyTorch.
  • Gain intuition on how to think about Private Enhancing Technologies (PETs) in the context of federated learning, and work through an example using Differential Privacy, which protects individual data points from being traced back to their source.
  • Learn about two types of differential privacy - central and local - along with the dual approach of clipping and noising to protect private data.
  • Explore the bandwidth requirements for federated learning and how you can optimize it by reducing the update size and communication frequency.

Enroll in Part One

Lesson Video Code
Introduction video
Why Federated Learning video code
Federated Training Process video code
Tuning video code
Data Privacy video code
Bandwidth video code
Conclusion video

In the second part, you’ll learn how to train powerful models with your own data in a federated way, called federated LLM fine-tuning:

  • Understand the importance of safely training LLMs using private data.
  • Learn about the limitations of current training data and how Federated LLM Fine-tuning can help overcome these challenges.
  • Build an LLM that is fine-tuned with private medical data to answer complex questions, where you’ll see the benefits of federated methods when using private data.
  • Learn how federated LLM fine-tuning works and how it simplifies access to private data, reduces bandwidth with Parameter-Efficient Fine-Tuning (PEFT), and increases privacy to training data with Differential Privacy.
  • Understand how LLMs can leak training data, how federated LLMs can lower this risk.

Enroll in Part Two

Lesson Video Code
Introduction video
Smarter LLMs with Private Data video
Centralized LLM Fine-tuning video code
Federated Fine-tuning for LLMs video code
Keeping LLMs Private video code
Conclusion video