Parameter-Efficient Multi-Task and Transfer Learning

The present disclosure provides systems and methods that enable parameter-efficient transfer learning, multi-task learning, and/or other forms of model re-purposing such as model personalization or domain adaptation. In particular, as one example, a computing system can obtain a machine-learned mode...

Full description

Saved in:
Bibliographic Details
Main Authors Mudrakarta, Pramod Kaushik, Zhmoginov, Andrey, Sandler, Mark, Howard, Andrew Gerald
Format Patent
LanguageEnglish
Published 02.04.2020
Subjects
Online AccessGet full text

Cover

More Information
Summary:The present disclosure provides systems and methods that enable parameter-efficient transfer learning, multi-task learning, and/or other forms of model re-purposing such as model personalization or domain adaptation. In particular, as one example, a computing system can obtain a machine-learned model that has been previously trained on a first training dataset to perform a first task. The machine-learned model can include a first set of learnable parameters. The computing system can modify the machine-learned model to include a model patch, where the model patch includes a second set of learnable parameters. The computing system can train the machine-learned model on a second training dataset to perform a second task that is different from the first task, which may include learning new values for the second set of learnable parameters included in the model patch while keeping at least some (e.g., all) of the first set of parameters fixed.
Bibliography:Application Number: US201916577698