Self-regulating Prompts: Foundational Model Adaptation without Forgetting

Prompt learning has emerged as an efficient alternative for fine-tuning foundational models, such as CLIP, for various downstream tasks. Conventionally trained using the task-specific objective, i.e., cross-entropy loss, prompts tend to overfit downstream data distributions and find it challenging t...

Full description

Saved in:
Bibliographic Details
Published in2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023) p. 15144
Main Authors Khattak, Muhammad Uzair, Wasim, Syed Talal, Naseer, Muzammal, Khan, Salman, Yang, Ming-Hsuan, Khan, Fahad
Format Conference Proceeding
LanguageEnglish
Published 2023
SeriesIEEE International Conference on Computer Vision
Online AccessGet full text
ISBN9798350307184
9798350307191
DOI10.1109/ICCV51070.2023.01394

Cover

Abstract Prompt learning has emerged as an efficient alternative for fine-tuning foundational models, such as CLIP, for various downstream tasks. Conventionally trained using the task-specific objective, i.e., cross-entropy loss, prompts tend to overfit downstream data distributions and find it challenging to capture task-agnostic general features from the frozen CLIP. This leads to the loss of the model's original generalization capability. To address this issue, our work introduces a self-regularization framework for prompting called PromptSRC (Prompting with Self-regulating Constraints). PromptSRC guides the prompts to optimize for both task-specific and task-agnostic general representations using a three-pronged approach by: (a) regulating prompted representations via mutual agreement maximization with the frozen model, (b) regulating with selfensemble of prompts over the training trajectory to encode their complementary strengths, and (c) regulating with textual diversity to mitigate sample diversity imbalance with the visual branch. To the best of our knowledge, this is the first regularization framework for prompt learning that avoids overfitting by jointly attending to pre-trained model features, the training trajectory during prompting, and the textual diversity. PromptSRC explicitly steers the prompts to learn a representation space that maximizes performance on downstream tasks without compromising CLIP generalization. We perform extensive experiments on 4 benchmarks where PromptSRC overall performs favorably well compared to the existing methods. Our code and pre-trained models are publicly available at: https://github.com/muzairkhattak/PromptSRC.
AbstractList Prompt learning has emerged as an efficient alternative for fine-tuning foundational models, such as CLIP, for various downstream tasks. Conventionally trained using the task-specific objective, i.e., cross-entropy loss, prompts tend to overfit downstream data distributions and find it challenging to capture task-agnostic general features from the frozen CLIP. This leads to the loss of the model's original generalization capability. To address this issue, our work introduces a self-regularization framework for prompting called PromptSRC (Prompting with Self-regulating Constraints). PromptSRC guides the prompts to optimize for both task-specific and task-agnostic general representations using a three-pronged approach by: (a) regulating prompted representations via mutual agreement maximization with the frozen model, (b) regulating with selfensemble of prompts over the training trajectory to encode their complementary strengths, and (c) regulating with textual diversity to mitigate sample diversity imbalance with the visual branch. To the best of our knowledge, this is the first regularization framework for prompt learning that avoids overfitting by jointly attending to pre-trained model features, the training trajectory during prompting, and the textual diversity. PromptSRC explicitly steers the prompts to learn a representation space that maximizes performance on downstream tasks without compromising CLIP generalization. We perform extensive experiments on 4 benchmarks where PromptSRC overall performs favorably well compared to the existing methods. Our code and pre-trained models are publicly available at: https://github.com/muzairkhattak/PromptSRC.
Author Naseer, Muzammal
Khan, Fahad
Yang, Ming-Hsuan
Khan, Salman
Wasim, Syed Talal
Khattak, Muhammad Uzair
Author_xml – sequence: 1
  givenname: Muhammad Uzair
  surname: Khattak
  fullname: Khattak, Muhammad Uzair
  organization: Mohamed Bin Zayed Univ AI, U Arab Emirates
– sequence: 2
  givenname: Syed Talal
  surname: Wasim
  fullname: Wasim, Syed Talal
  organization: Mohamed Bin Zayed Univ AI, U Arab Emirates
– sequence: 3
  givenname: Muzammal
  surname: Naseer
  fullname: Naseer, Muzammal
  organization: Mohamed Bin Zayed Univ AI, U Arab Emirates
– sequence: 4
  givenname: Salman
  surname: Khan
  fullname: Khan, Salman
  organization: Mohamed Bin Zayed Univ AI, U Arab Emirates; Australian Natl Univ, Australia
– sequence: 5
  givenname: Ming-Hsuan
  surname: Yang
  fullname: Yang, Ming-Hsuan
  organization: Univ Calif, CA USA; Google Res, CA USA
– sequence: 6
  givenname: Fahad
  surname: Khan
  fullname: Khan, Fahad
  organization: Mohamed Bin Zayed Univ AI, U Arab Emirates
BackLink https://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-202552$$DView record from Swedish Publication Index
BookMark eNotkN1KxDAUhAMqqGvfwIu-QOtJ0rSJd6W668KKgrq3Id2c1kC3KW3K4ttbf64GPmaGYa7Jee97JOSWQkopqLttVe0FhQJSBoynQLnKzkikCiW5AA4FldkliabJ1SB4Llkm5BXZvmHXJCO2c2eC69v4dfTHIUz38drPvV2Y700XP3uLXVxaM4RfFJ9c-PRzWFxji-EneUMuGtNNGP3rinysH9-rp2T3stlW5S4ZGKMhqSliLRpqFWuQMWwMgJEHwSkqtDmT2ByKupaWKjRyWa9UrixXILkVYARfkeSvdzrhMNd6GN3RjF_aG6cf3L7UyyLduVkvNwjB-DfZv1ay
ContentType Conference Proceeding
DBID ADTPV
BNKNJ
DG8
DOI 10.1109/ICCV51070.2023.01394
DatabaseName SwePub
SwePub Conference
SWEPUB Linköpings universitet
DatabaseTitleList
DeliveryMethod fulltext_linktorsrc
ExternalDocumentID oai_DiVA_org_liu_202552
GroupedDBID 6IE
6IG
6IL
6IN
ABLEC
ABQGA
ADTPV
ALMA_UNASSIGNED_HOLDINGS
BEFXN
BFFAM
BGNUA
BKEBE
BNKNJ
BPEOZ
CBEJK
DG8
IEGSK
IJVOP
OCL
RIB
RIC
RIE
RIL
RIO
ID FETCH-LOGICAL-p221t-b1eeb5f1d92fe22efa00a8c531e9ed628efc7bb8d19ea89799969d39083d50a53
ISBN 9798350307184
9798350307191
IngestDate Thu Aug 21 06:50:12 EDT 2025
IsPeerReviewed false
IsScholarly true
Language English
LinkModel OpenURL
MergedId FETCHMERGED-LOGICAL-p221t-b1eeb5f1d92fe22efa00a8c531e9ed628efc7bb8d19ea89799969d39083d50a53
ParticipantIDs swepub_primary_oai_DiVA_org_liu_202552
PublicationCentury 2000
PublicationDate 2023
PublicationDateYYYYMMDD 2023-01-01
PublicationDate_xml – year: 2023
  text: 2023
PublicationDecade 2020
PublicationSeriesTitle IEEE International Conference on Computer Vision
PublicationTitle 2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023)
PublicationYear 2023
SSID ssib053682458
Score 2.6231413
Snippet Prompt learning has emerged as an efficient alternative for fine-tuning foundational models, such as CLIP, for various downstream tasks. Conventionally trained...
SourceID swepub
SourceType Open Access Repository
StartPage 15144
Title Self-regulating Prompts: Foundational Model Adaptation without Forgetting
URI https://urn.kb.se/resolve?urn=urn:nbn:se:liu:diva-202552
hasFullText 1
inHoldings 1
isFullTextHit
isPrint
link http://utb.summon.serialssolutions.com/2.0.0/link/0/eLvHCXMwnV1Lb9pAEF7RnHpr1VR9pfKh6sUytdev3dwiGpREgqAGrNystXdXQTIEEftQflN_ZGfWBhtCpbYXCxuw1jufZr4Zz4OQL7GbgRXWoRMrJp1A576TuVI6UQAeM4_zjHpYOzwaR1ez4OY-vO_1fnWylqoy6-ebo3Ul_yNVuAZyxSrZf5Ds7qZwAT6DfOEIEobjAfk9amdwJrmN3ho2uU2G9n5728HteHj5w0SQbsdwNprM4Gs7uUb1aYjlYJDY1PRU4F3k3KlCO-t6Rj0GEiZr0BmlSZ1rhzBhSS8O0bEvpFg1GYsY08U056EJs5fbxaI6fxBlKYziHVUPYrEQ0p5txLzNDBZP9Vjnu5_AgKeiELvMjzHY2RpXo2qDfy06d62jt6JYNCDHbVdPxoU9CHa2pY34fmQ7zMJOTG19N_ZB_U7so3GBOTBI1FNePfOrUcNAY-quks8NhOmvihsMyih2-3jXPtLgoDWIe623v8-TixS2LS3mVUrRAwOD_yJmXl0ruNVdoR8xGoSm32S7KBbsn3Ovqd6EZXw7toiDhrWG5ExfkdN2j6zJDnCvSU8t35DrA1hYDSzOrS4oLAMKqwWF1YDCakFxSmbDy-ngymnmcDgrSr3SyTylslB7klOtKFVauK5gOWhvxZWMKFM6j7OMSY8rwfA9MY-49Dmwexm6IvTfkpPl41K9I1aONbt5DqwcdotFgglfg9OANDzWvp-_J1_r509XdbOV9A8y-PC3P_xIXrbI-UROynWlzoBHltlnI8Df99JvNw
linkProvider IEEE
openUrl ctx_ver=Z39.88-2004&ctx_enc=info%3Aofi%2Fenc%3AUTF-8&rfr_id=info%3Asid%2Fsummon.serialssolutions.com&rft_val_fmt=info%3Aofi%2Ffmt%3Akev%3Amtx%3Abook&rft.genre=proceeding&rft.title=2023+IEEE%2FCVF+INTERNATIONAL+CONFERENCE+ON+COMPUTER+VISION+%28ICCV+2023%29&rft.atitle=Self-regulating+Prompts%3A+Foundational+Model+Adaptation+without+Forgetting&rft.au=Khattak%2C+Muhammad+Uzair&rft.au=Wasim%2C+Syed+Talal&rft.au=Naseer%2C+Muzammal&rft.au=Khan%2C+Salman&rft.series=IEEE+International+Conference+on+Computer+Vision&rft.date=2023-01-01&rft.isbn=9798350307191&rft.spage=15144&rft_id=info:doi/10.1109%2FICCV51070.2023.01394&rft.externalDocID=oai_DiVA_org_liu_202552
thumbnail_l http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=9798350307184/lc.gif&client=summon&freeimage=true
thumbnail_m http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=9798350307184/mc.gif&client=summon&freeimage=true
thumbnail_s http://covers-cdn.summon.serialssolutions.com/index.aspx?isbn=9798350307184/sc.gif&client=summon&freeimage=true