Paper | PromptStyler: Prompt-driven Style Generation for Source-free Domain Generalization

1 introduction

  • iccv23
  • task: source-free domain generalization

common practice of DG:

  1. Utilize multiple source domains for learning domain-invariant features, but it is not clear which source domains are ideal for DG.
  2. Furthermore, collecting such large-scale multi-source domain data is costly.

another point:
a large-scale pre-trained model might have already seen a great variety of domains.

2 this paper

main idea: improve the model's generalization capability by simulating various distribution shifts in the latent space of such a large-scale model without using any source domain data?

Some insights:

  1. as shown in Fig. motivation (a), text features could effectively represent their relevant image features.
  2. the cross-model transferability phenomenon [67]: a classifier trained using text features can run inference using image features.

this paper:

  • propose a prompt-driven style generation method (PromptStyler)
  • synthesize styles via learnable word vectors in the form of “a S∗ style of a” (as in Fig. motivation (b))).

motivation:

imotivation

method

  • the style distance between "a S* style of a" are large;
  • the style-content distance between "a S* style of a [class]" are small.
image.png

Training detail:
use the learned style word vector to synthesize style-content features (synthesized features simulate images of known contents with diverse unknown styles in the joint space)

  • use the text encoder to generate such style-content features
  • use learnable features as input to train a linear classifier by classification loss;
  • inference: image --> image encoder --> linear classifier

model:

  • CLIP

exp:

版权声明:
作者:Mr李
链接:https://www.techfm.club/p/77930.html
来源:TechFM
文章版权归作者所有,未经允许请勿转载。

THE END
分享
二维码
< <上一篇
下一篇>>