What am I working on currently and other ideas
Currently, for my masters thesis, I have been working on continual learning using CLIP. In this regard, I have come across the XTAIL 1
Two successful methods in this benchmark are RAIL - which uses a frozen ZS CLIP model to predict, and if the prediction is within a seen class, the ViT output is additionally passed to a froze random projection matrix, followed by nearest class mean classification otherwise gives the ZS CLIP output. Thus, this has to store all the seen classes to be able to classify, and is almost like separate branches for seen and unseen classes - not really adapting CLIP.
Another method is LADA, which keeps ViT Frozen, PEFT trains the text encoder for a particular task, and stores the text vectors, and again during inference, uses ZS clip to predict if seen class, if seen it uses the text features which were stored of seen classes.
Again this is like a separate branch for classifying the same thing.
Ideally, in CLIP model any set of class names can be passed and the CLIP model is able to predict in that label space, here we are constraining CLIP to only be able to predict properly within the seen label space, otherwise a ZS CLIP output is given.
I want to suggest a way where the CLIP model can itself be adapted without needing to store classifiers - so we can pass any order of classes to text embedding and the model predicts it - and doesnt undergo catastrophic forgetting on seen classes.
-
Xu, Yicheng, Yuxin Chen, Jiahao Nie, Yusong Wang, Huiping Zhuang, and Manabu Okumura. “Advancing cross-domain discriminability in continual learning of vision-language models.” Advances in Neural Information Processing Systems 37 (2024): 51552-51576. ↩