20.4.3.3.9 CLIP, Contrastive Language-Image Pre-Training

Chapter Contents (Back)
Visual-Language Model. Pre-training. CLIP.
See also Referring Expression Comprehension.
See also Vision-Language Models, Language-Vision Models, VQA.
See also Pre-Training.

Tung, F.[Frederick], Mori, G.[Greg],
Deep Neural Network Compression by In-Parallel Pruning-Quantization,
PAMI(42), No. 3, March 2020, pp. 568-579.
IEEE DOI 2002
BibRef
Earlier:
CLIP-Q: Deep Network Compression Learning by In-parallel Pruning-Quantization,
CVPR18(7873-7882)
IEEE DOI 1812
Quantization (signal), Image coding, Neural networks, Visualization, Training, Convolution, Network architecture, Bayesian optimization. Training, Task analysis, Optimization BibRef

Kwon, G.[Gihyun], Ye, J.C.[Jong Chul],
One-Shot Adaptation of GAN in Just One CLIP,
PAMI(45), No. 10, October 2023, pp. 12179-12191.
IEEE DOI 2310
BibRef

Han, B.H.[Bai-Hong], Jiang, X.Y.[Xiao-Yan], Fang, Z.J.[Zhi-Jun], Fujita, H.[Hamido], Gao, Y.B.[Yong-Bin],
F-SCP: An automatic prompt generation method for specific classes based on visual language pre-training models,
PR(147), 2024, pp. 110096.
Elsevier DOI 2312
Multi-modal, Vision language model, Prompt tuning, Large-scale pre-training model BibRef

Liu, B.[Bo], Lu, D.[Donghuan], Wei, D.[Dong], Wu, X.[Xian], Wang, Y.[Yan], Zhang, Y.[Yu], Zheng, Y.F.[Ye-Feng],
Improving Medical Vision-Language Contrastive Pretraining With Semantics-Aware Triage,
MedImg(42), No. 12, December 2023, pp. 3579-3589.
IEEE DOI 2312
BibRef

Gao, P.[Peng], Geng, S.J.[Shi-Jie], Zhang, R.R.[Ren-Rui], Ma, T.[Teli], Fang, R.Y.[Rong-Yao], Zhang, Y.F.[Yong-Feng], Li, H.S.[Hong-Sheng], Qiao, Y.[Yu],
CLIP-Adapter: Better Vision-Language Models with Feature Adapters,
IJCV(132), No. 2, February 2024, pp. 581-595.
Springer DOI 2402
BibRef

Liu, Y.H.[Yi-Hao], He, J.W.[Jing-Wen], Gu, J.J.[Jin-Jin], Kong, X.T.[Xiang-Tao], Qiao, Y.[Yu], Dong, C.[Chao],
DegAE: A New Pretraining Paradigm for Low-Level Vision,
CVPR23(23292-23303)
IEEE DOI 2309
BibRef

Dong, S.[Sijun], Wang, L.[Libo], Du, B.[Bo], Meng, X.L.[Xiao-Liang],
ChangeCLIP: Remote sensing change detection with multimodal vision-language representation learning,
PandRS(208), 2024, pp. 53-69.
Elsevier DOI Code:
WWW Link. 2402
Change detection, Vision-language representation learning, Multimodal artificial intelligence, Remote sensing BibRef

Peng, F.[Fang], Yang, X.S.[Xiao-Shan], Xiao, L.H.[Lin-Hui], Wang, Y.[Yaowei], Xu, C.S.[Chang-Sheng],
SgVA-CLIP: Semantic-Guided Visual Adapting of Vision-Language Models for Few-Shot Image Classification,
MultMed(26), 2024, pp. 3469-3480.
IEEE DOI 2402
Visualization, Adaptation models, Prototypes, Image classification, Learning systems, Task analysis, Feature extraction, Few-shot, vision-language models BibRef

Guo, Y.F.[Yu-Fei], Chen, Y.P.[Yuan-Pei], Ma, Z.[Zhe],
NeuroCLIP: Neuromorphic Data Understanding by CLIP and SNN,
SPLetters(31), 2024, pp. 246-250.
IEEE DOI 2402
Neuromorphics, Adaptation models, Visualization, Training, Neural networks, Membrane potentials, CLIP, neuromorphic data, spiking neural network BibRef

Xing, Y.H.[Ying-Hui], Wu, Q.[Qirui], Cheng, D.[De], Zhang, S.Z.[Shi-Zhou], Liang, G.Q.[Guo-Qiang], Wang, P.[Peng], Zhang, Y.N.[Yan-Ning],
Dual Modality Prompt Tuning for Vision-Language Pre-Trained Model,
MultMed(26), 2024, pp. 2056-2068.
IEEE DOI 2402
Visualization, Tuning, Task analysis, Adaptation models, Computational modeling, Feature extraction, Training, vision-language model BibRef

Xiao, L.H.[Lin-Hui], Yang, X.S.[Xiao-Shan], Peng, F.[Fang], Yan, M.[Ming], Wang, Y.[Yaowei], Xu, C.S.[Chang-Sheng],
CLIP-VG: Self-Paced Curriculum Adapting of CLIP for Visual Grounding,
MultMed(26), 2024, pp. 4334-4347.
IEEE DOI 2403
Grounding, Reliability, Adaptation models, Task analysis, Visualization, Data models, Annotations, Visual grounding, and vision-language models BibRef

Zhang, K.[Ke], Yang, Y.[Yan], Yu, J.[Jun], Jiang, H.[Hanliang], Fan, J.P.[Jian-Ping], Huang, Q.M.[Qing-Ming], Han, W.D.[Wei-Dong],
Multi-Task Paired Masking With Alignment Modeling for Medical Vision-Language Pre-Training,
MultMed(26), 2024, pp. 4706-4721.
IEEE DOI 2403
Task analysis, Image reconstruction, Visualization, Medical diagnostic imaging, Transformers, Semantics, Multitasking, cross-modal alignment BibRef

Zara, G.[Giacomo], Turrisi-da Costa, V.G.[Victor Guilherme], Roy, S.[Subhankar], Rota, P.[Paolo], Ricci, E.[Elisa],
Simplifying open-set video domain adaptation with contrastive learning,
CVIU(241), 2024, pp. 103953.
Elsevier DOI 2403
BibRef
Earlier: A1, A3, A4, A5, Only:
AutoLabel: CLIP-based framework for Open-Set Video Domain Adaptation,
CVPR23(11504-11513)
IEEE DOI 2309
Open-set video domain adaptation, Video Action Recognition, Contrastive learning BibRef

Wang, X.H.[Xiao-Han], Wang, W.G.[Wen-Guan], Shao, J.Y.[Jia-Yi], Yang, Y.[Yi],
Learning to Follow and Generate Instructions for Language-Capable Navigation,
PAMI(46), No. 5, May 2024, pp. 3334-3350.
IEEE DOI 2404
Navigation, Task analysis, Robot kinematics, Natural languages, Behavioral sciences, Training, Visualization, Multi-task learning, vision-language pre-training BibRef


Gupta, D.[Devaansh], Kharbanda, S.[Siddhant], Zhou, J.W.[Jia-Wei], Li, W.[Wanhua], Pfister, H.[Hanspeter], Wei, D.L.[Dong-Lai],
CLIPTrans: Transferring Visual Knowledge with Pre-trained Models for Multimodal Machine Translation,
ICCV23(2863-2874)
IEEE DOI 2401
BibRef

Fang, H.[Han], Yang, Z.F.[Zhi-Fei], Wei, Y.H.[Yu-Han], Zang, X.H.[Xiang-Hao], Ban, C.[Chao], Feng, Z.[Zerun], He, Z.J.[Zhong-Jiang], Li, Y.X.[Yong-Xiang], Sun, H.[Hao],
Alignment and Generation Adapter for Efficient Video-Text Understanding,
CLVL23(2783-2789)
IEEE DOI 2401
BibRef

Zhu, B.[Beier], Niu, Y.[Yulei], Han, Y.C.[Yu-Cheng], Wu, Y.[Yue], Zhang, H.W.[Han-Wang],
Prompt-aligned Gradient for Prompt Tuning,
ICCV23(15613-15623)
IEEE DOI 2401
BibRef

Yuan, H.J.[Hang-Jie], Zhang, S.W.[Shi-Wei], Wang, X.[Xiang], Albanie, S.[Samuel], Pan, Y.[Yining], Feng, T.[Tao], Jiang, J.W.[Jian-Wen], Ni, D.[Dong], Zhang, Y.[Yingya], Zhao, D.L.[De-Li],
RLIPv2: Fast Scaling of Relational Language-Image Pre-training,
ICCV23(21592-21604)
IEEE DOI Code:
WWW Link. 2401
BibRef

Wang, Z.[Ziyi], Yu, X.[Xumin], Rao, Y.M.[Yong-Ming], Zhou, J.[Jie], Lu, J.W.[Ji-Wen],
Take-A-Photo: 3D-to-2D Generative Pre-training of Point Cloud Models,
ICCV23(5617-5627)
IEEE DOI Code:
WWW Link. 2401
BibRef

Li, M.[Ming], Wu, J.[Jie], Wang, X.[Xionghui], Chen, C.[Chen], Qin, J.[Jie], Xiao, X.F.[Xue-Feng], Wang, R.[Rui], Zheng, M.[Min], Pan, X.[Xin],
AlignDet: Aligning Pre-training and Fine-tuning in Object Detection,
ICCV23(6843-6853)
IEEE DOI 2401
BibRef

Lee, S.[Suhyeon], Chung, H.J.[Hyung-Jin], Park, M.Y.[Min-Young], Park, J.[Jonghyuk], Ryu, W.S.[Wi-Sun], Ye, J.C.[Jong Chul],
Improving 3D Imaging with Pre-Trained Perpendicular 2D Diffusion Models,
ICCV23(10676-10686)
IEEE DOI Code:
WWW Link. 2401
BibRef

Ye, Q.H.[Qing-Hao], Xu, G.H.[Guo-Hai], Yan, M.[Ming], Xu, H.Y.[Hai-Yang], Qian, Q.[Qi], Zhang, J.[Ji], Huang, F.[Fei],
HiTeA: Hierarchical Temporal-Aware Video-Language Pre-training,
ICCV23(15359-15370)
IEEE DOI 2401
BibRef

Wu, C.Y.[Chao-Yi], Zhang, X.M.[Xiao-Man], Zhang, Y.[Ya], Wang, Y.F.[Yan-Feng], Xie, W.[Weidi],
MedKLIP: Medical Knowledge Enhanced Language-Image Pre-Training for X-ray Diagnosis,
ICCV23(21315-21326)
IEEE DOI 2401
BibRef

Yang, Q.S.[Qiu-Shi], Li, W.Y.[Wu-Yang], Li, B.[Baopu], Yuan, Y.X.[Yi-Xuan],
MRM: Masked Relation Modeling for Medical Image Pre-Training with Genetics,
ICCV23(21395-21405)
IEEE DOI Code:
WWW Link. 2401
BibRef

Ma, W.X.[Wen-Xuan], Li, S.[Shuang], Zhang, J.[JinMing], Liu, C.H.[Chi Harold], Kang, J.X.[Jing-Xuan], Wang, Y.L.[Yu-Lin], Huang, G.[Gao],
Borrowing Knowledge From Pre-trained Language Model: A New Data-efficient Visual Learning Paradigm,
ICCV23(18740-18751)
IEEE DOI Code:
WWW Link. 2401
BibRef

Ganugula, P.[Prajwal], Kumar, Y.S.S.S.S.[Y.S.S.S. Santosh], Reddy, N.K.S.[N.K. Sagar], Chellingi, P.[Prabhath], Thakur, A.[Avinash], Kasera, N.[Neeraj], Anand, C.S.[C. Shyam],
MOSAIC: Multi-Object Segmented Arbitrary Stylization Using CLIP,
NIVT23(892-903)
IEEE DOI 2401
BibRef

Luo, Z.Y.[Zi-Yang], Zhao, P.[Pu], Xu, C.[Can], Geng, X.[Xiubo], Shen, T.[Tao], Tao, C.Y.[Chong-Yang], Ma, J.[Jing], Lin, Q.W.[Qing-Wei], Jiang, D.X.[Da-Xin],
LexLIP: Lexicon-Bottlenecked Language-Image Pre-Training for Large-Scale Image-Text Sparse Retrieval,
ICCV23(11172-11183)
IEEE DOI 2401
BibRef

Zhai, X.H.[Xiao-Hua], Mustafa, B.[Basil], Kolesnikov, A.[Alexander], Beyer, L.[Lucas],
Sigmoid Loss for Language Image Pre-Training,
ICCV23(11941-11952)
IEEE DOI 2401
BibRef

Shtedritski, A.[Aleksandar], Rupprecht, C.[Christian], Vedaldi, A.[Andrea],
What does CLIP know about a red circle? Visual prompt engineering for VLMs,
ICCV23(11953-11963)
IEEE DOI 2401
BibRef

Zhu, Z.[Ziyu], Ma, X.J.[Xiao-Jian], Chen, Y.X.[Yi-Xin], Deng, Z.D.[Zhi-Dong], Huang, S.Y.[Si-Yuan], Li, Q.[Qing],
3D-VisTA: Pre-trained Transformer for 3D Vision and Text Alignment,
ICCV23(2899-2909)
IEEE DOI 2401
BibRef

Yang, K.C.[Kai-Cheng], Deng, J.K.[Jian-Kang], An, X.[Xiang], Li, J.W.[Jia-Wei], Feng, Z.[Ziyong], Guo, J.[Jia], Yang, J.[Jing], Liu, T.L.[Tong-Liang],
ALIP: Adaptive Language-Image Pre-training with Synthetic Caption,
ICCV23(2910-2919)
IEEE DOI Code:
WWW Link. 2401
BibRef

Yang, Y.F.[Yi-Fan], Huang, W.Q.[Wei-Quan], Wei, Y.X.[Yi-Xuan], Peng, H.[Houwen], Jiang, X.Y.[Xin-Yang], Jiang, H.Q.[Hui-Qiang], Wei, F.[Fangyun], Wang, Y.[Yin], Hu, H.[Han], Qiu, L.[Lili], Yang, Y.Q.[Yu-Qing],
Attentive Mask CLIP,
ICCV23(2759-2769)
IEEE DOI Code:
WWW Link. 2401
BibRef

Vinker, Y.[Yael], Alaluf, Y.[Yuval], Cohen-Or, D.[Daniel], Shamir, A.[Ariel],
CLIPascene: Scene Sketching with Different Types and Levels of Abstraction,
ICCV23(4123-4133)
IEEE DOI Code:
WWW Link. 2401
BibRef

Wei, Y.X.[Yi-Xuan], Hu, H.[Han], Xie, Z.[Zhenda], Liu, Z.[Ze], Zhang, Z.[Zheng], Cao, Y.[Yue], Bao, J.M.[Jian-Min], Chen, D.[Dong], Guo, B.[Baining],
Improving CLIP Fine-tuning Performance,
ICCV23(5416-5426)
IEEE DOI Code:
WWW Link. 2401
BibRef

Maniparambil, M.[Mayug], Vorster, C.[Chris], Molloy, D.[Derek], Murphy, N.[Noel], McGuinness, K.[Kevin], O'Connor, N.E.[Noel E.],
Enhancing CLIP with GPT-4: Harnessing Visual Descriptions as Prompts,
MMFM23(262-271)
IEEE DOI 2401
BibRef

Huang, T.Y.[Tian-Yu], Dong, B.[Bowen], Yang, Y.H.[Yun-Han], Huang, X.S.[Xiao-Shui], Lau, R.W.H.[Rynson W.H.], Ouyang, W.L.[Wan-Li], Zuo, W.M.[Wang-Meng],
CLIP2Point: Transfer CLIP to Point Cloud Classification with Image-Depth Pre-Training,
ICCV23(22100-22110)
IEEE DOI Code:
WWW Link. 2401
BibRef

Wu, K.[Kan], Peng, H.[Houwen], Zhou, Z.H.[Zheng-Hong], Xiao, B.[Bin], Liu, M.[Mengchen], Yuan, L.[Lu], Xuan, H.[Hong], Valenzuela, M.[Michael], Chen, X.S.[Xi Stephen], Wang, X.G.[Xing-Gang], Chao, H.Y.[Hong-Yang], Hu, H.[Han],
TinyCLIP: CLIP Distillation via Affinity Mimicking and Weight Inheritance,
ICCV23(21913-21923)
IEEE DOI 2401
BibRef

Deng, X.[Xinchi], Shi, H.[Han], Huang, R.[Runhui], Li, C.L.[Chang-Lin], Xu, H.[Hang], Han, J.H.[Jian-Hua], Kwok, J.[James], Zhao, S.[Shen], Zhang, W.[Wei], Liang, X.D.[Xiao-Dan],
GrowCLIP: Data-aware Automatic Model Growing for Large-scale Contrastive Language-Image Pre-training,
ICCV23(22121-22132)
IEEE DOI 2401
BibRef

Ranasinghe, K.[Kanchana], McKinzie, B.[Brandon], Ravi, S.[Sachin], Yang, Y.F.[Yin-Fei], Toshev, A.[Alexander], Shlens, J.[Jonathon],
Perceptual Grouping in Contrastive Vision-Language Models,
ICCV23(5548-5561)
IEEE DOI 2401
BibRef

Shao, B.[Bin], Liu, J.Z.[Jian-Zhuang], Pei, R.[Renjing], Xu, S.[Songcen], Dai, P.[Peng], Lu, J.W.[Ju-Wei], Li, W.[Weimian], Yan, Y.[Youliang],
HiVLP: Hierarchical Interactive Video-Language Pre-Training,
ICCV23(13710-13720)
IEEE DOI 2401
BibRef

Ali, M.[Muhammad], Khan, S.[Salman],
CLIP-Decoder: ZeroShot Multilabel Classification using Multimodal CLIP Aligned Representations,
VLAR23(4677-4681)
IEEE DOI 2401
BibRef

Singha, M.[Mainak], Pal, H.[Harsh], Jha, A.[Ankit], Banerjee, B.[Biplab],
AD-CLIP: Adapting Domains in Prompt Space Using CLIP,
OutDistri23(4357-4366)
IEEE DOI 2401
BibRef

Zhang, J.[Junbo], Dong, R.[Runpei], Ma, K.[Kaisheng],
CLIP-FO3D: Learning Free Open-world 3D Scene Representations from 2D Dense CLIP,
OpenSUN3D(2040-2051)
IEEE DOI 2401
BibRef

Auty, D.[Dylan], Mikolajczyk, K.[Krystian],
Learning to Prompt CLIP for Monocular Depth Estimation: Exploring the Limits of Human Language,
OpenSUN3D(2031-2049)
IEEE DOI Code:
WWW Link. 2401
BibRef

Hegde, D.[Deepti], Valanarasu, J.M.J.[Jeya Maria Jose], Patel, V.M.[Vishal M.],
CLIP goes 3D: Leveraging Prompt Tuning for Language Grounded 3D Recognition,
OpenSUN3D(2020-2030)
IEEE DOI Code:
WWW Link. 2401
BibRef

Xu, X.[Xin], Xiong, T.Y.[Tian-Yi], Ding, Z.[Zheng], Tu, Z.W.[Zhuo-Wen],
MasQCLIP for Open-Vocabulary Universal Image Segmentation,
ICCV23(887-898)
IEEE DOI Code:
WWW Link. 2401
BibRef

Wang, H.L.[Hua-Liang], Li, Y.[Yi], Yao, H.[Huifeng], Li, X.M.[Xiao-Meng],
CLIPN for Zero-Shot OOD Detection: Teaching CLIP to Say No,
ICCV23(1802-1812)
IEEE DOI Code:
WWW Link. 2401
BibRef

Zhu, X.Y.[Xiang-Yang], Zhang, R.[Renrui], He, B.[Bowei], Zhou, A.[Aojun], Wang, D.[Dong], Zhao, B.[Bin], Gao, P.[Peng],
Not All Features Matter: Enhancing Few-shot CLIP with Adaptive Prior Refinement,
ICCV23(2605-2615)
IEEE DOI Code:
WWW Link. 2401
BibRef

Paiss, R.[Roni], Ephrat, A.[Ariel], Tov, O.[Omer], Zada, S.[Shiran], Mosseri, I.[Inbar], Irani, M.[Michal], Dekel, T.[Tali],
Teaching CLIP to Count to Ten,
ICCV23(3147-3157)
IEEE DOI 2401
BibRef

Zhu, X.Y.[Xiang-Yang], Zhang, R.R.[Ren-Rui], He, B.[Bowei], Guo, Z.[Ziyu], Zeng, Z.[Ziyao], Qin, Z.[Zipeng], Zhang, S.H.[Shang-Hang], Gao, P.[Peng],
PointCLIP V2: Prompting CLIP and GPT for Powerful 3D Open-world Learning,
ICCV23(2639-2650)
IEEE DOI Code:
WWW Link. 2401
BibRef

Yuan, M.[Min], Lv, N.N.[Ning-Ning], Xie, Y.F.[Yu-Fei], Lu, F.X.[Fu-Xiang], Zhan, K.[Kun],
CLIP-FG: Selecting Discriminative Image Patches by Contrastive Language-Image Pre-Training for Fine-Grained Image Classification,
ICIP23(560-564)
IEEE DOI 2312
BibRef

Zeng, Z.[Ziyun], Ge, Y.Y.[Yu-Ying], Liu, X.H.[Xi-Hui], Chen, B.[Bin], Luo, P.[Ping], Xia, S.T.[Shu-Tao], Ge, Y.X.[Yi-Xiao],
Learning Transferable Spatiotemporal Representations from Natural Script Knowledge,
CVPR23(23079-23089)
IEEE DOI 2309

WWW Link. BibRef

Wang, J.P.[Jin-Peng], Ge, Y.X.[Yi-Xiao], Yan, R.[Rui], Ge, Y.Y.[Yu-Ying], Lin, K.Q.[Kevin Qinghong], Tsutsui, S.[Satoshi], Lin, X.D.[Xu-Dong], Cai, G.[Guanyu], Wu, J.P.[Jian-Ping], Shan, Y.[Ying], Qie, X.[Xiaohu], Shou, M.Z.[Mike Zheng],
All in One: Exploring Unified Video-Language Pre-Training,
CVPR23(6598-6608)
IEEE DOI 2309
BibRef

Ramrakhya, R.[Ram], Batra, D.[Dhruv], Wijmans, E.[Erik], Das, A.[Abhishek],
PIRLNav: Pretraining with Imitation and RL Finetuning for OBJECTNAV,
CVPR23(17896-17906)
IEEE DOI 2309
BibRef

Lin, X.D.[Xu-Dong], Tiwari, S.[Simran], Huang, S.Y.[Shi-Yuan], Li, M.[Manling], Shou, M.Z.[Mike Zheng], Ji, H.[Heng], Chang, S.F.[Shih-Fu],
Towards Fast Adaptation of Pretrained Contrastive Models for Multi-channel Video-Language Retrieval,
CVPR23(14846-14855)
IEEE DOI 2309
BibRef

Wang, H.C.[Hao-Chen], Du, X.D.[Xiao-Dan], Li, J.H.[Jia-Hao], Yeh, R.A.[Raymond A.], Shakhnarovich, G.[Greg],
Score Jacobian Chaining: Lifting Pretrained 2D Diffusion Models for 3D Generation,
CVPR23(12619-12629)
IEEE DOI 2309
BibRef

Jin, L.[Lei], Luo, G.[Gen], Zhou, Y.[Yiyi], Sun, X.S.[Xiao-Shuai], Jiang, G.[Guannan], Shu, A.[Annan], Ji, R.R.[Rong-Rong],
RefCLIP: A Universal Teacher for Weakly Supervised Referring Expression Comprehension,
CVPR23(01-10)
IEEE DOI 2309
BibRef

Saito, K.[Kuniaki], Sohn, K.[Kihyuk], Zhang, X.[Xiang], Li, C.L.[Chun-Liang], Lee, C.Y.[Chen-Yu], Saenko, K.[Kate], Pfister, T.[Tomas],
Prefix Conditioning Unifies Language and Label Supervision,
CVPR23(2861-2870)
IEEE DOI 2309
BibRef

Park, J.[Jaeyoo], Han, B.H.[Bo-Hyung],
Multi-Modal Representation Learning with Text-Driven Soft Masks,
CVPR23(2798-2807)
IEEE DOI 2309
BibRef

Jin, Z.[Zhao], Hayat, M.[Munawar], Yang, Y.W.[Yu-Wei], Guo, Y.L.[Yu-Lan], Lei, Y.J.[Yig-Jie],
Context-aware Alignment and Mutual Masking for 3D-Language Pre-training,
CVPR23(10984-10994)
IEEE DOI 2309
BibRef

Guo, Z.X.[Zi-Xian], Dong, B.[Bowen], Ji, Z.L.[Zhi-Long], Bai, J.F.[Jin-Feng], Guo, Y.[Yiwen], Zuo, W.M.[Wang-Meng],
Texts as Images in Prompt Tuning for Multi-Label Image Recognition,
CVPR23(2808-2817)
IEEE DOI 2309

WWW Link. BibRef

Cherti, M.[Mehdi], Beaumont, R.[Romain], Wightman, R.[Ross], Wortsman, M.[Mitchell], Ilharco, G.[Gabriel], Gordon, C.[Cade], Schuhmann, C.[Christoph], Schmidt, L.[Ludwig], Jitsev, J.[Jenia],
Reproducible Scaling Laws for Contrastive Language-Image Learning,
CVPR23(2818-2829)
IEEE DOI 2309
BibRef

Lei, J.[Jie], Li, L.J.[Lin-Jie], Zhou, L.[Luowei], Gan, Z.[Zhe], Berg, T.L.[Tamara L.], Bansal, M.[Mohit], Liu, J.J.[Jing-Jing],
Less is More: CLIPBERT for Video-and-Language Learning via Sparse Sampling,
CVPR21(7327-7337)
IEEE DOI 2111
Training, Computational modeling, Feature extraction, Knowledge discovery, Distance measurement BibRef

Zhou, J.H.[Jing-Hao], Dong, L.[Li], Gan, Z.[Zhe], Wang, L.J.[Li-Juan], Wei, F.[Furu],
Non-Contrastive Learning Meets Language-Image Pre-Training,
CVPR23(11028-11038)
IEEE DOI 2309
BibRef

Hu, Z.[Ziniu], Iscen, A.[Ahmet], Sun, C.[Chen], Wang, Z.[Zirui], Chang, K.W.[Kai-Wei], Sun, Y.Z.[Yi-Zhou], Schmid, C.[Cordelia], Ross, D.A.[David A.], Fathi, A.[Alireza],
Reveal: Retrieval-Augmented Visual-Language Pre-Training with Multi-Source Multimodal Knowledge Memory,
CVPR23(23369-23379)
IEEE DOI 2309
BibRef

Li, Y.H.[Yang-Hao], Fan, H.Q.[Hao-Qi], Hu, R.[Ronghang], Feichtenhofer, C.[Christoph], He, K.[Kaiming],
Scaling Language-Image Pre-Training via Masking,
CVPR23(23390-23400)
IEEE DOI 2309
BibRef

Jin, P.[Peng], Huang, J.[Jinfa], Xiong, P.F.[Peng-Fei], Tian, S.X.[Shang-Xuan], Liu, C.[Chang], Ji, X.Y.[Xiang-Yang], Yuan, L.[Li], Chen, J.[Jie],
Video-Text as Game Players: Hierarchical Banzhaf Interaction for Cross-Modal Representation Learning,
CVPR23(2472-2482)
IEEE DOI 2309

WWW Link. BibRef

Ye, S.Q.[Shu-Quan], Xie, Y.[Yujia], Chen, D.D.[Dong-Dong], Xu, Y.[Yichong], Yuan, L.[Lu], Zhu, C.G.[Chen-Guang], Liao, J.[Jing],
Improving Commonsense in Vision-Language Models via Knowledge Graph Riddles,
CVPR23(2634-2645)
IEEE DOI 2309
BibRef

Li, H.[Hao], Zhu, J.[Jinguo], Jiang, X.[Xiaohu], Zhu, X.[Xizhou], Li, H.S.[Hong-Sheng], Yuan, C.[Chun], Wang, X.H.[Xiao-Hua], Qiao, Y.[Yu], Wang, X.G.[Xiao-Gang], Wang, W.H.[Wen-Hai], Dai, J.F.[Ji-Feng],
Uni-Perceiver v2: A Generalist Model for Large-Scale Vision and Vision-Language Tasks,
CVPR23(2691-2700)
IEEE DOI 2309
BibRef

Wu, W.H.[Wen-Hao], Wang, X.H.[Xiao-Han], Luo, H.P.[Hai-Peng], Wang, J.D.[Jing-Dong], Yang, Y.[Yi], Ouyang, W.L.[Wan-Li],
Bidirectional Cross-Modal Knowledge Exploration for Video Recognition with Pre-trained Vision-Language Models,
CVPR23(6620-6630)
IEEE DOI 2309
BibRef

Seth, A.[Ashish], Hemani, M.[Mayur], Agarwal, C.[Chirag],
DeAR: Debiasing Vision-Language Models with Additive Residuals,
CVPR23(6820-6829)
IEEE DOI 2309
BibRef

Radenovic, F.[Filip], Dubey, A.[Abhimanyu], Kadian, A.[Abhishek], Mihaylov, T.[Todor], Vandenhende, S.[Simon], Patel, Y.[Yash], Wen, Y.[Yi], Ramanathan, V.[Vignesh], Mahajan, D.[Dhruv],
Filtering, Distillation, and Hard Negatives for Vision-Language Pre-Training,
CVPR23(6967-6977)
IEEE DOI 2309
BibRef

Yu, T.[Tao], Lu, Z.[Zhihe], Jin, X.[Xin], Chen, Z.B.[Zhi-Bo], Wang, X.C.[Xin-Chao],
Task Residual for Tuning Vision-Language Models,
CVPR23(10899-10909)
IEEE DOI 2309
BibRef

Ma, Z.X.[Zi-Xian], Hong, J.[Jerry], Gul, M.O.[Mustafa Omer], Gandhi, M.[Mona], Gao, I.[Irena], Krishna, R.[Ranjay],
@ CREPE: Can Vision-Language Foundation Models Reason Compositionally?,
CVPR23(10910-10921)
IEEE DOI 2309
BibRef

Yin, D.[Da], Gao, F.[Feng], Thattai, G.[Govind], Johnston, M.[Michael], Chang, K.W.[Kai-Wei],
GIVL: Improving Geographical Inclusivity of Vision-Language Models with Pre-Training Methods,
CVPR23(10951-10961)
IEEE DOI 2309
BibRef

Gao, C.[Chen], Peng, X.Y.[Xing-Yu], Yan, M.[Mi], Wang, H.[He], Yang, L.R.[Li-Rong], Ren, H.B.[Hai-Bing], Li, H.S.[Hong-Sheng], Liu, S.[Si],
Adaptive Zone-aware Hierarchical Planner for Vision-Language Navigation,
CVPR23(14911-14920)
IEEE DOI 2309
BibRef

Yeh, C.H.[Chun-Hsiao], Russell, B.[Bryan], Sivic, J.[Josef], Heilbron, F.C.[Fabian Caba], Jenni, S.[Simon],
Meta-Personalizing Vision-Language Models to Find Named Instances in Video,
CVPR23(19123-19132)
IEEE DOI 2309
BibRef

Gou, Y.H.[Yun-Hao], Ko, T.[Tom], Yang, H.[Hansi], Kwok, J.[James], Zhang, Y.[Yu], Wang, M.X.[Ming-Xuan],
Leveraging per Image-Token Consistency for Vision-Language Pre-Training,
CVPR23(19155-19164)
IEEE DOI 2309
BibRef

Wang, S.J.[Shi-Jie], Chang, J.L.[Jian-Long], Li, H.J.[Hao-Jie], Wang, Z.H.[Zhi-Hui], Ouyang, W.L.[Wan-Li], Tian, Q.[Qi],
Open-Set Fine-Grained Retrieval via Prompting Vision-Language Evaluator,
CVPR23(19381-19391)
IEEE DOI 2309
BibRef

Cheng, F.[Feng], Wang, X.[Xizi], Lei, J.[Jie], Crandall, D.[David], Bansal, M.[Mohit], Bertasius, G.[Gedas],
VindLU: A Recipe for Effective Video-and-Language Pretraining,
CVPR23(10739-10750)
IEEE DOI 2309
BibRef

Zhou, H.[Honglu], Martín-Martín, R.[Roberto], Kapadia, M.[Mubbasir], Savarese, S.[Silvio], Niebles, J.C.[Juan Carlos],
Procedure-Aware Pretraining for Instructional Video Understanding,
CVPR23(10727-10738)
IEEE DOI 2309
BibRef

Yang, A.[Antoine], Nagrani, A.[Arsha], Seo, P.H.[Paul Hongsuck], Miech, A.[Antoine], Pont-Tuset, J.[Jordi], Laptev, I.[Ivan], Sivic, J.[Josef], Schmid, C.[Cordelia],
Vid2Seq: Large-Scale Pretraining of a Visual Language Model for Dense Video Captioning,
CVPR23(10714-10726)
IEEE DOI 2309
BibRef

Ji, Y.[Yatai], Tu, R.C.[Rong-Cheng], Jiang, J.[Jie], Kong, W.J.[Wei-Jie], Cai, C.[Chengfei], Zhao, W.Z.[Wen-Zhe], Wang, H.F.[Hong-Fa], Yang, Y.[Yujiu], Liu, W.[Wei],
Seeing What You Miss: Vision-Language Pre-training with Semantic Completion Learning,
CVPR23(6789-6798)
IEEE DOI 2309
BibRef

Alper, M.[Morris], Fiman, M.[Michael], Averbuch-Elor, H.[Hadar],
Is BERT Blind? Exploring the Effect of Vision-and-Language Pretraining on Visual Language Understanding,
CVPR23(6778-6788)
IEEE DOI 2309
BibRef

Liu, M.Y.[Meng-Yin], Jiang, J.[Jie], Zhu, C.[Chao], Yin, X.C.[Xu-Cheng],
VLPD: Context-Aware Pedestrian Detection via Vision-Language Semantic Self-Supervision,
CVPR23(6662-6671)
IEEE DOI 2309
BibRef

Wei, Y.X.[Yi-Xuan], Cao, Y.[Yue], Zhang, Z.[Zheng], Peng, H.[Houwen], Yao, Z.L.[Zhu-Liang], Xie, Z.[Zhenda], Hu, H.[Han], Guo, B.[Baining],
iCLIP: Bridging Image Classification and Contrastive Language-Image Pre-training for Visual Recognition,
CVPR23(2776-2786)
IEEE DOI 2309
BibRef

Hyung, J.[Junha], Hwang, S.[Sungwon], Kim, D.[Daejin], Lee, H.[Hyunji], Choo, J.[Jaegul],
Local 3D Editing via 3D Distillation of CLIP Knowledge,
CVPR23(12674-12684)
IEEE DOI 2309
BibRef

Mu, N.[Norman], Kirillov, A.[Alexander], Wagner, D.[David], Xie, S.[Saining],
SLIP: Self-supervision Meets Language-Image Pre-training,
ECCV22(XXVI:529-544).
Springer DOI 2211
BibRef

Crowson, K.[Katherine], Biderman, S.[Stella], Kornis, D.[Daniel], Stander, D.[Dashiell], Hallahan, E.[Eric], Castricato, L.[Louis], Raff, E.[Edward],
VQGAN-CLIP: Open Domain Image Generation and Editing with Natural Language Guidance,
ECCV22(XXXVII:88-105).
Springer DOI 2211
BibRef

Wu, X.S.[Xiao-Shi], Zhu, F.[Feng], Zhao, R.[Rui], Li, H.S.[Hong-Sheng],
CORA: Adapting CLIP for Open-Vocabulary Detection with Region Prompting and Anchor Pre-Matching,
CVPR23(7031-7040)
IEEE DOI 2309
BibRef

Dong, X.Y.[Xiao-Yi], Bao, J.M.[Jian-Min], Zheng, Y.[Yinglin], Zhang, T.[Ting], Chen, D.D.[Dong-Dong], Yang, H.[Hao], Zeng, M.[Ming], Zhang, W.M.[Wei-Ming], Yuan, L.[Lu], Chen, D.[Dong], Wen, F.[Fang], Yu, N.H.[Neng-Hai],
MaskCLIP: Masked Self-Distillation Advances Contrastive Language-Image Pretraining,
CVPR23(10995-11005)
IEEE DOI 2309
BibRef

Xie, C.W.[Chen-Wei], Sun, S.Y.[Si-Yang], Xiong, X.[Xiong], Zheng, Y.[Yun], Zhao, D.L.[De-Li], Zhou, J.R.[Jing-Ren],
RA-CLIP: Retrieval Augmented Contrastive Language-Image Pre-Training,
CVPR23(19265-19274)
IEEE DOI 2309
BibRef

Chen, P.J.[Pei-Jie], Li, Q.[Qi], Biaz, S.[Saad], Bui, T.[Trung], Nguyen, A.[Anh],
gScoreCAM: What Objects Is CLIP Looking At?,
ACCV22(IV:588-604).
Springer DOI 2307
Analysis of OpenAI's CLIP. BibRef

Wang, R.[Runqi], Duan, X.Y.[Xiao-Yue], Kang, G.L.[Guo-Liang], Liu, J.Z.[Jian-Zhuang], Lin, S.H.[Shao-Hui], Xu, S.[Songcen], Lv, J.[Jinhu], Zhang, B.C.[Bao-Chang],
AttriCLIP: A Non-Incremental Learner for Incremental Knowledge Learning,
CVPR23(3654-3663)
IEEE DOI 2309

WWW Link. BibRef

Rasheed, H.[Hanoona], Khattak, M.U.[Muhammad Uzair], Maaz, M.[Muhammad], Khan, S.[Salman], Khan, F.S.[Fahad Shahbaz],
Fine-tuned CLIP Models are Efficient Video Learners,
CVPR23(6545-6554)
IEEE DOI 2309

WWW Link. BibRef

Liu, R.[Ruyang], Huang, J.J.[Jing-Jia], Li, G.[Ge], Feng, J.S.[Jia-Shi], Wu, X.L.[Xing-Long], Li, T.H.[Thomas H.],
Revisiting Temporal Modeling for CLIP-Based Image-to-Video Knowledge Transferring,
CVPR23(6555-6564)
IEEE DOI 2309
BibRef

Yu, W.W.[Wen-Wen], Liu, Y.L.[Yu-Liang], Hua, W.[Wei], Jiang, D.Q.[De-Qiang], Ren, B.[Bo], Bai, X.[Xiang],
Turning a CLIP Model into a Scene Text Detector,
CVPR23(6978-6988)
IEEE DOI 2309
BibRef

Tschannen, M.[Michael], Mustafa, B.[Basil], Houlsby, N.[Neil],
CLIPPO: Image-and-Language Understanding from Pixels Only,
CVPR23(11006-11017)
IEEE DOI 2309
BibRef

Zhou, Z.Q.[Zi-Qin], Lei, Y.J.[Yig-Jie], Zhang, B.[Bowen], Liu, L.Q.[Ling-Qiao], Liu, Y.F.[Yi-Fan],
ZegCLIP: Towards Adapting CLIP for Zero-shot Semantic Segmentation,
CVPR23(11175-11185)
IEEE DOI 2309
BibRef

He, W.B.[Wen-Bin], Jamonnak, S.[Suphanut], Gou, L.[Liang], Ren, L.[Liu],
CLIP-S4: Language-Guided Self-Supervised Semantic Segmentation,
CVPR23(11207-11216)
IEEE DOI 2309
BibRef

Huang, Z.X.[Zi-Xuan], Jampani, V.[Varun], Thai, A.[Anh], Li, Y.Z.[Yuan-Zhen], Stojanov, S.[Stefan], Rehg, J.M.[James M.],
ShapeClipper: Scalable 3D Shape Learning from Single-View Images via Geometric and CLIP-Based Consistency,
CVPR23(12912-12922)
IEEE DOI 2309
BibRef

Tao, M.[Ming], Bao, B.K.[Bing-Kun], Tang, H.[Hao], Xu, C.S.[Chang-Sheng],
GALIP: Generative Adversarial CLIPs for Text-to-Image Synthesis,
CVPR23(14214-14223)
IEEE DOI 2309
BibRef

Zeng, Y.[Yihan], Jiang, C.[Chenhan], Mao, J.[Jiageng], Han, J.H.[Jian-Hua], Ye, C.Q.[Chao-Qiang], Huang, Q.Q.[Qing-Qiu], Yeung, D.Y.[Dit-Yan], Yang, Z.[Zhen], Liang, X.D.[Xiao-Dan], Xu, H.[Hang],
CLIP2: Contrastive Language-Image-Point Pretraining from Real-World Point Cloud Data,
CVPR23(15244-15253)
IEEE DOI 2309
BibRef

Lin, Y.Q.[Yu-Qi], Chen, M.H.[Ming-Hao], Wang, W.X.[Wen-Xiao], Wu, B.[Boxi], Li, K.[Ke], Lin, B.B.[Bin-Bin], Liu, H.F.[Hai-Feng], He, X.F.[Xiao-Fei],
CLIP is Also an Efficient Segmenter: A Text-Driven Approach for Weakly Supervised Semantic Segmentation,
CVPR23(15305-15314)
IEEE DOI 2309
BibRef

Sanghi, A.[Aditya], Fu, R.[Rao], Liu, V.[Vivian], Willis, K.D.D.[Karl D.D.], Shayani, H.[Hooman], Khasahmadi, A.H.[Amir H.], Sridhar, S.[Srinath], Ritchie, D.[Daniel],
CLIP-Sculptor: Zero-Shot Generation of High-Fidelity and Diverse Shapes from Natural Language,
CVPR23(18339-18348)
IEEE DOI 2309
BibRef

Pei, R.J.[Ren-Jing], Liu, J.Z.[Jian-Zhuang], Li, W.[Weimian], Shao, B.[Bin], Xu, S.[Songcen], Dai, P.[Peng], Lu, J.W.[Ju-Wei], Yan, Y.[Youliang],
CLIPPING: Distilling CLIP-Based Models with a Student Base for Video-Language Retrieval,
CVPR23(18983-18992)
IEEE DOI 2309
BibRef

Jeong, J.[Jongheon], Zou, Y.[Yang], Kim, T.[Taewan], Zhang, D.Q.[Dong-Qing], Ravichandran, A.[Avinash], Dabeer, O.[Onkar],
WinCLIP: Zero-/Few-Shot Anomaly Classification and Segmentation,
CVPR23(19606-19616)
IEEE DOI 2309
BibRef

Shamshad, F.[Fahad], Naseer, M.[Muzammal], Nandakumar, K.[Karthik],
CLIP2Protect: Protecting Facial Privacy Using Text-Guided Makeup via Adversarial Latent Search,
CVPR23(20595-20605)
IEEE DOI 2309
BibRef

Chen, Y.H.[Yi-Hao], Qi, X.[Xianbiao], Wang, J.A.[Jian-An], Zhang, L.[Lei],
DisCo-CLIP: A Distributed Contrastive Loss for Memory Efficient CLIP Training,
CVPR23(22648-22657)
IEEE DOI 2309
BibRef

Wasim, S.T.[Syed Talal], Naseer, M.[Muzammal], Khan, S.[Salman], Khan, F.S.[Fahad Shahbaz], Shah, M.[Mubarak],
Vita-CLIP: Video and text adaptive CLIP via Multimodal Prompting,
CVPR23(23034-23044)
IEEE DOI 2309
BibRef

Parelli, M.[Maria], Delitzas, A.[Alexandros], Hars, N.[Nikolas], Vlassis, G.[Georgios], Anagnostidis, S.[Sotirios], Bachmann, G.[Gregor], Hofmann, T.[Thomas],
CLIP-Guided Vision-Language Pre-training for Question Answering in 3D Scenes,
ODRUM23(5607-5612)
IEEE DOI 2309
BibRef

Ning, S.[Shan], Qiu, L.[Longtian], Liu, Y.F.[Yong-Fei], He, X.M.[Xu-Ming],
HOICLIP: Efficient Knowledge Transfer for HOI Detection with Vision-Language Models,
CVPR23(23507-23517)
IEEE DOI 2309
BibRef

Yao, L.W.[Le-Wei], Han, J.H.[Jian-Hua], Liang, X.D.[Xiao-Dan], Xu, D.[Dan], Zhang, W.[Wei], Li, Z.G.[Zhen-Guo], Xu, H.[Hang],
DetCLIPv2: Scalable Open-Vocabulary Object Detection Pre-training via Word-Region Alignment,
CVPR23(23497-23506)
IEEE DOI 2309
BibRef

Singha, M.[Mainak], Jha, A.[Ankit], Solanki, B.[Bhupendra], Bose, S.[Shirsha], Banerjee, B.[Biplab],
APPLeNet: Visual Attention Parameterized Prompt Learning for Few-Shot Remote Sensing Image Generalization using CLIP,
EarthVision23(2024-2034)
IEEE DOI 2309
BibRef

Zhu, J.[Jun], Jin, J.[Jiandong], Yang, Z.[Zihan], Wu, X.[Xiaohao], Wang, X.[Xiao],
Learning CLIP Guided Visual-Text Fusion Transformer for Video-based Pedestrian Attribute Recognition,
NFVLR23(2626-2629)
IEEE DOI 2309
BibRef

Gannamaneni, S.S.[Sujan Sai], Sadaghiani, A.[Arwin], Rao, R.P.[Rohil Prakash], Mock, M.[Michael], Akila, M.[Maram],
Investigating CLIP Performance for Meta-data Generation in AD Datasets,
SAIAD23(3840-3850)
IEEE DOI 2309
BibRef

Chen, R.N.[Run-Nan], Liu, Y.Q.[You-Quan], Kong, L.D.[Ling-Dong], Zhu, X.G.[Xin-Ge], Ma, Y.X.[Yue-Xin], Li, Y.[Yikang], Hou, Y.N.[Yue-Nan], Qiao, Y.[Yu], Wang, W.P.[Wen-Ping],
CLIP2Scene: Towards Label-efficient 3D Scene Understanding by CLIP,
CVPR23(7020-7030)
IEEE DOI 2309
BibRef

Ni, B.L.[Bo-Lin], Peng, H.W.[Hou-Wen], Chen, M.H.[Ming-Hao], Zhang, S.Y.[Song-Yang], Meng, G.F.[Gao-Feng], Fu, J.L.[Jian-Long], Xiang, S.M.[Shi-Ming], Ling, H.B.[Hai-Bin],
Expanding Language-Image Pretrained Models for General Video Recognition,
ECCV22(IV:1-18).
Springer DOI 2211

WWW Link. BibRef

Zhang, R.[Renrui], Zhang, W.[Wei], Fang, R.[Rongyao], Gao, P.[Peng], Li, K.C.[Kun-Chang], Dai, J.F.[Ji-Feng], Qiao, Y.[Yu], Li, H.S.[Hong-Sheng],
Tip-Adapter: Training-Free Adaption of CLIP for Few-Shot Classification,
ECCV22(XXXV:493-510).
Springer DOI 2211
BibRef

Yang, J.[Jinyu], Duan, J.L.[Jia-Li], Tran, S.[Son], Xu, Y.[Yi], Chanda, S.[Sampath], Chen, L.Q.[Li-Qun], Zeng, B.[Belinda], Chilimbi, T.[Trishul], Huang, J.Z.[Jun-Zhou],
Vision-Language Pre-Training with Triple Contrastive Learning,
CVPR22(15650-15659)
IEEE DOI 2210
Representation learning, Visualization, Question answering (information retrieval), Self- semi- meta- unsupervised learning BibRef

Guo, X.Y.[Xiao-Yuan], Duan, J.L.[Jia-Li], Kuo, C.C.J.[C.C. Jay], Gichoya, J.W.[Judy Wawira], Banerjee, I.[Imon],
Augmenting Vision Language Pretraining by Learning Codebook with Visual Semantics,
ICPR22(4779-4785)
IEEE DOI 2212
Representation learning, Bridges, Visualization, Vocabulary, Semantics, Buildings, Benchmark testing BibRef

Zhou, C.[Chong], Loy, C.C.[Chen Change], Dai, B.[Bo],
Extract Free Dense Labels from CLIP,
ECCV22(XXVIII:696-712).
Springer DOI 2211

WWW Link. Contrastive Language-Image Pre-training. BibRef

Lin, Z.[Ziyi], Geng, S.J.[Shi-Jie], Zhang, R.[Renrui], Gao, P.[Peng], de Melo, G.[Gerard], Wang, X.G.[Xiao-Gang], Dai, J.F.[Ji-Feng], Qiao, Y.[Yu], Li, H.S.[Hong-Sheng],
Frozen CLIP Models are Efficient Video Learners,
ECCV22(XXXV:388-404).
Springer DOI 2211
BibRef

Rao, Y.M.[Yong-Ming], Zhao, W.L.[Wen-Liang], Chen, G.Y.[Guang-Yi], Tang, Y.S.[Yan-Song], Zhu, Z.[Zheng], Huang, G.[Guan], Zhou, J.[Jie], Lu, J.W.[Ji-Wen],
DenseCLIP: Language-Guided Dense Prediction with Context-Aware Prompting,
CVPR22(18061-18070)
IEEE DOI 2210
Representation learning, Image segmentation, Visualization, Shape, Computational modeling, Semantics, Predictive models, grouping and shape analysis BibRef

Kwon, G.[Gihyun], Ye, J.C.[Jong Chul],
CLIPstyler: Image Style Transfer with a Single Text Condition,
CVPR22(18041-18050)
IEEE DOI 2210
Semantics, Modulation, Pattern recognition, Vision+language, Image and video synthesis and generation BibRef

Khandelwal, A.[Apoorv], Weihs, L.[Luca], Mottaghi, R.[Roozbeh], Kembhavi, A.[Aniruddha],
Simple but Effective: CLIP Embeddings for Embodied AI,
CVPR22(14809-14818)
IEEE DOI 2210
Contrastive Language Image Pretraining. Training, Measurement, Visualization, Navigation, Semantics, Robot vision systems, Robot vision, Navigation and autonomous driving BibRef

Ma, H.Y.[Hao-Yu], Zhao, H.[Handong], Lin, Z.[Zhe], Kale, A.[Ajinkya], Wang, Z.Y.[Zhang-Yang], Yu, T.[Tong], Gu, J.X.[Jiu-Xiang], Choudhary, S.[Sunav], Xie, X.H.[Xiao-Hui],
EI-CLIP: Entity-aware Interventional Contrastive Learning for E-commerce Cross-modal Retrieval,
CVPR22(18030-18040)
IEEE DOI 2210
Design methodology, Semantics, Clothing, Metadata, Benchmark testing, Search problems, BibRef

Barraco, M.[Manuele], Cornia, M.[Marcella], Cascianelli, S.[Silvia], Baraldi, L.[Lorenzo], Cucchiara, R.[Rita],
The Unreasonable Effectiveness of CLIP Features for Image Captioning: An Experimental Analysis,
MULA22(4661-4669)
IEEE DOI 2210
Visualization, Protocols, Detectors, Distance measurement, Data models BibRef

Tevet, G.[Guy], Gordon, B.[Brian], Hertz, A.[Amir], Bermano, A.H.[Amit H.], Cohen-Or, D.[Daniel],
MotionCLIP: Exposing Human Motion Generation to CLIP Space,
ECCV22(XXII:358-374).
Springer DOI 2211

WWW Link. Contrastive Language-Image Pre-training (CLIP) applied to motion. BibRef

Materzynska, J.[Joanna], Torralba, A.[Antonio], Bau, D.[David],
Disentangling visual and written concepts in CLIP,
CVPR22(16389-16398)
IEEE DOI 2210
Visualization, Image coding, Benchmark testing, Cognition, Pattern recognition, Task analysis, Vision + language, Visual reasoning BibRef

Li, M.[Manling], Xu, R.[Ruochen], Wang, S.[Shuohang], Zhou, L.[Luowei], Lin, X.D.[Xu-Dong], Zhu, C.G.[Chen-Guang], Zeng, M.[Michael], Ji, H.[Heng], Chang, S.F.[Shih-Fu],
CLIP-Event: Connecting Text and Images with Event Structures,
CVPR22(16399-16408)
IEEE DOI 2210
Codes, Computational modeling, Image retrieval, Benchmark testing, Information retrieval, Pattern recognition, Vision + language BibRef

Zhong, Y.[Yiwu], Yang, J.W.[Jian-Wei], Zhang, P.C.[Peng-Chuan], Li, C.Y.[Chun-Yuan], Codella, N.[Noel], Li, L.H.[Liunian Harold], Zhou, L.[Luowei], Dai, X.[Xiyang], Yuan, L.[Lu], Li, Y.[Yin], Gao, J.F.[Jian-Feng],
RegionCLIP: Region-based Language-Image Pretraining,
CVPR22(16772-16782)
IEEE DOI 2210
Representation learning, Visualization, Technological innovation, Image recognition, Text recognition, Transfer learning, Vision + language BibRef

Baldrati, A.[Alberto], Bertini, M.[Marco], Uricchio, T.[Tiberio], del Bimbo, A.[Alberto],
Effective conditioned and composed image retrieval combining CLIP-based features,
CVPR22(21434-21442)
IEEE DOI 2210
BibRef
And:
Conditioned and composed image retrieval combining and partially fine-tuning CLIP-based features,
ODRUM22(4955-4964)
IEEE DOI 2210
Visualization, Shape, Image retrieval, Pipelines, Search engines, Pattern recognition. Training, Visualization, Solids, Pattern recognition BibRef

Patashnik, O.[Or], Wu, Z.Z.[Zong-Ze], Shechtman, E.[Eli], Cohen-Or, D.[Daniel], Lischinski, D.[Dani],
StyleCLIP: Text-Driven Manipulation of StyleGAN Imagery,
ICCV21(2065-2074)
IEEE DOI 2203
Hair, Visualization, Image coding, Manuals, Optimization, Image and video synthesis, Vision applications and systems BibRef

Chapter on Implementations and Applications, Databases, QBIC, Video Analysis, Hardware and Software, Inspection continues in
Internet Label Information .


Last update:Apr 18, 2024 at 11:38:49