赞
踩
视觉领域的同学应该有所体会,获取大量标注数据是一件成本非常高的事。为了应对这个问题,研究者们通过借助无标注数据、图文数据或者多模态数据等,采用对比学习、掩码重建等学习方式预训练得到视觉基础模型,用于适应各种下游任务,比如物体检测、语义分割等。在过去一年中,由于LLM、多模态等领域的快速发展,更多新兴的计算机视觉基础模型被提出。
到目前为止,已发布的计算机视觉基础模型数目已经相当可观,对于视觉领域的同学来说,这些基础模型具有非常高的研究价值。为了方便同学们了解并掌握该领域的最新进展,发出属于自己的顶会,我今天就和大家分享一篇综述论文,该文作者对计算机视觉领域的基础模型进行了详细的梳理,涵盖了13大类算法模型,以及每一类模型的变种共85个,从最早的LeNet、ResNet到最新的SAM、GPT4等都有。
综述链接:https://arxiv.org/pdf/2307.13721.pdf
除此之外,学姐也帮大家整理了120篇21年-23年必读的CV领域算法模型的代表性论文,部分代码已开源。
尽管已有的方法表现不俗,但我们清楚,计算机视觉基础模型的发展仍然有巨大的进步空间,希望同学们能通过这份资料全面掌握计算机视觉领域的发展脉络,厘清每个模型的变化历史,并从中找到更优解。
Foundational Models Defining a New Era in Vision: A Survey and Outlook 2023
A of Large Language Models 2023
Harnessing the Power of LLMs in Practice: A Survey on ChatGPT and Beyond 2023
Multimodal Learning with Transformers: A Survey 2023
Self-Supervised Multimodal Learning: A Survey
Vision-and-Language Pretrained Models: A Survey 2022
A Survey of Vision-Language Pre-Trained Models 2022
Vision-Language Models for Vision Tasks: A Survey 2022
A Comprehensive Survey on Segment Anything Model for Vision and Beyond 2023
Vision-language pre-training: Basics, recent advances, and future trends 2022
Towards Open Vocabulary Learning: A Survey 2023
Transformer-Based Visual Segmentation: A Survey 2023
Scaling Up Visual and Vision-Language Representation Learning With Noisy Text Supervision 2021-02-11
Learning Transferable Visual Models From Natural Language Supervision 2021-02-26
WenLan: Bridging Vision and Language by Large-Scale Multi-Modal Pre-Training 2021-03-11
Open-vocabulary Object Detection via Vision and Language Knowledge Distillation 2021-04-28
CLIP2Video: Mastering Video-Text Retrieval via Image CLIP 2021-06-21
AudioCLIP: Extending CLIP to Image, Text and Audio 2021-06-24
Multimodal Few-Shot Learning with Frozen Language Models 2021-06-25
SimVLM: Simple Visual Language Model Pretraining with Weak Supervision 2021-08-24
LAION-400M: Open Dataset of CLIP-Filtered 400 Million Image-Text Pairs 2021-11-03
FILIP: Fine-grained Interactive Language-Image Pre-Training 2021-11-09
Florence: A New Foundation Model for Computer Vision 2021-11-22
Extract Free Dense Labels from CLIP 2021-12-02
FLAVA: A Foundational Language And Vision Alignment Model 2021-12-08
Image Segmentation Using Text and Image Prompts 2021-12-18
Scaling Open-Vocabulary Image Segmentation with Image-Level Labels 2021-12-22
GroupViT: Semantic Segmentation Emerges from Text Supervision 2022-02-22
CoCa: Contrastive Captioners are Image-Text Foundation Models 2022-05-04
Simple Open-Vocabulary Object Detection with Vision Transformers 2022-05-12
GIT: A Generative Image-to-text Transformer for Vision and Language 2022-05-27
Language Models are General-Purpose Interfaces 2022-06-13
Coarse-to-Fine Vision-Language Pre-training with Fusion in the Backbone 2022-06-15
A Unified Sequence Interface for Vision Tasks 2022-06-15
BridgeTower: Building Bridges Between Encoders in Vision-Language Representation Learning 2022-06-17
MineDojo: Building Open-Ended Embodied Agents with Internet-Scale Knowledge 2022-06-17
LM-Nav: Robotic Navigation with Large Pre-Trained Models of Language, Vision, and Action 2022-07-10
Masked Vision and Language Modeling for Multi-modal Representation Learning 2022-08-03
PaLI: A Jointly-Scaled Multilingual Language-Image Model 2022-09-14
VIMA: General Robot Manipulation with Multimodal Prompts 2022-10-06
Images Speak in Images: A Generalist Painter for In-Context Visual Learning 2022-12-05
InternVideo: General Video Foundation Models via Generative and Discriminative Learning 2022-12-07
Reproducible scaling laws for contrastive language-image learning 2022-12-14
Toward Building General Foundation Models for Language, Vision, and Vision-Language Understanding Tasks 2023-01-12
BLIP-2: Bootstrapping Language-Image Pre-training with Frozen Image Encoders and Large Language Models 2023-01-30
Grounding Language Models to Images for Multimodal Inputs and Outputs 2023-01-31
Language Is Not All You Need: Aligning Perception with Language Models 2023-02-27
Prismer: A Vision-Language Model with An Ensemble of Experts 2023-03-04
PaLM-E: An Embodied Multimodal Language Model 2023-03-06
Visual ChatGPT: Talking, Drawing and Editing with Visual Foundation Models 2023-03-08
Task and Motion Planning with Large Language Models for Object Rearrangement 2023-03-10
GPT-4 Technical Report 2023-03-15
EVA-02: A Visual Representation for Neon Genesis 2023-03-20
MM-REACT: Prompting ChatGPT for Multimodal Reasoning and Action 2023-03-20
Detecting Everything in the Open World: Towards Universal Object Detection 2023-03-21
Errors are Useful Prompts: Instruction Guided Task Programming with Verifier-Assisted Iterative Prompting 2023-03-24
EVA-CLIP: Improved Training Techniques for CLIP at Scale 2023-03-27
Unmasked Teacher: Towards Training-Efficient Video Foundation Models 2023-03-28
ViewRefer: Grasp the Multi-view Knowledge for 3D Visual Grounding with GPT and Prototype Guidance 2023-03-29
HuggingGPT: Solving AI Tasks with ChatGPT and its Friends in Hugging Face 2023-03-30
ERRA: An Embodied Representation and Reasoning Architecture for Long-horizon Language-conditioned Manipulation Tasks 2023-04-05
Segment Anything 2023-04-05
SegGPT: Segmenting Everything In Context 2023-04-06
ChatGPT Empowered Long-Step Robot Control in Various Environments: A Case Application 2023-04-08
Video ChatCaptioner: Towards Enriched Spatiotemporal Descriptions 2023-04-09
OpenAGI: When LLM Meets Domain Experts 2023-04-10
Graph-ToolFormer: To Empower LLMs with Graph Reasoning Ability via Prompt Augmented by ChatGPT 2023-04-10
Advancing Medical Imaging with Language Models: A Journey from N-grams to ChatGPT 2023-04-11
SAMM (Segment Any Medical Model): A 3D Slicer Integration to SAM 2023-04-12
Segment Everything Everywhere All at Once 2023-04-13
Visual Instruction Tuning 2023-04-17
Chameleon: Plug-and-Play Compositional Reasoning with Large Language Models 2023-04-19
MiniGPT-4: Enhancing Vision-Language Understanding with Advanced Large Language Models 2023-04-20
Can GPT-4 Perform Neural Architecture Search? 2023-04-21
Evaluating ChatGPT's Information Extraction Capabilities: An Assessment of Performance, Explainability, Calibration, and Faithfulness 2023-04-23
Track Anything: Segment Anything Meets Videos 2023-04-24
Segment Anything in Medical Images 2023-04-24
Medical SAM Adapter: Adapting Segment Anything Model for Medical Image Segmentation 2023-04-25
Learnable Ophthalmology SAM 2023-04-26
LLaMA-Adapter V2: Parameter-Efficient Visual Instruction Model 2023-04-28
Transfer Visual Prompt Generator across LLMs 2023-05-02
Caption Anything: Interactive Image Description with Diverse Multimodal Controls 2023-05-04
ImageBind: One Embedding Space To Bind Them All 2023-05-09
InstructBLIP: Towards General-purpose Vision-Language Models with Instruction Tuning 2023-05-11
Segment and Track Anything 2023-05-11
An Inverse Scaling Law for CLIP Training 2023-05-11
VisionLLM: Large Language Model is also an Open-Ended Decoder for Vision-Centric Tasks 2023-05-18
Cream: Visually-Situated Natural Language Understanding with Contrastive Reading Model and Frozen Large Language Models 2023-05-24
Voyager: An Open-Ended Embodied Agent with Large Language Models 2023-05-25
DeSAM: Decoupling Segment Anything Model for Generalizable Medical Image Segmentation 2023-06-01
Video-ChatGPT: Towards Detailed Video Understanding via Large Vision and Language Models 2023-06-08
Valley: Video Assistant with Large Language model Enhanced abilitY 2023-06-12
mPLUG-owl: Modularization empowers large language models with multimodality 2023-04-27
Image Captioners Are Scalable Vision Learners Too 2023-06-13
XrayGPT: Chest Radiographs Summarization using Medical Vision-Language Models 2023-06-13
ViP: A Differentially Private Foundation Model for Computer Vision 2023-06-15
COSA: Concatenated Sample Pretrained Vision-Language Foundation Model 2023-06-15
LVLM-eHub: A Comprehensive Evaluation Benchmark for Large Vision-Language Models 2023-06-15
Segment Any Point Cloud Sequences by Distilling Vision Foundation Models 2023-06-15
RemoteCLIP: A Vision Language Foundation Model for Remote Sensing 2023-06-19
LVM-Med: Learning Large-Scale Self-Supervised Vision Models for Medical Imaging via Second-order Graph Matching 2023-06-20
Fast Segment Anything 2023-06-21
TaCA: Upgrading Your Visual Foundation Model with Task-agnostic Compatible Adapter 2023-06-22
3DSAM-adapter: Holistic Adaptation of SAM from 2D to 3D for Promptable Medical Image Segmentation 2023-06-23
How to Efficiently Adapt Large Segmentation Model(SAM) to Medical Images 2023-06-23
Faster Segment Anything: Towards Lightweight SAM for Mobile Applications 2023-06-25
MedLSAM: Localize and Segment Anything Model for 3D Medical Images 2023-06-26
LVM-Med:LVM-Med: Learning Large-Scale Self-Supervised Vision Models for Medical Imaging via Second-order Graph Matching
Kosmos-2: Grounding Multimodal Large Language Models to the World 2023-06-26
ViNT: A Foundation Model for Visual Navigation 2023-06-26
CLIPA-v2: Scaling CLIP Training with 81.1% Zero-shot ImageNet Accuracy within a $10,000 Budget; An Extra $4,000 Unlocks 81.8% Accuracy 2023-06-27
Stone Needle: A General Multimodal Large-scale Model Framework towards Healthcare 2023-06-28
RSPrompter: Learning to Prompt for Remote Sensing Instance Segmentation based on Visual Foundation Model 2023-06-28
Towards Language Models That Can See: Computer Vision Through the LENS of Natural Language 2023-06-28
Foundation Model for Endoscopy Video Analysis via Large-scale Self-supervised Pre-train 2023-06-29
MIS-FM: 3D Medical Image Segmentation using Foundation Models Pretrained on a Large-Scale Unannotated Dataset 2023-06-29
RefSAM: Efficiently Adapting Segmenting Anything Model for Referring Video Object Segmentation 2023-07-03
SAM-DA: UAV Tracks Anything at Night with SAM-Powered Domain Adaptation 2023-07-03
Segment Anything Meets Point Tracking 2023-07-03
BuboGPT: Enabling Visual Grounding in Multi-Modal LLMs 2023-07-17
关注下方【学姐带你玩AI】声明:本文内容由网友自发贡献,转载请注明出处:【wpsshop】
Copyright © 2003-2013 www.wpsshop.cn 版权所有,并保留所有权利。