GitHub 中文社区
回车: Github搜索    Shift+回车: Google搜索
论坛
排行榜
趋势
登录

©2025 GitHub中文社区论坛GitHub官网网站地图GitHub官方翻译

  • X iconGitHub on X
  • Facebook iconGitHub on Facebook
  • Linkedin iconGitHub on LinkedIn
  • YouTube iconGitHub on YouTube
  • Twitch iconGitHub on Twitch
  • TikTok iconGitHub on TikTok
  • GitHub markGitHub’s organization on GitHub

编程语言

”vision-language“ 的搜索结果

moondream
@vikhyat

tiny vision language model

Python8.19 k
20 天前

相关主题

vision-language-modelvision-and-languagevlm深度学习multimodal-learningfoundation-modelsmoeAwesome Listsvision-language-transformerllm

Google   Bing   GitHub

sglang
@sgl-project

#大语言模型#SGLang is a fast serving framework for large language models and vision language models.

CUDAinferencellamallavallm
Python15.91 k
1 小时前
Salesforce
LAVIS
Salesforce@salesforce

#计算机科学#LAVIS - A One-stop Library for Language-Vision Intelligence

深度学习deep-learning-libraryimage-captioningsalesforcevision-and-language
Jupyter Notebook10.73 k
8 个月前
awesome-vision-language-pretraining-papers
@yuewang-cuhk

Recent Advances in Vision and Language PreTrained Models (VL-PTMs)

vision-and-languagepretrainingmultimodal-deep-learningbert
1.15 k
3 年前
VLM_survey
@jingyi0000

#计算机科学#Collection of AWESOME vision-language models for vision tasks

机器视觉深度学习knowledge-distillationsurveytransfer-learning
2.83 k
2 个月前
DeepSeek-VL
@deepseek-ai

DeepSeek-VL: Towards Real-World Vision-Language Understanding

vision-language-modelvision-language-pretrainingfoundation-models
Python3.91 k
1 年前
DeepSeek-VL2
@deepseek-ai

DeepSeek-VL2: Mixture-of-Experts Vision-Language Models for Advanced Multimodal Understanding

Python4.95 k
5 个月前
Meta Research
vilbert-multi-task存档
@facebookresearch • Meta

Multi Task Vision and Language

Jupyter Notebook813
3 年前
awesome-vision-and-language
@sangminwoo

#Awesome#A curated list of awesome vision and language resources (still under construction... stay tuned!)

Awesome Listsvision-and-languagemultimodal-learning
540
8 个月前
Salesforce
BLIP
Salesforce@salesforce

PyTorch code for BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation

vision-languagevision-and-language-pre-trainingimage-text-retrievalimage-captioningvisual-question-answering
Jupyter Notebook5.38 k
1 年前
BriVL
@BAAI-WuDao

Bridging Vision and Language Model

Python276
2 年前
awesome-vlm-architectures
@gokayfem

#Awesome#Famous Vision Language Models and Their Architectures

clipllavavlm
Markdown907
5 个月前
MoE-LLaVA
@PKU-YuanGroup

Mixture-of-Experts for Large Vision-Language Models

large-vision-language-modelmixture-of-expertsmoemulti-modal
Python2.18 k
7 个月前
CONCH
@mahmoodlab

Vision-Language Pathology Foundation Model - Nature Medicine

foundation-model
Python378
4 个月前
Awesome_Prompting_Papers_in_Computer_Vision
@ttengwang

A curated list of prompt-based paper in computer vision and vision-language learning.

prompt-learningadapterfew-shot-learningprompt-tuningzero-shot-learning
921
2 年前
VSA
@cnzzx

#计算机科学#Vision Search Assistant: Empower Vision-Language Models as Multimodal Search Engines

ai-search-engine人工智能深度学习大语言模型search
Python125
8 个月前
LongVA
@EvolvingLMMs-Lab

Long Context Transfer from Language to Vision

Python359
8 个月前
谷歌公司
tirg
谷歌公司@google

deep learning, image retrieval, vision and language

Python305
4 年前
苹果公司
ml-fastvlm
苹果公司@apple

This repository contains the official implementation of "FastVLM: Efficient Vision Encoding for Vision Language Models" - CVPR 2025

Python4.31 k
2 个月前
VL-RLHF
@TideDra

#大语言模型#A RLHF Infrastructure for Vision-Language Models

dpollmlmmmllmrlhf
Python177
8 个月前
Vary
@Ucas-HaoranWei

[ECCV 2024] Official code implementation of Vary: Scaling Up the Vision Vocabulary of Large Vision Language Models.

Python1.84 k
6 个月前
CoOp
@KaiyangZhou

Prompt Learning for Vision-Language Models (IJCV'22, CVPR'22)

foundation-modelsmultimodal-learningprompt-learning
Python2.01 k
1 年前
Salesforce
ALBEF存档
Salesforce@salesforce

Code for ALBEF: a new vision-language pre-training method

vision-and-languagerepresentation-learningcontrastive-learning
Python1.67 k
3 年前
loading...