[EMNLP 2023 Demo] Video-LLaMA: An Instruction-tuned Audio-Visual Language Model for Video Understanding
#大语言模型#(AAAI 2024) BLIVA: A Simple Multimodal LLM for Better Handling of Text-rich Visual Questions
Automate Fashion Image Captioning using BLIP-2. Automatic generating descriptions of clothes on shopping websites, which can help customers without fashion knowledge to better understand the features ...
#计算机科学#Implementation of Qformer from BLIP2 in Zeta Lego blocks.
Official implementation and dataset for the NAACL 2024 paper "ComCLIP: Training-Free Compositional Image and Text Matching"
[ACM MM 2024] Improving Composed Image Retrieval via Contrastive Learning with Scaling Positives and Negatives
The Multimodal Model for Vietnamese Visual Question Answering (ViVQA)
CLIP Interrogator, fully in HuggingFace Transformers 🤗, with LongCLIP & CLIP's own words and / or *your* own words!
Modifying LAVIS' BLIP2 Q-former with models pretrained on Japanese datasets.
#大语言模型#This repository is for profiling, extracting, visualizing and reusing generative AI weights to hopefully build more accurate AI models and audit/scan weights at rest to identify knowledge domains for ...
#数据仓库#Annotations on a Budget: Leveraging Geo-Data Similarity to Balance Model Performance and Annotation Cost
Caption images across your datasets with state of the art models from Hugging Face and Replicate!
Finetuning Large Visual Models on Visual Question Answering
caption generator using lavis and argostranslate
In this we explore into visual Question Answering Using Gemini LLM and image was in URL or any other extension
Too lazy to organize my desktop, make gpt + BLIP-2 do it /s
This project performs multimodal document analysis and query retrieval by downloading PDFs, converting pages to images, indexing them for semantic search, and analyzing retrieved images using visual-l...