Skip to content

The code for paper:Joint Visual and Text Prompting for Improved Object-Centric Perception with Multimodal Large Language Models

Notifications You must be signed in to change notification settings

jiangsongtao/VTprompt

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

4 Commits
 
 

Repository files navigation

VTprompt

This repository contains the code for the paper: "Joint Visual and Text Prompting for Improved Object-Centric Perception with Multimodal Large Language Models". image

Installation

Environment Setup

Please follow the instructions in Grounded Segment Anything to set up the environment.

Usage

  1. Building Vprompt
  2. Using Tprompt to prompt Multimodal Large Language Models for generating answers.

Evaluation Code and Usage Tutorial

We are currently in the process of organizing detailed evaluation code and usage tutorials. Please stay tuned for updates!

About

The code for paper:Joint Visual and Text Prompting for Improved Object-Centric Perception with Multimodal Large Language Models

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published