Skip to content

MLLM-Tool/MLLM-Tool

Folders and files

NameName
Last commit message
Last commit date

Latest commit

Β 

History

59 Commits
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 
Β 

Repository files navigation

MLLM-Tool: A Multimodal Large Language Model For Tool Agent Learning

Chenyu Wang, Weixin Luo, Qianyu Chen, Haonan Mai, Jindi Guo, Sixun Dong, Xiaohua (Michael) Xuan, Zhengxin Li, Lin Ma, Shenghua Gao.

ShanghaiTech University && Meituan && UniDT


This repository hosts the code, data and model weight of MLLM-Tool, the first tool agent MLLM that has the ability to perceive visual- and auditory- input information and recommend appropriate tools for multi-modal instructions.


πŸŽ‰ News

  • [2024.02.02] πŸ“’πŸ“’ We change the permissions for data and checkpoints, no longer need to apply to download them.
  • [2024.01.16] πŸš€πŸš€ Release the code of MLLM-Tool.
  • [2024.01.16] πŸ”¨πŸ§© Release the ToolMMBench dataset.
  • [2024.01.16] πŸ“’πŸ“’ Release the checkpoint of MLLM-Tool in Vicuna-7B, Vicuna-13B, Llama-7B, Llama-13B, Llama2-7B, Llama2-13B, Llama2Chat-7B, Llama2Chat-13B.

πŸ‘‰ TODO

  • Collect more data and release the v2 dataset.
  • Update MLLM-Tool in more types & sizes of LLMs.
  • Empower MLLM-Tool with retrieving open-set tools.
  • Release Demo and Interactive Website.
  • ...

Brief Introduction

Recently, the astonishing performance of large language models (LLMs) in natural language comprehension and generation tasks triggered lots of exploration of using them as central controllers to build agent systems. Multiple studies focus on bridging the LLMs to external tools to extend the application scenarios. However, the current LLMs' perceiving tool-use ability is limited to a single text query, which may result in ambiguity in understanding the users' real intentions. LLMs are expected to eliminate that by perceiving the visual- or auditory-grounded instructions' information. Therefore, in this paper, we propose MLLM-Tool, a system incorporating open-source LLMs and multimodal encoders so that the learned LLMs can be conscious of multi-modal input instruction and then select the function-matched tool correctly. To facilitate the evaluation of the model’s capability, we collect a dataset featured by consisting of multi-modal input tools from HuggingFace. Another important feature of our dataset is that our dataset also contains multiple potential choices for the same instruction due to the existence of identical functions and synonymous functions, which provides more potential solutions for the same query. The experiments reveal that our MLLM-Tool is capable of recommending appropriate tools for multi-modal instructions.

Getting Started

Table of Contents:


1. Code Structure

β”œβ”€β”€ data
β”‚   β”œβ”€β”€ IT_data_ins                           # instruction data
β”‚   β”‚   └── T+X-T_data                        # text+[image/audio/video] to text instruction data
β”‚   β”‚   β”‚   β”œβ”€β”€ mm_dataset                    # multimodal input data
β”‚   β”‚   β”‚   β”œβ”€β”€ audio_tx2t.json
β”‚   β”‚   β”‚   β”œβ”€β”€ image_tx2t.json
β”‚   β”‚   β”‚   β”œβ”€β”€ text_t2t.json
β”‚   β”‚   β”‚   β”œβ”€β”€ video_tx2t.json
β”‚   β”‚   β”‚   └── combined_data.json
β”œβ”€β”€ code
β”‚   β”œβ”€β”€ config
β”‚   β”‚   β”œβ”€β”€__init__.py
β”‚   β”‚   β”œβ”€β”€ base.yaml                         # the model configuration 
β”‚   β”‚   └── openllama_peft.yaml               # instruction-tuning configuration
β”‚   β”œβ”€β”€ dsconfig
β”‚   β”‚   └──  openllama_peft_stage_1.json      # deepspeed configuration for instruction-tuning training
β”‚   β”œβ”€β”€ dataset
β”‚   β”‚   β”œβ”€β”€__init__ .py
β”‚   β”‚   β”œβ”€β”€_sampler.py
β”‚   β”‚   β”œβ”€β”€_utils.py
β”‚   β”‚   β”œβ”€β”€ catalog.py                        # the catalog information of the dataset
β”‚   β”‚   β”œβ”€β”€ T+X-T_instruction_dataset.py      # process and load text+x-to-text instruction dataset
β”‚   β”‚   └── concat_dataset.py                 # process and load multiple datasets
β”‚   β”œβ”€β”€ model                     
β”‚   β”‚   β”œβ”€β”€ ImageBind                         # the code from ImageBind Model
β”‚   β”‚   β”œβ”€β”€__init__ .py 
β”‚   β”‚   β”œβ”€β”€ openllama.py                      # the main model file
β”‚   β”‚   β”œβ”€β”€ agent.py
β”‚   β”‚   └── modeling_llama.py
β”‚   β”œβ”€β”€ scripts
β”‚   β”‚   └── train.sh                          # training MLLM-Tool script
β”‚   β”œβ”€β”€ header.py
β”‚   β”œβ”€β”€ train_sft.py                          # training
β”‚   └── inference.py                          # inference
β”œβ”€β”€ pretrained_checkpoint                     # frozen params of pretrained modules
β”‚   β”œβ”€β”€ imagebind_ckpt
β”‚   β”‚   β”œβ”€β”€huge                               # version
β”‚   β”‚   β”‚   └──imagebind_huge.pth
β”‚   β”œβ”€β”€ LLM_ckpt
β”‚   β”‚   β”œβ”€β”€ vicuna_7b
β”‚   β”‚   β”‚   β”œβ”€β”€ config.json
β”‚   β”‚   β”‚   β”œβ”€β”€ pytorch_model-00001-of-00002.bin
β”‚   β”‚   β”‚   β”œβ”€β”€ tokenizer.model
β”‚   β”‚   β”‚   └── ...
β”‚   β”‚   └── ...
β”œβ”€β”€ LICENCE.md
β”œβ”€β”€ README.md
└── requirements.txt

2. Environment Preparation [Back to Top]

Please first clone the repo and install the required environment, which can be done by running the following commands:

conda create -n MLLM-Tool python=3.8

conda activate MLLM-Tool

# CUDA 11.7
pip install torch==2.0.1+cu117 torchvision==0.15.2+cu117 torchaudio==2.0.2 --index-url https://download.pytorch.org/whl/cu117

git clone https://github.com/MLLM-Tool/MLLM-Tool.git
cd MLLM-Tool

pip install -r requirements.txt 

conda install -c conda-forge cartopy
conda install -c conda-forge pycocotools

3. Training/Adapting Your Own MLLM-Tool

3.1. Preparing Pre-trained Checkpoint [Back to Top]

Please follow the instructions to prepare the ImageBind and Large Language Models(LLM) checkpoints.

Base Language Model Maximum Sequence Length Huggingface Delta Weights Address
Vicuna-7B 512 lmsys/vicuna-7b-v1.5
Vicuna-13B 512 lmsys/vicuna-13b-v1.5
Llama-7B 512 huggyllama/llama-7b
Llama-13B 512 huggyllama/llama-13b
Llama2-7B 512 meta-llama/Llama-2-7b-hf
Llama2-13B 512 meta-llama/Llama-2-13b-hf
Llama2-Chat-7B 512 meta-llama/Llama2-7b-chat-hf
Llama2-Chat-13B 512 meta-llama/Llama-2-13b-chat-hf

3.2. Preparing Dataset [Back to Top]

Please download the following datasets used for model training and testing from here:

After downloading the dataset, please put it under the path [data/IT_data_ins/T+X-T_data/]

3.3. Training MLLM-Tool [Back to Top]

First of all, please refer to the base configuration file [./code/config/base.yaml] for the basic system setting of overall modules.

Then, the training of MLLM-Tool starts with this script(We take the example of using Vicuna-7B as the backbone and use 4 GPUs):

cd ./code
bash scripts/train.sh

Specifying the command:

deepspeed --include localhost:0,1,2,3 --master_addr 127.0.0.1 --master_port 28459 train_sft.py \
    --model openllama_peft \
    --stage 1\
    --imagebind_ckpt_path ../pretrained_checkpoint/imagebind_ckpt/\
    --llm_ckpt_path ../pretrained_checkpoint/LLM_ckpt/vicuna_7b/\
    --max_tgt_len 512\
    --epochs 5\
    --save_path  ../ckpt/mllmtool_vicuna_7b/\
    --log_path ../ckpt/mllmtool_vicuna_7b/log/\
    --version v1

where the key arguments are:

  • --include: localhost:0 indicating the GPT cuda number 0 of deepspeed.
  • --stage: training stage.
  • --imagebind_ckpt_path: the directory which saves the pretrained imagebind weights.
  • --llm_ckpt_path: the directory which saves the pretrained large language model weights. You can replace the Vicuna-7B to any other LLMs.
  • --max_tgt_len: the maximum sequence length.
  • --epochs: the number of training epochs.
  • --save_path: the directory which saves the trained delta weights. This directory will be automatically created.
  • --log_path: the directory which saves the log file.
  • --version: the name of the checkpoint file.

The whole MLLM-Tool training involves:

  • Step-1: Instruction Tuning. This stage instruction-tune 1) the LLM via LoRA, 2) input projection layer on the ToolMMBench dataset.

    Just run the above train.sh script.

    Also refer to the running config file [./code/config/openllama_peft.yaml] and deepspeed config file [./code/dsconfig/openllama_peft_stage_1.yaml] for detailed configurations. Pay attention to the train_batch_size, train_micro_batch_size_per_gpu, gradient_accumulation_steps in the deepspeed config file [./code/dsconfig/openllama_peft_stage_1.yaml], you need to adjust the value to match your computing resources. In this work, we set 64, 4, 4 for 7B models and 32, 1, 8 for 13B models respectively.

Note: We extract the embedding of all the video data and store here, in case there may exists some problems when getting the video embeddings.

4. Evaluating your own MLLM-Tool system

4.1. Preparing Checkpoints[Back to Top]

You can either 1) use the params trained yourselves, or 2) download our checkpoints from here.

The checkpoints should be stored under the path [code/ckpt/]

4.2. Inference [Back to Top]

The inference of MLLM-Tool starts with this script(Again, we take the example of Vicuna-7B as backbone):

python inference.py

The prediction output would save in the format of JSON under the path [data/inference/]. Specifically, the format would be

   "question_id": idx,
   "questions": prompt,
   "response": output,

Contact

For any questions or feedback, feel free to contact Chenyu Wang.

Citation

If you find MLLM-Tool useful in your research or applications, please kindly cite:

@article{wang2024mllmtool,
  title={MLLM-Tool: A Multimodal Large Language Model For Tool Agent Learning},
  author={Chenyu Wang and Weixin Luo and Qianyu Chen and Haonan Mai and  Jindi Guo and Sixun Dong and Xiaohua (Michael) Xuan and Zhengxin  Li and Lin Ma and Shenghua Gao},
  journal={arXiv preprint arXiv:2401.10727},
  year={2024}
}

Acknowledgements

You may refer to related work that serves as foundations for our framework and code repository, Vicuna, Llama, Llama2, Llama2-Chat, ImageBind, We also partially draw inspiration from PandaGPT, MiniGPT-4, LLaVA, NeXT-GPT. Thanks for their wonderful work.

License Notices

This repository is under MIT License. MLLM-Tool is a research project intended for non-commercial use only. One must NOT use the code of MLLM-Tool for any illegal, harmful, violent, racist, or sexual purposes. One is strictly prohibited from engaging in any activity that will potentially violate these guidelines. Any potential commercial use of this code should be approved by the authors.