/SkySenseGPT

A Fine-Grained Instruction Tuning Dataset and Model for Remote Sensing Vision-Language Understanding

Primary LanguagePythonApache License 2.0Apache-2.0

License Paper Paper GitHub stars

SkySenseGPT: A Fine-Grained Instruction Tuning Dataset and Model for Remote Sensing Vision-Language Understanding

📢 News and Updates

🔥🔥🔥 Last Updated on 2024.07.22🔥🔥🔥

  • 2024.07.22: The FIT-RSFG benchmark have been uploaded here and the evaluation scripts have been released here! See Evaluation for details on how to evaluate.
  • 2024.07.20: The FIT-RS dataset (training set 1415k) categorized by tasks have been uploaded here.
  • 2024.07.01: The FIT-RS dataset (training set 1415k) have been uploaded here.
  • 2024.06.17: Our paper is available in arxiv!
  • 2024.06.07: First version.

📌 Introduction

[Paper][Dataset][Model][Code]

In this project, we propose the FIT-RS (Remote Sensing Fine-Grained Instruction Tuning) dataset, which contains 1,800,851 high-quality instruction samples covering various vision-language comprehension tasks. FIT-RS aims to enhance the fine-grained comprehension ability of Remote Sensing Large Multi-Modal Models (RSLMMs), specifically their ability to understand semantic relationships among objects in complex remote sensing scenes. Based on FIT-RS, we establish the FIT-RSFG (Remote Sensing Fine-Grained Comprehension) Benchmark to evaluate RSLMMs' ability in fine-grained understanding.

In addition, we constructed the FIT-RSRC (Remote Sensing Relation Comprehension) Benchmark, which adopts the common-used single-choice format and CircularEval strategy. It includes high-quality distractor options derived from commonsense word lists, as well as unanswerable questions, aiming to evaluate the Remote Sensing Relation Comprehension capabilities of LMMs.

🛠️ Table of Contents

⭐️ Dataset and Download

  • FIT-RS
  • FIT-RS is a large-scale fine-grained instruction tuning dataset, which contains 1,800,851 high quality instruction samples, aiming at enhancing the fine-grained comprehension ability of RSLMMs.

    Introduction

  • FIT-RSRC

    Given the current lack of a publicly available benchmark for comprehensive and quantitative evaluation of existing LMMs in remote sensing relation understanding, we propose the FIT-RSRC (Remote Sensing Relation Comprehension) benchmark. It is designed in the form of single-choice questions, containing four different types of questions and high-quality distractor options. Following the mainstream general benchmark, FIT-RSRC employs CircularEval as the evaluation strategy.

    Introduction

  • Download Links
    • FIT-RS: A fine-grained remote sensing instruction tuning dataset, containing 1800k instruction samples, 1415k for training.
    • FIT-RSFG: A fine-grained benchmark for remote sensing vision-language evaluation.
    • FIT-RSRC: A single-choice benchmark for remote sensing relation comprehension evaluation.
    • SkySenseGPT: A remote sensing large multi-modal model, capable of handling complex comprehension tasks like image-level scene graph generation.

Evaluation

  1. Download FIT-RSFG and FIT-RSRC Benchmarks.
  2. Install necessary packages as in the requirements.txt.
  3. See evaluation.sh for evaluation.

License

This project is released under the Apache 2.0 license.

Citation

Our FIT-RS dataset is built based on the STAR dataset. If you find this work helpful for your research, please consider giving this repo a star ⭐ and citing our paper:

@article{luo2024sky,
  title={SkySenseGPT: A Fine-Grained Instruction Tuning Dataset and Model for Remote Sensing Vision-Language Understanding},
  author={Luo, Junwei and Pang, Zhen and Zhang, Yongjun and Wang, Tingzhu and Wang, Linlin and Dang, Bo and Lao, Jiangwei and Wang, Jian and Chen, Jingdong and Tan, Yihua and Li, Yansheng},
  journal={arXiv preprint arXiv:2406.10100},
  year={2024}
}

@article{li2024scene,
  title={STAR: A First-Ever Dataset and A Large-Scale Benchmark for Scene Graph Generation in Large-Size Satellite Imagery},
  author={Li, Yansheng and Wang, Linlin and Wang, Tingzhu and Yang, Xue and Luo, Junwei and Wang, Qi and Deng, Youming and Wang, Wenbin and Sun, Xian and Li, Haifeng and Dang, Bo and Zhang, Yongjun and Yu, Yi and Yan Junchi},
  journal={arXiv preprint arXiv:2406.09410},
  year={2024}
}

We are thankful to LLaVA-1.5 and GeoChat for releasing their models and code as open-source contributions.