MangaVQA and MangaLMM: A Benchmark and Specialized Model for Multimodal Manga Understanding
Paper
β’
2505.20298
β’
Published
β’
9
This repository contains the MangaLMM model described in the paper MangaVQA and MangaLMM: A Benchmark and Specialized Model for Multimodal Manga Understanding.
Code: https://github.com/manga109/MangaLMM
Official demo: https://huggingface.co/spaces/yuki-imajuku/MangaLMM-Demo