Pinned Loading
-
OpenGVLab/ChartAst
OpenGVLab/ChartAst Public[ACL 2024] ChartAssistant is a chart-based vision-language model for universal chart comprehension and reasoning.
-
OpenGVLab/MMIU
OpenGVLab/MMIU PublicMMIU: Multimodal Multi-image Understanding for Evaluating Large Vision-Language Models
-
OpenGVLab/Multitask-Model-Selector
OpenGVLab/Multitask-Model-Selector Public[NIPS2023]Implementation of Foundation Model is Efficient Multimodal Multitask Model Selector
-
OpenGVLab/Multi-Modality-Arena
OpenGVLab/Multi-Modality-Arena PublicChatbot Arena meets multi-modality! Multi-Modality Arena allows you to benchmark vision-language models side-by-side while providing images as inputs. Supports MiniGPT-4, LLaMA-Adapter V2, LLaVA, B…
-
OpenGVLab/PhyGenBench
OpenGVLab/PhyGenBench PublicThe code and data of Paper: Towards World Simulator: Crafting Physical Commonsense-Based Benchmark for Video Generation
-
MMT-Bench
MMT-Bench PublicForked from OpenGVLab/MMT-Bench
[ICML2024] MMT-Bench: A Comprehensive Multimodal Benchmark for Evaluating Large Vision-Language Models Towards Multitask AGI
Python
If the problem persists, check the GitHub status page or contact support.