compiled a list of 2500+ vision benchmarks for VLMs

compiled a list of 2500+ vision benchmarks for VLMs

I love reading benchmark / eval papers. It's one of the best way to stay up-to-date with progress in Vision Language Models, and understand where they fall short.

Vision tasks vary quite a lot from one to another. For example:

  • vision tasks that require high-level semantic understanding of the image. Models do quite well in them. Popular general benchmarks like MMMU are good for that.
  • visual reasoning tasks where VLMs are given a visual puzzle (think IQ-style test). VLMs perform quite poorly on them. Barely above a random guess. Benchmarks such as VisuLogic are designed for this.
  • visual counting tasks. Models only get it right about 20% of the times. But they’re getting better. Evals such as UNICBench test 21+ VLMs across counting tasks with varying levels of difficulty.

Compiled a list of 2.5k+ vision benchmarks with data links and high-level summary that auto-updates every day with new benchmarks.

I'm thinking of maybe adding a simple website to semantically search through them. Will do if someone asks

submitted by /u/batatibatata
[link] [comments]

Leave a Comment

Your email address will not be published. Required fields are marked *

Scroll to Top