We-Math
Leaderboard
We-Math

a benchmark that evaluates large multimodal models (LMMs) on their ability to perform human-like mathematical reasoning.

a benchmark that evaluates large multimodal models (LMMs) on their ability to perform human-like mathematical reasoning.

Relevant Sites

Leave a Reply

Your email address will not be published. Required fields are marked *