[P] Sotabench: Benchmarking Every Open Source Model
We (the team behind Papers With Code) have been working on a new free service: https://sotabench.com – to benchmark models directly from GitHub.
The mission of sotabench is to benchmark and map out every open source ML model. On Papers With Code we collect code implementations and results from papers, but until now we had no way of knowing if the code is sufficient to produce the claimed results. This is why we created sotabench.
We hope this is going to help practitioners and researchers alike to find high quality model implementations and gain insight into the accuracy / speed tradeoff of models out there.
We’ve implemented a couple of benchmarks for now – but we’ve made it open and free for anyone to add benchmarks and connect their repositories (including forks). If you find it useful, feel free to just go for it and add repos.
Open for suggestions and feedback!