Announcement

Welcome to Codabench! New here? Check out our getting started guide to begin your journey with ease!

Join the Google group to connect with the community!

We provide a default CPU queue for small jobs (typically for result submission competitions). For code submission, organizers are expected to provide their own queue, see instructions. The execution time is currently capped at 20 minutes on the default public queue.

Get Started

Participate

Find benchmarks that pique your interest! A benchmark allows you to test new algorithms against reference datasets OR (inverted benchmark) submit challenging data to reference algorithms.

Organize

Organize a benchmark on Codabench. Start with our tutorial.

Contribute

Interested in joining the development team? Join us on Github or contact us directly.

204
Public Competitions
12216
Users
70276
Submissions
About Codabench

What is Codabench?

Codabench is an open source platform allowing you to organize AI benchmarks. It is flexible and powerful, yet easy to use. You define tasks (e.g. datasets and metrics of success), then interface for submissions of code (algorithms), add some documentation pages, make an upload and that's it! Your benchmark is created, ready to accept submissions of new algorithms. Everything can be fully customized, including the code of the scoring program. Organizers can even hook up their own compute workers to their benchmarks, enabling unlimited computing power. Participants can try out their methods, get real-time feedback and results on a competitive leaderboard, detailed plots and more.

In a unique twist, Codabench also allows you to create inverted benchmarks. Here, the roles of datasets and algorithms are interchanged. In this scenario, you set the reference algorithms and the participants contribute datasets.

What is CodaLab?

CodaLab Competitions is a powerful open source framework for running competitions using result or code submissions. You can participate in an existing competition or host your own competition for free.

Most competitions hosted on CodaLab are machine learning (data science) competitions, but it is NOT limited to this application domain. It can accommodate any problem for which a solution can be provided in the form of a zip archive containing a number of files to be evaluated quantitatively by a scoring program (provided by the organizers). The scoring program must return a numeric score, which is displayed on a leaderboard where the performances of participants are compared.

History of CodaLab

CodaLab was created in 2013 as a joint venture between Microsoft and Stanford University. Originally the vision was to create an ecosystem for conducting computational research in a more efficient, reproducible, and collaborative manner, combining worksheets and competitions. Worksheets capture complex research pipelines in a reproducible way and create "executable papers". Codabench is the continuity of CodaLab, a version 2 in which users can organize benchmarks.

Some competitions have been organized using worksheets, but the competition platform and the worksheet platform have both a large user base and can be used independently. In 2014, ChaLearn joined to co-develop CodaLab competitions. Since 2015, University Paris-Saclay is community lead of CodaLab competitions, under the direction of Isabelle Guyon, professor of big data. CodaLab and Codabench are administered by the LISN staff.

CodaLab in Research

CodaLab is used actively in research. In 2019/2020, 400 new challenges were launched. Recent popular challenges organized with CodaLab include the COVID-19 retweet prediction challenge,  the ECCV 2020 ChaLearn LAP Fair face recognition challenge, the 2020 DriveML Huawei Autonomous Vehicle Challenge, and high profile challenges include the 2 million Euro prize of the EU, organized by the See.4C consortium, the CIKM AnalytiCup 2017, which attracted 493 participants, MSCOCO (633 participants) and the ChaLearn AutoML challenge 2017 (687 participants).

Since 2016, CodaLab offers the possibility of organizing machine learning challenges with code submission. The simplest machine learning challenges require only the submission of results, which are compared to a solution (or key) by a scoring program. Result submission challenges are less computationally expensive than code submission challenges. However, they offer less possibilities. In particular, code submission allows conducting fair benchmarks by executing submitted code in the same condition for all participants.

CodaLab has been providing free resources for challenge organizers who want to run high impact events, within a pre-approved agreed upon budget. New since version 1.5: organizers can hook up their own compute workers to the backend of CodaLab to redirect the code submissions, enabling growth to big data competitions running at the expense of the organizers. For very special dedicated projects, CodaLab can be customized since it is an open source project.

News

Dec 05, 2024
We presented CodaLab and Codabench, in French, at JCAD 2024, highlighting the project evolution, main features, the decentralization of computation resources made possible by the compute workers system, and our vision for the future of the platforms.
Mar 12, 2024
mlcontests.com summarized the state of the competitive landscape and analyzed the 300+ competitions that took place in 2023. They compared winning solutions to figure out the best strategies to win at competitive ML. They also compared platforms: the launch of Codabench was mentioned, and CodaLab rose as one of the most influential platforms with the highest number of challenges organized, many of which were competitions affiliated with academic conferences.
Feb 07, 2024
The paper describing CodaLab Competitions was cited more than 100 times, according to Google Scholar!
Jan 20, 2024
Codabench exceeds 2,500 users and 60 public competitions!
Aug 17, 2023
We're thrilled to announce that our pioneer challenge organizers have launched the very first competitions on Codabench! Dive into the AutoML Cup, a contest where participants craft state-of-the-art automated ML algorithms to tackle tasks across different dimensionalities. Additionally, the Auto-Survey Challenge invites language model enthusiasts to design models capable of both writing scientific surveys and reviewing them. Join us in this innovative journey on Codabench!
Jun 23, 2023
After months of development efforts from many contributors, Codabench is finally ready to be used for new benchmarks and competitions. Let's go!
Aug 15, 2020
CodaLab Competitions exceeds 50,000 users, 1000 competitions (over 400 in the last year), and ~600 submissions per day! Click here for more statistics.
Cite Codabench in your research
@article{codabench,
    title = {Codabench: Flexible, easy-to-use, and reproducible meta-benchmark platform},
    author = {Zhen Xu and Sergio Escalera and Adrien Pavão and Magali Richard and 
                Wei-Wei Tu and Quanming Yao and Huan Zhao and Isabelle Guyon},
    journal = {Patterns},
    volume = {3},
    number = {7},
    pages = {100543},
    year = {2022},
    issn = {2666-3899},
    doi = {https://doi.org/10.1016/j.patter.2022.100543},
    url = {https://www.sciencedirect.com/science/article/pii/S2666389922001465}
}