Skip to content

Commit 1ee83e2

Browse files
committed
Update for rolling leaderboard
1 parent d86273c commit 1ee83e2

File tree

1 file changed

+9
-23
lines changed

1 file changed

+9
-23
lines changed

docs/DOCUMENTATION.md

+9-23
Original file line numberDiff line numberDiff line change
@@ -31,12 +31,12 @@
3131
- [How do I run this on my SLURM cluster?](#how-do-i-run-this-on-my-slurm-cluster)
3232
- [How can I run this on my AWS/GCP/Azure cloud project?](#how-can-i-run-this-on-my-awsgcpazure-cloud-project)
3333
- [Submitting](#submitting)
34+
- [How do I submit my algorithm to the benchmark?](#how-do-i-submit-my-algorithm-to-the-benchmark)
3435
- [Can I submit multiple times to the benchmark competition?](#can-i-submit-multiple-times-to-the-benchmark-competition)
3536
- [Can my submission be structured using multiple files?](#can-my-submission-be-structured-using-multiple-files)
3637
- [Can I install custom dependencies?](#can-i-install-custom-dependencies)
3738
- [How can I know if my code can be run on benchmarking hardware?](#how-can-i-know-if-my-code-can-be-run-on-benchmarking-hardware)
38-
- [Are we allowed to use our own hardware to self-report the results?](#are-we-allowed-to-use-our-own-hardware-to-self-report-the-results)
39-
- [What can I do if running the benchmark is too expensive for me?](#what-can-i-do-if-running-the-benchmark-is-too-expensive-for-me)
39+
- [This benchmark seems computationally expensive. Do I have to run it myself?](#this-benchmark-seems-computationally-expensive-do-i-have-to-run-it-myself)
4040
- [Can I submit previously published training algorithms as submissions?](#can-i-submit-previously-published-training-algorithms-as-submissions)
4141
- [Disclaimers](#disclaimers)
4242
- [Shared Data Pipelines between JAX and PyTorch](#shared-data-pipelines-between-jax-and-pytorch)
@@ -383,14 +383,6 @@ Valid submissions must rely on new algorithmic or mathematical ideas and should
383383

384384
</details>
385385

386-
##### Submissions vs. Baselines
387-
388-
Submitters may also submit algorithms marked as *baselines*. These baseline algorithms are not eligible for winning the competition or prize money but they are also not required to be "substantially different" from other submissions by the same submitters. Baseline algorithms will still appear on the leaderboard but will be clearly marked as such. We highly encourage the submission of baselines for educational purposes.
389-
390-
Baseline algorithms might, for example, include existing algorithms with different search spaces or learning rate schedules.
391-
Another example involves porting submissions to different frameworks. For instance, a participant may wish to assess their algorithm in both JAX and PyTorch to demonstrate the impact of the framework. However, in such cases, one of these submissions must be designated as eligible for prize consideration, while the other is marked as a baseline. This prevents circumventing of tuning rules and the spirit of the benchmark by creating additional "lottery tickets".
392-
Baselines might not be prioritized when using the compute resources by the sponsors of the benchmark.
393-
394386
##### Software dependencies
395387

396388
We require submissions to use specific versions of `PyTorch`/`JAX` as well as additional dependencies in order to facilitate fair comparisons. Submitters must build on top of these provided software packages, which might be provided as a `Docker` container. Additional dependencies can be added as long as they include a comment describing what was added and why. Submitters are free to add dependencies that support new algorithmic and mathematical ideas but they should not circumvent the intention of the benchmark to measure training speedups due to new training methods. For example, software engineering techniques that lead to faster implementations of existing software, e.g. using newer versions of `PyTorch` or `JAX`, are not allowed and these are described in more detail in the [Disallowed submissions](#disallowed-submissions) section.
@@ -590,11 +582,13 @@ new Compute Instance with the "Deep Learning on Linux" Image in Boot disk option
590582

591583
### Submitting
592584

593-
#### Can I submit multiple times to the benchmark competition?
585+
#### How do I submit my algorithm to the benchmark?
586+
587+
Please see our [How to Submit](/README.md#how-to-submit) section. You can submit your algorithm to the benchmark by opening a PR on the [submission repository](https://github.com/mlcommons/submissions_algorithms).
594588

595-
Our benchmark allows multiple submissions by the same team of submitters as long as they are substantially different. We disallow submitters from circumventing the purpose of the benchmark by, for example, submitting dozens of copies of the same submission with slightly different hyperparameters. Such a bulk submission would result in an unfair advantage on the randomized workloads and is not in the spirit of the benchmark.
589+
#### Can I submit multiple times to the benchmark competition?
596590

597-
Submitters may submit algorithms marked as *baselines*. These might include existing algorithms with different search spaces or learning rate schedules. These baseline algorithms are not eligible for winning the competition or prize money but they are also not required to be "substantially different" from other submissions by the same submitters. See the [Submissions vs. Baselines](#submissions-vs-baselines) Section.
591+
Our benchmark allows multiple submissions by the same team of submitters as long as they are substantially different. We discourage submitters from creating bulk submissions as this is not in the spirit of the benchmark.
598592

599593
#### Can my submission be structured using multiple files?
600594

@@ -610,17 +604,9 @@ To include your custom dependencies in your submission, please include them in a
610604
The benchmarking hardware specifications are documented in the [Benchmarking Hardware Section](#benchmarking-hardware). We recommend monitoring your submission's memory usage so that it does not exceed the available memory
611605
on the benchmarking hardware. We also recommend to do a dry run using a cloud instance.
612606

613-
#### Are we allowed to use our own hardware to self-report the results?
614-
615-
NOTE: Submitters are no longer required to self-report results for AlgoPerf competition v0.5.
616-
617-
You only have to use the benchmarking hardware for runs that are directly involved in the scoring procedure. This includes all runs for the self-tuning ruleset, but only the runs of the best hyperparameter configuration in each study for the external tuning ruleset. For example, you could use your own (different) hardware to tune your submission and identify the best hyperparameter configuration (in each study) and then only run this configuration (i.e. 5 runs, one for each study) on the benchmarking hardware.
618-
619-
#### What can I do if running the benchmark is too expensive for me?
620-
621-
NOTE: Submitters are no longer required to self-report results for AlgoPerf competition v0.5.
607+
#### This benchmark seems computationally expensive. Do I have to run it myself?
622608

623-
Submitters unable to self-fund scoring costs can instead self-report only on the [qualification set of workloads](/COMPETITION_RULES.md#qualification-set) that excludes some of the most expensive workloads. Based on this performance on the qualification set, the working group will provide - as funding allows - compute to evaluate and score the most promising submissions. Additionally, we encourage researchers to reach out to the [working group](mailto:[email protected]) to find potential collaborators with the resources to run larger, more comprehensive experiments for both developing and scoring submissions.
609+
Submitters are no longer required to self-report results to get on the AlgoPerf leaderboard. Instead, they can open a PR and the working group will score the most promising submissions, see our [How to Submit](/README.md#how-to-submit) section for more details. You can use self-reported results to provide evidence of performance on the benchmark. Even if you fully self-report, we will still verify the scores by rerunning the submission on our setup.
624610

625611
#### Can I submit previously published training algorithms as submissions?
626612

0 commit comments

Comments
 (0)