You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: docs/DOCUMENTATION.md
+9-23
Original file line number
Diff line number
Diff line change
@@ -31,12 +31,12 @@
31
31
-[How do I run this on my SLURM cluster?](#how-do-i-run-this-on-my-slurm-cluster)
32
32
-[How can I run this on my AWS/GCP/Azure cloud project?](#how-can-i-run-this-on-my-awsgcpazure-cloud-project)
33
33
-[Submitting](#submitting)
34
+
-[How do I submit my algorithm to the benchmark?](#how-do-i-submit-my-algorithm-to-the-benchmark)
34
35
-[Can I submit multiple times to the benchmark competition?](#can-i-submit-multiple-times-to-the-benchmark-competition)
35
36
-[Can my submission be structured using multiple files?](#can-my-submission-be-structured-using-multiple-files)
36
37
-[Can I install custom dependencies?](#can-i-install-custom-dependencies)
37
38
-[How can I know if my code can be run on benchmarking hardware?](#how-can-i-know-if-my-code-can-be-run-on-benchmarking-hardware)
38
-
-[Are we allowed to use our own hardware to self-report the results?](#are-we-allowed-to-use-our-own-hardware-to-self-report-the-results)
39
-
-[What can I do if running the benchmark is too expensive for me?](#what-can-i-do-if-running-the-benchmark-is-too-expensive-for-me)
39
+
-[This benchmark seems computationally expensive. Do I have to run it myself?](#this-benchmark-seems-computationally-expensive-do-i-have-to-run-it-myself)
40
40
-[Can I submit previously published training algorithms as submissions?](#can-i-submit-previously-published-training-algorithms-as-submissions)
41
41
-[Disclaimers](#disclaimers)
42
42
-[Shared Data Pipelines between JAX and PyTorch](#shared-data-pipelines-between-jax-and-pytorch)
@@ -383,14 +383,6 @@ Valid submissions must rely on new algorithmic or mathematical ideas and should
383
383
384
384
</details>
385
385
386
-
##### Submissions vs. Baselines
387
-
388
-
Submitters may also submit algorithms marked as*baselines*. These baseline algorithms are not eligible for winning the competition or prize money but they are also not required to be "substantially different"from other submissions by the same submitters. Baseline algorithms will still appear on the leaderboard but will be clearly marked as such. We highly encourage the submission of baselines for educational purposes.
389
-
390
-
Baseline algorithms might, for example, include existing algorithms with different search spaces or learning rate schedules.
391
-
Another example involves porting submissions to different frameworks. For instance, a participant may wish to assess their algorithm in both JAXand PyTorch to demonstrate the impact of the framework. However, in such cases, one of these submissions must be designated as eligible for prize consideration, while the other is marked as a baseline. This prevents circumventing of tuning rules and the spirit of the benchmark by creating additional "lottery tickets".
392
-
Baselines might not be prioritized when using the compute resources by the sponsors of the benchmark.
393
-
394
386
##### Software dependencies
395
387
396
388
We require submissions to use specific versions of `PyTorch`/`JAX`as well as additional dependencies in order to facilitate fair comparisons. Submitters must build on top of these provided software packages, which might be provided as a `Docker` container. Additional dependencies can be added aslongas they include a comment describing what was added and why. Submitters are free to add dependencies that support new algorithmic and mathematical ideas but they should not circumvent the intention of the benchmark to measure training speedups due to new training methods. For example, software engineering techniques that lead to faster implementations of existing software, e.g. using newer versions of `PyTorch`or`JAX`, are not allowed and these are described in more detail in the [Disallowed submissions](#disallowed-submissions) section.
@@ -590,11 +582,13 @@ new Compute Instance with the "Deep Learning on Linux" Image in Boot disk option
590
582
591
583
### Submitting
592
584
593
-
#### Can I submit multiple times to the benchmark competition?
585
+
#### How do I submit my algorithm to the benchmark?
586
+
587
+
Please see our [How to Submit](/README.md#how-to-submit) section. You can submit your algorithm to the benchmark by opening a PR on the [submission repository](https://github.com/mlcommons/submissions_algorithms).
594
588
595
-
Our benchmark allows multiple submissions by the same team of submitters aslongas they are substantially different. We disallow submitters from circumventing the purpose of the benchmark by, for example, submitting dozens of copies of the same submission with slightly different hyperparameters. Such a bulk submission would result in an unfair advantage on the randomized workloads andisnotin the spirit of the benchmark.
589
+
#### Can I submit multiple times to the benchmark competition?
596
590
597
-
Submitters may submit algorithms marked as*baselines*. These might include existing algorithms with different search spaces or learning rate schedules. These baseline algorithms are not eligible for winning the competition or prize money but they are also notrequired to be "substantially different"from other submissions by the same submitters. See the [Submissions vs. Baselines](#submissions-vs-baselines) Section.
591
+
Our benchmark allows multiple submissions by the same team of submitters aslongas they are substantially different. We discourage submitters from creating bulk submissions as this isnotinthe spirit of the benchmark.
598
592
599
593
#### Can my submission be structured using multiple files?
600
594
@@ -610,17 +604,9 @@ To include your custom dependencies in your submission, please include them in a
610
604
The benchmarking hardware specifications are documented in the [Benchmarking Hardware Section](#benchmarking-hardware). We recommend monitoring your submission's memory usage so that it does not exceed the available memory
611
605
on the benchmarking hardware. We also recommend to do a dry run using a cloud instance.
612
606
613
-
#### Are we allowed to use our own hardware to self-report the results?
614
-
615
-
NOTE: Submitters are no longer required to self-report results for AlgoPerf competition v0.5.
616
-
617
-
You only have to use the benchmarking hardware for runs that are directly involved in the scoring procedure. This includes all runs for the self-tuning ruleset, but only the runs of the best hyperparameter configuration in each study for the external tuning ruleset. For example, you could use your own (different) hardware to tune your submission and identify the best hyperparameter configuration (in each study) and then only run this configuration (i.e. 5 runs, one for each study) on the benchmarking hardware.
618
-
619
-
#### What can I do if running the benchmark is too expensive for me?
620
-
621
-
NOTE: Submitters are no longer required to self-report results for AlgoPerf competition v0.5.
607
+
#### This benchmark seems computationally expensive. Do I have to run it myself?
622
608
623
-
Submitters unable to self-fund scoring costs can instead self-report only on the [qualification set of workloads](/COMPETITION_RULES.md#qualification-set) that excludes some of the most expensive workloads. Based on this performance on the qualification set, the working group will provide - as funding allows - compute to evaluate and score the most promising submissions. Additionally, we encourage researchers to reach out to the [working group](mailto:[email protected]) to find potential collaborators with the resources to run larger, more comprehensive experiments for both developing and scoring submissions.
609
+
Submitters are no longer required to self-report results to get on the AlgoPerf leaderboard. Instead, they can open a PRandthe working group will score the most promising submissions, see our [How to Submit](/README.md#how-to-submit) section for more details. You can use self-reported results to provide evidence of performance on the benchmark. Even if you fully self-report, we will still verify the scores by rerunning the submission on our setup.
624
610
625
611
#### Can I submit previously published training algorithms as submissions?
0 commit comments