|
1 | 1 | name: SYCL E2E |
2 | 2 |
|
3 | 3 | on: |
4 | | - # workflow_call: |
5 | | - # inputs: |
6 | | - # name: |
7 | | - # type: string |
8 | | - # required: True |
| 4 | + workflow_call: |
| 5 | + inputs: |
| 6 | + name: |
| 7 | + type: string |
| 8 | + required: True |
9 | 9 |
|
10 | | - # runner: |
11 | | - # type: string |
12 | | - # required: True |
13 | | - # image: |
14 | | - # type: string |
15 | | - # required: False |
16 | | - # image_options: |
17 | | - # type: string |
18 | | - # required: True |
| 10 | + runner: |
| 11 | + type: string |
| 12 | + required: True |
| 13 | + image: |
| 14 | + type: string |
| 15 | + required: False |
| 16 | + image_options: |
| 17 | + type: string |
| 18 | + required: True |
19 | 19 |
|
20 | | - # target_devices: |
21 | | - # type: string |
22 | | - # required: False |
23 | | - # extra_cmake_args: |
24 | | - # type: string |
25 | | - # required: False |
26 | | - # tests_selector: |
27 | | - # description: | |
28 | | - # Three possible options: "e2e", "cts", and "benchmarks". |
29 | | - # type: string |
30 | | - # default: "e2e" |
| 20 | + target_devices: |
| 21 | + type: string |
| 22 | + required: False |
| 23 | + extra_cmake_args: |
| 24 | + type: string |
| 25 | + required: False |
| 26 | + tests_selector: |
| 27 | + description: | |
| 28 | + Three possible options: "e2e", "cts", and "benchmarks". |
| 29 | + type: string |
| 30 | + default: "e2e" |
31 | 31 |
|
32 | | - # extra_lit_opts: |
33 | | - # description: | |
34 | | - # Extra options to be added to LIT_OPTS. |
35 | | - # type: string |
36 | | - # default: '' |
| 32 | + extra_lit_opts: |
| 33 | + description: | |
| 34 | + Extra options to be added to LIT_OPTS. |
| 35 | + type: string |
| 36 | + default: '' |
37 | 37 |
|
38 | | - # repo_ref: |
39 | | - # type: string |
40 | | - # required: False |
41 | | - # description: | |
42 | | - # Commit SHA or branch to checkout the intel/llvm repo. |
43 | | - # tests_ref: |
44 | | - # type: string |
45 | | - # required: False |
46 | | - # description: Commit SHA or branch to checkout e2e/cts tests. |
| 38 | + repo_ref: |
| 39 | + type: string |
| 40 | + required: False |
| 41 | + description: | |
| 42 | + Commit SHA or branch to checkout the intel/llvm repo. |
| 43 | + tests_ref: |
| 44 | + type: string |
| 45 | + required: False |
| 46 | + description: Commit SHA or branch to checkout e2e/cts tests. |
47 | 47 |
|
48 | | - # toolchain_artifact: |
49 | | - # type: string |
50 | | - # default: '' |
51 | | - # required: False |
52 | | - # toolchain_artifact_filename: |
53 | | - # type: string |
54 | | - # default: '' |
55 | | - # required: False |
56 | | - # toolchain_decompress_command: |
57 | | - # type: string |
58 | | - # default: '' |
59 | | - # required: False |
| 48 | + toolchain_artifact: |
| 49 | + type: string |
| 50 | + default: '' |
| 51 | + required: False |
| 52 | + toolchain_artifact_filename: |
| 53 | + type: string |
| 54 | + default: '' |
| 55 | + required: False |
| 56 | + toolchain_decompress_command: |
| 57 | + type: string |
| 58 | + default: '' |
| 59 | + required: False |
60 | 60 |
|
61 | | - # binaries_artifact: |
62 | | - # description: | |
63 | | - # When set in modes other than `run-only` results in artifact upload. |
64 | | - # For `run-only` mode, if specified, means downloading pre-built |
65 | | - # binaries from the artifact. If set to special value `in-container` |
66 | | - # then the binaries are expected to be stored in the container image |
67 | | - # instead of coming from an artifact. |
68 | | - # type: string |
69 | | - # default: '' |
70 | | - # required: False |
71 | | - # testing_mode: |
72 | | - # description: | |
73 | | - # Testing mode to run E2E/CTS tests in, can be either `full`, `build-only` |
74 | | - # or `run-only`. |
75 | | - # type: string |
76 | | - # default: 'full' |
77 | | - # retention-days: |
78 | | - # description: 'E2E/SYCL-CTS binaries artifact retention period.' |
79 | | - # type: string |
80 | | - # default: 1 |
| 61 | + binaries_artifact: |
| 62 | + description: | |
| 63 | + When set in modes other than `run-only` results in artifact upload. |
| 64 | + For `run-only` mode, if specified, means downloading pre-built |
| 65 | + binaries from the artifact. If set to special value `in-container` |
| 66 | + then the binaries are expected to be stored in the container image |
| 67 | + instead of coming from an artifact. |
| 68 | + type: string |
| 69 | + default: '' |
| 70 | + required: False |
| 71 | + testing_mode: |
| 72 | + description: | |
| 73 | + Testing mode to run E2E/CTS tests in, can be either `full`, `build-only` |
| 74 | + or `run-only`. |
| 75 | + type: string |
| 76 | + default: 'full' |
| 77 | + retention-days: |
| 78 | + description: 'E2E/SYCL-CTS binaries artifact retention period.' |
| 79 | + type: string |
| 80 | + default: 1 |
81 | 81 |
|
82 | | - # install_igc_driver: |
83 | | - # type: string |
84 | | - # required: False |
85 | | - # install_dev_igc_driver: |
86 | | - # type: string |
87 | | - # required: False |
88 | | - # env: |
89 | | - # type: string |
90 | | - # default: '{}' |
91 | | - # required: False |
| 82 | + install_igc_driver: |
| 83 | + type: string |
| 84 | + required: False |
| 85 | + install_dev_igc_driver: |
| 86 | + type: string |
| 87 | + required: False |
| 88 | + env: |
| 89 | + type: string |
| 90 | + default: '{}' |
| 91 | + required: False |
92 | 92 |
|
93 | | - # skip_run: |
94 | | - # type: string |
95 | | - # default: 'false' |
96 | | - # required: False |
| 93 | + skip_run: |
| 94 | + type: string |
| 95 | + default: 'false' |
| 96 | + required: False |
97 | 97 |
|
98 | | - # benchmark_upload_results: |
99 | | - # description: | |
100 | | - # Set to true to upload results to git repository storing benchmarking |
101 | | - # results. |
102 | | - # type: string |
103 | | - # default: 'false' |
104 | | - # required: False |
105 | | - # benchmark_save_name: |
106 | | - # description: | |
107 | | - # Save name to use for benchmark results: Save names are stored in |
108 | | - # metadata of result file, and are used to identify benchmark results in |
109 | | - # the same series (e.g. same configuration, same device, etc.). |
| 98 | + benchmark_upload_results: |
| 99 | + description: | |
| 100 | + Set to true to upload results to git repository storing benchmarking |
| 101 | + results. |
| 102 | + type: string |
| 103 | + default: 'false' |
| 104 | + required: False |
| 105 | + benchmark_save_name: |
| 106 | + description: | |
| 107 | + Save name to use for benchmark results: Save names are stored in |
| 108 | + metadata of result file, and are used to identify benchmark results in |
| 109 | + the same series (e.g. same configuration, same device, etc.). |
110 | 110 |
|
111 | | - # Note: Currently, benchmark result filenames are in the format of |
112 | | - # <benchmark_save_name>_<Device>_<Backend>_YYYYMMDD_HHMMSS.json |
113 | | - # type: string |
114 | | - # default: '' |
115 | | - # required: False |
116 | | - # benchmark_preset: |
117 | | - # description: | |
118 | | - # Name of benchmark preset to run. |
| 111 | + Note: Currently, benchmark result filenames are in the format of |
| 112 | + <benchmark_save_name>_<Device>_<Backend>_YYYYMMDD_HHMMSS.json |
| 113 | + type: string |
| 114 | + default: '' |
| 115 | + required: False |
| 116 | + benchmark_preset: |
| 117 | + description: | |
| 118 | + Name of benchmark preset to run. |
119 | 119 |
|
120 | | - # See /devops/scripts/benchmarks/presets.py for all presets available. |
121 | | - # type: string |
122 | | - # default: 'Minimal' |
123 | | - # required: False |
124 | | - # # dry-run is passed only to compare.py (to not fail on regression), not to main.py (where such flag would omit all benchmark runs) |
125 | | - # benchmark_dry_run: |
126 | | - # description: | |
127 | | - # Whether or not to fail the workflow upon a regression. |
128 | | - # type: string |
129 | | - # default: 'false' |
130 | | - # required: False |
131 | | - # benchmark_exit_on_failure: |
132 | | - # description: | |
133 | | - # Whether or not to fail the workflow upon a failure. |
134 | | - # type: string |
135 | | - # default: 'false' |
136 | | - # required: False |
| 120 | + See /devops/scripts/benchmarks/presets.py for all presets available. |
| 121 | + type: string |
| 122 | + default: 'Minimal' |
| 123 | + required: False |
| 124 | + # dry-run is passed only to compare.py (to not fail on regression), not to main.py (where such flag would omit all benchmark runs) |
| 125 | + benchmark_dry_run: |
| 126 | + description: | |
| 127 | + Whether or not to fail the workflow upon a regression. |
| 128 | + type: string |
| 129 | + default: 'false' |
| 130 | + required: False |
| 131 | + benchmark_exit_on_failure: |
| 132 | + description: | |
| 133 | + Whether or not to fail the workflow upon a failure. |
| 134 | + type: string |
| 135 | + default: 'false' |
| 136 | + required: False |
137 | 137 |
|
138 | | - # in_workflow_call_mode: |
139 | | - # description: | |
140 | | - # Unlike `github.event_type` that captures the original trigger, we use |
141 | | - # this to determine how this particular workflow has been called. This |
142 | | - # isn't supposed to be specified by a caller and only relies on the |
143 | | - # default value. |
144 | | - # default: 'true' |
145 | | - # required: False |
146 | | - # type: string |
| 138 | + in_workflow_call_mode: |
| 139 | + description: | |
| 140 | + Unlike `github.event_type` that captures the original trigger, we use |
| 141 | + this to determine how this particular workflow has been called. This |
| 142 | + isn't supposed to be specified by a caller and only relies on the |
| 143 | + default value. |
| 144 | + default: 'true' |
| 145 | + required: False |
| 146 | + type: string |
147 | 147 |
|
148 | 148 | workflow_dispatch: |
149 | 149 | inputs: |
|
0 commit comments