Skip to content

Commit 1b25848

Browse files
committed
Resolve comments
1 parent f973a4e commit 1b25848

1 file changed

Lines changed: 12 additions & 11 deletions

File tree

  • sdks/python/apache_beam/testing/benchmarks/inference

sdks/python/apache_beam/testing/benchmarks/inference/README.md

Lines changed: 12 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -129,8 +129,8 @@ Full pipeline implementation is available [here](https://github.com/apache/beam/
129129
2. Create the benchmark implementation
130130

131131
- Location: sdks/python/apache_beam/testing/benchmarks/inference (e.g., pytorch_sentiment_benchmarks.py)
132-
- Inherit from DataflowCostBenchmark class.
133-
- Ensure the expected 'pcollection' parameter is passed to your builder. This parameter could be obtained from GCP Dataflow Jobs -> Your Job Page.
132+
- Inherit from DataflowCostBenchmark class.
133+
- Ensure the 'pcollection' parameter is passed to the `DataflowCostBenchmark` constructor. This is the name of the PCollection for which to measure throughput, and you can find this name in the Dataflow UI job graph.
134134
- Keep naming consistent with other benchmarks.
135135

136136
3. Add an options txt file
@@ -150,7 +150,6 @@ Full pipeline implementation is available [here](https://github.com/apache/beam/
150150
--publish_to_big_query=true
151151
--metrics_dataset=beam_run_inference
152152
--metrics_table=your_table
153-
--input_options={}
154153
--influx_measurement=your-measurement
155154
--device=CPU
156155
--runner=DataflowRunner
@@ -160,7 +159,7 @@ Full pipeline implementation is available [here](https://github.com/apache/beam/
160159

161160
- Workflow: .github/workflows/beam_Inference_Python_Benchmarks_Dataflow.yml
162161
- Add your argument-file-path to the matrix.
163-
- Add a step that runs your <pipeline_name>_benchmarks.py with -PloadTest.args=$YOUR_ARGUMENTS.
162+
- Add a step that runs your <pipeline_name>_benchmarks.py with -PloadTest.args=$YOUR_ARGUMENTS. Which are the arguments created in previous step.
164163

165164
5. Test on your fork
166165

@@ -176,8 +175,8 @@ Full pipeline implementation is available [here](https://github.com/apache/beam/
176175

177176
- Create: website/www/site/content/en/performance/<pipeline_name>/_index.md (short title/description).
178177
- Update: website/www/site/data/performance.yaml — add your pipeline and five chart entries with:
179-
- - looker_folder_id
180-
- - public_slug_id (from Looker, see below)
178+
- looker_folder_id
179+
- public_slug_id (from Looker, see below)
181180

182181
8. Create Looker content (5 charts)
183182

@@ -187,10 +186,12 @@ Full pipeline implementation is available [here](https://github.com/apache/beam/
187186
- Save changes → Publish to production.
188187
- From Explore, open each, set fields/filters for your pipeline, Run, then Save as Look (in your folder).
189188
- Open each Look:
190-
- - Copy Look ID
191-
- - Add Look IDs to .test-infra/tools/refresh_looker_metrics.py.
192-
- - Exit Development mode → Edit Settings → Allow public access.
193-
- - Copy public_slug_id and paste into website/performance.yml.
194-
- - Run .test-infra/tools/refresh_looker_metrics.py script or manually download as PNG via the public slug and upload to GCS: gs://public_looker_explores_us_a3853f40/FOLDER_ID/<look_slug>.png
189+
- Copy Look ID
190+
- Add Look IDs to .test-infra/tools/refresh_looker_metrics.py.
191+
- Exit Development mode → Edit Settings → Allow public access.
192+
- Copy public_slug_id and paste into website/performance.yml.
193+
- Run .test-infra/tools/refresh_looker_metrics.py script or manually download as PNG via the public slug and upload to GCS: gs://public_looker_explores_us_a3853f40/FOLDER_ID/<look_slug>.png
195194

196195
9. Open a PR
196+
197+
- Example: https://github.com/apache/beam/pull/34577

0 commit comments

Comments
 (0)