Skip to content
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
12 changes: 4 additions & 8 deletions .github/workflows/performance-regression.yml
Original file line number Diff line number Diff line change
Expand Up @@ -14,8 +14,7 @@ env:
spec_defaults: '{warmups:100,iterations:100}'
specs: '{command:read,format:ion_text} {command:write,format:ion_text} {command:read,format:ion_binary} {command:write,format:ion_binary}'
test_data_id: 'generated-test-data'
run_cli: 'python amazon/ionbenchmark/ion_benchmark_cli.py'
run_cli_new: 'python src-python/amazon/ionbenchmark/ion_benchmark_cli.py'
run_cli: 'python src-python/amazon/ionbenchmark/ion_benchmark_cli.py'


jobs:
Expand Down Expand Up @@ -104,10 +103,7 @@ jobs:
# Generates performance results for the previous commit
- name: Create a virtual environment for baseline
working-directory: ./baseline
run: |
pip install -r requirements.txt
[ -e "requirements_benchmark.txt" ] && pip install -r requirements_benchmark.txt # include benchmark requirements if they exist.
pip install .
run: pip install '.[test,benchmarking]'

- name: Run baseline performance benchmark
id: 'baseline'
Expand All @@ -130,7 +126,7 @@ jobs:
id: 'new'
working-directory: ./new
run: |
${{env.run_cli_new}} spec '${{env.specs}}' -d '${{env.spec_defaults}}' \
${{env.run_cli}} spec '${{env.specs}}' -d '${{env.spec_defaults}}' \
-O '{input_file:"${{steps.download.outputs.download-path}}/${{ matrix.test-data }}.10n"}' \
-o "$PWD/report.ion" -r '${{env.report_statistics}}'
echo "::group::Ion Report"
Expand All @@ -141,4 +137,4 @@ jobs:
# Compare results and identify regression
- name: Detect performance regression
working-directory: ./new
run: ${{env.run_cli_new}} compare --fail ${{steps.baseline.outputs.report}} ${{steps.new.outputs.report}} -c '${{env.compare_statistics}}'
run: ${{env.run_cli}} compare --fail ${{steps.baseline.outputs.report}} ${{steps.new.outputs.report}} -c '${{env.compare_statistics}}'
Loading