2021-02-22 17:28:32 +08:00
|
|
|
name: Tests
|
|
|
|
|
|
|
|
on:
|
|
|
|
pull_request:
|
2023-10-13 21:44:11 +08:00
|
|
|
paths-ignore:
|
2024-03-29 02:46:55 +08:00
|
|
|
- ".github/workflows/migration-tests.yml"
|
2024-03-29 02:53:25 +08:00
|
|
|
- ".github/dependabot.yml"
|
|
|
|
- ".github/labeler.yml"
|
2023-10-13 21:44:11 +08:00
|
|
|
- "migrations/**"
|
2021-02-22 17:28:32 +08:00
|
|
|
push:
|
|
|
|
branches:
|
|
|
|
- main
|
2022-03-18 10:15:48 +08:00
|
|
|
- beta
|
|
|
|
- stable
|
2023-10-13 21:44:11 +08:00
|
|
|
paths-ignore:
|
2024-03-29 02:46:55 +08:00
|
|
|
- ".github/workflows/migration-tests.yml"
|
2024-03-29 02:53:25 +08:00
|
|
|
- ".github/dependabot.yml"
|
|
|
|
- ".github/labeler.yml"
|
2023-10-13 21:44:11 +08:00
|
|
|
- "migrations/**"
|
2021-02-22 17:28:32 +08:00
|
|
|
|
2021-11-26 04:44:40 +08:00
|
|
|
concurrency:
|
2021-11-26 06:31:05 +08:00
|
|
|
group: tests-${{ format('{0}-{1}', github.head_ref || github.run_number, github.job) }}
|
2021-11-26 04:44:40 +08:00
|
|
|
cancel-in-progress: true
|
|
|
|
|
2022-07-30 23:22:03 +08:00
|
|
|
permissions:
|
|
|
|
contents: read
|
|
|
|
|
2021-02-22 17:28:32 +08:00
|
|
|
jobs:
|
|
|
|
build:
|
2023-05-12 20:00:04 +08:00
|
|
|
if: github.event_name == 'pull_request' || github.repository != 'discourse/discourse-private-mirror'
|
2024-02-26 18:53:32 +08:00
|
|
|
name: ${{ matrix.target }} ${{ matrix.build_type }} # Update fetch-job-id step if changing this
|
2024-09-30 11:18:20 +08:00
|
|
|
runs-on: ${{ (github.repository != 'discourse/discourse' && 'ubuntu-22.04-8core') || 'debian-12' }}
|
2024-03-28 09:20:26 +08:00
|
|
|
container: discourse/discourse_test:slim${{ (matrix.build_type == 'frontend' || matrix.build_type == 'system') && '-browsers' || '' }}
|
2022-09-22 01:13:13 +08:00
|
|
|
timeout-minutes: 20
|
2021-02-22 17:28:32 +08:00
|
|
|
|
|
|
|
env:
|
|
|
|
DISCOURSE_HOSTNAME: www.example.com
|
|
|
|
RAILS_ENV: test
|
|
|
|
PGUSER: discourse
|
|
|
|
PGPASSWORD: discourse
|
2023-01-26 21:26:02 +08:00
|
|
|
USES_PARALLEL_DATABASES: ${{ matrix.build_type == 'backend' || matrix.build_type == 'system' }}
|
2023-05-29 10:41:24 +08:00
|
|
|
CAPYBARA_DEFAULT_MAX_WAIT_TIME: 10
|
2023-08-23 09:18:33 +08:00
|
|
|
MINIO_RUNNER_LOG_LEVEL: DEBUG
|
2024-10-21 22:53:52 +08:00
|
|
|
DISCOURSE_TURBO_RSPEC_RETRY_AND_LOG_FLAKY_TESTS: ${{ (matrix.build_type == 'system' || matrix.build_type == 'backend') && '1' }}
|
2024-05-08 17:28:07 +08:00
|
|
|
CHEAP_SOURCE_MAPS: "1"
|
2024-08-20 13:12:33 +08:00
|
|
|
TESTEM_DEFAULT_BROWSER: Chrome
|
2024-09-18 10:20:10 +08:00
|
|
|
MINIO_RUNNER_INSTALL_DIR: /home/discourse/.minio_runner
|
2021-02-22 17:28:32 +08:00
|
|
|
|
|
|
|
strategy:
|
|
|
|
fail-fast: false
|
|
|
|
|
|
|
|
matrix:
|
DEV: Minimal first pass of rails system test setup (#16311)
This commit introduces rails system tests run with chromedriver, selenium,
and headless chrome to our testing toolbox.
We use the `webdrivers` gem and `selenium-webdriver` which is what
the latest Rails uses so the tests run locally and in CI out of the box.
You can use `SELENIUM_VERBOSE_DRIVER_LOGS=1` to show extra
verbose logs of what selenium is doing to communicate with the system
tests.
By default JS logs are verbose so errors from JS are shown when
running system tests, you can disable this with
`SELENIUM_DISABLE_VERBOSE_JS_LOGS=1`
You can use `SELENIUM_HEADLESS=0` to run the system
tests inside a chrome browser instead of headless, which can be useful to debug things
and see what the spec sees. See note above about `bin/ember-cli` to avoid
surprises.
I have modified `bin/turbo_rspec` to exclude `spec/system` by default,
support for parallel system specs is a little shaky right now and we don't
want them slowing down the turbo by default either.
### PageObjects and System Tests
To make querying and inspecting parts of the page easier
and more reusable inbetween system tests, we are using the
concept of [PageObjects](https://www.selenium.dev/documentation/test_practices/encouraged/page_object_models/) in
our system tests. A "Page" here is generally corresponds to
an overarching ember route, e.g. "Topic" for `/t/324345/some-topic`,
and this contains logic for querying components within the topic
such as "Posts".
I have also split "Modals" into their own entity. Further down the
line we may want to explore creating independent "Component"
contexts.
Capybara DSL should be included in each PageObject class,
reference for this can be found at https://rubydoc.info/github/teamcapybara/capybara/master#the-dsl
For system tests, since they are so slow, we want to focus on
the "happy path" and not do every different possible context
and branch check using them. They are meant to be overarching
tests that check a number of things are correct using the full stack
from JS and ember to rails to ruby and then the database.
### CI Setup
Whenever a system spec fails, a screenshot
is taken and a build artifact is produced _after the entire CI run is complete_,
which can be downloaded from the Actions UI in the repo.
Most importantly, a step to build the Ember app using Ember CLI
is needed, otherwise the JS assets cannot be found by capybara:
```
- name: Build Ember CLI
run: bin/ember-cli --build
```
A new `--build` argument has been added to `bin/ember-cli` for this
case, which is not needed locally if you already have the discourse
rails server running via `bin/ember-cli -u` since the whole server is built and
set up by default.
Co-authored-by: David Taylor <david@taylorhq.com>
2022-09-28 09:48:16 +08:00
|
|
|
build_type: [backend, frontend, system, annotations]
|
2023-11-16 07:11:35 +08:00
|
|
|
target: [core, plugins, themes]
|
2021-07-06 16:47:16 +08:00
|
|
|
exclude:
|
|
|
|
- build_type: annotations
|
|
|
|
target: plugins
|
2023-11-16 07:11:35 +08:00
|
|
|
- build_type: annotations
|
|
|
|
target: themes
|
|
|
|
- build_type: backend
|
|
|
|
target: themes
|
2022-01-19 18:41:52 +08:00
|
|
|
- build_type: frontend
|
|
|
|
target: core # Handled by core_frontend_tests job (below)
|
2023-12-15 22:33:04 +08:00
|
|
|
include:
|
|
|
|
- build_type: system
|
|
|
|
target: chat
|
2021-02-22 17:28:32 +08:00
|
|
|
|
|
|
|
steps:
|
2023-01-30 23:39:43 +08:00
|
|
|
- name: Set working directory owner
|
|
|
|
run: chown root:root .
|
|
|
|
|
2024-08-26 08:42:35 +08:00
|
|
|
- name: Set PARALLEL_TEST_PROCESSORS for system tests
|
2024-08-23 16:56:09 +08:00
|
|
|
if: matrix.build_type == 'system'
|
|
|
|
run: |
|
|
|
|
echo "PARALLEL_TEST_PROCESSORS=$(($(nproc) / 2))" >> $GITHUB_ENV
|
|
|
|
|
2024-08-26 08:42:35 +08:00
|
|
|
- name: Set QUNIT_PARALLEL for QUnit tests
|
|
|
|
if: matrix.build_type == 'frontend'
|
|
|
|
run: |
|
2024-09-11 17:39:26 +08:00
|
|
|
if [ "${{ matrix.target }}" = "themes" ]; then
|
|
|
|
echo "QUNIT_PARALLEL=2" >> $GITHUB_ENV
|
|
|
|
else
|
|
|
|
echo "QUNIT_PARALLEL=$(($(nproc) / 2))" >> $GITHUB_ENV
|
|
|
|
fi
|
2024-08-26 08:42:35 +08:00
|
|
|
|
2023-09-11 17:42:02 +08:00
|
|
|
- uses: actions/checkout@v4
|
2021-02-22 17:28:32 +08:00
|
|
|
with:
|
|
|
|
fetch-depth: 1
|
|
|
|
|
|
|
|
- name: Setup Git
|
|
|
|
run: |
|
|
|
|
git config --global user.email "ci@ci.invalid"
|
|
|
|
git config --global user.name "Discourse CI"
|
|
|
|
|
2021-09-08 20:01:37 +08:00
|
|
|
- name: Start redis
|
|
|
|
run: |
|
|
|
|
redis-server /etc/redis/redis.conf &
|
2021-02-22 17:28:32 +08:00
|
|
|
|
2021-12-15 01:20:06 +08:00
|
|
|
- name: Start Postgres
|
|
|
|
run: |
|
|
|
|
chown -R postgres /var/run/postgresql
|
|
|
|
sudo -E -u postgres script/start_test_db.rb
|
|
|
|
sudo -u postgres psql -c "CREATE ROLE $PGUSER LOGIN SUPERUSER PASSWORD '$PGPASSWORD';"
|
|
|
|
|
2024-03-28 10:25:58 +08:00
|
|
|
- name: Container envs
|
|
|
|
id: container-envs
|
|
|
|
run: |
|
|
|
|
echo "ruby_version=$RUBY_VERSION" >> $GITHUB_OUTPUT
|
|
|
|
echo "debian_release=$DEBIAN_RELEASE" >> $GITHUB_OUTPUT
|
|
|
|
shell: bash
|
|
|
|
|
2021-02-22 17:28:32 +08:00
|
|
|
- name: Bundler cache
|
2024-01-22 18:50:56 +08:00
|
|
|
uses: actions/cache@v4
|
2021-02-22 17:28:32 +08:00
|
|
|
with:
|
|
|
|
path: vendor/bundle
|
2024-03-28 10:25:58 +08:00
|
|
|
key: ${{ runner.os }}-${{ steps.container-envs.outputs.ruby_version }}-${{ steps.container-envs.outputs.debian_release }}-gem-${{ hashFiles('**/Gemfile.lock') }}-cachev2
|
2021-02-22 17:28:32 +08:00
|
|
|
|
|
|
|
- name: Setup gems
|
|
|
|
run: |
|
2022-01-13 05:50:04 +08:00
|
|
|
gem install bundler --conservative -v $(awk '/BUNDLED WITH/ { getline; gsub(/ /,""); print $0 }' Gemfile.lock)
|
2021-02-22 17:28:32 +08:00
|
|
|
bundle config --local path vendor/bundle
|
|
|
|
bundle config --local deployment true
|
|
|
|
bundle config --local without development
|
2024-03-27 10:19:09 +08:00
|
|
|
bundle install --jobs $(($(nproc) - 1))
|
2021-02-22 17:28:32 +08:00
|
|
|
bundle clean
|
|
|
|
|
2024-09-03 17:51:07 +08:00
|
|
|
- name: pnpm install
|
|
|
|
run: pnpm install --frozen-lockfile
|
2021-02-22 17:28:32 +08:00
|
|
|
|
|
|
|
- name: Checkout official plugins
|
|
|
|
if: matrix.target == 'plugins'
|
|
|
|
run: bin/rake plugin:install_all_official
|
|
|
|
|
2022-03-18 10:49:58 +08:00
|
|
|
- name: Pull compatible versions of plugins
|
2024-10-07 14:30:03 +08:00
|
|
|
if: matrix.target == 'plugins' && (github.ref_name != 'main' && github.base_ref != 'main')
|
2022-03-18 10:49:58 +08:00
|
|
|
run: bin/rake plugin:pull_compatible_all
|
|
|
|
|
2023-11-27 07:22:32 +08:00
|
|
|
- name: Plugin gems cache
|
2024-01-22 18:50:56 +08:00
|
|
|
uses: actions/cache@v4
|
2023-11-27 07:22:32 +08:00
|
|
|
with:
|
|
|
|
path: plugins/*/gems
|
2024-03-28 10:25:58 +08:00
|
|
|
key: ${{ runner.os }}-plugin-gems-${{ steps.container-envs.outputs.ruby_version }}-${{ steps.container-envs.outputs.debian_release }}-${{ hashFiles('plugins/*/plugin.rb') }}
|
2023-11-27 07:22:32 +08:00
|
|
|
|
2023-11-16 07:11:35 +08:00
|
|
|
- name: Checkout official themes
|
2024-05-21 10:38:41 +08:00
|
|
|
if: matrix.target == 'themes'
|
2024-06-04 21:16:44 +08:00
|
|
|
run: bin/rake themes:clone_all_official themes:pull_compatible_all
|
2024-05-21 10:38:41 +08:00
|
|
|
|
2024-08-20 13:12:33 +08:00
|
|
|
- name: Add hosts to /etc/hosts, otherwise Chrome cannot reach minio
|
2023-12-07 11:46:14 +08:00
|
|
|
if: matrix.build_type == 'system' && matrix.target == 'core'
|
2023-08-23 09:18:33 +08:00
|
|
|
run: |
|
|
|
|
echo "127.0.0.1 minio.local" | sudo tee -a /etc/hosts
|
|
|
|
echo "127.0.0.1 discoursetest.minio.local" | sudo tee -a /etc/hosts
|
|
|
|
|
2024-08-29 10:40:34 +08:00
|
|
|
- name: Get CPU cores
|
|
|
|
id: cpu-info
|
|
|
|
run: echo "cpu-cores=$(nproc)" >> $GITHUB_OUTPUT
|
|
|
|
|
2021-12-14 17:40:16 +08:00
|
|
|
- name: Fetch app state cache
|
2024-01-22 18:50:56 +08:00
|
|
|
uses: actions/cache@v4
|
2021-12-14 17:40:16 +08:00
|
|
|
id: app-cache
|
|
|
|
with:
|
|
|
|
path: tmp/app-cache
|
2023-05-12 20:00:04 +08:00
|
|
|
key: >-
|
2024-08-29 10:40:34 +08:00
|
|
|
${{ runner.os }}-
|
|
|
|
${{ steps.cpu-info.outputs.cpu-cores }}-
|
2021-12-14 17:40:16 +08:00
|
|
|
${{ hashFiles('.github/workflows/tests.yml') }}-
|
|
|
|
${{ hashFiles('db/**/*', 'plugins/**/db/**/*') }}-
|
2024-01-29 09:57:58 +08:00
|
|
|
${{ hashFiles('config/environments/test.rb') }}-
|
2024-08-29 13:03:21 +08:00
|
|
|
${{ env.USES_PARALLEL_DATABASES }}-
|
|
|
|
${{ env.PARALLEL_TEST_PROCESSORS }}-
|
2021-12-14 17:40:16 +08:00
|
|
|
|
|
|
|
- name: Restore database from cache
|
|
|
|
if: steps.app-cache.outputs.cache-hit == 'true'
|
2023-11-16 23:55:41 +08:00
|
|
|
run: script/silence_successful_output psql --quiet -o /dev/null -f tmp/app-cache/cache.sql postgres
|
2021-12-14 17:40:16 +08:00
|
|
|
|
|
|
|
- name: Restore uploads from cache
|
|
|
|
if: steps.app-cache.outputs.cache-hit == 'true'
|
|
|
|
run: rm -rf public/uploads && cp -r tmp/app-cache/uploads public/uploads
|
|
|
|
|
|
|
|
- name: Create and migrate database
|
|
|
|
if: steps.app-cache.outputs.cache-hit != 'true'
|
2021-02-22 17:28:32 +08:00
|
|
|
run: |
|
|
|
|
bin/rake db:create
|
2023-11-16 23:55:41 +08:00
|
|
|
script/silence_successful_output bin/rake db:migrate
|
2021-02-22 17:28:32 +08:00
|
|
|
|
2021-12-14 17:40:16 +08:00
|
|
|
- name: Create and migrate parallel databases
|
|
|
|
if: >-
|
|
|
|
env.USES_PARALLEL_DATABASES == 'true' &&
|
|
|
|
steps.app-cache.outputs.cache-hit != 'true'
|
2021-02-22 17:28:32 +08:00
|
|
|
run: |
|
|
|
|
bin/rake parallel:create
|
2023-11-16 23:55:41 +08:00
|
|
|
script/silence_successful_output bin/rake parallel:migrate
|
2021-02-22 17:28:32 +08:00
|
|
|
|
2021-12-14 17:40:16 +08:00
|
|
|
- name: Dump database for cache
|
|
|
|
if: steps.app-cache.outputs.cache-hit != 'true'
|
|
|
|
run: mkdir -p tmp/app-cache && pg_dumpall > tmp/app-cache/cache.sql
|
|
|
|
|
|
|
|
- name: Dump uploads for cache
|
|
|
|
if: steps.app-cache.outputs.cache-hit != 'true'
|
|
|
|
run: rm -rf tmp/app-cache/uploads && cp -r public/uploads tmp/app-cache/uploads
|
|
|
|
|
2022-09-22 05:13:25 +08:00
|
|
|
- name: Fetch turbo_rspec_runtime.log cache
|
2024-01-22 18:50:56 +08:00
|
|
|
uses: actions/cache@v4
|
2022-09-22 05:13:25 +08:00
|
|
|
id: test-runtime-cache
|
2023-06-12 09:07:17 +08:00
|
|
|
if: matrix.build_type == 'backend' || matrix.build_type == 'system'
|
2022-09-22 05:13:25 +08:00
|
|
|
with:
|
|
|
|
path: tmp/turbo_rspec_runtime.log
|
2023-06-12 09:07:17 +08:00
|
|
|
key: rspec-runtime-${{ matrix.build_type }}-${{ matrix.target }}-${{ github.run_id }}
|
|
|
|
restore-keys: rspec-runtime-${{ matrix.build_type }}-${{ matrix.target }}-
|
2022-09-22 05:13:25 +08:00
|
|
|
|
2023-06-16 21:33:14 +08:00
|
|
|
- name: Check Zeitwerk reloading
|
|
|
|
if: matrix.build_type == 'backend'
|
|
|
|
env:
|
|
|
|
LOAD_PLUGINS: ${{ (matrix.target == 'plugins') && '1' || '0' }}
|
|
|
|
run: |
|
|
|
|
if ! bin/rails runner 'Rails.application.reloader.reload!'; then
|
|
|
|
echo
|
|
|
|
echo "---------------------------------------------"
|
|
|
|
echo
|
|
|
|
echo "::error::Zeitwerk reload failed - the app will not be able to reload properly in development."
|
|
|
|
echo "To reproduce locally, run \`bin/rails runner 'Rails.application.reloader.reload!'\`."
|
|
|
|
echo
|
|
|
|
exit 1
|
|
|
|
fi
|
|
|
|
|
2021-02-22 17:28:32 +08:00
|
|
|
- name: Core RSpec
|
|
|
|
if: matrix.build_type == 'backend' && matrix.target == 'core'
|
2024-08-26 08:52:17 +08:00
|
|
|
run: bin/turbo_rspec --use-runtime-info --verbose --format=documentation --profile=50
|
2021-02-22 17:28:32 +08:00
|
|
|
|
|
|
|
- name: Plugin RSpec
|
|
|
|
if: matrix.build_type == 'backend' && matrix.target == 'plugins'
|
2024-08-26 08:52:17 +08:00
|
|
|
run: bin/rake plugin:turbo_spec['*','--verbose --format=documentation --use-runtime-info --profile=50']
|
2021-02-22 17:28:32 +08:00
|
|
|
|
2022-06-20 22:33:05 +08:00
|
|
|
- name: Plugin QUnit
|
2022-07-17 04:55:39 +08:00
|
|
|
if: matrix.build_type == 'frontend' && matrix.target == 'plugins'
|
2024-09-06 19:08:42 +08:00
|
|
|
run: QUNIT_WRITE_EXECUTION_FILE=1 bin/rake plugin:qunit['*']
|
2021-02-22 17:28:32 +08:00
|
|
|
timeout-minutes: 30
|
2021-07-06 16:47:16 +08:00
|
|
|
|
2023-11-17 07:17:32 +08:00
|
|
|
- name: Theme QUnit
|
|
|
|
if: matrix.build_type == 'frontend' && matrix.target == 'themes'
|
2024-06-24 07:14:26 +08:00
|
|
|
run: DISCOURSE_DEV_DB=discourse_test bin/rake themes:qunit_all_official
|
|
|
|
timeout-minutes: 10
|
2023-11-17 07:17:32 +08:00
|
|
|
|
2023-12-18 18:47:22 +08:00
|
|
|
- uses: actions/upload-artifact@v4
|
2023-10-16 18:12:13 +08:00
|
|
|
if: always() && matrix.build_type == 'frontend' && matrix.target == 'plugins'
|
2023-10-10 23:29:28 +08:00
|
|
|
with:
|
2023-12-18 18:47:22 +08:00
|
|
|
name: ember-exam-execution-plugins-frontend-${{ hashFiles('./app/assets/javascripts/discourse/test-execution-*.json') }}
|
2023-10-10 23:29:28 +08:00
|
|
|
path: ./app/assets/javascripts/discourse/test-execution-*.json
|
|
|
|
|
DEV: Minimal first pass of rails system test setup (#16311)
This commit introduces rails system tests run with chromedriver, selenium,
and headless chrome to our testing toolbox.
We use the `webdrivers` gem and `selenium-webdriver` which is what
the latest Rails uses so the tests run locally and in CI out of the box.
You can use `SELENIUM_VERBOSE_DRIVER_LOGS=1` to show extra
verbose logs of what selenium is doing to communicate with the system
tests.
By default JS logs are verbose so errors from JS are shown when
running system tests, you can disable this with
`SELENIUM_DISABLE_VERBOSE_JS_LOGS=1`
You can use `SELENIUM_HEADLESS=0` to run the system
tests inside a chrome browser instead of headless, which can be useful to debug things
and see what the spec sees. See note above about `bin/ember-cli` to avoid
surprises.
I have modified `bin/turbo_rspec` to exclude `spec/system` by default,
support for parallel system specs is a little shaky right now and we don't
want them slowing down the turbo by default either.
### PageObjects and System Tests
To make querying and inspecting parts of the page easier
and more reusable inbetween system tests, we are using the
concept of [PageObjects](https://www.selenium.dev/documentation/test_practices/encouraged/page_object_models/) in
our system tests. A "Page" here is generally corresponds to
an overarching ember route, e.g. "Topic" for `/t/324345/some-topic`,
and this contains logic for querying components within the topic
such as "Posts".
I have also split "Modals" into their own entity. Further down the
line we may want to explore creating independent "Component"
contexts.
Capybara DSL should be included in each PageObject class,
reference for this can be found at https://rubydoc.info/github/teamcapybara/capybara/master#the-dsl
For system tests, since they are so slow, we want to focus on
the "happy path" and not do every different possible context
and branch check using them. They are meant to be overarching
tests that check a number of things are correct using the full stack
from JS and ember to rails to ruby and then the database.
### CI Setup
Whenever a system spec fails, a screenshot
is taken and a build artifact is produced _after the entire CI run is complete_,
which can be downloaded from the Actions UI in the repo.
Most importantly, a step to build the Ember app using Ember CLI
is needed, otherwise the JS assets cannot be found by capybara:
```
- name: Build Ember CLI
run: bin/ember-cli --build
```
A new `--build` argument has been added to `bin/ember-cli` for this
case, which is not needed locally if you already have the discourse
rails server running via `bin/ember-cli -u` since the whole server is built and
set up by default.
Co-authored-by: David Taylor <david@taylorhq.com>
2022-09-28 09:48:16 +08:00
|
|
|
- name: Ember Build for System Tests
|
|
|
|
if: matrix.build_type == 'system'
|
|
|
|
run: bin/ember-cli --build
|
|
|
|
|
2024-09-18 10:20:10 +08:00
|
|
|
- name: Minio cache
|
|
|
|
if: matrix.build_type == 'system' && matrix.target == 'core'
|
|
|
|
uses: actions/cache@v4
|
|
|
|
with:
|
|
|
|
path: ${{ env.MINIO_RUNNER_INSTALL_DIR }}
|
|
|
|
key: ${{ runner.os }}-${{ steps.container-envs.outputs.ruby_version }}-${{ steps.container-envs.outputs.debian_release }}-gem-${{ hashFiles('**/Gemfile.lock') }}
|
|
|
|
|
2023-12-07 11:46:14 +08:00
|
|
|
- name: Ensure latest minio binary installed for Core System Tests
|
|
|
|
if: matrix.build_type == 'system' && matrix.target == 'core'
|
2023-10-24 10:43:14 +08:00
|
|
|
run: bundle exec ruby script/install_minio_binaries.rb
|
|
|
|
|
DEV: Minimal first pass of rails system test setup (#16311)
This commit introduces rails system tests run with chromedriver, selenium,
and headless chrome to our testing toolbox.
We use the `webdrivers` gem and `selenium-webdriver` which is what
the latest Rails uses so the tests run locally and in CI out of the box.
You can use `SELENIUM_VERBOSE_DRIVER_LOGS=1` to show extra
verbose logs of what selenium is doing to communicate with the system
tests.
By default JS logs are verbose so errors from JS are shown when
running system tests, you can disable this with
`SELENIUM_DISABLE_VERBOSE_JS_LOGS=1`
You can use `SELENIUM_HEADLESS=0` to run the system
tests inside a chrome browser instead of headless, which can be useful to debug things
and see what the spec sees. See note above about `bin/ember-cli` to avoid
surprises.
I have modified `bin/turbo_rspec` to exclude `spec/system` by default,
support for parallel system specs is a little shaky right now and we don't
want them slowing down the turbo by default either.
### PageObjects and System Tests
To make querying and inspecting parts of the page easier
and more reusable inbetween system tests, we are using the
concept of [PageObjects](https://www.selenium.dev/documentation/test_practices/encouraged/page_object_models/) in
our system tests. A "Page" here is generally corresponds to
an overarching ember route, e.g. "Topic" for `/t/324345/some-topic`,
and this contains logic for querying components within the topic
such as "Posts".
I have also split "Modals" into their own entity. Further down the
line we may want to explore creating independent "Component"
contexts.
Capybara DSL should be included in each PageObject class,
reference for this can be found at https://rubydoc.info/github/teamcapybara/capybara/master#the-dsl
For system tests, since they are so slow, we want to focus on
the "happy path" and not do every different possible context
and branch check using them. They are meant to be overarching
tests that check a number of things are correct using the full stack
from JS and ember to rails to ruby and then the database.
### CI Setup
Whenever a system spec fails, a screenshot
is taken and a build artifact is produced _after the entire CI run is complete_,
which can be downloaded from the Actions UI in the repo.
Most importantly, a step to build the Ember app using Ember CLI
is needed, otherwise the JS assets cannot be found by capybara:
```
- name: Build Ember CLI
run: bin/ember-cli --build
```
A new `--build` argument has been added to `bin/ember-cli` for this
case, which is not needed locally if you already have the discourse
rails server running via `bin/ember-cli -u` since the whole server is built and
set up by default.
Co-authored-by: David Taylor <david@taylorhq.com>
2022-09-28 09:48:16 +08:00
|
|
|
- name: Core System Tests
|
|
|
|
if: matrix.build_type == 'system' && matrix.target == 'core'
|
2024-02-08 10:35:55 +08:00
|
|
|
env:
|
|
|
|
CHECKOUT_TIMEOUT: 10
|
2024-08-23 16:56:09 +08:00
|
|
|
run: RAILS_ENABLE_TEST_LOG=1 RAILS_TEST_LOG_LEVEL=error bin/turbo_rspec --use-runtime-info --profile=50 --verbose --format documentation spec/system
|
DEV: Minimal first pass of rails system test setup (#16311)
This commit introduces rails system tests run with chromedriver, selenium,
and headless chrome to our testing toolbox.
We use the `webdrivers` gem and `selenium-webdriver` which is what
the latest Rails uses so the tests run locally and in CI out of the box.
You can use `SELENIUM_VERBOSE_DRIVER_LOGS=1` to show extra
verbose logs of what selenium is doing to communicate with the system
tests.
By default JS logs are verbose so errors from JS are shown when
running system tests, you can disable this with
`SELENIUM_DISABLE_VERBOSE_JS_LOGS=1`
You can use `SELENIUM_HEADLESS=0` to run the system
tests inside a chrome browser instead of headless, which can be useful to debug things
and see what the spec sees. See note above about `bin/ember-cli` to avoid
surprises.
I have modified `bin/turbo_rspec` to exclude `spec/system` by default,
support for parallel system specs is a little shaky right now and we don't
want them slowing down the turbo by default either.
### PageObjects and System Tests
To make querying and inspecting parts of the page easier
and more reusable inbetween system tests, we are using the
concept of [PageObjects](https://www.selenium.dev/documentation/test_practices/encouraged/page_object_models/) in
our system tests. A "Page" here is generally corresponds to
an overarching ember route, e.g. "Topic" for `/t/324345/some-topic`,
and this contains logic for querying components within the topic
such as "Posts".
I have also split "Modals" into their own entity. Further down the
line we may want to explore creating independent "Component"
contexts.
Capybara DSL should be included in each PageObject class,
reference for this can be found at https://rubydoc.info/github/teamcapybara/capybara/master#the-dsl
For system tests, since they are so slow, we want to focus on
the "happy path" and not do every different possible context
and branch check using them. They are meant to be overarching
tests that check a number of things are correct using the full stack
from JS and ember to rails to ruby and then the database.
### CI Setup
Whenever a system spec fails, a screenshot
is taken and a build artifact is produced _after the entire CI run is complete_,
which can be downloaded from the Actions UI in the repo.
Most importantly, a step to build the Ember app using Ember CLI
is needed, otherwise the JS assets cannot be found by capybara:
```
- name: Build Ember CLI
run: bin/ember-cli --build
```
A new `--build` argument has been added to `bin/ember-cli` for this
case, which is not needed locally if you already have the discourse
rails server running via `bin/ember-cli -u` since the whole server is built and
set up by default.
Co-authored-by: David Taylor <david@taylorhq.com>
2022-09-28 09:48:16 +08:00
|
|
|
|
|
|
|
- name: Plugin System Tests
|
|
|
|
if: matrix.build_type == 'system' && matrix.target == 'plugins'
|
2024-02-08 10:35:55 +08:00
|
|
|
env:
|
|
|
|
CHECKOUT_TIMEOUT: 10
|
2023-10-25 19:58:36 +08:00
|
|
|
run: |
|
|
|
|
GLOBIGNORE="plugins/chat/*";
|
2024-08-23 16:56:09 +08:00
|
|
|
LOAD_PLUGINS=1 RAILS_ENABLE_TEST_LOG=1 RAILS_TEST_LOG_LEVEL=error bin/turbo_rspec --use-runtime-info --profile=50 --verbose --format documentation plugins/*/spec/system
|
2023-10-25 19:58:36 +08:00
|
|
|
shell: bash
|
|
|
|
timeout-minutes: 30
|
2023-11-16 07:11:35 +08:00
|
|
|
|
2023-10-25 19:58:36 +08:00
|
|
|
- name: Chat System Tests
|
|
|
|
if: matrix.build_type == 'system' && matrix.target == 'chat'
|
2024-02-08 10:35:55 +08:00
|
|
|
env:
|
|
|
|
CHECKOUT_TIMEOUT: 10
|
2024-08-23 16:56:09 +08:00
|
|
|
run: LOAD_PLUGINS=1 RAILS_ENABLE_TEST_LOG=1 RAILS_TEST_LOG_LEVEL=error bin/turbo_rspec --use-runtime-info --profile=50 --verbose --format documentation plugins/chat/spec/system
|
2023-04-26 16:08:10 +08:00
|
|
|
timeout-minutes: 30
|
DEV: Minimal first pass of rails system test setup (#16311)
This commit introduces rails system tests run with chromedriver, selenium,
and headless chrome to our testing toolbox.
We use the `webdrivers` gem and `selenium-webdriver` which is what
the latest Rails uses so the tests run locally and in CI out of the box.
You can use `SELENIUM_VERBOSE_DRIVER_LOGS=1` to show extra
verbose logs of what selenium is doing to communicate with the system
tests.
By default JS logs are verbose so errors from JS are shown when
running system tests, you can disable this with
`SELENIUM_DISABLE_VERBOSE_JS_LOGS=1`
You can use `SELENIUM_HEADLESS=0` to run the system
tests inside a chrome browser instead of headless, which can be useful to debug things
and see what the spec sees. See note above about `bin/ember-cli` to avoid
surprises.
I have modified `bin/turbo_rspec` to exclude `spec/system` by default,
support for parallel system specs is a little shaky right now and we don't
want them slowing down the turbo by default either.
### PageObjects and System Tests
To make querying and inspecting parts of the page easier
and more reusable inbetween system tests, we are using the
concept of [PageObjects](https://www.selenium.dev/documentation/test_practices/encouraged/page_object_models/) in
our system tests. A "Page" here is generally corresponds to
an overarching ember route, e.g. "Topic" for `/t/324345/some-topic`,
and this contains logic for querying components within the topic
such as "Posts".
I have also split "Modals" into their own entity. Further down the
line we may want to explore creating independent "Component"
contexts.
Capybara DSL should be included in each PageObject class,
reference for this can be found at https://rubydoc.info/github/teamcapybara/capybara/master#the-dsl
For system tests, since they are so slow, we want to focus on
the "happy path" and not do every different possible context
and branch check using them. They are meant to be overarching
tests that check a number of things are correct using the full stack
from JS and ember to rails to ruby and then the database.
### CI Setup
Whenever a system spec fails, a screenshot
is taken and a build artifact is produced _after the entire CI run is complete_,
which can be downloaded from the Actions UI in the repo.
Most importantly, a step to build the Ember app using Ember CLI
is needed, otherwise the JS assets cannot be found by capybara:
```
- name: Build Ember CLI
run: bin/ember-cli --build
```
A new `--build` argument has been added to `bin/ember-cli` for this
case, which is not needed locally if you already have the discourse
rails server running via `bin/ember-cli -u` since the whole server is built and
set up by default.
Co-authored-by: David Taylor <david@taylorhq.com>
2022-09-28 09:48:16 +08:00
|
|
|
|
2023-11-16 07:11:35 +08:00
|
|
|
- name: Theme System Tests
|
|
|
|
if: matrix.build_type == 'system' && matrix.target == 'themes'
|
2024-02-08 10:35:55 +08:00
|
|
|
env:
|
|
|
|
CHECKOUT_TIMEOUT: 10
|
2023-11-16 07:11:35 +08:00
|
|
|
run: |
|
2024-08-23 16:56:09 +08:00
|
|
|
RAILS_ENABLE_TEST_LOG=1 RAILS_TEST_LOG_LEVEL=error bin/turbo_rspec --profile=50 --verbose --format documentation tmp/themes/*/spec/system
|
2023-11-16 07:11:35 +08:00
|
|
|
shell: bash
|
|
|
|
timeout-minutes: 30
|
|
|
|
|
DEV: Introduce automatic reruns to RSpec tests on Github actions (#24811)
What motivated this change?
Our builds on Github actions have been extremely flaky mostly due to system tests. This has led to a drop in confidence
in our test suite where our developers tend to assume that a failed job is due to a flaky system test. As a result, we
have had occurrences where changes that resulted in legitimate test failures are merged into the `main` branch because developers
assumed it was a flaky test.
What does this change do?
This change seeks to reduce the flakiness of our builds on Github Actions by automatically re-running RSpec tests once when
they fail. If a failed test passes subsequently in the re-run, we mark the test as flaky by logging it into a file on disk
which is then uploaded as an artifact of the Github workflow run. We understand that automatically re-runs will lead to
lower accuracy of our tests but we accept this as an acceptable trade-off since a fragile build has a much greater impact
on our developers' time. Internally, the Discourse development team will be running a service to fetch the flaky tests
which have been logged for internal monitoring.
How is the change implemented?
1. A `--retry-and-log-flaky-tests` CLI flag is added to the `bin/turbo_rspec` CLI which will then initialize `TurboTests::Runner`
with the `retry_and_log_flaky_tests` kwarg set to `true`.
2. When the `retry_and_log_flaky_tests` kwarg is set to `true` for `TurboTests::Runner`, we will register an additional
formatter `Flaky::FailuresLoggerFormatter` to the `TurboTests::Reporter` in the `TurboTests::Runner#run` method.
The `Flaky::FailuresLoggerFormatter` has a simple job of logging all failed examples to a file on disk when running all the
tests. The details of the failed example which are logged can be found in `TurboTests::Flaky::FailedExample.to_h`.
3. Once all the tests have been run once, we check the result for any failed examples and if there are, we read the file on
disk to fetch the `location_rerun_location` of the failed examples which is then used to run the tests in a new RSpec process.
In the rerun, we configure a `TurboTests::Flaky::FlakyDetectorFormatter` with RSpec which removes all failed examples from the log file on disk since those examples are not flaky tests. Note that if there are too many failed examples on the first run, we will deem the failures to likely not be due to flaky tests and not re-run the test failures. As of writing, the threshold of failed examples is set to 10. If there are more than 10 failed examples, we will not re-run the failures.
2023-12-13 07:18:27 +08:00
|
|
|
- name: Check for failed system test screenshots
|
|
|
|
id: check-failed-system-test-screenshots
|
|
|
|
if: always() && matrix.build_type == 'system'
|
|
|
|
run: |
|
|
|
|
if [ -d tmp/capybara ] && [ "$(ls -A tmp/capybara/)" ]; then
|
|
|
|
echo "exists=true" >> $GITHUB_OUTPUT
|
|
|
|
else
|
|
|
|
echo "exists=false" >> $GITHUB_OUTPUT
|
|
|
|
fi
|
|
|
|
shell: bash
|
|
|
|
|
DEV: Minimal first pass of rails system test setup (#16311)
This commit introduces rails system tests run with chromedriver, selenium,
and headless chrome to our testing toolbox.
We use the `webdrivers` gem and `selenium-webdriver` which is what
the latest Rails uses so the tests run locally and in CI out of the box.
You can use `SELENIUM_VERBOSE_DRIVER_LOGS=1` to show extra
verbose logs of what selenium is doing to communicate with the system
tests.
By default JS logs are verbose so errors from JS are shown when
running system tests, you can disable this with
`SELENIUM_DISABLE_VERBOSE_JS_LOGS=1`
You can use `SELENIUM_HEADLESS=0` to run the system
tests inside a chrome browser instead of headless, which can be useful to debug things
and see what the spec sees. See note above about `bin/ember-cli` to avoid
surprises.
I have modified `bin/turbo_rspec` to exclude `spec/system` by default,
support for parallel system specs is a little shaky right now and we don't
want them slowing down the turbo by default either.
### PageObjects and System Tests
To make querying and inspecting parts of the page easier
and more reusable inbetween system tests, we are using the
concept of [PageObjects](https://www.selenium.dev/documentation/test_practices/encouraged/page_object_models/) in
our system tests. A "Page" here is generally corresponds to
an overarching ember route, e.g. "Topic" for `/t/324345/some-topic`,
and this contains logic for querying components within the topic
such as "Posts".
I have also split "Modals" into their own entity. Further down the
line we may want to explore creating independent "Component"
contexts.
Capybara DSL should be included in each PageObject class,
reference for this can be found at https://rubydoc.info/github/teamcapybara/capybara/master#the-dsl
For system tests, since they are so slow, we want to focus on
the "happy path" and not do every different possible context
and branch check using them. They are meant to be overarching
tests that check a number of things are correct using the full stack
from JS and ember to rails to ruby and then the database.
### CI Setup
Whenever a system spec fails, a screenshot
is taken and a build artifact is produced _after the entire CI run is complete_,
which can be downloaded from the Actions UI in the repo.
Most importantly, a step to build the Ember app using Ember CLI
is needed, otherwise the JS assets cannot be found by capybara:
```
- name: Build Ember CLI
run: bin/ember-cli --build
```
A new `--build` argument has been added to `bin/ember-cli` for this
case, which is not needed locally if you already have the discourse
rails server running via `bin/ember-cli -u` since the whole server is built and
set up by default.
Co-authored-by: David Taylor <david@taylorhq.com>
2022-09-28 09:48:16 +08:00
|
|
|
- name: Upload failed system test screenshots
|
2024-11-11 19:37:53 +08:00
|
|
|
uses: actions/upload-artifact@v4
|
DEV: Introduce automatic reruns to RSpec tests on Github actions (#24811)
What motivated this change?
Our builds on Github actions have been extremely flaky mostly due to system tests. This has led to a drop in confidence
in our test suite where our developers tend to assume that a failed job is due to a flaky system test. As a result, we
have had occurrences where changes that resulted in legitimate test failures are merged into the `main` branch because developers
assumed it was a flaky test.
What does this change do?
This change seeks to reduce the flakiness of our builds on Github Actions by automatically re-running RSpec tests once when
they fail. If a failed test passes subsequently in the re-run, we mark the test as flaky by logging it into a file on disk
which is then uploaded as an artifact of the Github workflow run. We understand that automatically re-runs will lead to
lower accuracy of our tests but we accept this as an acceptable trade-off since a fragile build has a much greater impact
on our developers' time. Internally, the Discourse development team will be running a service to fetch the flaky tests
which have been logged for internal monitoring.
How is the change implemented?
1. A `--retry-and-log-flaky-tests` CLI flag is added to the `bin/turbo_rspec` CLI which will then initialize `TurboTests::Runner`
with the `retry_and_log_flaky_tests` kwarg set to `true`.
2. When the `retry_and_log_flaky_tests` kwarg is set to `true` for `TurboTests::Runner`, we will register an additional
formatter `Flaky::FailuresLoggerFormatter` to the `TurboTests::Reporter` in the `TurboTests::Runner#run` method.
The `Flaky::FailuresLoggerFormatter` has a simple job of logging all failed examples to a file on disk when running all the
tests. The details of the failed example which are logged can be found in `TurboTests::Flaky::FailedExample.to_h`.
3. Once all the tests have been run once, we check the result for any failed examples and if there are, we read the file on
disk to fetch the `location_rerun_location` of the failed examples which is then used to run the tests in a new RSpec process.
In the rerun, we configure a `TurboTests::Flaky::FlakyDetectorFormatter` with RSpec which removes all failed examples from the log file on disk since those examples are not flaky tests. Note that if there are too many failed examples on the first run, we will deem the failures to likely not be due to flaky tests and not re-run the test failures. As of writing, the threshold of failed examples is set to 10. If there are more than 10 failed examples, we will not re-run the failures.
2023-12-13 07:18:27 +08:00
|
|
|
if: always() && steps.check-failed-system-test-screenshots.outputs.exists == 'true'
|
DEV: Minimal first pass of rails system test setup (#16311)
This commit introduces rails system tests run with chromedriver, selenium,
and headless chrome to our testing toolbox.
We use the `webdrivers` gem and `selenium-webdriver` which is what
the latest Rails uses so the tests run locally and in CI out of the box.
You can use `SELENIUM_VERBOSE_DRIVER_LOGS=1` to show extra
verbose logs of what selenium is doing to communicate with the system
tests.
By default JS logs are verbose so errors from JS are shown when
running system tests, you can disable this with
`SELENIUM_DISABLE_VERBOSE_JS_LOGS=1`
You can use `SELENIUM_HEADLESS=0` to run the system
tests inside a chrome browser instead of headless, which can be useful to debug things
and see what the spec sees. See note above about `bin/ember-cli` to avoid
surprises.
I have modified `bin/turbo_rspec` to exclude `spec/system` by default,
support for parallel system specs is a little shaky right now and we don't
want them slowing down the turbo by default either.
### PageObjects and System Tests
To make querying and inspecting parts of the page easier
and more reusable inbetween system tests, we are using the
concept of [PageObjects](https://www.selenium.dev/documentation/test_practices/encouraged/page_object_models/) in
our system tests. A "Page" here is generally corresponds to
an overarching ember route, e.g. "Topic" for `/t/324345/some-topic`,
and this contains logic for querying components within the topic
such as "Posts".
I have also split "Modals" into their own entity. Further down the
line we may want to explore creating independent "Component"
contexts.
Capybara DSL should be included in each PageObject class,
reference for this can be found at https://rubydoc.info/github/teamcapybara/capybara/master#the-dsl
For system tests, since they are so slow, we want to focus on
the "happy path" and not do every different possible context
and branch check using them. They are meant to be overarching
tests that check a number of things are correct using the full stack
from JS and ember to rails to ruby and then the database.
### CI Setup
Whenever a system spec fails, a screenshot
is taken and a build artifact is produced _after the entire CI run is complete_,
which can be downloaded from the Actions UI in the repo.
Most importantly, a step to build the Ember app using Ember CLI
is needed, otherwise the JS assets cannot be found by capybara:
```
- name: Build Ember CLI
run: bin/ember-cli --build
```
A new `--build` argument has been added to `bin/ember-cli` for this
case, which is not needed locally if you already have the discourse
rails server running via `bin/ember-cli -u` since the whole server is built and
set up by default.
Co-authored-by: David Taylor <david@taylorhq.com>
2022-09-28 09:48:16 +08:00
|
|
|
with:
|
2024-11-11 19:37:53 +08:00
|
|
|
name: failed-system-test-screenshots-${{ matrix.build_type }}-${{ matrix.target }}
|
2022-12-13 13:36:30 +08:00
|
|
|
path: tmp/capybara/*.png
|
DEV: Minimal first pass of rails system test setup (#16311)
This commit introduces rails system tests run with chromedriver, selenium,
and headless chrome to our testing toolbox.
We use the `webdrivers` gem and `selenium-webdriver` which is what
the latest Rails uses so the tests run locally and in CI out of the box.
You can use `SELENIUM_VERBOSE_DRIVER_LOGS=1` to show extra
verbose logs of what selenium is doing to communicate with the system
tests.
By default JS logs are verbose so errors from JS are shown when
running system tests, you can disable this with
`SELENIUM_DISABLE_VERBOSE_JS_LOGS=1`
You can use `SELENIUM_HEADLESS=0` to run the system
tests inside a chrome browser instead of headless, which can be useful to debug things
and see what the spec sees. See note above about `bin/ember-cli` to avoid
surprises.
I have modified `bin/turbo_rspec` to exclude `spec/system` by default,
support for parallel system specs is a little shaky right now and we don't
want them slowing down the turbo by default either.
### PageObjects and System Tests
To make querying and inspecting parts of the page easier
and more reusable inbetween system tests, we are using the
concept of [PageObjects](https://www.selenium.dev/documentation/test_practices/encouraged/page_object_models/) in
our system tests. A "Page" here is generally corresponds to
an overarching ember route, e.g. "Topic" for `/t/324345/some-topic`,
and this contains logic for querying components within the topic
such as "Posts".
I have also split "Modals" into their own entity. Further down the
line we may want to explore creating independent "Component"
contexts.
Capybara DSL should be included in each PageObject class,
reference for this can be found at https://rubydoc.info/github/teamcapybara/capybara/master#the-dsl
For system tests, since they are so slow, we want to focus on
the "happy path" and not do every different possible context
and branch check using them. They are meant to be overarching
tests that check a number of things are correct using the full stack
from JS and ember to rails to ruby and then the database.
### CI Setup
Whenever a system spec fails, a screenshot
is taken and a build artifact is produced _after the entire CI run is complete_,
which can be downloaded from the Actions UI in the repo.
Most importantly, a step to build the Ember app using Ember CLI
is needed, otherwise the JS assets cannot be found by capybara:
```
- name: Build Ember CLI
run: bin/ember-cli --build
```
A new `--build` argument has been added to `bin/ember-cli` for this
case, which is not needed locally if you already have the discourse
rails server running via `bin/ember-cli -u` since the whole server is built and
set up by default.
Co-authored-by: David Taylor <david@taylorhq.com>
2022-09-28 09:48:16 +08:00
|
|
|
|
DEV: Introduce automatic reruns to RSpec tests on Github actions (#24811)
What motivated this change?
Our builds on Github actions have been extremely flaky mostly due to system tests. This has led to a drop in confidence
in our test suite where our developers tend to assume that a failed job is due to a flaky system test. As a result, we
have had occurrences where changes that resulted in legitimate test failures are merged into the `main` branch because developers
assumed it was a flaky test.
What does this change do?
This change seeks to reduce the flakiness of our builds on Github Actions by automatically re-running RSpec tests once when
they fail. If a failed test passes subsequently in the re-run, we mark the test as flaky by logging it into a file on disk
which is then uploaded as an artifact of the Github workflow run. We understand that automatically re-runs will lead to
lower accuracy of our tests but we accept this as an acceptable trade-off since a fragile build has a much greater impact
on our developers' time. Internally, the Discourse development team will be running a service to fetch the flaky tests
which have been logged for internal monitoring.
How is the change implemented?
1. A `--retry-and-log-flaky-tests` CLI flag is added to the `bin/turbo_rspec` CLI which will then initialize `TurboTests::Runner`
with the `retry_and_log_flaky_tests` kwarg set to `true`.
2. When the `retry_and_log_flaky_tests` kwarg is set to `true` for `TurboTests::Runner`, we will register an additional
formatter `Flaky::FailuresLoggerFormatter` to the `TurboTests::Reporter` in the `TurboTests::Runner#run` method.
The `Flaky::FailuresLoggerFormatter` has a simple job of logging all failed examples to a file on disk when running all the
tests. The details of the failed example which are logged can be found in `TurboTests::Flaky::FailedExample.to_h`.
3. Once all the tests have been run once, we check the result for any failed examples and if there are, we read the file on
disk to fetch the `location_rerun_location` of the failed examples which is then used to run the tests in a new RSpec process.
In the rerun, we configure a `TurboTests::Flaky::FlakyDetectorFormatter` with RSpec which removes all failed examples from the log file on disk since those examples are not flaky tests. Note that if there are too many failed examples on the first run, we will deem the failures to likely not be due to flaky tests and not re-run the test failures. As of writing, the threshold of failed examples is set to 10. If there are more than 10 failed examples, we will not re-run the failures.
2023-12-13 07:18:27 +08:00
|
|
|
- name: Check for flaky tests report
|
|
|
|
id: check-flaky-spec-report
|
2024-09-13 14:59:49 +08:00
|
|
|
if: always() && github.repository == 'discourse/discourse' && ${{ env.DISCOURSE_TURBO_RSPEC_RETRY_AND_LOG_FLAKY_TESTS == '1' }}
|
DEV: Introduce automatic reruns to RSpec tests on Github actions (#24811)
What motivated this change?
Our builds on Github actions have been extremely flaky mostly due to system tests. This has led to a drop in confidence
in our test suite where our developers tend to assume that a failed job is due to a flaky system test. As a result, we
have had occurrences where changes that resulted in legitimate test failures are merged into the `main` branch because developers
assumed it was a flaky test.
What does this change do?
This change seeks to reduce the flakiness of our builds on Github Actions by automatically re-running RSpec tests once when
they fail. If a failed test passes subsequently in the re-run, we mark the test as flaky by logging it into a file on disk
which is then uploaded as an artifact of the Github workflow run. We understand that automatically re-runs will lead to
lower accuracy of our tests but we accept this as an acceptable trade-off since a fragile build has a much greater impact
on our developers' time. Internally, the Discourse development team will be running a service to fetch the flaky tests
which have been logged for internal monitoring.
How is the change implemented?
1. A `--retry-and-log-flaky-tests` CLI flag is added to the `bin/turbo_rspec` CLI which will then initialize `TurboTests::Runner`
with the `retry_and_log_flaky_tests` kwarg set to `true`.
2. When the `retry_and_log_flaky_tests` kwarg is set to `true` for `TurboTests::Runner`, we will register an additional
formatter `Flaky::FailuresLoggerFormatter` to the `TurboTests::Reporter` in the `TurboTests::Runner#run` method.
The `Flaky::FailuresLoggerFormatter` has a simple job of logging all failed examples to a file on disk when running all the
tests. The details of the failed example which are logged can be found in `TurboTests::Flaky::FailedExample.to_h`.
3. Once all the tests have been run once, we check the result for any failed examples and if there are, we read the file on
disk to fetch the `location_rerun_location` of the failed examples which is then used to run the tests in a new RSpec process.
In the rerun, we configure a `TurboTests::Flaky::FlakyDetectorFormatter` with RSpec which removes all failed examples from the log file on disk since those examples are not flaky tests. Note that if there are too many failed examples on the first run, we will deem the failures to likely not be due to flaky tests and not re-run the test failures. As of writing, the threshold of failed examples is set to 10. If there are more than 10 failed examples, we will not re-run the failures.
2023-12-13 07:18:27 +08:00
|
|
|
run: |
|
|
|
|
if [ -f tmp/turbo_rspec_flaky_tests.json ]; then
|
|
|
|
echo "exists=true" >> $GITHUB_OUTPUT
|
|
|
|
else
|
|
|
|
echo "exists=false" >> $GITHUB_OUTPUT
|
|
|
|
fi
|
|
|
|
shell: bash
|
|
|
|
|
2023-12-18 15:59:41 +08:00
|
|
|
- name: Fetch Job ID
|
|
|
|
id: fetch-job-id
|
|
|
|
if: always() && steps.check-flaky-spec-report.outputs.exists == 'true'
|
|
|
|
run: |
|
2024-02-26 18:53:32 +08:00
|
|
|
job_id=$(ruby script/get_github_workflow_run_job_id.rb ${{ github.run_id }} ${{ github.run_attempt }} '${{ matrix.target }} ${{ matrix.build_type }}')
|
2023-12-18 15:59:41 +08:00
|
|
|
echo "job_id=$job_id" >> $GITHUB_OUTPUT
|
|
|
|
|
DEV: Introduce automatic reruns to RSpec tests on Github actions (#24811)
What motivated this change?
Our builds on Github actions have been extremely flaky mostly due to system tests. This has led to a drop in confidence
in our test suite where our developers tend to assume that a failed job is due to a flaky system test. As a result, we
have had occurrences where changes that resulted in legitimate test failures are merged into the `main` branch because developers
assumed it was a flaky test.
What does this change do?
This change seeks to reduce the flakiness of our builds on Github Actions by automatically re-running RSpec tests once when
they fail. If a failed test passes subsequently in the re-run, we mark the test as flaky by logging it into a file on disk
which is then uploaded as an artifact of the Github workflow run. We understand that automatically re-runs will lead to
lower accuracy of our tests but we accept this as an acceptable trade-off since a fragile build has a much greater impact
on our developers' time. Internally, the Discourse development team will be running a service to fetch the flaky tests
which have been logged for internal monitoring.
How is the change implemented?
1. A `--retry-and-log-flaky-tests` CLI flag is added to the `bin/turbo_rspec` CLI which will then initialize `TurboTests::Runner`
with the `retry_and_log_flaky_tests` kwarg set to `true`.
2. When the `retry_and_log_flaky_tests` kwarg is set to `true` for `TurboTests::Runner`, we will register an additional
formatter `Flaky::FailuresLoggerFormatter` to the `TurboTests::Reporter` in the `TurboTests::Runner#run` method.
The `Flaky::FailuresLoggerFormatter` has a simple job of logging all failed examples to a file on disk when running all the
tests. The details of the failed example which are logged can be found in `TurboTests::Flaky::FailedExample.to_h`.
3. Once all the tests have been run once, we check the result for any failed examples and if there are, we read the file on
disk to fetch the `location_rerun_location` of the failed examples which is then used to run the tests in a new RSpec process.
In the rerun, we configure a `TurboTests::Flaky::FlakyDetectorFormatter` with RSpec which removes all failed examples from the log file on disk since those examples are not flaky tests. Note that if there are too many failed examples on the first run, we will deem the failures to likely not be due to flaky tests and not re-run the test failures. As of writing, the threshold of failed examples is set to 10. If there are more than 10 failed examples, we will not re-run the failures.
2023-12-13 07:18:27 +08:00
|
|
|
- name: Create flaky tests report artifact
|
|
|
|
if: always() && steps.check-flaky-spec-report.outputs.exists == 'true'
|
2023-12-18 15:59:41 +08:00
|
|
|
run: cp tmp/turbo_rspec_flaky_tests.json tmp/turbo_rspec_flaky_tests-${{ matrix.build_type }}-${{ matrix.target }}-${{ steps.fetch-job-id.outputs.job_id }}.json
|
DEV: Introduce automatic reruns to RSpec tests on Github actions (#24811)
What motivated this change?
Our builds on Github actions have been extremely flaky mostly due to system tests. This has led to a drop in confidence
in our test suite where our developers tend to assume that a failed job is due to a flaky system test. As a result, we
have had occurrences where changes that resulted in legitimate test failures are merged into the `main` branch because developers
assumed it was a flaky test.
What does this change do?
This change seeks to reduce the flakiness of our builds on Github Actions by automatically re-running RSpec tests once when
they fail. If a failed test passes subsequently in the re-run, we mark the test as flaky by logging it into a file on disk
which is then uploaded as an artifact of the Github workflow run. We understand that automatically re-runs will lead to
lower accuracy of our tests but we accept this as an acceptable trade-off since a fragile build has a much greater impact
on our developers' time. Internally, the Discourse development team will be running a service to fetch the flaky tests
which have been logged for internal monitoring.
How is the change implemented?
1. A `--retry-and-log-flaky-tests` CLI flag is added to the `bin/turbo_rspec` CLI which will then initialize `TurboTests::Runner`
with the `retry_and_log_flaky_tests` kwarg set to `true`.
2. When the `retry_and_log_flaky_tests` kwarg is set to `true` for `TurboTests::Runner`, we will register an additional
formatter `Flaky::FailuresLoggerFormatter` to the `TurboTests::Reporter` in the `TurboTests::Runner#run` method.
The `Flaky::FailuresLoggerFormatter` has a simple job of logging all failed examples to a file on disk when running all the
tests. The details of the failed example which are logged can be found in `TurboTests::Flaky::FailedExample.to_h`.
3. Once all the tests have been run once, we check the result for any failed examples and if there are, we read the file on
disk to fetch the `location_rerun_location` of the failed examples which is then used to run the tests in a new RSpec process.
In the rerun, we configure a `TurboTests::Flaky::FlakyDetectorFormatter` with RSpec which removes all failed examples from the log file on disk since those examples are not flaky tests. Note that if there are too many failed examples on the first run, we will deem the failures to likely not be due to flaky tests and not re-run the test failures. As of writing, the threshold of failed examples is set to 10. If there are more than 10 failed examples, we will not re-run the failures.
2023-12-13 07:18:27 +08:00
|
|
|
|
|
|
|
- name: Upload flaky tests report
|
2024-11-11 19:37:53 +08:00
|
|
|
uses: actions/upload-artifact@v4
|
DEV: Introduce automatic reruns to RSpec tests on Github actions (#24811)
What motivated this change?
Our builds on Github actions have been extremely flaky mostly due to system tests. This has led to a drop in confidence
in our test suite where our developers tend to assume that a failed job is due to a flaky system test. As a result, we
have had occurrences where changes that resulted in legitimate test failures are merged into the `main` branch because developers
assumed it was a flaky test.
What does this change do?
This change seeks to reduce the flakiness of our builds on Github Actions by automatically re-running RSpec tests once when
they fail. If a failed test passes subsequently in the re-run, we mark the test as flaky by logging it into a file on disk
which is then uploaded as an artifact of the Github workflow run. We understand that automatically re-runs will lead to
lower accuracy of our tests but we accept this as an acceptable trade-off since a fragile build has a much greater impact
on our developers' time. Internally, the Discourse development team will be running a service to fetch the flaky tests
which have been logged for internal monitoring.
How is the change implemented?
1. A `--retry-and-log-flaky-tests` CLI flag is added to the `bin/turbo_rspec` CLI which will then initialize `TurboTests::Runner`
with the `retry_and_log_flaky_tests` kwarg set to `true`.
2. When the `retry_and_log_flaky_tests` kwarg is set to `true` for `TurboTests::Runner`, we will register an additional
formatter `Flaky::FailuresLoggerFormatter` to the `TurboTests::Reporter` in the `TurboTests::Runner#run` method.
The `Flaky::FailuresLoggerFormatter` has a simple job of logging all failed examples to a file on disk when running all the
tests. The details of the failed example which are logged can be found in `TurboTests::Flaky::FailedExample.to_h`.
3. Once all the tests have been run once, we check the result for any failed examples and if there are, we read the file on
disk to fetch the `location_rerun_location` of the failed examples which is then used to run the tests in a new RSpec process.
In the rerun, we configure a `TurboTests::Flaky::FlakyDetectorFormatter` with RSpec which removes all failed examples from the log file on disk since those examples are not flaky tests. Note that if there are too many failed examples on the first run, we will deem the failures to likely not be due to flaky tests and not re-run the test failures. As of writing, the threshold of failed examples is set to 10. If there are more than 10 failed examples, we will not re-run the failures.
2023-12-13 07:18:27 +08:00
|
|
|
if: always() && steps.check-flaky-spec-report.outputs.exists == 'true'
|
|
|
|
with:
|
2024-11-11 19:37:53 +08:00
|
|
|
name: flaky-test-reports-${{ matrix.build_type }}-${{ matrix.target }}
|
2023-12-18 15:59:41 +08:00
|
|
|
path: tmp/turbo_rspec_flaky_tests-${{ matrix.build_type }}-${{ matrix.target }}-${{ steps.fetch-job-id.outputs.job_id }}.json
|
DEV: Introduce automatic reruns to RSpec tests on Github actions (#24811)
What motivated this change?
Our builds on Github actions have been extremely flaky mostly due to system tests. This has led to a drop in confidence
in our test suite where our developers tend to assume that a failed job is due to a flaky system test. As a result, we
have had occurrences where changes that resulted in legitimate test failures are merged into the `main` branch because developers
assumed it was a flaky test.
What does this change do?
This change seeks to reduce the flakiness of our builds on Github Actions by automatically re-running RSpec tests once when
they fail. If a failed test passes subsequently in the re-run, we mark the test as flaky by logging it into a file on disk
which is then uploaded as an artifact of the Github workflow run. We understand that automatically re-runs will lead to
lower accuracy of our tests but we accept this as an acceptable trade-off since a fragile build has a much greater impact
on our developers' time. Internally, the Discourse development team will be running a service to fetch the flaky tests
which have been logged for internal monitoring.
How is the change implemented?
1. A `--retry-and-log-flaky-tests` CLI flag is added to the `bin/turbo_rspec` CLI which will then initialize `TurboTests::Runner`
with the `retry_and_log_flaky_tests` kwarg set to `true`.
2. When the `retry_and_log_flaky_tests` kwarg is set to `true` for `TurboTests::Runner`, we will register an additional
formatter `Flaky::FailuresLoggerFormatter` to the `TurboTests::Reporter` in the `TurboTests::Runner#run` method.
The `Flaky::FailuresLoggerFormatter` has a simple job of logging all failed examples to a file on disk when running all the
tests. The details of the failed example which are logged can be found in `TurboTests::Flaky::FailedExample.to_h`.
3. Once all the tests have been run once, we check the result for any failed examples and if there are, we read the file on
disk to fetch the `location_rerun_location` of the failed examples which is then used to run the tests in a new RSpec process.
In the rerun, we configure a `TurboTests::Flaky::FlakyDetectorFormatter` with RSpec which removes all failed examples from the log file on disk since those examples are not flaky tests. Note that if there are too many failed examples on the first run, we will deem the failures to likely not be due to flaky tests and not re-run the test failures. As of writing, the threshold of failed examples is set to 10. If there are more than 10 failed examples, we will not re-run the failures.
2023-12-13 07:18:27 +08:00
|
|
|
|
2021-07-06 16:47:16 +08:00
|
|
|
- name: Check Annotations
|
|
|
|
if: matrix.build_type == 'annotations'
|
|
|
|
run: |
|
|
|
|
bin/rake annotate:ensure_all_indexes
|
|
|
|
bin/annotate --models --model-dir app/models
|
|
|
|
|
|
|
|
if [ ! -z "$(git status --porcelain app/models/)" ]; then
|
|
|
|
echo "Core annotations are not up to date. To resolve, run:"
|
|
|
|
echo " bin/rake annotate:clean"
|
|
|
|
echo
|
|
|
|
echo "Or manually apply the diff printed below:"
|
|
|
|
echo "---------------------------------------------"
|
|
|
|
git -c color.ui=always diff app/models/
|
|
|
|
exit 1
|
|
|
|
fi
|
|
|
|
timeout-minutes: 30
|
2022-01-19 18:41:52 +08:00
|
|
|
|
|
|
|
core_frontend_tests:
|
2023-05-12 20:00:04 +08:00
|
|
|
if: github.event_name == 'pull_request' || github.repository != 'discourse/discourse-private-mirror'
|
2024-02-26 18:53:32 +08:00
|
|
|
name: core frontend (${{ matrix.browser }})
|
2024-09-30 11:18:20 +08:00
|
|
|
runs-on: ${{ (github.repository != 'discourse/discourse' && 'ubuntu-22.04-8core') || 'debian-12' }}
|
2022-09-22 01:13:13 +08:00
|
|
|
container:
|
|
|
|
image: discourse/discourse_test:slim-browsers
|
|
|
|
options: --user discourse
|
|
|
|
|
2022-08-14 23:30:15 +08:00
|
|
|
timeout-minutes: 35
|
2022-01-19 18:41:52 +08:00
|
|
|
|
|
|
|
strategy:
|
|
|
|
fail-fast: false
|
|
|
|
matrix:
|
2024-08-20 13:12:33 +08:00
|
|
|
browser: ["Chrome", "Firefox ESR", "Firefox Evergreen"]
|
2022-09-21 22:34:26 +08:00
|
|
|
|
|
|
|
env:
|
|
|
|
TESTEM_BROWSER: ${{ (startsWith(matrix.browser, 'Firefox') && 'Firefox') || matrix.browser }}
|
|
|
|
TESTEM_FIREFOX_PATH: ${{ (matrix.browser == 'Firefox Evergreen') && '/opt/firefox-evergreen/firefox' }}
|
2024-05-08 17:28:07 +08:00
|
|
|
CHEAP_SOURCE_MAPS: "1"
|
2022-01-19 18:41:52 +08:00
|
|
|
|
|
|
|
steps:
|
2023-09-11 17:42:02 +08:00
|
|
|
- uses: actions/checkout@v4
|
2022-01-19 18:41:52 +08:00
|
|
|
with:
|
|
|
|
fetch-depth: 1
|
|
|
|
|
|
|
|
- name: Setup Git
|
|
|
|
run: |
|
|
|
|
git config --global user.email "ci@ci.invalid"
|
|
|
|
git config --global user.name "Discourse CI"
|
|
|
|
|
2024-09-03 17:51:07 +08:00
|
|
|
- name: pnpm install
|
|
|
|
run: pnpm install --frozen-lockfile
|
2022-01-19 18:41:52 +08:00
|
|
|
|
|
|
|
- name: Ember Build
|
|
|
|
working-directory: ./app/assets/javascripts/discourse
|
|
|
|
run: |
|
2022-09-22 01:13:13 +08:00
|
|
|
mkdir /tmp/emberbuild
|
2024-09-03 17:51:07 +08:00
|
|
|
pnpm ember build --environment=test -o /tmp/emberbuild
|
2022-01-19 18:41:52 +08:00
|
|
|
|
2022-09-22 01:13:13 +08:00
|
|
|
- name: Core QUnit
|
2022-01-19 18:41:52 +08:00
|
|
|
working-directory: ./app/assets/javascripts/discourse
|
2024-08-26 08:42:35 +08:00
|
|
|
run: |
|
2024-09-17 08:01:40 +08:00
|
|
|
pnpm ember exam --path /tmp/emberbuild --load-balance --parallel=$(($(nproc) / 2)) --launch "${{ env.TESTEM_BROWSER }}" --write-execution-file --random
|
2022-08-14 23:30:15 +08:00
|
|
|
timeout-minutes: 15
|
2022-01-19 18:41:52 +08:00
|
|
|
|
2023-12-18 18:47:22 +08:00
|
|
|
- uses: actions/upload-artifact@v4
|
2022-09-22 01:13:13 +08:00
|
|
|
if: ${{ always() }}
|
|
|
|
with:
|
2023-12-18 18:47:22 +08:00
|
|
|
name: ember-exam-execution-${{ matrix.browser }}-${{ hashFiles('./app/assets/javascripts/discourse/test-execution-*.json') }}
|
2022-09-22 01:13:13 +08:00
|
|
|
path: ./app/assets/javascripts/discourse/test-execution-*.json
|
2024-11-11 19:37:53 +08:00
|
|
|
|
|
|
|
merge:
|
|
|
|
if: github.repository == 'discourse/discourse' && github.ref == 'refs/heads/main'
|
|
|
|
runs-on: debian-12
|
|
|
|
needs: build
|
|
|
|
steps:
|
|
|
|
- name: Merge Artifacts
|
|
|
|
uses: actions/upload-artifact/merge@v4
|
|
|
|
with:
|
|
|
|
name: failed-system-test-screenshots
|
|
|
|
pattern: failed-system-test-screenshots-*
|
|
|
|
delete-merged: true
|
|
|
|
separate-directories: false
|
|
|
|
# Don't fail the job if there aren't any artifacts to merge.
|
|
|
|
continue-on-error: true
|
|
|
|
|
|
|
|
- name: Merge Artifacts
|
|
|
|
uses: actions/upload-artifact/merge@v4
|
|
|
|
with:
|
|
|
|
name: flaky-test-reports
|
|
|
|
pattern: flaky-test-reports-*
|
|
|
|
delete-merged: true
|
|
|
|
separate-directories: false
|
|
|
|
# Don't fail the job if there aren't any artifacts to merge.
|
|
|
|
continue-on-error: true
|