When distributing a columnar table, as well as changing options on a distributed columnar table, this patch will forward the settings from the coordinator to the workers. For propagating options changes on an already distributed table this change is pretty straight forward. Before applying the change in options locally we will create a `DDLJob` that contains a call to `alter_columnar_table_set(...)` for every shard placement with all settings of the current table. This goes both for setting an option as well as resetting. This will reset the values to the defaults configured on the coordinator. Having the effect that the coordinator is authoritative on the settings and makes sure the shards have the same settings set as the table on the coordinator. When a columnar table is distributed it is using the `TableDDLCommand` infra structure to create a new kind of `TableDDLCommand`. This new type, called a `TableDDLCommandFunction` contains a context and 2 function pointers to execute. One function returns the command as applied on the table, the second function will return the sql command to apply to a shard with a given shard id. The schema name is ignored as it will use the fully qualified name of the shard in the same schema as the base table. |
||
---|---|---|
.. | ||
bin | ||
data | ||
expected | ||
input | ||
mitmscripts | ||
output | ||
spec | ||
sql | ||
upgrade | ||
.gitignore | ||
Makefile | ||
Pipfile | ||
Pipfile.lock | ||
README.md | ||
after_citus_upgrade_coord_schedule | ||
after_pg_upgrade_schedule | ||
base_isolation_schedule | ||
base_schedule | ||
before_citus_upgrade_coord_schedule | ||
before_pg_upgrade_schedule | ||
columnar_am_isolation_schedule | ||
columnar_am_schedule | ||
failure_base_schedule | ||
failure_schedule | ||
isolation_schedule | ||
log_test_times | ||
minimal_schedule | ||
multi_follower_schedule | ||
multi_mx_schedule | ||
multi_schedule | ||
multi_schedule_hyperscale | ||
multi_schedule_hyperscale_superuser | ||
mx_base_schedule | ||
mx_minimal_schedule | ||
pg_regress_multi.pl | ||
worker_schedule |
README.md
How our testing works
We use the test tooling of postgres to run our tests. This tooling is very
simple but effective. The basics it runs a series of .sql
scripts, gets
their output and stores that in results/$sqlfilename.out
. It then compares the
actual output to the expected output with a simple diff
command:
diff results/$sqlfilename.out expected/$sqlfilename.out
Schedules
Which sql scripts to run is defined in a schedule file, e.g. multi_schedule
,
multi_mx_schedule
.
Makefile
In our Makefile
we have rules to run the different types of test schedules.
You can run them from the root of the repository like so:
# e.g. the multi_schedule
make install -j9 && make -C src/test/regress/ check-multi
Take a look at the makefile for a list of all the testing targets.
Running a specific test
Often you want to run a specific test and don't want to run everything. You can use one of the following commands to do so:
# If your tests needs almost no setup you can use check-minimal
make install -j9 && make -C src/test/regress/ check-minimal EXTRA_TESTS='multi_utility_warnings'
# Often tests need some testing data, if you get missing table errors using
# check-minimal you should try check-base
make install -j9 && make -C src/test/regress/ check-base EXTRA_TESTS='with_prepare'
# Sometimes this is still not enough and some other test needs to be run before
# the test you want to run. You can do so by adding it to EXTRA_TESTS too.
make install -j9 && make -C src/test/regress/ check-base EXTRA_TESTS='add_coordinator coordinator_shouldhaveshards'
Normalization
The output of tests is sadly not completely predictable. Still we want to
compare the output of different runs and error when the important things are
different. We do this by not using the regular system diff
to compare files.
Instead we use src/test/regress/bin/diff
which does the following things:
- Change the
$sqlfilename.out
file by running it throughsed
using thesrc/test/regress/bin/normalize.sed
file. This does stuff like replacing numbers that keep changing across runs with anXXX
string, e.g. portnumbers or transaction numbers. - Backup the original output to
$sqlfilename.out.unmodified
in case it's needed for debugging - Compare the changed
results
andexpected
files with the systemdiff
command.
Updating the expected test output
Sometimes you add a test to an existing file, or test output changes in a way
that's not bad (possibly even good if support for queries is added). In those
cases you want to update the expected test output.
The way to do this is very simple, you run the test and copy the new .out file
in the results
directory to the expected
directory, e.g.:
make install -j9 && make -C src/test/regress/ check-minimal EXTRA_TESTS='multi_utility_warnings'
cp src/test/regress/{results,expected}/multi_utility_warnings.out
Adding a new test file
Adding a new test file is quite simple:
- Write the SQL file in the
sql
directory - Add it to a schedule file, to make sure it's run in CI
- Run the test
- Check that the output is as expected
- Copy the
.out
file fromresults
toexpected
Isolation testing
See src/test/regress/spec/README.md
Upgrade testing
See src/test/regress/upgrade/README.md
Failure testing
See src/test/regress/mitmscripts/README.md
Perl test setup script
To automatically setup a citus cluster in tests we use our
src/test/regress/pg_regress_multi.pl
script. This sets up a citus cluster and
then starts the standard postgres test tooling. You almost never have to change
this file.