-- -- PG17 -- SHOW server_version \gset SELECT substring(:'server_version', '\d+')::int >= 17 AS server_version_ge_17 \gset -- PG17 has the capabilty to pull up a correlated ANY subquery to a join if -- the subquery only refers to its immediate parent query. Previously, the -- subquery needed to be implemented as a SubPlan node, typically as a -- filter on a scan or join node. This PG17 capability enables Citus to -- run queries with correlated subqueries in certain cases, as shown here. -- Relevant PG commit: -- https://git.postgresql.org/gitweb/?p=postgresql.git;a=commitdiff;h=9f1337639 -- This feature is tested for all PG versions, not just PG17; each test query with -- a correlated subquery should fail with PG version < 17.0, but the test query -- rewritten to reflect how PG17 optimizes it should succeed with PG < 17.0 CREATE SCHEMA pg17_corr_subq_folding; SET search_path TO pg17_corr_subq_folding; SET citus.next_shard_id TO 20240017; SET citus.shard_count TO 2; SET citus.shard_replication_factor TO 1; CREATE TABLE test (x int, y int); SELECT create_distributed_table('test', 'x'); create_distributed_table --------------------------------------------------------------------- (1 row) INSERT INTO test VALUES (1,1), (2,2); -- Query 1: WHERE clause has a correlated subquery with a UNION. PG17 can plan -- this as a nested loop join with the subquery as the inner. The correlation -- is on the distribution column so the join can be pushed down by Citus. explain (costs off) SELECT * FROM test a WHERE x IN (SELECT x FROM test b UNION SELECT y FROM test c WHERE a.x = c.x) ORDER BY 1,2; QUERY PLAN --------------------------------------------------------------------- Sort Sort Key: remote_scan.x, remote_scan.y -> Custom Scan (Citus Adaptive) Task Count: 2 Tasks Shown: One of 2 -> Task Node: host=localhost port=xxxxx dbname=regression -> Nested Loop -> Seq Scan on test_20240017 a -> Subquery Scan on "ANY_subquery" Filter: (a.x = "ANY_subquery".x) -> HashAggregate Group Key: b.x -> Append -> Seq Scan on test_20240017 b -> Seq Scan on test_20240017 c Filter: (a.x = x) (17 rows) SET client_min_messages TO DEBUG2; SELECT * FROM test a WHERE x IN (SELECT x FROM test b UNION SELECT y FROM test c WHERE a.x = c.x) ORDER BY 1,2; DEBUG: Router planner cannot handle multi-shard select queries x | y --------------------------------------------------------------------- 1 | 1 2 | 2 (2 rows) RESET client_min_messages; -- Query 1 rewritten with subquery pulled up to a join, as done by PG17 planner; -- this query can be run without issues by Citus with older (pre PG17) PGs. explain (costs off) SELECT a.* FROM test a JOIN LATERAL (SELECT x FROM test b UNION SELECT y FROM test c WHERE a.x = c.x) dt1 ON a.x = dt1.x ORDER BY 1,2; QUERY PLAN --------------------------------------------------------------------- Sort Sort Key: remote_scan.x, remote_scan.y -> Custom Scan (Citus Adaptive) Task Count: 2 Tasks Shown: One of 2 -> Task Node: host=localhost port=xxxxx dbname=regression -> Nested Loop -> Seq Scan on test_20240017 a -> Subquery Scan on dt1 Filter: (a.x = dt1.x) -> HashAggregate Group Key: b.x -> Append -> Seq Scan on test_20240017 b -> Seq Scan on test_20240017 c Filter: (a.x = x) (17 rows) SET client_min_messages TO DEBUG2; SELECT a.* FROM test a JOIN LATERAL (SELECT x FROM test b UNION SELECT y FROM test c WHERE a.x = c.x) dt1 ON a.x = dt1.x ORDER BY 1,2; DEBUG: Router planner cannot handle multi-shard select queries x | y --------------------------------------------------------------------- 1 | 1 2 | 2 (2 rows) RESET client_min_messages; CREATE TABLE users (user_id int, time int, dept int, info bigint); CREATE TABLE events (user_id int, time int, event_type int, payload text); select create_distributed_table('users', 'user_id'); create_distributed_table --------------------------------------------------------------------- (1 row) select create_distributed_table('events', 'user_id'); create_distributed_table --------------------------------------------------------------------- (1 row) insert into users select i, 2021 + (i % 3), i % 5, 99999 * i from generate_series(1, 10) i; insert into events select i % 10 + 1, 2021 + (i % 3), i %11, md5((i*i)::text) from generate_series(1, 100) i; -- Query 2. In Citus correlated subqueries can not be used in the WHERE -- clause but if the subquery can be pulled up to a join it becomes possible -- for Citus to run the query, per this example. Pre PG17 the suqbuery -- was implemented as a SubPlan filter on the events table scan. EXPLAIN (costs off) WITH event_id AS(SELECT user_id AS events_user_id, time AS events_time, event_type FROM events) SELECT Count(*) FROM event_id WHERE (events_user_id) IN (SELECT user_id FROM users WHERE users.time = events_time); QUERY PLAN --------------------------------------------------------------------- Aggregate -> Custom Scan (Citus Adaptive) Task Count: 2 Tasks Shown: One of 2 -> Task Node: host=localhost port=xxxxx dbname=regression -> Aggregate -> Hash Join Hash Cond: ((events."time" = users."time") AND (events.user_id = users.user_id)) -> Seq Scan on events_20240021 events -> Hash -> HashAggregate Group Key: users."time", users.user_id -> Seq Scan on users_20240019 users (14 rows) SET client_min_messages TO DEBUG2; WITH event_id AS(SELECT user_id AS events_user_id, time AS events_time, event_type FROM events) SELECT Count(*) FROM event_id WHERE (events_user_id) IN (SELECT user_id FROM users WHERE users.time = events_time); DEBUG: CTE event_id is going to be inlined via distributed planning DEBUG: Router planner cannot handle multi-shard select queries count --------------------------------------------------------------------- 31 (1 row) RESET client_min_messages; -- Query 2 rewritten with subquery pulled up to a join, as done by pg17 planner. Citus -- Citus is able to run this query with previous pg versions. Note that the CTE can be -- disregarded because it is inlined, being only referenced once. EXPLAIN (COSTS OFF) SELECT Count(*) FROM (SELECT user_id AS events_user_id, time AS events_time, event_type FROM events) dt1 INNER JOIN (SELECT distinct user_id, time FROM users) dt ON events_user_id = dt.user_id and events_time = dt.time; QUERY PLAN --------------------------------------------------------------------- Aggregate -> Custom Scan (Citus Adaptive) Task Count: 2 Tasks Shown: One of 2 -> Task Node: host=localhost port=xxxxx dbname=regression -> Aggregate -> Hash Join Hash Cond: ((events.user_id = users.user_id) AND (events."time" = users."time")) -> Seq Scan on events_20240021 events -> Hash -> HashAggregate Group Key: users.user_id, users."time" -> Seq Scan on users_20240019 users (14 rows) SET client_min_messages TO DEBUG2; SELECT Count(*) FROM (SELECT user_id AS events_user_id, time AS events_time, event_type FROM events) dt1 INNER JOIN (SELECT distinct user_id, time FROM users) dt ON events_user_id = dt.user_id and events_time = dt.time; DEBUG: Router planner cannot handle multi-shard select queries count --------------------------------------------------------------------- 31 (1 row) RESET client_min_messages; -- Query 3: another example where recursive planning was prevented due to -- correlated subqueries, but with PG17 folding the subquery to a join it is -- possible for Citus to plan and run the query. EXPLAIN (costs off) SELECT dept, sum(user_id) FROM (SELECT users.dept, users.user_id FROM users, events as d1 WHERE d1.user_id = users.user_id AND users.dept IN (3,4) AND users.user_id IN (SELECT s2.user_id FROM users as s2 GROUP BY d1.user_id, s2.user_id)) dt GROUP BY dept; QUERY PLAN --------------------------------------------------------------------- HashAggregate Group Key: remote_scan.dept -> Custom Scan (Citus Adaptive) Task Count: 2 Tasks Shown: One of 2 -> Task Node: host=localhost port=xxxxx dbname=regression -> GroupAggregate Group Key: users.dept -> Sort Sort Key: users.dept -> Nested Loop Semi Join -> Hash Join Hash Cond: (d1.user_id = users.user_id) -> Seq Scan on events_20240021 d1 -> Hash -> Seq Scan on users_20240019 users Filter: (dept = ANY ('{3,4}'::integer[])) -> Subquery Scan on "ANY_subquery" Filter: (d1.user_id = "ANY_subquery".user_id) -> HashAggregate Group Key: s2.user_id -> Seq Scan on users_20240019 s2 (23 rows) SET client_min_messages TO DEBUG2; SELECT dept, sum(user_id) FROM (SELECT users.dept, users.user_id FROM users, events as d1 WHERE d1.user_id = users.user_id AND users.dept IN (3,4) AND users.user_id IN (SELECT s2.user_id FROM users as s2 GROUP BY d1.user_id, s2.user_id)) dt GROUP BY dept; DEBUG: Router planner cannot handle multi-shard select queries dept | sum --------------------------------------------------------------------- 3 | 110 4 | 130 (2 rows) RESET client_min_messages; -- Query 3 rewritten in a similar way to how the PG17 pulls up the subquery; -- the join is on the distribution key so Citus can push down. EXPLAIN (costs off) SELECT dept, sum(user_id) FROM (SELECT users.dept, users.user_id FROM users, events as d1 JOIN LATERAL (SELECT s2.user_id FROM users as s2 GROUP BY s2.user_id HAVING d1.user_id IS NOT NULL) as d2 ON 1=1 WHERE d1.user_id = users.user_id AND users.dept IN (3,4) AND users.user_id = d2.user_id) dt GROUP BY dept; QUERY PLAN --------------------------------------------------------------------- HashAggregate Group Key: remote_scan.dept -> Custom Scan (Citus Adaptive) Task Count: 2 Tasks Shown: One of 2 -> Task Node: host=localhost port=xxxxx dbname=regression -> GroupAggregate Group Key: users.dept -> Sort Sort Key: users.dept -> Nested Loop -> Hash Join Hash Cond: (d1.user_id = users.user_id) -> Seq Scan on events_20240021 d1 -> Hash -> Seq Scan on users_20240019 users Filter: (dept = ANY ('{3,4}'::integer[])) -> Subquery Scan on d2 Filter: (d1.user_id = d2.user_id) -> HashAggregate Group Key: s2.user_id -> Result One-Time Filter: (d1.user_id IS NOT NULL) -> Seq Scan on users_20240019 s2 (25 rows) SET client_min_messages TO DEBUG2; SELECT dept, sum(user_id) FROM (SELECT users.dept, users.user_id FROM users, events as d1 JOIN LATERAL (SELECT s2.user_id FROM users as s2 GROUP BY s2.user_id HAVING d1.user_id IS NOT NULL) as d2 ON 1=1 WHERE d1.user_id = users.user_id AND users.dept IN (3,4) AND users.user_id = d2.user_id) dt GROUP BY dept; DEBUG: Router planner cannot handle multi-shard select queries dept | sum --------------------------------------------------------------------- 3 | 110 4 | 130 (2 rows) RESET client_min_messages; RESET search_path; DROP SCHEMA pg17_corr_subq_folding CASCADE; NOTICE: drop cascades to 3 other objects DETAIL: drop cascades to table pg17_corr_subq_folding.test drop cascades to table pg17_corr_subq_folding.users drop cascades to table pg17_corr_subq_folding.events \if :server_version_ge_17 \else \q \endif -- PG17-specific tests go here. -- CREATE SCHEMA pg17; SET search_path to pg17; -- Test specifying access method on partitioned tables. PG17 feature, added by: -- https://git.postgresql.org/gitweb/?p=postgresql.git;a=commitdiff;h=374c7a229 -- The following tests were failing tests in tableam but will pass on PG >= 17. -- There is some set-up duplication of tableam, and this test can be returned -- to tableam when 17 is the minimum supported PG version. SELECT public.run_command_on_coordinator_and_workers($Q$ SET citus.enable_ddl_propagation TO off; CREATE FUNCTION fake_am_handler(internal) RETURNS table_am_handler AS 'citus' LANGUAGE C; CREATE ACCESS METHOD fake_am TYPE TABLE HANDLER fake_am_handler; $Q$); run_command_on_coordinator_and_workers --------------------------------------------------------------------- (1 row) -- Since Citus assumes access methods are part of the extension, make fake_am -- owned manually to be able to pass checks on Citus while distributing tables. ALTER EXTENSION citus ADD ACCESS METHOD fake_am; CREATE TABLE test_partitioned(id int, p int, val int) PARTITION BY RANGE (p) USING fake_am; -- Test that children inherit access method from parent CREATE TABLE test_partitioned_p1 PARTITION OF test_partitioned FOR VALUES FROM (1) TO (10); CREATE TABLE test_partitioned_p2 PARTITION OF test_partitioned FOR VALUES FROM (11) TO (20); INSERT INTO test_partitioned VALUES (1, 5, -1), (2, 15, -2); WARNING: fake_tuple_insert WARNING: fake_tuple_insert INSERT INTO test_partitioned VALUES (3, 6, -6), (4, 16, -4); WARNING: fake_tuple_insert WARNING: fake_tuple_insert SELECT count(1) FROM test_partitioned_p1; WARNING: fake_scan_getnextslot WARNING: fake_scan_getnextslot WARNING: fake_scan_getnextslot count --------------------------------------------------------------------- 2 (1 row) SELECT count(1) FROM test_partitioned_p2; WARNING: fake_scan_getnextslot WARNING: fake_scan_getnextslot WARNING: fake_scan_getnextslot count --------------------------------------------------------------------- 2 (1 row) -- Both child table partitions inherit fake_am SELECT c.relname, am.amname FROM pg_class c, pg_am am WHERE c.relam = am.oid AND c.oid IN ('test_partitioned_p1'::regclass, 'test_partitioned_p2'::regclass) ORDER BY c.relname; relname | amname --------------------------------------------------------------------- test_partitioned_p1 | fake_am test_partitioned_p2 | fake_am (2 rows) -- Clean up DROP TABLE test_partitioned; ALTER EXTENSION citus DROP ACCESS METHOD fake_am; SELECT public.run_command_on_coordinator_and_workers($Q$ RESET citus.enable_ddl_propagation; $Q$); run_command_on_coordinator_and_workers --------------------------------------------------------------------- (1 row) -- End of testing specifying access method on partitioned tables. -- MAINTAIN privilege tests CREATE ROLE regress_maintain; CREATE ROLE regress_no_maintain; ALTER ROLE regress_maintain WITH login; GRANT USAGE ON SCHEMA pg17 TO regress_maintain; ALTER ROLE regress_no_maintain WITH login; GRANT USAGE ON SCHEMA pg17 TO regress_no_maintain; SET citus.shard_count TO 1; -- For consistent remote command logging CREATE TABLE dist_test(a int, b int); SELECT create_distributed_table('dist_test', 'a'); create_distributed_table --------------------------------------------------------------------- (1 row) INSERT INTO dist_test SELECT i % 10, i FROM generate_series(1, 100) t(i); SET citus.log_remote_commands TO on; SET citus.grep_remote_commands = '%maintain%'; GRANT MAINTAIN ON dist_test TO regress_maintain; NOTICE: issuing GRANT maintain ON dist_test TO regress_maintain DETAIL: on server postgres@localhost:xxxxx connectionId: xxxxxxx NOTICE: issuing GRANT maintain ON dist_test TO regress_maintain DETAIL: on server postgres@localhost:xxxxx connectionId: xxxxxxx NOTICE: issuing SELECT worker_apply_shard_ddl_command (20240023, 'pg17', 'GRANT maintain ON dist_test TO regress_maintain') DETAIL: on server postgres@localhost:xxxxx connectionId: xxxxxxx RESET citus.grep_remote_commands; SET ROLE regress_no_maintain; -- Current role does not have MAINTAIN privileges on dist_test ANALYZE dist_test; WARNING: permission denied to analyze "dist_test", skipping it NOTICE: issuing ANALYZE pg17.dist_test_20240023 DETAIL: on server regress_no_maintain@localhost:xxxxx connectionId: xxxxxxx VACUUM dist_test; WARNING: permission denied to vacuum "dist_test", skipping it NOTICE: issuing VACUUM pg17.dist_test_20240023 DETAIL: on server regress_no_maintain@localhost:xxxxx connectionId: xxxxxxx SET ROLE regress_maintain; -- Current role has MAINTAIN privileges on dist_test ANALYZE dist_test; NOTICE: issuing ANALYZE pg17.dist_test_20240023 DETAIL: on server regress_maintain@localhost:xxxxx connectionId: xxxxxxx VACUUM dist_test; NOTICE: issuing VACUUM pg17.dist_test_20240023 DETAIL: on server regress_maintain@localhost:xxxxx connectionId: xxxxxxx -- Take away regress_maintain's MAINTAIN privileges on dist_test RESET ROLE; SET citus.grep_remote_commands = '%maintain%'; REVOKE MAINTAIN ON dist_test FROM regress_maintain; NOTICE: issuing REVOKE maintain ON dist_test FROM regress_maintain DETAIL: on server postgres@localhost:xxxxx connectionId: xxxxxxx NOTICE: issuing REVOKE maintain ON dist_test FROM regress_maintain DETAIL: on server postgres@localhost:xxxxx connectionId: xxxxxxx NOTICE: issuing SELECT worker_apply_shard_ddl_command (20240023, 'pg17', 'REVOKE maintain ON dist_test FROM regress_maintain') DETAIL: on server postgres@localhost:xxxxx connectionId: xxxxxxx RESET citus.grep_remote_commands; SET ROLE regress_maintain; -- Current role does not have MAINTAIN privileges on dist_test ANALYZE dist_test; WARNING: permission denied to analyze "dist_test", skipping it NOTICE: issuing ANALYZE pg17.dist_test_20240023 DETAIL: on server regress_maintain@localhost:xxxxx connectionId: xxxxxxx VACUUM dist_test; WARNING: permission denied to vacuum "dist_test", skipping it NOTICE: issuing VACUUM pg17.dist_test_20240023 DETAIL: on server regress_maintain@localhost:xxxxx connectionId: xxxxxxx RESET ROLE; -- End of MAINTAIN privilege tests RESET citus.log_remote_commands; RESET citus.next_shard_id; RESET citus.shard_count; RESET citus.shard_replication_factor; DROP SCHEMA pg17 CASCADE; NOTICE: drop cascades to 3 other objects DETAIL: drop cascades to function fake_am_handler(internal) drop cascades to access method fake_am drop cascades to table dist_test DROP ROLE regress_maintain; DROP ROLE regress_no_maintain;